init.c 13.4 KB
Newer Older
1 2
/*
 * linux/arch/sh/mm/init.c
L
Linus Torvalds 已提交
3 4
 *
 *  Copyright (C) 1999  Niibe Yutaka
5
 *  Copyright (C) 2002 - 2011  Paul Mundt
L
Linus Torvalds 已提交
6 7 8 9 10 11 12
 *
 *  Based on linux/arch/i386/mm/init.c:
 *   Copyright (C) 1995  Linus Torvalds
 */
#include <linux/mm.h>
#include <linux/swap.h>
#include <linux/init.h>
13
#include <linux/gfp.h>
L
Linus Torvalds 已提交
14
#include <linux/bootmem.h>
P
Paul Mundt 已提交
15
#include <linux/proc_fs.h>
16
#include <linux/pagemap.h>
17 18
#include <linux/percpu.h>
#include <linux/io.h>
Y
Yinghai Lu 已提交
19
#include <linux/memblock.h>
20
#include <linux/dma-mapping.h>
21
#include <linux/export.h>
L
Linus Torvalds 已提交
22
#include <asm/mmu_context.h>
P
Paul Mundt 已提交
23
#include <asm/mmzone.h>
P
Paul Mundt 已提交
24
#include <asm/kexec.h>
L
Linus Torvalds 已提交
25 26
#include <asm/tlb.h>
#include <asm/cacheflush.h>
27
#include <asm/sections.h>
P
Paul Mundt 已提交
28
#include <asm/setup.h>
L
Linus Torvalds 已提交
29
#include <asm/cache.h>
30
#include <asm/sizes.h>
L
Linus Torvalds 已提交
31 32

pgd_t swapper_pg_dir[PTRS_PER_PGD];
33

34 35
void __init generic_mem_init(void)
{
Y
Yinghai Lu 已提交
36
	memblock_add(__MEMORY_START, __MEMORY_SIZE);
37 38
}

P
Paul Mundt 已提交
39 40 41 42 43
void __init __weak plat_mem_setup(void)
{
	/* Nothing to see here, move along. */
}

44
#ifdef CONFIG_MMU
45
static pte_t *__get_pte_phys(unsigned long addr)
L
Linus Torvalds 已提交
46 47
{
	pgd_t *pgd;
48
	pud_t *pud;
L
Linus Torvalds 已提交
49 50
	pmd_t *pmd;

S
Stuart Menefy 已提交
51
	pgd = pgd_offset_k(addr);
L
Linus Torvalds 已提交
52 53
	if (pgd_none(*pgd)) {
		pgd_ERROR(*pgd);
54
		return NULL;
L
Linus Torvalds 已提交
55 56
	}

S
Stuart Menefy 已提交
57 58 59
	pud = pud_alloc(NULL, pgd, addr);
	if (unlikely(!pud)) {
		pud_ERROR(*pud);
60
		return NULL;
61 62
	}

S
Stuart Menefy 已提交
63 64 65
	pmd = pmd_alloc(NULL, pud, addr);
	if (unlikely(!pmd)) {
		pmd_ERROR(*pmd);
66
		return NULL;
L
Linus Torvalds 已提交
67 68
	}

69
	return pte_offset_kernel(pmd, addr);
70 71 72 73 74 75 76
}

static void set_pte_phys(unsigned long addr, unsigned long phys, pgprot_t prot)
{
	pte_t *pte;

	pte = __get_pte_phys(addr);
L
Linus Torvalds 已提交
77 78 79 80 81 82
	if (!pte_none(*pte)) {
		pte_ERROR(*pte);
		return;
	}

	set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, prot));
83
	local_flush_tlb_one(get_asid(), addr);
84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99

	if (pgprot_val(prot) & _PAGE_WIRED)
		tlb_wire_entry(NULL, addr, *pte);
}

static void clear_pte_phys(unsigned long addr, pgprot_t prot)
{
	pte_t *pte;

	pte = __get_pte_phys(addr);

	if (pgprot_val(prot) & _PAGE_WIRED)
		tlb_unwire_entry();

	set_pte(pte, pfn_pte(0, __pgprot(0)));
	local_flush_tlb_one(get_asid(), addr);
L
Linus Torvalds 已提交
100 101 102 103 104 105 106 107 108 109 110 111 112
}

void __set_fixmap(enum fixed_addresses idx, unsigned long phys, pgprot_t prot)
{
	unsigned long address = __fix_to_virt(idx);

	if (idx >= __end_of_fixed_addresses) {
		BUG();
		return;
	}

	set_pte_phys(address, phys, prot);
}
113

114 115 116 117 118 119 120 121 122 123 124 125
void __clear_fixmap(enum fixed_addresses idx, pgprot_t prot)
{
	unsigned long address = __fix_to_virt(idx);

	if (idx >= __end_of_fixed_addresses) {
		BUG();
		return;
	}

	clear_pte_phys(address, prot);
}

126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157
static pmd_t * __init one_md_table_init(pud_t *pud)
{
	if (pud_none(*pud)) {
		pmd_t *pmd;

		pmd = alloc_bootmem_pages(PAGE_SIZE);
		pud_populate(&init_mm, pud, pmd);
		BUG_ON(pmd != pmd_offset(pud, 0));
	}

	return pmd_offset(pud, 0);
}

static pte_t * __init one_page_table_init(pmd_t *pmd)
{
	if (pmd_none(*pmd)) {
		pte_t *pte;

		pte = alloc_bootmem_pages(PAGE_SIZE);
		pmd_populate_kernel(&init_mm, pmd, pte);
		BUG_ON(pte != pte_offset_kernel(pmd, 0));
	}

	return pte_offset_kernel(pmd, 0);
}

static pte_t * __init page_table_kmap_check(pte_t *pte, pmd_t *pmd,
					    unsigned long vaddr, pte_t *lastpte)
{
	return pte;
}

158 159 160 161 162 163
void __init page_table_range_init(unsigned long start, unsigned long end,
					 pgd_t *pgd_base)
{
	pgd_t *pgd;
	pud_t *pud;
	pmd_t *pmd;
164
	pte_t *pte = NULL;
165
	int i, j, k;
166 167
	unsigned long vaddr;

168 169 170 171 172 173 174 175 176
	vaddr = start;
	i = __pgd_offset(vaddr);
	j = __pud_offset(vaddr);
	k = __pmd_offset(vaddr);
	pgd = pgd_base + i;

	for ( ; (i < PTRS_PER_PGD) && (vaddr != end); pgd++, i++) {
		pud = (pud_t *)pgd;
		for ( ; (j < PTRS_PER_PUD) && (vaddr != end); pud++, j++) {
177 178
			pmd = one_md_table_init(pud);
#ifndef __PAGETABLE_PMD_FOLDED
179 180
			pmd += k;
#endif
181
			for (; (k < PTRS_PER_PMD) && (vaddr != end); pmd++, k++) {
182 183
				pte = page_table_kmap_check(one_page_table_init(pmd),
							    pmd, vaddr, pte);
184 185 186
				vaddr += PMD_SIZE;
			}
			k = 0;
187
		}
188
		j = 0;
189 190
	}
}
191
#endif	/* CONFIG_MMU */
L
Linus Torvalds 已提交
192

P
Paul Mundt 已提交
193 194 195 196 197 198 199 200 201 202
void __init allocate_pgdat(unsigned int nid)
{
	unsigned long start_pfn, end_pfn;
#ifdef CONFIG_NEED_MULTIPLE_NODES
	unsigned long phys;
#endif

	get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);

#ifdef CONFIG_NEED_MULTIPLE_NODES
Y
Yinghai Lu 已提交
203
	phys = __memblock_alloc_base(sizeof(struct pglist_data),
P
Paul Mundt 已提交
204 205 206
				SMP_CACHE_BYTES, end_pfn << PAGE_SHIFT);
	/* Retry with all of system memory */
	if (!phys)
Y
Yinghai Lu 已提交
207 208
		phys = __memblock_alloc_base(sizeof(struct pglist_data),
					SMP_CACHE_BYTES, memblock_end_of_DRAM());
P
Paul Mundt 已提交
209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237
	if (!phys)
		panic("Can't allocate pgdat for node %d\n", nid);

	NODE_DATA(nid) = __va(phys);
	memset(NODE_DATA(nid), 0, sizeof(struct pglist_data));

	NODE_DATA(nid)->bdata = &bootmem_node_data[nid];
#endif

	NODE_DATA(nid)->node_start_pfn = start_pfn;
	NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn;
}

static void __init bootmem_init_one_node(unsigned int nid)
{
	unsigned long total_pages, paddr;
	unsigned long end_pfn;
	struct pglist_data *p;

	p = NODE_DATA(nid);

	/* Nothing to do.. */
	if (!p->node_spanned_pages)
		return;

	end_pfn = p->node_start_pfn + p->node_spanned_pages;

	total_pages = bootmem_bootmap_pages(p->node_spanned_pages);

Y
Yinghai Lu 已提交
238
	paddr = memblock_alloc(total_pages << PAGE_SHIFT, PAGE_SIZE);
P
Paul Mundt 已提交
239 240 241 242 243 244 245 246 247 248 249 250 251
	if (!paddr)
		panic("Can't allocate bootmap for nid[%d]\n", nid);

	init_bootmem_node(p, paddr >> PAGE_SHIFT, p->node_start_pfn, end_pfn);

	free_bootmem_with_active_regions(nid, end_pfn);

	/*
	 * XXX Handle initial reservations for the system memory node
	 * only for the moment, we'll refactor this later for handling
	 * reservations in other nodes.
	 */
	if (nid == 0) {
252 253
		struct memblock_region *reg;

P
Paul Mundt 已提交
254
		/* Reserve the sections we're already using. */
255 256 257
		for_each_memblock(reserved, reg) {
			reserve_bootmem(reg->base, reg->size, BOOTMEM_DEFAULT);
		}
P
Paul Mundt 已提交
258 259 260 261 262 263 264
	}

	sparse_memory_present_with_active_regions(nid);
}

static void __init do_init_bootmem(void)
{
265
	struct memblock_region *reg;
P
Paul Mundt 已提交
266 267 268
	int i;

	/* Add active regions with valid PFNs. */
269
	for_each_memblock(memory, reg) {
P
Paul Mundt 已提交
270
		unsigned long start_pfn, end_pfn;
271 272
		start_pfn = memblock_region_memory_base_pfn(reg);
		end_pfn = memblock_region_memory_end_pfn(reg);
P
Paul Mundt 已提交
273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290
		__add_active_range(0, start_pfn, end_pfn);
	}

	/* All of system RAM sits in node 0 for the non-NUMA case */
	allocate_pgdat(0);
	node_set_online(0);

	plat_mem_setup();

	for_each_online_node(i)
		bootmem_init_one_node(i);

	sparse_init();
}

static void __init early_reserve_mem(void)
{
	unsigned long start_pfn;
S
Simon Horman 已提交
291 292
	u32 zero_base = (u32)__MEMORY_START + (u32)PHYSICAL_OFFSET;
	u32 start = zero_base + (u32)CONFIG_ZERO_PAGE_OFFSET;
P
Paul Mundt 已提交
293 294 295 296 297 298 299 300 301 302 303 304 305

	/*
	 * Partially used pages are not usable - thus
	 * we are rounding upwards:
	 */
	start_pfn = PFN_UP(__pa(_end));

	/*
	 * Reserve the kernel text and Reserve the bootmem bitmap. We do
	 * this in two steps (first step was init_bootmem()), because
	 * this catches the (definitely buggy) case of us accidentally
	 * initializing the bootmem allocator with an invalid RAM area.
	 */
S
Simon Horman 已提交
306
	memblock_reserve(start, (PFN_PHYS(start_pfn) + PAGE_SIZE - 1) - start);
P
Paul Mundt 已提交
307 308 309 310 311

	/*
	 * Reserve physical pages below CONFIG_ZERO_PAGE_OFFSET.
	 */
	if (CONFIG_ZERO_PAGE_OFFSET != 0)
S
Simon Horman 已提交
312
		memblock_reserve(zero_base, CONFIG_ZERO_PAGE_OFFSET);
P
Paul Mundt 已提交
313 314 315 316 317 318 319 320

	/*
	 * Handle additional early reservations
	 */
	check_for_initrd();
	reserve_crashkernel();
}

L
Linus Torvalds 已提交
321 322
void __init paging_init(void)
{
323
	unsigned long max_zone_pfns[MAX_NR_ZONES];
324
	unsigned long vaddr, end;
325
	int nid;
L
Linus Torvalds 已提交
326

P
Paul Mundt 已提交
327 328 329 330
	sh_mv.mv_mem_init();

	early_reserve_mem();

331 332 333 334 335 336 337
	/*
	 * Once the early reservations are out of the way, give the
	 * platforms a chance to kick out some memory.
	 */
	if (sh_mv.mv_mem_reserve)
		sh_mv.mv_mem_reserve();

Y
Yinghai Lu 已提交
338
	memblock_enforce_memory_limit(memory_limit);
339
	memblock_allow_resize();
P
Paul Mundt 已提交
340

Y
Yinghai Lu 已提交
341
	memblock_dump_all();
P
Paul Mundt 已提交
342 343 344 345

	/*
	 * Determine low and high memory ranges:
	 */
Y
Yinghai Lu 已提交
346
	max_low_pfn = max_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT;
P
Paul Mundt 已提交
347 348 349 350 351
	min_low_pfn = __MEMORY_START >> PAGE_SHIFT;

	nodes_clear(node_online_map);

	memory_start = (unsigned long)__va(__MEMORY_START);
Y
Yinghai Lu 已提交
352
	memory_end = memory_start + (memory_limit ?: memblock_phys_mem_size());
P
Paul Mundt 已提交
353 354 355 356 357 358

	uncached_init();
	pmb_init();
	do_init_bootmem();
	ioremap_fixed_init();

359 360 361 362
	/* We don't need to map the kernel through the TLB, as
	 * it is permanatly mapped using P1. So clear the
	 * entire pgd. */
	memset(swapper_pg_dir, 0, sizeof(swapper_pg_dir));
L
Linus Torvalds 已提交
363

364 365 366 367
	/* Set an initial value for the MMU.TTB so we don't have to
	 * check for a null value. */
	set_TTB(swapper_pg_dir);

368 369
	/*
	 * Populate the relevant portions of swapper_pg_dir so that
370
	 * we can use the fixmap entries without calling kmalloc.
371 372 373
	 * pte's will be filled in by __set_fixmap().
	 */
	vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK;
374 375
	end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK;
	page_table_range_init(vaddr, end, swapper_pg_dir);
376 377

	kmap_coherent_init();
378

379 380
	memset(max_zone_pfns, 0, sizeof(max_zone_pfns));

381 382 383 384
	for_each_online_node(nid) {
		pg_data_t *pgdat = NODE_DATA(nid);
		unsigned long low, start_pfn;

385
		start_pfn = pgdat->bdata->node_min_pfn;
386 387
		low = pgdat->bdata->node_low_pfn;

388 389
		if (max_zone_pfns[ZONE_NORMAL] < low)
			max_zone_pfns[ZONE_NORMAL] = low;
390 391 392 393

		printk("Node %u: start_pfn = 0x%lx, low = 0x%lx\n",
		       nid, start_pfn, low);
	}
394 395

	free_area_init_nodes(max_zone_pfns);
L
Linus Torvalds 已提交
396 397
}

398 399 400 401 402 403 404 405
/*
 * Early initialization for any I/O MMUs we might have.
 */
static void __init iommu_init(void)
{
	no_iommu_init();
}

406 407
unsigned int mem_init_done = 0;

L
Linus Torvalds 已提交
408 409
void __init mem_init(void)
{
P
Paul Mundt 已提交
410
	int codesize, datasize, initsize;
411
	int nid;
L
Linus Torvalds 已提交
412

413 414
	iommu_init();

415 416 417
	num_physpages = 0;
	high_memory = NULL;

418 419 420 421 422 423 424 425 426 427 428
	for_each_online_node(nid) {
		pg_data_t *pgdat = NODE_DATA(nid);
		unsigned long node_pages = 0;
		void *node_high_memory;

		num_physpages += pgdat->node_present_pages;

		if (pgdat->node_spanned_pages)
			node_pages = free_all_bootmem_node(pgdat);

		totalram_pages += node_pages;
L
Linus Torvalds 已提交
429

430 431 432
		node_high_memory = (void *)__va((pgdat->node_start_pfn +
						 pgdat->node_spanned_pages) <<
						 PAGE_SHIFT);
433 434 435
		if (node_high_memory > high_memory)
			high_memory = node_high_memory;
	}
L
Linus Torvalds 已提交
436

437 438 439
	/* Set this up early, so we can take care of the zero page */
	cpu_cache_init();

L
Linus Torvalds 已提交
440 441 442 443
	/* clear the zero-page */
	memset(empty_zero_page, 0, PAGE_SIZE);
	__flush_wback_region(empty_zero_page, PAGE_SIZE);

444 445
	vsyscall_init();

L
Linus Torvalds 已提交
446 447 448 449
	codesize =  (unsigned long) &_etext - (unsigned long) &_text;
	datasize =  (unsigned long) &_edata - (unsigned long) &_etext;
	initsize =  (unsigned long) &__init_end - (unsigned long) &__init_begin;

P
Paul Mundt 已提交
450
	printk(KERN_INFO "Memory: %luk/%luk available (%dk kernel code, "
P
Paul Mundt 已提交
451
	       "%dk data, %dk init)\n",
452
		nr_free_pages() << (PAGE_SHIFT-10),
453
		num_physpages << (PAGE_SHIFT-10),
L
Linus Torvalds 已提交
454 455 456 457
		codesize >> 10,
		datasize >> 10,
		initsize >> 10);

458 459 460 461 462 463
	printk(KERN_INFO "virtual kernel memory layout:\n"
		"    fixmap  : 0x%08lx - 0x%08lx   (%4ld kB)\n"
#ifdef CONFIG_HIGHMEM
		"    pkmap   : 0x%08lx - 0x%08lx   (%4ld kB)\n"
#endif
		"    vmalloc : 0x%08lx - 0x%08lx   (%4ld MB)\n"
464
		"    lowmem  : 0x%08lx - 0x%08lx   (%4ld MB) (cached)\n"
465
#ifdef CONFIG_UNCACHED_MAPPING
466
		"            : 0x%08lx - 0x%08lx   (%4ld MB) (uncached)\n"
467
#endif
468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484
		"      .init : 0x%08lx - 0x%08lx   (%4ld kB)\n"
		"      .data : 0x%08lx - 0x%08lx   (%4ld kB)\n"
		"      .text : 0x%08lx - 0x%08lx   (%4ld kB)\n",
		FIXADDR_START, FIXADDR_TOP,
		(FIXADDR_TOP - FIXADDR_START) >> 10,

#ifdef CONFIG_HIGHMEM
		PKMAP_BASE, PKMAP_BASE+LAST_PKMAP*PAGE_SIZE,
		(LAST_PKMAP*PAGE_SIZE) >> 10,
#endif

		(unsigned long)VMALLOC_START, VMALLOC_END,
		(VMALLOC_END - VMALLOC_START) >> 20,

		(unsigned long)memory_start, (unsigned long)high_memory,
		((unsigned long)high_memory - (unsigned long)memory_start) >> 20,

485
#ifdef CONFIG_UNCACHED_MAPPING
P
Paul Mundt 已提交
486
		uncached_start, uncached_end, uncached_size >> 20,
487
#endif
488

489 490 491 492 493 494 495 496 497
		(unsigned long)&__init_begin, (unsigned long)&__init_end,
		((unsigned long)&__init_end -
		 (unsigned long)&__init_begin) >> 10,

		(unsigned long)&_etext, (unsigned long)&_edata,
		((unsigned long)&_edata - (unsigned long)&_etext) >> 10,

		(unsigned long)&_text, (unsigned long)&_etext,
		((unsigned long)&_etext - (unsigned long)&_text) >> 10);
498 499

	mem_init_done = 1;
L
Linus Torvalds 已提交
500 501 502 503 504
}

void free_initmem(void)
{
	unsigned long addr;
505

L
Linus Torvalds 已提交
506 507 508
	addr = (unsigned long)(&__init_begin);
	for (; addr < (unsigned long)(&__init_end); addr += PAGE_SIZE) {
		ClearPageReserved(virt_to_page(addr));
509
		init_page_count(virt_to_page(addr));
L
Linus Torvalds 已提交
510 511 512
		free_page(addr);
		totalram_pages++;
	}
513 514 515
	printk("Freeing unused kernel memory: %ldk freed\n",
	       ((unsigned long)&__init_end -
	        (unsigned long)&__init_begin) >> 10);
L
Linus Torvalds 已提交
516 517 518 519 520 521 522 523
}

#ifdef CONFIG_BLK_DEV_INITRD
void free_initrd_mem(unsigned long start, unsigned long end)
{
	unsigned long p;
	for (p = start; p < end; p += PAGE_SIZE) {
		ClearPageReserved(virt_to_page(p));
524
		init_page_count(virt_to_page(p));
L
Linus Torvalds 已提交
525 526 527
		free_page(p);
		totalram_pages++;
	}
528
	printk("Freeing initrd memory: %ldk freed\n", (end - start) >> 10);
L
Linus Torvalds 已提交
529 530
}
#endif
531 532 533 534 535 536 537 538 539 540 541 542

#ifdef CONFIG_MEMORY_HOTPLUG
int arch_add_memory(int nid, u64 start, u64 size)
{
	pg_data_t *pgdat;
	unsigned long start_pfn = start >> PAGE_SHIFT;
	unsigned long nr_pages = size >> PAGE_SHIFT;
	int ret;

	pgdat = NODE_DATA(nid);

	/* We only have ZONE_NORMAL, so this is easy.. */
543 544
	ret = __add_pages(nid, pgdat->node_zones + ZONE_NORMAL,
				start_pfn, nr_pages);
545
	if (unlikely(ret))
546
		printk("%s: Failed, __add_pages() == %d\n", __func__, ret);
547 548 549 550 551

	return ret;
}
EXPORT_SYMBOL_GPL(arch_add_memory);

552
#ifdef CONFIG_NUMA
553 554 555 556 557 558 559
int memory_add_physaddr_to_nid(u64 addr)
{
	/* Node 0 for now.. */
	return 0;
}
EXPORT_SYMBOL_GPL(memory_add_physaddr_to_nid);
#endif
M
Matt Fleming 已提交
560

561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577
#ifdef CONFIG_MEMORY_HOTREMOVE
int arch_remove_memory(u64 start, u64 size)
{
	unsigned long start_pfn = start >> PAGE_SHIFT;
	unsigned long nr_pages = size >> PAGE_SHIFT;
	struct zone *zone;
	int ret;

	zone = page_zone(pfn_to_page(start_pfn));
	ret = __remove_pages(zone, start_pfn, nr_pages);
	if (unlikely(ret))
		pr_warn("%s: Failed, __remove_pages() == %d\n", __func__,
			ret);

	return ret;
}
#endif
578
#endif /* CONFIG_MEMORY_HOTPLUG */