init.c 12.1 KB
Newer Older
1 2
/*
 * linux/arch/sh/mm/init.c
L
Linus Torvalds 已提交
3 4
 *
 *  Copyright (C) 1999  Niibe Yutaka
5
 *  Copyright (C) 2002 - 2011  Paul Mundt
L
Linus Torvalds 已提交
6 7 8 9 10 11 12
 *
 *  Based on linux/arch/i386/mm/init.c:
 *   Copyright (C) 1995  Linus Torvalds
 */
#include <linux/mm.h>
#include <linux/swap.h>
#include <linux/init.h>
13
#include <linux/gfp.h>
L
Linus Torvalds 已提交
14
#include <linux/bootmem.h>
P
Paul Mundt 已提交
15
#include <linux/proc_fs.h>
16
#include <linux/pagemap.h>
17 18
#include <linux/percpu.h>
#include <linux/io.h>
Y
Yinghai Lu 已提交
19
#include <linux/memblock.h>
20
#include <linux/dma-mapping.h>
21
#include <linux/export.h>
L
Linus Torvalds 已提交
22
#include <asm/mmu_context.h>
P
Paul Mundt 已提交
23
#include <asm/mmzone.h>
P
Paul Mundt 已提交
24
#include <asm/kexec.h>
L
Linus Torvalds 已提交
25 26
#include <asm/tlb.h>
#include <asm/cacheflush.h>
27
#include <asm/sections.h>
P
Paul Mundt 已提交
28
#include <asm/setup.h>
L
Linus Torvalds 已提交
29
#include <asm/cache.h>
30
#include <asm/sizes.h>
L
Linus Torvalds 已提交
31 32

pgd_t swapper_pg_dir[PTRS_PER_PGD];
33

34 35
void __init generic_mem_init(void)
{
Y
Yinghai Lu 已提交
36
	memblock_add(__MEMORY_START, __MEMORY_SIZE);
37 38
}

P
Paul Mundt 已提交
39 40 41 42 43
void __init __weak plat_mem_setup(void)
{
	/* Nothing to see here, move along. */
}

44
#ifdef CONFIG_MMU
45
static pte_t *__get_pte_phys(unsigned long addr)
L
Linus Torvalds 已提交
46 47
{
	pgd_t *pgd;
48
	pud_t *pud;
L
Linus Torvalds 已提交
49 50
	pmd_t *pmd;

S
Stuart Menefy 已提交
51
	pgd = pgd_offset_k(addr);
L
Linus Torvalds 已提交
52 53
	if (pgd_none(*pgd)) {
		pgd_ERROR(*pgd);
54
		return NULL;
L
Linus Torvalds 已提交
55 56
	}

S
Stuart Menefy 已提交
57 58 59
	pud = pud_alloc(NULL, pgd, addr);
	if (unlikely(!pud)) {
		pud_ERROR(*pud);
60
		return NULL;
61 62
	}

S
Stuart Menefy 已提交
63 64 65
	pmd = pmd_alloc(NULL, pud, addr);
	if (unlikely(!pmd)) {
		pmd_ERROR(*pmd);
66
		return NULL;
L
Linus Torvalds 已提交
67 68
	}

69
	return pte_offset_kernel(pmd, addr);
70 71 72 73 74 75 76
}

static void set_pte_phys(unsigned long addr, unsigned long phys, pgprot_t prot)
{
	pte_t *pte;

	pte = __get_pte_phys(addr);
L
Linus Torvalds 已提交
77 78 79 80 81 82
	if (!pte_none(*pte)) {
		pte_ERROR(*pte);
		return;
	}

	set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, prot));
83
	local_flush_tlb_one(get_asid(), addr);
84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99

	if (pgprot_val(prot) & _PAGE_WIRED)
		tlb_wire_entry(NULL, addr, *pte);
}

static void clear_pte_phys(unsigned long addr, pgprot_t prot)
{
	pte_t *pte;

	pte = __get_pte_phys(addr);

	if (pgprot_val(prot) & _PAGE_WIRED)
		tlb_unwire_entry();

	set_pte(pte, pfn_pte(0, __pgprot(0)));
	local_flush_tlb_one(get_asid(), addr);
L
Linus Torvalds 已提交
100 101 102 103 104 105 106 107 108 109 110 111 112
}

void __set_fixmap(enum fixed_addresses idx, unsigned long phys, pgprot_t prot)
{
	unsigned long address = __fix_to_virt(idx);

	if (idx >= __end_of_fixed_addresses) {
		BUG();
		return;
	}

	set_pte_phys(address, phys, prot);
}
113

114 115 116 117 118 119 120 121 122 123 124 125
void __clear_fixmap(enum fixed_addresses idx, pgprot_t prot)
{
	unsigned long address = __fix_to_virt(idx);

	if (idx >= __end_of_fixed_addresses) {
		BUG();
		return;
	}

	clear_pte_phys(address, prot);
}

126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157
static pmd_t * __init one_md_table_init(pud_t *pud)
{
	if (pud_none(*pud)) {
		pmd_t *pmd;

		pmd = alloc_bootmem_pages(PAGE_SIZE);
		pud_populate(&init_mm, pud, pmd);
		BUG_ON(pmd != pmd_offset(pud, 0));
	}

	return pmd_offset(pud, 0);
}

static pte_t * __init one_page_table_init(pmd_t *pmd)
{
	if (pmd_none(*pmd)) {
		pte_t *pte;

		pte = alloc_bootmem_pages(PAGE_SIZE);
		pmd_populate_kernel(&init_mm, pmd, pte);
		BUG_ON(pte != pte_offset_kernel(pmd, 0));
	}

	return pte_offset_kernel(pmd, 0);
}

static pte_t * __init page_table_kmap_check(pte_t *pte, pmd_t *pmd,
					    unsigned long vaddr, pte_t *lastpte)
{
	return pte;
}

158 159 160 161 162 163
void __init page_table_range_init(unsigned long start, unsigned long end,
					 pgd_t *pgd_base)
{
	pgd_t *pgd;
	pud_t *pud;
	pmd_t *pmd;
164
	pte_t *pte = NULL;
165
	int i, j, k;
166 167
	unsigned long vaddr;

168 169 170 171 172 173 174 175 176
	vaddr = start;
	i = __pgd_offset(vaddr);
	j = __pud_offset(vaddr);
	k = __pmd_offset(vaddr);
	pgd = pgd_base + i;

	for ( ; (i < PTRS_PER_PGD) && (vaddr != end); pgd++, i++) {
		pud = (pud_t *)pgd;
		for ( ; (j < PTRS_PER_PUD) && (vaddr != end); pud++, j++) {
177 178
			pmd = one_md_table_init(pud);
#ifndef __PAGETABLE_PMD_FOLDED
179 180
			pmd += k;
#endif
181
			for (; (k < PTRS_PER_PMD) && (vaddr != end); pmd++, k++) {
182 183
				pte = page_table_kmap_check(one_page_table_init(pmd),
							    pmd, vaddr, pte);
184 185 186
				vaddr += PMD_SIZE;
			}
			k = 0;
187
		}
188
		j = 0;
189 190
	}
}
191
#endif	/* CONFIG_MMU */
L
Linus Torvalds 已提交
192

P
Paul Mundt 已提交
193 194 195 196 197 198 199 200 201 202
void __init allocate_pgdat(unsigned int nid)
{
	unsigned long start_pfn, end_pfn;
#ifdef CONFIG_NEED_MULTIPLE_NODES
	unsigned long phys;
#endif

	get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);

#ifdef CONFIG_NEED_MULTIPLE_NODES
Y
Yinghai Lu 已提交
203
	phys = __memblock_alloc_base(sizeof(struct pglist_data),
P
Paul Mundt 已提交
204 205 206
				SMP_CACHE_BYTES, end_pfn << PAGE_SHIFT);
	/* Retry with all of system memory */
	if (!phys)
Y
Yinghai Lu 已提交
207 208
		phys = __memblock_alloc_base(sizeof(struct pglist_data),
					SMP_CACHE_BYTES, memblock_end_of_DRAM());
P
Paul Mundt 已提交
209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237
	if (!phys)
		panic("Can't allocate pgdat for node %d\n", nid);

	NODE_DATA(nid) = __va(phys);
	memset(NODE_DATA(nid), 0, sizeof(struct pglist_data));

	NODE_DATA(nid)->bdata = &bootmem_node_data[nid];
#endif

	NODE_DATA(nid)->node_start_pfn = start_pfn;
	NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn;
}

static void __init bootmem_init_one_node(unsigned int nid)
{
	unsigned long total_pages, paddr;
	unsigned long end_pfn;
	struct pglist_data *p;

	p = NODE_DATA(nid);

	/* Nothing to do.. */
	if (!p->node_spanned_pages)
		return;

	end_pfn = p->node_start_pfn + p->node_spanned_pages;

	total_pages = bootmem_bootmap_pages(p->node_spanned_pages);

Y
Yinghai Lu 已提交
238
	paddr = memblock_alloc(total_pages << PAGE_SHIFT, PAGE_SIZE);
P
Paul Mundt 已提交
239 240 241 242 243 244 245 246 247 248 249 250 251
	if (!paddr)
		panic("Can't allocate bootmap for nid[%d]\n", nid);

	init_bootmem_node(p, paddr >> PAGE_SHIFT, p->node_start_pfn, end_pfn);

	free_bootmem_with_active_regions(nid, end_pfn);

	/*
	 * XXX Handle initial reservations for the system memory node
	 * only for the moment, we'll refactor this later for handling
	 * reservations in other nodes.
	 */
	if (nid == 0) {
252 253
		struct memblock_region *reg;

P
Paul Mundt 已提交
254
		/* Reserve the sections we're already using. */
255 256 257
		for_each_memblock(reserved, reg) {
			reserve_bootmem(reg->base, reg->size, BOOTMEM_DEFAULT);
		}
P
Paul Mundt 已提交
258 259 260 261 262 263 264
	}

	sparse_memory_present_with_active_regions(nid);
}

static void __init do_init_bootmem(void)
{
265
	struct memblock_region *reg;
P
Paul Mundt 已提交
266 267 268
	int i;

	/* Add active regions with valid PFNs. */
269
	for_each_memblock(memory, reg) {
P
Paul Mundt 已提交
270
		unsigned long start_pfn, end_pfn;
271 272
		start_pfn = memblock_region_memory_base_pfn(reg);
		end_pfn = memblock_region_memory_end_pfn(reg);
P
Paul Mundt 已提交
273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290
		__add_active_range(0, start_pfn, end_pfn);
	}

	/* All of system RAM sits in node 0 for the non-NUMA case */
	allocate_pgdat(0);
	node_set_online(0);

	plat_mem_setup();

	for_each_online_node(i)
		bootmem_init_one_node(i);

	sparse_init();
}

static void __init early_reserve_mem(void)
{
	unsigned long start_pfn;
S
Simon Horman 已提交
291 292
	u32 zero_base = (u32)__MEMORY_START + (u32)PHYSICAL_OFFSET;
	u32 start = zero_base + (u32)CONFIG_ZERO_PAGE_OFFSET;
P
Paul Mundt 已提交
293 294 295 296 297 298 299 300 301 302 303 304 305

	/*
	 * Partially used pages are not usable - thus
	 * we are rounding upwards:
	 */
	start_pfn = PFN_UP(__pa(_end));

	/*
	 * Reserve the kernel text and Reserve the bootmem bitmap. We do
	 * this in two steps (first step was init_bootmem()), because
	 * this catches the (definitely buggy) case of us accidentally
	 * initializing the bootmem allocator with an invalid RAM area.
	 */
S
Simon Horman 已提交
306
	memblock_reserve(start, (PFN_PHYS(start_pfn) + PAGE_SIZE - 1) - start);
P
Paul Mundt 已提交
307 308 309 310 311

	/*
	 * Reserve physical pages below CONFIG_ZERO_PAGE_OFFSET.
	 */
	if (CONFIG_ZERO_PAGE_OFFSET != 0)
S
Simon Horman 已提交
312
		memblock_reserve(zero_base, CONFIG_ZERO_PAGE_OFFSET);
P
Paul Mundt 已提交
313 314 315 316 317 318 319 320

	/*
	 * Handle additional early reservations
	 */
	check_for_initrd();
	reserve_crashkernel();
}

L
Linus Torvalds 已提交
321 322
void __init paging_init(void)
{
323
	unsigned long max_zone_pfns[MAX_NR_ZONES];
324
	unsigned long vaddr, end;
325
	int nid;
L
Linus Torvalds 已提交
326

P
Paul Mundt 已提交
327 328 329 330
	sh_mv.mv_mem_init();

	early_reserve_mem();

331 332 333 334 335 336 337
	/*
	 * Once the early reservations are out of the way, give the
	 * platforms a chance to kick out some memory.
	 */
	if (sh_mv.mv_mem_reserve)
		sh_mv.mv_mem_reserve();

Y
Yinghai Lu 已提交
338
	memblock_enforce_memory_limit(memory_limit);
339
	memblock_allow_resize();
P
Paul Mundt 已提交
340

Y
Yinghai Lu 已提交
341
	memblock_dump_all();
P
Paul Mundt 已提交
342 343 344 345

	/*
	 * Determine low and high memory ranges:
	 */
Y
Yinghai Lu 已提交
346
	max_low_pfn = max_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT;
P
Paul Mundt 已提交
347 348 349 350 351
	min_low_pfn = __MEMORY_START >> PAGE_SHIFT;

	nodes_clear(node_online_map);

	memory_start = (unsigned long)__va(__MEMORY_START);
Y
Yinghai Lu 已提交
352
	memory_end = memory_start + (memory_limit ?: memblock_phys_mem_size());
P
Paul Mundt 已提交
353 354 355 356 357 358

	uncached_init();
	pmb_init();
	do_init_bootmem();
	ioremap_fixed_init();

359 360 361 362
	/* We don't need to map the kernel through the TLB, as
	 * it is permanatly mapped using P1. So clear the
	 * entire pgd. */
	memset(swapper_pg_dir, 0, sizeof(swapper_pg_dir));
L
Linus Torvalds 已提交
363

364 365 366 367
	/* Set an initial value for the MMU.TTB so we don't have to
	 * check for a null value. */
	set_TTB(swapper_pg_dir);

368 369
	/*
	 * Populate the relevant portions of swapper_pg_dir so that
370
	 * we can use the fixmap entries without calling kmalloc.
371 372 373
	 * pte's will be filled in by __set_fixmap().
	 */
	vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK;
374 375
	end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK;
	page_table_range_init(vaddr, end, swapper_pg_dir);
376 377

	kmap_coherent_init();
378

379 380
	memset(max_zone_pfns, 0, sizeof(max_zone_pfns));

381 382 383 384
	for_each_online_node(nid) {
		pg_data_t *pgdat = NODE_DATA(nid);
		unsigned long low, start_pfn;

385
		start_pfn = pgdat->bdata->node_min_pfn;
386 387
		low = pgdat->bdata->node_low_pfn;

388 389
		if (max_zone_pfns[ZONE_NORMAL] < low)
			max_zone_pfns[ZONE_NORMAL] = low;
390 391 392 393

		printk("Node %u: start_pfn = 0x%lx, low = 0x%lx\n",
		       nid, start_pfn, low);
	}
394 395

	free_area_init_nodes(max_zone_pfns);
L
Linus Torvalds 已提交
396 397
}

398 399 400 401 402 403 404 405
/*
 * Early initialization for any I/O MMUs we might have.
 */
static void __init iommu_init(void)
{
	no_iommu_init();
}

406 407
unsigned int mem_init_done = 0;

L
Linus Torvalds 已提交
408 409
void __init mem_init(void)
{
410
	pg_data_t *pgdat;
L
Linus Torvalds 已提交
411

412 413
	iommu_init();

414
	high_memory = NULL;
415 416 417
	for_each_online_pgdat(pgdat)
		high_memory = max_t(void *, high_memory,
				    __va(pgdat_end_pfn(pgdat) << PAGE_SHIFT));
418

419
	free_all_bootmem();
L
Linus Torvalds 已提交
420

421 422 423
	/* Set this up early, so we can take care of the zero page */
	cpu_cache_init();

L
Linus Torvalds 已提交
424 425 426 427
	/* clear the zero-page */
	memset(empty_zero_page, 0, PAGE_SIZE);
	__flush_wback_region(empty_zero_page, PAGE_SIZE);

428 429
	vsyscall_init();

430 431
	mem_init_print_info(NULL);
	pr_info("virtual kernel memory layout:\n"
432 433 434 435 436
		"    fixmap  : 0x%08lx - 0x%08lx   (%4ld kB)\n"
#ifdef CONFIG_HIGHMEM
		"    pkmap   : 0x%08lx - 0x%08lx   (%4ld kB)\n"
#endif
		"    vmalloc : 0x%08lx - 0x%08lx   (%4ld MB)\n"
437
		"    lowmem  : 0x%08lx - 0x%08lx   (%4ld MB) (cached)\n"
438
#ifdef CONFIG_UNCACHED_MAPPING
439
		"            : 0x%08lx - 0x%08lx   (%4ld MB) (uncached)\n"
440
#endif
441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457
		"      .init : 0x%08lx - 0x%08lx   (%4ld kB)\n"
		"      .data : 0x%08lx - 0x%08lx   (%4ld kB)\n"
		"      .text : 0x%08lx - 0x%08lx   (%4ld kB)\n",
		FIXADDR_START, FIXADDR_TOP,
		(FIXADDR_TOP - FIXADDR_START) >> 10,

#ifdef CONFIG_HIGHMEM
		PKMAP_BASE, PKMAP_BASE+LAST_PKMAP*PAGE_SIZE,
		(LAST_PKMAP*PAGE_SIZE) >> 10,
#endif

		(unsigned long)VMALLOC_START, VMALLOC_END,
		(VMALLOC_END - VMALLOC_START) >> 20,

		(unsigned long)memory_start, (unsigned long)high_memory,
		((unsigned long)high_memory - (unsigned long)memory_start) >> 20,

458
#ifdef CONFIG_UNCACHED_MAPPING
P
Paul Mundt 已提交
459
		uncached_start, uncached_end, uncached_size >> 20,
460
#endif
461

462 463 464 465 466 467 468 469 470
		(unsigned long)&__init_begin, (unsigned long)&__init_end,
		((unsigned long)&__init_end -
		 (unsigned long)&__init_begin) >> 10,

		(unsigned long)&_etext, (unsigned long)&_edata,
		((unsigned long)&_edata - (unsigned long)&_etext) >> 10,

		(unsigned long)&_text, (unsigned long)&_etext,
		((unsigned long)&_etext - (unsigned long)&_text) >> 10);
471 472

	mem_init_done = 1;
L
Linus Torvalds 已提交
473 474 475 476
}

void free_initmem(void)
{
477
	free_initmem_default(-1);
L
Linus Torvalds 已提交
478 479 480 481 482
}

#ifdef CONFIG_BLK_DEV_INITRD
void free_initrd_mem(unsigned long start, unsigned long end)
{
483
	free_reserved_area((void *)start, (void *)end, -1, "initrd");
L
Linus Torvalds 已提交
484 485
}
#endif
486 487 488 489 490 491 492 493 494 495 496 497

#ifdef CONFIG_MEMORY_HOTPLUG
int arch_add_memory(int nid, u64 start, u64 size)
{
	pg_data_t *pgdat;
	unsigned long start_pfn = start >> PAGE_SHIFT;
	unsigned long nr_pages = size >> PAGE_SHIFT;
	int ret;

	pgdat = NODE_DATA(nid);

	/* We only have ZONE_NORMAL, so this is easy.. */
498 499
	ret = __add_pages(nid, pgdat->node_zones + ZONE_NORMAL,
				start_pfn, nr_pages);
500
	if (unlikely(ret))
501
		printk("%s: Failed, __add_pages() == %d\n", __func__, ret);
502 503 504 505 506

	return ret;
}
EXPORT_SYMBOL_GPL(arch_add_memory);

507
#ifdef CONFIG_NUMA
508 509 510 511 512 513 514
int memory_add_physaddr_to_nid(u64 addr)
{
	/* Node 0 for now.. */
	return 0;
}
EXPORT_SYMBOL_GPL(memory_add_physaddr_to_nid);
#endif
M
Matt Fleming 已提交
515

516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532
#ifdef CONFIG_MEMORY_HOTREMOVE
int arch_remove_memory(u64 start, u64 size)
{
	unsigned long start_pfn = start >> PAGE_SHIFT;
	unsigned long nr_pages = size >> PAGE_SHIFT;
	struct zone *zone;
	int ret;

	zone = page_zone(pfn_to_page(start_pfn));
	ret = __remove_pages(zone, start_pfn, nr_pages);
	if (unlikely(ret))
		pr_warn("%s: Failed, __remove_pages() == %d\n", __func__,
			ret);

	return ret;
}
#endif
533
#endif /* CONFIG_MEMORY_HOTPLUG */