init.c 6.6 KB
Newer Older
L
Linus Torvalds 已提交
1 2
/*
 *  S390 version
3
 *    Copyright IBM Corp. 1999
L
Linus Torvalds 已提交
4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
 *    Author(s): Hartmut Penner (hp@de.ibm.com)
 *
 *  Derived from "arch/i386/mm/init.c"
 *    Copyright (C) 1995  Linus Torvalds
 */

#include <linux/signal.h>
#include <linux/sched.h>
#include <linux/kernel.h>
#include <linux/errno.h>
#include <linux/string.h>
#include <linux/types.h>
#include <linux/ptrace.h>
#include <linux/mman.h>
#include <linux/mm.h>
#include <linux/swap.h>
#include <linux/smp.h>
#include <linux/init.h>
#include <linux/pagemap.h>
#include <linux/bootmem.h>
24
#include <linux/pfn.h>
25
#include <linux/poison.h>
26
#include <linux/initrd.h>
27
#include <linux/export.h>
28
#include <linux/gfp.h>
L
Linus Torvalds 已提交
29 30 31 32 33 34 35 36
#include <asm/processor.h>
#include <asm/uaccess.h>
#include <asm/pgtable.h>
#include <asm/pgalloc.h>
#include <asm/dma.h>
#include <asm/lowcore.h>
#include <asm/tlb.h>
#include <asm/tlbflush.h>
37
#include <asm/sections.h>
38
#include <asm/ctl_reg.h>
L
Linus Torvalds 已提交
39 40

pgd_t swapper_pg_dir[PTRS_PER_PGD] __attribute__((__aligned__(PAGE_SIZE)));
41

42
unsigned long empty_zero_page, zero_page_mask;
43
EXPORT_SYMBOL(empty_zero_page);
L
Linus Torvalds 已提交
44

45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87
static unsigned long setup_zero_pages(void)
{
	struct cpuid cpu_id;
	unsigned int order;
	unsigned long size;
	struct page *page;
	int i;

	get_cpu_id(&cpu_id);
	switch (cpu_id.machine) {
	case 0x9672:	/* g5 */
	case 0x2064:	/* z900 */
	case 0x2066:	/* z900 */
	case 0x2084:	/* z990 */
	case 0x2086:	/* z990 */
	case 0x2094:	/* z9-109 */
	case 0x2096:	/* z9-109 */
		order = 0;
		break;
	case 0x2097:	/* z10 */
	case 0x2098:	/* z10 */
	default:
		order = 2;
		break;
	}

	empty_zero_page = __get_free_pages(GFP_KERNEL | __GFP_ZERO, order);
	if (!empty_zero_page)
		panic("Out of memory in setup_zero_pages");

	page = virt_to_page((void *) empty_zero_page);
	split_page(page, order);
	for (i = 1 << order; i > 0; i--) {
		SetPageReserved(page);
		page++;
	}

	size = PAGE_SIZE << order;
	zero_page_mask = (size - 1) & PAGE_MASK;

	return 1UL << order;
}

L
Linus Torvalds 已提交
88 89 90 91 92
/*
 * paging_init() sets up the page tables
 */
void __init paging_init(void)
{
93
	unsigned long max_zone_pfns[MAX_NR_ZONES];
94
	unsigned long pgd_type, asce_bits;
95

96
	init_mm.pgd = swapper_pg_dir;
H
Heiko Carstens 已提交
97
#ifdef CONFIG_64BIT
98 99 100 101 102 103 104
	if (VMALLOC_END > (1UL << 42)) {
		asce_bits = _ASCE_TYPE_REGION2 | _ASCE_TABLE_LENGTH;
		pgd_type = _REGION2_ENTRY_EMPTY;
	} else {
		asce_bits = _ASCE_TYPE_REGION3 | _ASCE_TABLE_LENGTH;
		pgd_type = _REGION3_ENTRY_EMPTY;
	}
H
Heiko Carstens 已提交
105
#else
106
	asce_bits = _ASCE_TABLE_LENGTH;
107
	pgd_type = _SEGMENT_ENTRY_EMPTY;
H
Heiko Carstens 已提交
108
#endif
109
	S390_lowcore.kernel_asce = (__pa(init_mm.pgd) & PAGE_MASK) | asce_bits;
110 111
	clear_table((unsigned long *) init_mm.pgd, pgd_type,
		    sizeof(unsigned long)*2048);
H
Heiko Carstens 已提交
112
	vmem_map_init();
L
Linus Torvalds 已提交
113 114

        /* enable virtual mapping in kernel mode */
115 116 117
	__ctl_load(S390_lowcore.kernel_asce, 1, 1);
	__ctl_load(S390_lowcore.kernel_asce, 7, 7);
	__ctl_load(S390_lowcore.kernel_asce, 13, 13);
D
David Howells 已提交
118
	arch_local_irq_restore(4UL << (BITS_PER_LONG - 8));
L
Linus Torvalds 已提交
119

120 121
	atomic_set(&init_mm.context.attach_count, 1);

122 123
	sparse_memory_present_with_active_regions(MAX_NUMNODES);
	sparse_init();
124 125 126 127
	memset(max_zone_pfns, 0, sizeof(max_zone_pfns));
	max_zone_pfns[ZONE_DMA] = PFN_DOWN(MAX_DMA_ADDRESS);
	max_zone_pfns[ZONE_NORMAL] = max_low_pfn;
	free_area_init_nodes(max_zone_pfns);
128
	fault_init();
L
Linus Torvalds 已提交
129 130 131 132 133 134 135 136 137
}

void __init mem_init(void)
{
	unsigned long codesize, reservedpages, datasize, initsize;

        max_mapnr = num_physpages = max_low_pfn;
        high_memory = (void *) __va(max_low_pfn * PAGE_SIZE);

138 139 140
	/* Setup guest page hinting */
	cmma_init();

L
Linus Torvalds 已提交
141 142
	/* this will put all low memory onto the freelists */
	totalram_pages += free_all_bootmem();
143
	totalram_pages -= setup_zero_pages();	/* Setup zeroed pages. */
L
Linus Torvalds 已提交
144 145 146 147 148 149 150

	reservedpages = 0;

	codesize =  (unsigned long) &_etext - (unsigned long) &_text;
	datasize =  (unsigned long) &_edata - (unsigned long) &_etext;
	initsize =  (unsigned long) &__init_end - (unsigned long) &__init_begin;
        printk("Memory: %luk/%luk available (%ldk kernel code, %ldk reserved, %ldk data, %ldk init)\n",
151
		nr_free_pages() << (PAGE_SHIFT-10),
L
Linus Torvalds 已提交
152 153 154 155 156
                max_mapnr << (PAGE_SHIFT-10),
                codesize >> 10,
                reservedpages << (PAGE_SHIFT-10),
                datasize >>10,
                initsize >> 10);
157
	printk("Write protected kernel read-only data: %#lx - %#lx\n",
158 159
	       (unsigned long)&_stext,
	       PFN_ALIGN((unsigned long)&_eshared) - 1);
L
Linus Torvalds 已提交
160 161
}

162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178
#ifdef CONFIG_DEBUG_PAGEALLOC
void kernel_map_pages(struct page *page, int numpages, int enable)
{
	pgd_t *pgd;
	pud_t *pud;
	pmd_t *pmd;
	pte_t *pte;
	unsigned long address;
	int i;

	for (i = 0; i < numpages; i++) {
		address = page_to_phys(page + i);
		pgd = pgd_offset_k(address);
		pud = pud_offset(pgd, address);
		pmd = pmd_offset(pud, address);
		pte = pte_offset_kernel(pmd, address);
		if (!enable) {
179 180
			__ptep_ipte(address, pte);
			pte_val(*pte) = _PAGE_TYPE_EMPTY;
181 182 183 184 185 186 187 188 189
			continue;
		}
		*pte = mk_pte_phys(address, __pgprot(_PAGE_TYPE_RW));
		/* Flush cpu write queue. */
		mb();
	}
}
#endif

190
void free_init_pages(char *what, unsigned long begin, unsigned long end)
L
Linus Torvalds 已提交
191
{
192
	unsigned long addr = begin;
L
Linus Torvalds 已提交
193

194 195 196
	if (begin >= end)
		return;
	for (; addr < end; addr += PAGE_SIZE) {
L
Linus Torvalds 已提交
197
		ClearPageReserved(virt_to_page(addr));
198
		init_page_count(virt_to_page(addr));
199 200
		memset((void *)(addr & PAGE_MASK), POISON_FREE_INITMEM,
		       PAGE_SIZE);
L
Linus Torvalds 已提交
201 202
		free_page(addr);
		totalram_pages++;
203 204 205 206 207 208 209 210 211
	}
	printk(KERN_INFO "Freeing %s: %luk freed\n", what, (end - begin) >> 10);
}

void free_initmem(void)
{
	free_init_pages("unused kernel memory",
			(unsigned long)&__init_begin,
			(unsigned long)&__init_end);
L
Linus Torvalds 已提交
212 213 214 215 216
}

#ifdef CONFIG_BLK_DEV_INITRD
void free_initrd_mem(unsigned long start, unsigned long end)
{
217
	free_init_pages("initrd memory", start, end);
L
Linus Torvalds 已提交
218 219
}
#endif
220 221 222 223

#ifdef CONFIG_MEMORY_HOTPLUG
int arch_add_memory(int nid, u64 start, u64 size)
{
224 225 226
	unsigned long zone_start_pfn, zone_end_pfn, nr_pages;
	unsigned long start_pfn = PFN_DOWN(start);
	unsigned long size_pages = PFN_DOWN(size);
227 228 229 230 231 232
	struct zone *zone;
	int rc;

	rc = vmem_add_mapping(start, size);
	if (rc)
		return rc;
233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255
	for_each_zone(zone) {
		if (zone_idx(zone) != ZONE_MOVABLE) {
			/* Add range within existing zone limits */
			zone_start_pfn = zone->zone_start_pfn;
			zone_end_pfn = zone->zone_start_pfn +
				       zone->spanned_pages;
		} else {
			/* Add remaining range to ZONE_MOVABLE */
			zone_start_pfn = start_pfn;
			zone_end_pfn = start_pfn + size_pages;
		}
		if (start_pfn < zone_start_pfn || start_pfn >= zone_end_pfn)
			continue;
		nr_pages = (start_pfn + size_pages > zone_end_pfn) ?
			   zone_end_pfn - start_pfn : size_pages;
		rc = __add_pages(nid, zone, start_pfn, nr_pages);
		if (rc)
			break;
		start_pfn += nr_pages;
		size_pages -= nr_pages;
		if (!size_pages)
			break;
	}
256 257 258 259 260
	if (rc)
		vmem_remove_mapping(start, size);
	return rc;
}
#endif /* CONFIG_MEMORY_HOTPLUG */