xref: /linux/arch/s390/kernel/setup.c (revision ed3174d93c342b8b2eeba6bbd124707d55304a7b)
1 /*
2  *  arch/s390/kernel/setup.c
3  *
4  *  S390 version
5  *    Copyright (C) 1999,2000 IBM Deutschland Entwicklung GmbH, IBM Corporation
6  *    Author(s): Hartmut Penner (hp@de.ibm.com),
7  *               Martin Schwidefsky (schwidefsky@de.ibm.com)
8  *
9  *  Derived from "arch/i386/kernel/setup.c"
10  *    Copyright (C) 1995, Linus Torvalds
11  */
12 
13 /*
14  * This file handles the architecture-dependent parts of initialization
15  */
16 
17 #include <linux/errno.h>
18 #include <linux/module.h>
19 #include <linux/sched.h>
20 #include <linux/kernel.h>
21 #include <linux/mm.h>
22 #include <linux/stddef.h>
23 #include <linux/unistd.h>
24 #include <linux/ptrace.h>
25 #include <linux/slab.h>
26 #include <linux/user.h>
27 #include <linux/tty.h>
28 #include <linux/ioport.h>
29 #include <linux/delay.h>
30 #include <linux/init.h>
31 #include <linux/initrd.h>
32 #include <linux/bootmem.h>
33 #include <linux/root_dev.h>
34 #include <linux/console.h>
35 #include <linux/seq_file.h>
36 #include <linux/kernel_stat.h>
37 #include <linux/device.h>
38 #include <linux/notifier.h>
39 #include <linux/pfn.h>
40 #include <linux/ctype.h>
41 #include <linux/reboot.h>
42 
43 #include <asm/ipl.h>
44 #include <asm/uaccess.h>
45 #include <asm/system.h>
46 #include <asm/smp.h>
47 #include <asm/mmu_context.h>
48 #include <asm/cpcmd.h>
49 #include <asm/lowcore.h>
50 #include <asm/irq.h>
51 #include <asm/page.h>
52 #include <asm/ptrace.h>
53 #include <asm/sections.h>
54 #include <asm/ebcdic.h>
55 #include <asm/compat.h>
56 
57 long psw_kernel_bits	= (PSW_BASE_BITS | PSW_MASK_DAT | PSW_ASC_PRIMARY |
58 			   PSW_MASK_MCHECK | PSW_DEFAULT_KEY);
59 long psw_user_bits	= (PSW_BASE_BITS | PSW_MASK_DAT | PSW_ASC_HOME |
60 			   PSW_MASK_IO | PSW_MASK_EXT | PSW_MASK_MCHECK |
61 			   PSW_MASK_PSTATE | PSW_DEFAULT_KEY);
62 
63 /*
64  * User copy operations.
65  */
66 struct uaccess_ops uaccess;
67 EXPORT_SYMBOL(uaccess);
68 
69 /*
70  * Machine setup..
71  */
72 unsigned int console_mode = 0;
73 unsigned int console_devno = -1;
74 unsigned int console_irq = -1;
75 unsigned long machine_flags = 0;
76 unsigned long elf_hwcap = 0;
77 char elf_platform[ELF_PLATFORM_SIZE];
78 
79 struct mem_chunk __meminitdata memory_chunk[MEMORY_CHUNKS];
80 volatile int __cpu_logical_map[NR_CPUS]; /* logical cpu to cpu address */
81 static unsigned long __initdata memory_end;
82 
83 /*
84  * This is set up by the setup-routine at boot-time
85  * for S390 need to find out, what we have to setup
86  * using address 0x10400 ...
87  */
88 
89 #include <asm/setup.h>
90 
91 static struct resource code_resource = {
92 	.name  = "Kernel code",
93 	.flags = IORESOURCE_BUSY | IORESOURCE_MEM,
94 };
95 
96 static struct resource data_resource = {
97 	.name = "Kernel data",
98 	.flags = IORESOURCE_BUSY | IORESOURCE_MEM,
99 };
100 
101 /*
102  * cpu_init() initializes state that is per-CPU.
103  */
104 void __cpuinit cpu_init(void)
105 {
106         int addr = hard_smp_processor_id();
107 
108         /*
109          * Store processor id in lowcore (used e.g. in timer_interrupt)
110          */
111 	get_cpu_id(&S390_lowcore.cpu_data.cpu_id);
112         S390_lowcore.cpu_data.cpu_addr = addr;
113 
114         /*
115          * Force FPU initialization:
116          */
117         clear_thread_flag(TIF_USEDFPU);
118         clear_used_math();
119 
120 	atomic_inc(&init_mm.mm_count);
121 	current->active_mm = &init_mm;
122         if (current->mm)
123                 BUG();
124         enter_lazy_tlb(&init_mm, current);
125 }
126 
127 /*
128  * condev= and conmode= setup parameter.
129  */
130 
131 static int __init condev_setup(char *str)
132 {
133 	int vdev;
134 
135 	vdev = simple_strtoul(str, &str, 0);
136 	if (vdev >= 0 && vdev < 65536) {
137 		console_devno = vdev;
138 		console_irq = -1;
139 	}
140 	return 1;
141 }
142 
143 __setup("condev=", condev_setup);
144 
145 static int __init conmode_setup(char *str)
146 {
147 #if defined(CONFIG_SCLP_CONSOLE) || defined(CONFIG_SCLP_VT220_CONSOLE)
148 	if (strncmp(str, "hwc", 4) == 0 || strncmp(str, "sclp", 5) == 0)
149                 SET_CONSOLE_SCLP;
150 #endif
151 #if defined(CONFIG_TN3215_CONSOLE)
152 	if (strncmp(str, "3215", 5) == 0)
153 		SET_CONSOLE_3215;
154 #endif
155 #if defined(CONFIG_TN3270_CONSOLE)
156 	if (strncmp(str, "3270", 5) == 0)
157 		SET_CONSOLE_3270;
158 #endif
159         return 1;
160 }
161 
162 __setup("conmode=", conmode_setup);
163 
164 static void __init conmode_default(void)
165 {
166 	char query_buffer[1024];
167 	char *ptr;
168 
169         if (MACHINE_IS_VM) {
170 		cpcmd("QUERY CONSOLE", query_buffer, 1024, NULL);
171 		console_devno = simple_strtoul(query_buffer + 5, NULL, 16);
172 		ptr = strstr(query_buffer, "SUBCHANNEL =");
173 		console_irq = simple_strtoul(ptr + 13, NULL, 16);
174 		cpcmd("QUERY TERM", query_buffer, 1024, NULL);
175 		ptr = strstr(query_buffer, "CONMODE");
176 		/*
177 		 * Set the conmode to 3215 so that the device recognition
178 		 * will set the cu_type of the console to 3215. If the
179 		 * conmode is 3270 and we don't set it back then both
180 		 * 3215 and the 3270 driver will try to access the console
181 		 * device (3215 as console and 3270 as normal tty).
182 		 */
183 		cpcmd("TERM CONMODE 3215", NULL, 0, NULL);
184 		if (ptr == NULL) {
185 #if defined(CONFIG_SCLP_CONSOLE) || defined(CONFIG_SCLP_VT220_CONSOLE)
186 			SET_CONSOLE_SCLP;
187 #endif
188 			return;
189 		}
190 		if (strncmp(ptr + 8, "3270", 4) == 0) {
191 #if defined(CONFIG_TN3270_CONSOLE)
192 			SET_CONSOLE_3270;
193 #elif defined(CONFIG_TN3215_CONSOLE)
194 			SET_CONSOLE_3215;
195 #elif defined(CONFIG_SCLP_CONSOLE) || defined(CONFIG_SCLP_VT220_CONSOLE)
196 			SET_CONSOLE_SCLP;
197 #endif
198 		} else if (strncmp(ptr + 8, "3215", 4) == 0) {
199 #if defined(CONFIG_TN3215_CONSOLE)
200 			SET_CONSOLE_3215;
201 #elif defined(CONFIG_TN3270_CONSOLE)
202 			SET_CONSOLE_3270;
203 #elif defined(CONFIG_SCLP_CONSOLE) || defined(CONFIG_SCLP_VT220_CONSOLE)
204 			SET_CONSOLE_SCLP;
205 #endif
206 		}
207         } else if (MACHINE_IS_P390) {
208 #if defined(CONFIG_TN3215_CONSOLE)
209 		SET_CONSOLE_3215;
210 #elif defined(CONFIG_TN3270_CONSOLE)
211 		SET_CONSOLE_3270;
212 #endif
213 	} else {
214 #if defined(CONFIG_SCLP_CONSOLE) || defined(CONFIG_SCLP_VT220_CONSOLE)
215 		SET_CONSOLE_SCLP;
216 #endif
217 	}
218 }
219 
220 #if defined(CONFIG_ZFCPDUMP) || defined(CONFIG_ZFCPDUMP_MODULE)
221 static void __init setup_zfcpdump(unsigned int console_devno)
222 {
223 	static char str[64];
224 
225 	if (ipl_info.type != IPL_TYPE_FCP_DUMP)
226 		return;
227 	if (console_devno != -1)
228 		sprintf(str, "cio_ignore=all,!0.0.%04x,!0.0.%04x",
229 			ipl_info.data.fcp.dev_id.devno, console_devno);
230 	else
231 		sprintf(str, "cio_ignore=all,!0.0.%04x",
232 			ipl_info.data.fcp.dev_id.devno);
233 	strcat(COMMAND_LINE, " ");
234 	strcat(COMMAND_LINE, str);
235 	console_loglevel = 2;
236 }
237 #else
238 static inline void setup_zfcpdump(unsigned int console_devno) {}
239 #endif /* CONFIG_ZFCPDUMP */
240 
241  /*
242  * Reboot, halt and power_off stubs. They just call _machine_restart,
243  * _machine_halt or _machine_power_off.
244  */
245 
246 void machine_restart(char *command)
247 {
248 	if ((!in_interrupt() && !in_atomic()) || oops_in_progress)
249 		/*
250 		 * Only unblank the console if we are called in enabled
251 		 * context or a bust_spinlocks cleared the way for us.
252 		 */
253 		console_unblank();
254 	_machine_restart(command);
255 }
256 
257 void machine_halt(void)
258 {
259 	if (!in_interrupt() || oops_in_progress)
260 		/*
261 		 * Only unblank the console if we are called in enabled
262 		 * context or a bust_spinlocks cleared the way for us.
263 		 */
264 		console_unblank();
265 	_machine_halt();
266 }
267 
268 void machine_power_off(void)
269 {
270 	if (!in_interrupt() || oops_in_progress)
271 		/*
272 		 * Only unblank the console if we are called in enabled
273 		 * context or a bust_spinlocks cleared the way for us.
274 		 */
275 		console_unblank();
276 	_machine_power_off();
277 }
278 
279 /*
280  * Dummy power off function.
281  */
282 void (*pm_power_off)(void) = machine_power_off;
283 
284 static int __init early_parse_mem(char *p)
285 {
286 	memory_end = memparse(p, &p);
287 	return 0;
288 }
289 early_param("mem", early_parse_mem);
290 
291 /*
292  * "ipldelay=XXX[sm]" sets ipl delay in seconds or minutes
293  */
294 static int __init early_parse_ipldelay(char *p)
295 {
296 	unsigned long delay = 0;
297 
298 	delay = simple_strtoul(p, &p, 0);
299 
300 	switch (*p) {
301 	case 's':
302 	case 'S':
303 		delay *= 1000000;
304 		break;
305 	case 'm':
306 	case 'M':
307 		delay *= 60 * 1000000;
308 	}
309 
310 	/* now wait for the requested amount of time */
311 	udelay(delay);
312 
313 	return 0;
314 }
315 early_param("ipldelay", early_parse_ipldelay);
316 
317 #ifdef CONFIG_S390_SWITCH_AMODE
318 unsigned int switch_amode = 0;
319 EXPORT_SYMBOL_GPL(switch_amode);
320 
321 static void set_amode_and_uaccess(unsigned long user_amode,
322 				  unsigned long user32_amode)
323 {
324 	psw_user_bits = PSW_BASE_BITS | PSW_MASK_DAT | user_amode |
325 			PSW_MASK_IO | PSW_MASK_EXT | PSW_MASK_MCHECK |
326 			PSW_MASK_PSTATE | PSW_DEFAULT_KEY;
327 #ifdef CONFIG_COMPAT
328 	psw_user32_bits = PSW_BASE32_BITS | PSW_MASK_DAT | user_amode |
329 			  PSW_MASK_IO | PSW_MASK_EXT | PSW_MASK_MCHECK |
330 			  PSW_MASK_PSTATE | PSW_DEFAULT_KEY;
331 	psw32_user_bits = PSW32_BASE_BITS | PSW32_MASK_DAT | user32_amode |
332 			  PSW32_MASK_IO | PSW32_MASK_EXT | PSW32_MASK_MCHECK |
333 			  PSW32_MASK_PSTATE;
334 #endif
335 	psw_kernel_bits = PSW_BASE_BITS | PSW_MASK_DAT | PSW_ASC_HOME |
336 			  PSW_MASK_MCHECK | PSW_DEFAULT_KEY;
337 
338 	if (MACHINE_HAS_MVCOS) {
339 		printk("mvcos available.\n");
340 		memcpy(&uaccess, &uaccess_mvcos_switch, sizeof(uaccess));
341 	} else {
342 		printk("mvcos not available.\n");
343 		memcpy(&uaccess, &uaccess_pt, sizeof(uaccess));
344 	}
345 }
346 
347 /*
348  * Switch kernel/user addressing modes?
349  */
350 static int __init early_parse_switch_amode(char *p)
351 {
352 	switch_amode = 1;
353 	return 0;
354 }
355 early_param("switch_amode", early_parse_switch_amode);
356 
357 #else /* CONFIG_S390_SWITCH_AMODE */
358 static inline void set_amode_and_uaccess(unsigned long user_amode,
359 					 unsigned long user32_amode)
360 {
361 }
362 #endif /* CONFIG_S390_SWITCH_AMODE */
363 
364 #ifdef CONFIG_S390_EXEC_PROTECT
365 unsigned int s390_noexec = 0;
366 EXPORT_SYMBOL_GPL(s390_noexec);
367 
368 /*
369  * Enable execute protection?
370  */
371 static int __init early_parse_noexec(char *p)
372 {
373 	if (!strncmp(p, "off", 3))
374 		return 0;
375 	switch_amode = 1;
376 	s390_noexec = 1;
377 	return 0;
378 }
379 early_param("noexec", early_parse_noexec);
380 #endif /* CONFIG_S390_EXEC_PROTECT */
381 
382 static void setup_addressing_mode(void)
383 {
384 	if (s390_noexec) {
385 		printk("S390 execute protection active, ");
386 		set_amode_and_uaccess(PSW_ASC_SECONDARY, PSW32_ASC_SECONDARY);
387 	} else if (switch_amode) {
388 		printk("S390 address spaces switched, ");
389 		set_amode_and_uaccess(PSW_ASC_PRIMARY, PSW32_ASC_PRIMARY);
390 	}
391 #ifdef CONFIG_TRACE_IRQFLAGS
392 	sysc_restore_trace_psw.mask = psw_kernel_bits & ~PSW_MASK_MCHECK;
393 	io_restore_trace_psw.mask = psw_kernel_bits & ~PSW_MASK_MCHECK;
394 #endif
395 }
396 
397 static void __init
398 setup_lowcore(void)
399 {
400 	struct _lowcore *lc;
401 	int lc_pages;
402 
403 	/*
404 	 * Setup lowcore for boot cpu
405 	 */
406 	lc_pages = sizeof(void *) == 8 ? 2 : 1;
407 	lc = (struct _lowcore *)
408 		__alloc_bootmem(lc_pages * PAGE_SIZE, lc_pages * PAGE_SIZE, 0);
409 	memset(lc, 0, lc_pages * PAGE_SIZE);
410 	lc->restart_psw.mask = PSW_BASE_BITS | PSW_DEFAULT_KEY;
411 	lc->restart_psw.addr =
412 		PSW_ADDR_AMODE | (unsigned long) restart_int_handler;
413 	if (switch_amode)
414 		lc->restart_psw.mask |= PSW_ASC_HOME;
415 	lc->external_new_psw.mask = psw_kernel_bits;
416 	lc->external_new_psw.addr =
417 		PSW_ADDR_AMODE | (unsigned long) ext_int_handler;
418 	lc->svc_new_psw.mask = psw_kernel_bits | PSW_MASK_IO | PSW_MASK_EXT;
419 	lc->svc_new_psw.addr = PSW_ADDR_AMODE | (unsigned long) system_call;
420 	lc->program_new_psw.mask = psw_kernel_bits;
421 	lc->program_new_psw.addr =
422 		PSW_ADDR_AMODE | (unsigned long)pgm_check_handler;
423 	lc->mcck_new_psw.mask =
424 		psw_kernel_bits & ~PSW_MASK_MCHECK & ~PSW_MASK_DAT;
425 	lc->mcck_new_psw.addr =
426 		PSW_ADDR_AMODE | (unsigned long) mcck_int_handler;
427 	lc->io_new_psw.mask = psw_kernel_bits;
428 	lc->io_new_psw.addr = PSW_ADDR_AMODE | (unsigned long) io_int_handler;
429 	lc->ipl_device = S390_lowcore.ipl_device;
430 	lc->jiffy_timer = -1LL;
431 	lc->kernel_stack = ((unsigned long) &init_thread_union) + THREAD_SIZE;
432 	lc->async_stack = (unsigned long)
433 		__alloc_bootmem(ASYNC_SIZE, ASYNC_SIZE, 0) + ASYNC_SIZE;
434 	lc->panic_stack = (unsigned long)
435 		__alloc_bootmem(PAGE_SIZE, PAGE_SIZE, 0) + PAGE_SIZE;
436 	lc->current_task = (unsigned long) init_thread_union.thread_info.task;
437 	lc->thread_info = (unsigned long) &init_thread_union;
438 #ifndef CONFIG_64BIT
439 	if (MACHINE_HAS_IEEE) {
440 		lc->extended_save_area_addr = (__u32)
441 			__alloc_bootmem(PAGE_SIZE, PAGE_SIZE, 0);
442 		/* enable extended save area */
443 		__ctl_set_bit(14, 29);
444 	}
445 #endif
446 	set_prefix((u32)(unsigned long) lc);
447 }
448 
449 static void __init
450 setup_resources(void)
451 {
452 	struct resource *res, *sub_res;
453 	int i;
454 
455 	code_resource.start = (unsigned long) &_text;
456 	code_resource.end = (unsigned long) &_etext - 1;
457 	data_resource.start = (unsigned long) &_etext;
458 	data_resource.end = (unsigned long) &_edata - 1;
459 
460 	for (i = 0; i < MEMORY_CHUNKS; i++) {
461 		if (!memory_chunk[i].size)
462 			continue;
463 		res = alloc_bootmem_low(sizeof(struct resource));
464 		res->flags = IORESOURCE_BUSY | IORESOURCE_MEM;
465 		switch (memory_chunk[i].type) {
466 		case CHUNK_READ_WRITE:
467 			res->name = "System RAM";
468 			break;
469 		case CHUNK_READ_ONLY:
470 			res->name = "System ROM";
471 			res->flags |= IORESOURCE_READONLY;
472 			break;
473 		default:
474 			res->name = "reserved";
475 		}
476 		res->start = memory_chunk[i].addr;
477 		res->end = memory_chunk[i].addr +  memory_chunk[i].size - 1;
478 		request_resource(&iomem_resource, res);
479 
480 		if (code_resource.start >= res->start  &&
481 			code_resource.start <= res->end &&
482 			code_resource.end > res->end) {
483 			sub_res = alloc_bootmem_low(sizeof(struct resource));
484 			memcpy(sub_res, &code_resource,
485 				sizeof(struct resource));
486 			sub_res->end = res->end;
487 			code_resource.start = res->end + 1;
488 			request_resource(res, sub_res);
489 		}
490 
491 		if (code_resource.start >= res->start &&
492 			code_resource.start <= res->end &&
493 			code_resource.end <= res->end)
494 			request_resource(res, &code_resource);
495 
496 		if (data_resource.start >= res->start &&
497 			data_resource.start <= res->end &&
498 			data_resource.end > res->end) {
499 			sub_res = alloc_bootmem_low(sizeof(struct resource));
500 			memcpy(sub_res, &data_resource,
501 				sizeof(struct resource));
502 			sub_res->end = res->end;
503 			data_resource.start = res->end + 1;
504 			request_resource(res, sub_res);
505 		}
506 
507 		if (data_resource.start >= res->start &&
508 			data_resource.start <= res->end &&
509 			data_resource.end <= res->end)
510 			request_resource(res, &data_resource);
511 	}
512 }
513 
514 unsigned long real_memory_size;
515 EXPORT_SYMBOL_GPL(real_memory_size);
516 
517 static void __init setup_memory_end(void)
518 {
519 	unsigned long memory_size;
520 	unsigned long max_mem;
521 	int i;
522 
523 #if defined(CONFIG_ZFCPDUMP) || defined(CONFIG_ZFCPDUMP_MODULE)
524 	if (ipl_info.type == IPL_TYPE_FCP_DUMP)
525 		memory_end = ZFCPDUMP_HSA_SIZE;
526 #endif
527 	memory_size = 0;
528 	memory_end &= PAGE_MASK;
529 
530 	max_mem = memory_end ? min(VMEM_MAX_PHYS, memory_end) : VMEM_MAX_PHYS;
531 	memory_end = min(max_mem, memory_end);
532 
533 	/*
534 	 * Make sure all chunks are MAX_ORDER aligned so we don't need the
535 	 * extra checks that HOLES_IN_ZONE would require.
536 	 */
537 	for (i = 0; i < MEMORY_CHUNKS; i++) {
538 		unsigned long start, end;
539 		struct mem_chunk *chunk;
540 		unsigned long align;
541 
542 		chunk = &memory_chunk[i];
543 		align = 1UL << (MAX_ORDER + PAGE_SHIFT - 1);
544 		start = (chunk->addr + align - 1) & ~(align - 1);
545 		end = (chunk->addr + chunk->size) & ~(align - 1);
546 		if (start >= end)
547 			memset(chunk, 0, sizeof(*chunk));
548 		else {
549 			chunk->addr = start;
550 			chunk->size = end - start;
551 		}
552 	}
553 
554 	for (i = 0; i < MEMORY_CHUNKS; i++) {
555 		struct mem_chunk *chunk = &memory_chunk[i];
556 
557 		real_memory_size = max(real_memory_size,
558 				       chunk->addr + chunk->size);
559 		if (chunk->addr >= max_mem) {
560 			memset(chunk, 0, sizeof(*chunk));
561 			continue;
562 		}
563 		if (chunk->addr + chunk->size > max_mem)
564 			chunk->size = max_mem - chunk->addr;
565 		memory_size = max(memory_size, chunk->addr + chunk->size);
566 	}
567 	if (!memory_end)
568 		memory_end = memory_size;
569 }
570 
571 static void __init
572 setup_memory(void)
573 {
574         unsigned long bootmap_size;
575 	unsigned long start_pfn, end_pfn;
576 	int i;
577 
578 	/*
579 	 * partially used pages are not usable - thus
580 	 * we are rounding upwards:
581 	 */
582 	start_pfn = PFN_UP(__pa(&_end));
583 	end_pfn = max_pfn = PFN_DOWN(memory_end);
584 
585 #ifdef CONFIG_BLK_DEV_INITRD
586 	/*
587 	 * Move the initrd in case the bitmap of the bootmem allocater
588 	 * would overwrite it.
589 	 */
590 
591 	if (INITRD_START && INITRD_SIZE) {
592 		unsigned long bmap_size;
593 		unsigned long start;
594 
595 		bmap_size = bootmem_bootmap_pages(end_pfn - start_pfn + 1);
596 		bmap_size = PFN_PHYS(bmap_size);
597 
598 		if (PFN_PHYS(start_pfn) + bmap_size > INITRD_START) {
599 			start = PFN_PHYS(start_pfn) + bmap_size + PAGE_SIZE;
600 
601 			if (start + INITRD_SIZE > memory_end) {
602 				printk("initrd extends beyond end of memory "
603 				       "(0x%08lx > 0x%08lx)\n"
604 				       "disabling initrd\n",
605 				       start + INITRD_SIZE, memory_end);
606 				INITRD_START = INITRD_SIZE = 0;
607 			} else {
608 				printk("Moving initrd (0x%08lx -> 0x%08lx, "
609 				       "size: %ld)\n",
610 				       INITRD_START, start, INITRD_SIZE);
611 				memmove((void *) start, (void *) INITRD_START,
612 					INITRD_SIZE);
613 				INITRD_START = start;
614 			}
615 		}
616 	}
617 #endif
618 
619 	/*
620 	 * Initialize the boot-time allocator
621 	 */
622 	bootmap_size = init_bootmem(start_pfn, end_pfn);
623 
624 	/*
625 	 * Register RAM areas with the bootmem allocator.
626 	 */
627 
628 	for (i = 0; i < MEMORY_CHUNKS && memory_chunk[i].size > 0; i++) {
629 		unsigned long start_chunk, end_chunk, pfn;
630 
631 		if (memory_chunk[i].type != CHUNK_READ_WRITE)
632 			continue;
633 		start_chunk = PFN_DOWN(memory_chunk[i].addr);
634 		end_chunk = start_chunk + PFN_DOWN(memory_chunk[i].size) - 1;
635 		end_chunk = min(end_chunk, end_pfn);
636 		if (start_chunk >= end_chunk)
637 			continue;
638 		add_active_range(0, start_chunk, end_chunk);
639 		pfn = max(start_chunk, start_pfn);
640 		for (; pfn <= end_chunk; pfn++)
641 			page_set_storage_key(PFN_PHYS(pfn), PAGE_DEFAULT_KEY);
642 	}
643 
644 	psw_set_key(PAGE_DEFAULT_KEY);
645 
646 	free_bootmem_with_active_regions(0, max_pfn);
647 
648 	/*
649 	 * Reserve memory used for lowcore/command line/kernel image.
650 	 */
651 	reserve_bootmem(0, (unsigned long)_ehead, BOOTMEM_DEFAULT);
652 	reserve_bootmem((unsigned long)_stext,
653 			PFN_PHYS(start_pfn) - (unsigned long)_stext,
654 			BOOTMEM_DEFAULT);
655 	/*
656 	 * Reserve the bootmem bitmap itself as well. We do this in two
657 	 * steps (first step was init_bootmem()) because this catches
658 	 * the (very unlikely) case of us accidentally initializing the
659 	 * bootmem allocator with an invalid RAM area.
660 	 */
661 	reserve_bootmem(start_pfn << PAGE_SHIFT, bootmap_size,
662 			BOOTMEM_DEFAULT);
663 
664 #ifdef CONFIG_BLK_DEV_INITRD
665 	if (INITRD_START && INITRD_SIZE) {
666 		if (INITRD_START + INITRD_SIZE <= memory_end) {
667 			reserve_bootmem(INITRD_START, INITRD_SIZE,
668 					BOOTMEM_DEFAULT);
669 			initrd_start = INITRD_START;
670 			initrd_end = initrd_start + INITRD_SIZE;
671 		} else {
672 			printk("initrd extends beyond end of memory "
673 			       "(0x%08lx > 0x%08lx)\ndisabling initrd\n",
674 			       initrd_start + INITRD_SIZE, memory_end);
675 			initrd_start = initrd_end = 0;
676 		}
677 	}
678 #endif
679 }
680 
681 static __init unsigned int stfl(void)
682 {
683 	asm volatile(
684 		"	.insn	s,0xb2b10000,0(0)\n" /* stfl */
685 		"0:\n"
686 		EX_TABLE(0b,0b));
687 	return S390_lowcore.stfl_fac_list;
688 }
689 
690 static __init int stfle(unsigned long long *list, int doublewords)
691 {
692 	typedef struct { unsigned long long _[doublewords]; } addrtype;
693 	register unsigned long __nr asm("0") = doublewords - 1;
694 
695 	asm volatile(".insn s,0xb2b00000,%0" /* stfle */
696 		     : "=m" (*(addrtype *) list), "+d" (__nr) : : "cc");
697 	return __nr + 1;
698 }
699 
700 /*
701  * Setup hardware capabilities.
702  */
703 static void __init setup_hwcaps(void)
704 {
705 	static const int stfl_bits[6] = { 0, 2, 7, 17, 19, 21 };
706 	struct cpuinfo_S390 *cpuinfo = &S390_lowcore.cpu_data;
707 	unsigned long long facility_list_extended;
708 	unsigned int facility_list;
709 	int i;
710 
711 	facility_list = stfl();
712 	/*
713 	 * The store facility list bits numbers as found in the principles
714 	 * of operation are numbered with bit 1UL<<31 as number 0 to
715 	 * bit 1UL<<0 as number 31.
716 	 *   Bit 0: instructions named N3, "backported" to esa-mode
717 	 *   Bit 2: z/Architecture mode is active
718 	 *   Bit 7: the store-facility-list-extended facility is installed
719 	 *   Bit 17: the message-security assist is installed
720 	 *   Bit 19: the long-displacement facility is installed
721 	 *   Bit 21: the extended-immediate facility is installed
722 	 * These get translated to:
723 	 *   HWCAP_S390_ESAN3 bit 0, HWCAP_S390_ZARCH bit 1,
724 	 *   HWCAP_S390_STFLE bit 2, HWCAP_S390_MSA bit 3,
725 	 *   HWCAP_S390_LDISP bit 4, and HWCAP_S390_EIMM bit 5.
726 	 */
727 	for (i = 0; i < 6; i++)
728 		if (facility_list & (1UL << (31 - stfl_bits[i])))
729 			elf_hwcap |= 1UL << i;
730 
731 	/*
732 	 * Check for additional facilities with store-facility-list-extended.
733 	 * stfle stores doublewords (8 byte) with bit 1ULL<<63 as bit 0
734 	 * and 1ULL<<0 as bit 63. Bits 0-31 contain the same information
735 	 * as stored by stfl, bits 32-xxx contain additional facilities.
736 	 * How many facility words are stored depends on the number of
737 	 * doublewords passed to the instruction. The additional facilites
738 	 * are:
739 	 *   Bit 43: decimal floating point facility is installed
740 	 * translated to:
741 	 *   HWCAP_S390_DFP bit 6.
742 	 */
743 	if ((elf_hwcap & (1UL << 2)) &&
744 	    stfle(&facility_list_extended, 1) > 0) {
745 		if (facility_list_extended & (1ULL << (64 - 43)))
746 			elf_hwcap |= 1UL << 6;
747 	}
748 
749 	switch (cpuinfo->cpu_id.machine) {
750 	case 0x9672:
751 #if !defined(CONFIG_64BIT)
752 	default:	/* Use "g5" as default for 31 bit kernels. */
753 #endif
754 		strcpy(elf_platform, "g5");
755 		break;
756 	case 0x2064:
757 	case 0x2066:
758 #if defined(CONFIG_64BIT)
759 	default:	/* Use "z900" as default for 64 bit kernels. */
760 #endif
761 		strcpy(elf_platform, "z900");
762 		break;
763 	case 0x2084:
764 	case 0x2086:
765 		strcpy(elf_platform, "z990");
766 		break;
767 	case 0x2094:
768 		strcpy(elf_platform, "z9-109");
769 		break;
770 	}
771 }
772 
773 /*
774  * Setup function called from init/main.c just after the banner
775  * was printed.
776  */
777 
778 void __init
779 setup_arch(char **cmdline_p)
780 {
781         /*
782          * print what head.S has found out about the machine
783          */
784 #ifndef CONFIG_64BIT
785 	printk((MACHINE_IS_VM) ?
786 	       "We are running under VM (31 bit mode)\n" :
787 	       "We are running native (31 bit mode)\n");
788 	printk((MACHINE_HAS_IEEE) ?
789 	       "This machine has an IEEE fpu\n" :
790 	       "This machine has no IEEE fpu\n");
791 #else /* CONFIG_64BIT */
792 	printk((MACHINE_IS_VM) ?
793 	       "We are running under VM (64 bit mode)\n" :
794 	       "We are running native (64 bit mode)\n");
795 #endif /* CONFIG_64BIT */
796 
797 	/* Save unparsed command line copy for /proc/cmdline */
798 	strlcpy(boot_command_line, COMMAND_LINE, COMMAND_LINE_SIZE);
799 
800 	*cmdline_p = COMMAND_LINE;
801 	*(*cmdline_p + COMMAND_LINE_SIZE - 1) = '\0';
802 
803         ROOT_DEV = Root_RAM0;
804 
805 	init_mm.start_code = PAGE_OFFSET;
806 	init_mm.end_code = (unsigned long) &_etext;
807 	init_mm.end_data = (unsigned long) &_edata;
808 	init_mm.brk = (unsigned long) &_end;
809 
810 	if (MACHINE_HAS_MVCOS)
811 		memcpy(&uaccess, &uaccess_mvcos, sizeof(uaccess));
812 	else
813 		memcpy(&uaccess, &uaccess_std, sizeof(uaccess));
814 
815 	parse_early_param();
816 
817 	setup_ipl();
818 	setup_memory_end();
819 	setup_addressing_mode();
820 	setup_memory();
821 	setup_resources();
822 	setup_lowcore();
823 
824         cpu_init();
825         __cpu_logical_map[0] = S390_lowcore.cpu_data.cpu_addr;
826 
827 	/*
828 	 * Setup capabilities (ELF_HWCAP & ELF_PLATFORM).
829 	 */
830 	setup_hwcaps();
831 
832 	/*
833 	 * Create kernel page tables and switch to virtual addressing.
834 	 */
835         paging_init();
836 
837         /* Setup default console */
838 	conmode_default();
839 
840 	/* Setup zfcpdump support */
841 	setup_zfcpdump(console_devno);
842 }
843 
844 void __cpuinit print_cpu_info(struct cpuinfo_S390 *cpuinfo)
845 {
846    printk(KERN_INFO "cpu %d "
847 #ifdef CONFIG_SMP
848            "phys_idx=%d "
849 #endif
850            "vers=%02X ident=%06X machine=%04X unused=%04X\n",
851            cpuinfo->cpu_nr,
852 #ifdef CONFIG_SMP
853            cpuinfo->cpu_addr,
854 #endif
855            cpuinfo->cpu_id.version,
856            cpuinfo->cpu_id.ident,
857            cpuinfo->cpu_id.machine,
858            cpuinfo->cpu_id.unused);
859 }
860 
861 /*
862  * show_cpuinfo - Get information on one CPU for use by procfs.
863  */
864 
865 static int show_cpuinfo(struct seq_file *m, void *v)
866 {
867 	static const char *hwcap_str[7] = {
868 		"esan3", "zarch", "stfle", "msa", "ldisp", "eimm", "dfp"
869 	};
870         struct cpuinfo_S390 *cpuinfo;
871 	unsigned long n = (unsigned long) v - 1;
872 	int i;
873 
874 	s390_adjust_jiffies();
875 	preempt_disable();
876 	if (!n) {
877 		seq_printf(m, "vendor_id       : IBM/S390\n"
878 			       "# processors    : %i\n"
879 			       "bogomips per cpu: %lu.%02lu\n",
880 			       num_online_cpus(), loops_per_jiffy/(500000/HZ),
881 			       (loops_per_jiffy/(5000/HZ))%100);
882 		seq_puts(m, "features\t: ");
883 		for (i = 0; i < 7; i++)
884 			if (hwcap_str[i] && (elf_hwcap & (1UL << i)))
885 				seq_printf(m, "%s ", hwcap_str[i]);
886 		seq_puts(m, "\n");
887 	}
888 
889 	if (cpu_online(n)) {
890 #ifdef CONFIG_SMP
891 		if (smp_processor_id() == n)
892 			cpuinfo = &S390_lowcore.cpu_data;
893 		else
894 			cpuinfo = &lowcore_ptr[n]->cpu_data;
895 #else
896 		cpuinfo = &S390_lowcore.cpu_data;
897 #endif
898 		seq_printf(m, "processor %li: "
899 			       "version = %02X,  "
900 			       "identification = %06X,  "
901 			       "machine = %04X\n",
902 			       n, cpuinfo->cpu_id.version,
903 			       cpuinfo->cpu_id.ident,
904 			       cpuinfo->cpu_id.machine);
905 	}
906 	preempt_enable();
907         return 0;
908 }
909 
910 static void *c_start(struct seq_file *m, loff_t *pos)
911 {
912 	return *pos < NR_CPUS ? (void *)((unsigned long) *pos + 1) : NULL;
913 }
914 static void *c_next(struct seq_file *m, void *v, loff_t *pos)
915 {
916 	++*pos;
917 	return c_start(m, pos);
918 }
919 static void c_stop(struct seq_file *m, void *v)
920 {
921 }
922 const struct seq_operations cpuinfo_op = {
923 	.start	= c_start,
924 	.next	= c_next,
925 	.stop	= c_stop,
926 	.show	= show_cpuinfo,
927 };
928 
929