xref: /linux/arch/sh/kernel/process.c (revision 87c2ce3b9305b9b723faeedf6e32ef703ec9b33a)
1 /* $Id: process.c,v 1.28 2004/05/05 16:54:23 lethal Exp $
2  *
3  *  linux/arch/sh/kernel/process.c
4  *
5  *  Copyright (C) 1995  Linus Torvalds
6  *
7  *  SuperH version:  Copyright (C) 1999, 2000  Niibe Yutaka & Kaz Kojima
8  */
9 
10 /*
11  * This file handles the architecture-dependent parts of process handling..
12  */
13 
14 #include <linux/module.h>
15 #include <linux/unistd.h>
16 #include <linux/mm.h>
17 #include <linux/elfcore.h>
18 #include <linux/slab.h>
19 #include <linux/a.out.h>
20 #include <linux/ptrace.h>
21 #include <linux/platform.h>
22 #include <linux/kallsyms.h>
23 
24 #include <asm/io.h>
25 #include <asm/uaccess.h>
26 #include <asm/mmu_context.h>
27 #include <asm/elf.h>
28 #if defined(CONFIG_SH_HS7751RVOIP)
29 #include <asm/hs7751rvoip/hs7751rvoip.h>
30 #elif defined(CONFIG_SH_RTS7751R2D)
31 #include <asm/rts7751r2d/rts7751r2d.h>
32 #endif
33 
34 static int hlt_counter=0;
35 
36 int ubc_usercnt = 0;
37 
38 #define HARD_IDLE_TIMEOUT (HZ / 3)
39 
40 void disable_hlt(void)
41 {
42 	hlt_counter++;
43 }
44 
45 EXPORT_SYMBOL(disable_hlt);
46 
47 void enable_hlt(void)
48 {
49 	hlt_counter--;
50 }
51 
52 EXPORT_SYMBOL(enable_hlt);
53 
54 void cpu_idle(void)
55 {
56 	/* endless idle loop with no priority at all */
57 	while (1) {
58 		if (hlt_counter) {
59 			while (!need_resched())
60 				cpu_relax();
61 		} else {
62 			while (!need_resched())
63 				cpu_sleep();
64 		}
65 
66 		preempt_enable_no_resched();
67 		schedule();
68 		preempt_disable();
69 	}
70 }
71 
72 void machine_restart(char * __unused)
73 {
74 	/* SR.BL=1 and invoke address error to let CPU reset (manual reset) */
75 	asm volatile("ldc %0, sr\n\t"
76 		     "mov.l @%1, %0" : : "r" (0x10000000), "r" (0x80000001));
77 }
78 
79 void machine_halt(void)
80 {
81 #if defined(CONFIG_SH_HS7751RVOIP)
82 	unsigned short value;
83 
84 	value = ctrl_inw(PA_OUTPORTR);
85 	ctrl_outw((value & 0xffdf), PA_OUTPORTR);
86 #elif defined(CONFIG_SH_RTS7751R2D)
87 	ctrl_outw(0x0001, PA_POWOFF);
88 #endif
89 	while (1)
90 		cpu_sleep();
91 }
92 
93 void machine_power_off(void)
94 {
95 #if defined(CONFIG_SH_HS7751RVOIP)
96 	unsigned short value;
97 
98 	value = ctrl_inw(PA_OUTPORTR);
99 	ctrl_outw((value & 0xffdf), PA_OUTPORTR);
100 #elif defined(CONFIG_SH_RTS7751R2D)
101 	ctrl_outw(0x0001, PA_POWOFF);
102 #endif
103 }
104 
105 void show_regs(struct pt_regs * regs)
106 {
107 	printk("\n");
108 	printk("Pid : %d, Comm: %20s\n", current->pid, current->comm);
109 	print_symbol("PC is at %s\n", regs->pc);
110 	printk("PC  : %08lx SP  : %08lx SR  : %08lx ",
111 	       regs->pc, regs->regs[15], regs->sr);
112 #ifdef CONFIG_MMU
113 	printk("TEA : %08x    ", ctrl_inl(MMU_TEA));
114 #else
115 	printk("                  ");
116 #endif
117 	printk("%s\n", print_tainted());
118 
119 	printk("R0  : %08lx R1  : %08lx R2  : %08lx R3  : %08lx\n",
120 	       regs->regs[0],regs->regs[1],
121 	       regs->regs[2],regs->regs[3]);
122 	printk("R4  : %08lx R5  : %08lx R6  : %08lx R7  : %08lx\n",
123 	       regs->regs[4],regs->regs[5],
124 	       regs->regs[6],regs->regs[7]);
125 	printk("R8  : %08lx R9  : %08lx R10 : %08lx R11 : %08lx\n",
126 	       regs->regs[8],regs->regs[9],
127 	       regs->regs[10],regs->regs[11]);
128 	printk("R12 : %08lx R13 : %08lx R14 : %08lx\n",
129 	       regs->regs[12],regs->regs[13],
130 	       regs->regs[14]);
131 	printk("MACH: %08lx MACL: %08lx GBR : %08lx PR  : %08lx\n",
132 	       regs->mach, regs->macl, regs->gbr, regs->pr);
133 
134 	/*
135 	 * If we're in kernel mode, dump the stack too..
136 	 */
137 	if (!user_mode(regs)) {
138 		extern void show_task(unsigned long *sp);
139 		unsigned long sp = regs->regs[15];
140 
141 		show_task((unsigned long *)sp);
142 	}
143 }
144 
145 /*
146  * Create a kernel thread
147  */
148 
149 /*
150  * This is the mechanism for creating a new kernel thread.
151  *
152  */
153 extern void kernel_thread_helper(void);
154 __asm__(".align 5\n"
155 	"kernel_thread_helper:\n\t"
156 	"jsr	@r5\n\t"
157 	" nop\n\t"
158 	"mov.l	1f, r1\n\t"
159 	"jsr	@r1\n\t"
160 	" mov	r0, r4\n\t"
161 	".align 2\n\t"
162 	"1:.long do_exit");
163 
164 int kernel_thread(int (*fn)(void *), void * arg, unsigned long flags)
165 {	/* Don't use this in BL=1(cli).  Or else, CPU resets! */
166 	struct pt_regs regs;
167 
168 	memset(&regs, 0, sizeof(regs));
169 	regs.regs[4] = (unsigned long) arg;
170 	regs.regs[5] = (unsigned long) fn;
171 
172 	regs.pc = (unsigned long) kernel_thread_helper;
173 	regs.sr = (1 << 30);
174 
175 	/* Ok, create the new process.. */
176 	return do_fork(flags | CLONE_VM | CLONE_UNTRACED, 0, &regs, 0, NULL, NULL);
177 }
178 
179 /*
180  * Free current thread data structures etc..
181  */
182 void exit_thread(void)
183 {
184 	if (current->thread.ubc_pc) {
185 		current->thread.ubc_pc = 0;
186 		ubc_usercnt -= 1;
187 	}
188 }
189 
190 void flush_thread(void)
191 {
192 #if defined(CONFIG_SH_FPU)
193 	struct task_struct *tsk = current;
194 	struct pt_regs *regs = (struct pt_regs *)
195 				((unsigned long)tsk->thread_info
196 				 + THREAD_SIZE - sizeof(struct pt_regs)
197 				 - sizeof(unsigned long));
198 
199 	/* Forget lazy FPU state */
200 	clear_fpu(tsk, regs);
201 	clear_used_math();
202 #endif
203 }
204 
205 void release_thread(struct task_struct *dead_task)
206 {
207 	/* do nothing */
208 }
209 
210 /* Fill in the fpu structure for a core dump.. */
211 int dump_fpu(struct pt_regs *regs, elf_fpregset_t *fpu)
212 {
213 	int fpvalid = 0;
214 
215 #if defined(CONFIG_SH_FPU)
216 	struct task_struct *tsk = current;
217 
218 	fpvalid = !!tsk_used_math(tsk);
219 	if (fpvalid) {
220 		unlazy_fpu(tsk, regs);
221 		memcpy(fpu, &tsk->thread.fpu.hard, sizeof(*fpu));
222 	}
223 #endif
224 
225 	return fpvalid;
226 }
227 
228 /*
229  * Capture the user space registers if the task is not running (in user space)
230  */
231 int dump_task_regs(struct task_struct *tsk, elf_gregset_t *regs)
232 {
233 	struct pt_regs ptregs;
234 
235 	ptregs = *(struct pt_regs *)
236 		((unsigned long)tsk->thread_info + THREAD_SIZE
237 		 - sizeof(struct pt_regs)
238 #ifdef CONFIG_SH_DSP
239 		 - sizeof(struct pt_dspregs)
240 #endif
241 		 - sizeof(unsigned long));
242 	elf_core_copy_regs(regs, &ptregs);
243 
244 	return 1;
245 }
246 
247 int
248 dump_task_fpu (struct task_struct *tsk, elf_fpregset_t *fpu)
249 {
250 	int fpvalid = 0;
251 
252 #if defined(CONFIG_SH_FPU)
253 	fpvalid = !!tsk_used_math(tsk);
254 	if (fpvalid) {
255 		struct pt_regs *regs = (struct pt_regs *)
256 					((unsigned long)tsk->thread_info
257 					 + THREAD_SIZE - sizeof(struct pt_regs)
258 					 - sizeof(unsigned long));
259 		unlazy_fpu(tsk, regs);
260 		memcpy(fpu, &tsk->thread.fpu.hard, sizeof(*fpu));
261 	}
262 #endif
263 
264 	return fpvalid;
265 }
266 
267 asmlinkage void ret_from_fork(void);
268 
269 int copy_thread(int nr, unsigned long clone_flags, unsigned long usp,
270 		unsigned long unused,
271 		struct task_struct *p, struct pt_regs *regs)
272 {
273 	struct pt_regs *childregs;
274 #if defined(CONFIG_SH_FPU)
275 	struct task_struct *tsk = current;
276 
277 	unlazy_fpu(tsk, regs);
278 	p->thread.fpu = tsk->thread.fpu;
279 	copy_to_stopped_child_used_math(p);
280 #endif
281 
282 	childregs = ((struct pt_regs *)
283 		(THREAD_SIZE + (unsigned long) p->thread_info)
284 #ifdef CONFIG_SH_DSP
285 		- sizeof(struct pt_dspregs)
286 #endif
287 		- sizeof(unsigned long)) - 1;
288 	*childregs = *regs;
289 
290 	if (user_mode(regs)) {
291 		childregs->regs[15] = usp;
292 	} else {
293 		childregs->regs[15] = (unsigned long)p->thread_info + THREAD_SIZE;
294 	}
295         if (clone_flags & CLONE_SETTLS) {
296 		childregs->gbr = childregs->regs[0];
297 	}
298 	childregs->regs[0] = 0; /* Set return value for child */
299 
300 	p->thread.sp = (unsigned long) childregs;
301 	p->thread.pc = (unsigned long) ret_from_fork;
302 
303 	p->thread.ubc_pc = 0;
304 
305 	return 0;
306 }
307 
308 /* Tracing by user break controller.  */
309 static void
310 ubc_set_tracing(int asid, unsigned long pc)
311 {
312 	ctrl_outl(pc, UBC_BARA);
313 
314 	/* We don't have any ASID settings for the SH-2! */
315 	if (cpu_data->type != CPU_SH7604)
316 		ctrl_outb(asid, UBC_BASRA);
317 
318 	ctrl_outl(0, UBC_BAMRA);
319 
320 	if (cpu_data->type == CPU_SH7729) {
321 		ctrl_outw(BBR_INST | BBR_READ | BBR_CPU, UBC_BBRA);
322 		ctrl_outl(BRCR_PCBA | BRCR_PCTE, UBC_BRCR);
323 	} else {
324 		ctrl_outw(BBR_INST | BBR_READ, UBC_BBRA);
325 		ctrl_outw(BRCR_PCBA, UBC_BRCR);
326 	}
327 }
328 
329 /*
330  *	switch_to(x,y) should switch tasks from x to y.
331  *
332  */
333 struct task_struct *__switch_to(struct task_struct *prev, struct task_struct *next)
334 {
335 #if defined(CONFIG_SH_FPU)
336 	struct pt_regs *regs = (struct pt_regs *)
337 				((unsigned long)prev->thread_info
338 				 + THREAD_SIZE - sizeof(struct pt_regs)
339 				 - sizeof(unsigned long));
340 	unlazy_fpu(prev, regs);
341 #endif
342 
343 #ifdef CONFIG_PREEMPT
344 	{
345 		unsigned long flags;
346 		struct pt_regs *regs;
347 
348 		local_irq_save(flags);
349 		regs = (struct pt_regs *)
350 			((unsigned long)prev->thread_info
351 			 + THREAD_SIZE - sizeof(struct pt_regs)
352 #ifdef CONFIG_SH_DSP
353 			 - sizeof(struct pt_dspregs)
354 #endif
355 			 - sizeof(unsigned long));
356 		if (user_mode(regs) && regs->regs[15] >= 0xc0000000) {
357 			int offset = (int)regs->regs[15];
358 
359 			/* Reset stack pointer: clear critical region mark */
360 			regs->regs[15] = regs->regs[1];
361 			if (regs->pc < regs->regs[0])
362 				/* Go to rewind point */
363 				regs->pc = regs->regs[0] + offset;
364 		}
365 		local_irq_restore(flags);
366 	}
367 #endif
368 
369 	/*
370 	 * Restore the kernel mode register
371 	 *   	k7 (r7_bank1)
372 	 */
373 	asm volatile("ldc	%0, r7_bank"
374 		     : /* no output */
375 		     : "r" (next->thread_info));
376 
377 #ifdef CONFIG_MMU
378 	/* If no tasks are using the UBC, we're done */
379 	if (ubc_usercnt == 0)
380 		/* If no tasks are using the UBC, we're done */;
381 	else if (next->thread.ubc_pc && next->mm) {
382 		ubc_set_tracing(next->mm->context & MMU_CONTEXT_ASID_MASK,
383 				next->thread.ubc_pc);
384 	} else {
385 		ctrl_outw(0, UBC_BBRA);
386 		ctrl_outw(0, UBC_BBRB);
387 	}
388 #endif
389 
390 	return prev;
391 }
392 
393 asmlinkage int sys_fork(unsigned long r4, unsigned long r5,
394 			unsigned long r6, unsigned long r7,
395 			struct pt_regs regs)
396 {
397 #ifdef CONFIG_MMU
398 	return do_fork(SIGCHLD, regs.regs[15], &regs, 0, NULL, NULL);
399 #else
400 	/* fork almost works, enough to trick you into looking elsewhere :-( */
401 	return -EINVAL;
402 #endif
403 }
404 
405 asmlinkage int sys_clone(unsigned long clone_flags, unsigned long newsp,
406 			 unsigned long parent_tidptr,
407 			 unsigned long child_tidptr,
408 			 struct pt_regs regs)
409 {
410 	if (!newsp)
411 		newsp = regs.regs[15];
412 	return do_fork(clone_flags, newsp, &regs, 0,
413 			(int __user *)parent_tidptr, (int __user *)child_tidptr);
414 }
415 
416 /*
417  * This is trivial, and on the face of it looks like it
418  * could equally well be done in user mode.
419  *
420  * Not so, for quite unobvious reasons - register pressure.
421  * In user mode vfork() cannot have a stack frame, and if
422  * done by calling the "clone()" system call directly, you
423  * do not have enough call-clobbered registers to hold all
424  * the information you need.
425  */
426 asmlinkage int sys_vfork(unsigned long r4, unsigned long r5,
427 			 unsigned long r6, unsigned long r7,
428 			 struct pt_regs regs)
429 {
430 	return do_fork(CLONE_VFORK | CLONE_VM | SIGCHLD, regs.regs[15], &regs,
431 		       0, NULL, NULL);
432 }
433 
434 /*
435  * sys_execve() executes a new program.
436  */
437 asmlinkage int sys_execve(char *ufilename, char **uargv,
438 			  char **uenvp, unsigned long r7,
439 			  struct pt_regs regs)
440 {
441 	int error;
442 	char *filename;
443 
444 	filename = getname((char __user *)ufilename);
445 	error = PTR_ERR(filename);
446 	if (IS_ERR(filename))
447 		goto out;
448 
449 	error = do_execve(filename,
450 			  (char __user * __user *)uargv,
451 			  (char __user * __user *)uenvp,
452 			  &regs);
453 	if (error == 0) {
454 		task_lock(current);
455 		current->ptrace &= ~PT_DTRACE;
456 		task_unlock(current);
457 	}
458 	putname(filename);
459 out:
460 	return error;
461 }
462 
463 unsigned long get_wchan(struct task_struct *p)
464 {
465 	unsigned long schedule_frame;
466 	unsigned long pc;
467 
468 	if (!p || p == current || p->state == TASK_RUNNING)
469 		return 0;
470 
471 	/*
472 	 * The same comment as on the Alpha applies here, too ...
473 	 */
474 	pc = thread_saved_pc(p);
475 	if (in_sched_functions(pc)) {
476 		schedule_frame = ((unsigned long *)(long)p->thread.sp)[1];
477 		return (unsigned long)((unsigned long *)schedule_frame)[1];
478 	}
479 	return pc;
480 }
481 
482 asmlinkage void break_point_trap(unsigned long r4, unsigned long r5,
483 				 unsigned long r6, unsigned long r7,
484 				 struct pt_regs regs)
485 {
486 	/* Clear tracing.  */
487 	ctrl_outw(0, UBC_BBRA);
488 	ctrl_outw(0, UBC_BBRB);
489 	current->thread.ubc_pc = 0;
490 	ubc_usercnt -= 1;
491 
492 	force_sig(SIGTRAP, current);
493 }
494 
495 asmlinkage void break_point_trap_software(unsigned long r4, unsigned long r5,
496 					  unsigned long r6, unsigned long r7,
497 					  struct pt_regs regs)
498 {
499 	regs.pc -= 2;
500 	force_sig(SIGTRAP, current);
501 }
502