1 /* 2 * linux/arch/arm/kernel/traps.c 3 * 4 * Copyright (C) 1995-2002 Russell King 5 * Fragments that appear the same as linux/arch/i386/kernel/traps.c (C) Linus Torvalds 6 * 7 * This program is free software; you can redistribute it and/or modify 8 * it under the terms of the GNU General Public License version 2 as 9 * published by the Free Software Foundation. 10 * 11 * 'traps.c' handles hardware exceptions after we have saved some state in 12 * 'linux/arch/arm/lib/traps.S'. Mostly a debugging aid, but will probably 13 * kill the offending process. 14 */ 15 #include <linux/module.h> 16 #include <linux/signal.h> 17 #include <linux/spinlock.h> 18 #include <linux/personality.h> 19 #include <linux/ptrace.h> 20 #include <linux/kallsyms.h> 21 #include <linux/delay.h> 22 #include <linux/init.h> 23 24 #include <asm/atomic.h> 25 #include <asm/cacheflush.h> 26 #include <asm/system.h> 27 #include <asm/uaccess.h> 28 #include <asm/unistd.h> 29 #include <asm/traps.h> 30 #include <asm/io.h> 31 32 #include "ptrace.h" 33 #include "signal.h" 34 35 const char *processor_modes[]= 36 { "USER_26", "FIQ_26" , "IRQ_26" , "SVC_26" , "UK4_26" , "UK5_26" , "UK6_26" , "UK7_26" , 37 "UK8_26" , "UK9_26" , "UK10_26", "UK11_26", "UK12_26", "UK13_26", "UK14_26", "UK15_26", 38 "USER_32", "FIQ_32" , "IRQ_32" , "SVC_32" , "UK4_32" , "UK5_32" , "UK6_32" , "ABT_32" , 39 "UK8_32" , "UK9_32" , "UK10_32", "UND_32" , "UK12_32", "UK13_32", "UK14_32", "SYS_32" 40 }; 41 42 static const char *handler[]= { "prefetch abort", "data abort", "address exception", "interrupt" }; 43 44 #ifdef CONFIG_DEBUG_USER 45 unsigned int user_debug; 46 47 static int __init user_debug_setup(char *str) 48 { 49 get_option(&str, &user_debug); 50 return 1; 51 } 52 __setup("user_debug=", user_debug_setup); 53 #endif 54 55 void dump_backtrace_entry(unsigned long where, unsigned long from) 56 { 57 #ifdef CONFIG_KALLSYMS 58 printk("[<%08lx>] ", where); 59 print_symbol("(%s) ", where); 60 printk("from [<%08lx>] ", from); 61 print_symbol("(%s)\n", from); 62 #else 63 printk("Function entered at [<%08lx>] from [<%08lx>]\n", where, from); 64 #endif 65 } 66 67 /* 68 * Stack pointers should always be within the kernels view of 69 * physical memory. If it is not there, then we can't dump 70 * out any information relating to the stack. 71 */ 72 static int verify_stack(unsigned long sp) 73 { 74 if (sp < PAGE_OFFSET || (sp > (unsigned long)high_memory && high_memory != 0)) 75 return -EFAULT; 76 77 return 0; 78 } 79 80 /* 81 * Dump out the contents of some memory nicely... 82 */ 83 static void dump_mem(const char *str, unsigned long bottom, unsigned long top) 84 { 85 unsigned long p = bottom & ~31; 86 mm_segment_t fs; 87 int i; 88 89 /* 90 * We need to switch to kernel mode so that we can use __get_user 91 * to safely read from kernel space. Note that we now dump the 92 * code first, just in case the backtrace kills us. 93 */ 94 fs = get_fs(); 95 set_fs(KERNEL_DS); 96 97 printk("%s(0x%08lx to 0x%08lx)\n", str, bottom, top); 98 99 for (p = bottom & ~31; p < top;) { 100 printk("%04lx: ", p & 0xffff); 101 102 for (i = 0; i < 8; i++, p += 4) { 103 unsigned int val; 104 105 if (p < bottom || p >= top) 106 printk(" "); 107 else { 108 __get_user(val, (unsigned long *)p); 109 printk("%08x ", val); 110 } 111 } 112 printk ("\n"); 113 } 114 115 set_fs(fs); 116 } 117 118 static void dump_instr(struct pt_regs *regs) 119 { 120 unsigned long addr = instruction_pointer(regs); 121 const int thumb = thumb_mode(regs); 122 const int width = thumb ? 4 : 8; 123 mm_segment_t fs; 124 int i; 125 126 /* 127 * We need to switch to kernel mode so that we can use __get_user 128 * to safely read from kernel space. Note that we now dump the 129 * code first, just in case the backtrace kills us. 130 */ 131 fs = get_fs(); 132 set_fs(KERNEL_DS); 133 134 printk("Code: "); 135 for (i = -4; i < 1; i++) { 136 unsigned int val, bad; 137 138 if (thumb) 139 bad = __get_user(val, &((u16 *)addr)[i]); 140 else 141 bad = __get_user(val, &((u32 *)addr)[i]); 142 143 if (!bad) 144 printk(i == 0 ? "(%0*x) " : "%0*x ", width, val); 145 else { 146 printk("bad PC value."); 147 break; 148 } 149 } 150 printk("\n"); 151 152 set_fs(fs); 153 } 154 155 static void dump_backtrace(struct pt_regs *regs, struct task_struct *tsk) 156 { 157 unsigned int fp; 158 int ok = 1; 159 160 printk("Backtrace: "); 161 fp = regs->ARM_fp; 162 if (!fp) { 163 printk("no frame pointer"); 164 ok = 0; 165 } else if (verify_stack(fp)) { 166 printk("invalid frame pointer 0x%08x", fp); 167 ok = 0; 168 } else if (fp < (unsigned long)end_of_stack(tsk)) 169 printk("frame pointer underflow"); 170 printk("\n"); 171 172 if (ok) 173 c_backtrace(fp, processor_mode(regs)); 174 } 175 176 void dump_stack(void) 177 { 178 #ifdef CONFIG_DEBUG_ERRORS 179 __backtrace(); 180 #endif 181 } 182 183 EXPORT_SYMBOL(dump_stack); 184 185 void show_stack(struct task_struct *tsk, unsigned long *sp) 186 { 187 unsigned long fp; 188 189 if (!tsk) 190 tsk = current; 191 192 if (tsk != current) 193 fp = thread_saved_fp(tsk); 194 else 195 asm("mov %0, fp" : "=r" (fp) : : "cc"); 196 197 c_backtrace(fp, 0x10); 198 barrier(); 199 } 200 201 static void __die(const char *str, int err, struct thread_info *thread, struct pt_regs *regs) 202 { 203 struct task_struct *tsk = thread->task; 204 static int die_counter; 205 206 printk("Internal error: %s: %x [#%d]\n", str, err, ++die_counter); 207 print_modules(); 208 __show_regs(regs); 209 printk("Process %s (pid: %d, stack limit = 0x%p)\n", 210 tsk->comm, tsk->pid, thread + 1); 211 212 if (!user_mode(regs) || in_interrupt()) { 213 dump_mem("Stack: ", regs->ARM_sp, 214 THREAD_SIZE + (unsigned long)task_stack_page(tsk)); 215 dump_backtrace(regs, tsk); 216 dump_instr(regs); 217 } 218 } 219 220 DEFINE_SPINLOCK(die_lock); 221 222 /* 223 * This function is protected against re-entrancy. 224 */ 225 NORET_TYPE void die(const char *str, struct pt_regs *regs, int err) 226 { 227 struct thread_info *thread = current_thread_info(); 228 229 console_verbose(); 230 spin_lock_irq(&die_lock); 231 bust_spinlocks(1); 232 __die(str, err, thread, regs); 233 bust_spinlocks(0); 234 spin_unlock_irq(&die_lock); 235 236 if (panic_on_oops) 237 panic("Fatal exception"); 238 239 do_exit(SIGSEGV); 240 } 241 242 void notify_die(const char *str, struct pt_regs *regs, struct siginfo *info, 243 unsigned long err, unsigned long trap) 244 { 245 if (user_mode(regs)) { 246 current->thread.error_code = err; 247 current->thread.trap_no = trap; 248 249 force_sig_info(info->si_signo, info, current); 250 } else { 251 die(str, regs, err); 252 } 253 } 254 255 static LIST_HEAD(undef_hook); 256 static DEFINE_SPINLOCK(undef_lock); 257 258 void register_undef_hook(struct undef_hook *hook) 259 { 260 unsigned long flags; 261 262 spin_lock_irqsave(&undef_lock, flags); 263 list_add(&hook->node, &undef_hook); 264 spin_unlock_irqrestore(&undef_lock, flags); 265 } 266 267 void unregister_undef_hook(struct undef_hook *hook) 268 { 269 unsigned long flags; 270 271 spin_lock_irqsave(&undef_lock, flags); 272 list_del(&hook->node); 273 spin_unlock_irqrestore(&undef_lock, flags); 274 } 275 276 asmlinkage void do_undefinstr(struct pt_regs *regs) 277 { 278 unsigned int correction = thumb_mode(regs) ? 2 : 4; 279 unsigned int instr; 280 struct undef_hook *hook; 281 siginfo_t info; 282 void __user *pc; 283 284 /* 285 * According to the ARM ARM, PC is 2 or 4 bytes ahead, 286 * depending whether we're in Thumb mode or not. 287 * Correct this offset. 288 */ 289 regs->ARM_pc -= correction; 290 291 pc = (void __user *)instruction_pointer(regs); 292 if (thumb_mode(regs)) { 293 get_user(instr, (u16 __user *)pc); 294 } else { 295 get_user(instr, (u32 __user *)pc); 296 } 297 298 spin_lock_irq(&undef_lock); 299 list_for_each_entry(hook, &undef_hook, node) { 300 if ((instr & hook->instr_mask) == hook->instr_val && 301 (regs->ARM_cpsr & hook->cpsr_mask) == hook->cpsr_val) { 302 if (hook->fn(regs, instr) == 0) { 303 spin_unlock_irq(&undef_lock); 304 return; 305 } 306 } 307 } 308 spin_unlock_irq(&undef_lock); 309 310 #ifdef CONFIG_DEBUG_USER 311 if (user_debug & UDBG_UNDEFINED) { 312 printk(KERN_INFO "%s (%d): undefined instruction: pc=%p\n", 313 current->comm, current->pid, pc); 314 dump_instr(regs); 315 } 316 #endif 317 318 info.si_signo = SIGILL; 319 info.si_errno = 0; 320 info.si_code = ILL_ILLOPC; 321 info.si_addr = pc; 322 323 notify_die("Oops - undefined instruction", regs, &info, 0, 6); 324 } 325 326 asmlinkage void do_unexp_fiq (struct pt_regs *regs) 327 { 328 #ifndef CONFIG_IGNORE_FIQ 329 printk("Hmm. Unexpected FIQ received, but trying to continue\n"); 330 printk("You may have a hardware problem...\n"); 331 #endif 332 } 333 334 /* 335 * bad_mode handles the impossible case in the vectors. If you see one of 336 * these, then it's extremely serious, and could mean you have buggy hardware. 337 * It never returns, and never tries to sync. We hope that we can at least 338 * dump out some state information... 339 */ 340 asmlinkage void bad_mode(struct pt_regs *regs, int reason, int proc_mode) 341 { 342 console_verbose(); 343 344 printk(KERN_CRIT "Bad mode in %s handler detected: mode %s\n", 345 handler[reason], processor_modes[proc_mode]); 346 347 die("Oops - bad mode", regs, 0); 348 local_irq_disable(); 349 panic("bad mode"); 350 } 351 352 static int bad_syscall(int n, struct pt_regs *regs) 353 { 354 struct thread_info *thread = current_thread_info(); 355 siginfo_t info; 356 357 if (current->personality != PER_LINUX && 358 current->personality != PER_LINUX_32BIT && 359 thread->exec_domain->handler) { 360 thread->exec_domain->handler(n, regs); 361 return regs->ARM_r0; 362 } 363 364 #ifdef CONFIG_DEBUG_USER 365 if (user_debug & UDBG_SYSCALL) { 366 printk(KERN_ERR "[%d] %s: obsolete system call %08x.\n", 367 current->pid, current->comm, n); 368 dump_instr(regs); 369 } 370 #endif 371 372 info.si_signo = SIGILL; 373 info.si_errno = 0; 374 info.si_code = ILL_ILLTRP; 375 info.si_addr = (void __user *)instruction_pointer(regs) - 376 (thumb_mode(regs) ? 2 : 4); 377 378 notify_die("Oops - bad syscall", regs, &info, n, 0); 379 380 return regs->ARM_r0; 381 } 382 383 static inline void 384 do_cache_op(unsigned long start, unsigned long end, int flags) 385 { 386 struct vm_area_struct *vma; 387 388 if (end < start || flags) 389 return; 390 391 vma = find_vma(current->active_mm, start); 392 if (vma && vma->vm_start < end) { 393 if (start < vma->vm_start) 394 start = vma->vm_start; 395 if (end > vma->vm_end) 396 end = vma->vm_end; 397 398 flush_cache_user_range(vma, start, end); 399 } 400 } 401 402 /* 403 * Handle all unrecognised system calls. 404 * 0x9f0000 - 0x9fffff are some more esoteric system calls 405 */ 406 #define NR(x) ((__ARM_NR_##x) - __ARM_NR_BASE) 407 asmlinkage int arm_syscall(int no, struct pt_regs *regs) 408 { 409 struct thread_info *thread = current_thread_info(); 410 siginfo_t info; 411 412 if ((no >> 16) != (__ARM_NR_BASE>> 16)) 413 return bad_syscall(no, regs); 414 415 switch (no & 0xffff) { 416 case 0: /* branch through 0 */ 417 info.si_signo = SIGSEGV; 418 info.si_errno = 0; 419 info.si_code = SEGV_MAPERR; 420 info.si_addr = NULL; 421 422 notify_die("branch through zero", regs, &info, 0, 0); 423 return 0; 424 425 case NR(breakpoint): /* SWI BREAK_POINT */ 426 regs->ARM_pc -= thumb_mode(regs) ? 2 : 4; 427 ptrace_break(current, regs); 428 return regs->ARM_r0; 429 430 /* 431 * Flush a region from virtual address 'r0' to virtual address 'r1' 432 * _exclusive_. There is no alignment requirement on either address; 433 * user space does not need to know the hardware cache layout. 434 * 435 * r2 contains flags. It should ALWAYS be passed as ZERO until it 436 * is defined to be something else. For now we ignore it, but may 437 * the fires of hell burn in your belly if you break this rule. ;) 438 * 439 * (at a later date, we may want to allow this call to not flush 440 * various aspects of the cache. Passing '0' will guarantee that 441 * everything necessary gets flushed to maintain consistency in 442 * the specified region). 443 */ 444 case NR(cacheflush): 445 do_cache_op(regs->ARM_r0, regs->ARM_r1, regs->ARM_r2); 446 return 0; 447 448 case NR(usr26): 449 if (!(elf_hwcap & HWCAP_26BIT)) 450 break; 451 regs->ARM_cpsr &= ~MODE32_BIT; 452 return regs->ARM_r0; 453 454 case NR(usr32): 455 if (!(elf_hwcap & HWCAP_26BIT)) 456 break; 457 regs->ARM_cpsr |= MODE32_BIT; 458 return regs->ARM_r0; 459 460 case NR(set_tls): 461 thread->tp_value = regs->ARM_r0; 462 #if defined(CONFIG_HAS_TLS_REG) 463 asm ("mcr p15, 0, %0, c13, c0, 3" : : "r" (regs->ARM_r0) ); 464 #elif !defined(CONFIG_TLS_REG_EMUL) 465 /* 466 * User space must never try to access this directly. 467 * Expect your app to break eventually if you do so. 468 * The user helper at 0xffff0fe0 must be used instead. 469 * (see entry-armv.S for details) 470 */ 471 *((unsigned int *)0xffff0ff0) = regs->ARM_r0; 472 #endif 473 return 0; 474 475 #ifdef CONFIG_NEEDS_SYSCALL_FOR_CMPXCHG 476 /* 477 * Atomically store r1 in *r2 if *r2 is equal to r0 for user space. 478 * Return zero in r0 if *MEM was changed or non-zero if no exchange 479 * happened. Also set the user C flag accordingly. 480 * If access permissions have to be fixed up then non-zero is 481 * returned and the operation has to be re-attempted. 482 * 483 * *NOTE*: This is a ghost syscall private to the kernel. Only the 484 * __kuser_cmpxchg code in entry-armv.S should be aware of its 485 * existence. Don't ever use this from user code. 486 */ 487 case 0xfff0: 488 { 489 extern void do_DataAbort(unsigned long addr, unsigned int fsr, 490 struct pt_regs *regs); 491 unsigned long val; 492 unsigned long addr = regs->ARM_r2; 493 struct mm_struct *mm = current->mm; 494 pgd_t *pgd; pmd_t *pmd; pte_t *pte; 495 spinlock_t *ptl; 496 497 regs->ARM_cpsr &= ~PSR_C_BIT; 498 down_read(&mm->mmap_sem); 499 pgd = pgd_offset(mm, addr); 500 if (!pgd_present(*pgd)) 501 goto bad_access; 502 pmd = pmd_offset(pgd, addr); 503 if (!pmd_present(*pmd)) 504 goto bad_access; 505 pte = pte_offset_map_lock(mm, pmd, addr, &ptl); 506 if (!pte_present(*pte) || !pte_dirty(*pte)) { 507 pte_unmap_unlock(pte, ptl); 508 goto bad_access; 509 } 510 val = *(unsigned long *)addr; 511 val -= regs->ARM_r0; 512 if (val == 0) { 513 *(unsigned long *)addr = regs->ARM_r1; 514 regs->ARM_cpsr |= PSR_C_BIT; 515 } 516 pte_unmap_unlock(pte, ptl); 517 up_read(&mm->mmap_sem); 518 return val; 519 520 bad_access: 521 up_read(&mm->mmap_sem); 522 /* simulate a write access fault */ 523 do_DataAbort(addr, 15 + (1 << 11), regs); 524 return -1; 525 } 526 #endif 527 528 default: 529 /* Calls 9f00xx..9f07ff are defined to return -ENOSYS 530 if not implemented, rather than raising SIGILL. This 531 way the calling program can gracefully determine whether 532 a feature is supported. */ 533 if (no <= 0x7ff) 534 return -ENOSYS; 535 break; 536 } 537 #ifdef CONFIG_DEBUG_USER 538 /* 539 * experience shows that these seem to indicate that 540 * something catastrophic has happened 541 */ 542 if (user_debug & UDBG_SYSCALL) { 543 printk("[%d] %s: arm syscall %d\n", 544 current->pid, current->comm, no); 545 dump_instr(regs); 546 if (user_mode(regs)) { 547 __show_regs(regs); 548 c_backtrace(regs->ARM_fp, processor_mode(regs)); 549 } 550 } 551 #endif 552 info.si_signo = SIGILL; 553 info.si_errno = 0; 554 info.si_code = ILL_ILLTRP; 555 info.si_addr = (void __user *)instruction_pointer(regs) - 556 (thumb_mode(regs) ? 2 : 4); 557 558 notify_die("Oops - bad syscall(2)", regs, &info, no, 0); 559 return 0; 560 } 561 562 #ifdef CONFIG_TLS_REG_EMUL 563 564 /* 565 * We might be running on an ARMv6+ processor which should have the TLS 566 * register but for some reason we can't use it, or maybe an SMP system 567 * using a pre-ARMv6 processor (there are apparently a few prototypes like 568 * that in existence) and therefore access to that register must be 569 * emulated. 570 */ 571 572 static int get_tp_trap(struct pt_regs *regs, unsigned int instr) 573 { 574 int reg = (instr >> 12) & 15; 575 if (reg == 15) 576 return 1; 577 regs->uregs[reg] = current_thread_info()->tp_value; 578 regs->ARM_pc += 4; 579 return 0; 580 } 581 582 static struct undef_hook arm_mrc_hook = { 583 .instr_mask = 0x0fff0fff, 584 .instr_val = 0x0e1d0f70, 585 .cpsr_mask = PSR_T_BIT, 586 .cpsr_val = 0, 587 .fn = get_tp_trap, 588 }; 589 590 static int __init arm_mrc_hook_init(void) 591 { 592 register_undef_hook(&arm_mrc_hook); 593 return 0; 594 } 595 596 late_initcall(arm_mrc_hook_init); 597 598 #endif 599 600 void __bad_xchg(volatile void *ptr, int size) 601 { 602 printk("xchg: bad data size: pc 0x%p, ptr 0x%p, size %d\n", 603 __builtin_return_address(0), ptr, size); 604 BUG(); 605 } 606 EXPORT_SYMBOL(__bad_xchg); 607 608 /* 609 * A data abort trap was taken, but we did not handle the instruction. 610 * Try to abort the user program, or panic if it was the kernel. 611 */ 612 asmlinkage void 613 baddataabort(int code, unsigned long instr, struct pt_regs *regs) 614 { 615 unsigned long addr = instruction_pointer(regs); 616 siginfo_t info; 617 618 #ifdef CONFIG_DEBUG_USER 619 if (user_debug & UDBG_BADABORT) { 620 printk(KERN_ERR "[%d] %s: bad data abort: code %d instr 0x%08lx\n", 621 current->pid, current->comm, code, instr); 622 dump_instr(regs); 623 show_pte(current->mm, addr); 624 } 625 #endif 626 627 info.si_signo = SIGILL; 628 info.si_errno = 0; 629 info.si_code = ILL_ILLOPC; 630 info.si_addr = (void __user *)addr; 631 632 notify_die("unknown data abort code", regs, &info, instr, 0); 633 } 634 635 void __attribute__((noreturn)) __bug(const char *file, int line) 636 { 637 printk(KERN_CRIT"kernel BUG at %s:%d!\n", file, line); 638 *(int *)0 = 0; 639 640 /* Avoid "noreturn function does return" */ 641 for (;;); 642 } 643 EXPORT_SYMBOL(__bug); 644 645 void __readwrite_bug(const char *fn) 646 { 647 printk("%s called, but not implemented\n", fn); 648 BUG(); 649 } 650 EXPORT_SYMBOL(__readwrite_bug); 651 652 void __pte_error(const char *file, int line, unsigned long val) 653 { 654 printk("%s:%d: bad pte %08lx.\n", file, line, val); 655 } 656 657 void __pmd_error(const char *file, int line, unsigned long val) 658 { 659 printk("%s:%d: bad pmd %08lx.\n", file, line, val); 660 } 661 662 void __pgd_error(const char *file, int line, unsigned long val) 663 { 664 printk("%s:%d: bad pgd %08lx.\n", file, line, val); 665 } 666 667 asmlinkage void __div0(void) 668 { 669 printk("Division by zero in kernel.\n"); 670 dump_stack(); 671 } 672 EXPORT_SYMBOL(__div0); 673 674 void abort(void) 675 { 676 BUG(); 677 678 /* if that doesn't kill us, halt */ 679 panic("Oops failed to kill thread"); 680 } 681 EXPORT_SYMBOL(abort); 682 683 void __init trap_init(void) 684 { 685 unsigned long vectors = CONFIG_VECTORS_BASE; 686 extern char __stubs_start[], __stubs_end[]; 687 extern char __vectors_start[], __vectors_end[]; 688 extern char __kuser_helper_start[], __kuser_helper_end[]; 689 int kuser_sz = __kuser_helper_end - __kuser_helper_start; 690 691 /* 692 * Copy the vectors, stubs and kuser helpers (in entry-armv.S) 693 * into the vector page, mapped at 0xffff0000, and ensure these 694 * are visible to the instruction stream. 695 */ 696 memcpy((void *)vectors, __vectors_start, __vectors_end - __vectors_start); 697 memcpy((void *)vectors + 0x200, __stubs_start, __stubs_end - __stubs_start); 698 memcpy((void *)vectors + 0x1000 - kuser_sz, __kuser_helper_start, kuser_sz); 699 700 /* 701 * Copy signal return handlers into the vector page, and 702 * set sigreturn to be a pointer to these. 703 */ 704 memcpy((void *)KERN_SIGRETURN_CODE, sigreturn_codes, 705 sizeof(sigreturn_codes)); 706 707 flush_icache_range(vectors, vectors + PAGE_SIZE); 708 modify_domain(DOMAIN_USER, DOMAIN_CLIENT); 709 } 710