1/* 2 * Low-level exception handling code 3 * 4 * Copyright (C) 2012 ARM Ltd. 5 * Authors: Catalin Marinas <catalin.marinas@arm.com> 6 * Will Deacon <will.deacon@arm.com> 7 * 8 * This program is free software; you can redistribute it and/or modify 9 * it under the terms of the GNU General Public License version 2 as 10 * published by the Free Software Foundation. 11 * 12 * This program is distributed in the hope that it will be useful, 13 * but WITHOUT ANY WARRANTY; without even the implied warranty of 14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 15 * GNU General Public License for more details. 16 * 17 * You should have received a copy of the GNU General Public License 18 * along with this program. If not, see <http://www.gnu.org/licenses/>. 19 */ 20 21#include <linux/init.h> 22#include <linux/linkage.h> 23 24#include <asm/assembler.h> 25#include <asm/asm-offsets.h> 26#include <asm/errno.h> 27#include <asm/thread_info.h> 28#include <asm/unistd.h> 29#include <asm/unistd32.h> 30 31/* 32 * Bad Abort numbers 33 *----------------- 34 */ 35#define BAD_SYNC 0 36#define BAD_IRQ 1 37#define BAD_FIQ 2 38#define BAD_ERROR 3 39 40 .macro kernel_entry, el, regsize = 64 41 sub sp, sp, #S_FRAME_SIZE - S_LR // room for LR, SP, SPSR, ELR 42 .if \regsize == 32 43 mov w0, w0 // zero upper 32 bits of x0 44 .endif 45 push x28, x29 46 push x26, x27 47 push x24, x25 48 push x22, x23 49 push x20, x21 50 push x18, x19 51 push x16, x17 52 push x14, x15 53 push x12, x13 54 push x10, x11 55 push x8, x9 56 push x6, x7 57 push x4, x5 58 push x2, x3 59 push x0, x1 60 .if \el == 0 61 mrs x21, sp_el0 62 .else 63 add x21, sp, #S_FRAME_SIZE 64 .endif 65 mrs x22, elr_el1 66 mrs x23, spsr_el1 67 stp lr, x21, [sp, #S_LR] 68 stp x22, x23, [sp, #S_PC] 69 70 /* 71 * Set syscallno to -1 by default (overridden later if real syscall). 72 */ 73 .if \el == 0 74 mvn x21, xzr 75 str x21, [sp, #S_SYSCALLNO] 76 .endif 77 78 /* 79 * Registers that may be useful after this macro is invoked: 80 * 81 * x21 - aborted SP 82 * x22 - aborted PC 83 * x23 - aborted PSTATE 84 */ 85 .endm 86 87 .macro kernel_exit, el, ret = 0 88 ldp x21, x22, [sp, #S_PC] // load ELR, SPSR 89 .if \el == 0 90 ldr x23, [sp, #S_SP] // load return stack pointer 91 .endif 92 .if \ret 93 ldr x1, [sp, #S_X1] // preserve x0 (syscall return) 94 add sp, sp, S_X2 95 .else 96 pop x0, x1 97 .endif 98 pop x2, x3 // load the rest of the registers 99 pop x4, x5 100 pop x6, x7 101 pop x8, x9 102 msr elr_el1, x21 // set up the return data 103 msr spsr_el1, x22 104 .if \el == 0 105 msr sp_el0, x23 106 .endif 107 pop x10, x11 108 pop x12, x13 109 pop x14, x15 110 pop x16, x17 111 pop x18, x19 112 pop x20, x21 113 pop x22, x23 114 pop x24, x25 115 pop x26, x27 116 pop x28, x29 117 ldr lr, [sp], #S_FRAME_SIZE - S_LR // load LR and restore SP 118 eret // return to kernel 119 .endm 120 121 .macro get_thread_info, rd 122 mov \rd, sp 123 and \rd, \rd, #~((1 << 13) - 1) // top of 8K stack 124 .endm 125 126/* 127 * These are the registers used in the syscall handler, and allow us to 128 * have in theory up to 7 arguments to a function - x0 to x6. 129 * 130 * x7 is reserved for the system call number in 32-bit mode. 131 */ 132sc_nr .req x25 // number of system calls 133scno .req x26 // syscall number 134stbl .req x27 // syscall table pointer 135tsk .req x28 // current thread_info 136 137/* 138 * Interrupt handling. 139 */ 140 .macro irq_handler 141 ldr x1, handle_arch_irq 142 mov x0, sp 143 blr x1 144 .endm 145 146 .text 147 148/* 149 * Exception vectors. 150 */ 151 .macro ventry label 152 .align 7 153 b \label 154 .endm 155 156 .align 11 157ENTRY(vectors) 158 ventry el1_sync_invalid // Synchronous EL1t 159 ventry el1_irq_invalid // IRQ EL1t 160 ventry el1_fiq_invalid // FIQ EL1t 161 ventry el1_error_invalid // Error EL1t 162 163 ventry el1_sync // Synchronous EL1h 164 ventry el1_irq // IRQ EL1h 165 ventry el1_fiq_invalid // FIQ EL1h 166 ventry el1_error_invalid // Error EL1h 167 168 ventry el0_sync // Synchronous 64-bit EL0 169 ventry el0_irq // IRQ 64-bit EL0 170 ventry el0_fiq_invalid // FIQ 64-bit EL0 171 ventry el0_error_invalid // Error 64-bit EL0 172 173#ifdef CONFIG_COMPAT 174 ventry el0_sync_compat // Synchronous 32-bit EL0 175 ventry el0_irq_compat // IRQ 32-bit EL0 176 ventry el0_fiq_invalid_compat // FIQ 32-bit EL0 177 ventry el0_error_invalid_compat // Error 32-bit EL0 178#else 179 ventry el0_sync_invalid // Synchronous 32-bit EL0 180 ventry el0_irq_invalid // IRQ 32-bit EL0 181 ventry el0_fiq_invalid // FIQ 32-bit EL0 182 ventry el0_error_invalid // Error 32-bit EL0 183#endif 184END(vectors) 185 186/* 187 * Invalid mode handlers 188 */ 189 .macro inv_entry, el, reason, regsize = 64 190 kernel_entry el, \regsize 191 mov x0, sp 192 mov x1, #\reason 193 mrs x2, esr_el1 194 b bad_mode 195 .endm 196 197el0_sync_invalid: 198 inv_entry 0, BAD_SYNC 199ENDPROC(el0_sync_invalid) 200 201el0_irq_invalid: 202 inv_entry 0, BAD_IRQ 203ENDPROC(el0_irq_invalid) 204 205el0_fiq_invalid: 206 inv_entry 0, BAD_FIQ 207ENDPROC(el0_fiq_invalid) 208 209el0_error_invalid: 210 inv_entry 0, BAD_ERROR 211ENDPROC(el0_error_invalid) 212 213#ifdef CONFIG_COMPAT 214el0_fiq_invalid_compat: 215 inv_entry 0, BAD_FIQ, 32 216ENDPROC(el0_fiq_invalid_compat) 217 218el0_error_invalid_compat: 219 inv_entry 0, BAD_ERROR, 32 220ENDPROC(el0_error_invalid_compat) 221#endif 222 223el1_sync_invalid: 224 inv_entry 1, BAD_SYNC 225ENDPROC(el1_sync_invalid) 226 227el1_irq_invalid: 228 inv_entry 1, BAD_IRQ 229ENDPROC(el1_irq_invalid) 230 231el1_fiq_invalid: 232 inv_entry 1, BAD_FIQ 233ENDPROC(el1_fiq_invalid) 234 235el1_error_invalid: 236 inv_entry 1, BAD_ERROR 237ENDPROC(el1_error_invalid) 238 239/* 240 * EL1 mode handlers. 241 */ 242 .align 6 243el1_sync: 244 kernel_entry 1 245 mrs x1, esr_el1 // read the syndrome register 246 lsr x24, x1, #26 // exception class 247 cmp x24, #0x25 // data abort in EL1 248 b.eq el1_da 249 cmp x24, #0x18 // configurable trap 250 b.eq el1_undef 251 cmp x24, #0x26 // stack alignment exception 252 b.eq el1_sp_pc 253 cmp x24, #0x22 // pc alignment exception 254 b.eq el1_sp_pc 255 cmp x24, #0x00 // unknown exception in EL1 256 b.eq el1_undef 257 cmp x24, #0x30 // debug exception in EL1 258 b.ge el1_dbg 259 b el1_inv 260el1_da: 261 /* 262 * Data abort handling 263 */ 264 mrs x0, far_el1 265 enable_dbg_if_not_stepping x2 266 // re-enable interrupts if they were enabled in the aborted context 267 tbnz x23, #7, 1f // PSR_I_BIT 268 enable_irq 2691: 270 mov x2, sp // struct pt_regs 271 bl do_mem_abort 272 273 // disable interrupts before pulling preserved data off the stack 274 disable_irq 275 kernel_exit 1 276el1_sp_pc: 277 /* 278 * Stack or PC alignment exception handling 279 */ 280 mrs x0, far_el1 281 mov x1, x25 282 mov x2, sp 283 b do_sp_pc_abort 284el1_undef: 285 /* 286 * Undefined instruction 287 */ 288 mov x0, sp 289 b do_undefinstr 290el1_dbg: 291 /* 292 * Debug exception handling 293 */ 294 tbz x24, #0, el1_inv // EL1 only 295 mrs x0, far_el1 296 mov x2, sp // struct pt_regs 297 bl do_debug_exception 298 299 kernel_exit 1 300el1_inv: 301 // TODO: add support for undefined instructions in kernel mode 302 mov x0, sp 303 mov x1, #BAD_SYNC 304 mrs x2, esr_el1 305 b bad_mode 306ENDPROC(el1_sync) 307 308 .align 6 309el1_irq: 310 kernel_entry 1 311 enable_dbg_if_not_stepping x0 312#ifdef CONFIG_TRACE_IRQFLAGS 313 bl trace_hardirqs_off 314#endif 315#ifdef CONFIG_PREEMPT 316 get_thread_info tsk 317 ldr x24, [tsk, #TI_PREEMPT] // get preempt count 318 add x0, x24, #1 // increment it 319 str x0, [tsk, #TI_PREEMPT] 320#endif 321 irq_handler 322#ifdef CONFIG_PREEMPT 323 str x24, [tsk, #TI_PREEMPT] // restore preempt count 324 cbnz x24, 1f // preempt count != 0 325 ldr x0, [tsk, #TI_FLAGS] // get flags 326 tbz x0, #TIF_NEED_RESCHED, 1f // needs rescheduling? 327 bl el1_preempt 3281: 329#endif 330#ifdef CONFIG_TRACE_IRQFLAGS 331 bl trace_hardirqs_on 332#endif 333 kernel_exit 1 334ENDPROC(el1_irq) 335 336#ifdef CONFIG_PREEMPT 337el1_preempt: 338 mov x24, lr 3391: enable_dbg 340 bl preempt_schedule_irq // irq en/disable is done inside 341 ldr x0, [tsk, #TI_FLAGS] // get new tasks TI_FLAGS 342 tbnz x0, #TIF_NEED_RESCHED, 1b // needs rescheduling? 343 ret x24 344#endif 345 346/* 347 * EL0 mode handlers. 348 */ 349 .align 6 350el0_sync: 351 kernel_entry 0 352 mrs x25, esr_el1 // read the syndrome register 353 lsr x24, x25, #26 // exception class 354 cmp x24, #0x15 // SVC in 64-bit state 355 b.eq el0_svc 356 adr lr, ret_from_exception 357 cmp x24, #0x24 // data abort in EL0 358 b.eq el0_da 359 cmp x24, #0x20 // instruction abort in EL0 360 b.eq el0_ia 361 cmp x24, #0x07 // FP/ASIMD access 362 b.eq el0_fpsimd_acc 363 cmp x24, #0x2c // FP/ASIMD exception 364 b.eq el0_fpsimd_exc 365 cmp x24, #0x18 // configurable trap 366 b.eq el0_undef 367 cmp x24, #0x26 // stack alignment exception 368 b.eq el0_sp_pc 369 cmp x24, #0x22 // pc alignment exception 370 b.eq el0_sp_pc 371 cmp x24, #0x00 // unknown exception in EL0 372 b.eq el0_undef 373 cmp x24, #0x30 // debug exception in EL0 374 b.ge el0_dbg 375 b el0_inv 376 377#ifdef CONFIG_COMPAT 378 .align 6 379el0_sync_compat: 380 kernel_entry 0, 32 381 mrs x25, esr_el1 // read the syndrome register 382 lsr x24, x25, #26 // exception class 383 cmp x24, #0x11 // SVC in 32-bit state 384 b.eq el0_svc_compat 385 adr lr, ret_from_exception 386 cmp x24, #0x24 // data abort in EL0 387 b.eq el0_da 388 cmp x24, #0x20 // instruction abort in EL0 389 b.eq el0_ia 390 cmp x24, #0x07 // FP/ASIMD access 391 b.eq el0_fpsimd_acc 392 cmp x24, #0x28 // FP/ASIMD exception 393 b.eq el0_fpsimd_exc 394 cmp x24, #0x00 // unknown exception in EL0 395 b.eq el0_undef 396 cmp x24, #0x30 // debug exception in EL0 397 b.ge el0_dbg 398 b el0_inv 399el0_svc_compat: 400 /* 401 * AArch32 syscall handling 402 */ 403 adr stbl, compat_sys_call_table // load compat syscall table pointer 404 uxtw scno, w7 // syscall number in w7 (r7) 405 mov sc_nr, #__NR_compat_syscalls 406 b el0_svc_naked 407 408 .align 6 409el0_irq_compat: 410 kernel_entry 0, 32 411 b el0_irq_naked 412#endif 413 414el0_da: 415 /* 416 * Data abort handling 417 */ 418 mrs x0, far_el1 419 disable_step x1 420 isb 421 enable_dbg 422 // enable interrupts before calling the main handler 423 enable_irq 424 mov x1, x25 425 mov x2, sp 426 b do_mem_abort 427el0_ia: 428 /* 429 * Instruction abort handling 430 */ 431 mrs x0, far_el1 432 disable_step x1 433 isb 434 enable_dbg 435 // enable interrupts before calling the main handler 436 enable_irq 437 orr x1, x25, #1 << 24 // use reserved ISS bit for instruction aborts 438 mov x2, sp 439 b do_mem_abort 440el0_fpsimd_acc: 441 /* 442 * Floating Point or Advanced SIMD access 443 */ 444 mov x0, x25 445 mov x1, sp 446 b do_fpsimd_acc 447el0_fpsimd_exc: 448 /* 449 * Floating Point or Advanced SIMD exception 450 */ 451 mov x0, x25 452 mov x1, sp 453 b do_fpsimd_exc 454el0_sp_pc: 455 /* 456 * Stack or PC alignment exception handling 457 */ 458 mrs x0, far_el1 459 disable_step x1 460 isb 461 enable_dbg 462 // enable interrupts before calling the main handler 463 enable_irq 464 mov x1, x25 465 mov x2, sp 466 b do_sp_pc_abort 467el0_undef: 468 /* 469 * Undefined instruction 470 */ 471 mov x0, sp 472 b do_undefinstr 473el0_dbg: 474 /* 475 * Debug exception handling 476 */ 477 tbnz x24, #0, el0_inv // EL0 only 478 mrs x0, far_el1 479 disable_step x1 480 mov x1, x25 481 mov x2, sp 482 b do_debug_exception 483el0_inv: 484 mov x0, sp 485 mov x1, #BAD_SYNC 486 mrs x2, esr_el1 487 b bad_mode 488ENDPROC(el0_sync) 489 490 .align 6 491el0_irq: 492 kernel_entry 0 493el0_irq_naked: 494 disable_step x1 495 isb 496 enable_dbg 497#ifdef CONFIG_TRACE_IRQFLAGS 498 bl trace_hardirqs_off 499#endif 500 get_thread_info tsk 501#ifdef CONFIG_PREEMPT 502 ldr x24, [tsk, #TI_PREEMPT] // get preempt count 503 add x23, x24, #1 // increment it 504 str x23, [tsk, #TI_PREEMPT] 505#endif 506 irq_handler 507#ifdef CONFIG_PREEMPT 508 ldr x0, [tsk, #TI_PREEMPT] 509 str x24, [tsk, #TI_PREEMPT] 510 cmp x0, x23 511 b.eq 1f 512 mov x1, #0 513 str x1, [x1] // BUG 5141: 515#endif 516#ifdef CONFIG_TRACE_IRQFLAGS 517 bl trace_hardirqs_on 518#endif 519 b ret_to_user 520ENDPROC(el0_irq) 521 522/* 523 * This is the return code to user mode for abort handlers 524 */ 525ret_from_exception: 526 get_thread_info tsk 527 b ret_to_user 528ENDPROC(ret_from_exception) 529 530/* 531 * Register switch for AArch64. The callee-saved registers need to be saved 532 * and restored. On entry: 533 * x0 = previous task_struct (must be preserved across the switch) 534 * x1 = next task_struct 535 * Previous and next are guaranteed not to be the same. 536 * 537 */ 538ENTRY(cpu_switch_to) 539 add x8, x0, #THREAD_CPU_CONTEXT 540 mov x9, sp 541 stp x19, x20, [x8], #16 // store callee-saved registers 542 stp x21, x22, [x8], #16 543 stp x23, x24, [x8], #16 544 stp x25, x26, [x8], #16 545 stp x27, x28, [x8], #16 546 stp x29, x9, [x8], #16 547 str lr, [x8] 548 add x8, x1, #THREAD_CPU_CONTEXT 549 ldp x19, x20, [x8], #16 // restore callee-saved registers 550 ldp x21, x22, [x8], #16 551 ldp x23, x24, [x8], #16 552 ldp x25, x26, [x8], #16 553 ldp x27, x28, [x8], #16 554 ldp x29, x9, [x8], #16 555 ldr lr, [x8] 556 mov sp, x9 557 ret 558ENDPROC(cpu_switch_to) 559 560/* 561 * This is the fast syscall return path. We do as little as possible here, 562 * and this includes saving x0 back into the kernel stack. 563 */ 564ret_fast_syscall: 565 disable_irq // disable interrupts 566 ldr x1, [tsk, #TI_FLAGS] 567 and x2, x1, #_TIF_WORK_MASK 568 cbnz x2, fast_work_pending 569 tbz x1, #TIF_SINGLESTEP, fast_exit 570 disable_dbg 571 enable_step x2 572fast_exit: 573 kernel_exit 0, ret = 1 574 575/* 576 * Ok, we need to do extra processing, enter the slow path. 577 */ 578fast_work_pending: 579 str x0, [sp, #S_X0] // returned x0 580work_pending: 581 tbnz x1, #TIF_NEED_RESCHED, work_resched 582 /* TIF_SIGPENDING or TIF_NOTIFY_RESUME case */ 583 ldr x2, [sp, #S_PSTATE] 584 mov x0, sp // 'regs' 585 tst x2, #PSR_MODE_MASK // user mode regs? 586 b.ne no_work_pending // returning to kernel 587 enable_irq // enable interrupts for do_notify_resume() 588 bl do_notify_resume 589 b ret_to_user 590work_resched: 591 enable_dbg 592 bl schedule 593 594/* 595 * "slow" syscall return path. 596 */ 597ENTRY(ret_to_user) 598 disable_irq // disable interrupts 599 ldr x1, [tsk, #TI_FLAGS] 600 and x2, x1, #_TIF_WORK_MASK 601 cbnz x2, work_pending 602 tbz x1, #TIF_SINGLESTEP, no_work_pending 603 disable_dbg 604 enable_step x2 605no_work_pending: 606 kernel_exit 0, ret = 0 607ENDPROC(ret_to_user) 608 609/* 610 * This is how we return from a fork. 611 */ 612ENTRY(ret_from_fork) 613 bl schedule_tail 614 get_thread_info tsk 615 b ret_to_user 616ENDPROC(ret_from_fork) 617 618/* 619 * SVC handler. 620 */ 621 .align 6 622el0_svc: 623 adrp stbl, sys_call_table // load syscall table pointer 624 uxtw scno, w8 // syscall number in w8 625 mov sc_nr, #__NR_syscalls 626el0_svc_naked: // compat entry point 627 stp x0, scno, [sp, #S_ORIG_X0] // save the original x0 and syscall number 628 disable_step x16 629 isb 630 enable_dbg 631 enable_irq 632 633 get_thread_info tsk 634 ldr x16, [tsk, #TI_FLAGS] // check for syscall tracing 635 tbnz x16, #TIF_SYSCALL_TRACE, __sys_trace // are we tracing syscalls? 636 adr lr, ret_fast_syscall // return address 637 cmp scno, sc_nr // check upper syscall limit 638 b.hs ni_sys 639 ldr x16, [stbl, scno, lsl #3] // address in the syscall table 640 br x16 // call sys_* routine 641ni_sys: 642 mov x0, sp 643 b do_ni_syscall 644ENDPROC(el0_svc) 645 646 /* 647 * This is the really slow path. We're going to be doing context 648 * switches, and waiting for our parent to respond. 649 */ 650__sys_trace: 651 mov x1, sp 652 mov w0, #0 // trace entry 653 bl syscall_trace 654 adr lr, __sys_trace_return // return address 655 uxtw scno, w0 // syscall number (possibly new) 656 mov x1, sp // pointer to regs 657 cmp scno, sc_nr // check upper syscall limit 658 b.hs ni_sys 659 ldp x0, x1, [sp] // restore the syscall args 660 ldp x2, x3, [sp, #S_X2] 661 ldp x4, x5, [sp, #S_X4] 662 ldp x6, x7, [sp, #S_X6] 663 ldr x16, [stbl, scno, lsl #3] // address in the syscall table 664 br x16 // call sys_* routine 665 666__sys_trace_return: 667 str x0, [sp] // save returned x0 668 mov x1, sp 669 mov w0, #1 // trace exit 670 bl syscall_trace 671 b ret_to_user 672 673/* 674 * Special system call wrappers. 675 */ 676ENTRY(sys_execve_wrapper) 677 mov x3, sp 678 b sys_execve 679ENDPROC(sys_execve_wrapper) 680 681ENTRY(sys_clone_wrapper) 682 mov x5, sp 683 b sys_clone 684ENDPROC(sys_clone_wrapper) 685 686ENTRY(sys_rt_sigreturn_wrapper) 687 mov x0, sp 688 b sys_rt_sigreturn 689ENDPROC(sys_rt_sigreturn_wrapper) 690 691ENTRY(sys_sigaltstack_wrapper) 692 ldr x2, [sp, #S_SP] 693 b sys_sigaltstack 694ENDPROC(sys_sigaltstack_wrapper) 695 696ENTRY(handle_arch_irq) 697 .quad 0 698