1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * linux/arch/arm/kernel/signal.c 4 * 5 * Copyright (C) 1995-2009 Russell King 6 */ 7 #include <linux/errno.h> 8 #include <linux/random.h> 9 #include <linux/signal.h> 10 #include <linux/personality.h> 11 #include <linux/uaccess.h> 12 #include <linux/resume_user_mode.h> 13 #include <linux/uprobes.h> 14 #include <linux/syscalls.h> 15 16 #include <asm/elf.h> 17 #include <asm/cacheflush.h> 18 #include <asm/traps.h> 19 #include <asm/unistd.h> 20 #include <asm/vfp.h> 21 #include <asm/syscalls.h> 22 23 #include "signal.h" 24 25 extern const unsigned long sigreturn_codes[17]; 26 27 static unsigned long signal_return_offset; 28 29 #ifdef CONFIG_IWMMXT 30 31 static int preserve_iwmmxt_context(struct iwmmxt_sigframe __user *frame) 32 { 33 char kbuf[sizeof(*frame) + 8]; 34 struct iwmmxt_sigframe *kframe; 35 int err = 0; 36 37 /* the iWMMXt context must be 64 bit aligned */ 38 kframe = (struct iwmmxt_sigframe *)((unsigned long)(kbuf + 8) & ~7); 39 40 if (test_thread_flag(TIF_USING_IWMMXT)) { 41 kframe->magic = IWMMXT_MAGIC; 42 kframe->size = IWMMXT_STORAGE_SIZE; 43 iwmmxt_task_copy(current_thread_info(), &kframe->storage); 44 } else { 45 /* 46 * For bug-compatibility with older kernels, some space 47 * has to be reserved for iWMMXt even if it's not used. 48 * Set the magic and size appropriately so that properly 49 * written userspace can skip it reliably: 50 */ 51 *kframe = (struct iwmmxt_sigframe) { 52 .magic = DUMMY_MAGIC, 53 .size = IWMMXT_STORAGE_SIZE, 54 }; 55 } 56 57 err = __copy_to_user(frame, kframe, sizeof(*kframe)); 58 59 return err; 60 } 61 62 static int restore_iwmmxt_context(char __user **auxp) 63 { 64 struct iwmmxt_sigframe __user *frame = 65 (struct iwmmxt_sigframe __user *)*auxp; 66 char kbuf[sizeof(*frame) + 8]; 67 struct iwmmxt_sigframe *kframe; 68 69 /* the iWMMXt context must be 64 bit aligned */ 70 kframe = (struct iwmmxt_sigframe *)((unsigned long)(kbuf + 8) & ~7); 71 if (__copy_from_user(kframe, frame, sizeof(*frame))) 72 return -1; 73 74 /* 75 * For non-iWMMXt threads: a single iwmmxt_sigframe-sized dummy 76 * block is discarded for compatibility with setup_sigframe() if 77 * present, but we don't mandate its presence. If some other 78 * magic is here, it's not for us: 79 */ 80 if (!test_thread_flag(TIF_USING_IWMMXT) && 81 kframe->magic != DUMMY_MAGIC) 82 return 0; 83 84 if (kframe->size != IWMMXT_STORAGE_SIZE) 85 return -1; 86 87 if (test_thread_flag(TIF_USING_IWMMXT)) { 88 if (kframe->magic != IWMMXT_MAGIC) 89 return -1; 90 91 iwmmxt_task_restore(current_thread_info(), &kframe->storage); 92 } 93 94 *auxp += IWMMXT_STORAGE_SIZE; 95 return 0; 96 } 97 98 #endif 99 100 #ifdef CONFIG_VFP 101 102 static int preserve_vfp_context(struct vfp_sigframe __user *frame) 103 { 104 struct vfp_sigframe kframe; 105 int err = 0; 106 107 memset(&kframe, 0, sizeof(kframe)); 108 kframe.magic = VFP_MAGIC; 109 kframe.size = VFP_STORAGE_SIZE; 110 111 err = vfp_preserve_user_clear_hwstate(&kframe.ufp, &kframe.ufp_exc); 112 if (err) 113 return err; 114 115 return __copy_to_user(frame, &kframe, sizeof(kframe)); 116 } 117 118 static int restore_vfp_context(char __user **auxp) 119 { 120 struct vfp_sigframe frame; 121 int err; 122 123 err = __copy_from_user(&frame, *auxp, sizeof(frame)); 124 if (err) 125 return err; 126 127 if (frame.magic != VFP_MAGIC || frame.size != VFP_STORAGE_SIZE) 128 return -EINVAL; 129 130 *auxp += sizeof(frame); 131 return vfp_restore_user_hwstate(&frame.ufp, &frame.ufp_exc); 132 } 133 134 #endif 135 136 /* 137 * Do a signal return; undo the signal stack. These are aligned to 64-bit. 138 */ 139 140 static int restore_sigframe(struct pt_regs *regs, struct sigframe __user *sf) 141 { 142 struct sigcontext context; 143 char __user *aux; 144 sigset_t set; 145 int err; 146 147 err = __copy_from_user(&set, &sf->uc.uc_sigmask, sizeof(set)); 148 if (err == 0) 149 set_current_blocked(&set); 150 151 err |= __copy_from_user(&context, &sf->uc.uc_mcontext, sizeof(context)); 152 if (err == 0) { 153 regs->ARM_r0 = context.arm_r0; 154 regs->ARM_r1 = context.arm_r1; 155 regs->ARM_r2 = context.arm_r2; 156 regs->ARM_r3 = context.arm_r3; 157 regs->ARM_r4 = context.arm_r4; 158 regs->ARM_r5 = context.arm_r5; 159 regs->ARM_r6 = context.arm_r6; 160 regs->ARM_r7 = context.arm_r7; 161 regs->ARM_r8 = context.arm_r8; 162 regs->ARM_r9 = context.arm_r9; 163 regs->ARM_r10 = context.arm_r10; 164 regs->ARM_fp = context.arm_fp; 165 regs->ARM_ip = context.arm_ip; 166 regs->ARM_sp = context.arm_sp; 167 regs->ARM_lr = context.arm_lr; 168 regs->ARM_pc = context.arm_pc; 169 regs->ARM_cpsr = context.arm_cpsr; 170 } 171 172 err |= !valid_user_regs(regs); 173 174 aux = (char __user *) sf->uc.uc_regspace; 175 #ifdef CONFIG_IWMMXT 176 if (err == 0) 177 err |= restore_iwmmxt_context(&aux); 178 #endif 179 #ifdef CONFIG_VFP 180 if (err == 0) 181 err |= restore_vfp_context(&aux); 182 #endif 183 184 return err; 185 } 186 187 asmlinkage int sys_sigreturn(struct pt_regs *regs) 188 { 189 struct sigframe __user *frame; 190 191 /* Always make any pending restarted system calls return -EINTR */ 192 current->restart_block.fn = do_no_restart_syscall; 193 194 /* 195 * Since we stacked the signal on a 64-bit boundary, 196 * then 'sp' should be word aligned here. If it's 197 * not, then the user is trying to mess with us. 198 */ 199 if (regs->ARM_sp & 7) 200 goto badframe; 201 202 frame = (struct sigframe __user *)regs->ARM_sp; 203 204 if (!access_ok(frame, sizeof (*frame))) 205 goto badframe; 206 207 if (restore_sigframe(regs, frame)) 208 goto badframe; 209 210 return regs->ARM_r0; 211 212 badframe: 213 force_sig(SIGSEGV); 214 return 0; 215 } 216 217 asmlinkage int sys_rt_sigreturn(struct pt_regs *regs) 218 { 219 struct rt_sigframe __user *frame; 220 221 /* Always make any pending restarted system calls return -EINTR */ 222 current->restart_block.fn = do_no_restart_syscall; 223 224 /* 225 * Since we stacked the signal on a 64-bit boundary, 226 * then 'sp' should be word aligned here. If it's 227 * not, then the user is trying to mess with us. 228 */ 229 if (regs->ARM_sp & 7) 230 goto badframe; 231 232 frame = (struct rt_sigframe __user *)regs->ARM_sp; 233 234 if (!access_ok(frame, sizeof (*frame))) 235 goto badframe; 236 237 if (restore_sigframe(regs, &frame->sig)) 238 goto badframe; 239 240 if (restore_altstack(&frame->sig.uc.uc_stack)) 241 goto badframe; 242 243 return regs->ARM_r0; 244 245 badframe: 246 force_sig(SIGSEGV); 247 return 0; 248 } 249 250 static int 251 setup_sigframe(struct sigframe __user *sf, struct pt_regs *regs, sigset_t *set) 252 { 253 struct aux_sigframe __user *aux; 254 struct sigcontext context; 255 int err = 0; 256 257 context = (struct sigcontext) { 258 .arm_r0 = regs->ARM_r0, 259 .arm_r1 = regs->ARM_r1, 260 .arm_r2 = regs->ARM_r2, 261 .arm_r3 = regs->ARM_r3, 262 .arm_r4 = regs->ARM_r4, 263 .arm_r5 = regs->ARM_r5, 264 .arm_r6 = regs->ARM_r6, 265 .arm_r7 = regs->ARM_r7, 266 .arm_r8 = regs->ARM_r8, 267 .arm_r9 = regs->ARM_r9, 268 .arm_r10 = regs->ARM_r10, 269 .arm_fp = regs->ARM_fp, 270 .arm_ip = regs->ARM_ip, 271 .arm_sp = regs->ARM_sp, 272 .arm_lr = regs->ARM_lr, 273 .arm_pc = regs->ARM_pc, 274 .arm_cpsr = regs->ARM_cpsr, 275 276 .trap_no = current->thread.trap_no, 277 .error_code = current->thread.error_code, 278 .fault_address = current->thread.address, 279 .oldmask = set->sig[0], 280 }; 281 282 err |= __copy_to_user(&sf->uc.uc_mcontext, &context, sizeof(context)); 283 284 err |= __copy_to_user(&sf->uc.uc_sigmask, set, sizeof(*set)); 285 286 aux = (struct aux_sigframe __user *) sf->uc.uc_regspace; 287 #ifdef CONFIG_IWMMXT 288 if (err == 0) 289 err |= preserve_iwmmxt_context(&aux->iwmmxt); 290 #endif 291 #ifdef CONFIG_VFP 292 if (err == 0) 293 err |= preserve_vfp_context(&aux->vfp); 294 #endif 295 err |= __put_user(0, &aux->end_magic); 296 297 return err; 298 } 299 300 static inline void __user * 301 get_sigframe(struct ksignal *ksig, struct pt_regs *regs, int framesize) 302 { 303 unsigned long sp = sigsp(regs->ARM_sp, ksig); 304 void __user *frame; 305 306 /* 307 * ATPCS B01 mandates 8-byte alignment 308 */ 309 frame = (void __user *)((sp - framesize) & ~7); 310 311 /* 312 * Check that we can actually write to the signal frame. 313 */ 314 if (!access_ok(frame, framesize)) 315 frame = NULL; 316 317 return frame; 318 } 319 320 static int 321 setup_return(struct pt_regs *regs, struct ksignal *ksig, 322 unsigned long __user *rc, void __user *frame) 323 { 324 unsigned long handler = (unsigned long)ksig->ka.sa.sa_handler; 325 unsigned long handler_fdpic_GOT = 0; 326 unsigned long retcode; 327 unsigned int idx, thumb = 0; 328 unsigned long cpsr = regs->ARM_cpsr & ~(PSR_f | PSR_E_BIT); 329 bool fdpic = IS_ENABLED(CONFIG_BINFMT_ELF_FDPIC) && 330 (current->personality & FDPIC_FUNCPTRS); 331 332 if (fdpic) { 333 unsigned long __user *fdpic_func_desc = 334 (unsigned long __user *)handler; 335 if (__get_user(handler, &fdpic_func_desc[0]) || 336 __get_user(handler_fdpic_GOT, &fdpic_func_desc[1])) 337 return 1; 338 } 339 340 if (IS_ENABLED(CONFIG_CPU_ENDIAN_BE8)) 341 cpsr |= PSR_E_BIT; 342 343 /* 344 * Maybe we need to deliver a 32-bit signal to a 26-bit task. 345 */ 346 if (ksig->ka.sa.sa_flags & SA_THIRTYTWO) 347 cpsr = (cpsr & ~MODE_MASK) | USR_MODE; 348 349 #ifdef CONFIG_ARM_THUMB 350 if (elf_hwcap & HWCAP_THUMB) { 351 /* 352 * The LSB of the handler determines if we're going to 353 * be using THUMB or ARM mode for this signal handler. 354 */ 355 thumb = handler & 1; 356 357 /* 358 * Clear the If-Then Thumb-2 execution state. ARM spec 359 * requires this to be all 000s in ARM mode. Snapdragon 360 * S4/Krait misbehaves on a Thumb=>ARM signal transition 361 * without this. 362 * 363 * We must do this whenever we are running on a Thumb-2 364 * capable CPU, which includes ARMv6T2. However, we elect 365 * to always do this to simplify the code; this field is 366 * marked UNK/SBZP for older architectures. 367 */ 368 cpsr &= ~PSR_IT_MASK; 369 370 if (thumb) { 371 cpsr |= PSR_T_BIT; 372 } else 373 cpsr &= ~PSR_T_BIT; 374 } 375 #endif 376 377 if (ksig->ka.sa.sa_flags & SA_RESTORER) { 378 retcode = (unsigned long)ksig->ka.sa.sa_restorer; 379 if (fdpic) { 380 /* 381 * We need code to load the function descriptor. 382 * That code follows the standard sigreturn code 383 * (6 words), and is made of 3 + 2 words for each 384 * variant. The 4th copied word is the actual FD 385 * address that the assembly code expects. 386 */ 387 idx = 6 + thumb * 3; 388 if (ksig->ka.sa.sa_flags & SA_SIGINFO) 389 idx += 5; 390 if (__put_user(sigreturn_codes[idx], rc ) || 391 __put_user(sigreturn_codes[idx+1], rc+1) || 392 __put_user(sigreturn_codes[idx+2], rc+2) || 393 __put_user(retcode, rc+3)) 394 return 1; 395 goto rc_finish; 396 } 397 } else { 398 idx = thumb << 1; 399 if (ksig->ka.sa.sa_flags & SA_SIGINFO) 400 idx += 3; 401 402 /* 403 * Put the sigreturn code on the stack no matter which return 404 * mechanism we use in order to remain ABI compliant 405 */ 406 if (__put_user(sigreturn_codes[idx], rc) || 407 __put_user(sigreturn_codes[idx+1], rc+1)) 408 return 1; 409 410 rc_finish: 411 #ifdef CONFIG_MMU 412 if (cpsr & MODE32_BIT) { 413 struct mm_struct *mm = current->mm; 414 415 /* 416 * 32-bit code can use the signal return page 417 * except when the MPU has protected the vectors 418 * page from PL0 419 */ 420 retcode = mm->context.sigpage + signal_return_offset + 421 (idx << 2) + thumb; 422 } else 423 #endif 424 { 425 /* 426 * Ensure that the instruction cache sees 427 * the return code written onto the stack. 428 */ 429 flush_icache_range((unsigned long)rc, 430 (unsigned long)(rc + 3)); 431 432 retcode = ((unsigned long)rc) + thumb; 433 } 434 } 435 436 regs->ARM_r0 = ksig->sig; 437 regs->ARM_sp = (unsigned long)frame; 438 regs->ARM_lr = retcode; 439 regs->ARM_pc = handler; 440 if (fdpic) 441 regs->ARM_r9 = handler_fdpic_GOT; 442 regs->ARM_cpsr = cpsr; 443 444 return 0; 445 } 446 447 static int 448 setup_frame(struct ksignal *ksig, sigset_t *set, struct pt_regs *regs) 449 { 450 struct sigframe __user *frame = get_sigframe(ksig, regs, sizeof(*frame)); 451 int err = 0; 452 453 if (!frame) 454 return 1; 455 456 /* 457 * Set uc.uc_flags to a value which sc.trap_no would never have. 458 */ 459 err = __put_user(0x5ac3c35a, &frame->uc.uc_flags); 460 461 err |= setup_sigframe(frame, regs, set); 462 if (err == 0) 463 err = setup_return(regs, ksig, frame->retcode, frame); 464 465 return err; 466 } 467 468 static int 469 setup_rt_frame(struct ksignal *ksig, sigset_t *set, struct pt_regs *regs) 470 { 471 struct rt_sigframe __user *frame = get_sigframe(ksig, regs, sizeof(*frame)); 472 int err = 0; 473 474 if (!frame) 475 return 1; 476 477 err |= copy_siginfo_to_user(&frame->info, &ksig->info); 478 479 err |= __put_user(0, &frame->sig.uc.uc_flags); 480 err |= __put_user(NULL, &frame->sig.uc.uc_link); 481 482 err |= __save_altstack(&frame->sig.uc.uc_stack, regs->ARM_sp); 483 err |= setup_sigframe(&frame->sig, regs, set); 484 if (err == 0) 485 err = setup_return(regs, ksig, frame->sig.retcode, frame); 486 487 if (err == 0) { 488 /* 489 * For realtime signals we must also set the second and third 490 * arguments for the signal handler. 491 * -- Peter Maydell <pmaydell@chiark.greenend.org.uk> 2000-12-06 492 */ 493 regs->ARM_r1 = (unsigned long)&frame->info; 494 regs->ARM_r2 = (unsigned long)&frame->sig.uc; 495 } 496 497 return err; 498 } 499 500 /* 501 * OK, we're invoking a handler 502 */ 503 static void handle_signal(struct ksignal *ksig, struct pt_regs *regs) 504 { 505 sigset_t *oldset = sigmask_to_save(); 506 int ret; 507 508 /* 509 * Perform fixup for the pre-signal frame. 510 */ 511 rseq_signal_deliver(ksig, regs); 512 513 /* 514 * Set up the stack frame 515 */ 516 if (ksig->ka.sa.sa_flags & SA_SIGINFO) 517 ret = setup_rt_frame(ksig, oldset, regs); 518 else 519 ret = setup_frame(ksig, oldset, regs); 520 521 /* 522 * Check that the resulting registers are actually sane. 523 */ 524 ret |= !valid_user_regs(regs); 525 526 signal_setup_done(ret, ksig, 0); 527 } 528 529 /* 530 * Note that 'init' is a special process: it doesn't get signals it doesn't 531 * want to handle. Thus you cannot kill init even with a SIGKILL even by 532 * mistake. 533 * 534 * Note that we go through the signals twice: once to check the signals that 535 * the kernel can handle, and then we build all the user-level signal handling 536 * stack-frames in one go after that. 537 */ 538 static int do_signal(struct pt_regs *regs, int syscall) 539 { 540 unsigned int retval = 0, continue_addr = 0, restart_addr = 0; 541 struct ksignal ksig; 542 int restart = 0; 543 544 /* 545 * If we were from a system call, check for system call restarting... 546 */ 547 if (syscall) { 548 continue_addr = regs->ARM_pc; 549 restart_addr = continue_addr - (thumb_mode(regs) ? 2 : 4); 550 retval = regs->ARM_r0; 551 552 /* 553 * Prepare for system call restart. We do this here so that a 554 * debugger will see the already changed PSW. 555 */ 556 switch (retval) { 557 case -ERESTART_RESTARTBLOCK: 558 restart -= 2; 559 fallthrough; 560 case -ERESTARTNOHAND: 561 case -ERESTARTSYS: 562 case -ERESTARTNOINTR: 563 restart++; 564 regs->ARM_r0 = regs->ARM_ORIG_r0; 565 regs->ARM_pc = restart_addr; 566 break; 567 } 568 } 569 570 /* 571 * Get the signal to deliver. When running under ptrace, at this 572 * point the debugger may change all our registers ... 573 */ 574 /* 575 * Depending on the signal settings we may need to revert the 576 * decision to restart the system call. But skip this if a 577 * debugger has chosen to restart at a different PC. 578 */ 579 if (get_signal(&ksig)) { 580 /* handler */ 581 if (unlikely(restart) && regs->ARM_pc == restart_addr) { 582 if (retval == -ERESTARTNOHAND || 583 retval == -ERESTART_RESTARTBLOCK 584 || (retval == -ERESTARTSYS 585 && !(ksig.ka.sa.sa_flags & SA_RESTART))) { 586 regs->ARM_r0 = -EINTR; 587 regs->ARM_pc = continue_addr; 588 } 589 } 590 handle_signal(&ksig, regs); 591 } else { 592 /* no handler */ 593 restore_saved_sigmask(); 594 if (unlikely(restart) && regs->ARM_pc == restart_addr) { 595 regs->ARM_pc = continue_addr; 596 return restart; 597 } 598 } 599 return 0; 600 } 601 602 asmlinkage int 603 do_work_pending(struct pt_regs *regs, unsigned int thread_flags, int syscall) 604 { 605 /* 606 * The assembly code enters us with IRQs off, but it hasn't 607 * informed the tracing code of that for efficiency reasons. 608 * Update the trace code with the current status. 609 */ 610 trace_hardirqs_off(); 611 do { 612 if (likely(thread_flags & _TIF_NEED_RESCHED)) { 613 schedule(); 614 } else { 615 if (unlikely(!user_mode(regs))) 616 return 0; 617 local_irq_enable(); 618 if (thread_flags & (_TIF_SIGPENDING | _TIF_NOTIFY_SIGNAL)) { 619 int restart = do_signal(regs, syscall); 620 if (unlikely(restart)) { 621 /* 622 * Restart without handlers. 623 * Deal with it without leaving 624 * the kernel space. 625 */ 626 return restart; 627 } 628 syscall = 0; 629 } else if (thread_flags & _TIF_UPROBE) { 630 uprobe_notify_resume(regs); 631 } else { 632 resume_user_mode_work(regs); 633 } 634 } 635 local_irq_disable(); 636 thread_flags = read_thread_flags(); 637 } while (thread_flags & _TIF_WORK_MASK); 638 return 0; 639 } 640 641 struct page *get_signal_page(void) 642 { 643 unsigned long ptr; 644 unsigned offset; 645 struct page *page; 646 void *addr; 647 648 page = alloc_pages(GFP_KERNEL, 0); 649 650 if (!page) 651 return NULL; 652 653 addr = page_address(page); 654 655 /* Poison the entire page */ 656 memset32(addr, __opcode_to_mem_arm(0xe7fddef1), 657 PAGE_SIZE / sizeof(u32)); 658 659 /* Give the signal return code some randomness */ 660 offset = 0x200 + (get_random_u16() & 0x7fc); 661 signal_return_offset = offset; 662 663 /* Copy signal return handlers into the page */ 664 memcpy(addr + offset, sigreturn_codes, sizeof(sigreturn_codes)); 665 666 /* Flush out all instructions in this page */ 667 ptr = (unsigned long)addr; 668 flush_icache_range(ptr, ptr + PAGE_SIZE); 669 670 return page; 671 } 672 673 #ifdef CONFIG_DEBUG_RSEQ 674 asmlinkage void do_rseq_syscall(struct pt_regs *regs) 675 { 676 rseq_syscall(regs); 677 } 678 #endif 679 680 /* 681 * Compile-time assertions for siginfo_t offsets. Check NSIG* as well, as 682 * changes likely come with new fields that should be added below. 683 */ 684 static_assert(NSIGILL == 11); 685 static_assert(NSIGFPE == 15); 686 static_assert(NSIGSEGV == 10); 687 static_assert(NSIGBUS == 5); 688 static_assert(NSIGTRAP == 6); 689 static_assert(NSIGCHLD == 6); 690 static_assert(NSIGSYS == 2); 691 static_assert(sizeof(siginfo_t) == 128); 692 static_assert(__alignof__(siginfo_t) == 4); 693 static_assert(offsetof(siginfo_t, si_signo) == 0x00); 694 static_assert(offsetof(siginfo_t, si_errno) == 0x04); 695 static_assert(offsetof(siginfo_t, si_code) == 0x08); 696 static_assert(offsetof(siginfo_t, si_pid) == 0x0c); 697 static_assert(offsetof(siginfo_t, si_uid) == 0x10); 698 static_assert(offsetof(siginfo_t, si_tid) == 0x0c); 699 static_assert(offsetof(siginfo_t, si_overrun) == 0x10); 700 static_assert(offsetof(siginfo_t, si_status) == 0x14); 701 static_assert(offsetof(siginfo_t, si_utime) == 0x18); 702 static_assert(offsetof(siginfo_t, si_stime) == 0x1c); 703 static_assert(offsetof(siginfo_t, si_value) == 0x14); 704 static_assert(offsetof(siginfo_t, si_int) == 0x14); 705 static_assert(offsetof(siginfo_t, si_ptr) == 0x14); 706 static_assert(offsetof(siginfo_t, si_addr) == 0x0c); 707 static_assert(offsetof(siginfo_t, si_addr_lsb) == 0x10); 708 static_assert(offsetof(siginfo_t, si_lower) == 0x14); 709 static_assert(offsetof(siginfo_t, si_upper) == 0x18); 710 static_assert(offsetof(siginfo_t, si_pkey) == 0x14); 711 static_assert(offsetof(siginfo_t, si_perf_data) == 0x10); 712 static_assert(offsetof(siginfo_t, si_perf_type) == 0x14); 713 static_assert(offsetof(siginfo_t, si_perf_flags) == 0x18); 714 static_assert(offsetof(siginfo_t, si_band) == 0x0c); 715 static_assert(offsetof(siginfo_t, si_fd) == 0x10); 716 static_assert(offsetof(siginfo_t, si_call_addr) == 0x0c); 717 static_assert(offsetof(siginfo_t, si_syscall) == 0x10); 718 static_assert(offsetof(siginfo_t, si_arch) == 0x14); 719