1 // SPDX-License-Identifier: GPL-2.0 2 // Copyright (C) 2017 Arm Ltd. 3 #define pr_fmt(fmt) "sdei: " fmt 4 5 #include <acpi/ghes.h> 6 #include <linux/acpi.h> 7 #include <linux/arm_sdei.h> 8 #include <linux/arm-smccc.h> 9 #include <linux/atomic.h> 10 #include <linux/bitops.h> 11 #include <linux/compiler.h> 12 #include <linux/cpuhotplug.h> 13 #include <linux/cpu.h> 14 #include <linux/cpu_pm.h> 15 #include <linux/errno.h> 16 #include <linux/hardirq.h> 17 #include <linux/kernel.h> 18 #include <linux/kprobes.h> 19 #include <linux/kvm_host.h> 20 #include <linux/list.h> 21 #include <linux/mutex.h> 22 #include <linux/notifier.h> 23 #include <linux/of.h> 24 #include <linux/of_platform.h> 25 #include <linux/percpu.h> 26 #include <linux/platform_device.h> 27 #include <linux/pm.h> 28 #include <linux/ptrace.h> 29 #include <linux/preempt.h> 30 #include <linux/reboot.h> 31 #include <linux/slab.h> 32 #include <linux/smp.h> 33 #include <linux/spinlock.h> 34 35 /* 36 * The call to use to reach the firmware. 37 */ 38 static asmlinkage void (*sdei_firmware_call)(unsigned long function_id, 39 unsigned long arg0, unsigned long arg1, 40 unsigned long arg2, unsigned long arg3, 41 unsigned long arg4, struct arm_smccc_res *res); 42 43 /* entry point from firmware to arch asm code */ 44 static unsigned long sdei_entry_point; 45 46 static int sdei_hp_state; 47 48 struct sdei_event { 49 /* These three are protected by the sdei_list_lock */ 50 struct list_head list; 51 bool reregister; 52 bool reenable; 53 54 u32 event_num; 55 u8 type; 56 u8 priority; 57 58 /* This pointer is handed to firmware as the event argument. */ 59 union { 60 /* Shared events */ 61 struct sdei_registered_event *registered; 62 63 /* CPU private events */ 64 struct sdei_registered_event __percpu *private_registered; 65 }; 66 }; 67 68 /* Take the mutex for any API call or modification. Take the mutex first. */ 69 static DEFINE_MUTEX(sdei_events_lock); 70 71 /* and then hold this when modifying the list */ 72 static DEFINE_SPINLOCK(sdei_list_lock); 73 static LIST_HEAD(sdei_list); 74 75 /* Private events are registered/enabled via IPI passing one of these */ 76 struct sdei_crosscall_args { 77 struct sdei_event *event; 78 atomic_t errors; 79 int first_error; 80 }; 81 82 #define CROSSCALL_INIT(arg, event) \ 83 do { \ 84 arg.event = event; \ 85 arg.first_error = 0; \ 86 atomic_set(&arg.errors, 0); \ 87 } while (0) 88 89 static inline int sdei_do_local_call(smp_call_func_t fn, 90 struct sdei_event *event) 91 { 92 struct sdei_crosscall_args arg; 93 94 CROSSCALL_INIT(arg, event); 95 fn(&arg); 96 97 return arg.first_error; 98 } 99 100 static inline int sdei_do_cross_call(smp_call_func_t fn, 101 struct sdei_event *event) 102 { 103 struct sdei_crosscall_args arg; 104 105 CROSSCALL_INIT(arg, event); 106 on_each_cpu(fn, &arg, true); 107 108 return arg.first_error; 109 } 110 111 static inline void 112 sdei_cross_call_return(struct sdei_crosscall_args *arg, int err) 113 { 114 if (err && (atomic_inc_return(&arg->errors) == 1)) 115 arg->first_error = err; 116 } 117 118 static int sdei_to_linux_errno(unsigned long sdei_err) 119 { 120 switch (sdei_err) { 121 case SDEI_NOT_SUPPORTED: 122 return -EOPNOTSUPP; 123 case SDEI_INVALID_PARAMETERS: 124 return -EINVAL; 125 case SDEI_DENIED: 126 return -EPERM; 127 case SDEI_PENDING: 128 return -EINPROGRESS; 129 case SDEI_OUT_OF_RESOURCE: 130 return -ENOMEM; 131 } 132 133 return 0; 134 } 135 136 static int invoke_sdei_fn(unsigned long function_id, unsigned long arg0, 137 unsigned long arg1, unsigned long arg2, 138 unsigned long arg3, unsigned long arg4, 139 u64 *result) 140 { 141 int err; 142 struct arm_smccc_res res; 143 144 if (sdei_firmware_call) { 145 sdei_firmware_call(function_id, arg0, arg1, arg2, arg3, arg4, 146 &res); 147 err = sdei_to_linux_errno(res.a0); 148 } else { 149 /* 150 * !sdei_firmware_call means we failed to probe or called 151 * sdei_mark_interface_broken(). -EIO is not an error returned 152 * by sdei_to_linux_errno() and is used to suppress messages 153 * from this driver. 154 */ 155 err = -EIO; 156 res.a0 = SDEI_NOT_SUPPORTED; 157 } 158 159 if (result) 160 *result = res.a0; 161 162 return err; 163 } 164 NOKPROBE_SYMBOL(invoke_sdei_fn); 165 166 static struct sdei_event *sdei_event_find(u32 event_num) 167 { 168 struct sdei_event *e, *found = NULL; 169 170 lockdep_assert_held(&sdei_events_lock); 171 172 spin_lock(&sdei_list_lock); 173 list_for_each_entry(e, &sdei_list, list) { 174 if (e->event_num == event_num) { 175 found = e; 176 break; 177 } 178 } 179 spin_unlock(&sdei_list_lock); 180 181 return found; 182 } 183 184 int sdei_api_event_context(u32 query, u64 *result) 185 { 186 return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_CONTEXT, query, 0, 0, 0, 0, 187 result); 188 } 189 NOKPROBE_SYMBOL(sdei_api_event_context); 190 191 static int sdei_api_event_get_info(u32 event, u32 info, u64 *result) 192 { 193 return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_GET_INFO, event, info, 0, 194 0, 0, result); 195 } 196 197 static struct sdei_event *sdei_event_create(u32 event_num, 198 sdei_event_callback *cb, 199 void *cb_arg) 200 { 201 int err; 202 u64 result; 203 struct sdei_event *event; 204 struct sdei_registered_event *reg; 205 206 lockdep_assert_held(&sdei_events_lock); 207 208 event = kzalloc(sizeof(*event), GFP_KERNEL); 209 if (!event) { 210 err = -ENOMEM; 211 goto fail; 212 } 213 214 INIT_LIST_HEAD(&event->list); 215 event->event_num = event_num; 216 217 err = sdei_api_event_get_info(event_num, SDEI_EVENT_INFO_EV_PRIORITY, 218 &result); 219 if (err) 220 goto fail; 221 event->priority = result; 222 223 err = sdei_api_event_get_info(event_num, SDEI_EVENT_INFO_EV_TYPE, 224 &result); 225 if (err) 226 goto fail; 227 event->type = result; 228 229 if (event->type == SDEI_EVENT_TYPE_SHARED) { 230 reg = kzalloc(sizeof(*reg), GFP_KERNEL); 231 if (!reg) { 232 err = -ENOMEM; 233 goto fail; 234 } 235 236 reg->event_num = event->event_num; 237 reg->priority = event->priority; 238 239 reg->callback = cb; 240 reg->callback_arg = cb_arg; 241 event->registered = reg; 242 } else { 243 int cpu; 244 struct sdei_registered_event __percpu *regs; 245 246 regs = alloc_percpu(struct sdei_registered_event); 247 if (!regs) { 248 err = -ENOMEM; 249 goto fail; 250 } 251 252 for_each_possible_cpu(cpu) { 253 reg = per_cpu_ptr(regs, cpu); 254 255 reg->event_num = event->event_num; 256 reg->priority = event->priority; 257 reg->callback = cb; 258 reg->callback_arg = cb_arg; 259 } 260 261 event->private_registered = regs; 262 } 263 264 spin_lock(&sdei_list_lock); 265 list_add(&event->list, &sdei_list); 266 spin_unlock(&sdei_list_lock); 267 268 return event; 269 270 fail: 271 kfree(event); 272 return ERR_PTR(err); 273 } 274 275 static void sdei_event_destroy_llocked(struct sdei_event *event) 276 { 277 lockdep_assert_held(&sdei_events_lock); 278 lockdep_assert_held(&sdei_list_lock); 279 280 list_del(&event->list); 281 282 if (event->type == SDEI_EVENT_TYPE_SHARED) 283 kfree(event->registered); 284 else 285 free_percpu(event->private_registered); 286 287 kfree(event); 288 } 289 290 static void sdei_event_destroy(struct sdei_event *event) 291 { 292 spin_lock(&sdei_list_lock); 293 sdei_event_destroy_llocked(event); 294 spin_unlock(&sdei_list_lock); 295 } 296 297 static int sdei_api_get_version(u64 *version) 298 { 299 return invoke_sdei_fn(SDEI_1_0_FN_SDEI_VERSION, 0, 0, 0, 0, 0, version); 300 } 301 302 int sdei_mask_local_cpu(void) 303 { 304 int err; 305 306 err = invoke_sdei_fn(SDEI_1_0_FN_SDEI_PE_MASK, 0, 0, 0, 0, 0, NULL); 307 if (err && err != -EIO) { 308 pr_warn_once("failed to mask CPU[%u]: %d\n", 309 smp_processor_id(), err); 310 return err; 311 } 312 313 return 0; 314 } 315 316 static void _ipi_mask_cpu(void *ignored) 317 { 318 WARN_ON_ONCE(preemptible()); 319 sdei_mask_local_cpu(); 320 } 321 322 int sdei_unmask_local_cpu(void) 323 { 324 int err; 325 326 err = invoke_sdei_fn(SDEI_1_0_FN_SDEI_PE_UNMASK, 0, 0, 0, 0, 0, NULL); 327 if (err && err != -EIO) { 328 pr_warn_once("failed to unmask CPU[%u]: %d\n", 329 smp_processor_id(), err); 330 return err; 331 } 332 333 return 0; 334 } 335 336 static void _ipi_unmask_cpu(void *ignored) 337 { 338 WARN_ON_ONCE(preemptible()); 339 sdei_unmask_local_cpu(); 340 } 341 342 static void _ipi_private_reset(void *ignored) 343 { 344 int err; 345 346 WARN_ON_ONCE(preemptible()); 347 348 err = invoke_sdei_fn(SDEI_1_0_FN_SDEI_PRIVATE_RESET, 0, 0, 0, 0, 0, 349 NULL); 350 if (err && err != -EIO) 351 pr_warn_once("failed to reset CPU[%u]: %d\n", 352 smp_processor_id(), err); 353 } 354 355 static int sdei_api_shared_reset(void) 356 { 357 return invoke_sdei_fn(SDEI_1_0_FN_SDEI_SHARED_RESET, 0, 0, 0, 0, 0, 358 NULL); 359 } 360 361 static void sdei_mark_interface_broken(void) 362 { 363 pr_err("disabling SDEI firmware interface\n"); 364 on_each_cpu(&_ipi_mask_cpu, NULL, true); 365 sdei_firmware_call = NULL; 366 } 367 368 static int sdei_platform_reset(void) 369 { 370 int err; 371 372 on_each_cpu(&_ipi_private_reset, NULL, true); 373 err = sdei_api_shared_reset(); 374 if (err) { 375 pr_err("Failed to reset platform: %d\n", err); 376 sdei_mark_interface_broken(); 377 } 378 379 return err; 380 } 381 382 static int sdei_api_event_enable(u32 event_num) 383 { 384 return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_ENABLE, event_num, 0, 0, 0, 385 0, NULL); 386 } 387 388 /* Called directly by the hotplug callbacks */ 389 static void _local_event_enable(void *data) 390 { 391 int err; 392 struct sdei_crosscall_args *arg = data; 393 394 err = sdei_api_event_enable(arg->event->event_num); 395 396 sdei_cross_call_return(arg, err); 397 } 398 399 int sdei_event_enable(u32 event_num) 400 { 401 int err = -EINVAL; 402 struct sdei_event *event; 403 404 mutex_lock(&sdei_events_lock); 405 event = sdei_event_find(event_num); 406 if (!event) { 407 mutex_unlock(&sdei_events_lock); 408 return -ENOENT; 409 } 410 411 412 cpus_read_lock(); 413 if (event->type == SDEI_EVENT_TYPE_SHARED) 414 err = sdei_api_event_enable(event->event_num); 415 else 416 err = sdei_do_cross_call(_local_event_enable, event); 417 418 if (!err) { 419 spin_lock(&sdei_list_lock); 420 event->reenable = true; 421 spin_unlock(&sdei_list_lock); 422 } 423 cpus_read_unlock(); 424 mutex_unlock(&sdei_events_lock); 425 426 return err; 427 } 428 429 static int sdei_api_event_disable(u32 event_num) 430 { 431 return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_DISABLE, event_num, 0, 0, 432 0, 0, NULL); 433 } 434 435 static void _ipi_event_disable(void *data) 436 { 437 int err; 438 struct sdei_crosscall_args *arg = data; 439 440 err = sdei_api_event_disable(arg->event->event_num); 441 442 sdei_cross_call_return(arg, err); 443 } 444 445 int sdei_event_disable(u32 event_num) 446 { 447 int err = -EINVAL; 448 struct sdei_event *event; 449 450 mutex_lock(&sdei_events_lock); 451 event = sdei_event_find(event_num); 452 if (!event) { 453 mutex_unlock(&sdei_events_lock); 454 return -ENOENT; 455 } 456 457 spin_lock(&sdei_list_lock); 458 event->reenable = false; 459 spin_unlock(&sdei_list_lock); 460 461 if (event->type == SDEI_EVENT_TYPE_SHARED) 462 err = sdei_api_event_disable(event->event_num); 463 else 464 err = sdei_do_cross_call(_ipi_event_disable, event); 465 mutex_unlock(&sdei_events_lock); 466 467 return err; 468 } 469 470 static int sdei_api_event_unregister(u32 event_num) 471 { 472 return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_UNREGISTER, event_num, 0, 473 0, 0, 0, NULL); 474 } 475 476 /* Called directly by the hotplug callbacks */ 477 static void _local_event_unregister(void *data) 478 { 479 int err; 480 struct sdei_crosscall_args *arg = data; 481 482 err = sdei_api_event_unregister(arg->event->event_num); 483 484 sdei_cross_call_return(arg, err); 485 } 486 487 int sdei_event_unregister(u32 event_num) 488 { 489 int err; 490 struct sdei_event *event; 491 492 WARN_ON(in_nmi()); 493 494 mutex_lock(&sdei_events_lock); 495 event = sdei_event_find(event_num); 496 if (!event) { 497 pr_warn("Event %u not registered\n", event_num); 498 err = -ENOENT; 499 goto unlock; 500 } 501 502 spin_lock(&sdei_list_lock); 503 event->reregister = false; 504 event->reenable = false; 505 spin_unlock(&sdei_list_lock); 506 507 if (event->type == SDEI_EVENT_TYPE_SHARED) 508 err = sdei_api_event_unregister(event->event_num); 509 else 510 err = sdei_do_cross_call(_local_event_unregister, event); 511 512 if (err) 513 goto unlock; 514 515 sdei_event_destroy(event); 516 unlock: 517 mutex_unlock(&sdei_events_lock); 518 519 return err; 520 } 521 522 /* 523 * unregister events, but don't destroy them as they are re-registered by 524 * sdei_reregister_shared(). 525 */ 526 static int sdei_unregister_shared(void) 527 { 528 int err = 0; 529 struct sdei_event *event; 530 531 mutex_lock(&sdei_events_lock); 532 spin_lock(&sdei_list_lock); 533 list_for_each_entry(event, &sdei_list, list) { 534 if (event->type != SDEI_EVENT_TYPE_SHARED) 535 continue; 536 537 err = sdei_api_event_unregister(event->event_num); 538 if (err) 539 break; 540 } 541 spin_unlock(&sdei_list_lock); 542 mutex_unlock(&sdei_events_lock); 543 544 return err; 545 } 546 547 static int sdei_api_event_register(u32 event_num, unsigned long entry_point, 548 void *arg, u64 flags, u64 affinity) 549 { 550 return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_REGISTER, event_num, 551 (unsigned long)entry_point, (unsigned long)arg, 552 flags, affinity, NULL); 553 } 554 555 /* Called directly by the hotplug callbacks */ 556 static void _local_event_register(void *data) 557 { 558 int err; 559 struct sdei_registered_event *reg; 560 struct sdei_crosscall_args *arg = data; 561 562 reg = per_cpu_ptr(arg->event->private_registered, smp_processor_id()); 563 err = sdei_api_event_register(arg->event->event_num, sdei_entry_point, 564 reg, 0, 0); 565 566 sdei_cross_call_return(arg, err); 567 } 568 569 int sdei_event_register(u32 event_num, sdei_event_callback *cb, void *arg) 570 { 571 int err; 572 struct sdei_event *event; 573 574 WARN_ON(in_nmi()); 575 576 mutex_lock(&sdei_events_lock); 577 if (sdei_event_find(event_num)) { 578 pr_warn("Event %u already registered\n", event_num); 579 err = -EBUSY; 580 goto unlock; 581 } 582 583 event = sdei_event_create(event_num, cb, arg); 584 if (IS_ERR(event)) { 585 err = PTR_ERR(event); 586 pr_warn("Failed to create event %u: %d\n", event_num, err); 587 goto unlock; 588 } 589 590 cpus_read_lock(); 591 if (event->type == SDEI_EVENT_TYPE_SHARED) { 592 err = sdei_api_event_register(event->event_num, 593 sdei_entry_point, 594 event->registered, 595 SDEI_EVENT_REGISTER_RM_ANY, 0); 596 } else { 597 err = sdei_do_cross_call(_local_event_register, event); 598 if (err) 599 sdei_do_cross_call(_local_event_unregister, event); 600 } 601 602 if (err) { 603 sdei_event_destroy(event); 604 pr_warn("Failed to register event %u: %d\n", event_num, err); 605 goto cpu_unlock; 606 } 607 608 spin_lock(&sdei_list_lock); 609 event->reregister = true; 610 spin_unlock(&sdei_list_lock); 611 cpu_unlock: 612 cpus_read_unlock(); 613 unlock: 614 mutex_unlock(&sdei_events_lock); 615 return err; 616 } 617 618 static int sdei_reregister_shared(void) 619 { 620 int err = 0; 621 struct sdei_event *event; 622 623 mutex_lock(&sdei_events_lock); 624 spin_lock(&sdei_list_lock); 625 list_for_each_entry(event, &sdei_list, list) { 626 if (event->type != SDEI_EVENT_TYPE_SHARED) 627 continue; 628 629 if (event->reregister) { 630 err = sdei_api_event_register(event->event_num, 631 sdei_entry_point, event->registered, 632 SDEI_EVENT_REGISTER_RM_ANY, 0); 633 if (err) { 634 pr_err("Failed to re-register event %u\n", 635 event->event_num); 636 sdei_event_destroy_llocked(event); 637 break; 638 } 639 } 640 641 if (event->reenable) { 642 err = sdei_api_event_enable(event->event_num); 643 if (err) { 644 pr_err("Failed to re-enable event %u\n", 645 event->event_num); 646 break; 647 } 648 } 649 } 650 spin_unlock(&sdei_list_lock); 651 mutex_unlock(&sdei_events_lock); 652 653 return err; 654 } 655 656 static int sdei_cpuhp_down(unsigned int cpu) 657 { 658 struct sdei_event *event; 659 int err; 660 661 /* un-register private events */ 662 spin_lock(&sdei_list_lock); 663 list_for_each_entry(event, &sdei_list, list) { 664 if (event->type == SDEI_EVENT_TYPE_SHARED) 665 continue; 666 667 err = sdei_do_local_call(_local_event_unregister, event); 668 if (err) { 669 pr_err("Failed to unregister event %u: %d\n", 670 event->event_num, err); 671 } 672 } 673 spin_unlock(&sdei_list_lock); 674 675 return sdei_mask_local_cpu(); 676 } 677 678 static int sdei_cpuhp_up(unsigned int cpu) 679 { 680 struct sdei_event *event; 681 int err; 682 683 /* re-register/enable private events */ 684 spin_lock(&sdei_list_lock); 685 list_for_each_entry(event, &sdei_list, list) { 686 if (event->type == SDEI_EVENT_TYPE_SHARED) 687 continue; 688 689 if (event->reregister) { 690 err = sdei_do_local_call(_local_event_register, event); 691 if (err) { 692 pr_err("Failed to re-register event %u: %d\n", 693 event->event_num, err); 694 } 695 } 696 697 if (event->reenable) { 698 err = sdei_do_local_call(_local_event_enable, event); 699 if (err) { 700 pr_err("Failed to re-enable event %u: %d\n", 701 event->event_num, err); 702 } 703 } 704 } 705 spin_unlock(&sdei_list_lock); 706 707 return sdei_unmask_local_cpu(); 708 } 709 710 /* When entering idle, mask/unmask events for this cpu */ 711 static int sdei_pm_notifier(struct notifier_block *nb, unsigned long action, 712 void *data) 713 { 714 int rv; 715 716 WARN_ON_ONCE(preemptible()); 717 718 switch (action) { 719 case CPU_PM_ENTER: 720 rv = sdei_mask_local_cpu(); 721 break; 722 case CPU_PM_EXIT: 723 case CPU_PM_ENTER_FAILED: 724 rv = sdei_unmask_local_cpu(); 725 break; 726 default: 727 return NOTIFY_DONE; 728 } 729 730 if (rv) 731 return notifier_from_errno(rv); 732 733 return NOTIFY_OK; 734 } 735 736 static struct notifier_block sdei_pm_nb = { 737 .notifier_call = sdei_pm_notifier, 738 }; 739 740 static int sdei_device_suspend(struct device *dev) 741 { 742 on_each_cpu(_ipi_mask_cpu, NULL, true); 743 744 return 0; 745 } 746 747 static int sdei_device_resume(struct device *dev) 748 { 749 on_each_cpu(_ipi_unmask_cpu, NULL, true); 750 751 return 0; 752 } 753 754 /* 755 * We need all events to be reregistered when we resume from hibernate. 756 * 757 * The sequence is freeze->thaw. Reboot. freeze->restore. We unregister 758 * events during freeze, then re-register and re-enable them during thaw 759 * and restore. 760 */ 761 static int sdei_device_freeze(struct device *dev) 762 { 763 int err; 764 765 /* unregister private events */ 766 cpuhp_remove_state(sdei_entry_point); 767 768 err = sdei_unregister_shared(); 769 if (err) 770 return err; 771 772 return 0; 773 } 774 775 static int sdei_device_thaw(struct device *dev) 776 { 777 int err; 778 779 /* re-register shared events */ 780 err = sdei_reregister_shared(); 781 if (err) { 782 pr_warn("Failed to re-register shared events...\n"); 783 sdei_mark_interface_broken(); 784 return err; 785 } 786 787 err = cpuhp_setup_state(CPUHP_AP_ONLINE_DYN, "SDEI", 788 &sdei_cpuhp_up, &sdei_cpuhp_down); 789 if (err < 0) { 790 pr_warn("Failed to re-register CPU hotplug notifier...\n"); 791 return err; 792 } 793 794 sdei_hp_state = err; 795 return 0; 796 } 797 798 static int sdei_device_restore(struct device *dev) 799 { 800 int err; 801 802 err = sdei_platform_reset(); 803 if (err) 804 return err; 805 806 return sdei_device_thaw(dev); 807 } 808 809 static const struct dev_pm_ops sdei_pm_ops = { 810 .suspend = sdei_device_suspend, 811 .resume = sdei_device_resume, 812 .freeze = sdei_device_freeze, 813 .thaw = sdei_device_thaw, 814 .restore = sdei_device_restore, 815 }; 816 817 /* 818 * Mask all CPUs and unregister all events on panic, reboot or kexec. 819 */ 820 static int sdei_reboot_notifier(struct notifier_block *nb, unsigned long action, 821 void *data) 822 { 823 /* 824 * We are going to reset the interface, after this there is no point 825 * doing work when we take CPUs offline. 826 */ 827 cpuhp_remove_state(sdei_hp_state); 828 829 sdei_platform_reset(); 830 831 return NOTIFY_OK; 832 } 833 834 static struct notifier_block sdei_reboot_nb = { 835 .notifier_call = sdei_reboot_notifier, 836 }; 837 838 static void sdei_smccc_smc(unsigned long function_id, 839 unsigned long arg0, unsigned long arg1, 840 unsigned long arg2, unsigned long arg3, 841 unsigned long arg4, struct arm_smccc_res *res) 842 { 843 arm_smccc_smc(function_id, arg0, arg1, arg2, arg3, arg4, 0, 0, res); 844 } 845 NOKPROBE_SYMBOL(sdei_smccc_smc); 846 847 static void sdei_smccc_hvc(unsigned long function_id, 848 unsigned long arg0, unsigned long arg1, 849 unsigned long arg2, unsigned long arg3, 850 unsigned long arg4, struct arm_smccc_res *res) 851 { 852 arm_smccc_hvc(function_id, arg0, arg1, arg2, arg3, arg4, 0, 0, res); 853 } 854 NOKPROBE_SYMBOL(sdei_smccc_hvc); 855 856 int sdei_register_ghes(struct ghes *ghes, sdei_event_callback *normal_cb, 857 sdei_event_callback *critical_cb) 858 { 859 int err; 860 u64 result; 861 u32 event_num; 862 sdei_event_callback *cb; 863 864 if (!IS_ENABLED(CONFIG_ACPI_APEI_GHES)) 865 return -EOPNOTSUPP; 866 867 event_num = ghes->generic->notify.vector; 868 if (event_num == 0) { 869 /* 870 * Event 0 is reserved by the specification for 871 * SDEI_EVENT_SIGNAL. 872 */ 873 return -EINVAL; 874 } 875 876 err = sdei_api_event_get_info(event_num, SDEI_EVENT_INFO_EV_PRIORITY, 877 &result); 878 if (err) 879 return err; 880 881 if (result == SDEI_EVENT_PRIORITY_CRITICAL) 882 cb = critical_cb; 883 else 884 cb = normal_cb; 885 886 err = sdei_event_register(event_num, cb, ghes); 887 if (!err) 888 err = sdei_event_enable(event_num); 889 890 return err; 891 } 892 893 int sdei_unregister_ghes(struct ghes *ghes) 894 { 895 int i; 896 int err; 897 u32 event_num = ghes->generic->notify.vector; 898 899 might_sleep(); 900 901 if (!IS_ENABLED(CONFIG_ACPI_APEI_GHES)) 902 return -EOPNOTSUPP; 903 904 /* 905 * The event may be running on another CPU. Disable it 906 * to stop new events, then try to unregister a few times. 907 */ 908 err = sdei_event_disable(event_num); 909 if (err) 910 return err; 911 912 for (i = 0; i < 3; i++) { 913 err = sdei_event_unregister(event_num); 914 if (err != -EINPROGRESS) 915 break; 916 917 schedule(); 918 } 919 920 return err; 921 } 922 923 static int sdei_get_conduit(struct platform_device *pdev) 924 { 925 const char *method; 926 struct device_node *np = pdev->dev.of_node; 927 928 sdei_firmware_call = NULL; 929 if (np) { 930 if (of_property_read_string(np, "method", &method)) { 931 pr_warn("missing \"method\" property\n"); 932 return SMCCC_CONDUIT_NONE; 933 } 934 935 if (!strcmp("hvc", method)) { 936 sdei_firmware_call = &sdei_smccc_hvc; 937 return SMCCC_CONDUIT_HVC; 938 } else if (!strcmp("smc", method)) { 939 sdei_firmware_call = &sdei_smccc_smc; 940 return SMCCC_CONDUIT_SMC; 941 } 942 943 pr_warn("invalid \"method\" property: %s\n", method); 944 } else if (!acpi_disabled) { 945 if (acpi_psci_use_hvc()) { 946 sdei_firmware_call = &sdei_smccc_hvc; 947 return SMCCC_CONDUIT_HVC; 948 } else { 949 sdei_firmware_call = &sdei_smccc_smc; 950 return SMCCC_CONDUIT_SMC; 951 } 952 } 953 954 return SMCCC_CONDUIT_NONE; 955 } 956 957 static int sdei_probe(struct platform_device *pdev) 958 { 959 int err; 960 u64 ver = 0; 961 int conduit; 962 963 conduit = sdei_get_conduit(pdev); 964 if (!sdei_firmware_call) 965 return 0; 966 967 err = sdei_api_get_version(&ver); 968 if (err) { 969 pr_err("Failed to get SDEI version: %d\n", err); 970 sdei_mark_interface_broken(); 971 return err; 972 } 973 974 pr_info("SDEIv%d.%d (0x%x) detected in firmware.\n", 975 (int)SDEI_VERSION_MAJOR(ver), (int)SDEI_VERSION_MINOR(ver), 976 (int)SDEI_VERSION_VENDOR(ver)); 977 978 if (SDEI_VERSION_MAJOR(ver) != 1) { 979 pr_warn("Conflicting SDEI version detected.\n"); 980 sdei_mark_interface_broken(); 981 return -EINVAL; 982 } 983 984 err = sdei_platform_reset(); 985 if (err) 986 return err; 987 988 sdei_entry_point = sdei_arch_get_entry_point(conduit); 989 if (!sdei_entry_point) { 990 /* Not supported due to hardware or boot configuration */ 991 sdei_mark_interface_broken(); 992 return 0; 993 } 994 995 err = cpu_pm_register_notifier(&sdei_pm_nb); 996 if (err) { 997 pr_warn("Failed to register CPU PM notifier...\n"); 998 goto error; 999 } 1000 1001 err = register_reboot_notifier(&sdei_reboot_nb); 1002 if (err) { 1003 pr_warn("Failed to register reboot notifier...\n"); 1004 goto remove_cpupm; 1005 } 1006 1007 err = cpuhp_setup_state(CPUHP_AP_ONLINE_DYN, "SDEI", 1008 &sdei_cpuhp_up, &sdei_cpuhp_down); 1009 if (err < 0) { 1010 pr_warn("Failed to register CPU hotplug notifier...\n"); 1011 goto remove_reboot; 1012 } 1013 1014 sdei_hp_state = err; 1015 1016 return 0; 1017 1018 remove_reboot: 1019 unregister_reboot_notifier(&sdei_reboot_nb); 1020 1021 remove_cpupm: 1022 cpu_pm_unregister_notifier(&sdei_pm_nb); 1023 1024 error: 1025 sdei_mark_interface_broken(); 1026 return err; 1027 } 1028 1029 static const struct of_device_id sdei_of_match[] = { 1030 { .compatible = "arm,sdei-1.0" }, 1031 {} 1032 }; 1033 1034 static struct platform_driver sdei_driver = { 1035 .driver = { 1036 .name = "sdei", 1037 .pm = &sdei_pm_ops, 1038 .of_match_table = sdei_of_match, 1039 }, 1040 .probe = sdei_probe, 1041 }; 1042 1043 static bool __init sdei_present_acpi(void) 1044 { 1045 acpi_status status; 1046 struct acpi_table_header *sdei_table_header; 1047 1048 if (acpi_disabled) 1049 return false; 1050 1051 status = acpi_get_table(ACPI_SIG_SDEI, 0, &sdei_table_header); 1052 if (ACPI_FAILURE(status) && status != AE_NOT_FOUND) { 1053 const char *msg = acpi_format_exception(status); 1054 1055 pr_info("Failed to get ACPI:SDEI table, %s\n", msg); 1056 } 1057 if (ACPI_FAILURE(status)) 1058 return false; 1059 1060 acpi_put_table(sdei_table_header); 1061 1062 return true; 1063 } 1064 1065 void __init sdei_init(void) 1066 { 1067 struct platform_device *pdev; 1068 int ret; 1069 1070 ret = platform_driver_register(&sdei_driver); 1071 if (ret || !sdei_present_acpi()) 1072 return; 1073 1074 pdev = platform_device_register_simple(sdei_driver.driver.name, 1075 0, NULL, 0); 1076 if (IS_ERR(pdev)) { 1077 ret = PTR_ERR(pdev); 1078 platform_driver_unregister(&sdei_driver); 1079 pr_info("Failed to register ACPI:SDEI platform device %d\n", 1080 ret); 1081 } 1082 } 1083 1084 int sdei_event_handler(struct pt_regs *regs, 1085 struct sdei_registered_event *arg) 1086 { 1087 int err; 1088 u32 event_num = arg->event_num; 1089 1090 err = arg->callback(event_num, regs, arg->callback_arg); 1091 if (err) 1092 pr_err_ratelimited("event %u on CPU %u failed with error: %d\n", 1093 event_num, smp_processor_id(), err); 1094 1095 return err; 1096 } 1097 NOKPROBE_SYMBOL(sdei_event_handler); 1098 1099 void sdei_handler_abort(void) 1100 { 1101 /* 1102 * If the crash happened in an SDEI event handler then we need to 1103 * finish the handler with the firmware so that we can have working 1104 * interrupts in the crash kernel. 1105 */ 1106 if (__this_cpu_read(sdei_active_critical_event)) { 1107 pr_warn("still in SDEI critical event context, attempting to finish handler.\n"); 1108 __sdei_handler_abort(); 1109 __this_cpu_write(sdei_active_critical_event, NULL); 1110 } 1111 if (__this_cpu_read(sdei_active_normal_event)) { 1112 pr_warn("still in SDEI normal event context, attempting to finish handler.\n"); 1113 __sdei_handler_abort(); 1114 __this_cpu_write(sdei_active_normal_event, NULL); 1115 } 1116 } 1117