1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * sysfs.c - ACPI sysfs interface to userspace. 4 */ 5 6 #define pr_fmt(fmt) "ACPI: " fmt 7 8 #include <linux/init.h> 9 #include <linux/kernel.h> 10 #include <linux/moduleparam.h> 11 #include <linux/acpi.h> 12 13 #include "internal.h" 14 15 #define _COMPONENT ACPI_SYSTEM_COMPONENT 16 ACPI_MODULE_NAME("sysfs"); 17 18 #ifdef CONFIG_ACPI_DEBUG 19 /* 20 * ACPI debug sysfs I/F, including: 21 * /sys/modules/acpi/parameters/debug_layer 22 * /sys/modules/acpi/parameters/debug_level 23 * /sys/modules/acpi/parameters/trace_method_name 24 * /sys/modules/acpi/parameters/trace_state 25 * /sys/modules/acpi/parameters/trace_debug_layer 26 * /sys/modules/acpi/parameters/trace_debug_level 27 */ 28 29 struct acpi_dlayer { 30 const char *name; 31 unsigned long value; 32 }; 33 struct acpi_dlevel { 34 const char *name; 35 unsigned long value; 36 }; 37 #define ACPI_DEBUG_INIT(v) { .name = #v, .value = v } 38 39 static const struct acpi_dlayer acpi_debug_layers[] = { 40 ACPI_DEBUG_INIT(ACPI_UTILITIES), 41 ACPI_DEBUG_INIT(ACPI_HARDWARE), 42 ACPI_DEBUG_INIT(ACPI_EVENTS), 43 ACPI_DEBUG_INIT(ACPI_TABLES), 44 ACPI_DEBUG_INIT(ACPI_NAMESPACE), 45 ACPI_DEBUG_INIT(ACPI_PARSER), 46 ACPI_DEBUG_INIT(ACPI_DISPATCHER), 47 ACPI_DEBUG_INIT(ACPI_EXECUTER), 48 ACPI_DEBUG_INIT(ACPI_RESOURCES), 49 ACPI_DEBUG_INIT(ACPI_CA_DEBUGGER), 50 ACPI_DEBUG_INIT(ACPI_OS_SERVICES), 51 ACPI_DEBUG_INIT(ACPI_CA_DISASSEMBLER), 52 ACPI_DEBUG_INIT(ACPI_COMPILER), 53 ACPI_DEBUG_INIT(ACPI_TOOLS), 54 55 ACPI_DEBUG_INIT(ACPI_BUS_COMPONENT), 56 ACPI_DEBUG_INIT(ACPI_AC_COMPONENT), 57 ACPI_DEBUG_INIT(ACPI_BATTERY_COMPONENT), 58 ACPI_DEBUG_INIT(ACPI_BUTTON_COMPONENT), 59 ACPI_DEBUG_INIT(ACPI_SBS_COMPONENT), 60 ACPI_DEBUG_INIT(ACPI_FAN_COMPONENT), 61 ACPI_DEBUG_INIT(ACPI_PCI_COMPONENT), 62 ACPI_DEBUG_INIT(ACPI_POWER_COMPONENT), 63 ACPI_DEBUG_INIT(ACPI_CONTAINER_COMPONENT), 64 ACPI_DEBUG_INIT(ACPI_SYSTEM_COMPONENT), 65 ACPI_DEBUG_INIT(ACPI_THERMAL_COMPONENT), 66 ACPI_DEBUG_INIT(ACPI_MEMORY_DEVICE_COMPONENT), 67 ACPI_DEBUG_INIT(ACPI_VIDEO_COMPONENT), 68 ACPI_DEBUG_INIT(ACPI_PROCESSOR_COMPONENT), 69 }; 70 71 static const struct acpi_dlevel acpi_debug_levels[] = { 72 ACPI_DEBUG_INIT(ACPI_LV_INIT), 73 ACPI_DEBUG_INIT(ACPI_LV_DEBUG_OBJECT), 74 ACPI_DEBUG_INIT(ACPI_LV_INFO), 75 ACPI_DEBUG_INIT(ACPI_LV_REPAIR), 76 ACPI_DEBUG_INIT(ACPI_LV_TRACE_POINT), 77 78 ACPI_DEBUG_INIT(ACPI_LV_INIT_NAMES), 79 ACPI_DEBUG_INIT(ACPI_LV_PARSE), 80 ACPI_DEBUG_INIT(ACPI_LV_LOAD), 81 ACPI_DEBUG_INIT(ACPI_LV_DISPATCH), 82 ACPI_DEBUG_INIT(ACPI_LV_EXEC), 83 ACPI_DEBUG_INIT(ACPI_LV_NAMES), 84 ACPI_DEBUG_INIT(ACPI_LV_OPREGION), 85 ACPI_DEBUG_INIT(ACPI_LV_BFIELD), 86 ACPI_DEBUG_INIT(ACPI_LV_TABLES), 87 ACPI_DEBUG_INIT(ACPI_LV_VALUES), 88 ACPI_DEBUG_INIT(ACPI_LV_OBJECTS), 89 ACPI_DEBUG_INIT(ACPI_LV_RESOURCES), 90 ACPI_DEBUG_INIT(ACPI_LV_USER_REQUESTS), 91 ACPI_DEBUG_INIT(ACPI_LV_PACKAGE), 92 93 ACPI_DEBUG_INIT(ACPI_LV_ALLOCATIONS), 94 ACPI_DEBUG_INIT(ACPI_LV_FUNCTIONS), 95 ACPI_DEBUG_INIT(ACPI_LV_OPTIMIZATIONS), 96 97 ACPI_DEBUG_INIT(ACPI_LV_MUTEX), 98 ACPI_DEBUG_INIT(ACPI_LV_THREADS), 99 ACPI_DEBUG_INIT(ACPI_LV_IO), 100 ACPI_DEBUG_INIT(ACPI_LV_INTERRUPTS), 101 102 ACPI_DEBUG_INIT(ACPI_LV_AML_DISASSEMBLE), 103 ACPI_DEBUG_INIT(ACPI_LV_VERBOSE_INFO), 104 ACPI_DEBUG_INIT(ACPI_LV_FULL_TABLES), 105 ACPI_DEBUG_INIT(ACPI_LV_EVENTS), 106 }; 107 108 static int param_get_debug_layer(char *buffer, const struct kernel_param *kp) 109 { 110 int result = 0; 111 int i; 112 113 result = sprintf(buffer, "%-25s\tHex SET\n", "Description"); 114 115 for (i = 0; i < ARRAY_SIZE(acpi_debug_layers); i++) { 116 result += sprintf(buffer + result, "%-25s\t0x%08lX [%c]\n", 117 acpi_debug_layers[i].name, 118 acpi_debug_layers[i].value, 119 (acpi_dbg_layer & acpi_debug_layers[i].value) 120 ? '*' : ' '); 121 } 122 result += 123 sprintf(buffer + result, "%-25s\t0x%08X [%c]\n", "ACPI_ALL_DRIVERS", 124 ACPI_ALL_DRIVERS, 125 (acpi_dbg_layer & ACPI_ALL_DRIVERS) == 126 ACPI_ALL_DRIVERS ? '*' : (acpi_dbg_layer & ACPI_ALL_DRIVERS) 127 == 0 ? ' ' : '-'); 128 result += 129 sprintf(buffer + result, 130 "--\ndebug_layer = 0x%08X ( * = enabled)\n", 131 acpi_dbg_layer); 132 133 return result; 134 } 135 136 static int param_get_debug_level(char *buffer, const struct kernel_param *kp) 137 { 138 int result = 0; 139 int i; 140 141 result = sprintf(buffer, "%-25s\tHex SET\n", "Description"); 142 143 for (i = 0; i < ARRAY_SIZE(acpi_debug_levels); i++) { 144 result += sprintf(buffer + result, "%-25s\t0x%08lX [%c]\n", 145 acpi_debug_levels[i].name, 146 acpi_debug_levels[i].value, 147 (acpi_dbg_level & acpi_debug_levels[i].value) 148 ? '*' : ' '); 149 } 150 result += 151 sprintf(buffer + result, "--\ndebug_level = 0x%08X (* = enabled)\n", 152 acpi_dbg_level); 153 154 return result; 155 } 156 157 static const struct kernel_param_ops param_ops_debug_layer = { 158 .set = param_set_uint, 159 .get = param_get_debug_layer, 160 }; 161 162 static const struct kernel_param_ops param_ops_debug_level = { 163 .set = param_set_uint, 164 .get = param_get_debug_level, 165 }; 166 167 module_param_cb(debug_layer, ¶m_ops_debug_layer, &acpi_dbg_layer, 0644); 168 module_param_cb(debug_level, ¶m_ops_debug_level, &acpi_dbg_level, 0644); 169 170 static char trace_method_name[1024]; 171 172 static int param_set_trace_method_name(const char *val, 173 const struct kernel_param *kp) 174 { 175 u32 saved_flags = 0; 176 bool is_abs_path = true; 177 178 if (*val != '\\') 179 is_abs_path = false; 180 181 if ((is_abs_path && strlen(val) > 1023) || 182 (!is_abs_path && strlen(val) > 1022)) { 183 pr_err("%s: string parameter too long\n", kp->name); 184 return -ENOSPC; 185 } 186 187 /* 188 * It's not safe to update acpi_gbl_trace_method_name without 189 * having the tracer stopped, so we save the original tracer 190 * state and disable it. 191 */ 192 saved_flags = acpi_gbl_trace_flags; 193 (void)acpi_debug_trace(NULL, 194 acpi_gbl_trace_dbg_level, 195 acpi_gbl_trace_dbg_layer, 196 0); 197 198 /* This is a hack. We can't kmalloc in early boot. */ 199 if (is_abs_path) 200 strcpy(trace_method_name, val); 201 else { 202 trace_method_name[0] = '\\'; 203 strcpy(trace_method_name+1, val); 204 } 205 206 /* Restore the original tracer state */ 207 (void)acpi_debug_trace(trace_method_name, 208 acpi_gbl_trace_dbg_level, 209 acpi_gbl_trace_dbg_layer, 210 saved_flags); 211 212 return 0; 213 } 214 215 static int param_get_trace_method_name(char *buffer, const struct kernel_param *kp) 216 { 217 return scnprintf(buffer, PAGE_SIZE, "%s", acpi_gbl_trace_method_name); 218 } 219 220 static const struct kernel_param_ops param_ops_trace_method = { 221 .set = param_set_trace_method_name, 222 .get = param_get_trace_method_name, 223 }; 224 225 static const struct kernel_param_ops param_ops_trace_attrib = { 226 .set = param_set_uint, 227 .get = param_get_uint, 228 }; 229 230 module_param_cb(trace_method_name, ¶m_ops_trace_method, &trace_method_name, 0644); 231 module_param_cb(trace_debug_layer, ¶m_ops_trace_attrib, &acpi_gbl_trace_dbg_layer, 0644); 232 module_param_cb(trace_debug_level, ¶m_ops_trace_attrib, &acpi_gbl_trace_dbg_level, 0644); 233 234 static int param_set_trace_state(const char *val, 235 const struct kernel_param *kp) 236 { 237 acpi_status status; 238 const char *method = trace_method_name; 239 u32 flags = 0; 240 241 /* So "xxx-once" comparison should go prior than "xxx" comparison */ 242 #define acpi_compare_param(val, key) \ 243 strncmp((val), (key), sizeof(key) - 1) 244 245 if (!acpi_compare_param(val, "enable")) { 246 method = NULL; 247 flags = ACPI_TRACE_ENABLED; 248 } else if (!acpi_compare_param(val, "disable")) 249 method = NULL; 250 else if (!acpi_compare_param(val, "method-once")) 251 flags = ACPI_TRACE_ENABLED | ACPI_TRACE_ONESHOT; 252 else if (!acpi_compare_param(val, "method")) 253 flags = ACPI_TRACE_ENABLED; 254 else if (!acpi_compare_param(val, "opcode-once")) 255 flags = ACPI_TRACE_ENABLED | ACPI_TRACE_ONESHOT | ACPI_TRACE_OPCODE; 256 else if (!acpi_compare_param(val, "opcode")) 257 flags = ACPI_TRACE_ENABLED | ACPI_TRACE_OPCODE; 258 else 259 return -EINVAL; 260 261 status = acpi_debug_trace(method, 262 acpi_gbl_trace_dbg_level, 263 acpi_gbl_trace_dbg_layer, 264 flags); 265 if (ACPI_FAILURE(status)) 266 return -EBUSY; 267 268 return 0; 269 } 270 271 static int param_get_trace_state(char *buffer, const struct kernel_param *kp) 272 { 273 if (!(acpi_gbl_trace_flags & ACPI_TRACE_ENABLED)) 274 return sprintf(buffer, "disable"); 275 else { 276 if (acpi_gbl_trace_method_name) { 277 if (acpi_gbl_trace_flags & ACPI_TRACE_ONESHOT) 278 return sprintf(buffer, "method-once"); 279 else 280 return sprintf(buffer, "method"); 281 } else 282 return sprintf(buffer, "enable"); 283 } 284 return 0; 285 } 286 287 module_param_call(trace_state, param_set_trace_state, param_get_trace_state, 288 NULL, 0644); 289 #endif /* CONFIG_ACPI_DEBUG */ 290 291 292 /* /sys/modules/acpi/parameters/aml_debug_output */ 293 294 module_param_named(aml_debug_output, acpi_gbl_enable_aml_debug_object, 295 byte, 0644); 296 MODULE_PARM_DESC(aml_debug_output, 297 "To enable/disable the ACPI Debug Object output."); 298 299 /* /sys/module/acpi/parameters/acpica_version */ 300 static int param_get_acpica_version(char *buffer, 301 const struct kernel_param *kp) 302 { 303 int result; 304 305 result = sprintf(buffer, "%x", ACPI_CA_VERSION); 306 307 return result; 308 } 309 310 module_param_call(acpica_version, NULL, param_get_acpica_version, NULL, 0444); 311 312 /* 313 * ACPI table sysfs I/F: 314 * /sys/firmware/acpi/tables/ 315 * /sys/firmware/acpi/tables/data/ 316 * /sys/firmware/acpi/tables/dynamic/ 317 */ 318 319 static LIST_HEAD(acpi_table_attr_list); 320 static struct kobject *tables_kobj; 321 static struct kobject *tables_data_kobj; 322 static struct kobject *dynamic_tables_kobj; 323 static struct kobject *hotplug_kobj; 324 325 #define ACPI_MAX_TABLE_INSTANCES 999 326 #define ACPI_INST_SIZE 4 /* including trailing 0 */ 327 328 struct acpi_table_attr { 329 struct bin_attribute attr; 330 char name[ACPI_NAME_SIZE]; 331 int instance; 332 char filename[ACPI_NAME_SIZE+ACPI_INST_SIZE]; 333 struct list_head node; 334 }; 335 336 struct acpi_data_attr { 337 struct bin_attribute attr; 338 u64 addr; 339 }; 340 341 static ssize_t acpi_table_show(struct file *filp, struct kobject *kobj, 342 struct bin_attribute *bin_attr, char *buf, 343 loff_t offset, size_t count) 344 { 345 struct acpi_table_attr *table_attr = 346 container_of(bin_attr, struct acpi_table_attr, attr); 347 struct acpi_table_header *table_header = NULL; 348 acpi_status status; 349 ssize_t rc; 350 351 status = acpi_get_table(table_attr->name, table_attr->instance, 352 &table_header); 353 if (ACPI_FAILURE(status)) 354 return -ENODEV; 355 356 rc = memory_read_from_buffer(buf, count, &offset, table_header, 357 table_header->length); 358 acpi_put_table(table_header); 359 return rc; 360 } 361 362 static int acpi_table_attr_init(struct kobject *tables_obj, 363 struct acpi_table_attr *table_attr, 364 struct acpi_table_header *table_header) 365 { 366 struct acpi_table_header *header = NULL; 367 struct acpi_table_attr *attr = NULL; 368 char instance_str[ACPI_INST_SIZE]; 369 370 sysfs_attr_init(&table_attr->attr.attr); 371 ACPI_MOVE_NAME(table_attr->name, table_header->signature); 372 373 list_for_each_entry(attr, &acpi_table_attr_list, node) { 374 if (ACPI_COMPARE_NAME(table_attr->name, attr->name)) 375 if (table_attr->instance < attr->instance) 376 table_attr->instance = attr->instance; 377 } 378 table_attr->instance++; 379 if (table_attr->instance > ACPI_MAX_TABLE_INSTANCES) { 380 pr_warn("%4.4s: too many table instances\n", 381 table_attr->name); 382 return -ERANGE; 383 } 384 385 ACPI_MOVE_NAME(table_attr->filename, table_header->signature); 386 table_attr->filename[ACPI_NAME_SIZE] = '\0'; 387 if (table_attr->instance > 1 || (table_attr->instance == 1 && 388 !acpi_get_table 389 (table_header->signature, 2, &header))) { 390 snprintf(instance_str, sizeof(instance_str), "%u", 391 table_attr->instance); 392 strcat(table_attr->filename, instance_str); 393 } 394 395 table_attr->attr.size = table_header->length; 396 table_attr->attr.read = acpi_table_show; 397 table_attr->attr.attr.name = table_attr->filename; 398 table_attr->attr.attr.mode = 0400; 399 400 return sysfs_create_bin_file(tables_obj, &table_attr->attr); 401 } 402 403 acpi_status acpi_sysfs_table_handler(u32 event, void *table, void *context) 404 { 405 struct acpi_table_attr *table_attr; 406 407 switch (event) { 408 case ACPI_TABLE_EVENT_INSTALL: 409 table_attr = 410 kzalloc(sizeof(struct acpi_table_attr), GFP_KERNEL); 411 if (!table_attr) 412 return AE_NO_MEMORY; 413 414 if (acpi_table_attr_init(dynamic_tables_kobj, 415 table_attr, table)) { 416 kfree(table_attr); 417 return AE_ERROR; 418 } 419 list_add_tail(&table_attr->node, &acpi_table_attr_list); 420 break; 421 case ACPI_TABLE_EVENT_LOAD: 422 case ACPI_TABLE_EVENT_UNLOAD: 423 case ACPI_TABLE_EVENT_UNINSTALL: 424 /* 425 * we do not need to do anything right now 426 * because the table is not deleted from the 427 * global table list when unloading it. 428 */ 429 break; 430 default: 431 return AE_BAD_PARAMETER; 432 } 433 return AE_OK; 434 } 435 436 static ssize_t acpi_data_show(struct file *filp, struct kobject *kobj, 437 struct bin_attribute *bin_attr, char *buf, 438 loff_t offset, size_t count) 439 { 440 struct acpi_data_attr *data_attr; 441 void __iomem *base; 442 ssize_t rc; 443 444 data_attr = container_of(bin_attr, struct acpi_data_attr, attr); 445 446 base = acpi_os_map_memory(data_attr->addr, data_attr->attr.size); 447 if (!base) 448 return -ENOMEM; 449 rc = memory_read_from_buffer(buf, count, &offset, base, 450 data_attr->attr.size); 451 acpi_os_unmap_memory(base, data_attr->attr.size); 452 453 return rc; 454 } 455 456 static int acpi_bert_data_init(void *th, struct acpi_data_attr *data_attr) 457 { 458 struct acpi_table_bert *bert = th; 459 460 if (bert->header.length < sizeof(struct acpi_table_bert) || 461 bert->region_length < sizeof(struct acpi_hest_generic_status)) { 462 kfree(data_attr); 463 return -EINVAL; 464 } 465 data_attr->addr = bert->address; 466 data_attr->attr.size = bert->region_length; 467 data_attr->attr.attr.name = "BERT"; 468 469 return sysfs_create_bin_file(tables_data_kobj, &data_attr->attr); 470 } 471 472 static struct acpi_data_obj { 473 char *name; 474 int (*fn)(void *, struct acpi_data_attr *); 475 } acpi_data_objs[] = { 476 { ACPI_SIG_BERT, acpi_bert_data_init }, 477 }; 478 479 #define NUM_ACPI_DATA_OBJS ARRAY_SIZE(acpi_data_objs) 480 481 static int acpi_table_data_init(struct acpi_table_header *th) 482 { 483 struct acpi_data_attr *data_attr; 484 int i; 485 486 for (i = 0; i < NUM_ACPI_DATA_OBJS; i++) { 487 if (ACPI_COMPARE_NAME(th->signature, acpi_data_objs[i].name)) { 488 data_attr = kzalloc(sizeof(*data_attr), GFP_KERNEL); 489 if (!data_attr) 490 return -ENOMEM; 491 sysfs_attr_init(&data_attr->attr.attr); 492 data_attr->attr.read = acpi_data_show; 493 data_attr->attr.attr.mode = 0400; 494 return acpi_data_objs[i].fn(th, data_attr); 495 } 496 } 497 return 0; 498 } 499 500 static int acpi_tables_sysfs_init(void) 501 { 502 struct acpi_table_attr *table_attr; 503 struct acpi_table_header *table_header = NULL; 504 int table_index; 505 acpi_status status; 506 int ret; 507 508 tables_kobj = kobject_create_and_add("tables", acpi_kobj); 509 if (!tables_kobj) 510 goto err; 511 512 tables_data_kobj = kobject_create_and_add("data", tables_kobj); 513 if (!tables_data_kobj) 514 goto err_tables_data; 515 516 dynamic_tables_kobj = kobject_create_and_add("dynamic", tables_kobj); 517 if (!dynamic_tables_kobj) 518 goto err_dynamic_tables; 519 520 for (table_index = 0;; table_index++) { 521 status = acpi_get_table_by_index(table_index, &table_header); 522 523 if (status == AE_BAD_PARAMETER) 524 break; 525 526 if (ACPI_FAILURE(status)) 527 continue; 528 529 table_attr = kzalloc(sizeof(*table_attr), GFP_KERNEL); 530 if (!table_attr) 531 return -ENOMEM; 532 533 ret = acpi_table_attr_init(tables_kobj, 534 table_attr, table_header); 535 if (ret) { 536 kfree(table_attr); 537 return ret; 538 } 539 list_add_tail(&table_attr->node, &acpi_table_attr_list); 540 acpi_table_data_init(table_header); 541 } 542 543 kobject_uevent(tables_kobj, KOBJ_ADD); 544 kobject_uevent(tables_data_kobj, KOBJ_ADD); 545 kobject_uevent(dynamic_tables_kobj, KOBJ_ADD); 546 547 return 0; 548 err_dynamic_tables: 549 kobject_put(tables_data_kobj); 550 err_tables_data: 551 kobject_put(tables_kobj); 552 err: 553 return -ENOMEM; 554 } 555 556 /* 557 * Detailed ACPI IRQ counters: 558 * /sys/firmware/acpi/interrupts/ 559 */ 560 561 u32 acpi_irq_handled; 562 u32 acpi_irq_not_handled; 563 564 #define COUNT_GPE 0 565 #define COUNT_SCI 1 /* acpi_irq_handled */ 566 #define COUNT_SCI_NOT 2 /* acpi_irq_not_handled */ 567 #define COUNT_ERROR 3 /* other */ 568 #define NUM_COUNTERS_EXTRA 4 569 570 struct event_counter { 571 u32 count; 572 u32 flags; 573 }; 574 575 static struct event_counter *all_counters; 576 static u32 num_gpes; 577 static u32 num_counters; 578 static struct attribute **all_attrs; 579 static u32 acpi_gpe_count; 580 581 static struct attribute_group interrupt_stats_attr_group = { 582 .name = "interrupts", 583 }; 584 585 static struct kobj_attribute *counter_attrs; 586 587 static void delete_gpe_attr_array(void) 588 { 589 struct event_counter *tmp = all_counters; 590 591 all_counters = NULL; 592 kfree(tmp); 593 594 if (counter_attrs) { 595 int i; 596 597 for (i = 0; i < num_gpes; i++) 598 kfree(counter_attrs[i].attr.name); 599 600 kfree(counter_attrs); 601 } 602 kfree(all_attrs); 603 604 return; 605 } 606 607 static void gpe_count(u32 gpe_number) 608 { 609 acpi_gpe_count++; 610 611 if (!all_counters) 612 return; 613 614 if (gpe_number < num_gpes) 615 all_counters[gpe_number].count++; 616 else 617 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + 618 COUNT_ERROR].count++; 619 620 return; 621 } 622 623 static void fixed_event_count(u32 event_number) 624 { 625 if (!all_counters) 626 return; 627 628 if (event_number < ACPI_NUM_FIXED_EVENTS) 629 all_counters[num_gpes + event_number].count++; 630 else 631 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + 632 COUNT_ERROR].count++; 633 634 return; 635 } 636 637 static void acpi_global_event_handler(u32 event_type, acpi_handle device, 638 u32 event_number, void *context) 639 { 640 if (event_type == ACPI_EVENT_TYPE_GPE) { 641 gpe_count(event_number); 642 pr_debug("GPE event 0x%02x\n", event_number); 643 } else if (event_type == ACPI_EVENT_TYPE_FIXED) { 644 fixed_event_count(event_number); 645 pr_debug("Fixed event 0x%02x\n", event_number); 646 } else { 647 pr_debug("Other event 0x%02x\n", event_number); 648 } 649 } 650 651 static int get_status(u32 index, acpi_event_status *status, 652 acpi_handle *handle) 653 { 654 int result; 655 656 if (index >= num_gpes + ACPI_NUM_FIXED_EVENTS) 657 return -EINVAL; 658 659 if (index < num_gpes) { 660 result = acpi_get_gpe_device(index, handle); 661 if (result) { 662 ACPI_EXCEPTION((AE_INFO, AE_NOT_FOUND, 663 "Invalid GPE 0x%x", index)); 664 return result; 665 } 666 result = acpi_get_gpe_status(*handle, index, status); 667 } else if (index < (num_gpes + ACPI_NUM_FIXED_EVENTS)) 668 result = acpi_get_event_status(index - num_gpes, status); 669 670 return result; 671 } 672 673 static ssize_t counter_show(struct kobject *kobj, 674 struct kobj_attribute *attr, char *buf) 675 { 676 int index = attr - counter_attrs; 677 int size; 678 acpi_handle handle; 679 acpi_event_status status; 680 int result = 0; 681 682 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI].count = 683 acpi_irq_handled; 684 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI_NOT].count = 685 acpi_irq_not_handled; 686 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_GPE].count = 687 acpi_gpe_count; 688 size = sprintf(buf, "%8u", all_counters[index].count); 689 690 /* "gpe_all" or "sci" */ 691 if (index >= num_gpes + ACPI_NUM_FIXED_EVENTS) 692 goto end; 693 694 result = get_status(index, &status, &handle); 695 if (result) 696 goto end; 697 698 if (status & ACPI_EVENT_FLAG_ENABLE_SET) 699 size += sprintf(buf + size, " EN"); 700 else 701 size += sprintf(buf + size, " "); 702 if (status & ACPI_EVENT_FLAG_STATUS_SET) 703 size += sprintf(buf + size, " STS"); 704 else 705 size += sprintf(buf + size, " "); 706 707 if (!(status & ACPI_EVENT_FLAG_HAS_HANDLER)) 708 size += sprintf(buf + size, " invalid "); 709 else if (status & ACPI_EVENT_FLAG_ENABLED) 710 size += sprintf(buf + size, " enabled "); 711 else if (status & ACPI_EVENT_FLAG_WAKE_ENABLED) 712 size += sprintf(buf + size, " wake_enabled"); 713 else 714 size += sprintf(buf + size, " disabled "); 715 if (status & ACPI_EVENT_FLAG_MASKED) 716 size += sprintf(buf + size, " masked "); 717 else 718 size += sprintf(buf + size, " unmasked"); 719 720 end: 721 size += sprintf(buf + size, "\n"); 722 return result ? result : size; 723 } 724 725 /* 726 * counter_set() sets the specified counter. 727 * setting the total "sci" file to any value clears all counters. 728 * enable/disable/clear a gpe/fixed event in user space. 729 */ 730 static ssize_t counter_set(struct kobject *kobj, 731 struct kobj_attribute *attr, const char *buf, 732 size_t size) 733 { 734 int index = attr - counter_attrs; 735 acpi_event_status status; 736 acpi_handle handle; 737 int result = 0; 738 unsigned long tmp; 739 740 if (index == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI) { 741 int i; 742 for (i = 0; i < num_counters; ++i) 743 all_counters[i].count = 0; 744 acpi_gpe_count = 0; 745 acpi_irq_handled = 0; 746 acpi_irq_not_handled = 0; 747 goto end; 748 } 749 750 /* show the event status for both GPEs and Fixed Events */ 751 result = get_status(index, &status, &handle); 752 if (result) 753 goto end; 754 755 if (!(status & ACPI_EVENT_FLAG_HAS_HANDLER)) { 756 printk(KERN_WARNING PREFIX 757 "Can not change Invalid GPE/Fixed Event status\n"); 758 return -EINVAL; 759 } 760 761 if (index < num_gpes) { 762 if (!strcmp(buf, "disable\n") && 763 (status & ACPI_EVENT_FLAG_ENABLED)) 764 result = acpi_disable_gpe(handle, index); 765 else if (!strcmp(buf, "enable\n") && 766 !(status & ACPI_EVENT_FLAG_ENABLED)) 767 result = acpi_enable_gpe(handle, index); 768 else if (!strcmp(buf, "clear\n") && 769 (status & ACPI_EVENT_FLAG_STATUS_SET)) 770 result = acpi_clear_gpe(handle, index); 771 else if (!strcmp(buf, "mask\n")) 772 result = acpi_mask_gpe(handle, index, TRUE); 773 else if (!strcmp(buf, "unmask\n")) 774 result = acpi_mask_gpe(handle, index, FALSE); 775 else if (!kstrtoul(buf, 0, &tmp)) 776 all_counters[index].count = tmp; 777 else 778 result = -EINVAL; 779 } else if (index < num_gpes + ACPI_NUM_FIXED_EVENTS) { 780 int event = index - num_gpes; 781 if (!strcmp(buf, "disable\n") && 782 (status & ACPI_EVENT_FLAG_ENABLE_SET)) 783 result = acpi_disable_event(event, ACPI_NOT_ISR); 784 else if (!strcmp(buf, "enable\n") && 785 !(status & ACPI_EVENT_FLAG_ENABLE_SET)) 786 result = acpi_enable_event(event, ACPI_NOT_ISR); 787 else if (!strcmp(buf, "clear\n") && 788 (status & ACPI_EVENT_FLAG_STATUS_SET)) 789 result = acpi_clear_event(event); 790 else if (!kstrtoul(buf, 0, &tmp)) 791 all_counters[index].count = tmp; 792 else 793 result = -EINVAL; 794 } else 795 all_counters[index].count = strtoul(buf, NULL, 0); 796 797 if (ACPI_FAILURE(result)) 798 result = -EINVAL; 799 end: 800 return result ? result : size; 801 } 802 803 /* 804 * A Quirk Mechanism for GPE Flooding Prevention: 805 * 806 * Quirks may be needed to prevent GPE flooding on a specific GPE. The 807 * flooding typically cannot be detected and automatically prevented by 808 * ACPI_GPE_DISPATCH_NONE check because there is a _Lxx/_Exx prepared in 809 * the AML tables. This normally indicates a feature gap in Linux, thus 810 * instead of providing endless quirk tables, we provide a boot parameter 811 * for those who want this quirk. For example, if the users want to prevent 812 * the GPE flooding for GPE 00, they need to specify the following boot 813 * parameter: 814 * acpi_mask_gpe=0x00 815 * The masking status can be modified by the following runtime controlling 816 * interface: 817 * echo unmask > /sys/firmware/acpi/interrupts/gpe00 818 */ 819 #define ACPI_MASKABLE_GPE_MAX 0xFF 820 static DECLARE_BITMAP(acpi_masked_gpes_map, ACPI_MASKABLE_GPE_MAX) __initdata; 821 822 static int __init acpi_gpe_set_masked_gpes(char *val) 823 { 824 u8 gpe; 825 826 if (kstrtou8(val, 0, &gpe) || gpe > ACPI_MASKABLE_GPE_MAX) 827 return -EINVAL; 828 set_bit(gpe, acpi_masked_gpes_map); 829 830 return 1; 831 } 832 __setup("acpi_mask_gpe=", acpi_gpe_set_masked_gpes); 833 834 void __init acpi_gpe_apply_masked_gpes(void) 835 { 836 acpi_handle handle; 837 acpi_status status; 838 u8 gpe; 839 840 for_each_set_bit(gpe, acpi_masked_gpes_map, ACPI_MASKABLE_GPE_MAX) { 841 status = acpi_get_gpe_device(gpe, &handle); 842 if (ACPI_SUCCESS(status)) { 843 pr_info("Masking GPE 0x%x.\n", gpe); 844 (void)acpi_mask_gpe(handle, gpe, TRUE); 845 } 846 } 847 } 848 849 void acpi_irq_stats_init(void) 850 { 851 acpi_status status; 852 int i; 853 854 if (all_counters) 855 return; 856 857 num_gpes = acpi_current_gpe_count; 858 num_counters = num_gpes + ACPI_NUM_FIXED_EVENTS + NUM_COUNTERS_EXTRA; 859 860 all_attrs = kcalloc(num_counters + 1, sizeof(struct attribute *), 861 GFP_KERNEL); 862 if (all_attrs == NULL) 863 return; 864 865 all_counters = kcalloc(num_counters, sizeof(struct event_counter), 866 GFP_KERNEL); 867 if (all_counters == NULL) 868 goto fail; 869 870 status = acpi_install_global_event_handler(acpi_global_event_handler, NULL); 871 if (ACPI_FAILURE(status)) 872 goto fail; 873 874 counter_attrs = kcalloc(num_counters, sizeof(struct kobj_attribute), 875 GFP_KERNEL); 876 if (counter_attrs == NULL) 877 goto fail; 878 879 for (i = 0; i < num_counters; ++i) { 880 char buffer[12]; 881 char *name; 882 883 if (i < num_gpes) 884 sprintf(buffer, "gpe%02X", i); 885 else if (i == num_gpes + ACPI_EVENT_PMTIMER) 886 sprintf(buffer, "ff_pmtimer"); 887 else if (i == num_gpes + ACPI_EVENT_GLOBAL) 888 sprintf(buffer, "ff_gbl_lock"); 889 else if (i == num_gpes + ACPI_EVENT_POWER_BUTTON) 890 sprintf(buffer, "ff_pwr_btn"); 891 else if (i == num_gpes + ACPI_EVENT_SLEEP_BUTTON) 892 sprintf(buffer, "ff_slp_btn"); 893 else if (i == num_gpes + ACPI_EVENT_RTC) 894 sprintf(buffer, "ff_rt_clk"); 895 else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_GPE) 896 sprintf(buffer, "gpe_all"); 897 else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI) 898 sprintf(buffer, "sci"); 899 else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI_NOT) 900 sprintf(buffer, "sci_not"); 901 else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_ERROR) 902 sprintf(buffer, "error"); 903 else 904 sprintf(buffer, "bug%02X", i); 905 906 name = kstrdup(buffer, GFP_KERNEL); 907 if (name == NULL) 908 goto fail; 909 910 sysfs_attr_init(&counter_attrs[i].attr); 911 counter_attrs[i].attr.name = name; 912 counter_attrs[i].attr.mode = 0644; 913 counter_attrs[i].show = counter_show; 914 counter_attrs[i].store = counter_set; 915 916 all_attrs[i] = &counter_attrs[i].attr; 917 } 918 919 interrupt_stats_attr_group.attrs = all_attrs; 920 if (!sysfs_create_group(acpi_kobj, &interrupt_stats_attr_group)) 921 return; 922 923 fail: 924 delete_gpe_attr_array(); 925 return; 926 } 927 928 static void __exit interrupt_stats_exit(void) 929 { 930 sysfs_remove_group(acpi_kobj, &interrupt_stats_attr_group); 931 932 delete_gpe_attr_array(); 933 934 return; 935 } 936 937 static ssize_t 938 acpi_show_profile(struct device *dev, struct device_attribute *attr, 939 char *buf) 940 { 941 return sprintf(buf, "%d\n", acpi_gbl_FADT.preferred_profile); 942 } 943 944 static const struct device_attribute pm_profile_attr = 945 __ATTR(pm_profile, S_IRUGO, acpi_show_profile, NULL); 946 947 static ssize_t hotplug_enabled_show(struct kobject *kobj, 948 struct kobj_attribute *attr, char *buf) 949 { 950 struct acpi_hotplug_profile *hotplug = to_acpi_hotplug_profile(kobj); 951 952 return sprintf(buf, "%d\n", hotplug->enabled); 953 } 954 955 static ssize_t hotplug_enabled_store(struct kobject *kobj, 956 struct kobj_attribute *attr, 957 const char *buf, size_t size) 958 { 959 struct acpi_hotplug_profile *hotplug = to_acpi_hotplug_profile(kobj); 960 unsigned int val; 961 962 if (kstrtouint(buf, 10, &val) || val > 1) 963 return -EINVAL; 964 965 acpi_scan_hotplug_enabled(hotplug, val); 966 return size; 967 } 968 969 static struct kobj_attribute hotplug_enabled_attr = 970 __ATTR(enabled, S_IRUGO | S_IWUSR, hotplug_enabled_show, 971 hotplug_enabled_store); 972 973 static struct attribute *hotplug_profile_attrs[] = { 974 &hotplug_enabled_attr.attr, 975 NULL 976 }; 977 978 static struct kobj_type acpi_hotplug_profile_ktype = { 979 .sysfs_ops = &kobj_sysfs_ops, 980 .default_attrs = hotplug_profile_attrs, 981 }; 982 983 void acpi_sysfs_add_hotplug_profile(struct acpi_hotplug_profile *hotplug, 984 const char *name) 985 { 986 int error; 987 988 if (!hotplug_kobj) 989 goto err_out; 990 991 error = kobject_init_and_add(&hotplug->kobj, 992 &acpi_hotplug_profile_ktype, hotplug_kobj, "%s", name); 993 if (error) 994 goto err_out; 995 996 kobject_uevent(&hotplug->kobj, KOBJ_ADD); 997 return; 998 999 err_out: 1000 pr_err(PREFIX "Unable to add hotplug profile '%s'\n", name); 1001 } 1002 1003 static ssize_t force_remove_show(struct kobject *kobj, 1004 struct kobj_attribute *attr, char *buf) 1005 { 1006 return sprintf(buf, "%d\n", 0); 1007 } 1008 1009 static ssize_t force_remove_store(struct kobject *kobj, 1010 struct kobj_attribute *attr, 1011 const char *buf, size_t size) 1012 { 1013 bool val; 1014 int ret; 1015 1016 ret = strtobool(buf, &val); 1017 if (ret < 0) 1018 return ret; 1019 1020 if (val) { 1021 pr_err("Enabling force_remove is not supported anymore. Please report to linux-acpi@vger.kernel.org if you depend on this functionality\n"); 1022 return -EINVAL; 1023 } 1024 return size; 1025 } 1026 1027 static const struct kobj_attribute force_remove_attr = 1028 __ATTR(force_remove, S_IRUGO | S_IWUSR, force_remove_show, 1029 force_remove_store); 1030 1031 int __init acpi_sysfs_init(void) 1032 { 1033 int result; 1034 1035 result = acpi_tables_sysfs_init(); 1036 if (result) 1037 return result; 1038 1039 hotplug_kobj = kobject_create_and_add("hotplug", acpi_kobj); 1040 if (!hotplug_kobj) 1041 return -ENOMEM; 1042 1043 result = sysfs_create_file(hotplug_kobj, &force_remove_attr.attr); 1044 if (result) 1045 return result; 1046 1047 result = sysfs_create_file(acpi_kobj, &pm_profile_attr.attr); 1048 return result; 1049 } 1050