1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * coretemp.c - Linux kernel module for hardware monitoring 4 * 5 * Copyright (C) 2007 Rudolf Marek <r.marek@assembler.cz> 6 * 7 * Inspired from many hwmon drivers 8 */ 9 10 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 11 12 #include <linux/module.h> 13 #include <linux/init.h> 14 #include <linux/slab.h> 15 #include <linux/jiffies.h> 16 #include <linux/hwmon.h> 17 #include <linux/sysfs.h> 18 #include <linux/hwmon-sysfs.h> 19 #include <linux/err.h> 20 #include <linux/mutex.h> 21 #include <linux/list.h> 22 #include <linux/platform_device.h> 23 #include <linux/cpu.h> 24 #include <linux/smp.h> 25 #include <linux/moduleparam.h> 26 #include <linux/pci.h> 27 #include <asm/msr.h> 28 #include <asm/processor.h> 29 #include <asm/cpu_device_id.h> 30 31 #define DRVNAME "coretemp" 32 33 /* 34 * force_tjmax only matters when TjMax can't be read from the CPU itself. 35 * When set, it replaces the driver's suboptimal heuristic. 36 */ 37 static int force_tjmax; 38 module_param_named(tjmax, force_tjmax, int, 0444); 39 MODULE_PARM_DESC(tjmax, "TjMax value in degrees Celsius"); 40 41 #define PKG_SYSFS_ATTR_NO 1 /* Sysfs attribute for package temp */ 42 #define BASE_SYSFS_ATTR_NO 2 /* Sysfs Base attr no for coretemp */ 43 #define NUM_REAL_CORES 128 /* Number of Real cores per cpu */ 44 #define CORETEMP_NAME_LENGTH 19 /* String Length of attrs */ 45 #define MAX_CORE_ATTRS 4 /* Maximum no of basic attrs */ 46 #define TOTAL_ATTRS (MAX_CORE_ATTRS + 1) 47 #define MAX_CORE_DATA (NUM_REAL_CORES + BASE_SYSFS_ATTR_NO) 48 49 #ifdef CONFIG_SMP 50 #define for_each_sibling(i, cpu) \ 51 for_each_cpu(i, topology_sibling_cpumask(cpu)) 52 #else 53 #define for_each_sibling(i, cpu) for (i = 0; false; ) 54 #endif 55 56 /* 57 * Per-Core Temperature Data 58 * @last_updated: The time when the current temperature value was updated 59 * earlier (in jiffies). 60 * @cpu_core_id: The CPU Core from which temperature values should be read 61 * This value is passed as "id" field to rdmsr/wrmsr functions. 62 * @status_reg: One of IA32_THERM_STATUS or IA32_PACKAGE_THERM_STATUS, 63 * from where the temperature values should be read. 64 * @attr_size: Total number of pre-core attrs displayed in the sysfs. 65 * @is_pkg_data: If this is 1, the temp_data holds pkgtemp data. 66 * Otherwise, temp_data holds coretemp data. 67 * @valid: If this is 1, the current temperature is valid. 68 */ 69 struct temp_data { 70 int temp; 71 int ttarget; 72 int tjmax; 73 unsigned long last_updated; 74 unsigned int cpu; 75 u32 cpu_core_id; 76 u32 status_reg; 77 int attr_size; 78 bool is_pkg_data; 79 bool valid; 80 struct sensor_device_attribute sd_attrs[TOTAL_ATTRS]; 81 char attr_name[TOTAL_ATTRS][CORETEMP_NAME_LENGTH]; 82 struct attribute *attrs[TOTAL_ATTRS + 1]; 83 struct attribute_group attr_group; 84 struct mutex update_lock; 85 }; 86 87 /* Platform Data per Physical CPU */ 88 struct platform_data { 89 struct device *hwmon_dev; 90 u16 pkg_id; 91 u16 cpu_map[NUM_REAL_CORES]; 92 struct ida ida; 93 struct cpumask cpumask; 94 struct temp_data *core_data[MAX_CORE_DATA]; 95 struct device_attribute name_attr; 96 }; 97 98 /* Keep track of how many zone pointers we allocated in init() */ 99 static int max_zones __read_mostly; 100 /* Array of zone pointers. Serialized by cpu hotplug lock */ 101 static struct platform_device **zone_devices; 102 103 static ssize_t show_label(struct device *dev, 104 struct device_attribute *devattr, char *buf) 105 { 106 struct sensor_device_attribute *attr = to_sensor_dev_attr(devattr); 107 struct platform_data *pdata = dev_get_drvdata(dev); 108 struct temp_data *tdata = pdata->core_data[attr->index]; 109 110 if (tdata->is_pkg_data) 111 return sprintf(buf, "Package id %u\n", pdata->pkg_id); 112 113 return sprintf(buf, "Core %u\n", tdata->cpu_core_id); 114 } 115 116 static ssize_t show_crit_alarm(struct device *dev, 117 struct device_attribute *devattr, char *buf) 118 { 119 u32 eax, edx; 120 struct sensor_device_attribute *attr = to_sensor_dev_attr(devattr); 121 struct platform_data *pdata = dev_get_drvdata(dev); 122 struct temp_data *tdata = pdata->core_data[attr->index]; 123 124 mutex_lock(&tdata->update_lock); 125 rdmsr_on_cpu(tdata->cpu, tdata->status_reg, &eax, &edx); 126 mutex_unlock(&tdata->update_lock); 127 128 return sprintf(buf, "%d\n", (eax >> 5) & 1); 129 } 130 131 static ssize_t show_tjmax(struct device *dev, 132 struct device_attribute *devattr, char *buf) 133 { 134 struct sensor_device_attribute *attr = to_sensor_dev_attr(devattr); 135 struct platform_data *pdata = dev_get_drvdata(dev); 136 137 return sprintf(buf, "%d\n", pdata->core_data[attr->index]->tjmax); 138 } 139 140 static ssize_t show_ttarget(struct device *dev, 141 struct device_attribute *devattr, char *buf) 142 { 143 struct sensor_device_attribute *attr = to_sensor_dev_attr(devattr); 144 struct platform_data *pdata = dev_get_drvdata(dev); 145 146 return sprintf(buf, "%d\n", pdata->core_data[attr->index]->ttarget); 147 } 148 149 static ssize_t show_temp(struct device *dev, 150 struct device_attribute *devattr, char *buf) 151 { 152 u32 eax, edx; 153 struct sensor_device_attribute *attr = to_sensor_dev_attr(devattr); 154 struct platform_data *pdata = dev_get_drvdata(dev); 155 struct temp_data *tdata = pdata->core_data[attr->index]; 156 157 mutex_lock(&tdata->update_lock); 158 159 /* Check whether the time interval has elapsed */ 160 if (!tdata->valid || time_after(jiffies, tdata->last_updated + HZ)) { 161 rdmsr_on_cpu(tdata->cpu, tdata->status_reg, &eax, &edx); 162 /* 163 * Ignore the valid bit. In all observed cases the register 164 * value is either low or zero if the valid bit is 0. 165 * Return it instead of reporting an error which doesn't 166 * really help at all. 167 */ 168 tdata->temp = tdata->tjmax - ((eax >> 16) & 0x7f) * 1000; 169 tdata->valid = true; 170 tdata->last_updated = jiffies; 171 } 172 173 mutex_unlock(&tdata->update_lock); 174 return sprintf(buf, "%d\n", tdata->temp); 175 } 176 177 struct tjmax_pci { 178 unsigned int device; 179 int tjmax; 180 }; 181 182 static const struct tjmax_pci tjmax_pci_table[] = { 183 { 0x0708, 110000 }, /* CE41x0 (Sodaville ) */ 184 { 0x0c72, 102000 }, /* Atom S1240 (Centerton) */ 185 { 0x0c73, 95000 }, /* Atom S1220 (Centerton) */ 186 { 0x0c75, 95000 }, /* Atom S1260 (Centerton) */ 187 }; 188 189 struct tjmax { 190 char const *id; 191 int tjmax; 192 }; 193 194 static const struct tjmax tjmax_table[] = { 195 { "CPU 230", 100000 }, /* Model 0x1c, stepping 2 */ 196 { "CPU 330", 125000 }, /* Model 0x1c, stepping 2 */ 197 }; 198 199 struct tjmax_model { 200 u8 model; 201 u8 mask; 202 int tjmax; 203 }; 204 205 #define ANY 0xff 206 207 static const struct tjmax_model tjmax_model_table[] = { 208 { 0x1c, 10, 100000 }, /* D4xx, K4xx, N4xx, D5xx, K5xx, N5xx */ 209 { 0x1c, ANY, 90000 }, /* Z5xx, N2xx, possibly others 210 * Note: Also matches 230 and 330, 211 * which are covered by tjmax_table 212 */ 213 { 0x26, ANY, 90000 }, /* Atom Tunnel Creek (Exx), Lincroft (Z6xx) 214 * Note: TjMax for E6xxT is 110C, but CPU type 215 * is undetectable by software 216 */ 217 { 0x27, ANY, 90000 }, /* Atom Medfield (Z2460) */ 218 { 0x35, ANY, 90000 }, /* Atom Clover Trail/Cloverview (Z27x0) */ 219 { 0x36, ANY, 100000 }, /* Atom Cedar Trail/Cedarview (N2xxx, D2xxx) 220 * Also matches S12x0 (stepping 9), covered by 221 * PCI table 222 */ 223 }; 224 225 static int adjust_tjmax(struct cpuinfo_x86 *c, u32 id, struct device *dev) 226 { 227 /* The 100C is default for both mobile and non mobile CPUs */ 228 229 int tjmax = 100000; 230 int tjmax_ee = 85000; 231 int usemsr_ee = 1; 232 int err; 233 u32 eax, edx; 234 int i; 235 u16 devfn = PCI_DEVFN(0, 0); 236 struct pci_dev *host_bridge = pci_get_domain_bus_and_slot(0, 0, devfn); 237 238 /* 239 * Explicit tjmax table entries override heuristics. 240 * First try PCI host bridge IDs, followed by model ID strings 241 * and model/stepping information. 242 */ 243 if (host_bridge && host_bridge->vendor == PCI_VENDOR_ID_INTEL) { 244 for (i = 0; i < ARRAY_SIZE(tjmax_pci_table); i++) { 245 if (host_bridge->device == tjmax_pci_table[i].device) { 246 pci_dev_put(host_bridge); 247 return tjmax_pci_table[i].tjmax; 248 } 249 } 250 } 251 pci_dev_put(host_bridge); 252 253 for (i = 0; i < ARRAY_SIZE(tjmax_table); i++) { 254 if (strstr(c->x86_model_id, tjmax_table[i].id)) 255 return tjmax_table[i].tjmax; 256 } 257 258 for (i = 0; i < ARRAY_SIZE(tjmax_model_table); i++) { 259 const struct tjmax_model *tm = &tjmax_model_table[i]; 260 if (c->x86_model == tm->model && 261 (tm->mask == ANY || c->x86_stepping == tm->mask)) 262 return tm->tjmax; 263 } 264 265 /* Early chips have no MSR for TjMax */ 266 267 if (c->x86_model == 0xf && c->x86_stepping < 4) 268 usemsr_ee = 0; 269 270 if (c->x86_model > 0xe && usemsr_ee) { 271 u8 platform_id; 272 273 /* 274 * Now we can detect the mobile CPU using Intel provided table 275 * http://softwarecommunity.intel.com/Wiki/Mobility/720.htm 276 * For Core2 cores, check MSR 0x17, bit 28 1 = Mobile CPU 277 */ 278 err = rdmsr_safe_on_cpu(id, 0x17, &eax, &edx); 279 if (err) { 280 dev_warn(dev, 281 "Unable to access MSR 0x17, assuming desktop" 282 " CPU\n"); 283 usemsr_ee = 0; 284 } else if (c->x86_model < 0x17 && !(eax & 0x10000000)) { 285 /* 286 * Trust bit 28 up to Penryn, I could not find any 287 * documentation on that; if you happen to know 288 * someone at Intel please ask 289 */ 290 usemsr_ee = 0; 291 } else { 292 /* Platform ID bits 52:50 (EDX starts at bit 32) */ 293 platform_id = (edx >> 18) & 0x7; 294 295 /* 296 * Mobile Penryn CPU seems to be platform ID 7 or 5 297 * (guesswork) 298 */ 299 if (c->x86_model == 0x17 && 300 (platform_id == 5 || platform_id == 7)) { 301 /* 302 * If MSR EE bit is set, set it to 90 degrees C, 303 * otherwise 105 degrees C 304 */ 305 tjmax_ee = 90000; 306 tjmax = 105000; 307 } 308 } 309 } 310 311 if (usemsr_ee) { 312 err = rdmsr_safe_on_cpu(id, 0xee, &eax, &edx); 313 if (err) { 314 dev_warn(dev, 315 "Unable to access MSR 0xEE, for Tjmax, left" 316 " at default\n"); 317 } else if (eax & 0x40000000) { 318 tjmax = tjmax_ee; 319 } 320 } else if (tjmax == 100000) { 321 /* 322 * If we don't use msr EE it means we are desktop CPU 323 * (with exeception of Atom) 324 */ 325 dev_warn(dev, "Using relative temperature scale!\n"); 326 } 327 328 return tjmax; 329 } 330 331 static bool cpu_has_tjmax(struct cpuinfo_x86 *c) 332 { 333 u8 model = c->x86_model; 334 335 return model > 0xe && 336 model != 0x1c && 337 model != 0x26 && 338 model != 0x27 && 339 model != 0x35 && 340 model != 0x36; 341 } 342 343 static int get_tjmax(struct cpuinfo_x86 *c, u32 id, struct device *dev) 344 { 345 int err; 346 u32 eax, edx; 347 u32 val; 348 349 /* 350 * A new feature of current Intel(R) processors, the 351 * IA32_TEMPERATURE_TARGET contains the TjMax value 352 */ 353 err = rdmsr_safe_on_cpu(id, MSR_IA32_TEMPERATURE_TARGET, &eax, &edx); 354 if (err) { 355 if (cpu_has_tjmax(c)) 356 dev_warn(dev, "Unable to read TjMax from CPU %u\n", id); 357 } else { 358 val = (eax >> 16) & 0xff; 359 /* 360 * If the TjMax is not plausible, an assumption 361 * will be used 362 */ 363 if (val) { 364 dev_dbg(dev, "TjMax is %d degrees C\n", val); 365 return val * 1000; 366 } 367 } 368 369 if (force_tjmax) { 370 dev_notice(dev, "TjMax forced to %d degrees C by user\n", 371 force_tjmax); 372 return force_tjmax * 1000; 373 } 374 375 /* 376 * An assumption is made for early CPUs and unreadable MSR. 377 * NOTE: the calculated value may not be correct. 378 */ 379 return adjust_tjmax(c, id, dev); 380 } 381 382 static int create_core_attrs(struct temp_data *tdata, struct device *dev, 383 int attr_no) 384 { 385 int i; 386 static ssize_t (*const rd_ptr[TOTAL_ATTRS]) (struct device *dev, 387 struct device_attribute *devattr, char *buf) = { 388 show_label, show_crit_alarm, show_temp, show_tjmax, 389 show_ttarget }; 390 static const char *const suffixes[TOTAL_ATTRS] = { 391 "label", "crit_alarm", "input", "crit", "max" 392 }; 393 394 for (i = 0; i < tdata->attr_size; i++) { 395 snprintf(tdata->attr_name[i], CORETEMP_NAME_LENGTH, 396 "temp%d_%s", attr_no, suffixes[i]); 397 sysfs_attr_init(&tdata->sd_attrs[i].dev_attr.attr); 398 tdata->sd_attrs[i].dev_attr.attr.name = tdata->attr_name[i]; 399 tdata->sd_attrs[i].dev_attr.attr.mode = 0444; 400 tdata->sd_attrs[i].dev_attr.show = rd_ptr[i]; 401 tdata->sd_attrs[i].index = attr_no; 402 tdata->attrs[i] = &tdata->sd_attrs[i].dev_attr.attr; 403 } 404 tdata->attr_group.attrs = tdata->attrs; 405 return sysfs_create_group(&dev->kobj, &tdata->attr_group); 406 } 407 408 409 static int chk_ucode_version(unsigned int cpu) 410 { 411 struct cpuinfo_x86 *c = &cpu_data(cpu); 412 413 /* 414 * Check if we have problem with errata AE18 of Core processors: 415 * Readings might stop update when processor visited too deep sleep, 416 * fixed for stepping D0 (6EC). 417 */ 418 if (c->x86_model == 0xe && c->x86_stepping < 0xc && c->microcode < 0x39) { 419 pr_err("Errata AE18 not fixed, update BIOS or microcode of the CPU!\n"); 420 return -ENODEV; 421 } 422 return 0; 423 } 424 425 static struct platform_device *coretemp_get_pdev(unsigned int cpu) 426 { 427 int id = topology_logical_die_id(cpu); 428 429 if (id >= 0 && id < max_zones) 430 return zone_devices[id]; 431 return NULL; 432 } 433 434 static struct temp_data *init_temp_data(unsigned int cpu, int pkg_flag) 435 { 436 struct temp_data *tdata; 437 438 tdata = kzalloc(sizeof(struct temp_data), GFP_KERNEL); 439 if (!tdata) 440 return NULL; 441 442 tdata->status_reg = pkg_flag ? MSR_IA32_PACKAGE_THERM_STATUS : 443 MSR_IA32_THERM_STATUS; 444 tdata->is_pkg_data = pkg_flag; 445 tdata->cpu = cpu; 446 tdata->cpu_core_id = topology_core_id(cpu); 447 tdata->attr_size = MAX_CORE_ATTRS; 448 mutex_init(&tdata->update_lock); 449 return tdata; 450 } 451 452 static int create_core_data(struct platform_device *pdev, unsigned int cpu, 453 int pkg_flag) 454 { 455 struct temp_data *tdata; 456 struct platform_data *pdata = platform_get_drvdata(pdev); 457 struct cpuinfo_x86 *c = &cpu_data(cpu); 458 u32 eax, edx; 459 int err, index, attr_no; 460 461 /* 462 * Find attr number for sysfs: 463 * We map the attr number to core id of the CPU 464 * The attr number is always core id + 2 465 * The Pkgtemp will always show up as temp1_*, if available 466 */ 467 if (pkg_flag) { 468 attr_no = PKG_SYSFS_ATTR_NO; 469 } else { 470 index = ida_alloc(&pdata->ida, GFP_KERNEL); 471 if (index < 0) 472 return index; 473 pdata->cpu_map[index] = topology_core_id(cpu); 474 attr_no = index + BASE_SYSFS_ATTR_NO; 475 } 476 477 if (attr_no > MAX_CORE_DATA - 1) { 478 err = -ERANGE; 479 goto ida_free; 480 } 481 482 tdata = init_temp_data(cpu, pkg_flag); 483 if (!tdata) { 484 err = -ENOMEM; 485 goto ida_free; 486 } 487 488 /* Test if we can access the status register */ 489 err = rdmsr_safe_on_cpu(cpu, tdata->status_reg, &eax, &edx); 490 if (err) 491 goto exit_free; 492 493 /* We can access status register. Get Critical Temperature */ 494 tdata->tjmax = get_tjmax(c, cpu, &pdev->dev); 495 496 /* 497 * Read the still undocumented bits 8:15 of IA32_TEMPERATURE_TARGET. 498 * The target temperature is available on older CPUs but not in this 499 * register. Atoms don't have the register at all. 500 */ 501 if (c->x86_model > 0xe && c->x86_model != 0x1c) { 502 err = rdmsr_safe_on_cpu(cpu, MSR_IA32_TEMPERATURE_TARGET, 503 &eax, &edx); 504 if (!err) { 505 tdata->ttarget 506 = tdata->tjmax - ((eax >> 8) & 0xff) * 1000; 507 tdata->attr_size++; 508 } 509 } 510 511 pdata->core_data[attr_no] = tdata; 512 513 /* Create sysfs interfaces */ 514 err = create_core_attrs(tdata, pdata->hwmon_dev, attr_no); 515 if (err) 516 goto exit_free; 517 518 return 0; 519 exit_free: 520 pdata->core_data[attr_no] = NULL; 521 kfree(tdata); 522 ida_free: 523 if (!pkg_flag) 524 ida_free(&pdata->ida, index); 525 return err; 526 } 527 528 static void 529 coretemp_add_core(struct platform_device *pdev, unsigned int cpu, int pkg_flag) 530 { 531 if (create_core_data(pdev, cpu, pkg_flag)) 532 dev_err(&pdev->dev, "Adding Core %u failed\n", cpu); 533 } 534 535 static void coretemp_remove_core(struct platform_data *pdata, int indx) 536 { 537 struct temp_data *tdata = pdata->core_data[indx]; 538 539 /* if we errored on add then this is already gone */ 540 if (!tdata) 541 return; 542 543 /* Remove the sysfs attributes */ 544 sysfs_remove_group(&pdata->hwmon_dev->kobj, &tdata->attr_group); 545 546 kfree(pdata->core_data[indx]); 547 pdata->core_data[indx] = NULL; 548 549 if (indx >= BASE_SYSFS_ATTR_NO) 550 ida_free(&pdata->ida, indx - BASE_SYSFS_ATTR_NO); 551 } 552 553 static int coretemp_probe(struct platform_device *pdev) 554 { 555 struct device *dev = &pdev->dev; 556 struct platform_data *pdata; 557 558 /* Initialize the per-zone data structures */ 559 pdata = devm_kzalloc(dev, sizeof(struct platform_data), GFP_KERNEL); 560 if (!pdata) 561 return -ENOMEM; 562 563 pdata->pkg_id = pdev->id; 564 ida_init(&pdata->ida); 565 platform_set_drvdata(pdev, pdata); 566 567 pdata->hwmon_dev = devm_hwmon_device_register_with_groups(dev, DRVNAME, 568 pdata, NULL); 569 return PTR_ERR_OR_ZERO(pdata->hwmon_dev); 570 } 571 572 static int coretemp_remove(struct platform_device *pdev) 573 { 574 struct platform_data *pdata = platform_get_drvdata(pdev); 575 int i; 576 577 for (i = MAX_CORE_DATA - 1; i >= 0; --i) 578 if (pdata->core_data[i]) 579 coretemp_remove_core(pdata, i); 580 581 ida_destroy(&pdata->ida); 582 return 0; 583 } 584 585 static struct platform_driver coretemp_driver = { 586 .driver = { 587 .name = DRVNAME, 588 }, 589 .probe = coretemp_probe, 590 .remove = coretemp_remove, 591 }; 592 593 static struct platform_device *coretemp_device_add(unsigned int cpu) 594 { 595 int err, zoneid = topology_logical_die_id(cpu); 596 struct platform_device *pdev; 597 598 if (zoneid < 0) 599 return ERR_PTR(-ENOMEM); 600 601 pdev = platform_device_alloc(DRVNAME, zoneid); 602 if (!pdev) 603 return ERR_PTR(-ENOMEM); 604 605 err = platform_device_add(pdev); 606 if (err) { 607 platform_device_put(pdev); 608 return ERR_PTR(err); 609 } 610 611 zone_devices[zoneid] = pdev; 612 return pdev; 613 } 614 615 static int coretemp_cpu_online(unsigned int cpu) 616 { 617 struct platform_device *pdev = coretemp_get_pdev(cpu); 618 struct cpuinfo_x86 *c = &cpu_data(cpu); 619 struct platform_data *pdata; 620 621 /* 622 * Don't execute this on resume as the offline callback did 623 * not get executed on suspend. 624 */ 625 if (cpuhp_tasks_frozen) 626 return 0; 627 628 /* 629 * CPUID.06H.EAX[0] indicates whether the CPU has thermal 630 * sensors. We check this bit only, all the early CPUs 631 * without thermal sensors will be filtered out. 632 */ 633 if (!cpu_has(c, X86_FEATURE_DTHERM)) 634 return -ENODEV; 635 636 if (!pdev) { 637 /* Check the microcode version of the CPU */ 638 if (chk_ucode_version(cpu)) 639 return -EINVAL; 640 641 /* 642 * Alright, we have DTS support. 643 * We are bringing the _first_ core in this pkg 644 * online. So, initialize per-pkg data structures and 645 * then bring this core online. 646 */ 647 pdev = coretemp_device_add(cpu); 648 if (IS_ERR(pdev)) 649 return PTR_ERR(pdev); 650 651 /* 652 * Check whether pkgtemp support is available. 653 * If so, add interfaces for pkgtemp. 654 */ 655 if (cpu_has(c, X86_FEATURE_PTS)) 656 coretemp_add_core(pdev, cpu, 1); 657 } 658 659 pdata = platform_get_drvdata(pdev); 660 /* 661 * Check whether a thread sibling is already online. If not add the 662 * interface for this CPU core. 663 */ 664 if (!cpumask_intersects(&pdata->cpumask, topology_sibling_cpumask(cpu))) 665 coretemp_add_core(pdev, cpu, 0); 666 667 cpumask_set_cpu(cpu, &pdata->cpumask); 668 return 0; 669 } 670 671 static int coretemp_cpu_offline(unsigned int cpu) 672 { 673 struct platform_device *pdev = coretemp_get_pdev(cpu); 674 struct platform_data *pd; 675 struct temp_data *tdata; 676 int i, indx = -1, target; 677 678 /* 679 * Don't execute this on suspend as the device remove locks 680 * up the machine. 681 */ 682 if (cpuhp_tasks_frozen) 683 return 0; 684 685 /* If the physical CPU device does not exist, just return */ 686 if (!pdev) 687 return 0; 688 689 pd = platform_get_drvdata(pdev); 690 691 for (i = 0; i < NUM_REAL_CORES; i++) { 692 if (pd->cpu_map[i] == topology_core_id(cpu)) { 693 indx = i + BASE_SYSFS_ATTR_NO; 694 break; 695 } 696 } 697 698 /* Too many cores and this core is not populated, just return */ 699 if (indx < 0) 700 return 0; 701 702 tdata = pd->core_data[indx]; 703 704 cpumask_clear_cpu(cpu, &pd->cpumask); 705 706 /* 707 * If this is the last thread sibling, remove the CPU core 708 * interface, If there is still a sibling online, transfer the 709 * target cpu of that core interface to it. 710 */ 711 target = cpumask_any_and(&pd->cpumask, topology_sibling_cpumask(cpu)); 712 if (target >= nr_cpu_ids) { 713 coretemp_remove_core(pd, indx); 714 } else if (tdata && tdata->cpu == cpu) { 715 mutex_lock(&tdata->update_lock); 716 tdata->cpu = target; 717 mutex_unlock(&tdata->update_lock); 718 } 719 720 /* 721 * If all cores in this pkg are offline, remove the device. This 722 * will invoke the platform driver remove function, which cleans up 723 * the rest. 724 */ 725 if (cpumask_empty(&pd->cpumask)) { 726 zone_devices[topology_logical_die_id(cpu)] = NULL; 727 platform_device_unregister(pdev); 728 return 0; 729 } 730 731 /* 732 * Check whether this core is the target for the package 733 * interface. We need to assign it to some other cpu. 734 */ 735 tdata = pd->core_data[PKG_SYSFS_ATTR_NO]; 736 if (tdata && tdata->cpu == cpu) { 737 target = cpumask_first(&pd->cpumask); 738 mutex_lock(&tdata->update_lock); 739 tdata->cpu = target; 740 mutex_unlock(&tdata->update_lock); 741 } 742 return 0; 743 } 744 static const struct x86_cpu_id __initconst coretemp_ids[] = { 745 X86_MATCH_VENDOR_FEATURE(INTEL, X86_FEATURE_DTHERM, NULL), 746 {} 747 }; 748 MODULE_DEVICE_TABLE(x86cpu, coretemp_ids); 749 750 static enum cpuhp_state coretemp_hp_online; 751 752 static int __init coretemp_init(void) 753 { 754 int err; 755 756 /* 757 * CPUID.06H.EAX[0] indicates whether the CPU has thermal 758 * sensors. We check this bit only, all the early CPUs 759 * without thermal sensors will be filtered out. 760 */ 761 if (!x86_match_cpu(coretemp_ids)) 762 return -ENODEV; 763 764 max_zones = topology_max_packages() * topology_max_die_per_package(); 765 zone_devices = kcalloc(max_zones, sizeof(struct platform_device *), 766 GFP_KERNEL); 767 if (!zone_devices) 768 return -ENOMEM; 769 770 err = platform_driver_register(&coretemp_driver); 771 if (err) 772 goto outzone; 773 774 err = cpuhp_setup_state(CPUHP_AP_ONLINE_DYN, "hwmon/coretemp:online", 775 coretemp_cpu_online, coretemp_cpu_offline); 776 if (err < 0) 777 goto outdrv; 778 coretemp_hp_online = err; 779 return 0; 780 781 outdrv: 782 platform_driver_unregister(&coretemp_driver); 783 outzone: 784 kfree(zone_devices); 785 return err; 786 } 787 module_init(coretemp_init) 788 789 static void __exit coretemp_exit(void) 790 { 791 cpuhp_remove_state(coretemp_hp_online); 792 platform_driver_unregister(&coretemp_driver); 793 kfree(zone_devices); 794 } 795 module_exit(coretemp_exit) 796 797 MODULE_AUTHOR("Rudolf Marek <r.marek@assembler.cz>"); 798 MODULE_DESCRIPTION("Intel Core temperature monitor"); 799 MODULE_LICENSE("GPL"); 800