1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* 23 * Copyright (c) 2010, Oracle and/or its affiliates. All rights reserved. 24 */ 25 /* 26 * Copyright (c) 2010, Intel Corporation. 27 * All rights reserved. 28 */ 29 /* 30 * Copyright 2013 Nexenta Systems, Inc. All rights reserved. 31 */ 32 33 #include <sys/processor.h> 34 #include <sys/time.h> 35 #include <sys/psm.h> 36 #include <sys/smp_impldefs.h> 37 #include <sys/cram.h> 38 #include <sys/acpi/acpi.h> 39 #include <sys/acpica.h> 40 #include <sys/psm_common.h> 41 #include <sys/pit.h> 42 #include <sys/ddi.h> 43 #include <sys/sunddi.h> 44 #include <sys/ddi_impldefs.h> 45 #include <sys/pci.h> 46 #include <sys/promif.h> 47 #include <sys/x86_archext.h> 48 #include <sys/cpc_impl.h> 49 #include <sys/uadmin.h> 50 #include <sys/panic.h> 51 #include <sys/debug.h> 52 #include <sys/archsystm.h> 53 #include <sys/trap.h> 54 #include <sys/machsystm.h> 55 #include <sys/sysmacros.h> 56 #include <sys/cpuvar.h> 57 #include <sys/rm_platter.h> 58 #include <sys/privregs.h> 59 #include <sys/note.h> 60 #include <sys/pci_intr_lib.h> 61 #include <sys/spl.h> 62 #include <sys/clock.h> 63 #include <sys/dditypes.h> 64 #include <sys/sunddi.h> 65 #include <sys/x_call.h> 66 #include <sys/reboot.h> 67 #include <sys/apix.h> 68 69 static int apix_get_avail_vector_oncpu(uint32_t, int, int); 70 static apix_vector_t *apix_init_vector(processorid_t, uchar_t); 71 static void apix_cleanup_vector(apix_vector_t *); 72 static void apix_insert_av(apix_vector_t *, void *, avfunc, caddr_t, caddr_t, 73 uint64_t *, int, dev_info_t *); 74 static void apix_remove_av(apix_vector_t *, struct autovec *); 75 static void apix_clear_dev_map(dev_info_t *, int, int); 76 static boolean_t apix_is_cpu_enabled(processorid_t); 77 static void apix_wait_till_seen(processorid_t, int); 78 79 #define GET_INTR_INUM(ihdlp) \ 80 (((ihdlp) != NULL) ? ((ddi_intr_handle_impl_t *)(ihdlp))->ih_inum : 0) 81 82 apix_rebind_info_t apix_rebindinfo = {0, 0, 0, NULL, 0, NULL}; 83 84 /* 85 * Allocate IPI 86 * 87 * Return vector number or 0 on error 88 */ 89 uchar_t 90 apix_alloc_ipi(int ipl) 91 { 92 apix_vector_t *vecp; 93 uchar_t vector; 94 int cpun; 95 int nproc; 96 97 APIX_ENTER_CPU_LOCK(0); 98 99 vector = apix_get_avail_vector_oncpu(0, APIX_IPI_MIN, APIX_IPI_MAX); 100 if (vector == 0) { 101 APIX_LEAVE_CPU_LOCK(0); 102 cmn_err(CE_WARN, "apix: no available IPI\n"); 103 apic_error |= APIC_ERR_GET_IPIVECT_FAIL; 104 return (0); 105 } 106 107 nproc = max(apic_nproc, apic_max_nproc); 108 for (cpun = 0; cpun < nproc; cpun++) { 109 vecp = xv_vector(cpun, vector); 110 if (vecp == NULL) { 111 vecp = kmem_zalloc(sizeof (apix_vector_t), KM_NOSLEEP); 112 if (vecp == NULL) { 113 cmn_err(CE_WARN, "apix: No memory for ipi"); 114 goto fail; 115 } 116 xv_vector(cpun, vector) = vecp; 117 } 118 vecp->v_state = APIX_STATE_ALLOCED; 119 vecp->v_type = APIX_TYPE_IPI; 120 vecp->v_cpuid = vecp->v_bound_cpuid = cpun; 121 vecp->v_vector = vector; 122 vecp->v_pri = ipl; 123 } 124 APIX_LEAVE_CPU_LOCK(0); 125 return (vector); 126 127 fail: 128 while (--cpun >= 0) 129 apix_cleanup_vector(xv_vector(cpun, vector)); 130 APIX_LEAVE_CPU_LOCK(0); 131 return (0); 132 } 133 134 /* 135 * Add IPI service routine 136 */ 137 static int 138 apix_add_ipi(int ipl, avfunc xxintr, char *name, int vector, 139 caddr_t arg1, caddr_t arg2) 140 { 141 int cpun; 142 apix_vector_t *vecp; 143 int nproc; 144 145 ASSERT(vector >= APIX_IPI_MIN && vector <= APIX_IPI_MAX); 146 147 nproc = max(apic_nproc, apic_max_nproc); 148 for (cpun = 0; cpun < nproc; cpun++) { 149 APIX_ENTER_CPU_LOCK(cpun); 150 vecp = xv_vector(cpun, vector); 151 apix_insert_av(vecp, NULL, xxintr, arg1, arg2, NULL, ipl, NULL); 152 vecp->v_state = APIX_STATE_ENABLED; 153 APIX_LEAVE_CPU_LOCK(cpun); 154 } 155 156 APIC_VERBOSE(IPI, (CE_CONT, "apix: add ipi for %s, vector %x " 157 "ipl %x\n", name, vector, ipl)); 158 159 return (1); 160 } 161 162 /* 163 * Find and return first free vector in range (start, end) 164 */ 165 static int 166 apix_get_avail_vector_oncpu(uint32_t cpuid, int start, int end) 167 { 168 int i; 169 apix_impl_t *apixp = apixs[cpuid]; 170 171 for (i = start; i <= end; i++) { 172 if (APIC_CHECK_RESERVE_VECTORS(i)) 173 continue; 174 if (IS_VECT_FREE(apixp->x_vectbl[i])) 175 return (i); 176 } 177 178 return (0); 179 } 180 181 /* 182 * Allocate a vector on specified cpu 183 * 184 * Return NULL on error 185 */ 186 static apix_vector_t * 187 apix_alloc_vector_oncpu(uint32_t cpuid, dev_info_t *dip, int inum, int type) 188 { 189 processorid_t tocpu = cpuid & ~IRQ_USER_BOUND; 190 apix_vector_t *vecp; 191 int vector; 192 193 ASSERT(APIX_CPU_LOCK_HELD(tocpu)); 194 195 /* find free vector */ 196 vector = apix_get_avail_vector_oncpu(tocpu, APIX_AVINTR_MIN, 197 APIX_AVINTR_MAX); 198 if (vector == 0) 199 return (NULL); 200 201 vecp = apix_init_vector(tocpu, vector); 202 vecp->v_type = (ushort_t)type; 203 vecp->v_inum = inum; 204 vecp->v_flags = (cpuid & IRQ_USER_BOUND) ? APIX_VECT_USER_BOUND : 0; 205 206 if (dip != NULL) 207 apix_set_dev_map(vecp, dip, inum); 208 209 return (vecp); 210 } 211 212 /* 213 * Allocates "count" contiguous MSI vectors starting at the proper alignment. 214 * Caller needs to make sure that count has to be power of 2 and should not 215 * be < 1. 216 * 217 * Return first vector number 218 */ 219 apix_vector_t * 220 apix_alloc_nvectors_oncpu(uint32_t cpuid, dev_info_t *dip, int inum, 221 int count, int type) 222 { 223 int i, msibits, start = 0, navail = 0; 224 apix_vector_t *vecp, *startp = NULL; 225 processorid_t tocpu = cpuid & ~IRQ_USER_BOUND; 226 uint_t flags; 227 228 ASSERT(APIX_CPU_LOCK_HELD(tocpu)); 229 230 /* 231 * msibits is the no. of lower order message data bits for the 232 * allocated MSI vectors and is used to calculate the aligned 233 * starting vector 234 */ 235 msibits = count - 1; 236 237 /* It has to be contiguous */ 238 for (i = APIX_AVINTR_MIN; i <= APIX_AVINTR_MAX; i++) { 239 if (!IS_VECT_FREE(xv_vector(tocpu, i))) 240 continue; 241 242 /* 243 * starting vector has to be aligned accordingly for 244 * multiple MSIs 245 */ 246 if (msibits) 247 i = (i + msibits) & ~msibits; 248 249 for (navail = 0, start = i; i <= APIX_AVINTR_MAX; i++) { 250 if (!IS_VECT_FREE(xv_vector(tocpu, i))) 251 break; 252 if (APIC_CHECK_RESERVE_VECTORS(i)) 253 break; 254 if (++navail == count) 255 goto done; 256 } 257 } 258 259 return (NULL); 260 261 done: 262 flags = (cpuid & IRQ_USER_BOUND) ? APIX_VECT_USER_BOUND : 0; 263 264 for (i = 0; i < count; i++) { 265 if ((vecp = apix_init_vector(tocpu, start + i)) == NULL) 266 goto fail; 267 268 vecp->v_type = (ushort_t)type; 269 vecp->v_inum = inum + i; 270 vecp->v_flags = flags; 271 272 if (dip != NULL) 273 apix_set_dev_map(vecp, dip, inum + i); 274 275 if (i == 0) 276 startp = vecp; 277 } 278 279 return (startp); 280 281 fail: 282 while (i-- > 0) { /* Free allocated vectors */ 283 vecp = xv_vector(tocpu, start + i); 284 apix_clear_dev_map(dip, inum + i, type); 285 apix_cleanup_vector(vecp); 286 } 287 return (NULL); 288 } 289 290 #define APIX_WRITE_MSI_DATA(_hdl, _cap, _ctrl, _v)\ 291 do {\ 292 if ((_ctrl) & PCI_MSI_64BIT_MASK)\ 293 pci_config_put16((_hdl), (_cap) + PCI_MSI_64BIT_DATA, (_v));\ 294 else\ 295 pci_config_put16((_hdl), (_cap) + PCI_MSI_32BIT_DATA, (_v));\ 296 _NOTE(CONSTCOND)} while (0) 297 298 static void 299 apix_pci_msi_enable_vector(apix_vector_t *vecp, dev_info_t *dip, int type, 300 int inum, int count, uchar_t vector, int target_apic_id) 301 { 302 uint64_t msi_addr, msi_data; 303 ushort_t msi_ctrl; 304 int i, cap_ptr = i_ddi_get_msi_msix_cap_ptr(dip); 305 ddi_acc_handle_t handle = i_ddi_get_pci_config_handle(dip); 306 msi_regs_t msi_regs; 307 void *intrmap_tbl[PCI_MSI_MAX_INTRS]; 308 309 DDI_INTR_IMPLDBG((CE_CONT, "apix_pci_msi_enable_vector: dip=0x%p\n" 310 "\tdriver = %s, inum=0x%x vector=0x%x apicid=0x%x\n", (void *)dip, 311 ddi_driver_name(dip), inum, vector, target_apic_id)); 312 313 ASSERT((handle != NULL) && (cap_ptr != 0)); 314 315 msi_regs.mr_data = vector; 316 msi_regs.mr_addr = target_apic_id; 317 318 for (i = 0; i < count; i++) 319 intrmap_tbl[i] = xv_intrmap_private(vecp->v_cpuid, vector + i); 320 apic_vt_ops->apic_intrmap_alloc_entry(intrmap_tbl, dip, type, 321 count, 0xff); 322 for (i = 0; i < count; i++) 323 xv_intrmap_private(vecp->v_cpuid, vector + i) = intrmap_tbl[i]; 324 325 apic_vt_ops->apic_intrmap_map_entry(vecp->v_intrmap_private, 326 (void *)&msi_regs, type, count); 327 apic_vt_ops->apic_intrmap_record_msi(vecp->v_intrmap_private, 328 &msi_regs); 329 330 /* MSI Address */ 331 msi_addr = msi_regs.mr_addr; 332 333 /* MSI Data: MSI is edge triggered according to spec */ 334 msi_data = msi_regs.mr_data; 335 336 DDI_INTR_IMPLDBG((CE_CONT, "apix_pci_msi_enable_vector: addr=0x%lx " 337 "data=0x%lx\n", (long)msi_addr, (long)msi_data)); 338 339 if (type == APIX_TYPE_MSI) { 340 msi_ctrl = pci_config_get16(handle, cap_ptr + PCI_MSI_CTRL); 341 342 /* Set the bits to inform how many MSIs are enabled */ 343 msi_ctrl |= ((highbit(count) - 1) << PCI_MSI_MME_SHIFT); 344 pci_config_put16(handle, cap_ptr + PCI_MSI_CTRL, msi_ctrl); 345 346 if ((vecp->v_flags & APIX_VECT_MASKABLE) == 0) 347 APIX_WRITE_MSI_DATA(handle, cap_ptr, msi_ctrl, 348 APIX_RESV_VECTOR); 349 350 pci_config_put32(handle, 351 cap_ptr + PCI_MSI_ADDR_OFFSET, msi_addr); 352 if (msi_ctrl & PCI_MSI_64BIT_MASK) 353 pci_config_put32(handle, 354 cap_ptr + PCI_MSI_ADDR_OFFSET + 4, msi_addr >> 32); 355 356 APIX_WRITE_MSI_DATA(handle, cap_ptr, msi_ctrl, msi_data); 357 } else if (type == APIX_TYPE_MSIX) { 358 uintptr_t off; 359 ddi_intr_msix_t *msix_p = i_ddi_get_msix(dip); 360 361 /* Offset into the "inum"th entry in the MSI-X table */ 362 off = (uintptr_t)msix_p->msix_tbl_addr + 363 (inum * PCI_MSIX_VECTOR_SIZE); 364 365 ddi_put32(msix_p->msix_tbl_hdl, 366 (uint32_t *)(off + PCI_MSIX_DATA_OFFSET), msi_data); 367 ddi_put64(msix_p->msix_tbl_hdl, 368 (uint64_t *)(off + PCI_MSIX_LOWER_ADDR_OFFSET), msi_addr); 369 } 370 } 371 372 static void 373 apix_pci_msi_enable_mode(dev_info_t *dip, int type, int inum) 374 { 375 ushort_t msi_ctrl; 376 int cap_ptr = i_ddi_get_msi_msix_cap_ptr(dip); 377 ddi_acc_handle_t handle = i_ddi_get_pci_config_handle(dip); 378 379 ASSERT((handle != NULL) && (cap_ptr != 0)); 380 381 if (type == APIX_TYPE_MSI) { 382 msi_ctrl = pci_config_get16(handle, cap_ptr + PCI_MSI_CTRL); 383 if ((msi_ctrl & PCI_MSI_ENABLE_BIT)) 384 return; 385 386 msi_ctrl |= PCI_MSI_ENABLE_BIT; 387 pci_config_put16(handle, cap_ptr + PCI_MSI_CTRL, msi_ctrl); 388 389 } else if (type == DDI_INTR_TYPE_MSIX) { 390 uintptr_t off; 391 uint32_t mask; 392 ddi_intr_msix_t *msix_p; 393 394 msix_p = i_ddi_get_msix(dip); 395 396 /* Offset into "inum"th entry in the MSI-X table & clear mask */ 397 off = (uintptr_t)msix_p->msix_tbl_addr + (inum * 398 PCI_MSIX_VECTOR_SIZE) + PCI_MSIX_VECTOR_CTRL_OFFSET; 399 400 mask = ddi_get32(msix_p->msix_tbl_hdl, (uint32_t *)off); 401 402 ddi_put32(msix_p->msix_tbl_hdl, (uint32_t *)off, (mask & ~1)); 403 404 msi_ctrl = pci_config_get16(handle, cap_ptr + PCI_MSIX_CTRL); 405 406 if (!(msi_ctrl & PCI_MSIX_ENABLE_BIT)) { 407 msi_ctrl |= PCI_MSIX_ENABLE_BIT; 408 pci_config_put16(handle, cap_ptr + PCI_MSIX_CTRL, 409 msi_ctrl); 410 } 411 } 412 } 413 414 /* 415 * Setup interrupt, pogramming IO-APIC or MSI/X address/data. 416 */ 417 void 418 apix_enable_vector(apix_vector_t *vecp) 419 { 420 int tocpu = vecp->v_cpuid, type = vecp->v_type; 421 apic_cpus_info_t *cpu_infop; 422 ulong_t iflag; 423 424 ASSERT(tocpu < apic_nproc); 425 426 cpu_infop = &apic_cpus[tocpu]; 427 if (vecp->v_flags & APIX_VECT_USER_BOUND) 428 cpu_infop->aci_bound++; 429 else 430 cpu_infop->aci_temp_bound++; 431 432 iflag = intr_clear(); 433 lock_set(&apic_ioapic_lock); 434 435 if (!DDI_INTR_IS_MSI_OR_MSIX(type)) { /* fixed */ 436 apix_intx_enable(vecp->v_inum); 437 } else { 438 int inum = vecp->v_inum; 439 dev_info_t *dip = APIX_GET_DIP(vecp); 440 int count = i_ddi_intr_get_current_nintrs(dip); 441 442 if (type == APIX_TYPE_MSI) { /* MSI */ 443 if (inum == apix_get_max_dev_inum(dip, type)) { 444 /* last one */ 445 uchar_t start_inum = inum + 1 - count; 446 uchar_t start_vect = vecp->v_vector + 1 - count; 447 apix_vector_t *start_vecp = 448 xv_vector(vecp->v_cpuid, start_vect); 449 450 APIC_VERBOSE(INTR, (CE_CONT, "apix: call " 451 "apix_pci_msi_enable_vector\n")); 452 apix_pci_msi_enable_vector(start_vecp, dip, 453 type, start_inum, count, start_vect, 454 cpu_infop->aci_local_id); 455 456 APIC_VERBOSE(INTR, (CE_CONT, "apix: call " 457 "apix_pci_msi_enable_mode\n")); 458 apix_pci_msi_enable_mode(dip, type, inum); 459 } 460 } else { /* MSI-X */ 461 apix_pci_msi_enable_vector(vecp, dip, 462 type, inum, 1, vecp->v_vector, 463 cpu_infop->aci_local_id); 464 apix_pci_msi_enable_mode(dip, type, inum); 465 } 466 } 467 vecp->v_state = APIX_STATE_ENABLED; 468 apic_redist_cpu_skip &= ~(1 << tocpu); 469 470 lock_clear(&apic_ioapic_lock); 471 intr_restore(iflag); 472 } 473 474 /* 475 * Disable the interrupt 476 */ 477 void 478 apix_disable_vector(apix_vector_t *vecp) 479 { 480 struct autovec *avp = vecp->v_autovect; 481 ulong_t iflag; 482 483 ASSERT(avp != NULL); 484 485 iflag = intr_clear(); 486 lock_set(&apic_ioapic_lock); 487 488 switch (vecp->v_type) { 489 case APIX_TYPE_MSI: 490 ASSERT(avp->av_vector != NULL && avp->av_dip != NULL); 491 /* 492 * Disable the MSI vector 493 * Make sure we only disable on the last 494 * of the multi-MSI support 495 */ 496 if (i_ddi_intr_get_current_nenables(avp->av_dip) == 1) { 497 apic_pci_msi_disable_mode(avp->av_dip, 498 DDI_INTR_TYPE_MSI); 499 } 500 break; 501 case APIX_TYPE_MSIX: 502 ASSERT(avp->av_vector != NULL && avp->av_dip != NULL); 503 /* 504 * Disable the MSI-X vector 505 * needs to clear its mask and addr/data for each MSI-X 506 */ 507 apic_pci_msi_unconfigure(avp->av_dip, DDI_INTR_TYPE_MSIX, 508 vecp->v_inum); 509 /* 510 * Make sure we only disable on the last MSI-X 511 */ 512 if (i_ddi_intr_get_current_nenables(avp->av_dip) == 1) { 513 apic_pci_msi_disable_mode(avp->av_dip, 514 DDI_INTR_TYPE_MSIX); 515 } 516 break; 517 default: 518 apix_intx_disable(vecp->v_inum); 519 break; 520 } 521 522 if (!(apic_cpus[vecp->v_cpuid].aci_status & APIC_CPU_SUSPEND)) 523 vecp->v_state = APIX_STATE_DISABLED; 524 apic_vt_ops->apic_intrmap_free_entry(&vecp->v_intrmap_private); 525 vecp->v_intrmap_private = NULL; 526 527 lock_clear(&apic_ioapic_lock); 528 intr_restore(iflag); 529 } 530 531 /* 532 * Mark vector as obsoleted or freed. The vector is marked 533 * obsoleted if there are pending requests on it. Otherwise, 534 * free the vector. The obsoleted vectors get freed after 535 * being serviced. 536 * 537 * Return 1 on being obosoleted and 0 on being freed. 538 */ 539 #define INTR_BUSY(_avp)\ 540 ((((volatile ushort_t)(_avp)->av_flags) &\ 541 (AV_PENTRY_PEND | AV_PENTRY_ONPROC)) != 0) 542 #define LOCAL_WITH_INTR_DISABLED(_cpuid)\ 543 ((_cpuid) == psm_get_cpu_id() && !interrupts_enabled()) 544 static uint64_t dummy_tick; 545 546 int 547 apix_obsolete_vector(apix_vector_t *vecp) 548 { 549 struct autovec *avp = vecp->v_autovect; 550 int repeats, tries, ipl, busy = 0, cpuid = vecp->v_cpuid; 551 apix_impl_t *apixp = apixs[cpuid]; 552 553 ASSERT(APIX_CPU_LOCK_HELD(cpuid)); 554 555 for (avp = vecp->v_autovect; avp != NULL; avp = avp->av_link) { 556 if (avp->av_vector == NULL) 557 continue; 558 559 if (LOCAL_WITH_INTR_DISABLED(cpuid)) { 560 int bit, index, irr; 561 562 if (INTR_BUSY(avp)) { 563 busy++; 564 continue; 565 } 566 567 /* check IRR for pending interrupts */ 568 index = vecp->v_vector / 32; 569 bit = vecp->v_vector % 32; 570 irr = apic_reg_ops->apic_read(APIC_IRR_REG + index); 571 if ((irr & (1 << bit)) != 0) 572 busy++; 573 574 if (!busy) 575 apix_remove_av(vecp, avp); 576 577 continue; 578 } 579 580 repeats = 0; 581 do { 582 repeats++; 583 for (tries = 0; tries < apic_max_reps_clear_pending; 584 tries++) 585 if (!INTR_BUSY(avp)) 586 break; 587 } while (INTR_BUSY(avp) && 588 (repeats < apic_max_reps_clear_pending)); 589 590 if (INTR_BUSY(avp)) 591 busy++; 592 else { 593 /* 594 * Interrupt is not in pending list or being serviced. 595 * However it might be cached in Local APIC's IRR 596 * register. It's impossible to check another CPU's 597 * IRR register. Then wait till lower levels finish 598 * running. 599 */ 600 for (ipl = 1; ipl < MIN(LOCK_LEVEL, vecp->v_pri); ipl++) 601 apix_wait_till_seen(cpuid, ipl); 602 if (INTR_BUSY(avp)) 603 busy++; 604 } 605 606 if (!busy) 607 apix_remove_av(vecp, avp); 608 } 609 610 if (busy) { 611 apix_vector_t *tp = apixp->x_obsoletes; 612 613 if (vecp->v_state == APIX_STATE_OBSOLETED) 614 return (1); 615 616 vecp->v_state = APIX_STATE_OBSOLETED; 617 vecp->v_next = NULL; 618 if (tp == NULL) 619 apixp->x_obsoletes = vecp; 620 else { 621 while (tp->v_next != NULL) 622 tp = tp->v_next; 623 tp->v_next = vecp; 624 } 625 return (1); 626 } 627 628 /* interrupt is not busy */ 629 if (vecp->v_state == APIX_STATE_OBSOLETED) { 630 /* remove from obsoleted list */ 631 apixp->x_obsoletes = vecp->v_next; 632 vecp->v_next = NULL; 633 } 634 apix_cleanup_vector(vecp); 635 return (0); 636 } 637 638 /* 639 * Duplicate number of continuous vectors to specified target vectors. 640 */ 641 static void 642 apix_dup_vectors(apix_vector_t *oldp, apix_vector_t *newp, int count) 643 { 644 struct autovec *avp; 645 apix_vector_t *fromp, *top; 646 processorid_t oldcpu = oldp->v_cpuid, newcpu = newp->v_cpuid; 647 uchar_t oldvec = oldp->v_vector, newvec = newp->v_vector; 648 int i, inum; 649 650 ASSERT(oldp->v_type != APIX_TYPE_IPI); 651 652 for (i = 0; i < count; i++) { 653 fromp = xv_vector(oldcpu, oldvec + i); 654 top = xv_vector(newcpu, newvec + i); 655 ASSERT(fromp != NULL && top != NULL); 656 657 /* copy over original one */ 658 top->v_state = fromp->v_state; 659 top->v_type = fromp->v_type; 660 top->v_bound_cpuid = fromp->v_bound_cpuid; 661 top->v_inum = fromp->v_inum; 662 top->v_flags = fromp->v_flags; 663 top->v_intrmap_private = fromp->v_intrmap_private; 664 665 for (avp = fromp->v_autovect; avp != NULL; avp = avp->av_link) { 666 if (avp->av_vector == NULL) 667 continue; 668 669 apix_insert_av(top, avp->av_intr_id, avp->av_vector, 670 avp->av_intarg1, avp->av_intarg2, avp->av_ticksp, 671 avp->av_prilevel, avp->av_dip); 672 673 if (fromp->v_type == APIX_TYPE_FIXED && 674 avp->av_dip != NULL) { 675 inum = GET_INTR_INUM(avp->av_intr_id); 676 apix_set_dev_map(top, avp->av_dip, inum); 677 } 678 } 679 680 if (DDI_INTR_IS_MSI_OR_MSIX(fromp->v_type) && 681 fromp->v_devp != NULL) 682 apix_set_dev_map(top, fromp->v_devp->dv_dip, 683 fromp->v_devp->dv_inum); 684 } 685 } 686 687 static apix_vector_t * 688 apix_init_vector(processorid_t cpuid, uchar_t vector) 689 { 690 apix_impl_t *apixp = apixs[cpuid]; 691 apix_vector_t *vecp = apixp->x_vectbl[vector]; 692 693 ASSERT(IS_VECT_FREE(vecp)); 694 695 if (vecp == NULL) { 696 vecp = kmem_zalloc(sizeof (apix_vector_t), KM_NOSLEEP); 697 if (vecp == NULL) { 698 cmn_err(CE_WARN, "apix: no memory to allocate vector"); 699 return (NULL); 700 } 701 apixp->x_vectbl[vector] = vecp; 702 } 703 vecp->v_state = APIX_STATE_ALLOCED; 704 vecp->v_cpuid = vecp->v_bound_cpuid = cpuid; 705 vecp->v_vector = vector; 706 707 return (vecp); 708 } 709 710 static void 711 apix_cleanup_vector(apix_vector_t *vecp) 712 { 713 ASSERT(vecp->v_share == 0); 714 vecp->v_bound_cpuid = IRQ_UNINIT; 715 vecp->v_state = APIX_STATE_FREED; 716 vecp->v_type = 0; 717 vecp->v_flags = 0; 718 vecp->v_busy = 0; 719 vecp->v_intrmap_private = NULL; 720 } 721 722 static void 723 apix_dprint_vector(apix_vector_t *vecp, dev_info_t *dip, int count) 724 { 725 #ifdef DEBUG 726 major_t major; 727 char *name, *drv_name; 728 int instance, len, t_len; 729 char mesg[1024] = "apix: "; 730 731 t_len = sizeof (mesg); 732 len = strlen(mesg); 733 if (dip != NULL) { 734 name = ddi_get_name(dip); 735 major = ddi_name_to_major(name); 736 drv_name = ddi_major_to_name(major); 737 instance = ddi_get_instance(dip); 738 (void) snprintf(mesg + len, t_len - len, "%s (%s) instance %d ", 739 name, drv_name, instance); 740 } 741 len = strlen(mesg); 742 743 switch (vecp->v_type) { 744 case APIX_TYPE_FIXED: 745 (void) snprintf(mesg + len, t_len - len, "irqno %d", 746 vecp->v_inum); 747 break; 748 case APIX_TYPE_MSI: 749 (void) snprintf(mesg + len, t_len - len, 750 "msi inum %d (count %d)", vecp->v_inum, count); 751 break; 752 case APIX_TYPE_MSIX: 753 (void) snprintf(mesg + len, t_len - len, "msi-x inum %d", 754 vecp->v_inum); 755 break; 756 default: 757 break; 758 759 } 760 761 APIC_VERBOSE(ALLOC, (CE_CONT, "%s allocated with vector 0x%x on " 762 "cpu %d\n", mesg, vecp->v_vector, vecp->v_cpuid)); 763 #endif /* DEBUG */ 764 } 765 766 /* 767 * Operations on avintr 768 */ 769 770 #define INIT_AUTOVEC(p, intr_id, f, arg1, arg2, ticksp, ipl, dip) \ 771 do { \ 772 (p)->av_intr_id = intr_id; \ 773 (p)->av_vector = f; \ 774 (p)->av_intarg1 = arg1; \ 775 (p)->av_intarg2 = arg2; \ 776 (p)->av_ticksp = ticksp; \ 777 (p)->av_prilevel = ipl; \ 778 (p)->av_dip = dip; \ 779 (p)->av_flags = 0; \ 780 _NOTE(CONSTCOND)} while (0) 781 782 /* 783 * Insert an interrupt service routine into chain by its priority from 784 * high to low 785 */ 786 static void 787 apix_insert_av(apix_vector_t *vecp, void *intr_id, avfunc f, caddr_t arg1, 788 caddr_t arg2, uint64_t *ticksp, int ipl, dev_info_t *dip) 789 { 790 struct autovec *p, *prep, *mem; 791 792 APIC_VERBOSE(INTR, (CE_CONT, "apix_insert_av: dip %p, vector 0x%x, " 793 "cpu %d\n", (void *)dip, vecp->v_vector, vecp->v_cpuid)); 794 795 mem = kmem_zalloc(sizeof (struct autovec), KM_SLEEP); 796 INIT_AUTOVEC(mem, intr_id, f, arg1, arg2, ticksp, ipl, dip); 797 if (vecp->v_type == APIX_TYPE_FIXED && apic_level_intr[vecp->v_inum]) 798 mem->av_flags |= AV_PENTRY_LEVEL; 799 800 vecp->v_share++; 801 vecp->v_pri = (ipl > vecp->v_pri) ? ipl : vecp->v_pri; 802 if (vecp->v_autovect == NULL) { /* Nothing on list - put it at head */ 803 vecp->v_autovect = mem; 804 return; 805 } 806 807 if (DDI_INTR_IS_MSI_OR_MSIX(vecp->v_type)) { /* MSI/X */ 808 ASSERT(vecp->v_share == 1); /* No sharing for MSI/X */ 809 810 INIT_AUTOVEC(vecp->v_autovect, intr_id, f, arg1, arg2, ticksp, 811 ipl, dip); 812 prep = vecp->v_autovect->av_link; 813 vecp->v_autovect->av_link = NULL; 814 815 /* Free the following autovect chain */ 816 while (prep != NULL) { 817 ASSERT(prep->av_vector == NULL); 818 819 p = prep; 820 prep = prep->av_link; 821 kmem_free(p, sizeof (struct autovec)); 822 } 823 824 kmem_free(mem, sizeof (struct autovec)); 825 return; 826 } 827 828 /* find where it goes in list */ 829 prep = NULL; 830 for (p = vecp->v_autovect; p != NULL; p = p->av_link) { 831 if (p->av_vector && p->av_prilevel <= ipl) 832 break; 833 prep = p; 834 } 835 if (prep != NULL) { 836 if (prep->av_vector == NULL) { /* freed struct available */ 837 INIT_AUTOVEC(prep, intr_id, f, arg1, arg2, 838 ticksp, ipl, dip); 839 prep->av_flags = mem->av_flags; 840 kmem_free(mem, sizeof (struct autovec)); 841 return; 842 } 843 844 mem->av_link = prep->av_link; 845 prep->av_link = mem; 846 } else { 847 /* insert new intpt at beginning of chain */ 848 mem->av_link = vecp->v_autovect; 849 vecp->v_autovect = mem; 850 } 851 } 852 853 /* 854 * After having made a change to an autovector list, wait until we have 855 * seen specified cpu not executing an interrupt at that level--so we 856 * know our change has taken effect completely (no old state in registers, 857 * etc). 858 */ 859 #define APIX_CPU_ENABLED(_cp) \ 860 (quiesce_active == 0 && \ 861 (((_cp)->cpu_flags & (CPU_QUIESCED|CPU_OFFLINE)) == 0)) 862 863 static void 864 apix_wait_till_seen(processorid_t cpuid, int ipl) 865 { 866 struct cpu *cp = cpu[cpuid]; 867 868 if (cp == NULL || LOCAL_WITH_INTR_DISABLED(cpuid)) 869 return; 870 871 /* 872 * Don't wait if the CPU is quiesced or offlined. This can happen 873 * when a CPU is running pause thread but hardware triggered an 874 * interrupt and the interrupt gets queued. 875 */ 876 for (;;) { 877 if (!INTR_ACTIVE((volatile struct cpu *)cpu[cpuid], ipl) && 878 (!APIX_CPU_ENABLED(cp) || 879 !INTR_PENDING((volatile apix_impl_t *)apixs[cpuid], ipl))) 880 return; 881 } 882 } 883 884 static void 885 apix_remove_av(apix_vector_t *vecp, struct autovec *target) 886 { 887 int hi_pri = 0; 888 struct autovec *p; 889 890 if (target == NULL) 891 return; 892 893 APIC_VERBOSE(INTR, (CE_CONT, "apix_remove_av: dip %p, vector 0x%x, " 894 "cpu %d\n", (void *)target->av_dip, vecp->v_vector, vecp->v_cpuid)); 895 896 for (p = vecp->v_autovect; p; p = p->av_link) { 897 if (p == target || p->av_vector == NULL) 898 continue; 899 hi_pri = (p->av_prilevel > hi_pri) ? p->av_prilevel : hi_pri; 900 } 901 902 vecp->v_share--; 903 vecp->v_pri = hi_pri; 904 905 /* 906 * This drops the handler from the chain, it can no longer be called. 907 * However, there is no guarantee that the handler is not currently 908 * still executing. 909 */ 910 target->av_vector = NULL; 911 /* 912 * There is a race where we could be just about to pick up the ticksp 913 * pointer to increment it after returning from the service routine 914 * in av_dispatch_autovect. Rather than NULL it out let's just point 915 * it off to something safe so that any final tick update attempt 916 * won't fault. 917 */ 918 target->av_ticksp = &dummy_tick; 919 apix_wait_till_seen(vecp->v_cpuid, target->av_prilevel); 920 } 921 922 static struct autovec * 923 apix_find_av(apix_vector_t *vecp, void *intr_id, avfunc f) 924 { 925 struct autovec *p; 926 927 for (p = vecp->v_autovect; p; p = p->av_link) { 928 if ((p->av_vector == f) && (p->av_intr_id == intr_id)) { 929 /* found the handler */ 930 return (p); 931 } 932 } 933 934 return (NULL); 935 } 936 937 static apix_vector_t * 938 apix_find_vector_by_avintr(void *intr_id, avfunc f) 939 { 940 apix_vector_t *vecp; 941 processorid_t n; 942 uchar_t v; 943 944 for (n = 0; n < apic_nproc; n++) { 945 if (!apix_is_cpu_enabled(n)) 946 continue; 947 948 for (v = APIX_AVINTR_MIN; v <= APIX_AVINTR_MIN; v++) { 949 vecp = xv_vector(n, v); 950 if (vecp == NULL || 951 vecp->v_state <= APIX_STATE_OBSOLETED) 952 continue; 953 954 if (apix_find_av(vecp, intr_id, f) != NULL) 955 return (vecp); 956 } 957 } 958 959 return (NULL); 960 } 961 962 /* 963 * Add interrupt service routine. 964 * 965 * For legacy interrupts (HPET timer, ACPI SCI), the vector is actually 966 * IRQ no. A vector is then allocated. Otherwise, the vector is already 967 * allocated. The input argument virt_vect is virtual vector of format 968 * APIX_VIRTVEC_VECTOR(cpuid, vector). 969 * 970 * Return 1 on success, 0 on failure. 971 */ 972 int 973 apix_add_avintr(void *intr_id, int ipl, avfunc xxintr, char *name, 974 int virt_vect, caddr_t arg1, caddr_t arg2, uint64_t *ticksp, 975 dev_info_t *dip) 976 { 977 int cpuid; 978 uchar_t v = (uchar_t)APIX_VIRTVEC_VECTOR(virt_vect); 979 apix_vector_t *vecp; 980 981 if (xxintr == NULL) { 982 cmn_err(CE_WARN, "Attempt to add null for %s " 983 "on vector 0x%x,0x%x", name, 984 APIX_VIRTVEC_CPU(virt_vect), 985 APIX_VIRTVEC_VECTOR(virt_vect)); 986 return (0); 987 } 988 989 if (v >= APIX_IPI_MIN) /* IPIs */ 990 return (apix_add_ipi(ipl, xxintr, name, v, arg1, arg2)); 991 992 if (!APIX_IS_VIRTVEC(virt_vect)) { /* got irq */ 993 int irqno = virt_vect; 994 int inum = GET_INTR_INUM(intr_id); 995 996 /* 997 * Senarios include: 998 * a. add_avintr() is called before irqp initialized (legacy) 999 * b. irqp is initialized, vector is not allocated (fixed) 1000 * c. irqp is initialized, vector is allocated (fixed & shared) 1001 */ 1002 if ((vecp = apix_alloc_intx(dip, inum, irqno)) == NULL) 1003 return (0); 1004 1005 cpuid = vecp->v_cpuid; 1006 v = vecp->v_vector; 1007 virt_vect = APIX_VIRTVECTOR(cpuid, v); 1008 } else { /* got virtual vector */ 1009 cpuid = APIX_VIRTVEC_CPU(virt_vect); 1010 vecp = xv_vector(cpuid, v); 1011 ASSERT(vecp != NULL); 1012 } 1013 1014 lock_set(&apix_lock); 1015 if (vecp->v_state <= APIX_STATE_OBSOLETED) { 1016 vecp = NULL; 1017 1018 /* 1019 * Basically the allocated but not enabled interrupts 1020 * will not get re-targeted. But MSIs in allocated state 1021 * could be re-targeted due to group re-targeting. 1022 */ 1023 if (intr_id != NULL && dip != NULL) { 1024 ddi_intr_handle_impl_t *hdlp = intr_id; 1025 vecp = apix_get_dev_map(dip, hdlp->ih_inum, 1026 hdlp->ih_type); 1027 ASSERT(vecp->v_state == APIX_STATE_ALLOCED); 1028 } 1029 if (vecp == NULL) { 1030 lock_clear(&apix_lock); 1031 cmn_err(CE_WARN, "Invalid interrupt 0x%x,0x%x " 1032 " for %p to add", cpuid, v, intr_id); 1033 return (0); 1034 } 1035 cpuid = vecp->v_cpuid; 1036 virt_vect = APIX_VIRTVECTOR(cpuid, vecp->v_vector); 1037 } 1038 1039 APIX_ENTER_CPU_LOCK(cpuid); 1040 apix_insert_av(vecp, intr_id, xxintr, arg1, arg2, ticksp, ipl, dip); 1041 APIX_LEAVE_CPU_LOCK(cpuid); 1042 1043 (void) apix_addspl(virt_vect, ipl, 0, 0); 1044 1045 lock_clear(&apix_lock); 1046 1047 return (1); 1048 } 1049 1050 /* 1051 * Remove avintr 1052 * 1053 * For fixed, if it's the last one of shared interrupts, free the vector. 1054 * For msi/x, only disable the interrupt but not free the vector, which 1055 * is freed by PSM_XXX_FREE_XXX. 1056 */ 1057 void 1058 apix_rem_avintr(void *intr_id, int ipl, avfunc xxintr, int virt_vect) 1059 { 1060 avfunc f; 1061 apix_vector_t *vecp; 1062 struct autovec *avp; 1063 processorid_t cpuid; 1064 1065 if ((f = xxintr) == NULL) 1066 return; 1067 1068 lock_set(&apix_lock); 1069 1070 if (!APIX_IS_VIRTVEC(virt_vect)) { /* got irq */ 1071 vecp = apix_intx_get_vector(virt_vect); 1072 virt_vect = APIX_VIRTVECTOR(vecp->v_cpuid, vecp->v_vector); 1073 } else /* got virtual vector */ 1074 vecp = xv_vector(APIX_VIRTVEC_CPU(virt_vect), 1075 APIX_VIRTVEC_VECTOR(virt_vect)); 1076 1077 if (vecp == NULL) { 1078 lock_clear(&apix_lock); 1079 cmn_err(CE_CONT, "Invalid interrupt 0x%x,0x%x to remove", 1080 APIX_VIRTVEC_CPU(virt_vect), 1081 APIX_VIRTVEC_VECTOR(virt_vect)); 1082 return; 1083 } 1084 1085 if (vecp->v_state <= APIX_STATE_OBSOLETED || 1086 ((avp = apix_find_av(vecp, intr_id, f)) == NULL)) { 1087 /* 1088 * It's possible that the interrupt is rebound to a 1089 * different cpu before rem_avintr() is called. Search 1090 * through all vectors once it happens. 1091 */ 1092 if ((vecp = apix_find_vector_by_avintr(intr_id, f)) 1093 == NULL) { 1094 lock_clear(&apix_lock); 1095 cmn_err(CE_CONT, "Unknown interrupt 0x%x,0x%x " 1096 "for %p to remove", APIX_VIRTVEC_CPU(virt_vect), 1097 APIX_VIRTVEC_VECTOR(virt_vect), intr_id); 1098 return; 1099 } 1100 virt_vect = APIX_VIRTVECTOR(vecp->v_cpuid, vecp->v_vector); 1101 avp = apix_find_av(vecp, intr_id, f); 1102 } 1103 cpuid = vecp->v_cpuid; 1104 1105 /* disable interrupt */ 1106 (void) apix_delspl(virt_vect, ipl, 0, 0); 1107 1108 /* remove ISR entry */ 1109 APIX_ENTER_CPU_LOCK(cpuid); 1110 apix_remove_av(vecp, avp); 1111 APIX_LEAVE_CPU_LOCK(cpuid); 1112 1113 lock_clear(&apix_lock); 1114 } 1115 1116 /* 1117 * Device to vector mapping table 1118 */ 1119 1120 static void 1121 apix_clear_dev_map(dev_info_t *dip, int inum, int type) 1122 { 1123 char *name; 1124 major_t major; 1125 apix_dev_vector_t *dvp, *prev = NULL; 1126 int found = 0; 1127 1128 name = ddi_get_name(dip); 1129 major = ddi_name_to_major(name); 1130 1131 mutex_enter(&apix_mutex); 1132 1133 for (dvp = apix_dev_vector[major]; dvp != NULL; 1134 prev = dvp, dvp = dvp->dv_next) { 1135 if (dvp->dv_dip == dip && dvp->dv_inum == inum && 1136 dvp->dv_type == type) { 1137 found++; 1138 break; 1139 } 1140 } 1141 1142 if (!found) { 1143 mutex_exit(&apix_mutex); 1144 return; 1145 } 1146 1147 if (prev != NULL) 1148 prev->dv_next = dvp->dv_next; 1149 1150 if (apix_dev_vector[major] == dvp) 1151 apix_dev_vector[major] = dvp->dv_next; 1152 1153 dvp->dv_vector->v_devp = NULL; 1154 1155 mutex_exit(&apix_mutex); 1156 1157 kmem_free(dvp, sizeof (apix_dev_vector_t)); 1158 } 1159 1160 void 1161 apix_set_dev_map(apix_vector_t *vecp, dev_info_t *dip, int inum) 1162 { 1163 apix_dev_vector_t *dvp; 1164 char *name; 1165 major_t major; 1166 uint32_t found = 0; 1167 1168 ASSERT(dip != NULL); 1169 name = ddi_get_name(dip); 1170 major = ddi_name_to_major(name); 1171 1172 mutex_enter(&apix_mutex); 1173 1174 for (dvp = apix_dev_vector[major]; dvp != NULL; 1175 dvp = dvp->dv_next) { 1176 if (dvp->dv_dip == dip && dvp->dv_inum == inum && 1177 dvp->dv_type == vecp->v_type) { 1178 found++; 1179 break; 1180 } 1181 } 1182 1183 if (found == 0) { /* not found */ 1184 dvp = kmem_zalloc(sizeof (apix_dev_vector_t), KM_SLEEP); 1185 dvp->dv_dip = dip; 1186 dvp->dv_inum = inum; 1187 dvp->dv_type = vecp->v_type; 1188 1189 dvp->dv_next = apix_dev_vector[major]; 1190 apix_dev_vector[major] = dvp; 1191 } 1192 dvp->dv_vector = vecp; 1193 vecp->v_devp = dvp; 1194 1195 mutex_exit(&apix_mutex); 1196 1197 DDI_INTR_IMPLDBG((CE_CONT, "apix_set_dev_map: dip=0x%p " 1198 "inum=0x%x vector=0x%x/0x%x\n", 1199 (void *)dip, inum, vecp->v_cpuid, vecp->v_vector)); 1200 } 1201 1202 apix_vector_t * 1203 apix_get_dev_map(dev_info_t *dip, int inum, int type) 1204 { 1205 char *name; 1206 major_t major; 1207 apix_dev_vector_t *dvp; 1208 apix_vector_t *vecp; 1209 1210 name = ddi_get_name(dip); 1211 if ((major = ddi_name_to_major(name)) == DDI_MAJOR_T_NONE) 1212 return (NULL); 1213 1214 mutex_enter(&apix_mutex); 1215 for (dvp = apix_dev_vector[major]; dvp != NULL; 1216 dvp = dvp->dv_next) { 1217 if (dvp->dv_dip == dip && dvp->dv_inum == inum && 1218 dvp->dv_type == type) { 1219 vecp = dvp->dv_vector; 1220 mutex_exit(&apix_mutex); 1221 return (vecp); 1222 } 1223 } 1224 mutex_exit(&apix_mutex); 1225 1226 return (NULL); 1227 } 1228 1229 /* 1230 * Get minimum inum for specified device, used for MSI 1231 */ 1232 int 1233 apix_get_min_dev_inum(dev_info_t *dip, int type) 1234 { 1235 char *name; 1236 major_t major; 1237 apix_dev_vector_t *dvp; 1238 int inum = -1; 1239 1240 name = ddi_get_name(dip); 1241 major = ddi_name_to_major(name); 1242 1243 mutex_enter(&apix_mutex); 1244 for (dvp = apix_dev_vector[major]; dvp != NULL; 1245 dvp = dvp->dv_next) { 1246 if (dvp->dv_dip == dip && dvp->dv_type == type) { 1247 if (inum == -1) 1248 inum = dvp->dv_inum; 1249 else 1250 inum = (dvp->dv_inum < inum) ? 1251 dvp->dv_inum : inum; 1252 } 1253 } 1254 mutex_exit(&apix_mutex); 1255 1256 return (inum); 1257 } 1258 1259 int 1260 apix_get_max_dev_inum(dev_info_t *dip, int type) 1261 { 1262 char *name; 1263 major_t major; 1264 apix_dev_vector_t *dvp; 1265 int inum = -1; 1266 1267 name = ddi_get_name(dip); 1268 major = ddi_name_to_major(name); 1269 1270 mutex_enter(&apix_mutex); 1271 for (dvp = apix_dev_vector[major]; dvp != NULL; 1272 dvp = dvp->dv_next) { 1273 if (dvp->dv_dip == dip && dvp->dv_type == type) { 1274 if (inum == -1) 1275 inum = dvp->dv_inum; 1276 else 1277 inum = (dvp->dv_inum > inum) ? 1278 dvp->dv_inum : inum; 1279 } 1280 } 1281 mutex_exit(&apix_mutex); 1282 1283 return (inum); 1284 } 1285 1286 /* 1287 * Major to cpu binding, for INTR_ROUND_ROBIN_WITH_AFFINITY cpu 1288 * binding policy 1289 */ 1290 1291 static uint32_t 1292 apix_get_dev_binding(dev_info_t *dip) 1293 { 1294 major_t major; 1295 char *name; 1296 uint32_t cpu = IRQ_UNINIT; 1297 1298 name = ddi_get_name(dip); 1299 major = ddi_name_to_major(name); 1300 if (major < devcnt) { 1301 mutex_enter(&apix_mutex); 1302 cpu = apix_major_to_cpu[major]; 1303 mutex_exit(&apix_mutex); 1304 } 1305 1306 return (cpu); 1307 } 1308 1309 static void 1310 apix_set_dev_binding(dev_info_t *dip, uint32_t cpu) 1311 { 1312 major_t major; 1313 char *name; 1314 1315 /* setup major to cpu mapping */ 1316 name = ddi_get_name(dip); 1317 major = ddi_name_to_major(name); 1318 if (apix_major_to_cpu[major] == IRQ_UNINIT) { 1319 mutex_enter(&apix_mutex); 1320 apix_major_to_cpu[major] = cpu; 1321 mutex_exit(&apix_mutex); 1322 } 1323 } 1324 1325 /* 1326 * return the cpu to which this intr should be bound. 1327 * Check properties or any other mechanism to see if user wants it 1328 * bound to a specific CPU. If so, return the cpu id with high bit set. 1329 * If not, use the policy to choose a cpu and return the id. 1330 */ 1331 uint32_t 1332 apix_bind_cpu(dev_info_t *dip) 1333 { 1334 int instance, instno, prop_len, bind_cpu, count; 1335 uint_t i, rc; 1336 major_t major; 1337 char *name, *drv_name, *prop_val, *cptr; 1338 char prop_name[32]; 1339 1340 lock_set(&apix_lock); 1341 1342 if (apic_intr_policy == INTR_LOWEST_PRIORITY) { 1343 cmn_err(CE_WARN, "apix: unsupported interrupt binding policy " 1344 "LOWEST PRIORITY, use ROUND ROBIN instead"); 1345 apic_intr_policy = INTR_ROUND_ROBIN; 1346 } 1347 1348 if (apic_nproc == 1) { 1349 lock_clear(&apix_lock); 1350 return (0); 1351 } 1352 1353 drv_name = NULL; 1354 rc = DDI_PROP_NOT_FOUND; 1355 major = (major_t)-1; 1356 if (dip != NULL) { 1357 name = ddi_get_name(dip); 1358 major = ddi_name_to_major(name); 1359 drv_name = ddi_major_to_name(major); 1360 instance = ddi_get_instance(dip); 1361 if (apic_intr_policy == INTR_ROUND_ROBIN_WITH_AFFINITY) { 1362 bind_cpu = apix_get_dev_binding(dip); 1363 if (bind_cpu != IRQ_UNINIT) { 1364 lock_clear(&apix_lock); 1365 return (bind_cpu); 1366 } 1367 } 1368 /* 1369 * search for "drvname"_intpt_bind_cpus property first, the 1370 * syntax of the property should be "a[,b,c,...]" where 1371 * instance 0 binds to cpu a, instance 1 binds to cpu b, 1372 * instance 3 binds to cpu c... 1373 * ddi_getlongprop() will search /option first, then / 1374 * if "drvname"_intpt_bind_cpus doesn't exist, then find 1375 * intpt_bind_cpus property. The syntax is the same, and 1376 * it applies to all the devices if its "drvname" specific 1377 * property doesn't exist 1378 */ 1379 (void) strcpy(prop_name, drv_name); 1380 (void) strcat(prop_name, "_intpt_bind_cpus"); 1381 rc = ddi_getlongprop(DDI_DEV_T_ANY, dip, 0, prop_name, 1382 (caddr_t)&prop_val, &prop_len); 1383 if (rc != DDI_PROP_SUCCESS) { 1384 rc = ddi_getlongprop(DDI_DEV_T_ANY, dip, 0, 1385 "intpt_bind_cpus", (caddr_t)&prop_val, &prop_len); 1386 } 1387 } 1388 if (rc == DDI_PROP_SUCCESS) { 1389 for (i = count = 0; i < (prop_len - 1); i++) 1390 if (prop_val[i] == ',') 1391 count++; 1392 if (prop_val[i-1] != ',') 1393 count++; 1394 /* 1395 * if somehow the binding instances defined in the 1396 * property are not enough for this instno., then 1397 * reuse the pattern for the next instance until 1398 * it reaches the requested instno 1399 */ 1400 instno = instance % count; 1401 i = 0; 1402 cptr = prop_val; 1403 while (i < instno) 1404 if (*cptr++ == ',') 1405 i++; 1406 bind_cpu = stoi(&cptr); 1407 kmem_free(prop_val, prop_len); 1408 /* if specific cpu is bogus, then default to cpu 0 */ 1409 if (bind_cpu >= apic_nproc) { 1410 cmn_err(CE_WARN, "apix: %s=%s: CPU %d not present", 1411 prop_name, prop_val, bind_cpu); 1412 bind_cpu = 0; 1413 } else { 1414 /* indicate that we are bound at user request */ 1415 bind_cpu |= IRQ_USER_BOUND; 1416 } 1417 /* 1418 * no need to check apic_cpus[].aci_status, if specific cpu is 1419 * not up, then post_cpu_start will handle it. 1420 */ 1421 } else { 1422 bind_cpu = apic_get_next_bind_cpu(); 1423 } 1424 1425 lock_clear(&apix_lock); 1426 1427 return ((uint32_t)bind_cpu); 1428 } 1429 1430 static boolean_t 1431 apix_is_cpu_enabled(processorid_t cpuid) 1432 { 1433 apic_cpus_info_t *cpu_infop; 1434 1435 cpu_infop = &apic_cpus[cpuid]; 1436 1437 if ((cpu_infop->aci_status & APIC_CPU_INTR_ENABLE) == 0) 1438 return (B_FALSE); 1439 1440 return (B_TRUE); 1441 } 1442 1443 /* 1444 * Must be called with apix_lock held. This function can be 1445 * called from above lock level by apix_intr_redistribute(). 1446 * 1447 * Arguments: 1448 * vecp : Vector to be rebound 1449 * tocpu : Target cpu. IRQ_UNINIT means target is vecp->v_cpuid. 1450 * count : Number of continuous vectors 1451 * 1452 * Return new vector being bound to 1453 */ 1454 apix_vector_t * 1455 apix_rebind(apix_vector_t *vecp, processorid_t newcpu, int count) 1456 { 1457 apix_vector_t *newp, *oldp; 1458 processorid_t oldcpu = vecp->v_cpuid; 1459 uchar_t newvec, oldvec = vecp->v_vector; 1460 int i; 1461 1462 ASSERT(LOCK_HELD(&apix_lock) && count > 0); 1463 1464 if (!apix_is_cpu_enabled(newcpu)) 1465 return (NULL); 1466 1467 if (vecp->v_cpuid == newcpu) /* rebind to the same cpu */ 1468 return (vecp); 1469 1470 APIX_ENTER_CPU_LOCK(oldcpu); 1471 APIX_ENTER_CPU_LOCK(newcpu); 1472 1473 /* allocate vector */ 1474 if (count == 1) 1475 newp = apix_alloc_vector_oncpu(newcpu, NULL, 0, vecp->v_type); 1476 else { 1477 ASSERT(vecp->v_type == APIX_TYPE_MSI); 1478 newp = apix_alloc_nvectors_oncpu(newcpu, NULL, 0, count, 1479 vecp->v_type); 1480 } 1481 if (newp == NULL) { 1482 APIX_LEAVE_CPU_LOCK(newcpu); 1483 APIX_LEAVE_CPU_LOCK(oldcpu); 1484 return (NULL); 1485 } 1486 1487 newvec = newp->v_vector; 1488 apix_dup_vectors(vecp, newp, count); 1489 1490 APIX_LEAVE_CPU_LOCK(newcpu); 1491 APIX_LEAVE_CPU_LOCK(oldcpu); 1492 1493 if (!DDI_INTR_IS_MSI_OR_MSIX(vecp->v_type)) { 1494 ASSERT(count == 1); 1495 if (apix_intx_rebind(vecp->v_inum, newcpu, newvec) != 0) { 1496 struct autovec *avp; 1497 int inum; 1498 1499 /* undo duplication */ 1500 APIX_ENTER_CPU_LOCK(oldcpu); 1501 APIX_ENTER_CPU_LOCK(newcpu); 1502 for (avp = newp->v_autovect; avp != NULL; 1503 avp = avp->av_link) { 1504 if (avp->av_dip != NULL) { 1505 inum = GET_INTR_INUM(avp->av_intr_id); 1506 apix_set_dev_map(vecp, avp->av_dip, 1507 inum); 1508 } 1509 apix_remove_av(newp, avp); 1510 } 1511 apix_cleanup_vector(newp); 1512 APIX_LEAVE_CPU_LOCK(newcpu); 1513 APIX_LEAVE_CPU_LOCK(oldcpu); 1514 APIC_VERBOSE(REBIND, (CE_CONT, "apix: rebind fixed " 1515 "interrupt 0x%x to cpu %d failed\n", 1516 vecp->v_inum, newcpu)); 1517 return (NULL); 1518 } 1519 1520 APIX_ENTER_CPU_LOCK(oldcpu); 1521 (void) apix_obsolete_vector(vecp); 1522 APIX_LEAVE_CPU_LOCK(oldcpu); 1523 APIC_VERBOSE(REBIND, (CE_CONT, "apix: rebind fixed interrupt" 1524 " 0x%x/0x%x to 0x%x/0x%x\n", 1525 oldcpu, oldvec, newcpu, newvec)); 1526 return (newp); 1527 } 1528 1529 for (i = 0; i < count; i++) { 1530 oldp = xv_vector(oldcpu, oldvec + i); 1531 newp = xv_vector(newcpu, newvec + i); 1532 1533 if (newp->v_share > 0) { 1534 APIX_SET_REBIND_INFO(oldp, newp); 1535 1536 apix_enable_vector(newp); 1537 1538 APIX_CLR_REBIND_INFO(); 1539 } 1540 1541 APIX_ENTER_CPU_LOCK(oldcpu); 1542 (void) apix_obsolete_vector(oldp); 1543 APIX_LEAVE_CPU_LOCK(oldcpu); 1544 } 1545 APIC_VERBOSE(REBIND, (CE_CONT, "apix: rebind vector 0x%x/0x%x " 1546 "to 0x%x/0x%x, count=%d\n", 1547 oldcpu, oldvec, newcpu, newvec, count)); 1548 1549 return (xv_vector(newcpu, newvec)); 1550 } 1551 1552 /* 1553 * Senarios include: 1554 * a. add_avintr() is called before irqp initialized (legacy) 1555 * b. irqp is initialized, vector is not allocated (fixed interrupts) 1556 * c. irqp is initialized, vector is allocated (shared interrupts) 1557 */ 1558 apix_vector_t * 1559 apix_alloc_intx(dev_info_t *dip, int inum, int irqno) 1560 { 1561 apic_irq_t *irqp; 1562 apix_vector_t *vecp; 1563 1564 /* 1565 * Allocate IRQ. Caller is later responsible for the 1566 * initialization 1567 */ 1568 mutex_enter(&airq_mutex); 1569 if ((irqp = apic_irq_table[irqno]) == NULL) { 1570 /* allocate irq */ 1571 irqp = kmem_zalloc(sizeof (apic_irq_t), KM_SLEEP); 1572 irqp->airq_mps_intr_index = FREE_INDEX; 1573 apic_irq_table[irqno] = irqp; 1574 } 1575 if (irqp->airq_mps_intr_index == FREE_INDEX) { 1576 irqp->airq_mps_intr_index = DEFAULT_INDEX; 1577 irqp->airq_cpu = IRQ_UNINIT; 1578 irqp->airq_origirq = (uchar_t)irqno; 1579 } 1580 1581 mutex_exit(&airq_mutex); 1582 1583 /* 1584 * allocate vector 1585 */ 1586 if (irqp->airq_cpu == IRQ_UNINIT) { 1587 uint32_t bindcpu, cpuid; 1588 1589 /* select cpu by system policy */ 1590 bindcpu = apix_bind_cpu(dip); 1591 cpuid = bindcpu & ~IRQ_USER_BOUND; 1592 1593 /* allocate vector */ 1594 APIX_ENTER_CPU_LOCK(cpuid); 1595 1596 if ((vecp = apix_alloc_vector_oncpu(bindcpu, dip, inum, 1597 APIX_TYPE_FIXED)) == NULL) { 1598 cmn_err(CE_WARN, "No interrupt vector for irq %x", 1599 irqno); 1600 APIX_LEAVE_CPU_LOCK(cpuid); 1601 return (NULL); 1602 } 1603 vecp->v_inum = irqno; 1604 vecp->v_flags |= APIX_VECT_MASKABLE; 1605 1606 apix_intx_set_vector(irqno, vecp->v_cpuid, vecp->v_vector); 1607 1608 APIX_LEAVE_CPU_LOCK(cpuid); 1609 } else { 1610 vecp = xv_vector(irqp->airq_cpu, irqp->airq_vector); 1611 ASSERT(!IS_VECT_FREE(vecp)); 1612 1613 if (dip != NULL) 1614 apix_set_dev_map(vecp, dip, inum); 1615 } 1616 1617 if ((dip != NULL) && 1618 (apic_intr_policy == INTR_ROUND_ROBIN_WITH_AFFINITY) && 1619 ((vecp->v_flags & APIX_VECT_USER_BOUND) == 0)) 1620 apix_set_dev_binding(dip, vecp->v_cpuid); 1621 1622 apix_dprint_vector(vecp, dip, 1); 1623 1624 return (vecp); 1625 } 1626 1627 int 1628 apix_alloc_msi(dev_info_t *dip, int inum, int count, int behavior) 1629 { 1630 int i, cap_ptr, rcount = count; 1631 apix_vector_t *vecp; 1632 processorid_t bindcpu, cpuid; 1633 ushort_t msi_ctrl; 1634 ddi_acc_handle_t handle; 1635 1636 DDI_INTR_IMPLDBG((CE_CONT, "apix_alloc_msi_vectors: dip=0x%p " 1637 "inum=0x%x count=0x%x behavior=%d\n", 1638 (void *)dip, inum, count, behavior)); 1639 1640 if (count > 1) { 1641 if (behavior == DDI_INTR_ALLOC_STRICT && 1642 apic_multi_msi_enable == 0) 1643 return (0); 1644 if (apic_multi_msi_enable == 0) 1645 count = 1; 1646 } 1647 1648 /* Check whether it supports per-vector masking */ 1649 cap_ptr = i_ddi_get_msi_msix_cap_ptr(dip); 1650 handle = i_ddi_get_pci_config_handle(dip); 1651 msi_ctrl = pci_config_get16(handle, cap_ptr + PCI_MSI_CTRL); 1652 1653 /* bind to cpu */ 1654 bindcpu = apix_bind_cpu(dip); 1655 cpuid = bindcpu & ~IRQ_USER_BOUND; 1656 1657 /* if not ISP2, then round it down */ 1658 if (!ISP2(rcount)) 1659 rcount = 1 << (highbit(rcount) - 1); 1660 1661 APIX_ENTER_CPU_LOCK(cpuid); 1662 for (vecp = NULL; rcount > 0; rcount >>= 1) { 1663 vecp = apix_alloc_nvectors_oncpu(bindcpu, dip, inum, rcount, 1664 APIX_TYPE_MSI); 1665 if (vecp != NULL || behavior == DDI_INTR_ALLOC_STRICT) 1666 break; 1667 } 1668 for (i = 0; vecp && i < rcount; i++) 1669 xv_vector(vecp->v_cpuid, vecp->v_vector + i)->v_flags |= 1670 (msi_ctrl & PCI_MSI_PVM_MASK) ? APIX_VECT_MASKABLE : 0; 1671 APIX_LEAVE_CPU_LOCK(cpuid); 1672 if (vecp == NULL) { 1673 APIC_VERBOSE(INTR, (CE_CONT, 1674 "apix_alloc_msi: no %d cont vectors found on cpu 0x%x\n", 1675 count, bindcpu)); 1676 return (0); 1677 } 1678 1679 /* major to cpu binding */ 1680 if ((apic_intr_policy == INTR_ROUND_ROBIN_WITH_AFFINITY) && 1681 ((vecp->v_flags & APIX_VECT_USER_BOUND) == 0)) 1682 apix_set_dev_binding(dip, vecp->v_cpuid); 1683 1684 apix_dprint_vector(vecp, dip, rcount); 1685 1686 return (rcount); 1687 } 1688 1689 int 1690 apix_alloc_msix(dev_info_t *dip, int inum, int count, int behavior) 1691 { 1692 apix_vector_t *vecp; 1693 processorid_t bindcpu, cpuid; 1694 int i; 1695 1696 for (i = 0; i < count; i++) { 1697 /* select cpu by system policy */ 1698 bindcpu = apix_bind_cpu(dip); 1699 cpuid = bindcpu & ~IRQ_USER_BOUND; 1700 1701 /* allocate vector */ 1702 APIX_ENTER_CPU_LOCK(cpuid); 1703 if ((vecp = apix_alloc_vector_oncpu(bindcpu, dip, inum + i, 1704 APIX_TYPE_MSIX)) == NULL) { 1705 APIX_LEAVE_CPU_LOCK(cpuid); 1706 APIC_VERBOSE(INTR, (CE_CONT, "apix_alloc_msix: " 1707 "allocate msix for device dip=%p, inum=%d on" 1708 " cpu %d failed", (void *)dip, inum + i, bindcpu)); 1709 break; 1710 } 1711 vecp->v_flags |= APIX_VECT_MASKABLE; 1712 APIX_LEAVE_CPU_LOCK(cpuid); 1713 1714 /* major to cpu mapping */ 1715 if ((i == 0) && 1716 (apic_intr_policy == INTR_ROUND_ROBIN_WITH_AFFINITY) && 1717 ((vecp->v_flags & APIX_VECT_USER_BOUND) == 0)) 1718 apix_set_dev_binding(dip, vecp->v_cpuid); 1719 1720 apix_dprint_vector(vecp, dip, 1); 1721 } 1722 1723 if (i < count && behavior == DDI_INTR_ALLOC_STRICT) { 1724 APIC_VERBOSE(INTR, (CE_WARN, "apix_alloc_msix: " 1725 "strictly allocate %d vectors failed, got %d\n", 1726 count, i)); 1727 apix_free_vectors(dip, inum, i, APIX_TYPE_MSIX); 1728 i = 0; 1729 } 1730 1731 return (i); 1732 } 1733 1734 /* 1735 * A rollback free for vectors allocated by apix_alloc_xxx(). 1736 */ 1737 void 1738 apix_free_vectors(dev_info_t *dip, int inum, int count, int type) 1739 { 1740 int i, cpuid; 1741 apix_vector_t *vecp; 1742 1743 DDI_INTR_IMPLDBG((CE_CONT, "apix_free_vectors: dip: %p inum: %x " 1744 "count: %x type: %x\n", 1745 (void *)dip, inum, count, type)); 1746 1747 lock_set(&apix_lock); 1748 1749 for (i = 0; i < count; i++, inum++) { 1750 if ((vecp = apix_get_dev_map(dip, inum, type)) == NULL) { 1751 lock_clear(&apix_lock); 1752 DDI_INTR_IMPLDBG((CE_CONT, "apix_free_vectors: " 1753 "dip=0x%p inum=0x%x type=0x%x apix_find_intr() " 1754 "failed\n", (void *)dip, inum, type)); 1755 continue; 1756 } 1757 1758 APIX_ENTER_CPU_LOCK(vecp->v_cpuid); 1759 cpuid = vecp->v_cpuid; 1760 1761 DDI_INTR_IMPLDBG((CE_CONT, "apix_free_vectors: " 1762 "dip=0x%p inum=0x%x type=0x%x vector 0x%x (share %d)\n", 1763 (void *)dip, inum, type, vecp->v_vector, vecp->v_share)); 1764 1765 /* tear down device interrupt to vector mapping */ 1766 apix_clear_dev_map(dip, inum, type); 1767 1768 if (vecp->v_type == APIX_TYPE_FIXED) { 1769 if (vecp->v_share > 0) { /* share IRQ line */ 1770 APIX_LEAVE_CPU_LOCK(cpuid); 1771 continue; 1772 } 1773 1774 /* Free apic_irq_table entry */ 1775 apix_intx_free(vecp->v_inum); 1776 } 1777 1778 /* free vector */ 1779 apix_cleanup_vector(vecp); 1780 1781 APIX_LEAVE_CPU_LOCK(cpuid); 1782 } 1783 1784 lock_clear(&apix_lock); 1785 } 1786 1787 /* 1788 * Must be called with apix_lock held 1789 */ 1790 apix_vector_t * 1791 apix_setup_io_intr(apix_vector_t *vecp) 1792 { 1793 processorid_t bindcpu; 1794 int ret; 1795 1796 ASSERT(LOCK_HELD(&apix_lock)); 1797 1798 /* 1799 * Interrupts are enabled on the CPU, programme IOAPIC RDT 1800 * entry or MSI/X address/data to enable the interrupt. 1801 */ 1802 if (apix_is_cpu_enabled(vecp->v_cpuid)) { 1803 apix_enable_vector(vecp); 1804 return (vecp); 1805 } 1806 1807 /* 1808 * CPU is not up or interrupts are disabled. Fall back to the 1809 * first avialable CPU. 1810 */ 1811 bindcpu = apic_find_cpu(APIC_CPU_INTR_ENABLE); 1812 1813 if (vecp->v_type == APIX_TYPE_MSI) 1814 return (apix_grp_set_cpu(vecp, bindcpu, &ret)); 1815 1816 return (apix_set_cpu(vecp, bindcpu, &ret)); 1817 } 1818 1819 /* 1820 * For interrupts which call add_avintr() before apic is initialized. 1821 * ioapix_setup_intr() will 1822 * - allocate vector 1823 * - copy over ISR 1824 */ 1825 static void 1826 ioapix_setup_intr(int irqno, iflag_t *flagp) 1827 { 1828 extern struct av_head autovect[]; 1829 apix_vector_t *vecp; 1830 apic_irq_t *irqp; 1831 uchar_t ioapicindex, ipin; 1832 ulong_t iflag; 1833 struct autovec *avp; 1834 1835 ioapicindex = acpi_find_ioapic(irqno); 1836 ASSERT(ioapicindex != 0xFF); 1837 ipin = irqno - apic_io_vectbase[ioapicindex]; 1838 1839 mutex_enter(&airq_mutex); 1840 irqp = apic_irq_table[irqno]; 1841 1842 /* 1843 * The irq table entry shouldn't exist unless the interrupts are shared. 1844 * In that case, make sure it matches what we would initialize it to. 1845 */ 1846 if (irqp != NULL) { 1847 ASSERT(irqp->airq_mps_intr_index == ACPI_INDEX); 1848 ASSERT(irqp->airq_intin_no == ipin && 1849 irqp->airq_ioapicindex == ioapicindex); 1850 vecp = xv_vector(irqp->airq_cpu, irqp->airq_vector); 1851 ASSERT(!IS_VECT_FREE(vecp)); 1852 mutex_exit(&airq_mutex); 1853 } else { 1854 irqp = kmem_zalloc(sizeof (apic_irq_t), KM_SLEEP); 1855 1856 irqp->airq_cpu = IRQ_UNINIT; 1857 irqp->airq_origirq = (uchar_t)irqno; 1858 irqp->airq_mps_intr_index = ACPI_INDEX; 1859 irqp->airq_ioapicindex = ioapicindex; 1860 irqp->airq_intin_no = ipin; 1861 irqp->airq_iflag = *flagp; 1862 irqp->airq_share++; 1863 1864 apic_irq_table[irqno] = irqp; 1865 mutex_exit(&airq_mutex); 1866 1867 vecp = apix_alloc_intx(NULL, 0, irqno); 1868 } 1869 1870 /* copy over autovect */ 1871 for (avp = autovect[irqno].avh_link; avp; avp = avp->av_link) 1872 apix_insert_av(vecp, avp->av_intr_id, avp->av_vector, 1873 avp->av_intarg1, avp->av_intarg2, avp->av_ticksp, 1874 avp->av_prilevel, avp->av_dip); 1875 1876 /* Program I/O APIC */ 1877 iflag = intr_clear(); 1878 lock_set(&apix_lock); 1879 1880 (void) apix_setup_io_intr(vecp); 1881 1882 lock_clear(&apix_lock); 1883 intr_restore(iflag); 1884 1885 APIC_VERBOSE_IOAPIC((CE_CONT, "apix: setup ioapic, irqno %x " 1886 "(ioapic %x, ipin %x) is bound to cpu %x, vector %x\n", 1887 irqno, ioapicindex, ipin, irqp->airq_cpu, irqp->airq_vector)); 1888 } 1889 1890 void 1891 ioapix_init_intr(int mask_apic) 1892 { 1893 int ioapicindex; 1894 int i, j; 1895 1896 /* mask interrupt vectors */ 1897 for (j = 0; j < apic_io_max && mask_apic; j++) { 1898 int intin_max; 1899 1900 ioapicindex = j; 1901 /* Bits 23-16 define the maximum redirection entries */ 1902 intin_max = (ioapic_read(ioapicindex, APIC_VERS_CMD) >> 16) 1903 & 0xff; 1904 for (i = 0; i <= intin_max; i++) 1905 ioapic_write(ioapicindex, APIC_RDT_CMD + 2 * i, 1906 AV_MASK); 1907 } 1908 1909 /* 1910 * Hack alert: deal with ACPI SCI interrupt chicken/egg here 1911 */ 1912 if (apic_sci_vect > 0) 1913 ioapix_setup_intr(apic_sci_vect, &apic_sci_flags); 1914 1915 /* 1916 * Hack alert: deal with ACPI HPET interrupt chicken/egg here. 1917 */ 1918 if (apic_hpet_vect > 0) 1919 ioapix_setup_intr(apic_hpet_vect, &apic_hpet_flags); 1920 } 1921