1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* 23 * Copyright (c) 2010, Oracle and/or its affiliates. All rights reserved. 24 */ 25 /* 26 * Copyright (c) 2010, Intel Corporation. 27 * All rights reserved. 28 */ 29 /* 30 * Copyright 2013 Nexenta Systems, Inc. All rights reserved. 31 * Copyright 2013 Pluribus Networks, Inc. 32 * Copyright 2019 Joyent, Inc. 33 */ 34 35 #include <sys/processor.h> 36 #include <sys/time.h> 37 #include <sys/psm.h> 38 #include <sys/smp_impldefs.h> 39 #include <sys/cram.h> 40 #include <sys/acpi/acpi.h> 41 #include <sys/acpica.h> 42 #include <sys/psm_common.h> 43 #include <sys/pit.h> 44 #include <sys/ddi.h> 45 #include <sys/sunddi.h> 46 #include <sys/ddi_impldefs.h> 47 #include <sys/pci.h> 48 #include <sys/promif.h> 49 #include <sys/x86_archext.h> 50 #include <sys/cpc_impl.h> 51 #include <sys/uadmin.h> 52 #include <sys/panic.h> 53 #include <sys/debug.h> 54 #include <sys/archsystm.h> 55 #include <sys/trap.h> 56 #include <sys/machsystm.h> 57 #include <sys/sysmacros.h> 58 #include <sys/cpuvar.h> 59 #include <sys/rm_platter.h> 60 #include <sys/privregs.h> 61 #include <sys/note.h> 62 #include <sys/pci_intr_lib.h> 63 #include <sys/spl.h> 64 #include <sys/clock.h> 65 #include <sys/dditypes.h> 66 #include <sys/sunddi.h> 67 #include <sys/x_call.h> 68 #include <sys/reboot.h> 69 #include <sys/apix.h> 70 #include <sys/smt.h> 71 72 static int apix_get_avail_vector_oncpu(uint32_t, int, int); 73 static apix_vector_t *apix_init_vector(processorid_t, uchar_t); 74 static void apix_cleanup_vector(apix_vector_t *); 75 static void apix_insert_av(apix_vector_t *, void *, avfunc, caddr_t, caddr_t, 76 uint64_t *, int, dev_info_t *); 77 static void apix_remove_av(apix_vector_t *, struct autovec *); 78 static void apix_clear_dev_map(dev_info_t *, int, int); 79 static boolean_t apix_is_cpu_enabled(processorid_t); 80 static void apix_wait_till_seen(processorid_t, int); 81 82 #define GET_INTR_INUM(ihdlp) \ 83 (((ihdlp) != NULL) ? ((ddi_intr_handle_impl_t *)(ihdlp))->ih_inum : 0) 84 85 apix_rebind_info_t apix_rebindinfo = {0, 0, 0, NULL, 0, NULL}; 86 87 /* 88 * Allocate IPI 89 * 90 * Return vector number or 0 on error 91 */ 92 uchar_t 93 apix_alloc_ipi(int ipl) 94 { 95 apix_vector_t *vecp; 96 uchar_t vector; 97 int cpun; 98 int nproc; 99 100 APIX_ENTER_CPU_LOCK(0); 101 102 vector = apix_get_avail_vector_oncpu(0, APIX_IPI_MIN, APIX_IPI_MAX); 103 if (vector == 0) { 104 APIX_LEAVE_CPU_LOCK(0); 105 cmn_err(CE_WARN, "apix: no available IPI\n"); 106 apic_error |= APIC_ERR_GET_IPIVECT_FAIL; 107 return (0); 108 } 109 110 nproc = max(apic_nproc, apic_max_nproc); 111 for (cpun = 0; cpun < nproc; cpun++) { 112 vecp = xv_vector(cpun, vector); 113 if (vecp == NULL) { 114 vecp = kmem_zalloc(sizeof (apix_vector_t), KM_NOSLEEP); 115 if (vecp == NULL) { 116 cmn_err(CE_WARN, "apix: No memory for ipi"); 117 goto fail; 118 } 119 xv_vector(cpun, vector) = vecp; 120 } 121 vecp->v_state = APIX_STATE_ALLOCED; 122 vecp->v_type = APIX_TYPE_IPI; 123 vecp->v_cpuid = vecp->v_bound_cpuid = cpun; 124 vecp->v_vector = vector; 125 vecp->v_pri = ipl; 126 } 127 APIX_LEAVE_CPU_LOCK(0); 128 return (vector); 129 130 fail: 131 while (--cpun >= 0) 132 apix_cleanup_vector(xv_vector(cpun, vector)); 133 APIX_LEAVE_CPU_LOCK(0); 134 return (0); 135 } 136 137 /* 138 * Add IPI service routine 139 */ 140 static int 141 apix_add_ipi(int ipl, avfunc xxintr, char *name, int vector, 142 caddr_t arg1, caddr_t arg2) 143 { 144 int cpun; 145 apix_vector_t *vecp; 146 int nproc; 147 148 ASSERT(vector >= APIX_IPI_MIN && vector <= APIX_IPI_MAX); 149 150 nproc = max(apic_nproc, apic_max_nproc); 151 for (cpun = 0; cpun < nproc; cpun++) { 152 APIX_ENTER_CPU_LOCK(cpun); 153 vecp = xv_vector(cpun, vector); 154 apix_insert_av(vecp, NULL, xxintr, arg1, arg2, NULL, ipl, NULL); 155 vecp->v_state = APIX_STATE_ENABLED; 156 APIX_LEAVE_CPU_LOCK(cpun); 157 } 158 159 APIC_VERBOSE(IPI, (CE_CONT, "apix: add ipi for %s, vector %x " 160 "ipl %x\n", name, vector, ipl)); 161 162 return (1); 163 } 164 165 /* 166 * Find and return first free vector in range (start, end) 167 */ 168 static int 169 apix_get_avail_vector_oncpu(uint32_t cpuid, int start, int end) 170 { 171 int i; 172 apix_impl_t *apixp = apixs[cpuid]; 173 174 for (i = start; i <= end; i++) { 175 if (APIC_CHECK_RESERVE_VECTORS(i)) 176 continue; 177 if (IS_VECT_FREE(apixp->x_vectbl[i])) 178 return (i); 179 } 180 181 return (0); 182 } 183 184 /* 185 * Allocate a vector on specified cpu 186 * 187 * Return NULL on error 188 */ 189 static apix_vector_t * 190 apix_alloc_vector_oncpu(uint32_t cpuid, dev_info_t *dip, int inum, int type) 191 { 192 processorid_t tocpu = cpuid & ~IRQ_USER_BOUND; 193 apix_vector_t *vecp; 194 int vector; 195 196 ASSERT(APIX_CPU_LOCK_HELD(tocpu)); 197 198 /* find free vector */ 199 vector = apix_get_avail_vector_oncpu(tocpu, APIX_AVINTR_MIN, 200 APIX_AVINTR_MAX); 201 if (vector == 0) 202 return (NULL); 203 204 vecp = apix_init_vector(tocpu, vector); 205 vecp->v_type = (ushort_t)type; 206 vecp->v_inum = inum; 207 vecp->v_flags = (cpuid & IRQ_USER_BOUND) ? APIX_VECT_USER_BOUND : 0; 208 209 if (dip != NULL) 210 apix_set_dev_map(vecp, dip, inum); 211 212 return (vecp); 213 } 214 215 /* 216 * Allocates "count" contiguous MSI vectors starting at the proper alignment. 217 * Caller needs to make sure that count has to be power of 2 and should not 218 * be < 1. 219 * 220 * Return first vector number 221 */ 222 apix_vector_t * 223 apix_alloc_nvectors_oncpu(uint32_t cpuid, dev_info_t *dip, int inum, 224 int count, int type) 225 { 226 int i, msibits, start = 0, navail = 0; 227 apix_vector_t *vecp, *startp = NULL; 228 processorid_t tocpu = cpuid & ~IRQ_USER_BOUND; 229 uint_t flags; 230 231 ASSERT(APIX_CPU_LOCK_HELD(tocpu)); 232 233 /* 234 * msibits is the no. of lower order message data bits for the 235 * allocated MSI vectors and is used to calculate the aligned 236 * starting vector 237 */ 238 msibits = count - 1; 239 240 /* It has to be contiguous */ 241 for (i = APIX_AVINTR_MIN; i <= APIX_AVINTR_MAX; i++) { 242 if (!IS_VECT_FREE(xv_vector(tocpu, i))) 243 continue; 244 245 /* 246 * starting vector has to be aligned accordingly for 247 * multiple MSIs 248 */ 249 if (msibits) 250 i = (i + msibits) & ~msibits; 251 252 for (navail = 0, start = i; i <= APIX_AVINTR_MAX; i++) { 253 if (!IS_VECT_FREE(xv_vector(tocpu, i))) 254 break; 255 if (APIC_CHECK_RESERVE_VECTORS(i)) 256 break; 257 if (++navail == count) 258 goto done; 259 } 260 } 261 262 return (NULL); 263 264 done: 265 flags = (cpuid & IRQ_USER_BOUND) ? APIX_VECT_USER_BOUND : 0; 266 267 for (i = 0; i < count; i++) { 268 if ((vecp = apix_init_vector(tocpu, start + i)) == NULL) 269 goto fail; 270 271 vecp->v_type = (ushort_t)type; 272 vecp->v_inum = inum + i; 273 vecp->v_flags = flags; 274 275 if (dip != NULL) 276 apix_set_dev_map(vecp, dip, inum + i); 277 278 if (i == 0) 279 startp = vecp; 280 } 281 282 return (startp); 283 284 fail: 285 while (i-- > 0) { /* Free allocated vectors */ 286 vecp = xv_vector(tocpu, start + i); 287 apix_clear_dev_map(dip, inum + i, type); 288 apix_cleanup_vector(vecp); 289 } 290 return (NULL); 291 } 292 293 #define APIX_WRITE_MSI_DATA(_hdl, _cap, _ctrl, _v)\ 294 do {\ 295 if ((_ctrl) & PCI_MSI_64BIT_MASK)\ 296 pci_config_put16((_hdl), (_cap) + PCI_MSI_64BIT_DATA, (_v));\ 297 else\ 298 pci_config_put16((_hdl), (_cap) + PCI_MSI_32BIT_DATA, (_v));\ 299 _NOTE(CONSTCOND)} while (0) 300 301 static void 302 apix_pci_msi_enable_vector(apix_vector_t *vecp, dev_info_t *dip, int type, 303 int inum, int count, uchar_t vector, int target_apic_id) 304 { 305 uint64_t msi_addr, msi_data; 306 ushort_t msi_ctrl; 307 int i, cap_ptr = i_ddi_get_msi_msix_cap_ptr(dip); 308 ddi_acc_handle_t handle = i_ddi_get_pci_config_handle(dip); 309 msi_regs_t msi_regs; 310 void *intrmap_tbl[PCI_MSI_MAX_INTRS]; 311 312 DDI_INTR_IMPLDBG((CE_CONT, "apix_pci_msi_enable_vector: dip=0x%p\n" 313 "\tdriver = %s, inum=0x%x vector=0x%x apicid=0x%x\n", (void *)dip, 314 ddi_driver_name(dip), inum, vector, target_apic_id)); 315 316 ASSERT((handle != NULL) && (cap_ptr != 0)); 317 318 msi_regs.mr_data = vector; 319 msi_regs.mr_addr = target_apic_id; 320 321 for (i = 0; i < count; i++) 322 intrmap_tbl[i] = xv_intrmap_private(vecp->v_cpuid, vector + i); 323 apic_vt_ops->apic_intrmap_alloc_entry(intrmap_tbl, dip, type, 324 count, 0xff); 325 for (i = 0; i < count; i++) 326 xv_intrmap_private(vecp->v_cpuid, vector + i) = intrmap_tbl[i]; 327 328 apic_vt_ops->apic_intrmap_map_entry(vecp->v_intrmap_private, 329 (void *)&msi_regs, type, count); 330 apic_vt_ops->apic_intrmap_record_msi(vecp->v_intrmap_private, 331 &msi_regs); 332 333 /* MSI Address */ 334 msi_addr = msi_regs.mr_addr; 335 336 /* MSI Data: MSI is edge triggered according to spec */ 337 msi_data = msi_regs.mr_data; 338 339 DDI_INTR_IMPLDBG((CE_CONT, "apix_pci_msi_enable_vector: addr=0x%lx " 340 "data=0x%lx\n", (long)msi_addr, (long)msi_data)); 341 342 if (type == APIX_TYPE_MSI) { 343 msi_ctrl = pci_config_get16(handle, cap_ptr + PCI_MSI_CTRL); 344 345 /* Set the bits to inform how many MSIs are enabled */ 346 msi_ctrl |= ((highbit(count) - 1) << PCI_MSI_MME_SHIFT); 347 pci_config_put16(handle, cap_ptr + PCI_MSI_CTRL, msi_ctrl); 348 349 if ((vecp->v_flags & APIX_VECT_MASKABLE) == 0) 350 APIX_WRITE_MSI_DATA(handle, cap_ptr, msi_ctrl, 351 APIX_RESV_VECTOR); 352 353 pci_config_put32(handle, 354 cap_ptr + PCI_MSI_ADDR_OFFSET, msi_addr); 355 if (msi_ctrl & PCI_MSI_64BIT_MASK) 356 pci_config_put32(handle, 357 cap_ptr + PCI_MSI_ADDR_OFFSET + 4, msi_addr >> 32); 358 359 APIX_WRITE_MSI_DATA(handle, cap_ptr, msi_ctrl, msi_data); 360 } else if (type == APIX_TYPE_MSIX) { 361 uintptr_t off; 362 ddi_intr_msix_t *msix_p = i_ddi_get_msix(dip); 363 364 /* Offset into the "inum"th entry in the MSI-X table */ 365 off = (uintptr_t)msix_p->msix_tbl_addr + 366 (inum * PCI_MSIX_VECTOR_SIZE); 367 368 ddi_put32(msix_p->msix_tbl_hdl, 369 (uint32_t *)(off + PCI_MSIX_DATA_OFFSET), msi_data); 370 ddi_put32(msix_p->msix_tbl_hdl, 371 (uint32_t *)(off + PCI_MSIX_LOWER_ADDR_OFFSET), msi_addr); 372 ddi_put32(msix_p->msix_tbl_hdl, 373 (uint32_t *)(off + PCI_MSIX_UPPER_ADDR_OFFSET), 374 msi_addr >> 32); 375 } 376 } 377 378 static void 379 apix_pci_msi_enable_mode(dev_info_t *dip, int type, int inum) 380 { 381 ushort_t msi_ctrl; 382 int cap_ptr = i_ddi_get_msi_msix_cap_ptr(dip); 383 ddi_acc_handle_t handle = i_ddi_get_pci_config_handle(dip); 384 385 ASSERT((handle != NULL) && (cap_ptr != 0)); 386 387 if (type == APIX_TYPE_MSI) { 388 msi_ctrl = pci_config_get16(handle, cap_ptr + PCI_MSI_CTRL); 389 if ((msi_ctrl & PCI_MSI_ENABLE_BIT)) 390 return; 391 392 msi_ctrl |= PCI_MSI_ENABLE_BIT; 393 pci_config_put16(handle, cap_ptr + PCI_MSI_CTRL, msi_ctrl); 394 395 } else if (type == DDI_INTR_TYPE_MSIX) { 396 uintptr_t off; 397 uint32_t mask; 398 ddi_intr_msix_t *msix_p; 399 400 msix_p = i_ddi_get_msix(dip); 401 402 /* Offset into "inum"th entry in the MSI-X table & clear mask */ 403 off = (uintptr_t)msix_p->msix_tbl_addr + (inum * 404 PCI_MSIX_VECTOR_SIZE) + PCI_MSIX_VECTOR_CTRL_OFFSET; 405 406 mask = ddi_get32(msix_p->msix_tbl_hdl, (uint32_t *)off); 407 408 ddi_put32(msix_p->msix_tbl_hdl, (uint32_t *)off, (mask & ~1)); 409 410 msi_ctrl = pci_config_get16(handle, cap_ptr + PCI_MSIX_CTRL); 411 412 if (!(msi_ctrl & PCI_MSIX_ENABLE_BIT)) { 413 msi_ctrl |= PCI_MSIX_ENABLE_BIT; 414 pci_config_put16(handle, cap_ptr + PCI_MSIX_CTRL, 415 msi_ctrl); 416 } 417 } 418 } 419 420 /* 421 * Setup interrupt, pogramming IO-APIC or MSI/X address/data. 422 */ 423 void 424 apix_enable_vector(apix_vector_t *vecp) 425 { 426 int tocpu = vecp->v_cpuid, type = vecp->v_type; 427 apic_cpus_info_t *cpu_infop; 428 ulong_t iflag; 429 430 ASSERT(tocpu < apic_nproc); 431 432 cpu_infop = &apic_cpus[tocpu]; 433 if (vecp->v_flags & APIX_VECT_USER_BOUND) 434 cpu_infop->aci_bound++; 435 else 436 cpu_infop->aci_temp_bound++; 437 438 iflag = intr_clear(); 439 lock_set(&apic_ioapic_lock); 440 441 if (!DDI_INTR_IS_MSI_OR_MSIX(type)) { /* fixed */ 442 apix_intx_enable(vecp->v_inum); 443 } else { 444 int inum = vecp->v_inum; 445 dev_info_t *dip = APIX_GET_DIP(vecp); 446 int count = i_ddi_intr_get_current_nintrs(dip); 447 448 if (type == APIX_TYPE_MSI) { /* MSI */ 449 if (inum == apix_get_max_dev_inum(dip, type)) { 450 /* last one */ 451 uchar_t start_inum = inum + 1 - count; 452 uchar_t start_vect = vecp->v_vector + 1 - count; 453 apix_vector_t *start_vecp = 454 xv_vector(vecp->v_cpuid, start_vect); 455 456 APIC_VERBOSE(INTR, (CE_CONT, "apix: call " 457 "apix_pci_msi_enable_vector\n")); 458 apix_pci_msi_enable_vector(start_vecp, dip, 459 type, start_inum, count, start_vect, 460 cpu_infop->aci_local_id); 461 462 APIC_VERBOSE(INTR, (CE_CONT, "apix: call " 463 "apix_pci_msi_enable_mode\n")); 464 apix_pci_msi_enable_mode(dip, type, inum); 465 } 466 } else { /* MSI-X */ 467 apix_pci_msi_enable_vector(vecp, dip, 468 type, inum, 1, vecp->v_vector, 469 cpu_infop->aci_local_id); 470 apix_pci_msi_enable_mode(dip, type, inum); 471 } 472 } 473 vecp->v_state = APIX_STATE_ENABLED; 474 apic_redist_cpu_skip &= ~(1 << tocpu); 475 476 lock_clear(&apic_ioapic_lock); 477 intr_restore(iflag); 478 } 479 480 /* 481 * Disable the interrupt 482 */ 483 void 484 apix_disable_vector(apix_vector_t *vecp) 485 { 486 struct autovec *avp = vecp->v_autovect; 487 ulong_t iflag; 488 489 ASSERT(avp != NULL); 490 491 iflag = intr_clear(); 492 lock_set(&apic_ioapic_lock); 493 494 switch (vecp->v_type) { 495 case APIX_TYPE_MSI: 496 ASSERT(avp->av_vector != NULL && avp->av_dip != NULL); 497 /* 498 * Disable the MSI vector 499 * Make sure we only disable on the last 500 * of the multi-MSI support 501 */ 502 if (i_ddi_intr_get_current_nenables(avp->av_dip) == 1) { 503 apic_pci_msi_disable_mode(avp->av_dip, 504 DDI_INTR_TYPE_MSI); 505 } 506 break; 507 case APIX_TYPE_MSIX: 508 ASSERT(avp->av_vector != NULL && avp->av_dip != NULL); 509 /* 510 * Disable the MSI-X vector 511 * needs to clear its mask and addr/data for each MSI-X 512 */ 513 apic_pci_msi_unconfigure(avp->av_dip, DDI_INTR_TYPE_MSIX, 514 vecp->v_inum); 515 /* 516 * Make sure we only disable on the last MSI-X 517 */ 518 if (i_ddi_intr_get_current_nenables(avp->av_dip) == 1) { 519 apic_pci_msi_disable_mode(avp->av_dip, 520 DDI_INTR_TYPE_MSIX); 521 } 522 break; 523 default: 524 apix_intx_disable(vecp->v_inum); 525 break; 526 } 527 528 if (!(apic_cpus[vecp->v_cpuid].aci_status & APIC_CPU_SUSPEND)) 529 vecp->v_state = APIX_STATE_DISABLED; 530 apic_vt_ops->apic_intrmap_free_entry(&vecp->v_intrmap_private); 531 vecp->v_intrmap_private = NULL; 532 533 lock_clear(&apic_ioapic_lock); 534 intr_restore(iflag); 535 } 536 537 /* 538 * Mark vector as obsoleted or freed. The vector is marked 539 * obsoleted if there are pending requests on it. Otherwise, 540 * free the vector. The obsoleted vectors get freed after 541 * being serviced. 542 * 543 * Return 1 on being obosoleted and 0 on being freed. 544 */ 545 #define INTR_BUSY(_avp)\ 546 ((((volatile ushort_t)(_avp)->av_flags) &\ 547 (AV_PENTRY_PEND | AV_PENTRY_ONPROC)) != 0) 548 #define LOCAL_WITH_INTR_DISABLED(_cpuid)\ 549 ((_cpuid) == psm_get_cpu_id() && !interrupts_enabled()) 550 static uint64_t dummy_tick; 551 552 int 553 apix_obsolete_vector(apix_vector_t *vecp) 554 { 555 struct autovec *avp = vecp->v_autovect; 556 int repeats, tries, ipl, busy = 0, cpuid = vecp->v_cpuid; 557 apix_impl_t *apixp = apixs[cpuid]; 558 559 ASSERT(APIX_CPU_LOCK_HELD(cpuid)); 560 561 for (avp = vecp->v_autovect; avp != NULL; avp = avp->av_link) { 562 if (avp->av_vector == NULL) 563 continue; 564 565 if (LOCAL_WITH_INTR_DISABLED(cpuid)) { 566 int bit, index, irr; 567 568 if (INTR_BUSY(avp)) { 569 busy++; 570 continue; 571 } 572 573 /* check IRR for pending interrupts */ 574 index = vecp->v_vector / 32; 575 bit = vecp->v_vector % 32; 576 irr = apic_reg_ops->apic_read(APIC_IRR_REG + index); 577 if ((irr & (1 << bit)) != 0) 578 busy++; 579 580 if (!busy) 581 apix_remove_av(vecp, avp); 582 583 continue; 584 } 585 586 repeats = 0; 587 do { 588 repeats++; 589 for (tries = 0; tries < apic_max_reps_clear_pending; 590 tries++) 591 if (!INTR_BUSY(avp)) 592 break; 593 } while (INTR_BUSY(avp) && 594 (repeats < apic_max_reps_clear_pending)); 595 596 if (INTR_BUSY(avp)) 597 busy++; 598 else { 599 /* 600 * Interrupt is not in pending list or being serviced. 601 * However it might be cached in Local APIC's IRR 602 * register. It's impossible to check another CPU's 603 * IRR register. Then wait till lower levels finish 604 * running. 605 */ 606 for (ipl = 1; ipl < MIN(LOCK_LEVEL, vecp->v_pri); ipl++) 607 apix_wait_till_seen(cpuid, ipl); 608 if (INTR_BUSY(avp)) 609 busy++; 610 } 611 612 if (!busy) 613 apix_remove_av(vecp, avp); 614 } 615 616 if (busy) { 617 apix_vector_t *tp = apixp->x_obsoletes; 618 619 if (vecp->v_state == APIX_STATE_OBSOLETED) 620 return (1); 621 622 vecp->v_state = APIX_STATE_OBSOLETED; 623 vecp->v_next = NULL; 624 if (tp == NULL) 625 apixp->x_obsoletes = vecp; 626 else { 627 while (tp->v_next != NULL) 628 tp = tp->v_next; 629 tp->v_next = vecp; 630 } 631 return (1); 632 } 633 634 /* interrupt is not busy */ 635 if (vecp->v_state == APIX_STATE_OBSOLETED) { 636 /* remove from obsoleted list */ 637 apixp->x_obsoletes = vecp->v_next; 638 vecp->v_next = NULL; 639 } 640 apix_cleanup_vector(vecp); 641 return (0); 642 } 643 644 /* 645 * Duplicate number of continuous vectors to specified target vectors. 646 */ 647 static void 648 apix_dup_vectors(apix_vector_t *oldp, apix_vector_t *newp, int count) 649 { 650 struct autovec *avp; 651 apix_vector_t *fromp, *top; 652 processorid_t oldcpu = oldp->v_cpuid, newcpu = newp->v_cpuid; 653 uchar_t oldvec = oldp->v_vector, newvec = newp->v_vector; 654 int i, inum; 655 656 ASSERT(oldp->v_type != APIX_TYPE_IPI); 657 658 for (i = 0; i < count; i++) { 659 fromp = xv_vector(oldcpu, oldvec + i); 660 top = xv_vector(newcpu, newvec + i); 661 ASSERT(fromp != NULL && top != NULL); 662 663 /* copy over original one */ 664 top->v_state = fromp->v_state; 665 top->v_type = fromp->v_type; 666 top->v_bound_cpuid = fromp->v_bound_cpuid; 667 top->v_inum = fromp->v_inum; 668 top->v_flags = fromp->v_flags; 669 top->v_intrmap_private = fromp->v_intrmap_private; 670 671 for (avp = fromp->v_autovect; avp != NULL; avp = avp->av_link) { 672 if (avp->av_vector == NULL) 673 continue; 674 675 apix_insert_av(top, avp->av_intr_id, avp->av_vector, 676 avp->av_intarg1, avp->av_intarg2, avp->av_ticksp, 677 avp->av_prilevel, avp->av_dip); 678 679 if (fromp->v_type == APIX_TYPE_FIXED && 680 avp->av_dip != NULL) { 681 inum = GET_INTR_INUM(avp->av_intr_id); 682 apix_set_dev_map(top, avp->av_dip, inum); 683 } 684 } 685 686 if (DDI_INTR_IS_MSI_OR_MSIX(fromp->v_type) && 687 fromp->v_devp != NULL) 688 apix_set_dev_map(top, fromp->v_devp->dv_dip, 689 fromp->v_devp->dv_inum); 690 } 691 } 692 693 static apix_vector_t * 694 apix_init_vector(processorid_t cpuid, uchar_t vector) 695 { 696 apix_impl_t *apixp = apixs[cpuid]; 697 apix_vector_t *vecp = apixp->x_vectbl[vector]; 698 699 ASSERT(IS_VECT_FREE(vecp)); 700 701 if (vecp == NULL) { 702 vecp = kmem_zalloc(sizeof (apix_vector_t), KM_NOSLEEP); 703 if (vecp == NULL) { 704 cmn_err(CE_WARN, "apix: no memory to allocate vector"); 705 return (NULL); 706 } 707 apixp->x_vectbl[vector] = vecp; 708 } 709 vecp->v_state = APIX_STATE_ALLOCED; 710 vecp->v_cpuid = vecp->v_bound_cpuid = cpuid; 711 vecp->v_vector = vector; 712 713 return (vecp); 714 } 715 716 static void 717 apix_cleanup_vector(apix_vector_t *vecp) 718 { 719 ASSERT(vecp->v_share == 0); 720 vecp->v_bound_cpuid = IRQ_UNINIT; 721 vecp->v_state = APIX_STATE_FREED; 722 vecp->v_type = 0; 723 vecp->v_flags = 0; 724 vecp->v_busy = 0; 725 vecp->v_intrmap_private = NULL; 726 } 727 728 static void 729 apix_dprint_vector(apix_vector_t *vecp, dev_info_t *dip, int count) 730 { 731 #ifdef DEBUG 732 major_t major; 733 char *name, *drv_name; 734 int instance, len, t_len; 735 char mesg[1024] = "apix: "; 736 737 t_len = sizeof (mesg); 738 len = strlen(mesg); 739 if (dip != NULL) { 740 name = ddi_get_name(dip); 741 major = ddi_name_to_major(name); 742 drv_name = ddi_major_to_name(major); 743 instance = ddi_get_instance(dip); 744 (void) snprintf(mesg + len, t_len - len, "%s (%s) instance %d ", 745 name, drv_name, instance); 746 } 747 len = strlen(mesg); 748 749 switch (vecp->v_type) { 750 case APIX_TYPE_FIXED: 751 (void) snprintf(mesg + len, t_len - len, "irqno %d", 752 vecp->v_inum); 753 break; 754 case APIX_TYPE_MSI: 755 (void) snprintf(mesg + len, t_len - len, 756 "msi inum %d (count %d)", vecp->v_inum, count); 757 break; 758 case APIX_TYPE_MSIX: 759 (void) snprintf(mesg + len, t_len - len, "msi-x inum %d", 760 vecp->v_inum); 761 break; 762 default: 763 break; 764 765 } 766 767 APIC_VERBOSE(ALLOC, (CE_CONT, "%s allocated with vector 0x%x on " 768 "cpu %d\n", mesg, vecp->v_vector, vecp->v_cpuid)); 769 #endif /* DEBUG */ 770 } 771 772 /* 773 * Operations on avintr 774 */ 775 776 #define INIT_AUTOVEC(p, intr_id, f, arg1, arg2, ticksp, ipl, dip) \ 777 do { \ 778 (p)->av_intr_id = intr_id; \ 779 (p)->av_vector = f; \ 780 (p)->av_intarg1 = arg1; \ 781 (p)->av_intarg2 = arg2; \ 782 (p)->av_ticksp = ticksp; \ 783 (p)->av_prilevel = ipl; \ 784 (p)->av_dip = dip; \ 785 (p)->av_flags = 0; \ 786 _NOTE(CONSTCOND)} while (0) 787 788 /* 789 * Insert an interrupt service routine into chain by its priority from 790 * high to low 791 */ 792 static void 793 apix_insert_av(apix_vector_t *vecp, void *intr_id, avfunc f, caddr_t arg1, 794 caddr_t arg2, uint64_t *ticksp, int ipl, dev_info_t *dip) 795 { 796 struct autovec *p, *prep, *mem; 797 798 APIC_VERBOSE(INTR, (CE_CONT, "apix_insert_av: dip %p, vector 0x%x, " 799 "cpu %d\n", (void *)dip, vecp->v_vector, vecp->v_cpuid)); 800 801 mem = kmem_zalloc(sizeof (struct autovec), KM_SLEEP); 802 INIT_AUTOVEC(mem, intr_id, f, arg1, arg2, ticksp, ipl, dip); 803 if (vecp->v_type == APIX_TYPE_FIXED && apic_level_intr[vecp->v_inum]) 804 mem->av_flags |= AV_PENTRY_LEVEL; 805 806 vecp->v_share++; 807 vecp->v_pri = (ipl > vecp->v_pri) ? ipl : vecp->v_pri; 808 809 smt_intr_alloc_pil(vecp->v_pri); 810 811 if (vecp->v_autovect == NULL) { /* Nothing on list - put it at head */ 812 vecp->v_autovect = mem; 813 return; 814 } 815 816 if (DDI_INTR_IS_MSI_OR_MSIX(vecp->v_type)) { /* MSI/X */ 817 ASSERT(vecp->v_share == 1); /* No sharing for MSI/X */ 818 819 INIT_AUTOVEC(vecp->v_autovect, intr_id, f, arg1, arg2, ticksp, 820 ipl, dip); 821 prep = vecp->v_autovect->av_link; 822 vecp->v_autovect->av_link = NULL; 823 824 /* Free the following autovect chain */ 825 while (prep != NULL) { 826 ASSERT(prep->av_vector == NULL); 827 828 p = prep; 829 prep = prep->av_link; 830 kmem_free(p, sizeof (struct autovec)); 831 } 832 833 kmem_free(mem, sizeof (struct autovec)); 834 return; 835 } 836 837 /* find where it goes in list */ 838 prep = NULL; 839 for (p = vecp->v_autovect; p != NULL; p = p->av_link) { 840 if (p->av_vector && p->av_prilevel <= ipl) 841 break; 842 prep = p; 843 } 844 if (prep != NULL) { 845 if (prep->av_vector == NULL) { /* freed struct available */ 846 INIT_AUTOVEC(prep, intr_id, f, arg1, arg2, 847 ticksp, ipl, dip); 848 prep->av_flags = mem->av_flags; 849 kmem_free(mem, sizeof (struct autovec)); 850 return; 851 } 852 853 mem->av_link = prep->av_link; 854 prep->av_link = mem; 855 } else { 856 /* insert new intpt at beginning of chain */ 857 mem->av_link = vecp->v_autovect; 858 vecp->v_autovect = mem; 859 } 860 } 861 862 /* 863 * After having made a change to an autovector list, wait until we have 864 * seen specified cpu not executing an interrupt at that level--so we 865 * know our change has taken effect completely (no old state in registers, 866 * etc). 867 */ 868 #define APIX_CPU_ENABLED(_cp) \ 869 (quiesce_active == 0 && \ 870 (((_cp)->cpu_flags & (CPU_QUIESCED|CPU_OFFLINE)) == 0)) 871 872 static void 873 apix_wait_till_seen(processorid_t cpuid, int ipl) 874 { 875 struct cpu *cp = cpu[cpuid]; 876 877 if (cp == NULL || LOCAL_WITH_INTR_DISABLED(cpuid)) 878 return; 879 880 /* 881 * Don't wait if the CPU is quiesced or offlined. This can happen 882 * when a CPU is running pause thread but hardware triggered an 883 * interrupt and the interrupt gets queued. 884 */ 885 for (;;) { 886 if (!INTR_ACTIVE((volatile struct cpu *)cpu[cpuid], ipl) && 887 (!APIX_CPU_ENABLED(cp) || 888 !INTR_PENDING((volatile apix_impl_t *)apixs[cpuid], ipl))) 889 return; 890 } 891 } 892 893 static void 894 apix_remove_av(apix_vector_t *vecp, struct autovec *target) 895 { 896 int hi_pri = 0; 897 struct autovec *p; 898 899 if (target == NULL) 900 return; 901 902 APIC_VERBOSE(INTR, (CE_CONT, "apix_remove_av: dip %p, vector 0x%x, " 903 "cpu %d\n", (void *)target->av_dip, vecp->v_vector, vecp->v_cpuid)); 904 905 for (p = vecp->v_autovect; p; p = p->av_link) { 906 if (p == target || p->av_vector == NULL) 907 continue; 908 hi_pri = (p->av_prilevel > hi_pri) ? p->av_prilevel : hi_pri; 909 } 910 911 vecp->v_share--; 912 vecp->v_pri = hi_pri; 913 914 /* 915 * This drops the handler from the chain, it can no longer be called. 916 * However, there is no guarantee that the handler is not currently 917 * still executing. 918 */ 919 target->av_vector = NULL; 920 /* 921 * There is a race where we could be just about to pick up the ticksp 922 * pointer to increment it after returning from the service routine 923 * in av_dispatch_autovect. Rather than NULL it out let's just point 924 * it off to something safe so that any final tick update attempt 925 * won't fault. 926 */ 927 target->av_ticksp = &dummy_tick; 928 apix_wait_till_seen(vecp->v_cpuid, target->av_prilevel); 929 } 930 931 static struct autovec * 932 apix_find_av(apix_vector_t *vecp, void *intr_id, avfunc f) 933 { 934 struct autovec *p; 935 936 for (p = vecp->v_autovect; p; p = p->av_link) { 937 if ((p->av_vector == f) && (p->av_intr_id == intr_id)) { 938 /* found the handler */ 939 return (p); 940 } 941 } 942 943 return (NULL); 944 } 945 946 static apix_vector_t * 947 apix_find_vector_by_avintr(void *intr_id, avfunc f) 948 { 949 apix_vector_t *vecp; 950 processorid_t n; 951 uchar_t v; 952 953 for (n = 0; n < apic_nproc; n++) { 954 if (!apix_is_cpu_enabled(n)) 955 continue; 956 957 for (v = APIX_AVINTR_MIN; v <= APIX_AVINTR_MIN; v++) { 958 vecp = xv_vector(n, v); 959 if (vecp == NULL || 960 vecp->v_state <= APIX_STATE_OBSOLETED) 961 continue; 962 963 if (apix_find_av(vecp, intr_id, f) != NULL) 964 return (vecp); 965 } 966 } 967 968 return (NULL); 969 } 970 971 /* 972 * Add interrupt service routine. 973 * 974 * For legacy interrupts (HPET timer, ACPI SCI), the vector is actually 975 * IRQ no. A vector is then allocated. Otherwise, the vector is already 976 * allocated. The input argument virt_vect is virtual vector of format 977 * APIX_VIRTVEC_VECTOR(cpuid, vector). 978 * 979 * Return 1 on success, 0 on failure. 980 */ 981 int 982 apix_add_avintr(void *intr_id, int ipl, avfunc xxintr, char *name, 983 int virt_vect, caddr_t arg1, caddr_t arg2, uint64_t *ticksp, 984 dev_info_t *dip) 985 { 986 int cpuid; 987 uchar_t v = (uchar_t)APIX_VIRTVEC_VECTOR(virt_vect); 988 apix_vector_t *vecp; 989 990 if (xxintr == NULL) { 991 cmn_err(CE_WARN, "Attempt to add null for %s " 992 "on vector 0x%x,0x%x", name, 993 APIX_VIRTVEC_CPU(virt_vect), 994 APIX_VIRTVEC_VECTOR(virt_vect)); 995 return (0); 996 } 997 998 if (v >= APIX_IPI_MIN) /* IPIs */ 999 return (apix_add_ipi(ipl, xxintr, name, v, arg1, arg2)); 1000 1001 if (!APIX_IS_VIRTVEC(virt_vect)) { /* got irq */ 1002 int irqno = virt_vect; 1003 int inum = GET_INTR_INUM(intr_id); 1004 1005 /* 1006 * Senarios include: 1007 * a. add_avintr() is called before irqp initialized (legacy) 1008 * b. irqp is initialized, vector is not allocated (fixed) 1009 * c. irqp is initialized, vector is allocated (fixed & shared) 1010 */ 1011 if ((vecp = apix_alloc_intx(dip, inum, irqno)) == NULL) 1012 return (0); 1013 1014 cpuid = vecp->v_cpuid; 1015 v = vecp->v_vector; 1016 virt_vect = APIX_VIRTVECTOR(cpuid, v); 1017 } else { /* got virtual vector */ 1018 cpuid = APIX_VIRTVEC_CPU(virt_vect); 1019 vecp = xv_vector(cpuid, v); 1020 ASSERT(vecp != NULL); 1021 } 1022 1023 lock_set(&apix_lock); 1024 if (vecp->v_state <= APIX_STATE_OBSOLETED) { 1025 vecp = NULL; 1026 1027 /* 1028 * Basically the allocated but not enabled interrupts 1029 * will not get re-targeted. But MSIs in allocated state 1030 * could be re-targeted due to group re-targeting. 1031 */ 1032 if (intr_id != NULL && dip != NULL) { 1033 ddi_intr_handle_impl_t *hdlp = intr_id; 1034 vecp = apix_get_dev_map(dip, hdlp->ih_inum, 1035 hdlp->ih_type); 1036 ASSERT(vecp->v_state == APIX_STATE_ALLOCED); 1037 } 1038 if (vecp == NULL) { 1039 lock_clear(&apix_lock); 1040 cmn_err(CE_WARN, "Invalid interrupt 0x%x,0x%x " 1041 " for %p to add", cpuid, v, intr_id); 1042 return (0); 1043 } 1044 cpuid = vecp->v_cpuid; 1045 virt_vect = APIX_VIRTVECTOR(cpuid, vecp->v_vector); 1046 } 1047 1048 APIX_ENTER_CPU_LOCK(cpuid); 1049 apix_insert_av(vecp, intr_id, xxintr, arg1, arg2, ticksp, ipl, dip); 1050 APIX_LEAVE_CPU_LOCK(cpuid); 1051 1052 (void) apix_addspl(virt_vect, ipl, 0, 0); 1053 1054 lock_clear(&apix_lock); 1055 1056 return (1); 1057 } 1058 1059 /* 1060 * Remove avintr 1061 * 1062 * For fixed, if it's the last one of shared interrupts, free the vector. 1063 * For msi/x, only disable the interrupt but not free the vector, which 1064 * is freed by PSM_XXX_FREE_XXX. 1065 */ 1066 void 1067 apix_rem_avintr(void *intr_id, int ipl, avfunc xxintr, int virt_vect) 1068 { 1069 avfunc f; 1070 apix_vector_t *vecp; 1071 struct autovec *avp; 1072 processorid_t cpuid; 1073 1074 if ((f = xxintr) == NULL) 1075 return; 1076 1077 lock_set(&apix_lock); 1078 1079 if (!APIX_IS_VIRTVEC(virt_vect)) { /* got irq */ 1080 vecp = apix_intx_get_vector(virt_vect); 1081 virt_vect = APIX_VIRTVECTOR(vecp->v_cpuid, vecp->v_vector); 1082 } else /* got virtual vector */ 1083 vecp = xv_vector(APIX_VIRTVEC_CPU(virt_vect), 1084 APIX_VIRTVEC_VECTOR(virt_vect)); 1085 1086 if (vecp == NULL) { 1087 lock_clear(&apix_lock); 1088 cmn_err(CE_CONT, "Invalid interrupt 0x%x,0x%x to remove", 1089 APIX_VIRTVEC_CPU(virt_vect), 1090 APIX_VIRTVEC_VECTOR(virt_vect)); 1091 return; 1092 } 1093 1094 if (vecp->v_state <= APIX_STATE_OBSOLETED || 1095 ((avp = apix_find_av(vecp, intr_id, f)) == NULL)) { 1096 /* 1097 * It's possible that the interrupt is rebound to a 1098 * different cpu before rem_avintr() is called. Search 1099 * through all vectors once it happens. 1100 */ 1101 if ((vecp = apix_find_vector_by_avintr(intr_id, f)) 1102 == NULL) { 1103 lock_clear(&apix_lock); 1104 cmn_err(CE_CONT, "Unknown interrupt 0x%x,0x%x " 1105 "for %p to remove", APIX_VIRTVEC_CPU(virt_vect), 1106 APIX_VIRTVEC_VECTOR(virt_vect), intr_id); 1107 return; 1108 } 1109 virt_vect = APIX_VIRTVECTOR(vecp->v_cpuid, vecp->v_vector); 1110 avp = apix_find_av(vecp, intr_id, f); 1111 } 1112 cpuid = vecp->v_cpuid; 1113 1114 /* disable interrupt */ 1115 (void) apix_delspl(virt_vect, ipl, 0, 0); 1116 1117 /* remove ISR entry */ 1118 APIX_ENTER_CPU_LOCK(cpuid); 1119 apix_remove_av(vecp, avp); 1120 APIX_LEAVE_CPU_LOCK(cpuid); 1121 1122 lock_clear(&apix_lock); 1123 } 1124 1125 /* 1126 * Device to vector mapping table 1127 */ 1128 1129 static void 1130 apix_clear_dev_map(dev_info_t *dip, int inum, int type) 1131 { 1132 char *name; 1133 major_t major; 1134 apix_dev_vector_t *dvp, *prev = NULL; 1135 int found = 0; 1136 1137 name = ddi_get_name(dip); 1138 major = ddi_name_to_major(name); 1139 1140 mutex_enter(&apix_mutex); 1141 1142 for (dvp = apix_dev_vector[major]; dvp != NULL; 1143 prev = dvp, dvp = dvp->dv_next) { 1144 if (dvp->dv_dip == dip && dvp->dv_inum == inum && 1145 dvp->dv_type == type) { 1146 found++; 1147 break; 1148 } 1149 } 1150 1151 if (!found) { 1152 mutex_exit(&apix_mutex); 1153 return; 1154 } 1155 1156 if (prev != NULL) 1157 prev->dv_next = dvp->dv_next; 1158 1159 if (apix_dev_vector[major] == dvp) 1160 apix_dev_vector[major] = dvp->dv_next; 1161 1162 dvp->dv_vector->v_devp = NULL; 1163 1164 mutex_exit(&apix_mutex); 1165 1166 kmem_free(dvp, sizeof (apix_dev_vector_t)); 1167 } 1168 1169 void 1170 apix_set_dev_map(apix_vector_t *vecp, dev_info_t *dip, int inum) 1171 { 1172 apix_dev_vector_t *dvp; 1173 char *name; 1174 major_t major; 1175 uint32_t found = 0; 1176 1177 ASSERT(dip != NULL); 1178 name = ddi_get_name(dip); 1179 major = ddi_name_to_major(name); 1180 1181 mutex_enter(&apix_mutex); 1182 1183 for (dvp = apix_dev_vector[major]; dvp != NULL; 1184 dvp = dvp->dv_next) { 1185 if (dvp->dv_dip == dip && dvp->dv_inum == inum && 1186 dvp->dv_type == vecp->v_type) { 1187 found++; 1188 break; 1189 } 1190 } 1191 1192 if (found == 0) { /* not found */ 1193 dvp = kmem_zalloc(sizeof (apix_dev_vector_t), KM_SLEEP); 1194 dvp->dv_dip = dip; 1195 dvp->dv_inum = inum; 1196 dvp->dv_type = vecp->v_type; 1197 1198 dvp->dv_next = apix_dev_vector[major]; 1199 apix_dev_vector[major] = dvp; 1200 } 1201 dvp->dv_vector = vecp; 1202 vecp->v_devp = dvp; 1203 1204 mutex_exit(&apix_mutex); 1205 1206 DDI_INTR_IMPLDBG((CE_CONT, "apix_set_dev_map: dip=0x%p " 1207 "inum=0x%x vector=0x%x/0x%x\n", 1208 (void *)dip, inum, vecp->v_cpuid, vecp->v_vector)); 1209 } 1210 1211 apix_vector_t * 1212 apix_get_dev_map(dev_info_t *dip, int inum, int type) 1213 { 1214 char *name; 1215 major_t major; 1216 apix_dev_vector_t *dvp; 1217 apix_vector_t *vecp; 1218 1219 name = ddi_get_name(dip); 1220 if ((major = ddi_name_to_major(name)) == DDI_MAJOR_T_NONE) 1221 return (NULL); 1222 1223 mutex_enter(&apix_mutex); 1224 for (dvp = apix_dev_vector[major]; dvp != NULL; 1225 dvp = dvp->dv_next) { 1226 if (dvp->dv_dip == dip && dvp->dv_inum == inum && 1227 dvp->dv_type == type) { 1228 vecp = dvp->dv_vector; 1229 mutex_exit(&apix_mutex); 1230 return (vecp); 1231 } 1232 } 1233 mutex_exit(&apix_mutex); 1234 1235 return (NULL); 1236 } 1237 1238 /* 1239 * Get minimum inum for specified device, used for MSI 1240 */ 1241 int 1242 apix_get_min_dev_inum(dev_info_t *dip, int type) 1243 { 1244 char *name; 1245 major_t major; 1246 apix_dev_vector_t *dvp; 1247 int inum = -1; 1248 1249 name = ddi_get_name(dip); 1250 major = ddi_name_to_major(name); 1251 1252 mutex_enter(&apix_mutex); 1253 for (dvp = apix_dev_vector[major]; dvp != NULL; 1254 dvp = dvp->dv_next) { 1255 if (dvp->dv_dip == dip && dvp->dv_type == type) { 1256 if (inum == -1) 1257 inum = dvp->dv_inum; 1258 else 1259 inum = (dvp->dv_inum < inum) ? 1260 dvp->dv_inum : inum; 1261 } 1262 } 1263 mutex_exit(&apix_mutex); 1264 1265 return (inum); 1266 } 1267 1268 int 1269 apix_get_max_dev_inum(dev_info_t *dip, int type) 1270 { 1271 char *name; 1272 major_t major; 1273 apix_dev_vector_t *dvp; 1274 int inum = -1; 1275 1276 name = ddi_get_name(dip); 1277 major = ddi_name_to_major(name); 1278 1279 mutex_enter(&apix_mutex); 1280 for (dvp = apix_dev_vector[major]; dvp != NULL; 1281 dvp = dvp->dv_next) { 1282 if (dvp->dv_dip == dip && dvp->dv_type == type) { 1283 if (inum == -1) 1284 inum = dvp->dv_inum; 1285 else 1286 inum = (dvp->dv_inum > inum) ? 1287 dvp->dv_inum : inum; 1288 } 1289 } 1290 mutex_exit(&apix_mutex); 1291 1292 return (inum); 1293 } 1294 1295 /* 1296 * Major to cpu binding, for INTR_ROUND_ROBIN_WITH_AFFINITY cpu 1297 * binding policy 1298 */ 1299 1300 static uint32_t 1301 apix_get_dev_binding(dev_info_t *dip) 1302 { 1303 major_t major; 1304 char *name; 1305 uint32_t cpu = IRQ_UNINIT; 1306 1307 name = ddi_get_name(dip); 1308 major = ddi_name_to_major(name); 1309 if (major < devcnt) { 1310 mutex_enter(&apix_mutex); 1311 cpu = apix_major_to_cpu[major]; 1312 mutex_exit(&apix_mutex); 1313 } 1314 1315 return (cpu); 1316 } 1317 1318 static void 1319 apix_set_dev_binding(dev_info_t *dip, uint32_t cpu) 1320 { 1321 major_t major; 1322 char *name; 1323 1324 /* setup major to cpu mapping */ 1325 name = ddi_get_name(dip); 1326 major = ddi_name_to_major(name); 1327 if (apix_major_to_cpu[major] == IRQ_UNINIT) { 1328 mutex_enter(&apix_mutex); 1329 apix_major_to_cpu[major] = cpu; 1330 mutex_exit(&apix_mutex); 1331 } 1332 } 1333 1334 /* 1335 * return the cpu to which this intr should be bound. 1336 * Check properties or any other mechanism to see if user wants it 1337 * bound to a specific CPU. If so, return the cpu id with high bit set. 1338 * If not, use the policy to choose a cpu and return the id. 1339 */ 1340 uint32_t 1341 apix_bind_cpu(dev_info_t *dip) 1342 { 1343 int instance, instno, prop_len, bind_cpu, count; 1344 uint_t i, rc; 1345 major_t major; 1346 char *name, *drv_name, *prop_val, *cptr; 1347 char prop_name[32]; 1348 1349 lock_set(&apix_lock); 1350 1351 if (apic_intr_policy == INTR_LOWEST_PRIORITY) { 1352 cmn_err(CE_WARN, "apix: unsupported interrupt binding policy " 1353 "LOWEST PRIORITY, use ROUND ROBIN instead"); 1354 apic_intr_policy = INTR_ROUND_ROBIN; 1355 } 1356 1357 if (apic_nproc == 1) { 1358 lock_clear(&apix_lock); 1359 return (0); 1360 } 1361 1362 drv_name = NULL; 1363 rc = DDI_PROP_NOT_FOUND; 1364 major = (major_t)-1; 1365 if (dip != NULL) { 1366 name = ddi_get_name(dip); 1367 major = ddi_name_to_major(name); 1368 drv_name = ddi_major_to_name(major); 1369 instance = ddi_get_instance(dip); 1370 if (apic_intr_policy == INTR_ROUND_ROBIN_WITH_AFFINITY) { 1371 bind_cpu = apix_get_dev_binding(dip); 1372 if (bind_cpu != IRQ_UNINIT) { 1373 lock_clear(&apix_lock); 1374 return (bind_cpu); 1375 } 1376 } 1377 /* 1378 * search for "drvname"_intpt_bind_cpus property first, the 1379 * syntax of the property should be "a[,b,c,...]" where 1380 * instance 0 binds to cpu a, instance 1 binds to cpu b, 1381 * instance 3 binds to cpu c... 1382 * ddi_getlongprop() will search /option first, then / 1383 * if "drvname"_intpt_bind_cpus doesn't exist, then find 1384 * intpt_bind_cpus property. The syntax is the same, and 1385 * it applies to all the devices if its "drvname" specific 1386 * property doesn't exist 1387 */ 1388 (void) strcpy(prop_name, drv_name); 1389 (void) strcat(prop_name, "_intpt_bind_cpus"); 1390 rc = ddi_getlongprop(DDI_DEV_T_ANY, dip, 0, prop_name, 1391 (caddr_t)&prop_val, &prop_len); 1392 if (rc != DDI_PROP_SUCCESS) { 1393 rc = ddi_getlongprop(DDI_DEV_T_ANY, dip, 0, 1394 "intpt_bind_cpus", (caddr_t)&prop_val, &prop_len); 1395 } 1396 } 1397 if (rc == DDI_PROP_SUCCESS) { 1398 for (i = count = 0; i < (prop_len - 1); i++) 1399 if (prop_val[i] == ',') 1400 count++; 1401 if (prop_val[i-1] != ',') 1402 count++; 1403 /* 1404 * if somehow the binding instances defined in the 1405 * property are not enough for this instno., then 1406 * reuse the pattern for the next instance until 1407 * it reaches the requested instno 1408 */ 1409 instno = instance % count; 1410 i = 0; 1411 cptr = prop_val; 1412 while (i < instno) 1413 if (*cptr++ == ',') 1414 i++; 1415 bind_cpu = stoi(&cptr); 1416 kmem_free(prop_val, prop_len); 1417 /* if specific cpu is bogus, then default to cpu 0 */ 1418 if (bind_cpu >= apic_nproc) { 1419 cmn_err(CE_WARN, "apix: %s=%s: CPU %d not present", 1420 prop_name, prop_val, bind_cpu); 1421 bind_cpu = 0; 1422 } else { 1423 /* indicate that we are bound at user request */ 1424 bind_cpu |= IRQ_USER_BOUND; 1425 } 1426 /* 1427 * no need to check apic_cpus[].aci_status, if specific cpu is 1428 * not up, then post_cpu_start will handle it. 1429 */ 1430 } else { 1431 bind_cpu = apic_get_next_bind_cpu(); 1432 } 1433 1434 lock_clear(&apix_lock); 1435 1436 return ((uint32_t)bind_cpu); 1437 } 1438 1439 static boolean_t 1440 apix_is_cpu_enabled(processorid_t cpuid) 1441 { 1442 apic_cpus_info_t *cpu_infop; 1443 1444 cpu_infop = &apic_cpus[cpuid]; 1445 1446 if ((cpu_infop->aci_status & APIC_CPU_INTR_ENABLE) == 0) 1447 return (B_FALSE); 1448 1449 return (B_TRUE); 1450 } 1451 1452 /* 1453 * Must be called with apix_lock held. This function can be 1454 * called from above lock level by apix_intr_redistribute(). 1455 * 1456 * Arguments: 1457 * vecp : Vector to be rebound 1458 * tocpu : Target cpu. IRQ_UNINIT means target is vecp->v_cpuid. 1459 * count : Number of continuous vectors 1460 * 1461 * Return new vector being bound to 1462 */ 1463 apix_vector_t * 1464 apix_rebind(apix_vector_t *vecp, processorid_t newcpu, int count) 1465 { 1466 apix_vector_t *newp, *oldp; 1467 processorid_t oldcpu = vecp->v_cpuid; 1468 uchar_t newvec, oldvec = vecp->v_vector; 1469 int i; 1470 1471 ASSERT(LOCK_HELD(&apix_lock) && count > 0); 1472 1473 if (!apix_is_cpu_enabled(newcpu)) 1474 return (NULL); 1475 1476 if (vecp->v_cpuid == newcpu) /* rebind to the same cpu */ 1477 return (vecp); 1478 1479 APIX_ENTER_CPU_LOCK(oldcpu); 1480 APIX_ENTER_CPU_LOCK(newcpu); 1481 1482 /* allocate vector */ 1483 if (count == 1) 1484 newp = apix_alloc_vector_oncpu(newcpu, NULL, 0, vecp->v_type); 1485 else { 1486 ASSERT(vecp->v_type == APIX_TYPE_MSI); 1487 newp = apix_alloc_nvectors_oncpu(newcpu, NULL, 0, count, 1488 vecp->v_type); 1489 } 1490 if (newp == NULL) { 1491 APIX_LEAVE_CPU_LOCK(newcpu); 1492 APIX_LEAVE_CPU_LOCK(oldcpu); 1493 return (NULL); 1494 } 1495 1496 newvec = newp->v_vector; 1497 apix_dup_vectors(vecp, newp, count); 1498 1499 APIX_LEAVE_CPU_LOCK(newcpu); 1500 APIX_LEAVE_CPU_LOCK(oldcpu); 1501 1502 if (!DDI_INTR_IS_MSI_OR_MSIX(vecp->v_type)) { 1503 ASSERT(count == 1); 1504 if (apix_intx_rebind(vecp->v_inum, newcpu, newvec) != 0) { 1505 struct autovec *avp; 1506 int inum; 1507 1508 /* undo duplication */ 1509 APIX_ENTER_CPU_LOCK(oldcpu); 1510 APIX_ENTER_CPU_LOCK(newcpu); 1511 for (avp = newp->v_autovect; avp != NULL; 1512 avp = avp->av_link) { 1513 if (avp->av_dip != NULL) { 1514 inum = GET_INTR_INUM(avp->av_intr_id); 1515 apix_set_dev_map(vecp, avp->av_dip, 1516 inum); 1517 } 1518 apix_remove_av(newp, avp); 1519 } 1520 apix_cleanup_vector(newp); 1521 APIX_LEAVE_CPU_LOCK(newcpu); 1522 APIX_LEAVE_CPU_LOCK(oldcpu); 1523 APIC_VERBOSE(REBIND, (CE_CONT, "apix: rebind fixed " 1524 "interrupt 0x%x to cpu %d failed\n", 1525 vecp->v_inum, newcpu)); 1526 return (NULL); 1527 } 1528 1529 APIX_ENTER_CPU_LOCK(oldcpu); 1530 (void) apix_obsolete_vector(vecp); 1531 APIX_LEAVE_CPU_LOCK(oldcpu); 1532 APIC_VERBOSE(REBIND, (CE_CONT, "apix: rebind fixed interrupt" 1533 " 0x%x/0x%x to 0x%x/0x%x\n", 1534 oldcpu, oldvec, newcpu, newvec)); 1535 return (newp); 1536 } 1537 1538 for (i = 0; i < count; i++) { 1539 oldp = xv_vector(oldcpu, oldvec + i); 1540 newp = xv_vector(newcpu, newvec + i); 1541 1542 if (newp->v_share > 0) { 1543 APIX_SET_REBIND_INFO(oldp, newp); 1544 1545 apix_enable_vector(newp); 1546 1547 APIX_CLR_REBIND_INFO(); 1548 } 1549 1550 APIX_ENTER_CPU_LOCK(oldcpu); 1551 (void) apix_obsolete_vector(oldp); 1552 APIX_LEAVE_CPU_LOCK(oldcpu); 1553 } 1554 APIC_VERBOSE(REBIND, (CE_CONT, "apix: rebind vector 0x%x/0x%x " 1555 "to 0x%x/0x%x, count=%d\n", 1556 oldcpu, oldvec, newcpu, newvec, count)); 1557 1558 return (xv_vector(newcpu, newvec)); 1559 } 1560 1561 /* 1562 * Senarios include: 1563 * a. add_avintr() is called before irqp initialized (legacy) 1564 * b. irqp is initialized, vector is not allocated (fixed interrupts) 1565 * c. irqp is initialized, vector is allocated (shared interrupts) 1566 */ 1567 apix_vector_t * 1568 apix_alloc_intx(dev_info_t *dip, int inum, int irqno) 1569 { 1570 apic_irq_t *irqp; 1571 apix_vector_t *vecp; 1572 1573 /* 1574 * Allocate IRQ. Caller is later responsible for the 1575 * initialization 1576 */ 1577 mutex_enter(&airq_mutex); 1578 if ((irqp = apic_irq_table[irqno]) == NULL) { 1579 /* allocate irq */ 1580 irqp = kmem_zalloc(sizeof (apic_irq_t), KM_SLEEP); 1581 irqp->airq_mps_intr_index = FREE_INDEX; 1582 apic_irq_table[irqno] = irqp; 1583 } 1584 if (irqp->airq_mps_intr_index == FREE_INDEX) { 1585 irqp->airq_mps_intr_index = DEFAULT_INDEX; 1586 irqp->airq_cpu = IRQ_UNINIT; 1587 irqp->airq_origirq = (uchar_t)irqno; 1588 } 1589 1590 mutex_exit(&airq_mutex); 1591 1592 /* 1593 * allocate vector 1594 */ 1595 if (irqp->airq_cpu == IRQ_UNINIT) { 1596 uint32_t bindcpu, cpuid; 1597 1598 /* select cpu by system policy */ 1599 bindcpu = apix_bind_cpu(dip); 1600 cpuid = bindcpu & ~IRQ_USER_BOUND; 1601 1602 /* allocate vector */ 1603 APIX_ENTER_CPU_LOCK(cpuid); 1604 1605 if ((vecp = apix_alloc_vector_oncpu(bindcpu, dip, inum, 1606 APIX_TYPE_FIXED)) == NULL) { 1607 cmn_err(CE_WARN, "No interrupt vector for irq %x", 1608 irqno); 1609 APIX_LEAVE_CPU_LOCK(cpuid); 1610 return (NULL); 1611 } 1612 vecp->v_inum = irqno; 1613 vecp->v_flags |= APIX_VECT_MASKABLE; 1614 1615 apix_intx_set_vector(irqno, vecp->v_cpuid, vecp->v_vector); 1616 1617 APIX_LEAVE_CPU_LOCK(cpuid); 1618 } else { 1619 vecp = xv_vector(irqp->airq_cpu, irqp->airq_vector); 1620 ASSERT(!IS_VECT_FREE(vecp)); 1621 1622 if (dip != NULL) 1623 apix_set_dev_map(vecp, dip, inum); 1624 } 1625 1626 if ((dip != NULL) && 1627 (apic_intr_policy == INTR_ROUND_ROBIN_WITH_AFFINITY) && 1628 ((vecp->v_flags & APIX_VECT_USER_BOUND) == 0)) 1629 apix_set_dev_binding(dip, vecp->v_cpuid); 1630 1631 apix_dprint_vector(vecp, dip, 1); 1632 1633 return (vecp); 1634 } 1635 1636 int 1637 apix_alloc_msi(dev_info_t *dip, int inum, int count, int behavior) 1638 { 1639 int i, cap_ptr, rcount = count; 1640 apix_vector_t *vecp; 1641 processorid_t bindcpu, cpuid; 1642 ushort_t msi_ctrl; 1643 ddi_acc_handle_t handle; 1644 1645 DDI_INTR_IMPLDBG((CE_CONT, "apix_alloc_msi_vectors: dip=0x%p " 1646 "inum=0x%x count=0x%x behavior=%d\n", 1647 (void *)dip, inum, count, behavior)); 1648 1649 if (count > 1) { 1650 if (behavior == DDI_INTR_ALLOC_STRICT && 1651 apic_multi_msi_enable == 0) 1652 return (0); 1653 if (apic_multi_msi_enable == 0) 1654 count = 1; 1655 } 1656 1657 /* Check whether it supports per-vector masking */ 1658 cap_ptr = i_ddi_get_msi_msix_cap_ptr(dip); 1659 handle = i_ddi_get_pci_config_handle(dip); 1660 msi_ctrl = pci_config_get16(handle, cap_ptr + PCI_MSI_CTRL); 1661 1662 /* bind to cpu */ 1663 bindcpu = apix_bind_cpu(dip); 1664 cpuid = bindcpu & ~IRQ_USER_BOUND; 1665 1666 /* if not ISP2, then round it down */ 1667 if (!ISP2(rcount)) 1668 rcount = 1 << (highbit(rcount) - 1); 1669 1670 APIX_ENTER_CPU_LOCK(cpuid); 1671 for (vecp = NULL; rcount > 0; rcount >>= 1) { 1672 vecp = apix_alloc_nvectors_oncpu(bindcpu, dip, inum, rcount, 1673 APIX_TYPE_MSI); 1674 if (vecp != NULL || behavior == DDI_INTR_ALLOC_STRICT) 1675 break; 1676 } 1677 for (i = 0; vecp && i < rcount; i++) 1678 xv_vector(vecp->v_cpuid, vecp->v_vector + i)->v_flags |= 1679 (msi_ctrl & PCI_MSI_PVM_MASK) ? APIX_VECT_MASKABLE : 0; 1680 APIX_LEAVE_CPU_LOCK(cpuid); 1681 if (vecp == NULL) { 1682 APIC_VERBOSE(INTR, (CE_CONT, 1683 "apix_alloc_msi: no %d cont vectors found on cpu 0x%x\n", 1684 count, bindcpu)); 1685 return (0); 1686 } 1687 1688 /* major to cpu binding */ 1689 if ((apic_intr_policy == INTR_ROUND_ROBIN_WITH_AFFINITY) && 1690 ((vecp->v_flags & APIX_VECT_USER_BOUND) == 0)) 1691 apix_set_dev_binding(dip, vecp->v_cpuid); 1692 1693 apix_dprint_vector(vecp, dip, rcount); 1694 1695 return (rcount); 1696 } 1697 1698 int 1699 apix_alloc_msix(dev_info_t *dip, int inum, int count, int behavior) 1700 { 1701 apix_vector_t *vecp; 1702 processorid_t bindcpu, cpuid; 1703 int i; 1704 1705 for (i = 0; i < count; i++) { 1706 /* select cpu by system policy */ 1707 bindcpu = apix_bind_cpu(dip); 1708 cpuid = bindcpu & ~IRQ_USER_BOUND; 1709 1710 /* allocate vector */ 1711 APIX_ENTER_CPU_LOCK(cpuid); 1712 if ((vecp = apix_alloc_vector_oncpu(bindcpu, dip, inum + i, 1713 APIX_TYPE_MSIX)) == NULL) { 1714 APIX_LEAVE_CPU_LOCK(cpuid); 1715 APIC_VERBOSE(INTR, (CE_CONT, "apix_alloc_msix: " 1716 "allocate msix for device dip=%p, inum=%d on" 1717 " cpu %d failed", (void *)dip, inum + i, bindcpu)); 1718 break; 1719 } 1720 vecp->v_flags |= APIX_VECT_MASKABLE; 1721 APIX_LEAVE_CPU_LOCK(cpuid); 1722 1723 /* major to cpu mapping */ 1724 if ((i == 0) && 1725 (apic_intr_policy == INTR_ROUND_ROBIN_WITH_AFFINITY) && 1726 ((vecp->v_flags & APIX_VECT_USER_BOUND) == 0)) 1727 apix_set_dev_binding(dip, vecp->v_cpuid); 1728 1729 apix_dprint_vector(vecp, dip, 1); 1730 } 1731 1732 if (i < count && behavior == DDI_INTR_ALLOC_STRICT) { 1733 APIC_VERBOSE(INTR, (CE_WARN, "apix_alloc_msix: " 1734 "strictly allocate %d vectors failed, got %d\n", 1735 count, i)); 1736 apix_free_vectors(dip, inum, i, APIX_TYPE_MSIX); 1737 i = 0; 1738 } 1739 1740 return (i); 1741 } 1742 1743 /* 1744 * A rollback free for vectors allocated by apix_alloc_xxx(). 1745 */ 1746 void 1747 apix_free_vectors(dev_info_t *dip, int inum, int count, int type) 1748 { 1749 int i, cpuid; 1750 apix_vector_t *vecp; 1751 1752 DDI_INTR_IMPLDBG((CE_CONT, "apix_free_vectors: dip: %p inum: %x " 1753 "count: %x type: %x\n", 1754 (void *)dip, inum, count, type)); 1755 1756 lock_set(&apix_lock); 1757 1758 for (i = 0; i < count; i++, inum++) { 1759 if ((vecp = apix_get_dev_map(dip, inum, type)) == NULL) { 1760 lock_clear(&apix_lock); 1761 DDI_INTR_IMPLDBG((CE_CONT, "apix_free_vectors: " 1762 "dip=0x%p inum=0x%x type=0x%x apix_find_intr() " 1763 "failed\n", (void *)dip, inum, type)); 1764 continue; 1765 } 1766 1767 APIX_ENTER_CPU_LOCK(vecp->v_cpuid); 1768 cpuid = vecp->v_cpuid; 1769 1770 DDI_INTR_IMPLDBG((CE_CONT, "apix_free_vectors: " 1771 "dip=0x%p inum=0x%x type=0x%x vector 0x%x (share %d)\n", 1772 (void *)dip, inum, type, vecp->v_vector, vecp->v_share)); 1773 1774 /* tear down device interrupt to vector mapping */ 1775 apix_clear_dev_map(dip, inum, type); 1776 1777 if (vecp->v_type == APIX_TYPE_FIXED) { 1778 if (vecp->v_share > 0) { /* share IRQ line */ 1779 APIX_LEAVE_CPU_LOCK(cpuid); 1780 continue; 1781 } 1782 1783 /* Free apic_irq_table entry */ 1784 apix_intx_free(vecp->v_inum); 1785 } 1786 1787 /* free vector */ 1788 apix_cleanup_vector(vecp); 1789 1790 APIX_LEAVE_CPU_LOCK(cpuid); 1791 } 1792 1793 lock_clear(&apix_lock); 1794 } 1795 1796 /* 1797 * Must be called with apix_lock held 1798 */ 1799 apix_vector_t * 1800 apix_setup_io_intr(apix_vector_t *vecp) 1801 { 1802 processorid_t bindcpu; 1803 int ret; 1804 1805 ASSERT(LOCK_HELD(&apix_lock)); 1806 1807 /* 1808 * Interrupts are enabled on the CPU, programme IOAPIC RDT 1809 * entry or MSI/X address/data to enable the interrupt. 1810 */ 1811 if (apix_is_cpu_enabled(vecp->v_cpuid)) { 1812 apix_enable_vector(vecp); 1813 return (vecp); 1814 } 1815 1816 /* 1817 * CPU is not up or interrupts are disabled. Fall back to the 1818 * first avialable CPU. 1819 */ 1820 bindcpu = apic_find_cpu(APIC_CPU_INTR_ENABLE); 1821 1822 if (vecp->v_type == APIX_TYPE_MSI) 1823 return (apix_grp_set_cpu(vecp, bindcpu, &ret)); 1824 1825 return (apix_set_cpu(vecp, bindcpu, &ret)); 1826 } 1827 1828 /* 1829 * For interrupts which call add_avintr() before apic is initialized. 1830 * ioapix_setup_intr() will 1831 * - allocate vector 1832 * - copy over ISR 1833 */ 1834 static void 1835 ioapix_setup_intr(int irqno, iflag_t *flagp) 1836 { 1837 extern struct av_head autovect[]; 1838 apix_vector_t *vecp; 1839 apic_irq_t *irqp; 1840 uchar_t ioapicindex, ipin; 1841 ulong_t iflag; 1842 struct autovec *avp; 1843 1844 ioapicindex = acpi_find_ioapic(irqno); 1845 ASSERT(ioapicindex != 0xFF); 1846 ipin = irqno - apic_io_vectbase[ioapicindex]; 1847 1848 mutex_enter(&airq_mutex); 1849 irqp = apic_irq_table[irqno]; 1850 1851 /* 1852 * The irq table entry shouldn't exist unless the interrupts are shared. 1853 * In that case, make sure it matches what we would initialize it to. 1854 */ 1855 if (irqp != NULL) { 1856 ASSERT(irqp->airq_mps_intr_index == ACPI_INDEX); 1857 ASSERT(irqp->airq_intin_no == ipin && 1858 irqp->airq_ioapicindex == ioapicindex); 1859 vecp = xv_vector(irqp->airq_cpu, irqp->airq_vector); 1860 ASSERT(!IS_VECT_FREE(vecp)); 1861 mutex_exit(&airq_mutex); 1862 } else { 1863 irqp = kmem_zalloc(sizeof (apic_irq_t), KM_SLEEP); 1864 1865 irqp->airq_cpu = IRQ_UNINIT; 1866 irqp->airq_origirq = (uchar_t)irqno; 1867 irqp->airq_mps_intr_index = ACPI_INDEX; 1868 irqp->airq_ioapicindex = ioapicindex; 1869 irqp->airq_intin_no = ipin; 1870 irqp->airq_iflag = *flagp; 1871 irqp->airq_share++; 1872 1873 apic_irq_table[irqno] = irqp; 1874 mutex_exit(&airq_mutex); 1875 1876 vecp = apix_alloc_intx(NULL, 0, irqno); 1877 } 1878 1879 /* copy over autovect */ 1880 for (avp = autovect[irqno].avh_link; avp; avp = avp->av_link) 1881 apix_insert_av(vecp, avp->av_intr_id, avp->av_vector, 1882 avp->av_intarg1, avp->av_intarg2, avp->av_ticksp, 1883 avp->av_prilevel, avp->av_dip); 1884 1885 /* Program I/O APIC */ 1886 iflag = intr_clear(); 1887 lock_set(&apix_lock); 1888 1889 (void) apix_setup_io_intr(vecp); 1890 1891 lock_clear(&apix_lock); 1892 intr_restore(iflag); 1893 1894 APIC_VERBOSE_IOAPIC((CE_CONT, "apix: setup ioapic, irqno %x " 1895 "(ioapic %x, ipin %x) is bound to cpu %x, vector %x\n", 1896 irqno, ioapicindex, ipin, irqp->airq_cpu, irqp->airq_vector)); 1897 } 1898 1899 void 1900 ioapix_init_intr(int mask_apic) 1901 { 1902 int ioapicindex; 1903 int i, j; 1904 1905 /* mask interrupt vectors */ 1906 for (j = 0; j < apic_io_max && mask_apic; j++) { 1907 int intin_max; 1908 1909 ioapicindex = j; 1910 /* Bits 23-16 define the maximum redirection entries */ 1911 intin_max = (ioapic_read(ioapicindex, APIC_VERS_CMD) >> 16) 1912 & 0xff; 1913 for (i = 0; i <= intin_max; i++) 1914 ioapic_write(ioapicindex, APIC_RDT_CMD + 2 * i, 1915 AV_MASK); 1916 } 1917 1918 /* 1919 * Hack alert: deal with ACPI SCI interrupt chicken/egg here 1920 */ 1921 if (apic_sci_vect > 0) 1922 ioapix_setup_intr(apic_sci_vect, &apic_sci_flags); 1923 1924 /* 1925 * Hack alert: deal with ACPI HPET interrupt chicken/egg here. 1926 */ 1927 if (apic_hpet_vect > 0) 1928 ioapix_setup_intr(apic_hpet_vect, &apic_hpet_flags); 1929 } 1930