1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* 23 * Copyright (c) 2010, Oracle and/or its affiliates. All rights reserved. 24 */ 25 /* 26 * Copyright (c) 2010, Intel Corporation. 27 * All rights reserved. 28 */ 29 30 #include <sys/processor.h> 31 #include <sys/time.h> 32 #include <sys/psm.h> 33 #include <sys/smp_impldefs.h> 34 #include <sys/cram.h> 35 #include <sys/acpi/acpi.h> 36 #include <sys/acpica.h> 37 #include <sys/psm_common.h> 38 #include <sys/pit.h> 39 #include <sys/ddi.h> 40 #include <sys/sunddi.h> 41 #include <sys/ddi_impldefs.h> 42 #include <sys/pci.h> 43 #include <sys/promif.h> 44 #include <sys/x86_archext.h> 45 #include <sys/cpc_impl.h> 46 #include <sys/uadmin.h> 47 #include <sys/panic.h> 48 #include <sys/debug.h> 49 #include <sys/archsystm.h> 50 #include <sys/trap.h> 51 #include <sys/machsystm.h> 52 #include <sys/sysmacros.h> 53 #include <sys/cpuvar.h> 54 #include <sys/rm_platter.h> 55 #include <sys/privregs.h> 56 #include <sys/note.h> 57 #include <sys/pci_intr_lib.h> 58 #include <sys/spl.h> 59 #include <sys/clock.h> 60 #include <sys/dditypes.h> 61 #include <sys/sunddi.h> 62 #include <sys/x_call.h> 63 #include <sys/reboot.h> 64 #include <sys/apix.h> 65 66 static int apix_get_avail_vector_oncpu(uint32_t, int, int); 67 static apix_vector_t *apix_init_vector(processorid_t, uchar_t); 68 static void apix_cleanup_vector(apix_vector_t *); 69 static void apix_insert_av(apix_vector_t *, void *, avfunc, caddr_t, caddr_t, 70 uint64_t *, int, dev_info_t *); 71 static void apix_remove_av(apix_vector_t *, struct autovec *); 72 static void apix_clear_dev_map(dev_info_t *, int, int); 73 static boolean_t apix_is_cpu_enabled(processorid_t); 74 static void apix_wait_till_seen(processorid_t, int); 75 76 #define GET_INTR_INUM(ihdlp) \ 77 (((ihdlp) != NULL) ? ((ddi_intr_handle_impl_t *)(ihdlp))->ih_inum : 0) 78 79 apix_rebind_info_t apix_rebindinfo = {0, 0, 0, NULL, 0, NULL}; 80 81 /* 82 * Allocate IPI 83 * 84 * Return vector number or 0 on error 85 */ 86 uchar_t 87 apix_alloc_ipi(int ipl) 88 { 89 apix_vector_t *vecp; 90 uchar_t vector; 91 int cpun; 92 int nproc; 93 94 APIX_ENTER_CPU_LOCK(0); 95 96 vector = apix_get_avail_vector_oncpu(0, APIX_IPI_MIN, APIX_IPI_MAX); 97 if (vector == 0) { 98 APIX_LEAVE_CPU_LOCK(0); 99 cmn_err(CE_WARN, "apix: no available IPI\n"); 100 apic_error |= APIC_ERR_GET_IPIVECT_FAIL; 101 return (0); 102 } 103 104 nproc = max(apic_nproc, apic_max_nproc); 105 for (cpun = 0; cpun < nproc; cpun++) { 106 vecp = xv_vector(cpun, vector); 107 if (vecp == NULL) { 108 vecp = kmem_zalloc(sizeof (apix_vector_t), KM_NOSLEEP); 109 if (vecp == NULL) { 110 cmn_err(CE_WARN, "apix: No memory for ipi"); 111 goto fail; 112 } 113 xv_vector(cpun, vector) = vecp; 114 } 115 vecp->v_state = APIX_STATE_ALLOCED; 116 vecp->v_type = APIX_TYPE_IPI; 117 vecp->v_cpuid = vecp->v_bound_cpuid = cpun; 118 vecp->v_vector = vector; 119 vecp->v_pri = ipl; 120 } 121 APIX_LEAVE_CPU_LOCK(0); 122 return (vector); 123 124 fail: 125 while (--cpun >= 0) 126 apix_cleanup_vector(xv_vector(cpun, vector)); 127 APIX_LEAVE_CPU_LOCK(0); 128 return (0); 129 } 130 131 /* 132 * Add IPI service routine 133 */ 134 static int 135 apix_add_ipi(int ipl, avfunc xxintr, char *name, int vector, 136 caddr_t arg1, caddr_t arg2) 137 { 138 int cpun; 139 apix_vector_t *vecp; 140 int nproc; 141 142 ASSERT(vector >= APIX_IPI_MIN && vector <= APIX_IPI_MAX); 143 144 nproc = max(apic_nproc, apic_max_nproc); 145 for (cpun = 0; cpun < nproc; cpun++) { 146 APIX_ENTER_CPU_LOCK(cpun); 147 vecp = xv_vector(cpun, vector); 148 apix_insert_av(vecp, NULL, xxintr, arg1, arg2, NULL, ipl, NULL); 149 vecp->v_state = APIX_STATE_ENABLED; 150 APIX_LEAVE_CPU_LOCK(cpun); 151 } 152 153 APIC_VERBOSE(IPI, (CE_CONT, "apix: add ipi for %s, vector %x " 154 "ipl %x\n", name, vector, ipl)); 155 156 return (1); 157 } 158 159 /* 160 * Find and return first free vector in range (start, end) 161 */ 162 static int 163 apix_get_avail_vector_oncpu(uint32_t cpuid, int start, int end) 164 { 165 int i; 166 apix_impl_t *apixp = apixs[cpuid]; 167 168 for (i = start; i <= end; i++) { 169 if (APIC_CHECK_RESERVE_VECTORS(i)) 170 continue; 171 if (IS_VECT_FREE(apixp->x_vectbl[i])) 172 return (i); 173 } 174 175 return (0); 176 } 177 178 /* 179 * Allocate a vector on specified cpu 180 * 181 * Return NULL on error 182 */ 183 static apix_vector_t * 184 apix_alloc_vector_oncpu(uint32_t cpuid, dev_info_t *dip, int inum, int type) 185 { 186 processorid_t tocpu = cpuid & ~IRQ_USER_BOUND; 187 apix_vector_t *vecp; 188 int vector; 189 190 ASSERT(APIX_CPU_LOCK_HELD(tocpu)); 191 192 /* find free vector */ 193 vector = apix_get_avail_vector_oncpu(tocpu, APIX_AVINTR_MIN, 194 APIX_AVINTR_MAX); 195 if (vector == 0) 196 return (NULL); 197 198 vecp = apix_init_vector(tocpu, vector); 199 vecp->v_type = (ushort_t)type; 200 vecp->v_inum = inum; 201 vecp->v_flags = (cpuid & IRQ_USER_BOUND) ? APIX_VECT_USER_BOUND : 0; 202 203 if (dip != NULL) 204 apix_set_dev_map(vecp, dip, inum); 205 206 return (vecp); 207 } 208 209 /* 210 * Allocates "count" contiguous MSI vectors starting at the proper alignment. 211 * Caller needs to make sure that count has to be power of 2 and should not 212 * be < 1. 213 * 214 * Return first vector number 215 */ 216 apix_vector_t * 217 apix_alloc_nvectors_oncpu(uint32_t cpuid, dev_info_t *dip, int inum, 218 int count, int type) 219 { 220 int i, msibits, start = 0, navail = 0; 221 apix_vector_t *vecp, *startp = NULL; 222 processorid_t tocpu = cpuid & ~IRQ_USER_BOUND; 223 uint_t flags; 224 225 ASSERT(APIX_CPU_LOCK_HELD(tocpu)); 226 227 /* 228 * msibits is the no. of lower order message data bits for the 229 * allocated MSI vectors and is used to calculate the aligned 230 * starting vector 231 */ 232 msibits = count - 1; 233 234 /* It has to be contiguous */ 235 for (i = APIX_AVINTR_MIN; i <= APIX_AVINTR_MAX; i++) { 236 if (!IS_VECT_FREE(xv_vector(tocpu, i))) 237 continue; 238 239 /* 240 * starting vector has to be aligned accordingly for 241 * multiple MSIs 242 */ 243 if (msibits) 244 i = (i + msibits) & ~msibits; 245 246 for (navail = 0, start = i; i <= APIX_AVINTR_MAX; i++) { 247 if (!IS_VECT_FREE(xv_vector(tocpu, i))) 248 break; 249 if (APIC_CHECK_RESERVE_VECTORS(i)) 250 break; 251 if (++navail == count) 252 goto done; 253 } 254 } 255 256 return (NULL); 257 258 done: 259 flags = (cpuid & IRQ_USER_BOUND) ? APIX_VECT_USER_BOUND : 0; 260 261 for (i = 0; i < count; i++) { 262 if ((vecp = apix_init_vector(tocpu, start + i)) == NULL) 263 goto fail; 264 265 vecp->v_type = (ushort_t)type; 266 vecp->v_inum = inum + i; 267 vecp->v_flags = flags; 268 269 if (dip != NULL) 270 apix_set_dev_map(vecp, dip, inum + i); 271 272 if (i == 0) 273 startp = vecp; 274 } 275 276 return (startp); 277 278 fail: 279 while (i-- > 0) { /* Free allocated vectors */ 280 vecp = xv_vector(tocpu, start + i); 281 apix_clear_dev_map(dip, inum + i, type); 282 apix_cleanup_vector(vecp); 283 } 284 return (NULL); 285 } 286 287 #define APIX_WRITE_MSI_DATA(_hdl, _cap, _ctrl, _v)\ 288 do {\ 289 if ((_ctrl) & PCI_MSI_64BIT_MASK)\ 290 pci_config_put16((_hdl), (_cap) + PCI_MSI_64BIT_DATA, (_v));\ 291 else\ 292 pci_config_put16((_hdl), (_cap) + PCI_MSI_32BIT_DATA, (_v));\ 293 _NOTE(CONSTCOND)} while (0) 294 295 static void 296 apix_pci_msi_enable_vector(apix_vector_t *vecp, dev_info_t *dip, int type, 297 int inum, int count, uchar_t vector, int target_apic_id) 298 { 299 uint64_t msi_addr, msi_data; 300 ushort_t msi_ctrl; 301 int i, cap_ptr = i_ddi_get_msi_msix_cap_ptr(dip); 302 ddi_acc_handle_t handle = i_ddi_get_pci_config_handle(dip); 303 msi_regs_t msi_regs; 304 void *intrmap_tbl[PCI_MSI_MAX_INTRS]; 305 306 DDI_INTR_IMPLDBG((CE_CONT, "apix_pci_msi_enable_vector: dip=0x%p\n" 307 "\tdriver = %s, inum=0x%x vector=0x%x apicid=0x%x\n", (void *)dip, 308 ddi_driver_name(dip), inum, vector, target_apic_id)); 309 310 ASSERT((handle != NULL) && (cap_ptr != 0)); 311 312 msi_regs.mr_data = vector; 313 msi_regs.mr_addr = target_apic_id; 314 315 for (i = 0; i < count; i++) 316 intrmap_tbl[i] = xv_intrmap_private(vecp->v_cpuid, vector + i); 317 apic_vt_ops->apic_intrmap_alloc_entry(intrmap_tbl, dip, type, 318 count, 0xff); 319 for (i = 0; i < count; i++) 320 xv_intrmap_private(vecp->v_cpuid, vector + i) = intrmap_tbl[i]; 321 322 apic_vt_ops->apic_intrmap_map_entry(vecp->v_intrmap_private, 323 (void *)&msi_regs, type, count); 324 apic_vt_ops->apic_intrmap_record_msi(vecp->v_intrmap_private, 325 &msi_regs); 326 327 /* MSI Address */ 328 msi_addr = msi_regs.mr_addr; 329 330 /* MSI Data: MSI is edge triggered according to spec */ 331 msi_data = msi_regs.mr_data; 332 333 DDI_INTR_IMPLDBG((CE_CONT, "apix_pci_msi_enable_vector: addr=0x%lx " 334 "data=0x%lx\n", (long)msi_addr, (long)msi_data)); 335 336 if (type == APIX_TYPE_MSI) { 337 msi_ctrl = pci_config_get16(handle, cap_ptr + PCI_MSI_CTRL); 338 339 /* Set the bits to inform how many MSIs are enabled */ 340 msi_ctrl |= ((highbit(count) - 1) << PCI_MSI_MME_SHIFT); 341 pci_config_put16(handle, cap_ptr + PCI_MSI_CTRL, msi_ctrl); 342 343 if ((vecp->v_flags & APIX_VECT_MASKABLE) == 0) 344 APIX_WRITE_MSI_DATA(handle, cap_ptr, msi_ctrl, 345 APIX_RESV_VECTOR); 346 347 pci_config_put32(handle, 348 cap_ptr + PCI_MSI_ADDR_OFFSET, msi_addr); 349 if (msi_ctrl & PCI_MSI_64BIT_MASK) 350 pci_config_put32(handle, 351 cap_ptr + PCI_MSI_ADDR_OFFSET + 4, msi_addr >> 32); 352 353 APIX_WRITE_MSI_DATA(handle, cap_ptr, msi_ctrl, msi_data); 354 } else if (type == APIX_TYPE_MSIX) { 355 uintptr_t off; 356 ddi_intr_msix_t *msix_p = i_ddi_get_msix(dip); 357 358 /* Offset into the "inum"th entry in the MSI-X table */ 359 off = (uintptr_t)msix_p->msix_tbl_addr + 360 (inum * PCI_MSIX_VECTOR_SIZE); 361 362 ddi_put32(msix_p->msix_tbl_hdl, 363 (uint32_t *)(off + PCI_MSIX_DATA_OFFSET), msi_data); 364 ddi_put64(msix_p->msix_tbl_hdl, 365 (uint64_t *)(off + PCI_MSIX_LOWER_ADDR_OFFSET), msi_addr); 366 } 367 } 368 369 static void 370 apix_pci_msi_enable_mode(dev_info_t *dip, int type, int inum) 371 { 372 ushort_t msi_ctrl; 373 int cap_ptr = i_ddi_get_msi_msix_cap_ptr(dip); 374 ddi_acc_handle_t handle = i_ddi_get_pci_config_handle(dip); 375 376 ASSERT((handle != NULL) && (cap_ptr != 0)); 377 378 if (type == APIX_TYPE_MSI) { 379 msi_ctrl = pci_config_get16(handle, cap_ptr + PCI_MSI_CTRL); 380 if ((msi_ctrl & PCI_MSI_ENABLE_BIT)) 381 return; 382 383 msi_ctrl |= PCI_MSI_ENABLE_BIT; 384 pci_config_put16(handle, cap_ptr + PCI_MSI_CTRL, msi_ctrl); 385 386 } else if (type == DDI_INTR_TYPE_MSIX) { 387 uintptr_t off; 388 uint32_t mask; 389 ddi_intr_msix_t *msix_p; 390 391 msix_p = i_ddi_get_msix(dip); 392 393 /* Offset into "inum"th entry in the MSI-X table & clear mask */ 394 off = (uintptr_t)msix_p->msix_tbl_addr + (inum * 395 PCI_MSIX_VECTOR_SIZE) + PCI_MSIX_VECTOR_CTRL_OFFSET; 396 397 mask = ddi_get32(msix_p->msix_tbl_hdl, (uint32_t *)off); 398 399 ddi_put32(msix_p->msix_tbl_hdl, (uint32_t *)off, (mask & ~1)); 400 401 msi_ctrl = pci_config_get16(handle, cap_ptr + PCI_MSIX_CTRL); 402 403 if (!(msi_ctrl & PCI_MSIX_ENABLE_BIT)) { 404 msi_ctrl |= PCI_MSIX_ENABLE_BIT; 405 pci_config_put16(handle, cap_ptr + PCI_MSIX_CTRL, 406 msi_ctrl); 407 } 408 } 409 } 410 411 /* 412 * Setup interrupt, pogramming IO-APIC or MSI/X address/data. 413 */ 414 void 415 apix_enable_vector(apix_vector_t *vecp) 416 { 417 int tocpu = vecp->v_cpuid, type = vecp->v_type; 418 apic_cpus_info_t *cpu_infop; 419 ulong_t iflag; 420 421 ASSERT(tocpu < apic_nproc); 422 423 cpu_infop = &apic_cpus[tocpu]; 424 if (vecp->v_flags & APIX_VECT_USER_BOUND) 425 cpu_infop->aci_bound++; 426 else 427 cpu_infop->aci_temp_bound++; 428 429 iflag = intr_clear(); 430 lock_set(&apic_ioapic_lock); 431 432 if (!DDI_INTR_IS_MSI_OR_MSIX(type)) { /* fixed */ 433 apix_intx_enable(vecp->v_inum); 434 } else { 435 int inum = vecp->v_inum; 436 dev_info_t *dip = APIX_GET_DIP(vecp); 437 int count = i_ddi_intr_get_current_nintrs(dip); 438 439 if (type == APIX_TYPE_MSI) { /* MSI */ 440 if (inum == apix_get_max_dev_inum(dip, type)) { 441 /* last one */ 442 uchar_t start_inum = inum + 1 - count; 443 uchar_t start_vect = vecp->v_vector + 1 - count; 444 apix_vector_t *start_vecp = 445 xv_vector(vecp->v_cpuid, start_vect); 446 447 APIC_VERBOSE(INTR, (CE_CONT, "apix: call " 448 "apix_pci_msi_enable_vector\n")); 449 apix_pci_msi_enable_vector(start_vecp, dip, 450 type, start_inum, count, start_vect, 451 cpu_infop->aci_local_id); 452 453 APIC_VERBOSE(INTR, (CE_CONT, "apix: call " 454 "apix_pci_msi_enable_mode\n")); 455 apix_pci_msi_enable_mode(dip, type, inum); 456 } 457 } else { /* MSI-X */ 458 apix_pci_msi_enable_vector(vecp, dip, 459 type, inum, 1, vecp->v_vector, 460 cpu_infop->aci_local_id); 461 apix_pci_msi_enable_mode(dip, type, inum); 462 } 463 } 464 vecp->v_state = APIX_STATE_ENABLED; 465 apic_redist_cpu_skip &= ~(1 << tocpu); 466 467 lock_clear(&apic_ioapic_lock); 468 intr_restore(iflag); 469 } 470 471 /* 472 * Disable the interrupt 473 */ 474 void 475 apix_disable_vector(apix_vector_t *vecp) 476 { 477 struct autovec *avp = vecp->v_autovect; 478 ulong_t iflag; 479 480 ASSERT(avp != NULL); 481 482 iflag = intr_clear(); 483 lock_set(&apic_ioapic_lock); 484 485 switch (vecp->v_type) { 486 case APIX_TYPE_MSI: 487 ASSERT(avp->av_vector != NULL && avp->av_dip != NULL); 488 /* 489 * Disable the MSI vector 490 * Make sure we only disable on the last 491 * of the multi-MSI support 492 */ 493 if (i_ddi_intr_get_current_nenables(avp->av_dip) == 1) { 494 apic_pci_msi_disable_mode(avp->av_dip, 495 DDI_INTR_TYPE_MSI); 496 } 497 break; 498 case APIX_TYPE_MSIX: 499 ASSERT(avp->av_vector != NULL && avp->av_dip != NULL); 500 /* 501 * Disable the MSI-X vector 502 * needs to clear its mask and addr/data for each MSI-X 503 */ 504 apic_pci_msi_unconfigure(avp->av_dip, DDI_INTR_TYPE_MSIX, 505 vecp->v_inum); 506 /* 507 * Make sure we only disable on the last MSI-X 508 */ 509 if (i_ddi_intr_get_current_nenables(avp->av_dip) == 1) { 510 apic_pci_msi_disable_mode(avp->av_dip, 511 DDI_INTR_TYPE_MSIX); 512 } 513 break; 514 default: 515 apix_intx_disable(vecp->v_inum); 516 break; 517 } 518 519 if (!(apic_cpus[vecp->v_cpuid].aci_status & APIC_CPU_SUSPEND)) 520 vecp->v_state = APIX_STATE_DISABLED; 521 apic_vt_ops->apic_intrmap_free_entry(&vecp->v_intrmap_private); 522 vecp->v_intrmap_private = NULL; 523 524 lock_clear(&apic_ioapic_lock); 525 intr_restore(iflag); 526 } 527 528 /* 529 * Mark vector as obsoleted or freed. The vector is marked 530 * obsoleted if there are pending requests on it. Otherwise, 531 * free the vector. The obsoleted vectors get freed after 532 * being serviced. 533 * 534 * Return 1 on being obosoleted and 0 on being freed. 535 */ 536 #define INTR_BUSY(_avp)\ 537 ((((volatile ushort_t)(_avp)->av_flags) &\ 538 (AV_PENTRY_PEND | AV_PENTRY_ONPROC)) != 0) 539 #define LOCAL_WITH_INTR_DISABLED(_cpuid)\ 540 ((_cpuid) == psm_get_cpu_id() && !interrupts_enabled()) 541 static uint64_t dummy_tick; 542 543 int 544 apix_obsolete_vector(apix_vector_t *vecp) 545 { 546 struct autovec *avp = vecp->v_autovect; 547 int repeats, tries, ipl, busy = 0, cpuid = vecp->v_cpuid; 548 apix_impl_t *apixp = apixs[cpuid]; 549 550 ASSERT(APIX_CPU_LOCK_HELD(cpuid)); 551 552 for (avp = vecp->v_autovect; avp != NULL; avp = avp->av_link) { 553 if (avp->av_vector == NULL) 554 continue; 555 556 if (LOCAL_WITH_INTR_DISABLED(cpuid)) { 557 int bit, index, irr; 558 559 if (INTR_BUSY(avp)) { 560 busy++; 561 continue; 562 } 563 564 /* check IRR for pending interrupts */ 565 index = vecp->v_vector / 32; 566 bit = vecp->v_vector % 32; 567 irr = apic_reg_ops->apic_read(APIC_IRR_REG + index); 568 if ((irr & (1 << bit)) != 0) 569 busy++; 570 571 if (!busy) 572 apix_remove_av(vecp, avp); 573 574 continue; 575 } 576 577 repeats = 0; 578 do { 579 repeats++; 580 for (tries = 0; tries < apic_max_reps_clear_pending; 581 tries++) 582 if (!INTR_BUSY(avp)) 583 break; 584 } while (INTR_BUSY(avp) && 585 (repeats < apic_max_reps_clear_pending)); 586 587 if (INTR_BUSY(avp)) 588 busy++; 589 else { 590 /* 591 * Interrupt is not in pending list or being serviced. 592 * However it might be cached in Local APIC's IRR 593 * register. It's impossible to check another CPU's 594 * IRR register. Then wait till lower levels finish 595 * running. 596 */ 597 for (ipl = 1; ipl < MIN(LOCK_LEVEL, vecp->v_pri); ipl++) 598 apix_wait_till_seen(cpuid, ipl); 599 if (INTR_BUSY(avp)) 600 busy++; 601 } 602 603 if (!busy) 604 apix_remove_av(vecp, avp); 605 } 606 607 if (busy) { 608 apix_vector_t *tp = apixp->x_obsoletes; 609 610 if (vecp->v_state == APIX_STATE_OBSOLETED) 611 return (1); 612 613 vecp->v_state = APIX_STATE_OBSOLETED; 614 vecp->v_next = NULL; 615 if (tp == NULL) 616 apixp->x_obsoletes = vecp; 617 else { 618 while (tp->v_next != NULL) 619 tp = tp->v_next; 620 tp->v_next = vecp; 621 } 622 return (1); 623 } 624 625 /* interrupt is not busy */ 626 if (vecp->v_state == APIX_STATE_OBSOLETED) { 627 /* remove from obsoleted list */ 628 apixp->x_obsoletes = vecp->v_next; 629 vecp->v_next = NULL; 630 } 631 apix_cleanup_vector(vecp); 632 return (0); 633 } 634 635 /* 636 * Duplicate number of continuous vectors to specified target vectors. 637 */ 638 static void 639 apix_dup_vectors(apix_vector_t *oldp, apix_vector_t *newp, int count) 640 { 641 struct autovec *avp; 642 apix_vector_t *fromp, *top; 643 processorid_t oldcpu = oldp->v_cpuid, newcpu = newp->v_cpuid; 644 uchar_t oldvec = oldp->v_vector, newvec = newp->v_vector; 645 int i, inum; 646 647 ASSERT(oldp->v_type != APIX_TYPE_IPI); 648 649 for (i = 0; i < count; i++) { 650 fromp = xv_vector(oldcpu, oldvec + i); 651 top = xv_vector(newcpu, newvec + i); 652 ASSERT(fromp != NULL && top != NULL); 653 654 /* copy over original one */ 655 top->v_state = fromp->v_state; 656 top->v_type = fromp->v_type; 657 top->v_bound_cpuid = fromp->v_bound_cpuid; 658 top->v_inum = fromp->v_inum; 659 top->v_flags = fromp->v_flags; 660 top->v_intrmap_private = fromp->v_intrmap_private; 661 662 for (avp = fromp->v_autovect; avp != NULL; avp = avp->av_link) { 663 if (avp->av_vector == NULL) 664 continue; 665 666 apix_insert_av(top, avp->av_intr_id, avp->av_vector, 667 avp->av_intarg1, avp->av_intarg2, avp->av_ticksp, 668 avp->av_prilevel, avp->av_dip); 669 670 if (fromp->v_type == APIX_TYPE_FIXED && 671 avp->av_dip != NULL) { 672 inum = GET_INTR_INUM(avp->av_intr_id); 673 apix_set_dev_map(top, avp->av_dip, inum); 674 } 675 } 676 677 if (DDI_INTR_IS_MSI_OR_MSIX(fromp->v_type) && 678 fromp->v_devp != NULL) 679 apix_set_dev_map(top, fromp->v_devp->dv_dip, 680 fromp->v_devp->dv_inum); 681 } 682 } 683 684 static apix_vector_t * 685 apix_init_vector(processorid_t cpuid, uchar_t vector) 686 { 687 apix_impl_t *apixp = apixs[cpuid]; 688 apix_vector_t *vecp = apixp->x_vectbl[vector]; 689 690 ASSERT(IS_VECT_FREE(vecp)); 691 692 if (vecp == NULL) { 693 vecp = kmem_zalloc(sizeof (apix_vector_t), KM_NOSLEEP); 694 if (vecp == NULL) { 695 cmn_err(CE_WARN, "apix: no memory to allocate vector"); 696 return (NULL); 697 } 698 apixp->x_vectbl[vector] = vecp; 699 } 700 vecp->v_state = APIX_STATE_ALLOCED; 701 vecp->v_cpuid = vecp->v_bound_cpuid = cpuid; 702 vecp->v_vector = vector; 703 704 return (vecp); 705 } 706 707 static void 708 apix_cleanup_vector(apix_vector_t *vecp) 709 { 710 ASSERT(vecp->v_share == 0); 711 vecp->v_bound_cpuid = IRQ_UNINIT; 712 vecp->v_state = APIX_STATE_FREED; 713 vecp->v_type = 0; 714 vecp->v_flags = 0; 715 vecp->v_busy = 0; 716 vecp->v_intrmap_private = NULL; 717 } 718 719 static void 720 apix_dprint_vector(apix_vector_t *vecp, dev_info_t *dip, int count) 721 { 722 #ifdef DEBUG 723 major_t major; 724 char *name, *drv_name; 725 int instance, len, t_len; 726 char mesg[1024] = "apix: "; 727 728 t_len = sizeof (mesg); 729 len = strlen(mesg); 730 if (dip != NULL) { 731 name = ddi_get_name(dip); 732 major = ddi_name_to_major(name); 733 drv_name = ddi_major_to_name(major); 734 instance = ddi_get_instance(dip); 735 (void) snprintf(mesg + len, t_len - len, "%s (%s) instance %d ", 736 name, drv_name, instance); 737 } 738 len = strlen(mesg); 739 740 switch (vecp->v_type) { 741 case APIX_TYPE_FIXED: 742 (void) snprintf(mesg + len, t_len - len, "irqno %d", 743 vecp->v_inum); 744 break; 745 case APIX_TYPE_MSI: 746 (void) snprintf(mesg + len, t_len - len, 747 "msi inum %d (count %d)", vecp->v_inum, count); 748 break; 749 case APIX_TYPE_MSIX: 750 (void) snprintf(mesg + len, t_len - len, "msi-x inum %d", 751 vecp->v_inum); 752 break; 753 default: 754 break; 755 756 } 757 758 APIC_VERBOSE(ALLOC, (CE_CONT, "%s allocated with vector 0x%x on " 759 "cpu %d\n", mesg, vecp->v_vector, vecp->v_cpuid)); 760 #endif /* DEBUG */ 761 } 762 763 /* 764 * Operations on avintr 765 */ 766 767 #define INIT_AUTOVEC(p, intr_id, f, arg1, arg2, ticksp, ipl, dip) \ 768 do { \ 769 (p)->av_intr_id = intr_id; \ 770 (p)->av_vector = f; \ 771 (p)->av_intarg1 = arg1; \ 772 (p)->av_intarg2 = arg2; \ 773 (p)->av_ticksp = ticksp; \ 774 (p)->av_prilevel = ipl; \ 775 (p)->av_dip = dip; \ 776 (p)->av_flags = 0; \ 777 _NOTE(CONSTCOND)} while (0) 778 779 /* 780 * Insert an interrupt service routine into chain by its priority from 781 * high to low 782 */ 783 static void 784 apix_insert_av(apix_vector_t *vecp, void *intr_id, avfunc f, caddr_t arg1, 785 caddr_t arg2, uint64_t *ticksp, int ipl, dev_info_t *dip) 786 { 787 struct autovec *p, *prep, *mem; 788 789 APIC_VERBOSE(INTR, (CE_CONT, "apix_insert_av: dip %p, vector 0x%x, " 790 "cpu %d\n", (void *)dip, vecp->v_vector, vecp->v_cpuid)); 791 792 mem = kmem_zalloc(sizeof (struct autovec), KM_SLEEP); 793 INIT_AUTOVEC(mem, intr_id, f, arg1, arg2, ticksp, ipl, dip); 794 if (vecp->v_type == APIX_TYPE_FIXED && apic_level_intr[vecp->v_inum]) 795 mem->av_flags |= AV_PENTRY_LEVEL; 796 797 vecp->v_share++; 798 vecp->v_pri = (ipl > vecp->v_pri) ? ipl : vecp->v_pri; 799 if (vecp->v_autovect == NULL) { /* Nothing on list - put it at head */ 800 vecp->v_autovect = mem; 801 return; 802 } 803 804 if (DDI_INTR_IS_MSI_OR_MSIX(vecp->v_type)) { /* MSI/X */ 805 ASSERT(vecp->v_share == 1); /* No sharing for MSI/X */ 806 807 INIT_AUTOVEC(vecp->v_autovect, intr_id, f, arg1, arg2, ticksp, 808 ipl, dip); 809 prep = vecp->v_autovect->av_link; 810 vecp->v_autovect->av_link = NULL; 811 812 /* Free the following autovect chain */ 813 while (prep != NULL) { 814 ASSERT(prep->av_vector == NULL); 815 816 p = prep; 817 prep = prep->av_link; 818 kmem_free(p, sizeof (struct autovec)); 819 } 820 821 kmem_free(mem, sizeof (struct autovec)); 822 return; 823 } 824 825 /* find where it goes in list */ 826 prep = NULL; 827 for (p = vecp->v_autovect; p != NULL; p = p->av_link) { 828 if (p->av_vector && p->av_prilevel <= ipl) 829 break; 830 prep = p; 831 } 832 if (prep != NULL) { 833 if (prep->av_vector == NULL) { /* freed struct available */ 834 INIT_AUTOVEC(prep, intr_id, f, arg1, arg2, 835 ticksp, ipl, dip); 836 prep->av_flags = mem->av_flags; 837 kmem_free(mem, sizeof (struct autovec)); 838 return; 839 } 840 841 mem->av_link = prep->av_link; 842 prep->av_link = mem; 843 } else { 844 /* insert new intpt at beginning of chain */ 845 mem->av_link = vecp->v_autovect; 846 vecp->v_autovect = mem; 847 } 848 } 849 850 /* 851 * After having made a change to an autovector list, wait until we have 852 * seen specified cpu not executing an interrupt at that level--so we 853 * know our change has taken effect completely (no old state in registers, 854 * etc). 855 */ 856 #define APIX_CPU_ENABLED(_cp) \ 857 (quiesce_active == 0 && \ 858 (((_cp)->cpu_flags & (CPU_QUIESCED|CPU_OFFLINE)) == 0)) 859 860 static void 861 apix_wait_till_seen(processorid_t cpuid, int ipl) 862 { 863 struct cpu *cp = cpu[cpuid]; 864 865 if (cp == NULL || LOCAL_WITH_INTR_DISABLED(cpuid)) 866 return; 867 868 /* 869 * Don't wait if the CPU is quiesced or offlined. This can happen 870 * when a CPU is running pause thread but hardware triggered an 871 * interrupt and the interrupt gets queued. 872 */ 873 for (;;) { 874 if (!INTR_ACTIVE((volatile struct cpu *)cpu[cpuid], ipl) && 875 (!APIX_CPU_ENABLED(cp) || 876 !INTR_PENDING((volatile apix_impl_t *)apixs[cpuid], ipl))) 877 return; 878 } 879 } 880 881 static void 882 apix_remove_av(apix_vector_t *vecp, struct autovec *target) 883 { 884 int hi_pri = 0; 885 struct autovec *p; 886 887 if (target == NULL) 888 return; 889 890 APIC_VERBOSE(INTR, (CE_CONT, "apix_remove_av: dip %p, vector 0x%x, " 891 "cpu %d\n", (void *)target->av_dip, vecp->v_vector, vecp->v_cpuid)); 892 893 for (p = vecp->v_autovect; p; p = p->av_link) { 894 if (p == target || p->av_vector == NULL) 895 continue; 896 hi_pri = (p->av_prilevel > hi_pri) ? p->av_prilevel : hi_pri; 897 } 898 899 vecp->v_share--; 900 vecp->v_pri = hi_pri; 901 902 /* 903 * This drops the handler from the chain, it can no longer be called. 904 * However, there is no guarantee that the handler is not currently 905 * still executing. 906 */ 907 target->av_vector = NULL; 908 /* 909 * There is a race where we could be just about to pick up the ticksp 910 * pointer to increment it after returning from the service routine 911 * in av_dispatch_autovect. Rather than NULL it out let's just point 912 * it off to something safe so that any final tick update attempt 913 * won't fault. 914 */ 915 target->av_ticksp = &dummy_tick; 916 apix_wait_till_seen(vecp->v_cpuid, target->av_prilevel); 917 } 918 919 static struct autovec * 920 apix_find_av(apix_vector_t *vecp, void *intr_id, avfunc f) 921 { 922 struct autovec *p; 923 924 for (p = vecp->v_autovect; p; p = p->av_link) { 925 if ((p->av_vector == f) && (p->av_intr_id == intr_id)) { 926 /* found the handler */ 927 return (p); 928 } 929 } 930 931 return (NULL); 932 } 933 934 static apix_vector_t * 935 apix_find_vector_by_avintr(void *intr_id, avfunc f) 936 { 937 apix_vector_t *vecp; 938 processorid_t n; 939 uchar_t v; 940 941 for (n = 0; n < apic_nproc; n++) { 942 if (!apix_is_cpu_enabled(n)) 943 continue; 944 945 for (v = APIX_AVINTR_MIN; v <= APIX_AVINTR_MIN; v++) { 946 vecp = xv_vector(n, v); 947 if (vecp == NULL || 948 vecp->v_state <= APIX_STATE_OBSOLETED) 949 continue; 950 951 if (apix_find_av(vecp, intr_id, f) != NULL) 952 return (vecp); 953 } 954 } 955 956 return (NULL); 957 } 958 959 /* 960 * Add interrupt service routine. 961 * 962 * For legacy interrupts (HPET timer, ACPI SCI), the vector is actually 963 * IRQ no. A vector is then allocated. Otherwise, the vector is already 964 * allocated. The input argument virt_vect is virtual vector of format 965 * APIX_VIRTVEC_VECTOR(cpuid, vector). 966 * 967 * Return 1 on success, 0 on failure. 968 */ 969 int 970 apix_add_avintr(void *intr_id, int ipl, avfunc xxintr, char *name, 971 int virt_vect, caddr_t arg1, caddr_t arg2, uint64_t *ticksp, 972 dev_info_t *dip) 973 { 974 int cpuid; 975 uchar_t v = (uchar_t)APIX_VIRTVEC_VECTOR(virt_vect); 976 apix_vector_t *vecp; 977 978 if (xxintr == NULL) { 979 cmn_err(CE_WARN, "Attempt to add null for %s " 980 "on vector 0x%x,0x%x", name, 981 APIX_VIRTVEC_CPU(virt_vect), 982 APIX_VIRTVEC_VECTOR(virt_vect)); 983 return (0); 984 } 985 986 if (v >= APIX_IPI_MIN) /* IPIs */ 987 return (apix_add_ipi(ipl, xxintr, name, v, arg1, arg2)); 988 989 if (!APIX_IS_VIRTVEC(virt_vect)) { /* got irq */ 990 int irqno = virt_vect; 991 int inum = GET_INTR_INUM(intr_id); 992 993 /* 994 * Senarios include: 995 * a. add_avintr() is called before irqp initialized (legacy) 996 * b. irqp is initialized, vector is not allocated (fixed) 997 * c. irqp is initialized, vector is allocated (fixed & shared) 998 */ 999 if ((vecp = apix_alloc_intx(dip, inum, irqno)) == NULL) 1000 return (0); 1001 1002 cpuid = vecp->v_cpuid; 1003 v = vecp->v_vector; 1004 virt_vect = APIX_VIRTVECTOR(cpuid, v); 1005 } else { /* got virtual vector */ 1006 cpuid = APIX_VIRTVEC_CPU(virt_vect); 1007 vecp = xv_vector(cpuid, v); 1008 ASSERT(vecp != NULL); 1009 } 1010 1011 lock_set(&apix_lock); 1012 if (vecp->v_state <= APIX_STATE_OBSOLETED) { 1013 vecp = NULL; 1014 1015 /* 1016 * Basically the allocated but not enabled interrupts 1017 * will not get re-targeted. But MSIs in allocated state 1018 * could be re-targeted due to group re-targeting. 1019 */ 1020 if (intr_id != NULL && dip != NULL) { 1021 ddi_intr_handle_impl_t *hdlp = intr_id; 1022 vecp = apix_get_dev_map(dip, hdlp->ih_inum, 1023 hdlp->ih_type); 1024 ASSERT(vecp->v_state == APIX_STATE_ALLOCED); 1025 } 1026 if (vecp == NULL) { 1027 lock_clear(&apix_lock); 1028 cmn_err(CE_WARN, "Invalid interrupt 0x%x,0x%x " 1029 " for %p to add", cpuid, v, intr_id); 1030 return (0); 1031 } 1032 cpuid = vecp->v_cpuid; 1033 virt_vect = APIX_VIRTVECTOR(cpuid, vecp->v_vector); 1034 } 1035 1036 APIX_ENTER_CPU_LOCK(cpuid); 1037 apix_insert_av(vecp, intr_id, xxintr, arg1, arg2, ticksp, ipl, dip); 1038 APIX_LEAVE_CPU_LOCK(cpuid); 1039 1040 (void) apix_addspl(virt_vect, ipl, 0, 0); 1041 1042 lock_clear(&apix_lock); 1043 1044 return (1); 1045 } 1046 1047 /* 1048 * Remove avintr 1049 * 1050 * For fixed, if it's the last one of shared interrupts, free the vector. 1051 * For msi/x, only disable the interrupt but not free the vector, which 1052 * is freed by PSM_XXX_FREE_XXX. 1053 */ 1054 void 1055 apix_rem_avintr(void *intr_id, int ipl, avfunc xxintr, int virt_vect) 1056 { 1057 avfunc f; 1058 apix_vector_t *vecp; 1059 struct autovec *avp; 1060 processorid_t cpuid; 1061 1062 if ((f = xxintr) == NULL) 1063 return; 1064 1065 lock_set(&apix_lock); 1066 1067 if (!APIX_IS_VIRTVEC(virt_vect)) { /* got irq */ 1068 vecp = apix_intx_get_vector(virt_vect); 1069 virt_vect = APIX_VIRTVECTOR(vecp->v_cpuid, vecp->v_vector); 1070 } else /* got virtual vector */ 1071 vecp = xv_vector(APIX_VIRTVEC_CPU(virt_vect), 1072 APIX_VIRTVEC_VECTOR(virt_vect)); 1073 1074 if (vecp == NULL) { 1075 lock_clear(&apix_lock); 1076 cmn_err(CE_CONT, "Invalid interrupt 0x%x,0x%x to remove", 1077 APIX_VIRTVEC_CPU(virt_vect), 1078 APIX_VIRTVEC_VECTOR(virt_vect)); 1079 return; 1080 } 1081 1082 if (vecp->v_state <= APIX_STATE_OBSOLETED || 1083 ((avp = apix_find_av(vecp, intr_id, f)) == NULL)) { 1084 /* 1085 * It's possible that the interrupt is rebound to a 1086 * different cpu before rem_avintr() is called. Search 1087 * through all vectors once it happens. 1088 */ 1089 if ((vecp = apix_find_vector_by_avintr(intr_id, f)) 1090 == NULL) { 1091 lock_clear(&apix_lock); 1092 cmn_err(CE_CONT, "Unknown interrupt 0x%x,0x%x " 1093 "for %p to remove", APIX_VIRTVEC_CPU(virt_vect), 1094 APIX_VIRTVEC_VECTOR(virt_vect), intr_id); 1095 return; 1096 } 1097 virt_vect = APIX_VIRTVECTOR(vecp->v_cpuid, vecp->v_vector); 1098 avp = apix_find_av(vecp, intr_id, f); 1099 } 1100 cpuid = vecp->v_cpuid; 1101 1102 /* disable interrupt */ 1103 (void) apix_delspl(virt_vect, ipl, 0, 0); 1104 1105 /* remove ISR entry */ 1106 APIX_ENTER_CPU_LOCK(cpuid); 1107 apix_remove_av(vecp, avp); 1108 APIX_LEAVE_CPU_LOCK(cpuid); 1109 1110 lock_clear(&apix_lock); 1111 } 1112 1113 /* 1114 * Device to vector mapping table 1115 */ 1116 1117 static void 1118 apix_clear_dev_map(dev_info_t *dip, int inum, int type) 1119 { 1120 char *name; 1121 major_t major; 1122 apix_dev_vector_t *dvp, *prev = NULL; 1123 int found = 0; 1124 1125 name = ddi_get_name(dip); 1126 major = ddi_name_to_major(name); 1127 1128 mutex_enter(&apix_mutex); 1129 1130 for (dvp = apix_dev_vector[major]; dvp != NULL; 1131 prev = dvp, dvp = dvp->dv_next) { 1132 if (dvp->dv_dip == dip && dvp->dv_inum == inum && 1133 dvp->dv_type == type) { 1134 found++; 1135 break; 1136 } 1137 } 1138 1139 if (!found) { 1140 mutex_exit(&apix_mutex); 1141 return; 1142 } 1143 1144 if (prev != NULL) 1145 prev->dv_next = dvp->dv_next; 1146 1147 if (apix_dev_vector[major] == dvp) 1148 apix_dev_vector[major] = dvp->dv_next; 1149 1150 dvp->dv_vector->v_devp = NULL; 1151 1152 mutex_exit(&apix_mutex); 1153 1154 kmem_free(dvp, sizeof (apix_dev_vector_t)); 1155 } 1156 1157 void 1158 apix_set_dev_map(apix_vector_t *vecp, dev_info_t *dip, int inum) 1159 { 1160 apix_dev_vector_t *dvp; 1161 char *name; 1162 major_t major; 1163 uint32_t found = 0; 1164 1165 ASSERT(dip != NULL); 1166 name = ddi_get_name(dip); 1167 major = ddi_name_to_major(name); 1168 1169 mutex_enter(&apix_mutex); 1170 1171 for (dvp = apix_dev_vector[major]; dvp != NULL; 1172 dvp = dvp->dv_next) { 1173 if (dvp->dv_dip == dip && dvp->dv_inum == inum && 1174 dvp->dv_type == vecp->v_type) { 1175 found++; 1176 break; 1177 } 1178 } 1179 1180 if (found == 0) { /* not found */ 1181 dvp = kmem_zalloc(sizeof (apix_dev_vector_t), KM_SLEEP); 1182 dvp->dv_dip = dip; 1183 dvp->dv_inum = inum; 1184 dvp->dv_type = vecp->v_type; 1185 1186 dvp->dv_next = apix_dev_vector[major]; 1187 apix_dev_vector[major] = dvp; 1188 } 1189 dvp->dv_vector = vecp; 1190 vecp->v_devp = dvp; 1191 1192 mutex_exit(&apix_mutex); 1193 1194 DDI_INTR_IMPLDBG((CE_CONT, "apix_set_dev_map: dip=0x%p " 1195 "inum=0x%x vector=0x%x/0x%x\n", 1196 (void *)dip, inum, vecp->v_cpuid, vecp->v_vector)); 1197 } 1198 1199 apix_vector_t * 1200 apix_get_dev_map(dev_info_t *dip, int inum, int type) 1201 { 1202 char *name; 1203 major_t major; 1204 apix_dev_vector_t *dvp; 1205 apix_vector_t *vecp; 1206 1207 name = ddi_get_name(dip); 1208 if ((major = ddi_name_to_major(name)) == DDI_MAJOR_T_NONE) 1209 return (NULL); 1210 1211 mutex_enter(&apix_mutex); 1212 for (dvp = apix_dev_vector[major]; dvp != NULL; 1213 dvp = dvp->dv_next) { 1214 if (dvp->dv_dip == dip && dvp->dv_inum == inum && 1215 dvp->dv_type == type) { 1216 vecp = dvp->dv_vector; 1217 mutex_exit(&apix_mutex); 1218 return (vecp); 1219 } 1220 } 1221 mutex_exit(&apix_mutex); 1222 1223 return (NULL); 1224 } 1225 1226 /* 1227 * Get minimum inum for specified device, used for MSI 1228 */ 1229 int 1230 apix_get_min_dev_inum(dev_info_t *dip, int type) 1231 { 1232 char *name; 1233 major_t major; 1234 apix_dev_vector_t *dvp; 1235 int inum = -1; 1236 1237 name = ddi_get_name(dip); 1238 major = ddi_name_to_major(name); 1239 1240 mutex_enter(&apix_mutex); 1241 for (dvp = apix_dev_vector[major]; dvp != NULL; 1242 dvp = dvp->dv_next) { 1243 if (dvp->dv_dip == dip && dvp->dv_type == type) { 1244 if (inum == -1) 1245 inum = dvp->dv_inum; 1246 else 1247 inum = (dvp->dv_inum < inum) ? 1248 dvp->dv_inum : inum; 1249 } 1250 } 1251 mutex_exit(&apix_mutex); 1252 1253 return (inum); 1254 } 1255 1256 int 1257 apix_get_max_dev_inum(dev_info_t *dip, int type) 1258 { 1259 char *name; 1260 major_t major; 1261 apix_dev_vector_t *dvp; 1262 int inum = -1; 1263 1264 name = ddi_get_name(dip); 1265 major = ddi_name_to_major(name); 1266 1267 mutex_enter(&apix_mutex); 1268 for (dvp = apix_dev_vector[major]; dvp != NULL; 1269 dvp = dvp->dv_next) { 1270 if (dvp->dv_dip == dip && dvp->dv_type == type) { 1271 if (inum == -1) 1272 inum = dvp->dv_inum; 1273 else 1274 inum = (dvp->dv_inum > inum) ? 1275 dvp->dv_inum : inum; 1276 } 1277 } 1278 mutex_exit(&apix_mutex); 1279 1280 return (inum); 1281 } 1282 1283 /* 1284 * Major to cpu binding, for INTR_ROUND_ROBIN_WITH_AFFINITY cpu 1285 * binding policy 1286 */ 1287 1288 static uint32_t 1289 apix_get_dev_binding(dev_info_t *dip) 1290 { 1291 major_t major; 1292 char *name; 1293 uint32_t cpu = IRQ_UNINIT; 1294 1295 name = ddi_get_name(dip); 1296 major = ddi_name_to_major(name); 1297 if (major < devcnt) { 1298 mutex_enter(&apix_mutex); 1299 cpu = apix_major_to_cpu[major]; 1300 mutex_exit(&apix_mutex); 1301 } 1302 1303 return (cpu); 1304 } 1305 1306 static void 1307 apix_set_dev_binding(dev_info_t *dip, uint32_t cpu) 1308 { 1309 major_t major; 1310 char *name; 1311 1312 /* setup major to cpu mapping */ 1313 name = ddi_get_name(dip); 1314 major = ddi_name_to_major(name); 1315 if (apix_major_to_cpu[major] == IRQ_UNINIT) { 1316 mutex_enter(&apix_mutex); 1317 apix_major_to_cpu[major] = cpu; 1318 mutex_exit(&apix_mutex); 1319 } 1320 } 1321 1322 /* 1323 * return the cpu to which this intr should be bound. 1324 * Check properties or any other mechanism to see if user wants it 1325 * bound to a specific CPU. If so, return the cpu id with high bit set. 1326 * If not, use the policy to choose a cpu and return the id. 1327 */ 1328 uint32_t 1329 apix_bind_cpu(dev_info_t *dip) 1330 { 1331 int instance, instno, prop_len, bind_cpu, count; 1332 uint_t i, rc; 1333 major_t major; 1334 char *name, *drv_name, *prop_val, *cptr; 1335 char prop_name[32]; 1336 1337 lock_set(&apix_lock); 1338 1339 if (apic_intr_policy == INTR_LOWEST_PRIORITY) { 1340 cmn_err(CE_WARN, "apix: unsupported interrupt binding policy " 1341 "LOWEST PRIORITY, use ROUND ROBIN instead"); 1342 apic_intr_policy = INTR_ROUND_ROBIN; 1343 } 1344 1345 if (apic_nproc == 1) { 1346 lock_clear(&apix_lock); 1347 return (0); 1348 } 1349 1350 drv_name = NULL; 1351 rc = DDI_PROP_NOT_FOUND; 1352 major = (major_t)-1; 1353 if (dip != NULL) { 1354 name = ddi_get_name(dip); 1355 major = ddi_name_to_major(name); 1356 drv_name = ddi_major_to_name(major); 1357 instance = ddi_get_instance(dip); 1358 if (apic_intr_policy == INTR_ROUND_ROBIN_WITH_AFFINITY) { 1359 bind_cpu = apix_get_dev_binding(dip); 1360 if (bind_cpu != IRQ_UNINIT) { 1361 lock_clear(&apix_lock); 1362 return (bind_cpu); 1363 } 1364 } 1365 /* 1366 * search for "drvname"_intpt_bind_cpus property first, the 1367 * syntax of the property should be "a[,b,c,...]" where 1368 * instance 0 binds to cpu a, instance 1 binds to cpu b, 1369 * instance 3 binds to cpu c... 1370 * ddi_getlongprop() will search /option first, then / 1371 * if "drvname"_intpt_bind_cpus doesn't exist, then find 1372 * intpt_bind_cpus property. The syntax is the same, and 1373 * it applies to all the devices if its "drvname" specific 1374 * property doesn't exist 1375 */ 1376 (void) strcpy(prop_name, drv_name); 1377 (void) strcat(prop_name, "_intpt_bind_cpus"); 1378 rc = ddi_getlongprop(DDI_DEV_T_ANY, dip, 0, prop_name, 1379 (caddr_t)&prop_val, &prop_len); 1380 if (rc != DDI_PROP_SUCCESS) { 1381 rc = ddi_getlongprop(DDI_DEV_T_ANY, dip, 0, 1382 "intpt_bind_cpus", (caddr_t)&prop_val, &prop_len); 1383 } 1384 } 1385 if (rc == DDI_PROP_SUCCESS) { 1386 for (i = count = 0; i < (prop_len - 1); i++) 1387 if (prop_val[i] == ',') 1388 count++; 1389 if (prop_val[i-1] != ',') 1390 count++; 1391 /* 1392 * if somehow the binding instances defined in the 1393 * property are not enough for this instno., then 1394 * reuse the pattern for the next instance until 1395 * it reaches the requested instno 1396 */ 1397 instno = instance % count; 1398 i = 0; 1399 cptr = prop_val; 1400 while (i < instno) 1401 if (*cptr++ == ',') 1402 i++; 1403 bind_cpu = stoi(&cptr); 1404 kmem_free(prop_val, prop_len); 1405 /* if specific cpu is bogus, then default to cpu 0 */ 1406 if (bind_cpu >= apic_nproc) { 1407 cmn_err(CE_WARN, "apix: %s=%s: CPU %d not present", 1408 prop_name, prop_val, bind_cpu); 1409 bind_cpu = 0; 1410 } else { 1411 /* indicate that we are bound at user request */ 1412 bind_cpu |= IRQ_USER_BOUND; 1413 } 1414 /* 1415 * no need to check apic_cpus[].aci_status, if specific cpu is 1416 * not up, then post_cpu_start will handle it. 1417 */ 1418 } else { 1419 bind_cpu = apic_get_next_bind_cpu(); 1420 } 1421 1422 lock_clear(&apix_lock); 1423 1424 return ((uint32_t)bind_cpu); 1425 } 1426 1427 static boolean_t 1428 apix_is_cpu_enabled(processorid_t cpuid) 1429 { 1430 apic_cpus_info_t *cpu_infop; 1431 1432 cpu_infop = &apic_cpus[cpuid]; 1433 1434 if ((cpu_infop->aci_status & APIC_CPU_INTR_ENABLE) == 0) 1435 return (B_FALSE); 1436 1437 return (B_TRUE); 1438 } 1439 1440 /* 1441 * Must be called with apix_lock held. This function can be 1442 * called from above lock level by apix_intr_redistribute(). 1443 * 1444 * Arguments: 1445 * vecp : Vector to be rebound 1446 * tocpu : Target cpu. IRQ_UNINIT means target is vecp->v_cpuid. 1447 * count : Number of continuous vectors 1448 * 1449 * Return new vector being bound to 1450 */ 1451 apix_vector_t * 1452 apix_rebind(apix_vector_t *vecp, processorid_t newcpu, int count) 1453 { 1454 apix_vector_t *newp, *oldp; 1455 processorid_t oldcpu = vecp->v_cpuid; 1456 uchar_t newvec, oldvec = vecp->v_vector; 1457 int i; 1458 1459 ASSERT(LOCK_HELD(&apix_lock) && count > 0); 1460 1461 if (!apix_is_cpu_enabled(newcpu)) 1462 return (NULL); 1463 1464 if (vecp->v_cpuid == newcpu) /* rebind to the same cpu */ 1465 return (vecp); 1466 1467 APIX_ENTER_CPU_LOCK(oldcpu); 1468 APIX_ENTER_CPU_LOCK(newcpu); 1469 1470 /* allocate vector */ 1471 if (count == 1) 1472 newp = apix_alloc_vector_oncpu(newcpu, NULL, 0, vecp->v_type); 1473 else { 1474 ASSERT(vecp->v_type == APIX_TYPE_MSI); 1475 newp = apix_alloc_nvectors_oncpu(newcpu, NULL, 0, count, 1476 vecp->v_type); 1477 } 1478 if (newp == NULL) { 1479 APIX_LEAVE_CPU_LOCK(newcpu); 1480 APIX_LEAVE_CPU_LOCK(oldcpu); 1481 return (NULL); 1482 } 1483 1484 newvec = newp->v_vector; 1485 apix_dup_vectors(vecp, newp, count); 1486 1487 APIX_LEAVE_CPU_LOCK(newcpu); 1488 APIX_LEAVE_CPU_LOCK(oldcpu); 1489 1490 if (!DDI_INTR_IS_MSI_OR_MSIX(vecp->v_type)) { 1491 ASSERT(count == 1); 1492 if (apix_intx_rebind(vecp->v_inum, newcpu, newvec) != 0) { 1493 struct autovec *avp; 1494 int inum; 1495 1496 /* undo duplication */ 1497 APIX_ENTER_CPU_LOCK(oldcpu); 1498 APIX_ENTER_CPU_LOCK(newcpu); 1499 for (avp = newp->v_autovect; avp != NULL; 1500 avp = avp->av_link) { 1501 if (avp->av_dip != NULL) { 1502 inum = GET_INTR_INUM(avp->av_intr_id); 1503 apix_set_dev_map(vecp, avp->av_dip, 1504 inum); 1505 } 1506 apix_remove_av(newp, avp); 1507 } 1508 apix_cleanup_vector(newp); 1509 APIX_LEAVE_CPU_LOCK(newcpu); 1510 APIX_LEAVE_CPU_LOCK(oldcpu); 1511 APIC_VERBOSE(REBIND, (CE_CONT, "apix: rebind fixed " 1512 "interrupt 0x%x to cpu %d failed\n", 1513 vecp->v_inum, newcpu)); 1514 return (NULL); 1515 } 1516 1517 APIX_ENTER_CPU_LOCK(oldcpu); 1518 (void) apix_obsolete_vector(vecp); 1519 APIX_LEAVE_CPU_LOCK(oldcpu); 1520 APIC_VERBOSE(REBIND, (CE_CONT, "apix: rebind fixed interrupt" 1521 " 0x%x/0x%x to 0x%x/0x%x\n", 1522 oldcpu, oldvec, newcpu, newvec)); 1523 return (newp); 1524 } 1525 1526 for (i = 0; i < count; i++) { 1527 oldp = xv_vector(oldcpu, oldvec + i); 1528 newp = xv_vector(newcpu, newvec + i); 1529 1530 if (newp->v_share > 0) { 1531 APIX_SET_REBIND_INFO(oldp, newp); 1532 1533 apix_enable_vector(newp); 1534 1535 APIX_CLR_REBIND_INFO(); 1536 } 1537 1538 APIX_ENTER_CPU_LOCK(oldcpu); 1539 (void) apix_obsolete_vector(oldp); 1540 APIX_LEAVE_CPU_LOCK(oldcpu); 1541 } 1542 APIC_VERBOSE(REBIND, (CE_CONT, "apix: rebind vector 0x%x/0x%x " 1543 "to 0x%x/0x%x, count=%d\n", 1544 oldcpu, oldvec, newcpu, newvec, count)); 1545 1546 return (xv_vector(newcpu, newvec)); 1547 } 1548 1549 /* 1550 * Senarios include: 1551 * a. add_avintr() is called before irqp initialized (legacy) 1552 * b. irqp is initialized, vector is not allocated (fixed interrupts) 1553 * c. irqp is initialized, vector is allocated (shared interrupts) 1554 */ 1555 apix_vector_t * 1556 apix_alloc_intx(dev_info_t *dip, int inum, int irqno) 1557 { 1558 apic_irq_t *irqp; 1559 apix_vector_t *vecp; 1560 1561 /* 1562 * Allocate IRQ. Caller is later responsible for the 1563 * initialization 1564 */ 1565 mutex_enter(&airq_mutex); 1566 if ((irqp = apic_irq_table[irqno]) == NULL) { 1567 /* allocate irq */ 1568 irqp = kmem_zalloc(sizeof (apic_irq_t), KM_SLEEP); 1569 irqp->airq_mps_intr_index = FREE_INDEX; 1570 apic_irq_table[irqno] = irqp; 1571 } 1572 if (irqp->airq_mps_intr_index == FREE_INDEX) { 1573 irqp->airq_mps_intr_index = DEFAULT_INDEX; 1574 irqp->airq_cpu = IRQ_UNINIT; 1575 irqp->airq_origirq = (uchar_t)irqno; 1576 } 1577 1578 mutex_exit(&airq_mutex); 1579 1580 /* 1581 * allocate vector 1582 */ 1583 if (irqp->airq_cpu == IRQ_UNINIT) { 1584 uint32_t bindcpu, cpuid; 1585 1586 /* select cpu by system policy */ 1587 bindcpu = apix_bind_cpu(dip); 1588 cpuid = bindcpu & ~IRQ_USER_BOUND; 1589 1590 /* allocate vector */ 1591 APIX_ENTER_CPU_LOCK(cpuid); 1592 1593 if ((vecp = apix_alloc_vector_oncpu(bindcpu, dip, inum, 1594 APIX_TYPE_FIXED)) == NULL) { 1595 cmn_err(CE_WARN, "No interrupt vector for irq %x", 1596 irqno); 1597 APIX_LEAVE_CPU_LOCK(cpuid); 1598 return (NULL); 1599 } 1600 vecp->v_inum = irqno; 1601 vecp->v_flags |= APIX_VECT_MASKABLE; 1602 1603 apix_intx_set_vector(irqno, vecp->v_cpuid, vecp->v_vector); 1604 1605 APIX_LEAVE_CPU_LOCK(cpuid); 1606 } else { 1607 vecp = xv_vector(irqp->airq_cpu, irqp->airq_vector); 1608 ASSERT(!IS_VECT_FREE(vecp)); 1609 1610 if (dip != NULL) 1611 apix_set_dev_map(vecp, dip, inum); 1612 } 1613 1614 if ((dip != NULL) && 1615 (apic_intr_policy == INTR_ROUND_ROBIN_WITH_AFFINITY) && 1616 ((vecp->v_flags & APIX_VECT_USER_BOUND) == 0)) 1617 apix_set_dev_binding(dip, vecp->v_cpuid); 1618 1619 apix_dprint_vector(vecp, dip, 1); 1620 1621 return (vecp); 1622 } 1623 1624 int 1625 apix_alloc_msi(dev_info_t *dip, int inum, int count, int behavior) 1626 { 1627 int i, cap_ptr, rcount = count; 1628 apix_vector_t *vecp; 1629 processorid_t bindcpu, cpuid; 1630 ushort_t msi_ctrl; 1631 ddi_acc_handle_t handle; 1632 1633 DDI_INTR_IMPLDBG((CE_CONT, "apix_alloc_msi_vectors: dip=0x%p " 1634 "inum=0x%x count=0x%x behavior=%d\n", 1635 (void *)dip, inum, count, behavior)); 1636 1637 if (count > 1) { 1638 if (behavior == DDI_INTR_ALLOC_STRICT && 1639 apic_multi_msi_enable == 0) 1640 return (0); 1641 if (apic_multi_msi_enable == 0) 1642 count = 1; 1643 } 1644 1645 /* Check whether it supports per-vector masking */ 1646 cap_ptr = i_ddi_get_msi_msix_cap_ptr(dip); 1647 handle = i_ddi_get_pci_config_handle(dip); 1648 msi_ctrl = pci_config_get16(handle, cap_ptr + PCI_MSI_CTRL); 1649 1650 /* bind to cpu */ 1651 bindcpu = apix_bind_cpu(dip); 1652 cpuid = bindcpu & ~IRQ_USER_BOUND; 1653 1654 /* if not ISP2, then round it down */ 1655 if (!ISP2(rcount)) 1656 rcount = 1 << (highbit(rcount) - 1); 1657 1658 APIX_ENTER_CPU_LOCK(cpuid); 1659 for (vecp = NULL; rcount > 0; rcount >>= 1) { 1660 vecp = apix_alloc_nvectors_oncpu(bindcpu, dip, inum, rcount, 1661 APIX_TYPE_MSI); 1662 if (vecp != NULL || behavior == DDI_INTR_ALLOC_STRICT) 1663 break; 1664 } 1665 for (i = 0; vecp && i < rcount; i++) 1666 xv_vector(vecp->v_cpuid, vecp->v_vector + i)->v_flags |= 1667 (msi_ctrl & PCI_MSI_PVM_MASK) ? APIX_VECT_MASKABLE : 0; 1668 APIX_LEAVE_CPU_LOCK(cpuid); 1669 if (vecp == NULL) { 1670 APIC_VERBOSE(INTR, (CE_CONT, 1671 "apix_alloc_msi: no %d cont vectors found on cpu 0x%x\n", 1672 count, bindcpu)); 1673 return (0); 1674 } 1675 1676 /* major to cpu binding */ 1677 if ((apic_intr_policy == INTR_ROUND_ROBIN_WITH_AFFINITY) && 1678 ((vecp->v_flags & APIX_VECT_USER_BOUND) == 0)) 1679 apix_set_dev_binding(dip, vecp->v_cpuid); 1680 1681 apix_dprint_vector(vecp, dip, rcount); 1682 1683 return (rcount); 1684 } 1685 1686 int 1687 apix_alloc_msix(dev_info_t *dip, int inum, int count, int behavior) 1688 { 1689 apix_vector_t *vecp; 1690 processorid_t bindcpu, cpuid; 1691 int i; 1692 1693 for (i = 0; i < count; i++) { 1694 /* select cpu by system policy */ 1695 bindcpu = apix_bind_cpu(dip); 1696 cpuid = bindcpu & ~IRQ_USER_BOUND; 1697 1698 /* allocate vector */ 1699 APIX_ENTER_CPU_LOCK(cpuid); 1700 if ((vecp = apix_alloc_vector_oncpu(bindcpu, dip, inum + i, 1701 APIX_TYPE_MSIX)) == NULL) { 1702 APIX_LEAVE_CPU_LOCK(cpuid); 1703 APIC_VERBOSE(INTR, (CE_CONT, "apix_alloc_msix: " 1704 "allocate msix for device dip=%p, inum=%d on" 1705 " cpu %d failed", (void *)dip, inum + i, bindcpu)); 1706 break; 1707 } 1708 vecp->v_flags |= APIX_VECT_MASKABLE; 1709 APIX_LEAVE_CPU_LOCK(cpuid); 1710 1711 /* major to cpu mapping */ 1712 if ((i == 0) && 1713 (apic_intr_policy == INTR_ROUND_ROBIN_WITH_AFFINITY) && 1714 ((vecp->v_flags & APIX_VECT_USER_BOUND) == 0)) 1715 apix_set_dev_binding(dip, vecp->v_cpuid); 1716 1717 apix_dprint_vector(vecp, dip, 1); 1718 } 1719 1720 if (i < count && behavior == DDI_INTR_ALLOC_STRICT) { 1721 APIC_VERBOSE(INTR, (CE_WARN, "apix_alloc_msix: " 1722 "strictly allocate %d vectors failed, got %d\n", 1723 count, i)); 1724 apix_free_vectors(dip, inum, i, APIX_TYPE_MSIX); 1725 i = 0; 1726 } 1727 1728 return (i); 1729 } 1730 1731 /* 1732 * A rollback free for vectors allocated by apix_alloc_xxx(). 1733 */ 1734 void 1735 apix_free_vectors(dev_info_t *dip, int inum, int count, int type) 1736 { 1737 int i, cpuid; 1738 apix_vector_t *vecp; 1739 1740 DDI_INTR_IMPLDBG((CE_CONT, "apix_free_vectors: dip: %p inum: %x " 1741 "count: %x type: %x\n", 1742 (void *)dip, inum, count, type)); 1743 1744 lock_set(&apix_lock); 1745 1746 for (i = 0; i < count; i++, inum++) { 1747 if ((vecp = apix_get_dev_map(dip, inum, type)) == NULL) { 1748 lock_clear(&apix_lock); 1749 DDI_INTR_IMPLDBG((CE_CONT, "apix_free_vectors: " 1750 "dip=0x%p inum=0x%x type=0x%x apix_find_intr() " 1751 "failed\n", (void *)dip, inum, type)); 1752 continue; 1753 } 1754 1755 APIX_ENTER_CPU_LOCK(vecp->v_cpuid); 1756 cpuid = vecp->v_cpuid; 1757 1758 DDI_INTR_IMPLDBG((CE_CONT, "apix_free_vectors: " 1759 "dip=0x%p inum=0x%x type=0x%x vector 0x%x (share %d)\n", 1760 (void *)dip, inum, type, vecp->v_vector, vecp->v_share)); 1761 1762 /* tear down device interrupt to vector mapping */ 1763 apix_clear_dev_map(dip, inum, type); 1764 1765 if (vecp->v_type == APIX_TYPE_FIXED) { 1766 if (vecp->v_share > 0) { /* share IRQ line */ 1767 APIX_LEAVE_CPU_LOCK(cpuid); 1768 continue; 1769 } 1770 1771 /* Free apic_irq_table entry */ 1772 apix_intx_free(vecp->v_inum); 1773 } 1774 1775 /* free vector */ 1776 apix_cleanup_vector(vecp); 1777 1778 APIX_LEAVE_CPU_LOCK(cpuid); 1779 } 1780 1781 lock_clear(&apix_lock); 1782 } 1783 1784 /* 1785 * Must be called with apix_lock held 1786 */ 1787 apix_vector_t * 1788 apix_setup_io_intr(apix_vector_t *vecp) 1789 { 1790 processorid_t bindcpu; 1791 int ret; 1792 1793 ASSERT(LOCK_HELD(&apix_lock)); 1794 1795 /* 1796 * Interrupts are enabled on the CPU, programme IOAPIC RDT 1797 * entry or MSI/X address/data to enable the interrupt. 1798 */ 1799 if (apix_is_cpu_enabled(vecp->v_cpuid)) { 1800 apix_enable_vector(vecp); 1801 return (vecp); 1802 } 1803 1804 /* 1805 * CPU is not up or interrupts are disabled. Fall back to the 1806 * first avialable CPU. 1807 */ 1808 bindcpu = apic_find_cpu(APIC_CPU_INTR_ENABLE); 1809 1810 if (vecp->v_type == APIX_TYPE_MSI) 1811 return (apix_grp_set_cpu(vecp, bindcpu, &ret)); 1812 1813 return (apix_set_cpu(vecp, bindcpu, &ret)); 1814 } 1815 1816 /* 1817 * For interrupts which call add_avintr() before apic is initialized. 1818 * ioapix_setup_intr() will 1819 * - allocate vector 1820 * - copy over ISR 1821 */ 1822 static void 1823 ioapix_setup_intr(int irqno, iflag_t *flagp) 1824 { 1825 extern struct av_head autovect[]; 1826 apix_vector_t *vecp; 1827 apic_irq_t *irqp; 1828 uchar_t ioapicindex, ipin; 1829 ulong_t iflag; 1830 struct autovec *avp; 1831 1832 irqp = apic_irq_table[irqno]; 1833 ioapicindex = acpi_find_ioapic(irqno); 1834 ASSERT(ioapicindex != 0xFF); 1835 ipin = irqno - apic_io_vectbase[ioapicindex]; 1836 1837 if ((irqp != NULL) && (irqp->airq_mps_intr_index == ACPI_INDEX)) { 1838 ASSERT(irqp->airq_intin_no == ipin && 1839 irqp->airq_ioapicindex == ioapicindex); 1840 vecp = xv_vector(irqp->airq_cpu, irqp->airq_vector); 1841 ASSERT(!IS_VECT_FREE(vecp)); 1842 } else { 1843 vecp = apix_alloc_intx(NULL, 0, irqno); 1844 1845 irqp = apic_irq_table[irqno]; 1846 irqp->airq_mps_intr_index = ACPI_INDEX; 1847 irqp->airq_ioapicindex = ioapicindex; 1848 irqp->airq_intin_no = ipin; 1849 irqp->airq_iflag = *flagp; 1850 irqp->airq_share++; 1851 apic_record_rdt_entry(irqp, irqno); 1852 } 1853 1854 /* copy over autovect */ 1855 for (avp = autovect[irqno].avh_link; avp; avp = avp->av_link) 1856 apix_insert_av(vecp, avp->av_intr_id, avp->av_vector, 1857 avp->av_intarg1, avp->av_intarg2, avp->av_ticksp, 1858 avp->av_prilevel, avp->av_dip); 1859 1860 /* Program I/O APIC */ 1861 iflag = intr_clear(); 1862 lock_set(&apix_lock); 1863 1864 (void) apix_setup_io_intr(vecp); 1865 1866 lock_clear(&apix_lock); 1867 intr_restore(iflag); 1868 1869 APIC_VERBOSE_IOAPIC((CE_CONT, "apix: setup ioapic, irqno %x " 1870 "(ioapic %x, ipin %x) is bound to cpu %x, vector %x\n", 1871 irqno, ioapicindex, ipin, irqp->airq_cpu, irqp->airq_vector)); 1872 } 1873 1874 void 1875 ioapix_init_intr(int mask_apic) 1876 { 1877 int ioapicindex; 1878 int i, j; 1879 1880 /* mask interrupt vectors */ 1881 for (j = 0; j < apic_io_max && mask_apic; j++) { 1882 int intin_max; 1883 1884 ioapicindex = j; 1885 /* Bits 23-16 define the maximum redirection entries */ 1886 intin_max = (ioapic_read(ioapicindex, APIC_VERS_CMD) >> 16) 1887 & 0xff; 1888 for (i = 0; i <= intin_max; i++) 1889 ioapic_write(ioapicindex, APIC_RDT_CMD + 2 * i, 1890 AV_MASK); 1891 } 1892 1893 /* 1894 * Hack alert: deal with ACPI SCI interrupt chicken/egg here 1895 */ 1896 if (apic_sci_vect > 0) 1897 ioapix_setup_intr(apic_sci_vect, &apic_sci_flags); 1898 1899 /* 1900 * Hack alert: deal with ACPI HPET interrupt chicken/egg here. 1901 */ 1902 if (apic_hpet_vect > 0) 1903 ioapix_setup_intr(apic_hpet_vect, &apic_hpet_flags); 1904 } 1905