xref: /titanic_44/usr/src/uts/i86pc/io/apix/apix_intr.c (revision 7014882c6a3672fd0e5d60200af8643ae53c5928)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright (c) 2010, Oracle and/or its affiliates. All rights reserved.
24  */
25 
26 #include <sys/cpuvar.h>
27 #include <sys/cpu_event.h>
28 #include <sys/param.h>
29 #include <sys/cmn_err.h>
30 #include <sys/t_lock.h>
31 #include <sys/kmem.h>
32 #include <sys/machlock.h>
33 #include <sys/systm.h>
34 #include <sys/archsystm.h>
35 #include <sys/atomic.h>
36 #include <sys/sdt.h>
37 #include <sys/processor.h>
38 #include <sys/time.h>
39 #include <sys/psm.h>
40 #include <sys/smp_impldefs.h>
41 #include <sys/cram.h>
42 #include <sys/apic.h>
43 #include <sys/pit.h>
44 #include <sys/ddi.h>
45 #include <sys/sunddi.h>
46 #include <sys/ddi_impldefs.h>
47 #include <sys/pci.h>
48 #include <sys/promif.h>
49 #include <sys/x86_archext.h>
50 #include <sys/cpc_impl.h>
51 #include <sys/uadmin.h>
52 #include <sys/panic.h>
53 #include <sys/debug.h>
54 #include <sys/trap.h>
55 #include <sys/machsystm.h>
56 #include <sys/sysmacros.h>
57 #include <sys/rm_platter.h>
58 #include <sys/privregs.h>
59 #include <sys/note.h>
60 #include <sys/pci_intr_lib.h>
61 #include <sys/spl.h>
62 #include <sys/clock.h>
63 #include <sys/dditypes.h>
64 #include <sys/sunddi.h>
65 #include <sys/x_call.h>
66 #include <sys/reboot.h>
67 #include <vm/hat_i86.h>
68 #include <sys/stack.h>
69 #include <sys/apix.h>
70 
71 static void apix_post_hardint(int);
72 
73 /*
74  * Insert an vector into the tail of the interrupt pending list
75  */
76 static __inline__ void
77 apix_insert_pending_av(apix_impl_t *apixp, struct autovec *avp, int ipl)
78 {
79 	struct autovec **head = apixp->x_intr_head;
80 	struct autovec **tail = apixp->x_intr_tail;
81 
82 	avp->av_ipl_link = NULL;
83 	if (tail[ipl] == NULL) {
84 		head[ipl] = tail[ipl] = avp;
85 		return;
86 	}
87 
88 	tail[ipl]->av_ipl_link = avp;
89 	tail[ipl] = avp;
90 }
91 
92 /*
93  * Remove and return an vector from the head of hardware interrupt
94  * pending list.
95  */
96 static __inline__ struct autovec *
97 apix_remove_pending_av(apix_impl_t *apixp, int ipl)
98 {
99 	struct cpu *cpu = CPU;
100 	struct autovec **head = apixp->x_intr_head;
101 	struct autovec **tail = apixp->x_intr_tail;
102 	struct autovec *avp = head[ipl];
103 
104 	if (avp == NULL)
105 		return (NULL);
106 
107 	if (avp->av_vector != NULL && avp->av_prilevel < cpu->cpu_base_spl) {
108 		/*
109 		 * If there is blocked higher level interrupts, return
110 		 * NULL to quit handling of current IPL level.
111 		 */
112 		apixp->x_intr_pending |= (1 << avp->av_prilevel);
113 		return (NULL);
114 	}
115 
116 	avp->av_flags &= ~AV_PENTRY_PEND;
117 	avp->av_flags |= AV_PENTRY_ONPROC;
118 	head[ipl] = avp->av_ipl_link;
119 	avp->av_ipl_link = NULL;
120 
121 	if (head[ipl] == NULL)
122 		tail[ipl] = NULL;
123 
124 	return (avp);
125 }
126 
127 /*
128  * add_pending_hardint:
129  *
130  * Add hardware interrupts to the interrupt pending list.
131  */
132 static void
133 apix_add_pending_hardint(int vector)
134 {
135 	uint32_t cpuid = psm_get_cpu_id();
136 	apix_impl_t *apixp = apixs[cpuid];
137 	apix_vector_t *vecp = apixp->x_vectbl[vector];
138 	struct autovec *p, *prevp = NULL;
139 	int ipl;
140 
141 	/*
142 	 * The MSI interrupt not supporting per-vector masking could
143 	 * be triggered on a false vector as a result of rebinding
144 	 * operation cannot programme MSI address & data atomically.
145 	 * Add ISR of this interrupt to the pending list for such
146 	 * suspicious interrupt.
147 	 */
148 	APIX_DO_FAKE_INTR(cpuid, vector);
149 	if (vecp == NULL)
150 		return;
151 
152 	for (p = vecp->v_autovect; p != NULL; p = p->av_link) {
153 		if (p->av_vector == NULL)
154 			continue;	/* skip freed entry */
155 
156 		ipl = p->av_prilevel;
157 		prevp = p;
158 
159 		/* set pending at specified priority level */
160 		apixp->x_intr_pending |= (1 << ipl);
161 
162 		if (p->av_flags & AV_PENTRY_PEND)
163 			continue;	/* already in the pending list */
164 		p->av_flags |= AV_PENTRY_PEND;
165 
166 		/* insert into pending list by it original IPL */
167 		apix_insert_pending_av(apixp, p, ipl);
168 	}
169 
170 	/* last one of the linked list */
171 	if (prevp && ((prevp->av_flags & AV_PENTRY_LEVEL) != 0))
172 		prevp->av_flags |= (vector & AV_PENTRY_VECTMASK);
173 }
174 
175 /*
176  * Walk pending hardware interrupts at given priority level, invoking
177  * each interrupt handler as we go.
178  */
179 extern uint64_t intr_get_time(void);
180 
181 static void
182 apix_dispatch_pending_autovect(uint_t ipl)
183 {
184 	uint32_t cpuid = psm_get_cpu_id();
185 	apix_impl_t *apixp = apixs[cpuid];
186 	struct autovec *av;
187 
188 	while ((av = apix_remove_pending_av(apixp, ipl)) != NULL) {
189 		uint_t r;
190 		uint_t (*intr)() = av->av_vector;
191 		caddr_t arg1 = av->av_intarg1;
192 		caddr_t arg2 = av->av_intarg2;
193 		dev_info_t *dip = av->av_dip;
194 		uchar_t vector = av->av_flags & AV_PENTRY_VECTMASK;
195 
196 		if (intr == NULL)
197 			continue;
198 
199 		/* Don't enable interrupts during x-calls */
200 		if (ipl != XC_HI_PIL)
201 			sti();
202 
203 		DTRACE_PROBE4(interrupt__start, dev_info_t *, dip,
204 		    void *, intr, caddr_t, arg1, caddr_t, arg2);
205 		r = (*intr)(arg1, arg2);
206 		DTRACE_PROBE4(interrupt__complete, dev_info_t *, dip,
207 		    void *, intr, caddr_t, arg1, uint_t, r);
208 
209 		if (av->av_ticksp && av->av_prilevel <= LOCK_LEVEL)
210 			atomic_add_64(av->av_ticksp, intr_get_time());
211 
212 		cli();
213 
214 		if (vector) {
215 			if ((av->av_flags & AV_PENTRY_PEND) == 0)
216 				av->av_flags &= ~AV_PENTRY_VECTMASK;
217 
218 			apix_post_hardint(vector);
219 		}
220 
221 		/* mark it as idle */
222 		av->av_flags &= ~AV_PENTRY_ONPROC;
223 	}
224 }
225 
226 static caddr_t
227 apix_do_softint_prolog(struct cpu *cpu, uint_t pil, uint_t oldpil,
228     caddr_t stackptr)
229 {
230 	kthread_t *t, *volatile it;
231 	struct machcpu *mcpu = &cpu->cpu_m;
232 	hrtime_t now;
233 
234 	UNREFERENCED_1PARAMETER(oldpil);
235 	ASSERT(pil > mcpu->mcpu_pri && pil > cpu->cpu_base_spl);
236 
237 	atomic_and_32((uint32_t *)&mcpu->mcpu_softinfo.st_pending, ~(1 << pil));
238 
239 	mcpu->mcpu_pri = pil;
240 
241 	now = tsc_read();
242 
243 	/*
244 	 * Get set to run interrupt thread.
245 	 * There should always be an interrupt thread since we
246 	 * allocate one for each level on the CPU.
247 	 */
248 	it = cpu->cpu_intr_thread;
249 	ASSERT(it != NULL);
250 	cpu->cpu_intr_thread = it->t_link;
251 
252 	/* t_intr_start could be zero due to cpu_intr_swtch_enter. */
253 	t = cpu->cpu_thread;
254 	if ((t->t_flag & T_INTR_THREAD) && t->t_intr_start != 0) {
255 		hrtime_t intrtime = now - t->t_intr_start;
256 		mcpu->intrstat[pil][0] += intrtime;
257 		cpu->cpu_intracct[cpu->cpu_mstate] += intrtime;
258 		t->t_intr_start = 0;
259 	}
260 
261 	/*
262 	 * Note that the code in kcpc_overflow_intr -relies- on the
263 	 * ordering of events here - in particular that t->t_lwp of
264 	 * the interrupt thread is set to the pinned thread *before*
265 	 * curthread is changed.
266 	 */
267 	it->t_lwp = t->t_lwp;
268 	it->t_state = TS_ONPROC;
269 
270 	/*
271 	 * Push interrupted thread onto list from new thread.
272 	 * Set the new thread as the current one.
273 	 * Set interrupted thread's T_SP because if it is the idle thread,
274 	 * resume() may use that stack between threads.
275 	 */
276 
277 	ASSERT(SA((uintptr_t)stackptr) == (uintptr_t)stackptr);
278 	t->t_sp = (uintptr_t)stackptr;
279 
280 	it->t_intr = t;
281 	cpu->cpu_thread = it;
282 
283 	/*
284 	 * Set bit for this pil in CPU's interrupt active bitmask.
285 	 */
286 	ASSERT((cpu->cpu_intr_actv & (1 << pil)) == 0);
287 	cpu->cpu_intr_actv |= (1 << pil);
288 
289 	/*
290 	 * Initialize thread priority level from intr_pri
291 	 */
292 	it->t_pil = (uchar_t)pil;
293 	it->t_pri = (pri_t)pil + intr_pri;
294 	it->t_intr_start = now;
295 
296 	return (it->t_stk);
297 }
298 
299 static void
300 apix_do_softint_epilog(struct cpu *cpu, uint_t oldpil)
301 {
302 	struct machcpu *mcpu = &cpu->cpu_m;
303 	kthread_t *t, *it;
304 	uint_t pil, basespl;
305 	hrtime_t intrtime;
306 	hrtime_t now = tsc_read();
307 
308 	it = cpu->cpu_thread;
309 	pil = it->t_pil;
310 
311 	cpu->cpu_stats.sys.intr[pil - 1]++;
312 
313 	ASSERT(cpu->cpu_intr_actv & (1 << pil));
314 	cpu->cpu_intr_actv &= ~(1 << pil);
315 
316 	intrtime = now - it->t_intr_start;
317 	mcpu->intrstat[pil][0] += intrtime;
318 	cpu->cpu_intracct[cpu->cpu_mstate] += intrtime;
319 
320 	/*
321 	 * If there is still an interrupted thread underneath this one
322 	 * then the interrupt was never blocked and the return is
323 	 * fairly simple.  Otherwise it isn't.
324 	 */
325 	if ((t = it->t_intr) == NULL) {
326 		/*
327 		 * Put thread back on the interrupt thread list.
328 		 * This was an interrupt thread, so set CPU's base SPL.
329 		 */
330 		set_base_spl();
331 		/* mcpu->mcpu_pri = cpu->cpu_base_spl; */
332 
333 		it->t_state = TS_FREE;
334 		it->t_link = cpu->cpu_intr_thread;
335 		cpu->cpu_intr_thread = it;
336 		(void) splhigh();
337 		sti();
338 		swtch();
339 		/*NOTREACHED*/
340 		panic("dosoftint_epilog: swtch returned");
341 	}
342 	it->t_link = cpu->cpu_intr_thread;
343 	cpu->cpu_intr_thread = it;
344 	it->t_state = TS_FREE;
345 	cpu->cpu_thread = t;
346 	if (t->t_flag & T_INTR_THREAD)
347 		t->t_intr_start = now;
348 	basespl = cpu->cpu_base_spl;
349 	pil = MAX(oldpil, basespl);
350 	mcpu->mcpu_pri = pil;
351 }
352 
353 /*
354  * Dispatch a soft interrupt
355  */
356 static void
357 apix_dispatch_softint(uint_t oldpil, uint_t arg2)
358 {
359 	struct cpu *cpu = CPU;
360 
361 	UNREFERENCED_1PARAMETER(arg2);
362 
363 	sti();
364 	av_dispatch_softvect((int)cpu->cpu_thread->t_pil);
365 	cli();
366 
367 	/*
368 	 * Must run softint_epilog() on the interrupt thread stack, since
369 	 * there may not be a return from it if the interrupt thread blocked.
370 	 */
371 	apix_do_softint_epilog(cpu, oldpil);
372 }
373 
374 /*
375  * Deliver any softints the current interrupt priority allows.
376  * Called with interrupts disabled.
377  */
378 int
379 apix_do_softint(struct regs *regs)
380 {
381 	struct cpu *cpu = CPU;
382 	int oldipl;
383 	int newipl;
384 	volatile uint16_t pending;
385 	caddr_t newsp;
386 
387 	while ((pending = cpu->cpu_softinfo.st_pending) != 0) {
388 		newipl = bsrw_insn(pending);
389 		oldipl = cpu->cpu_pri;
390 		if (newipl <= oldipl || newipl <= cpu->cpu_base_spl)
391 			return (-1);
392 
393 		newsp = apix_do_softint_prolog(cpu, newipl, oldipl,
394 		    (caddr_t)regs);
395 		ASSERT(newsp != NULL);
396 		switch_sp_and_call(newsp, apix_dispatch_softint, oldipl, 0);
397 	}
398 
399 	return (0);
400 }
401 
402 static int
403 apix_hilevel_intr_prolog(struct cpu *cpu, uint_t pil, uint_t oldpil,
404     struct regs *rp)
405 {
406 	struct machcpu *mcpu = &cpu->cpu_m;
407 	hrtime_t intrtime;
408 	hrtime_t now = tsc_read();
409 	apix_impl_t *apixp = apixs[cpu->cpu_id];
410 	uint_t mask;
411 
412 	ASSERT(pil > mcpu->mcpu_pri && pil > cpu->cpu_base_spl);
413 
414 	if (pil == CBE_HIGH_PIL) {	/* 14 */
415 		cpu->cpu_profile_pil = oldpil;
416 		if (USERMODE(rp->r_cs)) {
417 			cpu->cpu_profile_pc = 0;
418 			cpu->cpu_profile_upc = rp->r_pc;
419 			cpu->cpu_cpcprofile_pc = 0;
420 			cpu->cpu_cpcprofile_upc = rp->r_pc;
421 		} else {
422 			cpu->cpu_profile_pc = rp->r_pc;
423 			cpu->cpu_profile_upc = 0;
424 			cpu->cpu_cpcprofile_pc = rp->r_pc;
425 			cpu->cpu_cpcprofile_upc = 0;
426 		}
427 	}
428 
429 	mcpu->mcpu_pri = pil;
430 
431 	mask = cpu->cpu_intr_actv & CPU_INTR_ACTV_HIGH_LEVEL_MASK;
432 	if (mask != 0) {
433 		int nestpil;
434 
435 		/*
436 		 * We have interrupted another high-level interrupt.
437 		 * Load starting timestamp, compute interval, update
438 		 * cumulative counter.
439 		 */
440 		nestpil = bsrw_insn((uint16_t)mask);
441 		intrtime = now -
442 		    mcpu->pil_high_start[nestpil - (LOCK_LEVEL + 1)];
443 		mcpu->intrstat[nestpil][0] += intrtime;
444 		cpu->cpu_intracct[cpu->cpu_mstate] += intrtime;
445 	} else {
446 		kthread_t *t = cpu->cpu_thread;
447 
448 		/*
449 		 * See if we are interrupting a low-level interrupt thread.
450 		 * If so, account for its time slice only if its time stamp
451 		 * is non-zero.
452 		 */
453 		if ((t->t_flag & T_INTR_THREAD) != 0 && t->t_intr_start != 0) {
454 			intrtime = now - t->t_intr_start;
455 			mcpu->intrstat[t->t_pil][0] += intrtime;
456 			cpu->cpu_intracct[cpu->cpu_mstate] += intrtime;
457 			t->t_intr_start = 0;
458 		}
459 	}
460 
461 	/* store starting timestamp in CPu structure for this IPL */
462 	mcpu->pil_high_start[pil - (LOCK_LEVEL + 1)] = now;
463 
464 	if (pil == 15) {
465 		/*
466 		 * To support reentrant level 15 interrupts, we maintain a
467 		 * recursion count in the top half of cpu_intr_actv.  Only
468 		 * when this count hits zero do we clear the PIL 15 bit from
469 		 * the lower half of cpu_intr_actv.
470 		 */
471 		uint16_t *refcntp = (uint16_t *)&cpu->cpu_intr_actv + 1;
472 		(*refcntp)++;
473 	}
474 
475 	cpu->cpu_intr_actv |= (1 << pil);
476 	/* clear pending ipl level bit */
477 	apixp->x_intr_pending &= ~(1 << pil);
478 
479 	return (mask);
480 }
481 
482 static int
483 apix_hilevel_intr_epilog(struct cpu *cpu, uint_t oldpil)
484 {
485 	struct machcpu *mcpu = &cpu->cpu_m;
486 	uint_t mask, pil;
487 	hrtime_t intrtime;
488 	hrtime_t now = tsc_read();
489 
490 	pil = mcpu->mcpu_pri;
491 	cpu->cpu_stats.sys.intr[pil - 1]++;
492 
493 	ASSERT(cpu->cpu_intr_actv & (1 << pil));
494 
495 	if (pil == 15) {
496 		/*
497 		 * To support reentrant level 15 interrupts, we maintain a
498 		 * recursion count in the top half of cpu_intr_actv.  Only
499 		 * when this count hits zero do we clear the PIL 15 bit from
500 		 * the lower half of cpu_intr_actv.
501 		 */
502 		uint16_t *refcntp = (uint16_t *)&cpu->cpu_intr_actv + 1;
503 
504 		ASSERT(*refcntp > 0);
505 
506 		if (--(*refcntp) == 0)
507 			cpu->cpu_intr_actv &= ~(1 << pil);
508 	} else {
509 		cpu->cpu_intr_actv &= ~(1 << pil);
510 	}
511 
512 	ASSERT(mcpu->pil_high_start[pil - (LOCK_LEVEL + 1)] != 0);
513 
514 	intrtime = now - mcpu->pil_high_start[pil - (LOCK_LEVEL + 1)];
515 	mcpu->intrstat[pil][0] += intrtime;
516 	cpu->cpu_intracct[cpu->cpu_mstate] += intrtime;
517 
518 	/*
519 	 * Check for lower-pil nested high-level interrupt beneath
520 	 * current one.  If so, place a starting timestamp in its
521 	 * pil_high_start entry.
522 	 */
523 	mask = cpu->cpu_intr_actv & CPU_INTR_ACTV_HIGH_LEVEL_MASK;
524 	if (mask != 0) {
525 		int nestpil;
526 
527 		/*
528 		 * find PIL of nested interrupt
529 		 */
530 		nestpil = bsrw_insn((uint16_t)mask);
531 		ASSERT(nestpil < pil);
532 		mcpu->pil_high_start[nestpil - (LOCK_LEVEL + 1)] = now;
533 		/*
534 		 * (Another high-level interrupt is active below this one,
535 		 * so there is no need to check for an interrupt
536 		 * thread.  That will be done by the lowest priority
537 		 * high-level interrupt active.)
538 		 */
539 	} else {
540 		/*
541 		 * Check to see if there is a low-level interrupt active.
542 		 * If so, place a starting timestamp in the thread
543 		 * structure.
544 		 */
545 		kthread_t *t = cpu->cpu_thread;
546 
547 		if (t->t_flag & T_INTR_THREAD)
548 			t->t_intr_start = now;
549 	}
550 
551 	mcpu->mcpu_pri = oldpil;
552 	if (pil < CBE_HIGH_PIL)
553 		(void) (*setlvlx)(oldpil, 0);
554 
555 	return (mask);
556 }
557 
558 /*
559  * Dispatch a hilevel interrupt (one above LOCK_LEVEL)
560  */
561 static void
562 apix_dispatch_pending_hilevel(uint_t ipl, uint_t arg2)
563 {
564 	UNREFERENCED_1PARAMETER(arg2);
565 
566 	apix_dispatch_pending_autovect(ipl);
567 }
568 
569 static __inline__ int
570 apix_do_pending_hilevel(struct cpu *cpu, struct regs *rp)
571 {
572 	volatile uint16_t pending;
573 	uint_t newipl, oldipl;
574 	caddr_t newsp;
575 
576 	while ((pending = HILEVEL_PENDING(cpu)) != 0) {
577 		newipl = bsrw_insn(pending);
578 		ASSERT(newipl > LOCK_LEVEL && newipl > cpu->cpu_base_spl);
579 		oldipl = cpu->cpu_pri;
580 		if (newipl <= oldipl)
581 			return (-1);
582 
583 		/*
584 		 * High priority interrupts run on this cpu's interrupt stack.
585 		 */
586 		if (apix_hilevel_intr_prolog(cpu, newipl, oldipl, rp) == 0) {
587 			newsp = cpu->cpu_intr_stack;
588 			switch_sp_and_call(newsp, apix_dispatch_pending_hilevel,
589 			    newipl, 0);
590 		} else {	/* already on the interrupt stack */
591 			apix_dispatch_pending_hilevel(newipl, 0);
592 		}
593 		(void) apix_hilevel_intr_epilog(cpu, oldipl);
594 	}
595 
596 	return (0);
597 }
598 
599 /*
600  * Get an interrupt thread and swith to it. It's called from do_interrupt().
601  * The IF flag is cleared and thus all maskable interrupts are blocked at
602  * the time of calling.
603  */
604 static caddr_t
605 apix_intr_thread_prolog(struct cpu *cpu, uint_t pil, caddr_t stackptr)
606 {
607 	apix_impl_t *apixp = apixs[cpu->cpu_id];
608 	struct machcpu *mcpu = &cpu->cpu_m;
609 	hrtime_t now = tsc_read();
610 	kthread_t *t, *volatile it;
611 
612 	ASSERT(pil > mcpu->mcpu_pri && pil > cpu->cpu_base_spl);
613 
614 	apixp->x_intr_pending &= ~(1 << pil);
615 	ASSERT((cpu->cpu_intr_actv & (1 << pil)) == 0);
616 	cpu->cpu_intr_actv |= (1 << pil);
617 	mcpu->mcpu_pri = pil;
618 
619 	/*
620 	 * Get set to run interrupt thread.
621 	 * There should always be an interrupt thread since we
622 	 * allocate one for each level on the CPU.
623 	 */
624 	/* t_intr_start could be zero due to cpu_intr_swtch_enter. */
625 	t = cpu->cpu_thread;
626 	if ((t->t_flag & T_INTR_THREAD) && t->t_intr_start != 0) {
627 		hrtime_t intrtime = now - t->t_intr_start;
628 		mcpu->intrstat[pil][0] += intrtime;
629 		cpu->cpu_intracct[cpu->cpu_mstate] += intrtime;
630 		t->t_intr_start = 0;
631 	}
632 
633 	/*
634 	 * Push interrupted thread onto list from new thread.
635 	 * Set the new thread as the current one.
636 	 * Set interrupted thread's T_SP because if it is the idle thread,
637 	 * resume() may use that stack between threads.
638 	 */
639 
640 	ASSERT(SA((uintptr_t)stackptr) == (uintptr_t)stackptr);
641 
642 	t->t_sp = (uintptr_t)stackptr;	/* mark stack in curthread for resume */
643 
644 	/*
645 	 * Note that the code in kcpc_overflow_intr -relies- on the
646 	 * ordering of events here - in particular that t->t_lwp of
647 	 * the interrupt thread is set to the pinned thread *before*
648 	 * curthread is changed.
649 	 */
650 	it = cpu->cpu_intr_thread;
651 	cpu->cpu_intr_thread = it->t_link;
652 	it->t_intr = t;
653 	it->t_lwp = t->t_lwp;
654 
655 	/*
656 	 * (threads on the interrupt thread free list could have state
657 	 * preset to TS_ONPROC, but it helps in debugging if
658 	 * they're TS_FREE.)
659 	 */
660 	it->t_state = TS_ONPROC;
661 
662 	cpu->cpu_thread = it;
663 
664 	/*
665 	 * Initialize thread priority level from intr_pri
666 	 */
667 	it->t_pil = (uchar_t)pil;
668 	it->t_pri = (pri_t)pil + intr_pri;
669 	it->t_intr_start = now;
670 
671 	return (it->t_stk);
672 }
673 
674 static void
675 apix_intr_thread_epilog(struct cpu *cpu, uint_t oldpil)
676 {
677 	struct machcpu *mcpu = &cpu->cpu_m;
678 	kthread_t *t, *it = cpu->cpu_thread;
679 	uint_t pil, basespl;
680 	hrtime_t intrtime;
681 	hrtime_t now = tsc_read();
682 
683 	pil = it->t_pil;
684 	cpu->cpu_stats.sys.intr[pil - 1]++;
685 
686 	ASSERT(cpu->cpu_intr_actv & (1 << pil));
687 	cpu->cpu_intr_actv &= ~(1 << pil);
688 
689 	ASSERT(it->t_intr_start != 0);
690 	intrtime = now - it->t_intr_start;
691 	mcpu->intrstat[pil][0] += intrtime;
692 	cpu->cpu_intracct[cpu->cpu_mstate] += intrtime;
693 
694 	/*
695 	 * If there is still an interrupted thread underneath this one
696 	 * then the interrupt was never blocked and the return is
697 	 * fairly simple.  Otherwise it isn't.
698 	 */
699 	if ((t = it->t_intr) == NULL) {
700 		/*
701 		 * The interrupted thread is no longer pinned underneath
702 		 * the interrupt thread.  This means the interrupt must
703 		 * have blocked, and the interrupted thread has been
704 		 * unpinned, and has probably been running around the
705 		 * system for a while.
706 		 *
707 		 * Since there is no longer a thread under this one, put
708 		 * this interrupt thread back on the CPU's free list and
709 		 * resume the idle thread which will dispatch the next
710 		 * thread to run.
711 		 */
712 		cpu->cpu_stats.sys.intrblk++;
713 
714 		/*
715 		 * Put thread back on the interrupt thread list.
716 		 * This was an interrupt thread, so set CPU's base SPL.
717 		 */
718 		set_base_spl();
719 		basespl = cpu->cpu_base_spl;
720 		mcpu->mcpu_pri = basespl;
721 		(*setlvlx)(basespl, 0);
722 
723 		it->t_state = TS_FREE;
724 		/*
725 		 * Return interrupt thread to pool
726 		 */
727 		it->t_link = cpu->cpu_intr_thread;
728 		cpu->cpu_intr_thread = it;
729 
730 		(void) splhigh();
731 		sti();
732 		swtch();
733 		/*NOTREACHED*/
734 		panic("dosoftint_epilog: swtch returned");
735 	}
736 
737 	/*
738 	 * Return interrupt thread to the pool
739 	 */
740 	it->t_link = cpu->cpu_intr_thread;
741 	cpu->cpu_intr_thread = it;
742 	it->t_state = TS_FREE;
743 
744 	cpu->cpu_thread = t;
745 	if (t->t_flag & T_INTR_THREAD)
746 		t->t_intr_start = now;
747 	basespl = cpu->cpu_base_spl;
748 	mcpu->mcpu_pri = MAX(oldpil, basespl);
749 	(*setlvlx)(mcpu->mcpu_pri, 0);
750 }
751 
752 
753 static void
754 apix_dispatch_pending_hardint(uint_t oldpil, uint_t arg2)
755 {
756 	struct cpu *cpu = CPU;
757 
758 	UNREFERENCED_1PARAMETER(arg2);
759 
760 	apix_dispatch_pending_autovect((int)cpu->cpu_thread->t_pil);
761 
762 	/*
763 	 * Must run intr_thread_epilog() on the interrupt thread stack, since
764 	 * there may not be a return from it if the interrupt thread blocked.
765 	 */
766 	apix_intr_thread_epilog(cpu, oldpil);
767 }
768 
769 static __inline__ int
770 apix_do_pending_hardint(struct cpu *cpu, struct regs *rp)
771 {
772 	volatile uint16_t pending;
773 	uint_t newipl, oldipl;
774 	caddr_t newsp;
775 
776 	while ((pending = LOWLEVEL_PENDING(cpu)) != 0) {
777 		newipl = bsrw_insn(pending);
778 		ASSERT(newipl <= LOCK_LEVEL);
779 		oldipl = cpu->cpu_pri;
780 		if (newipl <= oldipl || newipl <= cpu->cpu_base_spl)
781 			return (-1);
782 
783 		/*
784 		 * Run this interrupt in a separate thread.
785 		 */
786 		newsp = apix_intr_thread_prolog(cpu, newipl, (caddr_t)rp);
787 		ASSERT(newsp != NULL);
788 		switch_sp_and_call(newsp, apix_dispatch_pending_hardint,
789 		    oldipl, 0);
790 	}
791 
792 	return (0);
793 }
794 
795 /*
796  * Unmask level triggered interrupts
797  */
798 static void
799 apix_post_hardint(int vector)
800 {
801 	apix_vector_t *vecp = xv_vector(psm_get_cpu_id(), vector);
802 	int irqno = vecp->v_inum;
803 
804 	ASSERT(vecp->v_type == APIX_TYPE_FIXED && apic_level_intr[irqno]);
805 
806 	apix_level_intr_post_dispatch(irqno);
807 }
808 
809 static void
810 apix_dispatch_by_vector(uint_t vector)
811 {
812 	struct cpu *cpu = CPU;
813 	apix_vector_t *vecp = xv_vector(cpu->cpu_id, vector);
814 	struct autovec *avp;
815 	uint_t r, (*intr)();
816 	caddr_t arg1, arg2;
817 	dev_info_t *dip;
818 
819 	if (vecp == NULL ||
820 	    (avp = vecp->v_autovect) == NULL || avp->av_vector == NULL)
821 		return;
822 
823 	avp->av_flags |= AV_PENTRY_ONPROC;
824 	intr = avp->av_vector;
825 	arg1 = avp->av_intarg1;
826 	arg2 = avp->av_intarg2;
827 	dip = avp->av_dip;
828 
829 	if (avp->av_prilevel != XC_HI_PIL)
830 		sti();
831 
832 	DTRACE_PROBE4(interrupt__start, dev_info_t *, dip,
833 	    void *, intr, caddr_t, arg1, caddr_t, arg2);
834 	r = (*intr)(arg1, arg2);
835 	DTRACE_PROBE4(interrupt__complete, dev_info_t *, dip,
836 	    void *, intr, caddr_t, arg1, uint_t, r);
837 
838 	cli();
839 	avp->av_flags &= ~AV_PENTRY_ONPROC;
840 }
841 
842 
843 static void
844 apix_dispatch_hilevel(uint_t vector, uint_t arg2)
845 {
846 	UNREFERENCED_1PARAMETER(arg2);
847 
848 	apix_dispatch_by_vector(vector);
849 }
850 
851 static void
852 apix_dispatch_lowlevel(uint_t vector, uint_t oldipl)
853 {
854 	struct cpu *cpu = CPU;
855 
856 	apix_dispatch_by_vector(vector);
857 
858 	/*
859 	 * Must run intr_thread_epilog() on the interrupt thread stack, since
860 	 * there may not be a return from it if the interrupt thread blocked.
861 	 */
862 	apix_intr_thread_epilog(cpu, oldipl);
863 }
864 
865 void
866 apix_do_interrupt(struct regs *rp, trap_trace_rec_t *ttp)
867 {
868 	struct cpu *cpu = CPU;
869 	int vector = rp->r_trapno, newipl, oldipl = cpu->cpu_pri, ret;
870 	apix_vector_t *vecp = NULL;
871 
872 #ifdef TRAPTRACE
873 	ttp->ttr_marker = TT_INTERRUPT;
874 	ttp->ttr_cpuid = cpu->cpu_id;
875 	ttp->ttr_ipl = 0xff;
876 	ttp->ttr_pri = (uchar_t)oldipl;
877 	ttp->ttr_spl = cpu->cpu_base_spl;
878 	ttp->ttr_vector = 0xff;
879 #endif	/* TRAPTRACE */
880 
881 	cpu_idle_exit(CPU_IDLE_CB_FLAG_INTR);
882 
883 	++*(uint16_t *)&cpu->cpu_m.mcpu_istamp;
884 
885 	/*
886 	 * If it's a softint go do it now.
887 	 */
888 	if (rp->r_trapno == T_SOFTINT) {
889 		/*
890 		 * It might be the case that when an interrupt is triggered,
891 		 * the spl is raised to high by splhigh(). Later when do_splx()
892 		 * is called to restore the spl, both hardware and software
893 		 * interrupt pending flags are check and an SOFTINT is faked
894 		 * accordingly.
895 		 */
896 		(void) apix_do_pending_hilevel(cpu, rp);
897 		(void) apix_do_pending_hardint(cpu, rp);
898 		(void) apix_do_softint(rp);
899 		ASSERT(!interrupts_enabled());
900 #ifdef TRAPTRACE
901 	ttp->ttr_vector = T_SOFTINT;
902 #endif
903 		return;
904 	}
905 
906 	/*
907 	 * Raise the interrupt priority. Send EOI to local APIC
908 	 */
909 	newipl = (*setlvl)(oldipl, (int *)&rp->r_trapno);
910 #ifdef TRAPTRACE
911 	ttp->ttr_ipl = (uchar_t)newipl;
912 #endif	/* TRAPTRACE */
913 
914 	/*
915 	 * Bail if it is a spurious interrupt
916 	 */
917 	if (newipl == -1)
918 		return;
919 
920 	vector = rp->r_trapno;
921 	vecp = xv_vector(cpu->cpu_id, vector);
922 #ifdef TRAPTRACE
923 	ttp->ttr_vector = (short)vector;
924 #endif	/* TRAPTRACE */
925 
926 	/*
927 	 * Direct dispatch for IPI, MSI, MSI-X
928 	 */
929 	if (vecp && vecp->v_type != APIX_TYPE_FIXED &&
930 	    newipl > MAX(oldipl, cpu->cpu_base_spl)) {
931 		caddr_t newsp;
932 
933 		if (newipl > LOCK_LEVEL) {
934 			if (apix_hilevel_intr_prolog(cpu, newipl, oldipl, rp)
935 			    == 0) {
936 				newsp = cpu->cpu_intr_stack;
937 				switch_sp_and_call(newsp, apix_dispatch_hilevel,
938 				    vector, 0);
939 			} else {
940 				apix_dispatch_hilevel(vector, 0);
941 			}
942 			(void) apix_hilevel_intr_epilog(cpu, oldipl);
943 		} else {
944 			newsp = apix_intr_thread_prolog(cpu, newipl,
945 			    (caddr_t)rp);
946 			switch_sp_and_call(newsp, apix_dispatch_lowlevel,
947 			    vector, oldipl);
948 		}
949 	} else {
950 		/* Add to per-pil pending queue */
951 		apix_add_pending_hardint(vector);
952 		if (newipl <= MAX(oldipl, cpu->cpu_base_spl) ||
953 		    !apixs[cpu->cpu_id]->x_intr_pending)
954 			return;
955 	}
956 
957 	if (apix_do_pending_hilevel(cpu, rp) < 0)
958 		return;
959 
960 	do {
961 		ret = apix_do_pending_hardint(cpu, rp);
962 
963 		/*
964 		 * Deliver any pending soft interrupts.
965 		 */
966 		(void) apix_do_softint(rp);
967 	} while (!ret && LOWLEVEL_PENDING(cpu));
968 }
969