xref: /freebsd/sys/x86/x86/cpu_machdep.c (revision a212f56d10d04508849cd817d2c595dd61107890)
1dfe7b3bfSKonstantin Belousov /*-
2dfe7b3bfSKonstantin Belousov  * Copyright (c) 2003 Peter Wemm.
3dfe7b3bfSKonstantin Belousov  * Copyright (c) 1992 Terrence R. Lambert.
4dfe7b3bfSKonstantin Belousov  * Copyright (c) 1982, 1987, 1990 The Regents of the University of California.
5dfe7b3bfSKonstantin Belousov  * All rights reserved.
6dfe7b3bfSKonstantin Belousov  *
7dfe7b3bfSKonstantin Belousov  * This code is derived from software contributed to Berkeley by
8dfe7b3bfSKonstantin Belousov  * William Jolitz.
9dfe7b3bfSKonstantin Belousov  *
10dfe7b3bfSKonstantin Belousov  * Redistribution and use in source and binary forms, with or without
11dfe7b3bfSKonstantin Belousov  * modification, are permitted provided that the following conditions
12dfe7b3bfSKonstantin Belousov  * are met:
13dfe7b3bfSKonstantin Belousov  * 1. Redistributions of source code must retain the above copyright
14dfe7b3bfSKonstantin Belousov  *    notice, this list of conditions and the following disclaimer.
15dfe7b3bfSKonstantin Belousov  * 2. Redistributions in binary form must reproduce the above copyright
16dfe7b3bfSKonstantin Belousov  *    notice, this list of conditions and the following disclaimer in the
17dfe7b3bfSKonstantin Belousov  *    documentation and/or other materials provided with the distribution.
18dfe7b3bfSKonstantin Belousov  * 3. All advertising materials mentioning features or use of this software
19dfe7b3bfSKonstantin Belousov  *    must display the following acknowledgement:
20dfe7b3bfSKonstantin Belousov  *	This product includes software developed by the University of
21dfe7b3bfSKonstantin Belousov  *	California, Berkeley and its contributors.
22dfe7b3bfSKonstantin Belousov  * 4. Neither the name of the University nor the names of its contributors
23dfe7b3bfSKonstantin Belousov  *    may be used to endorse or promote products derived from this software
24dfe7b3bfSKonstantin Belousov  *    without specific prior written permission.
25dfe7b3bfSKonstantin Belousov  *
26dfe7b3bfSKonstantin Belousov  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
27dfe7b3bfSKonstantin Belousov  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
28dfe7b3bfSKonstantin Belousov  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
29dfe7b3bfSKonstantin Belousov  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
30dfe7b3bfSKonstantin Belousov  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
31dfe7b3bfSKonstantin Belousov  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
32dfe7b3bfSKonstantin Belousov  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
33dfe7b3bfSKonstantin Belousov  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
34dfe7b3bfSKonstantin Belousov  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
35dfe7b3bfSKonstantin Belousov  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
36dfe7b3bfSKonstantin Belousov  * SUCH DAMAGE.
37dfe7b3bfSKonstantin Belousov  *
38dfe7b3bfSKonstantin Belousov  *	from: @(#)machdep.c	7.4 (Berkeley) 6/3/91
39dfe7b3bfSKonstantin Belousov  */
40dfe7b3bfSKonstantin Belousov 
41dfe7b3bfSKonstantin Belousov #include <sys/cdefs.h>
42dfe7b3bfSKonstantin Belousov __FBSDID("$FreeBSD$");
43dfe7b3bfSKonstantin Belousov 
447705dd4dSKonstantin Belousov #include "opt_acpi.h"
45dfe7b3bfSKonstantin Belousov #include "opt_atpic.h"
46dfe7b3bfSKonstantin Belousov #include "opt_cpu.h"
47dfe7b3bfSKonstantin Belousov #include "opt_ddb.h"
48dfe7b3bfSKonstantin Belousov #include "opt_inet.h"
49dfe7b3bfSKonstantin Belousov #include "opt_isa.h"
50835c2787SKonstantin Belousov #include "opt_kdb.h"
51dfe7b3bfSKonstantin Belousov #include "opt_kstack_pages.h"
52dfe7b3bfSKonstantin Belousov #include "opt_maxmem.h"
53dfe7b3bfSKonstantin Belousov #include "opt_mp_watchdog.h"
54dfe7b3bfSKonstantin Belousov #include "opt_platform.h"
55dfe7b3bfSKonstantin Belousov #ifdef __i386__
56dfe7b3bfSKonstantin Belousov #include "opt_apic.h"
57dfe7b3bfSKonstantin Belousov #endif
58dfe7b3bfSKonstantin Belousov 
59dfe7b3bfSKonstantin Belousov #include <sys/param.h>
60dfe7b3bfSKonstantin Belousov #include <sys/proc.h>
61dfe7b3bfSKonstantin Belousov #include <sys/systm.h>
62dfe7b3bfSKonstantin Belousov #include <sys/bus.h>
63dfe7b3bfSKonstantin Belousov #include <sys/cpu.h>
647355a02bSKonstantin Belousov #include <sys/domainset.h>
65dfe7b3bfSKonstantin Belousov #include <sys/kdb.h>
66dfe7b3bfSKonstantin Belousov #include <sys/kernel.h>
67dfe7b3bfSKonstantin Belousov #include <sys/ktr.h>
68dfe7b3bfSKonstantin Belousov #include <sys/lock.h>
69dfe7b3bfSKonstantin Belousov #include <sys/malloc.h>
70dfe7b3bfSKonstantin Belousov #include <sys/mutex.h>
71dfe7b3bfSKonstantin Belousov #include <sys/pcpu.h>
72dfe7b3bfSKonstantin Belousov #include <sys/rwlock.h>
73dfe7b3bfSKonstantin Belousov #include <sys/sched.h>
74dfe7b3bfSKonstantin Belousov #include <sys/smp.h>
75dfe7b3bfSKonstantin Belousov #include <sys/sysctl.h>
76dfe7b3bfSKonstantin Belousov 
77dfe7b3bfSKonstantin Belousov #include <machine/clock.h>
78dfe7b3bfSKonstantin Belousov #include <machine/cpu.h>
79dfe7b3bfSKonstantin Belousov #include <machine/cputypes.h>
80dfe7b3bfSKonstantin Belousov #include <machine/specialreg.h>
81dfe7b3bfSKonstantin Belousov #include <machine/md_var.h>
82dfe7b3bfSKonstantin Belousov #include <machine/mp_watchdog.h>
83dfe7b3bfSKonstantin Belousov #include <machine/tss.h>
84dfe7b3bfSKonstantin Belousov #ifdef SMP
85dfe7b3bfSKonstantin Belousov #include <machine/smp.h>
86dfe7b3bfSKonstantin Belousov #endif
873da25bdbSAndriy Gapon #ifdef CPU_ELAN
883da25bdbSAndriy Gapon #include <machine/elan_mmcr.h>
893da25bdbSAndriy Gapon #endif
90b57a73f8SKonstantin Belousov #include <x86/acpica_machdep.h>
91dfe7b3bfSKonstantin Belousov 
92dfe7b3bfSKonstantin Belousov #include <vm/vm.h>
93dfe7b3bfSKonstantin Belousov #include <vm/vm_extern.h>
94dfe7b3bfSKonstantin Belousov #include <vm/vm_kern.h>
95dfe7b3bfSKonstantin Belousov #include <vm/vm_page.h>
96dfe7b3bfSKonstantin Belousov #include <vm/vm_map.h>
97dfe7b3bfSKonstantin Belousov #include <vm/vm_object.h>
98dfe7b3bfSKonstantin Belousov #include <vm/vm_pager.h>
99dfe7b3bfSKonstantin Belousov #include <vm/vm_param.h>
100dfe7b3bfSKonstantin Belousov 
1018428d0f1SAndriy Gapon #include <isa/isareg.h>
1028428d0f1SAndriy Gapon 
1037705dd4dSKonstantin Belousov #include <contrib/dev/acpica/include/acpi.h>
1047705dd4dSKonstantin Belousov 
105d9e8bbb6SKonstantin Belousov #define	STATE_RUNNING	0x0
106d9e8bbb6SKonstantin Belousov #define	STATE_MWAIT	0x1
107d9e8bbb6SKonstantin Belousov #define	STATE_SLEEPING	0x2
108d9e8bbb6SKonstantin Belousov 
1098428d0f1SAndriy Gapon #ifdef SMP
1108428d0f1SAndriy Gapon static u_int	cpu_reset_proxyid;
1118428d0f1SAndriy Gapon static volatile u_int	cpu_reset_proxy_active;
1128428d0f1SAndriy Gapon #endif
1138428d0f1SAndriy Gapon 
114a2495c36SRoger Pau Monné char bootmethod[16];
115a2495c36SRoger Pau Monné SYSCTL_STRING(_machdep, OID_AUTO, bootmethod, CTLFLAG_RD, bootmethod, 0,
116a2495c36SRoger Pau Monné     "System firmware boot method");
117a2495c36SRoger Pau Monné 
118fa83f689SKonstantin Belousov struct msr_op_arg {
119fa83f689SKonstantin Belousov 	u_int msr;
120fa83f689SKonstantin Belousov 	int op;
121fa83f689SKonstantin Belousov 	uint64_t arg1;
122fa83f689SKonstantin Belousov };
123fa83f689SKonstantin Belousov 
124fa83f689SKonstantin Belousov static void
125fa83f689SKonstantin Belousov x86_msr_op_one(void *argp)
126fa83f689SKonstantin Belousov {
127fa83f689SKonstantin Belousov 	struct msr_op_arg *a;
128fa83f689SKonstantin Belousov 	uint64_t v;
129fa83f689SKonstantin Belousov 
130fa83f689SKonstantin Belousov 	a = argp;
131fa83f689SKonstantin Belousov 	switch (a->op) {
132fa83f689SKonstantin Belousov 	case MSR_OP_ANDNOT:
133fa83f689SKonstantin Belousov 		v = rdmsr(a->msr);
134fa83f689SKonstantin Belousov 		v &= ~a->arg1;
135fa83f689SKonstantin Belousov 		wrmsr(a->msr, v);
136fa83f689SKonstantin Belousov 		break;
137fa83f689SKonstantin Belousov 	case MSR_OP_OR:
138fa83f689SKonstantin Belousov 		v = rdmsr(a->msr);
139fa83f689SKonstantin Belousov 		v |= a->arg1;
140fa83f689SKonstantin Belousov 		wrmsr(a->msr, v);
141fa83f689SKonstantin Belousov 		break;
142fa83f689SKonstantin Belousov 	case MSR_OP_WRITE:
143fa83f689SKonstantin Belousov 		wrmsr(a->msr, a->arg1);
144fa83f689SKonstantin Belousov 		break;
145fa83f689SKonstantin Belousov 	}
146fa83f689SKonstantin Belousov }
147fa83f689SKonstantin Belousov 
148fa83f689SKonstantin Belousov #define	MSR_OP_EXMODE_MASK	0xf0000000
149fa83f689SKonstantin Belousov #define	MSR_OP_OP_MASK		0x000000ff
150fa83f689SKonstantin Belousov 
151fa83f689SKonstantin Belousov void
152fa83f689SKonstantin Belousov x86_msr_op(u_int msr, u_int op, uint64_t arg1)
153fa83f689SKonstantin Belousov {
154fa83f689SKonstantin Belousov 	struct thread *td;
155fa83f689SKonstantin Belousov 	struct msr_op_arg a;
156fa83f689SKonstantin Belousov 	u_int exmode;
157fa83f689SKonstantin Belousov 	int bound_cpu, i, is_bound;
158fa83f689SKonstantin Belousov 
159fa83f689SKonstantin Belousov 	a.op = op & MSR_OP_OP_MASK;
160fa83f689SKonstantin Belousov 	MPASS(a.op == MSR_OP_ANDNOT || a.op == MSR_OP_OR ||
161fa83f689SKonstantin Belousov 	    a.op == MSR_OP_WRITE);
162fa83f689SKonstantin Belousov 	exmode = op & MSR_OP_EXMODE_MASK;
163fa83f689SKonstantin Belousov 	MPASS(exmode == MSR_OP_LOCAL || exmode == MSR_OP_SCHED ||
164fa83f689SKonstantin Belousov 	    exmode == MSR_OP_RENDEZVOUS);
165fa83f689SKonstantin Belousov 	a.msr = msr;
166fa83f689SKonstantin Belousov 	a.arg1 = arg1;
167fa83f689SKonstantin Belousov 	switch (exmode) {
168fa83f689SKonstantin Belousov 	case MSR_OP_LOCAL:
169fa83f689SKonstantin Belousov 		x86_msr_op_one(&a);
170fa83f689SKonstantin Belousov 		break;
171fa83f689SKonstantin Belousov 	case MSR_OP_SCHED:
172fa83f689SKonstantin Belousov 		td = curthread;
173fa83f689SKonstantin Belousov 		thread_lock(td);
174fa83f689SKonstantin Belousov 		is_bound = sched_is_bound(td);
175fa83f689SKonstantin Belousov 		bound_cpu = td->td_oncpu;
176fa83f689SKonstantin Belousov 		CPU_FOREACH(i) {
177fa83f689SKonstantin Belousov 			sched_bind(td, i);
178fa83f689SKonstantin Belousov 			x86_msr_op_one(&a);
179fa83f689SKonstantin Belousov 		}
180fa83f689SKonstantin Belousov 		if (is_bound)
181fa83f689SKonstantin Belousov 			sched_bind(td, bound_cpu);
182fa83f689SKonstantin Belousov 		else
183fa83f689SKonstantin Belousov 			sched_unbind(td);
184fa83f689SKonstantin Belousov 		thread_unlock(td);
185fa83f689SKonstantin Belousov 		break;
186fa83f689SKonstantin Belousov 	case MSR_OP_RENDEZVOUS:
187fa83f689SKonstantin Belousov 		smp_rendezvous(NULL, x86_msr_op_one, NULL, &a);
188fa83f689SKonstantin Belousov 		break;
189fa83f689SKonstantin Belousov 	}
190fa83f689SKonstantin Belousov }
191fa83f689SKonstantin Belousov 
192665919aaSConrad Meyer /*
193665919aaSConrad Meyer  * Automatically initialized per CPU errata in cpu_idle_tun below.
194665919aaSConrad Meyer  */
195665919aaSConrad Meyer bool mwait_cpustop_broken = false;
196665919aaSConrad Meyer SYSCTL_BOOL(_machdep, OID_AUTO, mwait_cpustop_broken, CTLFLAG_RDTUN,
197665919aaSConrad Meyer     &mwait_cpustop_broken, 0,
198665919aaSConrad Meyer     "Can not reliably wake MONITOR/MWAIT cpus without interrupts");
1998428d0f1SAndriy Gapon 
200dfe7b3bfSKonstantin Belousov /*
201dfe7b3bfSKonstantin Belousov  * Flush the D-cache for non-DMA I/O so that the I-cache can
202dfe7b3bfSKonstantin Belousov  * be made coherent later.
203dfe7b3bfSKonstantin Belousov  */
204dfe7b3bfSKonstantin Belousov void
205dfe7b3bfSKonstantin Belousov cpu_flush_dcache(void *ptr, size_t len)
206dfe7b3bfSKonstantin Belousov {
207dfe7b3bfSKonstantin Belousov 	/* Not applicable */
208dfe7b3bfSKonstantin Belousov }
209dfe7b3bfSKonstantin Belousov 
210b57a73f8SKonstantin Belousov void
211b57a73f8SKonstantin Belousov acpi_cpu_c1(void)
212b57a73f8SKonstantin Belousov {
213b57a73f8SKonstantin Belousov 
214b57a73f8SKonstantin Belousov 	__asm __volatile("sti; hlt");
215b57a73f8SKonstantin Belousov }
216b57a73f8SKonstantin Belousov 
21719d4720bSJonathan T. Looney /*
21819d4720bSJonathan T. Looney  * Use mwait to pause execution while waiting for an interrupt or
21919d4720bSJonathan T. Looney  * another thread to signal that there is more work.
22019d4720bSJonathan T. Looney  *
22119d4720bSJonathan T. Looney  * NOTE: Interrupts will cause a wakeup; however, this function does
22219d4720bSJonathan T. Looney  * not enable interrupt handling. The caller is responsible to enable
22319d4720bSJonathan T. Looney  * interrupts.
22419d4720bSJonathan T. Looney  */
225b57a73f8SKonstantin Belousov void
226b57a73f8SKonstantin Belousov acpi_cpu_idle_mwait(uint32_t mwait_hint)
227b57a73f8SKonstantin Belousov {
228b57a73f8SKonstantin Belousov 	int *state;
2293621ba1eSKonstantin Belousov 	uint64_t v;
230b57a73f8SKonstantin Belousov 
231b57a73f8SKonstantin Belousov 	/*
232319117fdSKonstantin Belousov 	 * A comment in Linux patch claims that 'CPUs run faster with
233319117fdSKonstantin Belousov 	 * speculation protection disabled. All CPU threads in a core
234319117fdSKonstantin Belousov 	 * must disable speculation protection for it to be
235319117fdSKonstantin Belousov 	 * disabled. Disable it while we are idle so the other
236319117fdSKonstantin Belousov 	 * hyperthread can run fast.'
237319117fdSKonstantin Belousov 	 *
238b57a73f8SKonstantin Belousov 	 * XXXKIB.  Software coordination mode should be supported,
239b57a73f8SKonstantin Belousov 	 * but all Intel CPUs provide hardware coordination.
240b57a73f8SKonstantin Belousov 	 */
241d9e8bbb6SKonstantin Belousov 
24283dc49beSConrad Meyer 	state = &PCPU_PTR(monitorbuf)->idle_state;
243a5bd21d0SKonstantin Belousov 	KASSERT(atomic_load_int(state) == STATE_SLEEPING,
244d9e8bbb6SKonstantin Belousov 	    ("cpu_mwait_cx: wrong monitorbuf state"));
245a5bd21d0SKonstantin Belousov 	atomic_store_int(state, STATE_MWAIT);
2463621ba1eSKonstantin Belousov 	if (PCPU_GET(ibpb_set) || hw_ssb_active) {
2473621ba1eSKonstantin Belousov 		v = rdmsr(MSR_IA32_SPEC_CTRL);
2483621ba1eSKonstantin Belousov 		wrmsr(MSR_IA32_SPEC_CTRL, v & ~(IA32_SPEC_CTRL_IBRS |
2493621ba1eSKonstantin Belousov 		    IA32_SPEC_CTRL_STIBP | IA32_SPEC_CTRL_SSBD));
2503621ba1eSKonstantin Belousov 	} else {
2513621ba1eSKonstantin Belousov 		v = 0;
2523621ba1eSKonstantin Belousov 	}
253b57a73f8SKonstantin Belousov 	cpu_monitor(state, 0, 0);
254a5bd21d0SKonstantin Belousov 	if (atomic_load_int(state) == STATE_MWAIT)
255b57a73f8SKonstantin Belousov 		cpu_mwait(MWAIT_INTRBREAK, mwait_hint);
2563621ba1eSKonstantin Belousov 
2573621ba1eSKonstantin Belousov 	/*
2583621ba1eSKonstantin Belousov 	 * SSB cannot be disabled while we sleep, or rather, if it was
2593621ba1eSKonstantin Belousov 	 * disabled, the sysctl thread will bind to our cpu to tweak
2603621ba1eSKonstantin Belousov 	 * MSR.
2613621ba1eSKonstantin Belousov 	 */
2623621ba1eSKonstantin Belousov 	if (v != 0)
2633621ba1eSKonstantin Belousov 		wrmsr(MSR_IA32_SPEC_CTRL, v);
264d9e8bbb6SKonstantin Belousov 
265d9e8bbb6SKonstantin Belousov 	/*
266d9e8bbb6SKonstantin Belousov 	 * We should exit on any event that interrupts mwait, because
267d9e8bbb6SKonstantin Belousov 	 * that event might be a wanted interrupt.
268d9e8bbb6SKonstantin Belousov 	 */
269a5bd21d0SKonstantin Belousov 	atomic_store_int(state, STATE_RUNNING);
270b57a73f8SKonstantin Belousov }
271b57a73f8SKonstantin Belousov 
272dfe7b3bfSKonstantin Belousov /* Get current clock frequency for the given cpu id. */
273dfe7b3bfSKonstantin Belousov int
274dfe7b3bfSKonstantin Belousov cpu_est_clockrate(int cpu_id, uint64_t *rate)
275dfe7b3bfSKonstantin Belousov {
276dfe7b3bfSKonstantin Belousov 	uint64_t tsc1, tsc2;
277dfe7b3bfSKonstantin Belousov 	uint64_t acnt, mcnt, perf;
278dfe7b3bfSKonstantin Belousov 	register_t reg;
279dfe7b3bfSKonstantin Belousov 
280dfe7b3bfSKonstantin Belousov 	if (pcpu_find(cpu_id) == NULL || rate == NULL)
281dfe7b3bfSKonstantin Belousov 		return (EINVAL);
282dfe7b3bfSKonstantin Belousov #ifdef __i386__
283dfe7b3bfSKonstantin Belousov 	if ((cpu_feature & CPUID_TSC) == 0)
284dfe7b3bfSKonstantin Belousov 		return (EOPNOTSUPP);
285dfe7b3bfSKonstantin Belousov #endif
286dfe7b3bfSKonstantin Belousov 
287dfe7b3bfSKonstantin Belousov 	/*
288dfe7b3bfSKonstantin Belousov 	 * If TSC is P-state invariant and APERF/MPERF MSRs do not exist,
289dfe7b3bfSKonstantin Belousov 	 * DELAY(9) based logic fails.
290dfe7b3bfSKonstantin Belousov 	 */
291dfe7b3bfSKonstantin Belousov 	if (tsc_is_invariant && !tsc_perf_stat)
292dfe7b3bfSKonstantin Belousov 		return (EOPNOTSUPP);
293dfe7b3bfSKonstantin Belousov 
294dfe7b3bfSKonstantin Belousov #ifdef SMP
295dfe7b3bfSKonstantin Belousov 	if (smp_cpus > 1) {
296dfe7b3bfSKonstantin Belousov 		/* Schedule ourselves on the indicated cpu. */
297dfe7b3bfSKonstantin Belousov 		thread_lock(curthread);
298dfe7b3bfSKonstantin Belousov 		sched_bind(curthread, cpu_id);
299dfe7b3bfSKonstantin Belousov 		thread_unlock(curthread);
300dfe7b3bfSKonstantin Belousov 	}
301dfe7b3bfSKonstantin Belousov #endif
302dfe7b3bfSKonstantin Belousov 
303dfe7b3bfSKonstantin Belousov 	/* Calibrate by measuring a short delay. */
304dfe7b3bfSKonstantin Belousov 	reg = intr_disable();
305dfe7b3bfSKonstantin Belousov 	if (tsc_is_invariant) {
306dfe7b3bfSKonstantin Belousov 		wrmsr(MSR_MPERF, 0);
307dfe7b3bfSKonstantin Belousov 		wrmsr(MSR_APERF, 0);
308dfe7b3bfSKonstantin Belousov 		tsc1 = rdtsc();
309dfe7b3bfSKonstantin Belousov 		DELAY(1000);
310dfe7b3bfSKonstantin Belousov 		mcnt = rdmsr(MSR_MPERF);
311dfe7b3bfSKonstantin Belousov 		acnt = rdmsr(MSR_APERF);
312dfe7b3bfSKonstantin Belousov 		tsc2 = rdtsc();
313dfe7b3bfSKonstantin Belousov 		intr_restore(reg);
314dfe7b3bfSKonstantin Belousov 		perf = 1000 * acnt / mcnt;
315dfe7b3bfSKonstantin Belousov 		*rate = (tsc2 - tsc1) * perf;
316dfe7b3bfSKonstantin Belousov 	} else {
317dfe7b3bfSKonstantin Belousov 		tsc1 = rdtsc();
318dfe7b3bfSKonstantin Belousov 		DELAY(1000);
319dfe7b3bfSKonstantin Belousov 		tsc2 = rdtsc();
320dfe7b3bfSKonstantin Belousov 		intr_restore(reg);
321dfe7b3bfSKonstantin Belousov 		*rate = (tsc2 - tsc1) * 1000;
322dfe7b3bfSKonstantin Belousov 	}
323dfe7b3bfSKonstantin Belousov 
324dfe7b3bfSKonstantin Belousov #ifdef SMP
325dfe7b3bfSKonstantin Belousov 	if (smp_cpus > 1) {
326dfe7b3bfSKonstantin Belousov 		thread_lock(curthread);
327dfe7b3bfSKonstantin Belousov 		sched_unbind(curthread);
328dfe7b3bfSKonstantin Belousov 		thread_unlock(curthread);
329dfe7b3bfSKonstantin Belousov 	}
330dfe7b3bfSKonstantin Belousov #endif
331dfe7b3bfSKonstantin Belousov 
332dfe7b3bfSKonstantin Belousov 	return (0);
333dfe7b3bfSKonstantin Belousov }
334dfe7b3bfSKonstantin Belousov 
335dfe7b3bfSKonstantin Belousov /*
336dfe7b3bfSKonstantin Belousov  * Shutdown the CPU as much as possible
337dfe7b3bfSKonstantin Belousov  */
338dfe7b3bfSKonstantin Belousov void
339dfe7b3bfSKonstantin Belousov cpu_halt(void)
340dfe7b3bfSKonstantin Belousov {
341dfe7b3bfSKonstantin Belousov 	for (;;)
342dfe7b3bfSKonstantin Belousov 		halt();
343dfe7b3bfSKonstantin Belousov }
344dfe7b3bfSKonstantin Belousov 
3458428d0f1SAndriy Gapon static void
346b7b25af0SAndriy Gapon cpu_reset_real(void)
3478428d0f1SAndriy Gapon {
3488428d0f1SAndriy Gapon 	struct region_descriptor null_idt;
3498428d0f1SAndriy Gapon 	int b;
3508428d0f1SAndriy Gapon 
3518428d0f1SAndriy Gapon 	disable_intr();
3528428d0f1SAndriy Gapon #ifdef CPU_ELAN
3538428d0f1SAndriy Gapon 	if (elan_mmcr != NULL)
3548428d0f1SAndriy Gapon 		elan_mmcr->RESCFG = 1;
3558428d0f1SAndriy Gapon #endif
3568428d0f1SAndriy Gapon #ifdef __i386__
3578428d0f1SAndriy Gapon 	if (cpu == CPU_GEODE1100) {
3588428d0f1SAndriy Gapon 		/* Attempt Geode's own reset */
3598428d0f1SAndriy Gapon 		outl(0xcf8, 0x80009044ul);
3608428d0f1SAndriy Gapon 		outl(0xcfc, 0xf);
3618428d0f1SAndriy Gapon 	}
3628428d0f1SAndriy Gapon #endif
3638428d0f1SAndriy Gapon #if !defined(BROKEN_KEYBOARD_RESET)
3648428d0f1SAndriy Gapon 	/*
3658428d0f1SAndriy Gapon 	 * Attempt to do a CPU reset via the keyboard controller,
3668428d0f1SAndriy Gapon 	 * do not turn off GateA20, as any machine that fails
3678428d0f1SAndriy Gapon 	 * to do the reset here would then end up in no man's land.
3688428d0f1SAndriy Gapon 	 */
3698428d0f1SAndriy Gapon 	outb(IO_KBD + 4, 0xFE);
3708428d0f1SAndriy Gapon 	DELAY(500000);	/* wait 0.5 sec to see if that did it */
3718428d0f1SAndriy Gapon #endif
3728428d0f1SAndriy Gapon 
3738428d0f1SAndriy Gapon 	/*
3748428d0f1SAndriy Gapon 	 * Attempt to force a reset via the Reset Control register at
3758428d0f1SAndriy Gapon 	 * I/O port 0xcf9.  Bit 2 forces a system reset when it
3768428d0f1SAndriy Gapon 	 * transitions from 0 to 1.  Bit 1 selects the type of reset
3778428d0f1SAndriy Gapon 	 * to attempt: 0 selects a "soft" reset, and 1 selects a
3788428d0f1SAndriy Gapon 	 * "hard" reset.  We try a "hard" reset.  The first write sets
3798428d0f1SAndriy Gapon 	 * bit 1 to select a "hard" reset and clears bit 2.  The
3808428d0f1SAndriy Gapon 	 * second write forces a 0 -> 1 transition in bit 2 to trigger
3818428d0f1SAndriy Gapon 	 * a reset.
3828428d0f1SAndriy Gapon 	 */
3838428d0f1SAndriy Gapon 	outb(0xcf9, 0x2);
3848428d0f1SAndriy Gapon 	outb(0xcf9, 0x6);
3858428d0f1SAndriy Gapon 	DELAY(500000);  /* wait 0.5 sec to see if that did it */
3868428d0f1SAndriy Gapon 
3878428d0f1SAndriy Gapon 	/*
3888428d0f1SAndriy Gapon 	 * Attempt to force a reset via the Fast A20 and Init register
3898428d0f1SAndriy Gapon 	 * at I/O port 0x92.  Bit 1 serves as an alternate A20 gate.
3908428d0f1SAndriy Gapon 	 * Bit 0 asserts INIT# when set to 1.  We are careful to only
3918428d0f1SAndriy Gapon 	 * preserve bit 1 while setting bit 0.  We also must clear bit
3928428d0f1SAndriy Gapon 	 * 0 before setting it if it isn't already clear.
3938428d0f1SAndriy Gapon 	 */
3948428d0f1SAndriy Gapon 	b = inb(0x92);
3958428d0f1SAndriy Gapon 	if (b != 0xff) {
3968428d0f1SAndriy Gapon 		if ((b & 0x1) != 0)
3978428d0f1SAndriy Gapon 			outb(0x92, b & 0xfe);
3988428d0f1SAndriy Gapon 		outb(0x92, b | 0x1);
3998428d0f1SAndriy Gapon 		DELAY(500000);  /* wait 0.5 sec to see if that did it */
4008428d0f1SAndriy Gapon 	}
4018428d0f1SAndriy Gapon 
4028428d0f1SAndriy Gapon 	printf("No known reset method worked, attempting CPU shutdown\n");
4038428d0f1SAndriy Gapon 	DELAY(1000000); /* wait 1 sec for printf to complete */
4048428d0f1SAndriy Gapon 
4058428d0f1SAndriy Gapon 	/* Wipe the IDT. */
4068428d0f1SAndriy Gapon 	null_idt.rd_limit = 0;
4078428d0f1SAndriy Gapon 	null_idt.rd_base = 0;
4088428d0f1SAndriy Gapon 	lidt(&null_idt);
4098428d0f1SAndriy Gapon 
4108428d0f1SAndriy Gapon 	/* "good night, sweet prince .... <THUNK!>" */
4118428d0f1SAndriy Gapon 	breakpoint();
4128428d0f1SAndriy Gapon 
4138428d0f1SAndriy Gapon 	/* NOTREACHED */
4148428d0f1SAndriy Gapon 	while(1);
4158428d0f1SAndriy Gapon }
4168428d0f1SAndriy Gapon 
4178428d0f1SAndriy Gapon #ifdef SMP
4188428d0f1SAndriy Gapon static void
419b7b25af0SAndriy Gapon cpu_reset_proxy(void)
4208428d0f1SAndriy Gapon {
4218428d0f1SAndriy Gapon 
4228428d0f1SAndriy Gapon 	cpu_reset_proxy_active = 1;
4238428d0f1SAndriy Gapon 	while (cpu_reset_proxy_active == 1)
4248428d0f1SAndriy Gapon 		ia32_pause(); /* Wait for other cpu to see that we've started */
4258428d0f1SAndriy Gapon 
4268428d0f1SAndriy Gapon 	printf("cpu_reset_proxy: Stopped CPU %d\n", cpu_reset_proxyid);
4278428d0f1SAndriy Gapon 	DELAY(1000000);
4288428d0f1SAndriy Gapon 	cpu_reset_real();
4298428d0f1SAndriy Gapon }
4308428d0f1SAndriy Gapon #endif
4318428d0f1SAndriy Gapon 
4328428d0f1SAndriy Gapon void
433b7b25af0SAndriy Gapon cpu_reset(void)
4348428d0f1SAndriy Gapon {
4358428d0f1SAndriy Gapon #ifdef SMP
436665919aaSConrad Meyer 	struct monitorbuf *mb;
4378428d0f1SAndriy Gapon 	cpuset_t map;
4388428d0f1SAndriy Gapon 	u_int cnt;
4398428d0f1SAndriy Gapon 
4408428d0f1SAndriy Gapon 	if (smp_started) {
4418428d0f1SAndriy Gapon 		map = all_cpus;
4428428d0f1SAndriy Gapon 		CPU_CLR(PCPU_GET(cpuid), &map);
4439825eadfSRyan Libby 		CPU_ANDNOT(&map, &stopped_cpus);
4448428d0f1SAndriy Gapon 		if (!CPU_EMPTY(&map)) {
4458428d0f1SAndriy Gapon 			printf("cpu_reset: Stopping other CPUs\n");
4468428d0f1SAndriy Gapon 			stop_cpus(map);
4478428d0f1SAndriy Gapon 		}
4488428d0f1SAndriy Gapon 
4498428d0f1SAndriy Gapon 		if (PCPU_GET(cpuid) != 0) {
4508428d0f1SAndriy Gapon 			cpu_reset_proxyid = PCPU_GET(cpuid);
4518428d0f1SAndriy Gapon 			cpustop_restartfunc = cpu_reset_proxy;
4528428d0f1SAndriy Gapon 			cpu_reset_proxy_active = 0;
4538428d0f1SAndriy Gapon 			printf("cpu_reset: Restarting BSP\n");
4548428d0f1SAndriy Gapon 
4558428d0f1SAndriy Gapon 			/* Restart CPU #0. */
4568428d0f1SAndriy Gapon 			CPU_SETOF(0, &started_cpus);
457665919aaSConrad Meyer 			mb = &pcpu_find(0)->pc_monitorbuf;
458665919aaSConrad Meyer 			atomic_store_int(&mb->stop_state,
459665919aaSConrad Meyer 			    MONITOR_STOPSTATE_RUNNING);
4608428d0f1SAndriy Gapon 
4618428d0f1SAndriy Gapon 			cnt = 0;
4628428d0f1SAndriy Gapon 			while (cpu_reset_proxy_active == 0 && cnt < 10000000) {
4638428d0f1SAndriy Gapon 				ia32_pause();
4648428d0f1SAndriy Gapon 				cnt++;	/* Wait for BSP to announce restart */
4658428d0f1SAndriy Gapon 			}
4668428d0f1SAndriy Gapon 			if (cpu_reset_proxy_active == 0) {
4678428d0f1SAndriy Gapon 				printf("cpu_reset: Failed to restart BSP\n");
4688428d0f1SAndriy Gapon 			} else {
4698428d0f1SAndriy Gapon 				cpu_reset_proxy_active = 2;
4708428d0f1SAndriy Gapon 				while (1)
4718428d0f1SAndriy Gapon 					ia32_pause();
4728428d0f1SAndriy Gapon 				/* NOTREACHED */
4738428d0f1SAndriy Gapon 			}
4748428d0f1SAndriy Gapon 		}
4758428d0f1SAndriy Gapon 
4768428d0f1SAndriy Gapon 		DELAY(1000000);
4778428d0f1SAndriy Gapon 	}
4788428d0f1SAndriy Gapon #endif
4798428d0f1SAndriy Gapon 	cpu_reset_real();
4808428d0f1SAndriy Gapon 	/* NOTREACHED */
4818428d0f1SAndriy Gapon }
4828428d0f1SAndriy Gapon 
483b57a73f8SKonstantin Belousov bool
484b57a73f8SKonstantin Belousov cpu_mwait_usable(void)
485b57a73f8SKonstantin Belousov {
486b57a73f8SKonstantin Belousov 
487b57a73f8SKonstantin Belousov 	return ((cpu_feature2 & CPUID2_MON) != 0 && ((cpu_mon_mwait_flags &
488b57a73f8SKonstantin Belousov 	    (CPUID5_MON_MWAIT_EXT | CPUID5_MWAIT_INTRBREAK)) ==
489b57a73f8SKonstantin Belousov 	    (CPUID5_MON_MWAIT_EXT | CPUID5_MWAIT_INTRBREAK)));
490b57a73f8SKonstantin Belousov }
491b57a73f8SKonstantin Belousov 
492dfe7b3bfSKonstantin Belousov void (*cpu_idle_hook)(sbintime_t) = NULL;	/* ACPI idle hook. */
493d3ba71b2SKonstantin Belousov 
494d3ba71b2SKonstantin Belousov int cpu_amdc1e_bug = 0;			/* AMD C1E APIC workaround required. */
495d3ba71b2SKonstantin Belousov 
496dfe7b3bfSKonstantin Belousov static int	idle_mwait = 1;		/* Use MONITOR/MWAIT for short idle. */
497dfe7b3bfSKonstantin Belousov SYSCTL_INT(_machdep, OID_AUTO, idle_mwait, CTLFLAG_RWTUN, &idle_mwait,
498dfe7b3bfSKonstantin Belousov     0, "Use MONITOR/MWAIT for short idle");
499dfe7b3bfSKonstantin Belousov 
500dfe7b3bfSKonstantin Belousov static void
501dfe7b3bfSKonstantin Belousov cpu_idle_acpi(sbintime_t sbt)
502dfe7b3bfSKonstantin Belousov {
503dfe7b3bfSKonstantin Belousov 	int *state;
504dfe7b3bfSKonstantin Belousov 
50583dc49beSConrad Meyer 	state = &PCPU_PTR(monitorbuf)->idle_state;
506a5bd21d0SKonstantin Belousov 	atomic_store_int(state, STATE_SLEEPING);
507dfe7b3bfSKonstantin Belousov 
508dfe7b3bfSKonstantin Belousov 	/* See comments in cpu_idle_hlt(). */
509dfe7b3bfSKonstantin Belousov 	disable_intr();
510dfe7b3bfSKonstantin Belousov 	if (sched_runnable())
511dfe7b3bfSKonstantin Belousov 		enable_intr();
512dfe7b3bfSKonstantin Belousov 	else if (cpu_idle_hook)
513dfe7b3bfSKonstantin Belousov 		cpu_idle_hook(sbt);
514dfe7b3bfSKonstantin Belousov 	else
515b57a73f8SKonstantin Belousov 		acpi_cpu_c1();
516a5bd21d0SKonstantin Belousov 	atomic_store_int(state, STATE_RUNNING);
517dfe7b3bfSKonstantin Belousov }
518dfe7b3bfSKonstantin Belousov 
519dfe7b3bfSKonstantin Belousov static void
520dfe7b3bfSKonstantin Belousov cpu_idle_hlt(sbintime_t sbt)
521dfe7b3bfSKonstantin Belousov {
522dfe7b3bfSKonstantin Belousov 	int *state;
523dfe7b3bfSKonstantin Belousov 
52483dc49beSConrad Meyer 	state = &PCPU_PTR(monitorbuf)->idle_state;
525a5bd21d0SKonstantin Belousov 	atomic_store_int(state, STATE_SLEEPING);
526dfe7b3bfSKonstantin Belousov 
527dfe7b3bfSKonstantin Belousov 	/*
528dfe7b3bfSKonstantin Belousov 	 * Since we may be in a critical section from cpu_idle(), if
529dfe7b3bfSKonstantin Belousov 	 * an interrupt fires during that critical section we may have
530dfe7b3bfSKonstantin Belousov 	 * a pending preemption.  If the CPU halts, then that thread
531dfe7b3bfSKonstantin Belousov 	 * may not execute until a later interrupt awakens the CPU.
532dfe7b3bfSKonstantin Belousov 	 * To handle this race, check for a runnable thread after
533dfe7b3bfSKonstantin Belousov 	 * disabling interrupts and immediately return if one is
534dfe7b3bfSKonstantin Belousov 	 * found.  Also, we must absolutely guarentee that hlt is
535dfe7b3bfSKonstantin Belousov 	 * the next instruction after sti.  This ensures that any
536dfe7b3bfSKonstantin Belousov 	 * interrupt that fires after the call to disable_intr() will
537dfe7b3bfSKonstantin Belousov 	 * immediately awaken the CPU from hlt.  Finally, please note
538dfe7b3bfSKonstantin Belousov 	 * that on x86 this works fine because of interrupts enabled only
539dfe7b3bfSKonstantin Belousov 	 * after the instruction following sti takes place, while IF is set
540dfe7b3bfSKonstantin Belousov 	 * to 1 immediately, allowing hlt instruction to acknowledge the
541dfe7b3bfSKonstantin Belousov 	 * interrupt.
542dfe7b3bfSKonstantin Belousov 	 */
543dfe7b3bfSKonstantin Belousov 	disable_intr();
544dfe7b3bfSKonstantin Belousov 	if (sched_runnable())
545dfe7b3bfSKonstantin Belousov 		enable_intr();
546dfe7b3bfSKonstantin Belousov 	else
547b57a73f8SKonstantin Belousov 		acpi_cpu_c1();
548a5bd21d0SKonstantin Belousov 	atomic_store_int(state, STATE_RUNNING);
549dfe7b3bfSKonstantin Belousov }
550dfe7b3bfSKonstantin Belousov 
551dfe7b3bfSKonstantin Belousov static void
552dfe7b3bfSKonstantin Belousov cpu_idle_mwait(sbintime_t sbt)
553dfe7b3bfSKonstantin Belousov {
554dfe7b3bfSKonstantin Belousov 	int *state;
555dfe7b3bfSKonstantin Belousov 
55683dc49beSConrad Meyer 	state = &PCPU_PTR(monitorbuf)->idle_state;
557a5bd21d0SKonstantin Belousov 	atomic_store_int(state, STATE_MWAIT);
558dfe7b3bfSKonstantin Belousov 
559dfe7b3bfSKonstantin Belousov 	/* See comments in cpu_idle_hlt(). */
560dfe7b3bfSKonstantin Belousov 	disable_intr();
561dfe7b3bfSKonstantin Belousov 	if (sched_runnable()) {
562a5bd21d0SKonstantin Belousov 		atomic_store_int(state, STATE_RUNNING);
563dfe7b3bfSKonstantin Belousov 		enable_intr();
564dfe7b3bfSKonstantin Belousov 		return;
565dfe7b3bfSKonstantin Belousov 	}
566a5bd21d0SKonstantin Belousov 
567dfe7b3bfSKonstantin Belousov 	cpu_monitor(state, 0, 0);
568a5bd21d0SKonstantin Belousov 	if (atomic_load_int(state) == STATE_MWAIT)
569dfe7b3bfSKonstantin Belousov 		__asm __volatile("sti; mwait" : : "a" (MWAIT_C1), "c" (0));
570dfe7b3bfSKonstantin Belousov 	else
571dfe7b3bfSKonstantin Belousov 		enable_intr();
572a5bd21d0SKonstantin Belousov 	atomic_store_int(state, STATE_RUNNING);
573dfe7b3bfSKonstantin Belousov }
574dfe7b3bfSKonstantin Belousov 
575dfe7b3bfSKonstantin Belousov static void
576dfe7b3bfSKonstantin Belousov cpu_idle_spin(sbintime_t sbt)
577dfe7b3bfSKonstantin Belousov {
578dfe7b3bfSKonstantin Belousov 	int *state;
579dfe7b3bfSKonstantin Belousov 	int i;
580dfe7b3bfSKonstantin Belousov 
58183dc49beSConrad Meyer 	state = &PCPU_PTR(monitorbuf)->idle_state;
582a5bd21d0SKonstantin Belousov 	atomic_store_int(state, STATE_RUNNING);
583dfe7b3bfSKonstantin Belousov 
584dfe7b3bfSKonstantin Belousov 	/*
585dfe7b3bfSKonstantin Belousov 	 * The sched_runnable() call is racy but as long as there is
586dfe7b3bfSKonstantin Belousov 	 * a loop missing it one time will have just a little impact if any
587dfe7b3bfSKonstantin Belousov 	 * (and it is much better than missing the check at all).
588dfe7b3bfSKonstantin Belousov 	 */
589dfe7b3bfSKonstantin Belousov 	for (i = 0; i < 1000; i++) {
590dfe7b3bfSKonstantin Belousov 		if (sched_runnable())
591dfe7b3bfSKonstantin Belousov 			return;
592dfe7b3bfSKonstantin Belousov 		cpu_spinwait();
593dfe7b3bfSKonstantin Belousov 	}
594dfe7b3bfSKonstantin Belousov }
595dfe7b3bfSKonstantin Belousov 
596dfe7b3bfSKonstantin Belousov void (*cpu_idle_fn)(sbintime_t) = cpu_idle_acpi;
597dfe7b3bfSKonstantin Belousov 
598dfe7b3bfSKonstantin Belousov void
599dfe7b3bfSKonstantin Belousov cpu_idle(int busy)
600dfe7b3bfSKonstantin Belousov {
601dfe7b3bfSKonstantin Belousov 	uint64_t msr;
602dfe7b3bfSKonstantin Belousov 	sbintime_t sbt = -1;
603dfe7b3bfSKonstantin Belousov 
604dfe7b3bfSKonstantin Belousov 	CTR2(KTR_SPARE2, "cpu_idle(%d) at %d",
605dfe7b3bfSKonstantin Belousov 	    busy, curcpu);
606ed95805eSJohn Baldwin #ifdef MP_WATCHDOG
607dfe7b3bfSKonstantin Belousov 	ap_watchdog(PCPU_GET(cpuid));
608dfe7b3bfSKonstantin Belousov #endif
609ed95805eSJohn Baldwin 
610dfe7b3bfSKonstantin Belousov 	/* If we are busy - try to use fast methods. */
611dfe7b3bfSKonstantin Belousov 	if (busy) {
612dfe7b3bfSKonstantin Belousov 		if ((cpu_feature2 & CPUID2_MON) && idle_mwait) {
613dfe7b3bfSKonstantin Belousov 			cpu_idle_mwait(busy);
614dfe7b3bfSKonstantin Belousov 			goto out;
615dfe7b3bfSKonstantin Belousov 		}
616dfe7b3bfSKonstantin Belousov 	}
617dfe7b3bfSKonstantin Belousov 
618dfe7b3bfSKonstantin Belousov 	/* If we have time - switch timers into idle mode. */
619dfe7b3bfSKonstantin Belousov 	if (!busy) {
620dfe7b3bfSKonstantin Belousov 		critical_enter();
621dfe7b3bfSKonstantin Belousov 		sbt = cpu_idleclock();
622dfe7b3bfSKonstantin Belousov 	}
623dfe7b3bfSKonstantin Belousov 
624dfe7b3bfSKonstantin Belousov 	/* Apply AMD APIC timer C1E workaround. */
625d3ba71b2SKonstantin Belousov 	if (cpu_amdc1e_bug && cpu_disable_c3_sleep) {
626dfe7b3bfSKonstantin Belousov 		msr = rdmsr(MSR_AMDK8_IPM);
627d3ba71b2SKonstantin Belousov 		if ((msr & (AMDK8_SMIONCMPHALT | AMDK8_C1EONCMPHALT)) != 0)
628d3ba71b2SKonstantin Belousov 			wrmsr(MSR_AMDK8_IPM, msr & ~(AMDK8_SMIONCMPHALT |
629d3ba71b2SKonstantin Belousov 			    AMDK8_C1EONCMPHALT));
630dfe7b3bfSKonstantin Belousov 	}
631dfe7b3bfSKonstantin Belousov 
632dfe7b3bfSKonstantin Belousov 	/* Call main idle method. */
633dfe7b3bfSKonstantin Belousov 	cpu_idle_fn(sbt);
634dfe7b3bfSKonstantin Belousov 
635dfe7b3bfSKonstantin Belousov 	/* Switch timers back into active mode. */
636dfe7b3bfSKonstantin Belousov 	if (!busy) {
637dfe7b3bfSKonstantin Belousov 		cpu_activeclock();
638dfe7b3bfSKonstantin Belousov 		critical_exit();
639dfe7b3bfSKonstantin Belousov 	}
640dfe7b3bfSKonstantin Belousov out:
641dfe7b3bfSKonstantin Belousov 	CTR2(KTR_SPARE2, "cpu_idle(%d) at %d done",
642dfe7b3bfSKonstantin Belousov 	    busy, curcpu);
643dfe7b3bfSKonstantin Belousov }
644dfe7b3bfSKonstantin Belousov 
6453f3937b4SKonstantin Belousov static int cpu_idle_apl31_workaround;
6463f3937b4SKonstantin Belousov SYSCTL_INT(_machdep, OID_AUTO, idle_apl31, CTLFLAG_RW,
6473f3937b4SKonstantin Belousov     &cpu_idle_apl31_workaround, 0,
648160be7ccSKonstantin Belousov     "Apollo Lake APL31 MWAIT bug workaround");
6493f3937b4SKonstantin Belousov 
650dfe7b3bfSKonstantin Belousov int
651dfe7b3bfSKonstantin Belousov cpu_idle_wakeup(int cpu)
652dfe7b3bfSKonstantin Belousov {
65383dc49beSConrad Meyer 	struct monitorbuf *mb;
654dfe7b3bfSKonstantin Belousov 	int *state;
655dfe7b3bfSKonstantin Belousov 
65683dc49beSConrad Meyer 	mb = &pcpu_find(cpu)->pc_monitorbuf;
65783dc49beSConrad Meyer 	state = &mb->idle_state;
658a5bd21d0SKonstantin Belousov 	switch (atomic_load_int(state)) {
659a5bd21d0SKonstantin Belousov 	case STATE_SLEEPING:
660dfe7b3bfSKonstantin Belousov 		return (0);
661a5bd21d0SKonstantin Belousov 	case STATE_MWAIT:
662a5bd21d0SKonstantin Belousov 		atomic_store_int(state, STATE_RUNNING);
6633f3937b4SKonstantin Belousov 		return (cpu_idle_apl31_workaround ? 0 : 1);
664a5bd21d0SKonstantin Belousov 	case STATE_RUNNING:
665a5bd21d0SKonstantin Belousov 		return (1);
666a5bd21d0SKonstantin Belousov 	default:
667a5bd21d0SKonstantin Belousov 		panic("bad monitor state");
668a5bd21d0SKonstantin Belousov 		return (1);
669a5bd21d0SKonstantin Belousov 	}
670dfe7b3bfSKonstantin Belousov }
671dfe7b3bfSKonstantin Belousov 
672dfe7b3bfSKonstantin Belousov /*
673dfe7b3bfSKonstantin Belousov  * Ordered by speed/power consumption.
674dfe7b3bfSKonstantin Belousov  */
675a5f472c5SKonstantin Belousov static struct {
676dfe7b3bfSKonstantin Belousov 	void	*id_fn;
677dfe7b3bfSKonstantin Belousov 	char	*id_name;
678a5f472c5SKonstantin Belousov 	int	id_cpuid2_flag;
679dfe7b3bfSKonstantin Belousov } idle_tbl[] = {
680a5f472c5SKonstantin Belousov 	{ .id_fn = cpu_idle_spin, .id_name = "spin" },
681a5f472c5SKonstantin Belousov 	{ .id_fn = cpu_idle_mwait, .id_name = "mwait",
682a5f472c5SKonstantin Belousov 	    .id_cpuid2_flag = CPUID2_MON },
683a5f472c5SKonstantin Belousov 	{ .id_fn = cpu_idle_hlt, .id_name = "hlt" },
684a5f472c5SKonstantin Belousov 	{ .id_fn = cpu_idle_acpi, .id_name = "acpi" },
685dfe7b3bfSKonstantin Belousov };
686dfe7b3bfSKonstantin Belousov 
687dfe7b3bfSKonstantin Belousov static int
688dfe7b3bfSKonstantin Belousov idle_sysctl_available(SYSCTL_HANDLER_ARGS)
689dfe7b3bfSKonstantin Belousov {
690dfe7b3bfSKonstantin Belousov 	char *avail, *p;
691dfe7b3bfSKonstantin Belousov 	int error;
692dfe7b3bfSKonstantin Belousov 	int i;
693dfe7b3bfSKonstantin Belousov 
694dfe7b3bfSKonstantin Belousov 	avail = malloc(256, M_TEMP, M_WAITOK);
695dfe7b3bfSKonstantin Belousov 	p = avail;
696a5f472c5SKonstantin Belousov 	for (i = 0; i < nitems(idle_tbl); i++) {
697a5f472c5SKonstantin Belousov 		if (idle_tbl[i].id_cpuid2_flag != 0 &&
698a5f472c5SKonstantin Belousov 		    (cpu_feature2 & idle_tbl[i].id_cpuid2_flag) == 0)
699dfe7b3bfSKonstantin Belousov 			continue;
700dfe7b3bfSKonstantin Belousov 		if (strcmp(idle_tbl[i].id_name, "acpi") == 0 &&
701dfe7b3bfSKonstantin Belousov 		    cpu_idle_hook == NULL)
702dfe7b3bfSKonstantin Belousov 			continue;
703dfe7b3bfSKonstantin Belousov 		p += sprintf(p, "%s%s", p != avail ? ", " : "",
704dfe7b3bfSKonstantin Belousov 		    idle_tbl[i].id_name);
705dfe7b3bfSKonstantin Belousov 	}
706dfe7b3bfSKonstantin Belousov 	error = sysctl_handle_string(oidp, avail, 0, req);
707dfe7b3bfSKonstantin Belousov 	free(avail, M_TEMP);
708dfe7b3bfSKonstantin Belousov 	return (error);
709dfe7b3bfSKonstantin Belousov }
710dfe7b3bfSKonstantin Belousov 
7117029da5cSPawel Biernacki SYSCTL_PROC(_machdep, OID_AUTO, idle_available,
7127029da5cSPawel Biernacki     CTLTYPE_STRING | CTLFLAG_RD | CTLFLAG_NEEDGIANT,
7137029da5cSPawel Biernacki     0, 0, idle_sysctl_available, "A",
7147029da5cSPawel Biernacki     "list of available idle functions");
715dfe7b3bfSKonstantin Belousov 
71655ba21d4SKonstantin Belousov static bool
717a5f472c5SKonstantin Belousov cpu_idle_selector(const char *new_idle_name)
71855ba21d4SKonstantin Belousov {
71955ba21d4SKonstantin Belousov 	int i;
72055ba21d4SKonstantin Belousov 
721a5f472c5SKonstantin Belousov 	for (i = 0; i < nitems(idle_tbl); i++) {
722a5f472c5SKonstantin Belousov 		if (idle_tbl[i].id_cpuid2_flag != 0 &&
723a5f472c5SKonstantin Belousov 		    (cpu_feature2 & idle_tbl[i].id_cpuid2_flag) == 0)
72455ba21d4SKonstantin Belousov 			continue;
72555ba21d4SKonstantin Belousov 		if (strcmp(idle_tbl[i].id_name, "acpi") == 0 &&
72655ba21d4SKonstantin Belousov 		    cpu_idle_hook == NULL)
72755ba21d4SKonstantin Belousov 			continue;
72855ba21d4SKonstantin Belousov 		if (strcmp(idle_tbl[i].id_name, new_idle_name))
72955ba21d4SKonstantin Belousov 			continue;
73055ba21d4SKonstantin Belousov 		cpu_idle_fn = idle_tbl[i].id_fn;
73155ba21d4SKonstantin Belousov 		if (bootverbose)
73255ba21d4SKonstantin Belousov 			printf("CPU idle set to %s\n", idle_tbl[i].id_name);
73355ba21d4SKonstantin Belousov 		return (true);
73455ba21d4SKonstantin Belousov 	}
73555ba21d4SKonstantin Belousov 	return (false);
73655ba21d4SKonstantin Belousov }
73755ba21d4SKonstantin Belousov 
738dfe7b3bfSKonstantin Belousov static int
739a5f472c5SKonstantin Belousov cpu_idle_sysctl(SYSCTL_HANDLER_ARGS)
740dfe7b3bfSKonstantin Belousov {
74155ba21d4SKonstantin Belousov 	char buf[16], *p;
74255ba21d4SKonstantin Belousov 	int error, i;
743dfe7b3bfSKonstantin Belousov 
744dfe7b3bfSKonstantin Belousov 	p = "unknown";
745a5f472c5SKonstantin Belousov 	for (i = 0; i < nitems(idle_tbl); i++) {
746dfe7b3bfSKonstantin Belousov 		if (idle_tbl[i].id_fn == cpu_idle_fn) {
747dfe7b3bfSKonstantin Belousov 			p = idle_tbl[i].id_name;
748dfe7b3bfSKonstantin Belousov 			break;
749dfe7b3bfSKonstantin Belousov 		}
750dfe7b3bfSKonstantin Belousov 	}
751dfe7b3bfSKonstantin Belousov 	strncpy(buf, p, sizeof(buf));
752dfe7b3bfSKonstantin Belousov 	error = sysctl_handle_string(oidp, buf, sizeof(buf), req);
753dfe7b3bfSKonstantin Belousov 	if (error != 0 || req->newptr == NULL)
754dfe7b3bfSKonstantin Belousov 		return (error);
755a5f472c5SKonstantin Belousov 	return (cpu_idle_selector(buf) ? 0 : EINVAL);
756dfe7b3bfSKonstantin Belousov }
757dfe7b3bfSKonstantin Belousov 
7587029da5cSPawel Biernacki SYSCTL_PROC(_machdep, OID_AUTO, idle,
7597029da5cSPawel Biernacki     CTLTYPE_STRING | CTLFLAG_RW | CTLFLAG_NEEDGIANT,
7607029da5cSPawel Biernacki     0, 0, cpu_idle_sysctl, "A",
7617029da5cSPawel Biernacki     "currently selected idle function");
762835c2787SKonstantin Belousov 
76355ba21d4SKonstantin Belousov static void
764a5f472c5SKonstantin Belousov cpu_idle_tun(void *unused __unused)
76555ba21d4SKonstantin Belousov {
76655ba21d4SKonstantin Belousov 	char tunvar[16];
76755ba21d4SKonstantin Belousov 
76855ba21d4SKonstantin Belousov 	if (TUNABLE_STR_FETCH("machdep.idle", tunvar, sizeof(tunvar)))
769a5f472c5SKonstantin Belousov 		cpu_idle_selector(tunvar);
77045ed991dSKonstantin Belousov 	else if (cpu_vendor_id == CPU_VENDOR_AMD &&
77145ed991dSKonstantin Belousov 	    CPUID_TO_FAMILY(cpu_id) == 0x17 && CPUID_TO_MODEL(cpu_id) == 0x1) {
77245ed991dSKonstantin Belousov 		/* Ryzen erratas 1057, 1109. */
77345ed991dSKonstantin Belousov 		cpu_idle_selector("hlt");
77445ed991dSKonstantin Belousov 		idle_mwait = 0;
775665919aaSConrad Meyer 		mwait_cpustop_broken = true;
77645ed991dSKonstantin Belousov 	}
77745ed991dSKonstantin Belousov 
7783f3937b4SKonstantin Belousov 	if (cpu_vendor_id == CPU_VENDOR_INTEL && cpu_id == 0x506c9) {
7793f3937b4SKonstantin Belousov 		/*
780160be7ccSKonstantin Belousov 		 * Apollo Lake errata APL31 (public errata APL30).
781160be7ccSKonstantin Belousov 		 * Stores to the armed address range may not trigger
782160be7ccSKonstantin Belousov 		 * MWAIT to resume execution.  OS needs to use
783160be7ccSKonstantin Belousov 		 * interrupts to wake processors from MWAIT-induced
784160be7ccSKonstantin Belousov 		 * sleep states.
7853f3937b4SKonstantin Belousov 		 */
7863f3937b4SKonstantin Belousov 		cpu_idle_apl31_workaround = 1;
787665919aaSConrad Meyer 		mwait_cpustop_broken = true;
7883f3937b4SKonstantin Belousov 	}
7893f3937b4SKonstantin Belousov 	TUNABLE_INT_FETCH("machdep.idle_apl31", &cpu_idle_apl31_workaround);
79055ba21d4SKonstantin Belousov }
791a5f472c5SKonstantin Belousov SYSINIT(cpu_idle_tun, SI_SUB_CPU, SI_ORDER_MIDDLE, cpu_idle_tun, NULL);
79255ba21d4SKonstantin Belousov 
793ba0ced82SEric van Gyzen static int panic_on_nmi = 0xff;
794295f4b6cSKonstantin Belousov SYSCTL_INT(_machdep, OID_AUTO, panic_on_nmi, CTLFLAG_RWTUN,
795295f4b6cSKonstantin Belousov     &panic_on_nmi, 0,
796ba0ced82SEric van Gyzen     "Panic on NMI: 1 = H/W failure; 2 = unknown; 0xff = all");
797835c2787SKonstantin Belousov int nmi_is_broadcast = 1;
798835c2787SKonstantin Belousov SYSCTL_INT(_machdep, OID_AUTO, nmi_is_broadcast, CTLFLAG_RWTUN,
799835c2787SKonstantin Belousov     &nmi_is_broadcast, 0,
800835c2787SKonstantin Belousov     "Chipset NMI is broadcast");
801855e49f3SAlexander Motin int (*apei_nmi)(void);
802835c2787SKonstantin Belousov 
803295f4b6cSKonstantin Belousov void
804295f4b6cSKonstantin Belousov nmi_call_kdb(u_int cpu, u_int type, struct trapframe *frame)
805835c2787SKonstantin Belousov {
8060fb3a72aSAndriy Gapon 	bool claimed = false;
807835c2787SKonstantin Belousov 
8080fb3a72aSAndriy Gapon #ifdef DEV_ISA
809835c2787SKonstantin Belousov 	/* machine/parity/power fail/"kitchen sink" faults */
8100fb3a72aSAndriy Gapon 	if (isa_nmi(frame->tf_err)) {
8110fb3a72aSAndriy Gapon 		claimed = true;
812ba0ced82SEric van Gyzen 		if ((panic_on_nmi & 1) != 0)
8130fb3a72aSAndriy Gapon 			panic("NMI indicates hardware failure");
8140fb3a72aSAndriy Gapon 	}
8150fb3a72aSAndriy Gapon #endif /* DEV_ISA */
816ba0ced82SEric van Gyzen 
817855e49f3SAlexander Motin 	/* ACPI Platform Error Interfaces callback. */
818855e49f3SAlexander Motin 	if (apei_nmi != NULL && (*apei_nmi)())
819855e49f3SAlexander Motin 		claimed = true;
820855e49f3SAlexander Motin 
821835c2787SKonstantin Belousov 	/*
822ba0ced82SEric van Gyzen 	 * NMIs can be useful for debugging.  They can be hooked up to a
823ba0ced82SEric van Gyzen 	 * pushbutton, usually on an ISA, PCI, or PCIe card.  They can also be
824ba0ced82SEric van Gyzen 	 * generated by an IPMI BMC, either manually or in response to a
825ba0ced82SEric van Gyzen 	 * watchdog timeout.  For example, see the "power diag" command in
826ba0ced82SEric van Gyzen 	 * ports/sysutils/ipmitool.  They can also be generated by a
827ba0ced82SEric van Gyzen 	 * hypervisor; see "bhyvectl --inject-nmi".
828835c2787SKonstantin Belousov 	 */
829ba0ced82SEric van Gyzen 
830ba0ced82SEric van Gyzen #ifdef KDB
831ba0ced82SEric van Gyzen 	if (!claimed && (panic_on_nmi & 2) != 0) {
832ba0ced82SEric van Gyzen 		if (debugger_on_panic) {
833835c2787SKonstantin Belousov 			printf("NMI/cpu%d ... going to debugger\n", cpu);
834ba0ced82SEric van Gyzen 			claimed = kdb_trap(type, 0, frame);
835ba0ced82SEric van Gyzen 		}
836835c2787SKonstantin Belousov 	}
837835c2787SKonstantin Belousov #endif /* KDB */
838ba0ced82SEric van Gyzen 
839ba0ced82SEric van Gyzen 	if (!claimed && panic_on_nmi != 0)
840ba0ced82SEric van Gyzen 		panic("NMI");
841295f4b6cSKonstantin Belousov }
842835c2787SKonstantin Belousov 
843295f4b6cSKonstantin Belousov void
844295f4b6cSKonstantin Belousov nmi_handle_intr(u_int type, struct trapframe *frame)
845835c2787SKonstantin Belousov {
846835c2787SKonstantin Belousov 
847835c2787SKonstantin Belousov #ifdef SMP
848295f4b6cSKonstantin Belousov 	if (nmi_is_broadcast) {
849295f4b6cSKonstantin Belousov 		nmi_call_kdb_smp(type, frame);
850295f4b6cSKonstantin Belousov 		return;
851295f4b6cSKonstantin Belousov 	}
852835c2787SKonstantin Belousov #endif
8531d6dfd12SKonstantin Belousov 	nmi_call_kdb(PCPU_GET(cpuid), type, frame);
854835c2787SKonstantin Belousov }
855319117fdSKonstantin Belousov 
856a324b7f7SKonstantin Belousov static int hw_ibrs_active;
857a324b7f7SKonstantin Belousov int hw_ibrs_ibpb_active;
858319117fdSKonstantin Belousov int hw_ibrs_disable = 1;
859319117fdSKonstantin Belousov 
860319117fdSKonstantin Belousov SYSCTL_INT(_hw, OID_AUTO, ibrs_active, CTLFLAG_RD, &hw_ibrs_active, 0,
861b31b965eSKonstantin Belousov     "Indirect Branch Restricted Speculation active");
862319117fdSKonstantin Belousov 
8637029da5cSPawel Biernacki SYSCTL_NODE(_machdep_mitigations, OID_AUTO, ibrs,
8647029da5cSPawel Biernacki     CTLFLAG_RW | CTLFLAG_MPSAFE, 0,
865961aacb1SScott Long     "Indirect Branch Restricted Speculation active");
866961aacb1SScott Long 
867961aacb1SScott Long SYSCTL_INT(_machdep_mitigations_ibrs, OID_AUTO, active, CTLFLAG_RD,
868961aacb1SScott Long     &hw_ibrs_active, 0, "Indirect Branch Restricted Speculation active");
869961aacb1SScott Long 
870319117fdSKonstantin Belousov void
871a324b7f7SKonstantin Belousov hw_ibrs_recalculate(bool for_all_cpus)
872319117fdSKonstantin Belousov {
873319117fdSKonstantin Belousov 	if ((cpu_ia32_arch_caps & IA32_ARCH_CAP_IBRS_ALL) != 0) {
874a324b7f7SKonstantin Belousov 		x86_msr_op(MSR_IA32_SPEC_CTRL, (for_all_cpus ?
875a324b7f7SKonstantin Belousov 		    MSR_OP_RENDEZVOUS : MSR_OP_LOCAL) |
876a324b7f7SKonstantin Belousov 		    (hw_ibrs_disable != 0 ? MSR_OP_ANDNOT : MSR_OP_OR),
877fa83f689SKonstantin Belousov 		    IA32_SPEC_CTRL_IBRS);
878a324b7f7SKonstantin Belousov 		hw_ibrs_active = hw_ibrs_disable == 0;
879a324b7f7SKonstantin Belousov 		hw_ibrs_ibpb_active = 0;
880a324b7f7SKonstantin Belousov 	} else {
881a324b7f7SKonstantin Belousov 		hw_ibrs_active = hw_ibrs_ibpb_active = (cpu_stdext_feature3 &
882a324b7f7SKonstantin Belousov 		    CPUID_STDEXT3_IBPB) != 0 && !hw_ibrs_disable;
883319117fdSKonstantin Belousov 	}
884319117fdSKonstantin Belousov }
885319117fdSKonstantin Belousov 
886319117fdSKonstantin Belousov static int
887319117fdSKonstantin Belousov hw_ibrs_disable_handler(SYSCTL_HANDLER_ARGS)
888319117fdSKonstantin Belousov {
889319117fdSKonstantin Belousov 	int error, val;
890319117fdSKonstantin Belousov 
891319117fdSKonstantin Belousov 	val = hw_ibrs_disable;
892319117fdSKonstantin Belousov 	error = sysctl_handle_int(oidp, &val, 0, req);
893319117fdSKonstantin Belousov 	if (error != 0 || req->newptr == NULL)
894319117fdSKonstantin Belousov 		return (error);
895319117fdSKonstantin Belousov 	hw_ibrs_disable = val != 0;
896a324b7f7SKonstantin Belousov 	hw_ibrs_recalculate(true);
897319117fdSKonstantin Belousov 	return (0);
898319117fdSKonstantin Belousov }
899319117fdSKonstantin Belousov SYSCTL_PROC(_hw, OID_AUTO, ibrs_disable, CTLTYPE_INT | CTLFLAG_RWTUN |
900319117fdSKonstantin Belousov     CTLFLAG_NOFETCH | CTLFLAG_MPSAFE, NULL, 0, hw_ibrs_disable_handler, "I",
901b31b965eSKonstantin Belousov     "Disable Indirect Branch Restricted Speculation");
9028fbcc334SKonstantin Belousov 
903961aacb1SScott Long SYSCTL_PROC(_machdep_mitigations_ibrs, OID_AUTO, disable, CTLTYPE_INT |
904961aacb1SScott Long     CTLFLAG_RWTUN | CTLFLAG_NOFETCH | CTLFLAG_MPSAFE, NULL, 0,
905961aacb1SScott Long     hw_ibrs_disable_handler, "I",
906961aacb1SScott Long     "Disable Indirect Branch Restricted Speculation");
907961aacb1SScott Long 
9083621ba1eSKonstantin Belousov int hw_ssb_active;
9093621ba1eSKonstantin Belousov int hw_ssb_disable;
9103621ba1eSKonstantin Belousov 
9113621ba1eSKonstantin Belousov SYSCTL_INT(_hw, OID_AUTO, spec_store_bypass_disable_active, CTLFLAG_RD,
9123621ba1eSKonstantin Belousov     &hw_ssb_active, 0,
9133621ba1eSKonstantin Belousov     "Speculative Store Bypass Disable active");
9143621ba1eSKonstantin Belousov 
9157029da5cSPawel Biernacki SYSCTL_NODE(_machdep_mitigations, OID_AUTO, ssb,
9167029da5cSPawel Biernacki     CTLFLAG_RW | CTLFLAG_MPSAFE, 0,
917961aacb1SScott Long     "Speculative Store Bypass Disable active");
918961aacb1SScott Long 
919961aacb1SScott Long SYSCTL_INT(_machdep_mitigations_ssb, OID_AUTO, active, CTLFLAG_RD,
920961aacb1SScott Long     &hw_ssb_active, 0, "Speculative Store Bypass Disable active");
921961aacb1SScott Long 
9223621ba1eSKonstantin Belousov static void
9233621ba1eSKonstantin Belousov hw_ssb_set(bool enable, bool for_all_cpus)
9243621ba1eSKonstantin Belousov {
9253621ba1eSKonstantin Belousov 
9263621ba1eSKonstantin Belousov 	if ((cpu_stdext_feature3 & CPUID_STDEXT3_SSBD) == 0) {
9273621ba1eSKonstantin Belousov 		hw_ssb_active = 0;
9283621ba1eSKonstantin Belousov 		return;
9293621ba1eSKonstantin Belousov 	}
9303621ba1eSKonstantin Belousov 	hw_ssb_active = enable;
931fa83f689SKonstantin Belousov 	x86_msr_op(MSR_IA32_SPEC_CTRL,
932fa83f689SKonstantin Belousov 	    (enable ? MSR_OP_OR : MSR_OP_ANDNOT) |
933fa83f689SKonstantin Belousov 	    (for_all_cpus ? MSR_OP_SCHED : MSR_OP_LOCAL), IA32_SPEC_CTRL_SSBD);
9343621ba1eSKonstantin Belousov }
9353621ba1eSKonstantin Belousov 
9363621ba1eSKonstantin Belousov void
9373621ba1eSKonstantin Belousov hw_ssb_recalculate(bool all_cpus)
9383621ba1eSKonstantin Belousov {
9393621ba1eSKonstantin Belousov 
9403621ba1eSKonstantin Belousov 	switch (hw_ssb_disable) {
9413621ba1eSKonstantin Belousov 	default:
9423621ba1eSKonstantin Belousov 		hw_ssb_disable = 0;
9433621ba1eSKonstantin Belousov 		/* FALLTHROUGH */
9443621ba1eSKonstantin Belousov 	case 0: /* off */
9453621ba1eSKonstantin Belousov 		hw_ssb_set(false, all_cpus);
9463621ba1eSKonstantin Belousov 		break;
9473621ba1eSKonstantin Belousov 	case 1: /* on */
9483621ba1eSKonstantin Belousov 		hw_ssb_set(true, all_cpus);
9493621ba1eSKonstantin Belousov 		break;
9503621ba1eSKonstantin Belousov 	case 2: /* auto */
95123437573SKonstantin Belousov 		hw_ssb_set((cpu_ia32_arch_caps & IA32_ARCH_CAP_SSB_NO) != 0 ?
9523621ba1eSKonstantin Belousov 		    false : true, all_cpus);
9533621ba1eSKonstantin Belousov 		break;
9543621ba1eSKonstantin Belousov 	}
9553621ba1eSKonstantin Belousov }
9563621ba1eSKonstantin Belousov 
9573621ba1eSKonstantin Belousov static int
9583621ba1eSKonstantin Belousov hw_ssb_disable_handler(SYSCTL_HANDLER_ARGS)
9593621ba1eSKonstantin Belousov {
9603621ba1eSKonstantin Belousov 	int error, val;
9613621ba1eSKonstantin Belousov 
9623621ba1eSKonstantin Belousov 	val = hw_ssb_disable;
9633621ba1eSKonstantin Belousov 	error = sysctl_handle_int(oidp, &val, 0, req);
9643621ba1eSKonstantin Belousov 	if (error != 0 || req->newptr == NULL)
9653621ba1eSKonstantin Belousov 		return (error);
9663621ba1eSKonstantin Belousov 	hw_ssb_disable = val;
9673621ba1eSKonstantin Belousov 	hw_ssb_recalculate(true);
9683621ba1eSKonstantin Belousov 	return (0);
9693621ba1eSKonstantin Belousov }
9703621ba1eSKonstantin Belousov SYSCTL_PROC(_hw, OID_AUTO, spec_store_bypass_disable, CTLTYPE_INT |
9713621ba1eSKonstantin Belousov     CTLFLAG_RWTUN | CTLFLAG_NOFETCH | CTLFLAG_MPSAFE, NULL, 0,
9723621ba1eSKonstantin Belousov     hw_ssb_disable_handler, "I",
973*a212f56dSPiotr Pawel Stefaniak     "Speculative Store Bypass Disable (0 - off, 1 - on, 2 - auto)");
9743621ba1eSKonstantin Belousov 
975961aacb1SScott Long SYSCTL_PROC(_machdep_mitigations_ssb, OID_AUTO, disable, CTLTYPE_INT |
976961aacb1SScott Long     CTLFLAG_RWTUN | CTLFLAG_NOFETCH | CTLFLAG_MPSAFE, NULL, 0,
977961aacb1SScott Long     hw_ssb_disable_handler, "I",
978*a212f56dSPiotr Pawel Stefaniak     "Speculative Store Bypass Disable (0 - off, 1 - on, 2 - auto)");
979961aacb1SScott Long 
9807355a02bSKonstantin Belousov int hw_mds_disable;
9817355a02bSKonstantin Belousov 
9827355a02bSKonstantin Belousov /*
9837355a02bSKonstantin Belousov  * Handler for Microarchitectural Data Sampling issues.  Really not a
9847355a02bSKonstantin Belousov  * pointer to C function: on amd64 the code must not change any CPU
9857355a02bSKonstantin Belousov  * architectural state except possibly %rflags. Also, it is always
9867355a02bSKonstantin Belousov  * called with interrupts disabled.
9877355a02bSKonstantin Belousov  */
9887355a02bSKonstantin Belousov void mds_handler_void(void);
9897355a02bSKonstantin Belousov void mds_handler_verw(void);
9907355a02bSKonstantin Belousov void mds_handler_ivb(void);
9917355a02bSKonstantin Belousov void mds_handler_bdw(void);
9927355a02bSKonstantin Belousov void mds_handler_skl_sse(void);
9937355a02bSKonstantin Belousov void mds_handler_skl_avx(void);
9947355a02bSKonstantin Belousov void mds_handler_skl_avx512(void);
9957355a02bSKonstantin Belousov void mds_handler_silvermont(void);
996e2e0470dSKonstantin Belousov void (*mds_handler)(void) = mds_handler_void;
9977355a02bSKonstantin Belousov 
9987355a02bSKonstantin Belousov static int
9997355a02bSKonstantin Belousov sysctl_hw_mds_disable_state_handler(SYSCTL_HANDLER_ARGS)
10007355a02bSKonstantin Belousov {
10017355a02bSKonstantin Belousov 	const char *state;
10027355a02bSKonstantin Belousov 
10037355a02bSKonstantin Belousov 	if (mds_handler == mds_handler_void)
10047355a02bSKonstantin Belousov 		state = "inactive";
10057355a02bSKonstantin Belousov 	else if (mds_handler == mds_handler_verw)
10067355a02bSKonstantin Belousov 		state = "VERW";
10077355a02bSKonstantin Belousov 	else if (mds_handler == mds_handler_ivb)
10087355a02bSKonstantin Belousov 		state = "software IvyBridge";
10097355a02bSKonstantin Belousov 	else if (mds_handler == mds_handler_bdw)
10107355a02bSKonstantin Belousov 		state = "software Broadwell";
10117355a02bSKonstantin Belousov 	else if (mds_handler == mds_handler_skl_sse)
10127355a02bSKonstantin Belousov 		state = "software Skylake SSE";
10137355a02bSKonstantin Belousov 	else if (mds_handler == mds_handler_skl_avx)
10147355a02bSKonstantin Belousov 		state = "software Skylake AVX";
10157355a02bSKonstantin Belousov 	else if (mds_handler == mds_handler_skl_avx512)
10167355a02bSKonstantin Belousov 		state = "software Skylake AVX512";
10177355a02bSKonstantin Belousov 	else if (mds_handler == mds_handler_silvermont)
10187355a02bSKonstantin Belousov 		state = "software Silvermont";
10197355a02bSKonstantin Belousov 	else
10207355a02bSKonstantin Belousov 		state = "unknown";
10217355a02bSKonstantin Belousov 	return (SYSCTL_OUT(req, state, strlen(state)));
10227355a02bSKonstantin Belousov }
10237355a02bSKonstantin Belousov 
10247355a02bSKonstantin Belousov SYSCTL_PROC(_hw, OID_AUTO, mds_disable_state,
10257355a02bSKonstantin Belousov     CTLTYPE_STRING | CTLFLAG_RD | CTLFLAG_MPSAFE, NULL, 0,
10267355a02bSKonstantin Belousov     sysctl_hw_mds_disable_state_handler, "A",
10277355a02bSKonstantin Belousov     "Microarchitectural Data Sampling Mitigation state");
10287355a02bSKonstantin Belousov 
10297029da5cSPawel Biernacki SYSCTL_NODE(_machdep_mitigations, OID_AUTO, mds,
10307029da5cSPawel Biernacki     CTLFLAG_RW | CTLFLAG_MPSAFE, 0,
1031961aacb1SScott Long     "Microarchitectural Data Sampling Mitigation state");
1032961aacb1SScott Long 
1033961aacb1SScott Long SYSCTL_PROC(_machdep_mitigations_mds, OID_AUTO, state,
1034961aacb1SScott Long     CTLTYPE_STRING | CTLFLAG_RD | CTLFLAG_MPSAFE, NULL, 0,
1035961aacb1SScott Long     sysctl_hw_mds_disable_state_handler, "A",
1036961aacb1SScott Long     "Microarchitectural Data Sampling Mitigation state");
1037961aacb1SScott Long 
10387355a02bSKonstantin Belousov _Static_assert(__offsetof(struct pcpu, pc_mds_tmp) % 64 == 0, "MDS AVX512");
10397355a02bSKonstantin Belousov 
10407355a02bSKonstantin Belousov void
10417355a02bSKonstantin Belousov hw_mds_recalculate(void)
10427355a02bSKonstantin Belousov {
10437355a02bSKonstantin Belousov 	struct pcpu *pc;
10447355a02bSKonstantin Belousov 	vm_offset_t b64;
10457355a02bSKonstantin Belousov 	u_long xcr0;
10467355a02bSKonstantin Belousov 	int i;
10477355a02bSKonstantin Belousov 
10487355a02bSKonstantin Belousov 	/*
10497355a02bSKonstantin Belousov 	 * Allow user to force VERW variant even if MD_CLEAR is not
10507355a02bSKonstantin Belousov 	 * reported.  For instance, hypervisor might unknowingly
10517355a02bSKonstantin Belousov 	 * filter the cap out.
10527355a02bSKonstantin Belousov 	 * For the similar reasons, and for testing, allow to enable
105336e1ad61SKonstantin Belousov 	 * mitigation even when MDS_NO cap is set.
10547355a02bSKonstantin Belousov 	 */
10557355a02bSKonstantin Belousov 	if (cpu_vendor_id != CPU_VENDOR_INTEL || hw_mds_disable == 0 ||
105636e1ad61SKonstantin Belousov 	    ((cpu_ia32_arch_caps & IA32_ARCH_CAP_MDS_NO) != 0 &&
105736e1ad61SKonstantin Belousov 	    hw_mds_disable == 3)) {
10587355a02bSKonstantin Belousov 		mds_handler = mds_handler_void;
10597355a02bSKonstantin Belousov 	} else if (((cpu_stdext_feature3 & CPUID_STDEXT3_MD_CLEAR) != 0 &&
10607355a02bSKonstantin Belousov 	    hw_mds_disable == 3) || hw_mds_disable == 1) {
10617355a02bSKonstantin Belousov 		mds_handler = mds_handler_verw;
10627355a02bSKonstantin Belousov 	} else if (CPUID_TO_FAMILY(cpu_id) == 0x6 &&
10637355a02bSKonstantin Belousov 	    (CPUID_TO_MODEL(cpu_id) == 0x2e || CPUID_TO_MODEL(cpu_id) == 0x1e ||
10647355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x1f || CPUID_TO_MODEL(cpu_id) == 0x1a ||
10657355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x2f || CPUID_TO_MODEL(cpu_id) == 0x25 ||
10667355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x2c || CPUID_TO_MODEL(cpu_id) == 0x2d ||
10677355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x2a || CPUID_TO_MODEL(cpu_id) == 0x3e ||
10687355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x3a) &&
10697355a02bSKonstantin Belousov 	    (hw_mds_disable == 2 || hw_mds_disable == 3)) {
10707355a02bSKonstantin Belousov 		/*
10717355a02bSKonstantin Belousov 		 * Nehalem, SandyBridge, IvyBridge
10727355a02bSKonstantin Belousov 		 */
10737355a02bSKonstantin Belousov 		CPU_FOREACH(i) {
10747355a02bSKonstantin Belousov 			pc = pcpu_find(i);
10757355a02bSKonstantin Belousov 			if (pc->pc_mds_buf == NULL) {
10767355a02bSKonstantin Belousov 				pc->pc_mds_buf = malloc_domainset(672, M_TEMP,
10777355a02bSKonstantin Belousov 				    DOMAINSET_PREF(pc->pc_domain), M_WAITOK);
10787355a02bSKonstantin Belousov 				bzero(pc->pc_mds_buf, 16);
10797355a02bSKonstantin Belousov 			}
10807355a02bSKonstantin Belousov 		}
10817355a02bSKonstantin Belousov 		mds_handler = mds_handler_ivb;
10827355a02bSKonstantin Belousov 	} else if (CPUID_TO_FAMILY(cpu_id) == 0x6 &&
10837355a02bSKonstantin Belousov 	    (CPUID_TO_MODEL(cpu_id) == 0x3f || CPUID_TO_MODEL(cpu_id) == 0x3c ||
10847355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x45 || CPUID_TO_MODEL(cpu_id) == 0x46 ||
10857355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x56 || CPUID_TO_MODEL(cpu_id) == 0x4f ||
10867355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x47 || CPUID_TO_MODEL(cpu_id) == 0x3d) &&
10877355a02bSKonstantin Belousov 	    (hw_mds_disable == 2 || hw_mds_disable == 3)) {
10887355a02bSKonstantin Belousov 		/*
10897355a02bSKonstantin Belousov 		 * Haswell, Broadwell
10907355a02bSKonstantin Belousov 		 */
10917355a02bSKonstantin Belousov 		CPU_FOREACH(i) {
10927355a02bSKonstantin Belousov 			pc = pcpu_find(i);
10937355a02bSKonstantin Belousov 			if (pc->pc_mds_buf == NULL) {
10947355a02bSKonstantin Belousov 				pc->pc_mds_buf = malloc_domainset(1536, M_TEMP,
10957355a02bSKonstantin Belousov 				    DOMAINSET_PREF(pc->pc_domain), M_WAITOK);
10967355a02bSKonstantin Belousov 				bzero(pc->pc_mds_buf, 16);
10977355a02bSKonstantin Belousov 			}
10987355a02bSKonstantin Belousov 		}
10997355a02bSKonstantin Belousov 		mds_handler = mds_handler_bdw;
11007355a02bSKonstantin Belousov 	} else if (CPUID_TO_FAMILY(cpu_id) == 0x6 &&
11017355a02bSKonstantin Belousov 	    ((CPUID_TO_MODEL(cpu_id) == 0x55 && (cpu_id &
11027355a02bSKonstantin Belousov 	    CPUID_STEPPING) <= 5) ||
11037355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x4e || CPUID_TO_MODEL(cpu_id) == 0x5e ||
11047355a02bSKonstantin Belousov 	    (CPUID_TO_MODEL(cpu_id) == 0x8e && (cpu_id &
11057355a02bSKonstantin Belousov 	    CPUID_STEPPING) <= 0xb) ||
11067355a02bSKonstantin Belousov 	    (CPUID_TO_MODEL(cpu_id) == 0x9e && (cpu_id &
11077355a02bSKonstantin Belousov 	    CPUID_STEPPING) <= 0xc)) &&
11087355a02bSKonstantin Belousov 	    (hw_mds_disable == 2 || hw_mds_disable == 3)) {
11097355a02bSKonstantin Belousov 		/*
11107355a02bSKonstantin Belousov 		 * Skylake, KabyLake, CoffeeLake, WhiskeyLake,
11117355a02bSKonstantin Belousov 		 * CascadeLake
11127355a02bSKonstantin Belousov 		 */
11137355a02bSKonstantin Belousov 		CPU_FOREACH(i) {
11147355a02bSKonstantin Belousov 			pc = pcpu_find(i);
11157355a02bSKonstantin Belousov 			if (pc->pc_mds_buf == NULL) {
11167355a02bSKonstantin Belousov 				pc->pc_mds_buf = malloc_domainset(6 * 1024,
11177355a02bSKonstantin Belousov 				    M_TEMP, DOMAINSET_PREF(pc->pc_domain),
11187355a02bSKonstantin Belousov 				    M_WAITOK);
11197355a02bSKonstantin Belousov 				b64 = (vm_offset_t)malloc_domainset(64 + 63,
11207355a02bSKonstantin Belousov 				    M_TEMP, DOMAINSET_PREF(pc->pc_domain),
11217355a02bSKonstantin Belousov 				    M_WAITOK);
11227355a02bSKonstantin Belousov 				pc->pc_mds_buf64 = (void *)roundup2(b64, 64);
11237355a02bSKonstantin Belousov 				bzero(pc->pc_mds_buf64, 64);
11247355a02bSKonstantin Belousov 			}
11257355a02bSKonstantin Belousov 		}
11267355a02bSKonstantin Belousov 		xcr0 = rxcr(0);
11277355a02bSKonstantin Belousov 		if ((xcr0 & XFEATURE_ENABLED_ZMM_HI256) != 0 &&
112899a6085fSScott Long 		    (cpu_stdext_feature & CPUID_STDEXT_AVX512DQ) != 0)
11297355a02bSKonstantin Belousov 			mds_handler = mds_handler_skl_avx512;
11307355a02bSKonstantin Belousov 		else if ((xcr0 & XFEATURE_ENABLED_AVX) != 0 &&
11317355a02bSKonstantin Belousov 		    (cpu_feature2 & CPUID2_AVX) != 0)
11327355a02bSKonstantin Belousov 			mds_handler = mds_handler_skl_avx;
11337355a02bSKonstantin Belousov 		else
11347355a02bSKonstantin Belousov 			mds_handler = mds_handler_skl_sse;
11357355a02bSKonstantin Belousov 	} else if (CPUID_TO_FAMILY(cpu_id) == 0x6 &&
11367355a02bSKonstantin Belousov 	    ((CPUID_TO_MODEL(cpu_id) == 0x37 ||
11377355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x4a ||
11387355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x4c ||
11397355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x4d ||
11407355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x5a ||
11417355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x5d ||
11427355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x6e ||
11437355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x65 ||
11447355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x75 ||
11457355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x1c ||
11467355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x26 ||
11477355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x27 ||
11487355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x35 ||
11497355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x36 ||
11507355a02bSKonstantin Belousov 	    CPUID_TO_MODEL(cpu_id) == 0x7a))) {
11517355a02bSKonstantin Belousov 		/* Silvermont, Airmont */
11527355a02bSKonstantin Belousov 		CPU_FOREACH(i) {
11537355a02bSKonstantin Belousov 			pc = pcpu_find(i);
11547355a02bSKonstantin Belousov 			if (pc->pc_mds_buf == NULL)
11557355a02bSKonstantin Belousov 				pc->pc_mds_buf = malloc(256, M_TEMP, M_WAITOK);
11567355a02bSKonstantin Belousov 		}
11577355a02bSKonstantin Belousov 		mds_handler = mds_handler_silvermont;
11587355a02bSKonstantin Belousov 	} else {
11597355a02bSKonstantin Belousov 		hw_mds_disable = 0;
11607355a02bSKonstantin Belousov 		mds_handler = mds_handler_void;
11617355a02bSKonstantin Belousov 	}
11627355a02bSKonstantin Belousov }
11637355a02bSKonstantin Belousov 
116448ec6d3bSKonstantin Belousov static void
116548ec6d3bSKonstantin Belousov hw_mds_recalculate_boot(void *arg __unused)
116648ec6d3bSKonstantin Belousov {
116748ec6d3bSKonstantin Belousov 
116848ec6d3bSKonstantin Belousov 	hw_mds_recalculate();
116948ec6d3bSKonstantin Belousov }
117048ec6d3bSKonstantin Belousov SYSINIT(mds_recalc, SI_SUB_SMP, SI_ORDER_ANY, hw_mds_recalculate_boot, NULL);
117148ec6d3bSKonstantin Belousov 
11727355a02bSKonstantin Belousov static int
11737355a02bSKonstantin Belousov sysctl_mds_disable_handler(SYSCTL_HANDLER_ARGS)
11747355a02bSKonstantin Belousov {
11757355a02bSKonstantin Belousov 	int error, val;
11767355a02bSKonstantin Belousov 
11777355a02bSKonstantin Belousov 	val = hw_mds_disable;
11787355a02bSKonstantin Belousov 	error = sysctl_handle_int(oidp, &val, 0, req);
11797355a02bSKonstantin Belousov 	if (error != 0 || req->newptr == NULL)
11807355a02bSKonstantin Belousov 		return (error);
11817355a02bSKonstantin Belousov 	if (val < 0 || val > 3)
11827355a02bSKonstantin Belousov 		return (EINVAL);
11837355a02bSKonstantin Belousov 	hw_mds_disable = val;
11847355a02bSKonstantin Belousov 	hw_mds_recalculate();
11857355a02bSKonstantin Belousov 	return (0);
11867355a02bSKonstantin Belousov }
11877355a02bSKonstantin Belousov 
11887355a02bSKonstantin Belousov SYSCTL_PROC(_hw, OID_AUTO, mds_disable, CTLTYPE_INT |
11897355a02bSKonstantin Belousov     CTLFLAG_RWTUN | CTLFLAG_NOFETCH | CTLFLAG_MPSAFE, NULL, 0,
11907355a02bSKonstantin Belousov     sysctl_mds_disable_handler, "I",
11917355a02bSKonstantin Belousov     "Microarchitectural Data Sampling Mitigation "
1192*a212f56dSPiotr Pawel Stefaniak     "(0 - off, 1 - on VERW, 2 - on SW, 3 - on AUTO)");
11937355a02bSKonstantin Belousov 
1194961aacb1SScott Long SYSCTL_PROC(_machdep_mitigations_mds, OID_AUTO, disable, CTLTYPE_INT |
1195961aacb1SScott Long     CTLFLAG_RWTUN | CTLFLAG_NOFETCH | CTLFLAG_MPSAFE, NULL, 0,
1196961aacb1SScott Long     sysctl_mds_disable_handler, "I",
1197961aacb1SScott Long     "Microarchitectural Data Sampling Mitigation "
1198*a212f56dSPiotr Pawel Stefaniak     "(0 - off, 1 - on VERW, 2 - on SW, 3 - on AUTO)");
1199e3721601SScott Long 
1200e3721601SScott Long /*
1201e3721601SScott Long  * Intel Transactional Memory Asynchronous Abort Mitigation
1202e3721601SScott Long  * CVE-2019-11135
1203e3721601SScott Long  */
1204e3721601SScott Long int x86_taa_enable;
1205e3721601SScott Long int x86_taa_state;
1206e3721601SScott Long enum {
1207184b15ffSScott Long 	TAA_NONE	= 0,	/* No mitigation enabled */
1208184b15ffSScott Long 	TAA_TSX_DISABLE	= 1,	/* Disable TSX via MSR */
1209184b15ffSScott Long 	TAA_VERW	= 2,	/* Use VERW mitigation */
1210184b15ffSScott Long 	TAA_AUTO	= 3,	/* Automatically select the mitigation */
1211184b15ffSScott Long 
1212184b15ffSScott Long 	/* The states below are not selectable by the operator */
1213184b15ffSScott Long 
1214184b15ffSScott Long 	TAA_TAA_UC	= 4,	/* Mitigation present in microcode */
1215184b15ffSScott Long 	TAA_NOT_PRESENT	= 5	/* TSX is not present */
1216e3721601SScott Long };
1217e3721601SScott Long 
1218e3721601SScott Long static void
1219e3721601SScott Long taa_set(bool enable, bool all)
1220e3721601SScott Long {
1221e3721601SScott Long 
1222fa83f689SKonstantin Belousov 	x86_msr_op(MSR_IA32_TSX_CTRL,
1223fa83f689SKonstantin Belousov 	    (enable ? MSR_OP_OR : MSR_OP_ANDNOT) |
1224fa83f689SKonstantin Belousov 	    (all ? MSR_OP_RENDEZVOUS : MSR_OP_LOCAL),
1225fa83f689SKonstantin Belousov 	    IA32_TSX_CTRL_RTM_DISABLE | IA32_TSX_CTRL_TSX_CPUID_CLEAR);
1226e3721601SScott Long }
1227e3721601SScott Long 
1228e3721601SScott Long void
1229e3721601SScott Long x86_taa_recalculate(void)
1230e3721601SScott Long {
1231e3721601SScott Long 	static int taa_saved_mds_disable = 0;
1232e3721601SScott Long 	int taa_need = 0, taa_state = 0;
1233e3721601SScott Long 	int mds_disable = 0, need_mds_recalc = 0;
1234e3721601SScott Long 
1235e3721601SScott Long 	/* Check CPUID.07h.EBX.HLE and RTM for the presence of TSX */
1236e3721601SScott Long 	if ((cpu_stdext_feature & CPUID_STDEXT_HLE) == 0 ||
1237e3721601SScott Long 	    (cpu_stdext_feature & CPUID_STDEXT_RTM) == 0) {
1238e3721601SScott Long 		/* TSX is not present */
1239184b15ffSScott Long 		x86_taa_state = TAA_NOT_PRESENT;
1240e3721601SScott Long 		return;
1241e3721601SScott Long 	}
1242e3721601SScott Long 
1243e3721601SScott Long 	/* Check to see what mitigation options the CPU gives us */
1244e3721601SScott Long 	if (cpu_ia32_arch_caps & IA32_ARCH_CAP_TAA_NO) {
1245e3721601SScott Long 		/* CPU is not suseptible to TAA */
12460d423176SScott Long 		taa_need = TAA_TAA_UC;
1247e3721601SScott Long 	} else if (cpu_ia32_arch_caps & IA32_ARCH_CAP_TSX_CTRL) {
1248e3721601SScott Long 		/*
1249e3721601SScott Long 		 * CPU can turn off TSX.  This is the next best option
1250e3721601SScott Long 		 * if TAA_NO hardware mitigation isn't present
1251e3721601SScott Long 		 */
1252e3721601SScott Long 		taa_need = TAA_TSX_DISABLE;
1253e3721601SScott Long 	} else {
1254e3721601SScott Long 		/* No TSX/TAA specific remedies are available. */
1255e3721601SScott Long 		if (x86_taa_enable == TAA_TSX_DISABLE) {
1256e3721601SScott Long 			if (bootverbose)
1257e3721601SScott Long 				printf("TSX control not available\n");
1258e3721601SScott Long 			return;
1259e3721601SScott Long 		} else
1260e3721601SScott Long 			taa_need = TAA_VERW;
1261e3721601SScott Long 	}
1262e3721601SScott Long 
1263e3721601SScott Long 	/* Can we automatically take action, or are we being forced? */
1264e3721601SScott Long 	if (x86_taa_enable == TAA_AUTO)
1265e3721601SScott Long 		taa_state = taa_need;
1266e3721601SScott Long 	else
1267e3721601SScott Long 		taa_state = x86_taa_enable;
1268e3721601SScott Long 
1269e3721601SScott Long 	/* No state change, nothing to do */
1270e3721601SScott Long 	if (taa_state == x86_taa_state) {
1271e3721601SScott Long 		if (bootverbose)
1272e3721601SScott Long 			printf("No TSX change made\n");
1273e3721601SScott Long 		return;
1274e3721601SScott Long 	}
1275e3721601SScott Long 
1276e3721601SScott Long 	/* Does the MSR need to be turned on or off? */
1277e3721601SScott Long 	if (taa_state == TAA_TSX_DISABLE)
1278e3721601SScott Long 		taa_set(true, true);
1279e3721601SScott Long 	else if (x86_taa_state == TAA_TSX_DISABLE)
1280e3721601SScott Long 		taa_set(false, true);
1281e3721601SScott Long 
1282e3721601SScott Long 	/* Does MDS need to be set to turn on VERW? */
1283e3721601SScott Long 	if (taa_state == TAA_VERW) {
1284e3721601SScott Long 		taa_saved_mds_disable = hw_mds_disable;
1285e3721601SScott Long 		mds_disable = hw_mds_disable = 1;
1286e3721601SScott Long 		need_mds_recalc = 1;
1287e3721601SScott Long 	} else if (x86_taa_state == TAA_VERW) {
1288e3721601SScott Long 		mds_disable = hw_mds_disable = taa_saved_mds_disable;
1289e3721601SScott Long 		need_mds_recalc = 1;
1290e3721601SScott Long 	}
1291e3721601SScott Long 	if (need_mds_recalc) {
1292e3721601SScott Long 		hw_mds_recalculate();
1293e3721601SScott Long 		if (mds_disable != hw_mds_disable) {
1294e3721601SScott Long 			if (bootverbose)
1295e3721601SScott Long 				printf("Cannot change MDS state for TAA\n");
1296e3721601SScott Long 			/* Don't update our state */
1297e3721601SScott Long 			return;
1298e3721601SScott Long 		}
1299e3721601SScott Long 	}
1300e3721601SScott Long 
1301e3721601SScott Long 	x86_taa_state = taa_state;
1302e3721601SScott Long 	return;
1303e3721601SScott Long }
1304e3721601SScott Long 
1305e3721601SScott Long static void
1306e3721601SScott Long taa_recalculate_boot(void * arg __unused)
1307e3721601SScott Long {
1308e3721601SScott Long 
1309e3721601SScott Long 	x86_taa_recalculate();
1310e3721601SScott Long }
1311e3721601SScott Long SYSINIT(taa_recalc, SI_SUB_SMP, SI_ORDER_ANY, taa_recalculate_boot, NULL);
1312e3721601SScott Long 
13137029da5cSPawel Biernacki SYSCTL_NODE(_machdep_mitigations, OID_AUTO, taa,
13147029da5cSPawel Biernacki     CTLFLAG_RW | CTLFLAG_MPSAFE, 0,
1315e3721601SScott Long     "TSX Asynchronous Abort Mitigation");
1316e3721601SScott Long 
1317e3721601SScott Long static int
1318e3721601SScott Long sysctl_taa_handler(SYSCTL_HANDLER_ARGS)
1319e3721601SScott Long {
1320e3721601SScott Long 	int error, val;
1321e3721601SScott Long 
1322e3721601SScott Long 	val = x86_taa_enable;
1323e3721601SScott Long 	error = sysctl_handle_int(oidp, &val, 0, req);
1324e3721601SScott Long 	if (error != 0 || req->newptr == NULL)
1325e3721601SScott Long 		return (error);
1326e3721601SScott Long 	if (val < TAA_NONE || val > TAA_AUTO)
1327e3721601SScott Long 		return (EINVAL);
1328e3721601SScott Long 	x86_taa_enable = val;
1329e3721601SScott Long 	x86_taa_recalculate();
1330e3721601SScott Long 	return (0);
1331e3721601SScott Long }
1332e3721601SScott Long 
1333e3721601SScott Long SYSCTL_PROC(_machdep_mitigations_taa, OID_AUTO, enable, CTLTYPE_INT |
1334e3721601SScott Long     CTLFLAG_RWTUN | CTLFLAG_NOFETCH | CTLFLAG_MPSAFE, NULL, 0,
1335e3721601SScott Long     sysctl_taa_handler, "I",
1336e3721601SScott Long     "TAA Mitigation enablement control "
1337*a212f56dSPiotr Pawel Stefaniak     "(0 - off, 1 - disable TSX, 2 - VERW, 3 - on AUTO)");
1338e3721601SScott Long 
1339e3721601SScott Long static int
1340e3721601SScott Long sysctl_taa_state_handler(SYSCTL_HANDLER_ARGS)
1341e3721601SScott Long {
1342e3721601SScott Long 	const char *state;
1343e3721601SScott Long 
1344e3721601SScott Long 	switch (x86_taa_state) {
1345e3721601SScott Long 	case TAA_NONE:
1346e3721601SScott Long 		state = "inactive";
1347e3721601SScott Long 		break;
1348e3721601SScott Long 	case TAA_TSX_DISABLE:
1349e3721601SScott Long 		state = "TSX disabled";
1350e3721601SScott Long 		break;
1351e3721601SScott Long 	case TAA_VERW:
1352e3721601SScott Long 		state = "VERW";
1353e3721601SScott Long 		break;
1354184b15ffSScott Long 	case TAA_TAA_UC:
1355184b15ffSScott Long 		state = "Mitigated in microcode";
1356e3721601SScott Long 		break;
1357184b15ffSScott Long 	case TAA_NOT_PRESENT:
1358184b15ffSScott Long 		state = "TSX not present";
1359ee02bd9cSConrad Meyer 		break;
1360e3721601SScott Long 	default:
1361e3721601SScott Long 		state = "unknown";
1362e3721601SScott Long 	}
1363e3721601SScott Long 
1364e3721601SScott Long 	return (SYSCTL_OUT(req, state, strlen(state)));
1365e3721601SScott Long }
1366e3721601SScott Long 
1367e3721601SScott Long SYSCTL_PROC(_machdep_mitigations_taa, OID_AUTO, state,
1368e3721601SScott Long     CTLTYPE_STRING | CTLFLAG_RD | CTLFLAG_MPSAFE, NULL, 0,
1369e3721601SScott Long     sysctl_taa_state_handler, "A",
1370e3721601SScott Long     "TAA Mitigation state");
1371e3721601SScott Long 
1372ea602083SKonstantin Belousov int __read_frequently cpu_flush_rsb_ctxsw;
1373ea602083SKonstantin Belousov SYSCTL_INT(_machdep_mitigations, OID_AUTO, flush_rsb_ctxsw,
1374ea602083SKonstantin Belousov     CTLFLAG_RW | CTLFLAG_NOFETCH, &cpu_flush_rsb_ctxsw, 0,
1375ea602083SKonstantin Belousov     "Flush Return Stack Buffer on context switch");
1376ea602083SKonstantin Belousov 
137717edf152SKonstantin Belousov SYSCTL_NODE(_machdep_mitigations, OID_AUTO, rngds,
137817edf152SKonstantin Belousov     CTLFLAG_RW | CTLFLAG_MPSAFE, 0,
137917edf152SKonstantin Belousov     "MCU Optimization, disable RDSEED mitigation");
138017edf152SKonstantin Belousov 
138117edf152SKonstantin Belousov int x86_rngds_mitg_enable = 1;
138217edf152SKonstantin Belousov void
138317edf152SKonstantin Belousov x86_rngds_mitg_recalculate(bool all_cpus)
138417edf152SKonstantin Belousov {
138517edf152SKonstantin Belousov 	if ((cpu_stdext_feature3 & CPUID_STDEXT3_MCUOPT) == 0)
138617edf152SKonstantin Belousov 		return;
138717edf152SKonstantin Belousov 	x86_msr_op(MSR_IA32_MCU_OPT_CTRL,
138817edf152SKonstantin Belousov 	    (x86_rngds_mitg_enable ? MSR_OP_OR : MSR_OP_ANDNOT) |
138917edf152SKonstantin Belousov 	    (all_cpus ? MSR_OP_RENDEZVOUS : MSR_OP_LOCAL),
139017edf152SKonstantin Belousov 	    IA32_RNGDS_MITG_DIS);
139117edf152SKonstantin Belousov }
139217edf152SKonstantin Belousov 
139317edf152SKonstantin Belousov static int
139417edf152SKonstantin Belousov sysctl_rngds_mitg_enable_handler(SYSCTL_HANDLER_ARGS)
139517edf152SKonstantin Belousov {
139617edf152SKonstantin Belousov 	int error, val;
139717edf152SKonstantin Belousov 
139817edf152SKonstantin Belousov 	val = x86_rngds_mitg_enable;
139917edf152SKonstantin Belousov 	error = sysctl_handle_int(oidp, &val, 0, req);
140017edf152SKonstantin Belousov 	if (error != 0 || req->newptr == NULL)
140117edf152SKonstantin Belousov 		return (error);
140217edf152SKonstantin Belousov 	x86_rngds_mitg_enable = val;
140317edf152SKonstantin Belousov 	x86_rngds_mitg_recalculate(true);
140417edf152SKonstantin Belousov 	return (0);
140517edf152SKonstantin Belousov }
140617edf152SKonstantin Belousov SYSCTL_PROC(_machdep_mitigations_rngds, OID_AUTO, enable, CTLTYPE_INT |
140717edf152SKonstantin Belousov     CTLFLAG_RWTUN | CTLFLAG_NOFETCH | CTLFLAG_MPSAFE, NULL, 0,
140817edf152SKonstantin Belousov     sysctl_rngds_mitg_enable_handler, "I",
140917edf152SKonstantin Belousov     "MCU Optimization, disabling RDSEED mitigation control "
1410*a212f56dSPiotr Pawel Stefaniak     "(0 - mitigation disabled (RDSEED optimized), 1 - mitigation enabled)");
141117edf152SKonstantin Belousov 
141217edf152SKonstantin Belousov static int
141317edf152SKonstantin Belousov sysctl_rngds_state_handler(SYSCTL_HANDLER_ARGS)
141417edf152SKonstantin Belousov {
141517edf152SKonstantin Belousov 	const char *state;
141617edf152SKonstantin Belousov 
141717edf152SKonstantin Belousov 	if ((cpu_stdext_feature3 & CPUID_STDEXT3_MCUOPT) == 0) {
141817edf152SKonstantin Belousov 		state = "Not applicable";
141917edf152SKonstantin Belousov 	} else if (x86_rngds_mitg_enable == 0) {
142017edf152SKonstantin Belousov 		state = "RDSEED not serialized";
142117edf152SKonstantin Belousov 	} else {
142217edf152SKonstantin Belousov 		state = "Mitigated";
142317edf152SKonstantin Belousov 	}
142417edf152SKonstantin Belousov 	return (SYSCTL_OUT(req, state, strlen(state)));
142517edf152SKonstantin Belousov }
142617edf152SKonstantin Belousov SYSCTL_PROC(_machdep_mitigations_rngds, OID_AUTO, state,
142717edf152SKonstantin Belousov     CTLTYPE_STRING | CTLFLAG_RD | CTLFLAG_MPSAFE, NULL, 0,
142817edf152SKonstantin Belousov     sysctl_rngds_state_handler, "A",
142917edf152SKonstantin Belousov     "MCU Optimization state");
143017edf152SKonstantin Belousov 
14318fbcc334SKonstantin Belousov /*
14328fbcc334SKonstantin Belousov  * Enable and restore kernel text write permissions.
14338fbcc334SKonstantin Belousov  * Callers must ensure that disable_wp()/restore_wp() are executed
14348fbcc334SKonstantin Belousov  * without rescheduling on the same core.
14358fbcc334SKonstantin Belousov  */
14368fbcc334SKonstantin Belousov bool
14378fbcc334SKonstantin Belousov disable_wp(void)
14388fbcc334SKonstantin Belousov {
14398fbcc334SKonstantin Belousov 	u_int cr0;
14408fbcc334SKonstantin Belousov 
14418fbcc334SKonstantin Belousov 	cr0 = rcr0();
14428fbcc334SKonstantin Belousov 	if ((cr0 & CR0_WP) == 0)
14438fbcc334SKonstantin Belousov 		return (false);
14448fbcc334SKonstantin Belousov 	load_cr0(cr0 & ~CR0_WP);
14458fbcc334SKonstantin Belousov 	return (true);
14468fbcc334SKonstantin Belousov }
14478fbcc334SKonstantin Belousov 
14488fbcc334SKonstantin Belousov void
14498fbcc334SKonstantin Belousov restore_wp(bool old_wp)
14508fbcc334SKonstantin Belousov {
14518fbcc334SKonstantin Belousov 
14528fbcc334SKonstantin Belousov 	if (old_wp)
14538fbcc334SKonstantin Belousov 		load_cr0(rcr0() | CR0_WP);
14548fbcc334SKonstantin Belousov }
14558fbcc334SKonstantin Belousov 
14567705dd4dSKonstantin Belousov bool
14577705dd4dSKonstantin Belousov acpi_get_fadt_bootflags(uint16_t *flagsp)
14587705dd4dSKonstantin Belousov {
14597705dd4dSKonstantin Belousov #ifdef DEV_ACPI
14607705dd4dSKonstantin Belousov 	ACPI_TABLE_FADT *fadt;
14617705dd4dSKonstantin Belousov 	vm_paddr_t physaddr;
14627705dd4dSKonstantin Belousov 
14637705dd4dSKonstantin Belousov 	physaddr = acpi_find_table(ACPI_SIG_FADT);
14647705dd4dSKonstantin Belousov 	if (physaddr == 0)
14657705dd4dSKonstantin Belousov 		return (false);
14667705dd4dSKonstantin Belousov 	fadt = acpi_map_table(physaddr, ACPI_SIG_FADT);
14677705dd4dSKonstantin Belousov 	if (fadt == NULL)
14687705dd4dSKonstantin Belousov 		return (false);
14697705dd4dSKonstantin Belousov 	*flagsp = fadt->BootFlags;
14707705dd4dSKonstantin Belousov 	acpi_unmap_table(fadt);
14717705dd4dSKonstantin Belousov 	return (true);
14727705dd4dSKonstantin Belousov #else
14737705dd4dSKonstantin Belousov 	return (false);
14747705dd4dSKonstantin Belousov #endif
14757705dd4dSKonstantin Belousov }
1476