xref: /freebsd/sys/arm64/vmm/vmm_reset.c (revision d54631360276d3fdbaa9a7872f8af82f1f4287da)
1 /*-
2  * SPDX-License-Identifier: BSD-2-Clause
3  *
4  * Copyright (C) 2018 Alexandru Elisei <alexandru.elisei@gmail.com>
5  *
6  * Redistribution and use in source and binary forms, with or without
7  * modification, are permitted provided that the following conditions
8  * are met:
9  * 1. Redistributions of source code must retain the above copyright
10  *    notice, this list of conditions and the following disclaimer.
11  * 2. Redistributions in binary form must reproduce the above copyright
12  *    notice, this list of conditions and the following disclaimer in the
13  *    documentation and/or other materials provided with the distribution.
14  *
15  * THIS SOFTWARE IS PROVIDED BY AUTHOR AND CONTRIBUTORS ``AS IS'' AND
16  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18  * ARE DISCLAIMED.  IN NO EVENT SHALL AUTHOR OR CONTRIBUTORS BE LIABLE
19  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
20  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
21  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
22  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
23  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
24  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
25  * SUCH DAMAGE.
26  */
27 
28 #include <sys/cdefs.h>
29 #include <sys/types.h>
30 #include <sys/systm.h>
31 #include <sys/kernel.h>
32 #include <sys/lock.h>
33 
34 #include <machine/armreg.h>
35 #include <machine/cpu.h>
36 #include <machine/hypervisor.h>
37 
38 #include "arm64.h"
39 #include "reset.h"
40 
41 /*
42  * Make the architecturally UNKNOWN value 0. As a bonus, we don't have to
43  * manually set all those RES0 fields.
44  */
45 #define	ARCH_UNKNOWN		0
46 #define	set_arch_unknown(reg)	(memset(&(reg), ARCH_UNKNOWN, sizeof(reg)))
47 
48 void
reset_vm_el01_regs(void * vcpu)49 reset_vm_el01_regs(void *vcpu)
50 {
51 	struct hypctx *el2ctx;
52 
53 	el2ctx = vcpu;
54 
55 	set_arch_unknown(el2ctx->tf);
56 
57 	set_arch_unknown(el2ctx->actlr_el1);
58 	set_arch_unknown(el2ctx->afsr0_el1);
59 	set_arch_unknown(el2ctx->afsr1_el1);
60 	set_arch_unknown(el2ctx->amair_el1);
61 	set_arch_unknown(el2ctx->contextidr_el1);
62 	set_arch_unknown(el2ctx->cpacr_el1);
63 	set_arch_unknown(el2ctx->csselr_el1);
64 	set_arch_unknown(el2ctx->elr_el1);
65 	set_arch_unknown(el2ctx->esr_el1);
66 	set_arch_unknown(el2ctx->far_el1);
67 	set_arch_unknown(el2ctx->mair_el1);
68 	set_arch_unknown(el2ctx->mdccint_el1);
69 	set_arch_unknown(el2ctx->mdscr_el1);
70 	set_arch_unknown(el2ctx->par_el1);
71 
72 	/*
73 	 * Guest starts with:
74 	 * ~SCTLR_M: MMU off
75 	 * ~SCTLR_C: data cache off
76 	 * SCTLR_CP15BEN: memory barrier instruction enable from EL0; RAO/WI
77 	 * ~SCTLR_I: instruction cache off
78 	 */
79 	el2ctx->sctlr_el1 = SCTLR_RES1;
80 	el2ctx->sctlr_el1 &= ~SCTLR_M & ~SCTLR_C & ~SCTLR_I;
81 	el2ctx->sctlr_el1 |= SCTLR_CP15BEN;
82 
83 	set_arch_unknown(el2ctx->sp_el0);
84 	set_arch_unknown(el2ctx->tcr_el1);
85 	set_arch_unknown(el2ctx->tpidr_el0);
86 	set_arch_unknown(el2ctx->tpidr_el1);
87 	set_arch_unknown(el2ctx->tpidrro_el0);
88 	set_arch_unknown(el2ctx->ttbr0_el1);
89 	set_arch_unknown(el2ctx->ttbr1_el1);
90 	set_arch_unknown(el2ctx->vbar_el1);
91 	set_arch_unknown(el2ctx->spsr_el1);
92 
93 	set_arch_unknown(el2ctx->dbgbcr_el1);
94 	set_arch_unknown(el2ctx->dbgbvr_el1);
95 	set_arch_unknown(el2ctx->dbgwcr_el1);
96 	set_arch_unknown(el2ctx->dbgwvr_el1);
97 
98 	el2ctx->pmcr_el0 = READ_SPECIALREG(pmcr_el0) & PMCR_N_MASK;
99 	/* PMCR_LC is unknown when AArch32 is supported or RES1 otherwise */
100 	el2ctx->pmcr_el0 |= PMCR_LC;
101 	set_arch_unknown(el2ctx->pmccntr_el0);
102 	set_arch_unknown(el2ctx->pmccfiltr_el0);
103 	set_arch_unknown(el2ctx->pmcntenset_el0);
104 	set_arch_unknown(el2ctx->pmintenset_el1);
105 	set_arch_unknown(el2ctx->pmovsset_el0);
106 	set_arch_unknown(el2ctx->pmuserenr_el0);
107 	memset(el2ctx->pmevcntr_el0, 0, sizeof(el2ctx->pmevcntr_el0));
108 	memset(el2ctx->pmevtyper_el0, 0, sizeof(el2ctx->pmevtyper_el0));
109 }
110 
111 void
reset_vm_el2_regs(void * vcpu)112 reset_vm_el2_regs(void *vcpu)
113 {
114 	struct hypctx *el2ctx;
115 	uint64_t cpu_aff, vcpuid;
116 
117 	el2ctx = vcpu;
118 	vcpuid = vcpu_vcpuid(el2ctx->vcpu);
119 
120 	/*
121 	 * Set the Hypervisor Configuration Register:
122 	 *
123 	 * HCR_RW: use AArch64 for EL1
124 	 * HCR_TID3: handle ID registers in the vmm to privide a common
125 	 * set of featers on all vcpus
126 	 * HCR_TWI: Trap WFI to the hypervisor
127 	 * HCR_BSU_IS: barrier instructions apply to the inner shareable
128 	 * domain
129 	 * HCR_FB: broadcast maintenance operations
130 	 * HCR_AMO: route physical SError interrupts to EL2
131 	 * HCR_IMO: route physical IRQ interrupts to EL2
132 	 * HCR_FMO: route physical FIQ interrupts to EL2
133 	 * HCR_SWIO: turn set/way invalidate into set/way clean and
134 	 * invalidate
135 	 * HCR_VM: use stage 2 translation
136 	 */
137 	el2ctx->hcr_el2 = HCR_RW | HCR_TID3 | HCR_TWI | HCR_BSU_IS | HCR_FB |
138 	    HCR_AMO | HCR_IMO | HCR_FMO | HCR_SWIO | HCR_VM;
139 	if (in_vhe()) {
140 		el2ctx->hcr_el2 |= HCR_E2H;
141 	}
142 
143 	/* TODO: Trap all extensions we don't support */
144 	el2ctx->mdcr_el2 = 0;
145 	/* PMCR_EL0.N is read from MDCR_EL2.HPMN */
146 	el2ctx->mdcr_el2 |= (el2ctx->pmcr_el0 & PMCR_N_MASK) >> PMCR_N_SHIFT;
147 
148 	el2ctx->vmpidr_el2 = VMPIDR_EL2_RES1;
149 	/* The guest will detect a multi-core, single-threaded CPU */
150 	el2ctx->vmpidr_el2 &= ~VMPIDR_EL2_U & ~VMPIDR_EL2_MT;
151 	/*
152 	 * Generate the guest MPIDR value. We only support 16 CPUs at affinity
153 	 * level 0 to simplify the vgicv3 driver (see writing sgi1r_el1).
154 	 */
155 	cpu_aff = (vcpuid & 0xf) << MPIDR_AFF0_SHIFT |
156 	    ((vcpuid >> 4) & 0xff) << MPIDR_AFF1_SHIFT |
157 	    ((vcpuid >> 12) & 0xff) << MPIDR_AFF2_SHIFT |
158 	    ((vcpuid >> 20) & 0xff) << MPIDR_AFF3_SHIFT;
159 	el2ctx->vmpidr_el2 |= cpu_aff;
160 
161 	/* Use the same CPU identification information as the host */
162 	el2ctx->vpidr_el2 = CPU_IMPL_TO_MIDR(CPU_IMPL_ARM);
163 	el2ctx->vpidr_el2 |= CPU_VAR_TO_MIDR(0);
164 	el2ctx->vpidr_el2 |= CPU_ARCH_TO_MIDR(0xf);
165 	el2ctx->vpidr_el2 |= CPU_PART_TO_MIDR(CPU_PART_FOUNDATION);
166 	el2ctx->vpidr_el2 |= CPU_REV_TO_MIDR(0);
167 
168 	/*
169 	 * Don't trap accesses to CPACR_EL1, trace, SVE, Advanced SIMD
170 	 * and floating point functionality to EL2.
171 	 */
172 	if (in_vhe())
173 		el2ctx->cptr_el2 = CPTR_E2H_TRAP_ALL | CPTR_E2H_FPEN;
174 	else
175 		el2ctx->cptr_el2 = CPTR_TRAP_ALL & ~CPTR_TFP;
176 	el2ctx->cptr_el2 &= ~CPTR_TCPAC;
177 	/*
178 	 * Disable interrupts in the guest. The guest OS will re-enable
179 	 * them.
180 	 */
181 	el2ctx->tf.tf_spsr = PSR_D | PSR_A | PSR_I | PSR_F;
182 	/* Use the EL1 stack when taking exceptions to EL1 */
183 	el2ctx->tf.tf_spsr |= PSR_M_EL1h;
184 }
185