xref: /linux/arch/x86/hyperv/hv_vtl.c (revision dcb7600849ce9b3d9b3d2965f452287f06fc9093)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (c) 2023, Microsoft Corporation.
4  *
5  * Author:
6  *   Saurabh Sengar <ssengar@microsoft.com>
7  */
8 
9 #include <asm/apic.h>
10 #include <asm/boot.h>
11 #include <asm/desc.h>
12 #include <asm/i8259.h>
13 #include <asm/mshyperv.h>
14 #include <asm/realmode.h>
15 
16 extern struct boot_params boot_params;
17 static struct real_mode_header hv_vtl_real_mode_header;
18 
19 void __init hv_vtl_init_platform(void)
20 {
21 	pr_info("Linux runs in Hyper-V Virtual Trust Level\n");
22 
23 	x86_platform.realmode_reserve = x86_init_noop;
24 	x86_platform.realmode_init = x86_init_noop;
25 	x86_init.irqs.pre_vector_init = x86_init_noop;
26 	x86_init.timers.timer_init = x86_init_noop;
27 
28 	/* Avoid searching for BIOS MP tables */
29 	x86_init.mpparse.find_mptable = x86_init_noop;
30 	x86_init.mpparse.early_parse_smp_cfg = x86_init_noop;
31 	x86_init.mpparse.parse_smp_cfg = x86_init_noop;
32 
33 	x86_platform.get_wallclock = get_rtc_noop;
34 	x86_platform.set_wallclock = set_rtc_noop;
35 	x86_platform.get_nmi_reason = hv_get_nmi_reason;
36 
37 	x86_platform.legacy.i8042 = X86_LEGACY_I8042_PLATFORM_ABSENT;
38 	x86_platform.legacy.rtc = 0;
39 	x86_platform.legacy.warm_reset = 0;
40 	x86_platform.legacy.reserve_bios_regions = 0;
41 	x86_platform.legacy.devices.pnpbios = 0;
42 }
43 
44 static inline u64 hv_vtl_system_desc_base(struct ldttss_desc *desc)
45 {
46 	return ((u64)desc->base3 << 32) | ((u64)desc->base2 << 24) |
47 		(desc->base1 << 16) | desc->base0;
48 }
49 
50 static inline u32 hv_vtl_system_desc_limit(struct ldttss_desc *desc)
51 {
52 	return ((u32)desc->limit1 << 16) | (u32)desc->limit0;
53 }
54 
55 typedef void (*secondary_startup_64_fn)(void*, void*);
56 static void hv_vtl_ap_entry(void)
57 {
58 	((secondary_startup_64_fn)secondary_startup_64)(&boot_params, &boot_params);
59 }
60 
61 static int hv_vtl_bringup_vcpu(u32 target_vp_index, u64 eip_ignored)
62 {
63 	u64 status;
64 	int ret = 0;
65 	struct hv_enable_vp_vtl *input;
66 	unsigned long irq_flags;
67 
68 	struct desc_ptr gdt_ptr;
69 	struct desc_ptr idt_ptr;
70 
71 	struct ldttss_desc *tss;
72 	struct ldttss_desc *ldt;
73 	struct desc_struct *gdt;
74 
75 	u64 rsp = current->thread.sp;
76 	u64 rip = (u64)&hv_vtl_ap_entry;
77 
78 	native_store_gdt(&gdt_ptr);
79 	store_idt(&idt_ptr);
80 
81 	gdt = (struct desc_struct *)((void *)(gdt_ptr.address));
82 	tss = (struct ldttss_desc *)(gdt + GDT_ENTRY_TSS);
83 	ldt = (struct ldttss_desc *)(gdt + GDT_ENTRY_LDT);
84 
85 	local_irq_save(irq_flags);
86 
87 	input = *this_cpu_ptr(hyperv_pcpu_input_arg);
88 	memset(input, 0, sizeof(*input));
89 
90 	input->partition_id = HV_PARTITION_ID_SELF;
91 	input->vp_index = target_vp_index;
92 	input->target_vtl.target_vtl = HV_VTL_MGMT;
93 
94 	/*
95 	 * The x86_64 Linux kernel follows the 16-bit -> 32-bit -> 64-bit
96 	 * mode transition sequence after waking up an AP with SIPI whose
97 	 * vector points to the 16-bit AP startup trampoline code. Here in
98 	 * VTL2, we can't perform that sequence as the AP has to start in
99 	 * the 64-bit mode.
100 	 *
101 	 * To make this happen, we tell the hypervisor to load a valid 64-bit
102 	 * context (most of which is just magic numbers from the CPU manual)
103 	 * so that AP jumps right to the 64-bit entry of the kernel, and the
104 	 * control registers are loaded with values that let the AP fetch the
105 	 * code and data and carry on with work it gets assigned.
106 	 */
107 
108 	input->vp_context.rip = rip;
109 	input->vp_context.rsp = rsp;
110 	input->vp_context.rflags = 0x0000000000000002;
111 	input->vp_context.efer = __rdmsr(MSR_EFER);
112 	input->vp_context.cr0 = native_read_cr0();
113 	input->vp_context.cr3 = __native_read_cr3();
114 	input->vp_context.cr4 = native_read_cr4();
115 	input->vp_context.msr_cr_pat = __rdmsr(MSR_IA32_CR_PAT);
116 	input->vp_context.idtr.limit = idt_ptr.size;
117 	input->vp_context.idtr.base = idt_ptr.address;
118 	input->vp_context.gdtr.limit = gdt_ptr.size;
119 	input->vp_context.gdtr.base = gdt_ptr.address;
120 
121 	/* Non-system desc (64bit), long, code, present */
122 	input->vp_context.cs.selector = __KERNEL_CS;
123 	input->vp_context.cs.base = 0;
124 	input->vp_context.cs.limit = 0xffffffff;
125 	input->vp_context.cs.attributes = 0xa09b;
126 	/* Non-system desc (64bit), data, present, granularity, default */
127 	input->vp_context.ss.selector = __KERNEL_DS;
128 	input->vp_context.ss.base = 0;
129 	input->vp_context.ss.limit = 0xffffffff;
130 	input->vp_context.ss.attributes = 0xc093;
131 
132 	/* System desc (128bit), present, LDT */
133 	input->vp_context.ldtr.selector = GDT_ENTRY_LDT * 8;
134 	input->vp_context.ldtr.base = hv_vtl_system_desc_base(ldt);
135 	input->vp_context.ldtr.limit = hv_vtl_system_desc_limit(ldt);
136 	input->vp_context.ldtr.attributes = 0x82;
137 
138 	/* System desc (128bit), present, TSS, 0x8b - busy, 0x89 -- default */
139 	input->vp_context.tr.selector = GDT_ENTRY_TSS * 8;
140 	input->vp_context.tr.base = hv_vtl_system_desc_base(tss);
141 	input->vp_context.tr.limit = hv_vtl_system_desc_limit(tss);
142 	input->vp_context.tr.attributes = 0x8b;
143 
144 	status = hv_do_hypercall(HVCALL_ENABLE_VP_VTL, input, NULL);
145 
146 	if (!hv_result_success(status) &&
147 	    hv_result(status) != HV_STATUS_VTL_ALREADY_ENABLED) {
148 		pr_err("HVCALL_ENABLE_VP_VTL failed for VP : %d ! [Err: %#llx\n]",
149 		       target_vp_index, status);
150 		ret = -EINVAL;
151 		goto free_lock;
152 	}
153 
154 	status = hv_do_hypercall(HVCALL_START_VP, input, NULL);
155 
156 	if (!hv_result_success(status)) {
157 		pr_err("HVCALL_START_VP failed for VP : %d ! [Err: %#llx]\n",
158 		       target_vp_index, status);
159 		ret = -EINVAL;
160 	}
161 
162 free_lock:
163 	local_irq_restore(irq_flags);
164 
165 	return ret;
166 }
167 
168 static int hv_vtl_apicid_to_vp_id(u32 apic_id)
169 {
170 	u64 control;
171 	u64 status;
172 	unsigned long irq_flags;
173 	struct hv_get_vp_from_apic_id_in *input;
174 	u32 *output, ret;
175 
176 	local_irq_save(irq_flags);
177 
178 	input = *this_cpu_ptr(hyperv_pcpu_input_arg);
179 	memset(input, 0, sizeof(*input));
180 	input->partition_id = HV_PARTITION_ID_SELF;
181 	input->apic_ids[0] = apic_id;
182 
183 	output = (u32 *)input;
184 
185 	control = HV_HYPERCALL_REP_COMP_1 | HVCALL_GET_VP_ID_FROM_APIC_ID;
186 	status = hv_do_hypercall(control, input, output);
187 	ret = output[0];
188 
189 	local_irq_restore(irq_flags);
190 
191 	if (!hv_result_success(status)) {
192 		pr_err("failed to get vp id from apic id %d, status %#llx\n",
193 		       apic_id, status);
194 		return -EINVAL;
195 	}
196 
197 	return ret;
198 }
199 
200 static int hv_vtl_wakeup_secondary_cpu(u32 apicid, unsigned long start_eip)
201 {
202 	int vp_id;
203 
204 	pr_debug("Bringing up CPU with APIC ID %d in VTL2...\n", apicid);
205 	vp_id = hv_vtl_apicid_to_vp_id(apicid);
206 
207 	if (vp_id < 0) {
208 		pr_err("Couldn't find CPU with APIC ID %d\n", apicid);
209 		return -EINVAL;
210 	}
211 	if (vp_id > ms_hyperv.max_vp_index) {
212 		pr_err("Invalid CPU id %d for APIC ID %d\n", vp_id, apicid);
213 		return -EINVAL;
214 	}
215 
216 	return hv_vtl_bringup_vcpu(vp_id, start_eip);
217 }
218 
219 int __init hv_vtl_early_init(void)
220 {
221 	/*
222 	 * `boot_cpu_has` returns the runtime feature support,
223 	 * and here is the earliest it can be used.
224 	 */
225 	if (cpu_feature_enabled(X86_FEATURE_XSAVE))
226 		panic("XSAVE has to be disabled as it is not supported by this module.\n"
227 			  "Please add 'noxsave' to the kernel command line.\n");
228 
229 	real_mode_header = &hv_vtl_real_mode_header;
230 	apic_update_callback(wakeup_secondary_cpu_64, hv_vtl_wakeup_secondary_cpu);
231 
232 	return 0;
233 }
234