xref: /linux/arch/x86/include/asm/sev.h (revision c434e25b62f8efcfbb6bf1f7ce55960206c1137e)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 /*
3  * AMD Encrypted Register State Support
4  *
5  * Author: Joerg Roedel <jroedel@suse.de>
6  */
7 
8 #ifndef __ASM_ENCRYPTED_STATE_H
9 #define __ASM_ENCRYPTED_STATE_H
10 
11 #include <linux/types.h>
12 #include <linux/sev-guest.h>
13 
14 #include <asm/insn.h>
15 #include <asm/sev-common.h>
16 #include <asm/coco.h>
17 
18 #define GHCB_PROTOCOL_MIN	1ULL
19 #define GHCB_PROTOCOL_MAX	2ULL
20 #define GHCB_DEFAULT_USAGE	0ULL
21 
22 #define	VMGEXIT()			{ asm volatile("rep; vmmcall\n\r"); }
23 
24 struct boot_params;
25 
26 enum es_result {
27 	ES_OK,			/* All good */
28 	ES_UNSUPPORTED,		/* Requested operation not supported */
29 	ES_VMM_ERROR,		/* Unexpected state from the VMM */
30 	ES_DECODE_FAILED,	/* Instruction decoding failed */
31 	ES_EXCEPTION,		/* Instruction caused exception */
32 	ES_RETRY,		/* Retry instruction emulation */
33 };
34 
35 struct es_fault_info {
36 	unsigned long vector;
37 	unsigned long error_code;
38 	unsigned long cr2;
39 };
40 
41 struct pt_regs;
42 
43 /* ES instruction emulation context */
44 struct es_em_ctxt {
45 	struct pt_regs *regs;
46 	struct insn insn;
47 	struct es_fault_info fi;
48 };
49 
50 /*
51  * AMD SEV Confidential computing blob structure. The structure is
52  * defined in OVMF UEFI firmware header:
53  * https://github.com/tianocore/edk2/blob/master/OvmfPkg/Include/Guid/ConfidentialComputingSevSnpBlob.h
54  */
55 #define CC_BLOB_SEV_HDR_MAGIC	0x45444d41
56 struct cc_blob_sev_info {
57 	u32 magic;
58 	u16 version;
59 	u16 reserved;
60 	u64 secrets_phys;
61 	u32 secrets_len;
62 	u32 rsvd1;
63 	u64 cpuid_phys;
64 	u32 cpuid_len;
65 	u32 rsvd2;
66 } __packed;
67 
68 void do_vc_no_ghcb(struct pt_regs *regs, unsigned long exit_code);
69 
70 static inline u64 lower_bits(u64 val, unsigned int bits)
71 {
72 	u64 mask = (1ULL << bits) - 1;
73 
74 	return (val & mask);
75 }
76 
77 struct real_mode_header;
78 enum stack_type;
79 
80 /* Early IDT entry points for #VC handler */
81 extern void vc_no_ghcb(void);
82 extern void vc_boot_ghcb(void);
83 extern bool handle_vc_boot_ghcb(struct pt_regs *regs);
84 
85 /* PVALIDATE return codes */
86 #define PVALIDATE_FAIL_SIZEMISMATCH	6
87 
88 /* Software defined (when rFlags.CF = 1) */
89 #define PVALIDATE_FAIL_NOUPDATE		255
90 
91 /* RMUPDATE detected 4K page and 2MB page overlap. */
92 #define RMPUPDATE_FAIL_OVERLAP		4
93 
94 /* RMP page size */
95 #define RMP_PG_SIZE_4K			0
96 #define RMP_PG_SIZE_2M			1
97 #define RMP_TO_PG_LEVEL(level)		(((level) == RMP_PG_SIZE_4K) ? PG_LEVEL_4K : PG_LEVEL_2M)
98 #define PG_LEVEL_TO_RMP(level)		(((level) == PG_LEVEL_4K) ? RMP_PG_SIZE_4K : RMP_PG_SIZE_2M)
99 
100 struct rmp_state {
101 	u64 gpa;
102 	u8 assigned;
103 	u8 pagesize;
104 	u8 immutable;
105 	u8 rsvd;
106 	u32 asid;
107 } __packed;
108 
109 #define RMPADJUST_VMSA_PAGE_BIT		BIT(16)
110 
111 /* SNP Guest message request */
112 struct snp_req_data {
113 	unsigned long req_gpa;
114 	unsigned long resp_gpa;
115 	unsigned long data_gpa;
116 	unsigned int data_npages;
117 };
118 
119 struct sev_guest_platform_data {
120 	u64 secrets_gpa;
121 };
122 
123 /*
124  * The secrets page contains 96-bytes of reserved field that can be used by
125  * the guest OS. The guest OS uses the area to save the message sequence
126  * number for each VMPCK.
127  *
128  * See the GHCB spec section Secret page layout for the format for this area.
129  */
130 struct secrets_os_area {
131 	u32 msg_seqno_0;
132 	u32 msg_seqno_1;
133 	u32 msg_seqno_2;
134 	u32 msg_seqno_3;
135 	u64 ap_jump_table_pa;
136 	u8 rsvd[40];
137 	u8 guest_usage[32];
138 } __packed;
139 
140 #define VMPCK_KEY_LEN		32
141 
142 /* See the SNP spec version 0.9 for secrets page format */
143 struct snp_secrets_page {
144 	u32 version;
145 	u32 imien	: 1,
146 	    rsvd1	: 31;
147 	u32 fms;
148 	u32 rsvd2;
149 	u8 gosvw[16];
150 	u8 vmpck0[VMPCK_KEY_LEN];
151 	u8 vmpck1[VMPCK_KEY_LEN];
152 	u8 vmpck2[VMPCK_KEY_LEN];
153 	u8 vmpck3[VMPCK_KEY_LEN];
154 	struct secrets_os_area os_area;
155 
156 	u8 vmsa_tweak_bitmap[64];
157 
158 	/* SVSM fields */
159 	u64 svsm_base;
160 	u64 svsm_size;
161 	u64 svsm_caa;
162 	u32 svsm_max_version;
163 	u8 svsm_guest_vmpl;
164 	u8 rsvd3[3];
165 
166 	/* Remainder of page */
167 	u8 rsvd4[3744];
168 } __packed;
169 
170 /*
171  * The SVSM Calling Area (CA) related structures.
172  */
173 struct svsm_ca {
174 	u8 call_pending;
175 	u8 mem_available;
176 	u8 rsvd1[6];
177 
178 	u8 svsm_buffer[PAGE_SIZE - 8];
179 };
180 
181 #define SVSM_SUCCESS				0
182 #define SVSM_ERR_INCOMPLETE			0x80000000
183 #define SVSM_ERR_UNSUPPORTED_PROTOCOL		0x80000001
184 #define SVSM_ERR_UNSUPPORTED_CALL		0x80000002
185 #define SVSM_ERR_INVALID_ADDRESS		0x80000003
186 #define SVSM_ERR_INVALID_FORMAT			0x80000004
187 #define SVSM_ERR_INVALID_PARAMETER		0x80000005
188 #define SVSM_ERR_INVALID_REQUEST		0x80000006
189 #define SVSM_ERR_BUSY				0x80000007
190 #define SVSM_PVALIDATE_FAIL_SIZEMISMATCH	0x80001006
191 
192 /*
193  * The SVSM PVALIDATE related structures
194  */
195 struct svsm_pvalidate_entry {
196 	u64 page_size		: 2,
197 	    action		: 1,
198 	    ignore_cf		: 1,
199 	    rsvd		: 8,
200 	    pfn			: 52;
201 };
202 
203 struct svsm_pvalidate_call {
204 	u16 num_entries;
205 	u16 cur_index;
206 
207 	u8 rsvd1[4];
208 
209 	struct svsm_pvalidate_entry entry[];
210 };
211 
212 #define SVSM_PVALIDATE_MAX_COUNT	((sizeof_field(struct svsm_ca, svsm_buffer) -		\
213 					  offsetof(struct svsm_pvalidate_call, entry)) /	\
214 					 sizeof(struct svsm_pvalidate_entry))
215 
216 /*
217  * The SVSM Attestation related structures
218  */
219 struct svsm_loc_entry {
220 	u64 pa;
221 	u32 len;
222 	u8 rsvd[4];
223 };
224 
225 struct svsm_attest_call {
226 	struct svsm_loc_entry report_buf;
227 	struct svsm_loc_entry nonce;
228 	struct svsm_loc_entry manifest_buf;
229 	struct svsm_loc_entry certificates_buf;
230 
231 	/* For attesting a single service */
232 	u8 service_guid[16];
233 	u32 service_manifest_ver;
234 	u8 rsvd[4];
235 };
236 
237 /*
238  * SVSM protocol structure
239  */
240 struct svsm_call {
241 	struct svsm_ca *caa;
242 	u64 rax;
243 	u64 rcx;
244 	u64 rdx;
245 	u64 r8;
246 	u64 r9;
247 	u64 rax_out;
248 	u64 rcx_out;
249 	u64 rdx_out;
250 	u64 r8_out;
251 	u64 r9_out;
252 };
253 
254 #define SVSM_CORE_CALL(x)		((0ULL << 32) | (x))
255 #define SVSM_CORE_REMAP_CA		0
256 #define SVSM_CORE_PVALIDATE		1
257 #define SVSM_CORE_CREATE_VCPU		2
258 #define SVSM_CORE_DELETE_VCPU		3
259 
260 #define SVSM_ATTEST_CALL(x)		((1ULL << 32) | (x))
261 #define SVSM_ATTEST_SERVICES		0
262 #define SVSM_ATTEST_SINGLE_SERVICE	1
263 
264 #ifdef CONFIG_AMD_MEM_ENCRYPT
265 
266 extern u8 snp_vmpl;
267 
268 extern void __sev_es_ist_enter(struct pt_regs *regs);
269 extern void __sev_es_ist_exit(void);
270 static __always_inline void sev_es_ist_enter(struct pt_regs *regs)
271 {
272 	if (cc_vendor == CC_VENDOR_AMD &&
273 	    cc_platform_has(CC_ATTR_GUEST_STATE_ENCRYPT))
274 		__sev_es_ist_enter(regs);
275 }
276 static __always_inline void sev_es_ist_exit(void)
277 {
278 	if (cc_vendor == CC_VENDOR_AMD &&
279 	    cc_platform_has(CC_ATTR_GUEST_STATE_ENCRYPT))
280 		__sev_es_ist_exit();
281 }
282 extern int sev_es_setup_ap_jump_table(struct real_mode_header *rmh);
283 extern void __sev_es_nmi_complete(void);
284 static __always_inline void sev_es_nmi_complete(void)
285 {
286 	if (cc_vendor == CC_VENDOR_AMD &&
287 	    cc_platform_has(CC_ATTR_GUEST_STATE_ENCRYPT))
288 		__sev_es_nmi_complete();
289 }
290 extern int __init sev_es_efi_map_ghcbs(pgd_t *pgd);
291 extern void sev_enable(struct boot_params *bp);
292 
293 /*
294  * RMPADJUST modifies the RMP permissions of a page of a lesser-
295  * privileged (numerically higher) VMPL.
296  *
297  * If the guest is running at a higher-privilege than the privilege
298  * level the instruction is targeting, the instruction will succeed,
299  * otherwise, it will fail.
300  */
301 static inline int rmpadjust(unsigned long vaddr, bool rmp_psize, unsigned long attrs)
302 {
303 	int rc;
304 
305 	/* "rmpadjust" mnemonic support in binutils 2.36 and newer */
306 	asm volatile(".byte 0xF3,0x0F,0x01,0xFE\n\t"
307 		     : "=a"(rc)
308 		     : "a"(vaddr), "c"(rmp_psize), "d"(attrs)
309 		     : "memory", "cc");
310 
311 	return rc;
312 }
313 static inline int pvalidate(unsigned long vaddr, bool rmp_psize, bool validate)
314 {
315 	bool no_rmpupdate;
316 	int rc;
317 
318 	/* "pvalidate" mnemonic support in binutils 2.36 and newer */
319 	asm volatile(".byte 0xF2, 0x0F, 0x01, 0xFF\n\t"
320 		     CC_SET(c)
321 		     : CC_OUT(c) (no_rmpupdate), "=a"(rc)
322 		     : "a"(vaddr), "c"(rmp_psize), "d"(validate)
323 		     : "memory", "cc");
324 
325 	if (no_rmpupdate)
326 		return PVALIDATE_FAIL_NOUPDATE;
327 
328 	return rc;
329 }
330 
331 struct snp_guest_request_ioctl;
332 
333 void setup_ghcb(void);
334 void early_snp_set_memory_private(unsigned long vaddr, unsigned long paddr,
335 				  unsigned long npages);
336 void early_snp_set_memory_shared(unsigned long vaddr, unsigned long paddr,
337 				 unsigned long npages);
338 void snp_set_memory_shared(unsigned long vaddr, unsigned long npages);
339 void snp_set_memory_private(unsigned long vaddr, unsigned long npages);
340 void snp_set_wakeup_secondary_cpu(void);
341 bool snp_init(struct boot_params *bp);
342 void __noreturn snp_abort(void);
343 void snp_dmi_setup(void);
344 int snp_issue_guest_request(u64 exit_code, struct snp_req_data *input, struct snp_guest_request_ioctl *rio);
345 int snp_issue_svsm_attest_req(u64 call_id, struct svsm_call *call, struct svsm_attest_call *input);
346 void snp_accept_memory(phys_addr_t start, phys_addr_t end);
347 u64 snp_get_unsupported_features(u64 status);
348 u64 sev_get_status(void);
349 void sev_show_status(void);
350 void snp_update_svsm_ca(void);
351 
352 #else	/* !CONFIG_AMD_MEM_ENCRYPT */
353 
354 #define snp_vmpl 0
355 static inline void sev_es_ist_enter(struct pt_regs *regs) { }
356 static inline void sev_es_ist_exit(void) { }
357 static inline int sev_es_setup_ap_jump_table(struct real_mode_header *rmh) { return 0; }
358 static inline void sev_es_nmi_complete(void) { }
359 static inline int sev_es_efi_map_ghcbs(pgd_t *pgd) { return 0; }
360 static inline void sev_enable(struct boot_params *bp) { }
361 static inline int pvalidate(unsigned long vaddr, bool rmp_psize, bool validate) { return 0; }
362 static inline int rmpadjust(unsigned long vaddr, bool rmp_psize, unsigned long attrs) { return 0; }
363 static inline void setup_ghcb(void) { }
364 static inline void __init
365 early_snp_set_memory_private(unsigned long vaddr, unsigned long paddr, unsigned long npages) { }
366 static inline void __init
367 early_snp_set_memory_shared(unsigned long vaddr, unsigned long paddr, unsigned long npages) { }
368 static inline void snp_set_memory_shared(unsigned long vaddr, unsigned long npages) { }
369 static inline void snp_set_memory_private(unsigned long vaddr, unsigned long npages) { }
370 static inline void snp_set_wakeup_secondary_cpu(void) { }
371 static inline bool snp_init(struct boot_params *bp) { return false; }
372 static inline void snp_abort(void) { }
373 static inline void snp_dmi_setup(void) { }
374 static inline int snp_issue_guest_request(u64 exit_code, struct snp_req_data *input, struct snp_guest_request_ioctl *rio)
375 {
376 	return -ENOTTY;
377 }
378 static inline int snp_issue_svsm_attest_req(u64 call_id, struct svsm_call *call, struct svsm_attest_call *input)
379 {
380 	return -ENOTTY;
381 }
382 static inline void snp_accept_memory(phys_addr_t start, phys_addr_t end) { }
383 static inline u64 snp_get_unsupported_features(u64 status) { return 0; }
384 static inline u64 sev_get_status(void) { return 0; }
385 static inline void sev_show_status(void) { }
386 static inline void snp_update_svsm_ca(void) { }
387 
388 #endif	/* CONFIG_AMD_MEM_ENCRYPT */
389 
390 #ifdef CONFIG_KVM_AMD_SEV
391 bool snp_probe_rmptable_info(void);
392 int snp_lookup_rmpentry(u64 pfn, bool *assigned, int *level);
393 void snp_dump_hva_rmpentry(unsigned long address);
394 int psmash(u64 pfn);
395 int rmp_make_private(u64 pfn, u64 gpa, enum pg_level level, u32 asid, bool immutable);
396 int rmp_make_shared(u64 pfn, enum pg_level level);
397 void snp_leak_pages(u64 pfn, unsigned int npages);
398 void kdump_sev_callback(void);
399 void snp_fixup_e820_tables(void);
400 #else
401 static inline bool snp_probe_rmptable_info(void) { return false; }
402 static inline int snp_lookup_rmpentry(u64 pfn, bool *assigned, int *level) { return -ENODEV; }
403 static inline void snp_dump_hva_rmpentry(unsigned long address) {}
404 static inline int psmash(u64 pfn) { return -ENODEV; }
405 static inline int rmp_make_private(u64 pfn, u64 gpa, enum pg_level level, u32 asid,
406 				   bool immutable)
407 {
408 	return -ENODEV;
409 }
410 static inline int rmp_make_shared(u64 pfn, enum pg_level level) { return -ENODEV; }
411 static inline void snp_leak_pages(u64 pfn, unsigned int npages) {}
412 static inline void kdump_sev_callback(void) { }
413 static inline void snp_fixup_e820_tables(void) {}
414 #endif
415 
416 #endif
417