1 /* SPDX-License-Identifier: GPL-2.0 */
2 /*
3 * AMD Encrypted Register State Support
4 *
5 * Author: Joerg Roedel <jroedel@suse.de>
6 */
7
8 #ifndef __ASM_ENCRYPTED_STATE_H
9 #define __ASM_ENCRYPTED_STATE_H
10
11 #include <linux/types.h>
12 #include <linux/sev-guest.h>
13
14 #include <asm/insn.h>
15 #include <asm/sev-common.h>
16 #include <asm/coco.h>
17 #include <asm/set_memory.h>
18 #include <asm/svm.h>
19
20 #define GHCB_PROTOCOL_MIN 1ULL
21 #define GHCB_PROTOCOL_MAX 2ULL
22 #define GHCB_DEFAULT_USAGE 0ULL
23
24 #define VMGEXIT() { asm volatile("rep; vmmcall\n\r"); }
25
26 struct boot_params;
27
28 enum es_result {
29 ES_OK, /* All good */
30 ES_UNSUPPORTED, /* Requested operation not supported */
31 ES_VMM_ERROR, /* Unexpected state from the VMM */
32 ES_DECODE_FAILED, /* Instruction decoding failed */
33 ES_EXCEPTION, /* Instruction caused exception */
34 ES_RETRY, /* Retry instruction emulation */
35 };
36
37 struct es_fault_info {
38 unsigned long vector;
39 unsigned long error_code;
40 unsigned long cr2;
41 };
42
43 struct pt_regs;
44
45 /* ES instruction emulation context */
46 struct es_em_ctxt {
47 struct pt_regs *regs;
48 struct insn insn;
49 struct es_fault_info fi;
50 };
51
52 /*
53 * AMD SEV Confidential computing blob structure. The structure is
54 * defined in OVMF UEFI firmware header:
55 * https://github.com/tianocore/edk2/blob/master/OvmfPkg/Include/Guid/ConfidentialComputingSevSnpBlob.h
56 */
57 #define CC_BLOB_SEV_HDR_MAGIC 0x45444d41
58 struct cc_blob_sev_info {
59 u32 magic;
60 u16 version;
61 u16 reserved;
62 u64 secrets_phys;
63 u32 secrets_len;
64 u32 rsvd1;
65 u64 cpuid_phys;
66 u32 cpuid_len;
67 u32 rsvd2;
68 } __packed;
69
70 void do_vc_no_ghcb(struct pt_regs *regs, unsigned long exit_code);
71
lower_bits(u64 val,unsigned int bits)72 static inline u64 lower_bits(u64 val, unsigned int bits)
73 {
74 u64 mask = (1ULL << bits) - 1;
75
76 return (val & mask);
77 }
78
79 struct real_mode_header;
80 enum stack_type;
81
82 /* Early IDT entry points for #VC handler */
83 extern void vc_no_ghcb(void);
84 extern void vc_boot_ghcb(void);
85 extern bool handle_vc_boot_ghcb(struct pt_regs *regs);
86
87 /*
88 * Individual entries of the SNP CPUID table, as defined by the SNP
89 * Firmware ABI, Revision 0.9, Section 7.1, Table 14.
90 */
91 struct snp_cpuid_fn {
92 u32 eax_in;
93 u32 ecx_in;
94 u64 xcr0_in;
95 u64 xss_in;
96 u32 eax;
97 u32 ebx;
98 u32 ecx;
99 u32 edx;
100 u64 __reserved;
101 } __packed;
102
103 /*
104 * SNP CPUID table, as defined by the SNP Firmware ABI, Revision 0.9,
105 * Section 8.14.2.6. Also noted there is the SNP firmware-enforced limit
106 * of 64 entries per CPUID table.
107 */
108 #define SNP_CPUID_COUNT_MAX 64
109
110 struct snp_cpuid_table {
111 u32 count;
112 u32 __reserved1;
113 u64 __reserved2;
114 struct snp_cpuid_fn fn[SNP_CPUID_COUNT_MAX];
115 } __packed;
116
117 /* PVALIDATE return codes */
118 #define PVALIDATE_FAIL_SIZEMISMATCH 6
119
120 /* Software defined (when rFlags.CF = 1) */
121 #define PVALIDATE_FAIL_NOUPDATE 255
122
123 /* RMUPDATE detected 4K page and 2MB page overlap. */
124 #define RMPUPDATE_FAIL_OVERLAP 4
125
126 /* PSMASH failed due to concurrent access by another CPU */
127 #define PSMASH_FAIL_INUSE 3
128
129 /* RMP page size */
130 #define RMP_PG_SIZE_4K 0
131 #define RMP_PG_SIZE_2M 1
132 #define RMP_TO_PG_LEVEL(level) (((level) == RMP_PG_SIZE_4K) ? PG_LEVEL_4K : PG_LEVEL_2M)
133 #define PG_LEVEL_TO_RMP(level) (((level) == PG_LEVEL_4K) ? RMP_PG_SIZE_4K : RMP_PG_SIZE_2M)
134
135 struct rmp_state {
136 u64 gpa;
137 u8 assigned;
138 u8 pagesize;
139 u8 immutable;
140 u8 rsvd;
141 u32 asid;
142 } __packed;
143
144 #define RMPADJUST_VMSA_PAGE_BIT BIT(16)
145
146 /* SNP Guest message request */
147 struct snp_req_data {
148 unsigned long req_gpa;
149 unsigned long resp_gpa;
150 unsigned long data_gpa;
151 unsigned int data_npages;
152 };
153
154 #define MAX_AUTHTAG_LEN 32
155 #define AUTHTAG_LEN 16
156 #define AAD_LEN 48
157 #define MSG_HDR_VER 1
158
159 #define SNP_REQ_MAX_RETRY_DURATION (60*HZ)
160 #define SNP_REQ_RETRY_DELAY (2*HZ)
161
162 /* See SNP spec SNP_GUEST_REQUEST section for the structure */
163 enum msg_type {
164 SNP_MSG_TYPE_INVALID = 0,
165 SNP_MSG_CPUID_REQ,
166 SNP_MSG_CPUID_RSP,
167 SNP_MSG_KEY_REQ,
168 SNP_MSG_KEY_RSP,
169 SNP_MSG_REPORT_REQ,
170 SNP_MSG_REPORT_RSP,
171 SNP_MSG_EXPORT_REQ,
172 SNP_MSG_EXPORT_RSP,
173 SNP_MSG_IMPORT_REQ,
174 SNP_MSG_IMPORT_RSP,
175 SNP_MSG_ABSORB_REQ,
176 SNP_MSG_ABSORB_RSP,
177 SNP_MSG_VMRK_REQ,
178 SNP_MSG_VMRK_RSP,
179
180 SNP_MSG_TSC_INFO_REQ = 17,
181 SNP_MSG_TSC_INFO_RSP,
182
183 SNP_MSG_TYPE_MAX
184 };
185
186 enum aead_algo {
187 SNP_AEAD_INVALID,
188 SNP_AEAD_AES_256_GCM,
189 };
190
191 struct snp_guest_msg_hdr {
192 u8 authtag[MAX_AUTHTAG_LEN];
193 u64 msg_seqno;
194 u8 rsvd1[8];
195 u8 algo;
196 u8 hdr_version;
197 u16 hdr_sz;
198 u8 msg_type;
199 u8 msg_version;
200 u16 msg_sz;
201 u32 rsvd2;
202 u8 msg_vmpck;
203 u8 rsvd3[35];
204 } __packed;
205
206 struct snp_guest_msg {
207 struct snp_guest_msg_hdr hdr;
208 u8 payload[PAGE_SIZE - sizeof(struct snp_guest_msg_hdr)];
209 } __packed;
210
211 #define SNP_TSC_INFO_REQ_SZ 128
212
213 struct snp_tsc_info_req {
214 u8 rsvd[SNP_TSC_INFO_REQ_SZ];
215 } __packed;
216
217 struct snp_tsc_info_resp {
218 u32 status;
219 u32 rsvd1;
220 u64 tsc_scale;
221 u64 tsc_offset;
222 u32 tsc_factor;
223 u8 rsvd2[100];
224 } __packed;
225
226 struct snp_guest_req {
227 void *req_buf;
228 size_t req_sz;
229
230 void *resp_buf;
231 size_t resp_sz;
232
233 u64 exit_code;
234 unsigned int vmpck_id;
235 u8 msg_version;
236 u8 msg_type;
237
238 struct snp_req_data input;
239 void *certs_data;
240 };
241
242 /*
243 * The secrets page contains 96-bytes of reserved field that can be used by
244 * the guest OS. The guest OS uses the area to save the message sequence
245 * number for each VMPCK.
246 *
247 * See the GHCB spec section Secret page layout for the format for this area.
248 */
249 struct secrets_os_area {
250 u32 msg_seqno_0;
251 u32 msg_seqno_1;
252 u32 msg_seqno_2;
253 u32 msg_seqno_3;
254 u64 ap_jump_table_pa;
255 u8 rsvd[40];
256 u8 guest_usage[32];
257 } __packed;
258
259 #define VMPCK_KEY_LEN 32
260
261 /* See the SNP spec version 0.9 for secrets page format */
262 struct snp_secrets_page {
263 u32 version;
264 u32 imien : 1,
265 rsvd1 : 31;
266 u32 fms;
267 u32 rsvd2;
268 u8 gosvw[16];
269 u8 vmpck0[VMPCK_KEY_LEN];
270 u8 vmpck1[VMPCK_KEY_LEN];
271 u8 vmpck2[VMPCK_KEY_LEN];
272 u8 vmpck3[VMPCK_KEY_LEN];
273 struct secrets_os_area os_area;
274
275 u8 vmsa_tweak_bitmap[64];
276
277 /* SVSM fields */
278 u64 svsm_base;
279 u64 svsm_size;
280 u64 svsm_caa;
281 u32 svsm_max_version;
282 u8 svsm_guest_vmpl;
283 u8 rsvd3[3];
284
285 /* Remainder of page */
286 u8 rsvd4[3744];
287 } __packed;
288
289 struct snp_msg_desc {
290 /* request and response are in unencrypted memory */
291 struct snp_guest_msg *request, *response;
292
293 /*
294 * Avoid information leakage by double-buffering shared messages
295 * in fields that are in regular encrypted memory.
296 */
297 struct snp_guest_msg secret_request, secret_response;
298
299 struct snp_secrets_page *secrets;
300
301 struct aesgcm_ctx *ctx;
302
303 u32 *os_area_msg_seqno;
304 u8 *vmpck;
305 int vmpck_id;
306 };
307
308 /*
309 * The SVSM Calling Area (CA) related structures.
310 */
311 struct svsm_ca {
312 u8 call_pending;
313 u8 mem_available;
314 u8 rsvd1[6];
315
316 u8 svsm_buffer[PAGE_SIZE - 8];
317 };
318
319 #define SVSM_SUCCESS 0
320 #define SVSM_ERR_INCOMPLETE 0x80000000
321 #define SVSM_ERR_UNSUPPORTED_PROTOCOL 0x80000001
322 #define SVSM_ERR_UNSUPPORTED_CALL 0x80000002
323 #define SVSM_ERR_INVALID_ADDRESS 0x80000003
324 #define SVSM_ERR_INVALID_FORMAT 0x80000004
325 #define SVSM_ERR_INVALID_PARAMETER 0x80000005
326 #define SVSM_ERR_INVALID_REQUEST 0x80000006
327 #define SVSM_ERR_BUSY 0x80000007
328 #define SVSM_PVALIDATE_FAIL_SIZEMISMATCH 0x80001006
329
330 /*
331 * The SVSM PVALIDATE related structures
332 */
333 struct svsm_pvalidate_entry {
334 u64 page_size : 2,
335 action : 1,
336 ignore_cf : 1,
337 rsvd : 8,
338 pfn : 52;
339 };
340
341 struct svsm_pvalidate_call {
342 u16 num_entries;
343 u16 cur_index;
344
345 u8 rsvd1[4];
346
347 struct svsm_pvalidate_entry entry[];
348 };
349
350 #define SVSM_PVALIDATE_MAX_COUNT ((sizeof_field(struct svsm_ca, svsm_buffer) - \
351 offsetof(struct svsm_pvalidate_call, entry)) / \
352 sizeof(struct svsm_pvalidate_entry))
353
354 /*
355 * The SVSM Attestation related structures
356 */
357 struct svsm_loc_entry {
358 u64 pa;
359 u32 len;
360 u8 rsvd[4];
361 };
362
363 struct svsm_attest_call {
364 struct svsm_loc_entry report_buf;
365 struct svsm_loc_entry nonce;
366 struct svsm_loc_entry manifest_buf;
367 struct svsm_loc_entry certificates_buf;
368
369 /* For attesting a single service */
370 u8 service_guid[16];
371 u32 service_manifest_ver;
372 u8 rsvd[4];
373 };
374
375 /* PTE descriptor used for the prepare_pte_enc() operations. */
376 struct pte_enc_desc {
377 pte_t *kpte;
378 int pte_level;
379 bool encrypt;
380 /* pfn of the kpte above */
381 unsigned long pfn;
382 /* physical address of @pfn */
383 unsigned long pa;
384 /* virtual address of @pfn */
385 void *va;
386 /* memory covered by the pte */
387 unsigned long size;
388 pgprot_t new_pgprot;
389 };
390
391 /*
392 * SVSM protocol structure
393 */
394 struct svsm_call {
395 struct svsm_ca *caa;
396 u64 rax;
397 u64 rcx;
398 u64 rdx;
399 u64 r8;
400 u64 r9;
401 u64 rax_out;
402 u64 rcx_out;
403 u64 rdx_out;
404 u64 r8_out;
405 u64 r9_out;
406 };
407
408 #define SVSM_CORE_CALL(x) ((0ULL << 32) | (x))
409 #define SVSM_CORE_REMAP_CA 0
410 #define SVSM_CORE_PVALIDATE 1
411 #define SVSM_CORE_CREATE_VCPU 2
412 #define SVSM_CORE_DELETE_VCPU 3
413
414 #define SVSM_ATTEST_CALL(x) ((1ULL << 32) | (x))
415 #define SVSM_ATTEST_SERVICES 0
416 #define SVSM_ATTEST_SINGLE_SERVICE 1
417
418 #define SVSM_VTPM_CALL(x) ((2ULL << 32) | (x))
419 #define SVSM_VTPM_QUERY 0
420 #define SVSM_VTPM_CMD 1
421
422 #ifdef CONFIG_AMD_MEM_ENCRYPT
423
424 extern u8 snp_vmpl;
425
426 extern void __sev_es_ist_enter(struct pt_regs *regs);
427 extern void __sev_es_ist_exit(void);
sev_es_ist_enter(struct pt_regs * regs)428 static __always_inline void sev_es_ist_enter(struct pt_regs *regs)
429 {
430 if (cc_vendor == CC_VENDOR_AMD &&
431 cc_platform_has(CC_ATTR_GUEST_STATE_ENCRYPT))
432 __sev_es_ist_enter(regs);
433 }
sev_es_ist_exit(void)434 static __always_inline void sev_es_ist_exit(void)
435 {
436 if (cc_vendor == CC_VENDOR_AMD &&
437 cc_platform_has(CC_ATTR_GUEST_STATE_ENCRYPT))
438 __sev_es_ist_exit();
439 }
440 extern int sev_es_setup_ap_jump_table(struct real_mode_header *rmh);
441 extern void __sev_es_nmi_complete(void);
sev_es_nmi_complete(void)442 static __always_inline void sev_es_nmi_complete(void)
443 {
444 if (cc_vendor == CC_VENDOR_AMD &&
445 cc_platform_has(CC_ATTR_GUEST_STATE_ENCRYPT))
446 __sev_es_nmi_complete();
447 }
448 extern int __init sev_es_efi_map_ghcbs(pgd_t *pgd);
449 extern void sev_enable(struct boot_params *bp);
450
451 /*
452 * RMPADJUST modifies the RMP permissions of a page of a lesser-
453 * privileged (numerically higher) VMPL.
454 *
455 * If the guest is running at a higher-privilege than the privilege
456 * level the instruction is targeting, the instruction will succeed,
457 * otherwise, it will fail.
458 */
rmpadjust(unsigned long vaddr,bool rmp_psize,unsigned long attrs)459 static inline int rmpadjust(unsigned long vaddr, bool rmp_psize, unsigned long attrs)
460 {
461 int rc;
462
463 /* "rmpadjust" mnemonic support in binutils 2.36 and newer */
464 asm volatile(".byte 0xF3,0x0F,0x01,0xFE\n\t"
465 : "=a"(rc)
466 : "a"(vaddr), "c"(rmp_psize), "d"(attrs)
467 : "memory", "cc");
468
469 return rc;
470 }
pvalidate(unsigned long vaddr,bool rmp_psize,bool validate)471 static inline int pvalidate(unsigned long vaddr, bool rmp_psize, bool validate)
472 {
473 bool no_rmpupdate;
474 int rc;
475
476 /* "pvalidate" mnemonic support in binutils 2.36 and newer */
477 asm volatile(".byte 0xF2, 0x0F, 0x01, 0xFF\n\t"
478 CC_SET(c)
479 : CC_OUT(c) (no_rmpupdate), "=a"(rc)
480 : "a"(vaddr), "c"(rmp_psize), "d"(validate)
481 : "memory", "cc");
482
483 if (no_rmpupdate)
484 return PVALIDATE_FAIL_NOUPDATE;
485
486 return rc;
487 }
488
489 struct snp_guest_request_ioctl;
490
491 void setup_ghcb(void);
492 void early_snp_set_memory_private(unsigned long vaddr, unsigned long paddr,
493 unsigned long npages);
494 void early_snp_set_memory_shared(unsigned long vaddr, unsigned long paddr,
495 unsigned long npages);
496 void snp_set_memory_shared(unsigned long vaddr, unsigned long npages);
497 void snp_set_memory_private(unsigned long vaddr, unsigned long npages);
498 void snp_set_wakeup_secondary_cpu(void);
499 bool snp_init(struct boot_params *bp);
500 void __noreturn snp_abort(void);
501 void snp_dmi_setup(void);
502 int snp_issue_svsm_attest_req(u64 call_id, struct svsm_call *call, struct svsm_attest_call *input);
503 void snp_accept_memory(phys_addr_t start, phys_addr_t end);
504 u64 snp_get_unsupported_features(u64 status);
505 u64 sev_get_status(void);
506 void sev_show_status(void);
507 void snp_update_svsm_ca(void);
508 int prepare_pte_enc(struct pte_enc_desc *d);
509 void set_pte_enc_mask(pte_t *kpte, unsigned long pfn, pgprot_t new_prot);
510 void snp_kexec_finish(void);
511 void snp_kexec_begin(void);
512
513 int snp_msg_init(struct snp_msg_desc *mdesc, int vmpck_id);
514 struct snp_msg_desc *snp_msg_alloc(void);
515 void snp_msg_free(struct snp_msg_desc *mdesc);
516 int snp_send_guest_request(struct snp_msg_desc *mdesc, struct snp_guest_req *req,
517 struct snp_guest_request_ioctl *rio);
518
519 int snp_svsm_vtpm_send_command(u8 *buffer);
520
521 void __init snp_secure_tsc_prepare(void);
522 void __init snp_secure_tsc_init(void);
523
vc_ghcb_invalidate(struct ghcb * ghcb)524 static __always_inline void vc_ghcb_invalidate(struct ghcb *ghcb)
525 {
526 ghcb->save.sw_exit_code = 0;
527 __builtin_memset(ghcb->save.valid_bitmap, 0, sizeof(ghcb->save.valid_bitmap));
528 }
529
530 void vc_forward_exception(struct es_em_ctxt *ctxt);
531
532 /* I/O parameters for CPUID-related helpers */
533 struct cpuid_leaf {
534 u32 fn;
535 u32 subfn;
536 u32 eax;
537 u32 ebx;
538 u32 ecx;
539 u32 edx;
540 };
541
542 int snp_cpuid(struct ghcb *ghcb, struct es_em_ctxt *ctxt, struct cpuid_leaf *leaf);
543
544 void __noreturn sev_es_terminate(unsigned int set, unsigned int reason);
545 enum es_result sev_es_ghcb_hv_call(struct ghcb *ghcb,
546 struct es_em_ctxt *ctxt,
547 u64 exit_code, u64 exit_info_1,
548 u64 exit_info_2);
549
550 extern struct ghcb *boot_ghcb;
551
552 #else /* !CONFIG_AMD_MEM_ENCRYPT */
553
554 #define snp_vmpl 0
sev_es_ist_enter(struct pt_regs * regs)555 static inline void sev_es_ist_enter(struct pt_regs *regs) { }
sev_es_ist_exit(void)556 static inline void sev_es_ist_exit(void) { }
sev_es_setup_ap_jump_table(struct real_mode_header * rmh)557 static inline int sev_es_setup_ap_jump_table(struct real_mode_header *rmh) { return 0; }
sev_es_nmi_complete(void)558 static inline void sev_es_nmi_complete(void) { }
sev_es_efi_map_ghcbs(pgd_t * pgd)559 static inline int sev_es_efi_map_ghcbs(pgd_t *pgd) { return 0; }
sev_enable(struct boot_params * bp)560 static inline void sev_enable(struct boot_params *bp) { }
pvalidate(unsigned long vaddr,bool rmp_psize,bool validate)561 static inline int pvalidate(unsigned long vaddr, bool rmp_psize, bool validate) { return 0; }
rmpadjust(unsigned long vaddr,bool rmp_psize,unsigned long attrs)562 static inline int rmpadjust(unsigned long vaddr, bool rmp_psize, unsigned long attrs) { return 0; }
setup_ghcb(void)563 static inline void setup_ghcb(void) { }
564 static inline void __init
early_snp_set_memory_private(unsigned long vaddr,unsigned long paddr,unsigned long npages)565 early_snp_set_memory_private(unsigned long vaddr, unsigned long paddr, unsigned long npages) { }
566 static inline void __init
early_snp_set_memory_shared(unsigned long vaddr,unsigned long paddr,unsigned long npages)567 early_snp_set_memory_shared(unsigned long vaddr, unsigned long paddr, unsigned long npages) { }
snp_set_memory_shared(unsigned long vaddr,unsigned long npages)568 static inline void snp_set_memory_shared(unsigned long vaddr, unsigned long npages) { }
snp_set_memory_private(unsigned long vaddr,unsigned long npages)569 static inline void snp_set_memory_private(unsigned long vaddr, unsigned long npages) { }
snp_set_wakeup_secondary_cpu(void)570 static inline void snp_set_wakeup_secondary_cpu(void) { }
snp_init(struct boot_params * bp)571 static inline bool snp_init(struct boot_params *bp) { return false; }
snp_abort(void)572 static inline void snp_abort(void) { }
snp_dmi_setup(void)573 static inline void snp_dmi_setup(void) { }
snp_issue_svsm_attest_req(u64 call_id,struct svsm_call * call,struct svsm_attest_call * input)574 static inline int snp_issue_svsm_attest_req(u64 call_id, struct svsm_call *call, struct svsm_attest_call *input)
575 {
576 return -ENOTTY;
577 }
snp_accept_memory(phys_addr_t start,phys_addr_t end)578 static inline void snp_accept_memory(phys_addr_t start, phys_addr_t end) { }
snp_get_unsupported_features(u64 status)579 static inline u64 snp_get_unsupported_features(u64 status) { return 0; }
sev_get_status(void)580 static inline u64 sev_get_status(void) { return 0; }
sev_show_status(void)581 static inline void sev_show_status(void) { }
snp_update_svsm_ca(void)582 static inline void snp_update_svsm_ca(void) { }
prepare_pte_enc(struct pte_enc_desc * d)583 static inline int prepare_pte_enc(struct pte_enc_desc *d) { return 0; }
set_pte_enc_mask(pte_t * kpte,unsigned long pfn,pgprot_t new_prot)584 static inline void set_pte_enc_mask(pte_t *kpte, unsigned long pfn, pgprot_t new_prot) { }
snp_kexec_finish(void)585 static inline void snp_kexec_finish(void) { }
snp_kexec_begin(void)586 static inline void snp_kexec_begin(void) { }
snp_msg_init(struct snp_msg_desc * mdesc,int vmpck_id)587 static inline int snp_msg_init(struct snp_msg_desc *mdesc, int vmpck_id) { return -1; }
snp_msg_alloc(void)588 static inline struct snp_msg_desc *snp_msg_alloc(void) { return NULL; }
snp_msg_free(struct snp_msg_desc * mdesc)589 static inline void snp_msg_free(struct snp_msg_desc *mdesc) { }
snp_send_guest_request(struct snp_msg_desc * mdesc,struct snp_guest_req * req,struct snp_guest_request_ioctl * rio)590 static inline int snp_send_guest_request(struct snp_msg_desc *mdesc, struct snp_guest_req *req,
591 struct snp_guest_request_ioctl *rio) { return -ENODEV; }
snp_svsm_vtpm_send_command(u8 * buffer)592 static inline int snp_svsm_vtpm_send_command(u8 *buffer) { return -ENODEV; }
snp_secure_tsc_prepare(void)593 static inline void __init snp_secure_tsc_prepare(void) { }
snp_secure_tsc_init(void)594 static inline void __init snp_secure_tsc_init(void) { }
595
596 #endif /* CONFIG_AMD_MEM_ENCRYPT */
597
598 #ifdef CONFIG_KVM_AMD_SEV
599 bool snp_probe_rmptable_info(void);
600 int snp_rmptable_init(void);
601 int snp_lookup_rmpentry(u64 pfn, bool *assigned, int *level);
602 void snp_dump_hva_rmpentry(unsigned long address);
603 int psmash(u64 pfn);
604 int rmp_make_private(u64 pfn, u64 gpa, enum pg_level level, u32 asid, bool immutable);
605 int rmp_make_shared(u64 pfn, enum pg_level level);
606 void snp_leak_pages(u64 pfn, unsigned int npages);
607 void kdump_sev_callback(void);
608 void snp_fixup_e820_tables(void);
609 #else
snp_probe_rmptable_info(void)610 static inline bool snp_probe_rmptable_info(void) { return false; }
snp_rmptable_init(void)611 static inline int snp_rmptable_init(void) { return -ENOSYS; }
snp_lookup_rmpentry(u64 pfn,bool * assigned,int * level)612 static inline int snp_lookup_rmpentry(u64 pfn, bool *assigned, int *level) { return -ENODEV; }
snp_dump_hva_rmpentry(unsigned long address)613 static inline void snp_dump_hva_rmpentry(unsigned long address) {}
psmash(u64 pfn)614 static inline int psmash(u64 pfn) { return -ENODEV; }
rmp_make_private(u64 pfn,u64 gpa,enum pg_level level,u32 asid,bool immutable)615 static inline int rmp_make_private(u64 pfn, u64 gpa, enum pg_level level, u32 asid,
616 bool immutable)
617 {
618 return -ENODEV;
619 }
rmp_make_shared(u64 pfn,enum pg_level level)620 static inline int rmp_make_shared(u64 pfn, enum pg_level level) { return -ENODEV; }
snp_leak_pages(u64 pfn,unsigned int npages)621 static inline void snp_leak_pages(u64 pfn, unsigned int npages) {}
kdump_sev_callback(void)622 static inline void kdump_sev_callback(void) { }
snp_fixup_e820_tables(void)623 static inline void snp_fixup_e820_tables(void) {}
624 #endif
625
626 #endif
627