1 /* SPDX-License-Identifier: GPL-2.0 */ 2 /* 3 * AMD Encrypted Register State Support 4 * 5 * Author: Joerg Roedel <jroedel@suse.de> 6 */ 7 8 #ifndef __ASM_ENCRYPTED_STATE_H 9 #define __ASM_ENCRYPTED_STATE_H 10 11 #include <linux/types.h> 12 #include <linux/sev-guest.h> 13 14 #include <asm/insn.h> 15 #include <asm/sev-common.h> 16 #include <asm/coco.h> 17 18 #define GHCB_PROTOCOL_MIN 1ULL 19 #define GHCB_PROTOCOL_MAX 2ULL 20 #define GHCB_DEFAULT_USAGE 0ULL 21 22 #define VMGEXIT() { asm volatile("rep; vmmcall\n\r"); } 23 24 struct boot_params; 25 26 enum es_result { 27 ES_OK, /* All good */ 28 ES_UNSUPPORTED, /* Requested operation not supported */ 29 ES_VMM_ERROR, /* Unexpected state from the VMM */ 30 ES_DECODE_FAILED, /* Instruction decoding failed */ 31 ES_EXCEPTION, /* Instruction caused exception */ 32 ES_RETRY, /* Retry instruction emulation */ 33 }; 34 35 struct es_fault_info { 36 unsigned long vector; 37 unsigned long error_code; 38 unsigned long cr2; 39 }; 40 41 struct pt_regs; 42 43 /* ES instruction emulation context */ 44 struct es_em_ctxt { 45 struct pt_regs *regs; 46 struct insn insn; 47 struct es_fault_info fi; 48 }; 49 50 /* 51 * AMD SEV Confidential computing blob structure. The structure is 52 * defined in OVMF UEFI firmware header: 53 * https://github.com/tianocore/edk2/blob/master/OvmfPkg/Include/Guid/ConfidentialComputingSevSnpBlob.h 54 */ 55 #define CC_BLOB_SEV_HDR_MAGIC 0x45444d41 56 struct cc_blob_sev_info { 57 u32 magic; 58 u16 version; 59 u16 reserved; 60 u64 secrets_phys; 61 u32 secrets_len; 62 u32 rsvd1; 63 u64 cpuid_phys; 64 u32 cpuid_len; 65 u32 rsvd2; 66 } __packed; 67 68 void do_vc_no_ghcb(struct pt_regs *regs, unsigned long exit_code); 69 70 static inline u64 lower_bits(u64 val, unsigned int bits) 71 { 72 u64 mask = (1ULL << bits) - 1; 73 74 return (val & mask); 75 } 76 77 struct real_mode_header; 78 enum stack_type; 79 80 /* Early IDT entry points for #VC handler */ 81 extern void vc_no_ghcb(void); 82 extern void vc_boot_ghcb(void); 83 extern bool handle_vc_boot_ghcb(struct pt_regs *regs); 84 85 /* PVALIDATE return codes */ 86 #define PVALIDATE_FAIL_SIZEMISMATCH 6 87 88 /* Software defined (when rFlags.CF = 1) */ 89 #define PVALIDATE_FAIL_NOUPDATE 255 90 91 /* RMUPDATE detected 4K page and 2MB page overlap. */ 92 #define RMPUPDATE_FAIL_OVERLAP 4 93 94 /* PSMASH failed due to concurrent access by another CPU */ 95 #define PSMASH_FAIL_INUSE 3 96 97 /* RMP page size */ 98 #define RMP_PG_SIZE_4K 0 99 #define RMP_PG_SIZE_2M 1 100 #define RMP_TO_PG_LEVEL(level) (((level) == RMP_PG_SIZE_4K) ? PG_LEVEL_4K : PG_LEVEL_2M) 101 #define PG_LEVEL_TO_RMP(level) (((level) == PG_LEVEL_4K) ? RMP_PG_SIZE_4K : RMP_PG_SIZE_2M) 102 103 struct rmp_state { 104 u64 gpa; 105 u8 assigned; 106 u8 pagesize; 107 u8 immutable; 108 u8 rsvd; 109 u32 asid; 110 } __packed; 111 112 #define RMPADJUST_VMSA_PAGE_BIT BIT(16) 113 114 /* SNP Guest message request */ 115 struct snp_req_data { 116 unsigned long req_gpa; 117 unsigned long resp_gpa; 118 unsigned long data_gpa; 119 unsigned int data_npages; 120 }; 121 122 #define MAX_AUTHTAG_LEN 32 123 #define AUTHTAG_LEN 16 124 #define AAD_LEN 48 125 #define MSG_HDR_VER 1 126 127 /* See SNP spec SNP_GUEST_REQUEST section for the structure */ 128 enum msg_type { 129 SNP_MSG_TYPE_INVALID = 0, 130 SNP_MSG_CPUID_REQ, 131 SNP_MSG_CPUID_RSP, 132 SNP_MSG_KEY_REQ, 133 SNP_MSG_KEY_RSP, 134 SNP_MSG_REPORT_REQ, 135 SNP_MSG_REPORT_RSP, 136 SNP_MSG_EXPORT_REQ, 137 SNP_MSG_EXPORT_RSP, 138 SNP_MSG_IMPORT_REQ, 139 SNP_MSG_IMPORT_RSP, 140 SNP_MSG_ABSORB_REQ, 141 SNP_MSG_ABSORB_RSP, 142 SNP_MSG_VMRK_REQ, 143 SNP_MSG_VMRK_RSP, 144 145 SNP_MSG_TYPE_MAX 146 }; 147 148 enum aead_algo { 149 SNP_AEAD_INVALID, 150 SNP_AEAD_AES_256_GCM, 151 }; 152 153 struct snp_guest_msg_hdr { 154 u8 authtag[MAX_AUTHTAG_LEN]; 155 u64 msg_seqno; 156 u8 rsvd1[8]; 157 u8 algo; 158 u8 hdr_version; 159 u16 hdr_sz; 160 u8 msg_type; 161 u8 msg_version; 162 u16 msg_sz; 163 u32 rsvd2; 164 u8 msg_vmpck; 165 u8 rsvd3[35]; 166 } __packed; 167 168 struct snp_guest_msg { 169 struct snp_guest_msg_hdr hdr; 170 u8 payload[PAGE_SIZE - sizeof(struct snp_guest_msg_hdr)]; 171 } __packed; 172 173 struct sev_guest_platform_data { 174 u64 secrets_gpa; 175 }; 176 177 struct snp_guest_req { 178 void *req_buf; 179 size_t req_sz; 180 181 void *resp_buf; 182 size_t resp_sz; 183 184 u64 exit_code; 185 unsigned int vmpck_id; 186 u8 msg_version; 187 u8 msg_type; 188 }; 189 190 /* 191 * The secrets page contains 96-bytes of reserved field that can be used by 192 * the guest OS. The guest OS uses the area to save the message sequence 193 * number for each VMPCK. 194 * 195 * See the GHCB spec section Secret page layout for the format for this area. 196 */ 197 struct secrets_os_area { 198 u32 msg_seqno_0; 199 u32 msg_seqno_1; 200 u32 msg_seqno_2; 201 u32 msg_seqno_3; 202 u64 ap_jump_table_pa; 203 u8 rsvd[40]; 204 u8 guest_usage[32]; 205 } __packed; 206 207 #define VMPCK_KEY_LEN 32 208 209 /* See the SNP spec version 0.9 for secrets page format */ 210 struct snp_secrets_page { 211 u32 version; 212 u32 imien : 1, 213 rsvd1 : 31; 214 u32 fms; 215 u32 rsvd2; 216 u8 gosvw[16]; 217 u8 vmpck0[VMPCK_KEY_LEN]; 218 u8 vmpck1[VMPCK_KEY_LEN]; 219 u8 vmpck2[VMPCK_KEY_LEN]; 220 u8 vmpck3[VMPCK_KEY_LEN]; 221 struct secrets_os_area os_area; 222 223 u8 vmsa_tweak_bitmap[64]; 224 225 /* SVSM fields */ 226 u64 svsm_base; 227 u64 svsm_size; 228 u64 svsm_caa; 229 u32 svsm_max_version; 230 u8 svsm_guest_vmpl; 231 u8 rsvd3[3]; 232 233 /* Remainder of page */ 234 u8 rsvd4[3744]; 235 } __packed; 236 237 struct snp_msg_desc { 238 /* request and response are in unencrypted memory */ 239 struct snp_guest_msg *request, *response; 240 241 /* 242 * Avoid information leakage by double-buffering shared messages 243 * in fields that are in regular encrypted memory. 244 */ 245 struct snp_guest_msg secret_request, secret_response; 246 247 struct snp_secrets_page *secrets; 248 struct snp_req_data input; 249 250 void *certs_data; 251 252 struct aesgcm_ctx *ctx; 253 254 u32 *os_area_msg_seqno; 255 u8 *vmpck; 256 }; 257 258 /* 259 * The SVSM Calling Area (CA) related structures. 260 */ 261 struct svsm_ca { 262 u8 call_pending; 263 u8 mem_available; 264 u8 rsvd1[6]; 265 266 u8 svsm_buffer[PAGE_SIZE - 8]; 267 }; 268 269 #define SVSM_SUCCESS 0 270 #define SVSM_ERR_INCOMPLETE 0x80000000 271 #define SVSM_ERR_UNSUPPORTED_PROTOCOL 0x80000001 272 #define SVSM_ERR_UNSUPPORTED_CALL 0x80000002 273 #define SVSM_ERR_INVALID_ADDRESS 0x80000003 274 #define SVSM_ERR_INVALID_FORMAT 0x80000004 275 #define SVSM_ERR_INVALID_PARAMETER 0x80000005 276 #define SVSM_ERR_INVALID_REQUEST 0x80000006 277 #define SVSM_ERR_BUSY 0x80000007 278 #define SVSM_PVALIDATE_FAIL_SIZEMISMATCH 0x80001006 279 280 /* 281 * The SVSM PVALIDATE related structures 282 */ 283 struct svsm_pvalidate_entry { 284 u64 page_size : 2, 285 action : 1, 286 ignore_cf : 1, 287 rsvd : 8, 288 pfn : 52; 289 }; 290 291 struct svsm_pvalidate_call { 292 u16 num_entries; 293 u16 cur_index; 294 295 u8 rsvd1[4]; 296 297 struct svsm_pvalidate_entry entry[]; 298 }; 299 300 #define SVSM_PVALIDATE_MAX_COUNT ((sizeof_field(struct svsm_ca, svsm_buffer) - \ 301 offsetof(struct svsm_pvalidate_call, entry)) / \ 302 sizeof(struct svsm_pvalidate_entry)) 303 304 /* 305 * The SVSM Attestation related structures 306 */ 307 struct svsm_loc_entry { 308 u64 pa; 309 u32 len; 310 u8 rsvd[4]; 311 }; 312 313 struct svsm_attest_call { 314 struct svsm_loc_entry report_buf; 315 struct svsm_loc_entry nonce; 316 struct svsm_loc_entry manifest_buf; 317 struct svsm_loc_entry certificates_buf; 318 319 /* For attesting a single service */ 320 u8 service_guid[16]; 321 u32 service_manifest_ver; 322 u8 rsvd[4]; 323 }; 324 325 /* PTE descriptor used for the prepare_pte_enc() operations. */ 326 struct pte_enc_desc { 327 pte_t *kpte; 328 int pte_level; 329 bool encrypt; 330 /* pfn of the kpte above */ 331 unsigned long pfn; 332 /* physical address of @pfn */ 333 unsigned long pa; 334 /* virtual address of @pfn */ 335 void *va; 336 /* memory covered by the pte */ 337 unsigned long size; 338 pgprot_t new_pgprot; 339 }; 340 341 /* 342 * SVSM protocol structure 343 */ 344 struct svsm_call { 345 struct svsm_ca *caa; 346 u64 rax; 347 u64 rcx; 348 u64 rdx; 349 u64 r8; 350 u64 r9; 351 u64 rax_out; 352 u64 rcx_out; 353 u64 rdx_out; 354 u64 r8_out; 355 u64 r9_out; 356 }; 357 358 #define SVSM_CORE_CALL(x) ((0ULL << 32) | (x)) 359 #define SVSM_CORE_REMAP_CA 0 360 #define SVSM_CORE_PVALIDATE 1 361 #define SVSM_CORE_CREATE_VCPU 2 362 #define SVSM_CORE_DELETE_VCPU 3 363 364 #define SVSM_ATTEST_CALL(x) ((1ULL << 32) | (x)) 365 #define SVSM_ATTEST_SERVICES 0 366 #define SVSM_ATTEST_SINGLE_SERVICE 1 367 368 #ifdef CONFIG_AMD_MEM_ENCRYPT 369 370 extern u8 snp_vmpl; 371 372 extern void __sev_es_ist_enter(struct pt_regs *regs); 373 extern void __sev_es_ist_exit(void); 374 static __always_inline void sev_es_ist_enter(struct pt_regs *regs) 375 { 376 if (cc_vendor == CC_VENDOR_AMD && 377 cc_platform_has(CC_ATTR_GUEST_STATE_ENCRYPT)) 378 __sev_es_ist_enter(regs); 379 } 380 static __always_inline void sev_es_ist_exit(void) 381 { 382 if (cc_vendor == CC_VENDOR_AMD && 383 cc_platform_has(CC_ATTR_GUEST_STATE_ENCRYPT)) 384 __sev_es_ist_exit(); 385 } 386 extern int sev_es_setup_ap_jump_table(struct real_mode_header *rmh); 387 extern void __sev_es_nmi_complete(void); 388 static __always_inline void sev_es_nmi_complete(void) 389 { 390 if (cc_vendor == CC_VENDOR_AMD && 391 cc_platform_has(CC_ATTR_GUEST_STATE_ENCRYPT)) 392 __sev_es_nmi_complete(); 393 } 394 extern int __init sev_es_efi_map_ghcbs(pgd_t *pgd); 395 extern void sev_enable(struct boot_params *bp); 396 397 /* 398 * RMPADJUST modifies the RMP permissions of a page of a lesser- 399 * privileged (numerically higher) VMPL. 400 * 401 * If the guest is running at a higher-privilege than the privilege 402 * level the instruction is targeting, the instruction will succeed, 403 * otherwise, it will fail. 404 */ 405 static inline int rmpadjust(unsigned long vaddr, bool rmp_psize, unsigned long attrs) 406 { 407 int rc; 408 409 /* "rmpadjust" mnemonic support in binutils 2.36 and newer */ 410 asm volatile(".byte 0xF3,0x0F,0x01,0xFE\n\t" 411 : "=a"(rc) 412 : "a"(vaddr), "c"(rmp_psize), "d"(attrs) 413 : "memory", "cc"); 414 415 return rc; 416 } 417 static inline int pvalidate(unsigned long vaddr, bool rmp_psize, bool validate) 418 { 419 bool no_rmpupdate; 420 int rc; 421 422 /* "pvalidate" mnemonic support in binutils 2.36 and newer */ 423 asm volatile(".byte 0xF2, 0x0F, 0x01, 0xFF\n\t" 424 CC_SET(c) 425 : CC_OUT(c) (no_rmpupdate), "=a"(rc) 426 : "a"(vaddr), "c"(rmp_psize), "d"(validate) 427 : "memory", "cc"); 428 429 if (no_rmpupdate) 430 return PVALIDATE_FAIL_NOUPDATE; 431 432 return rc; 433 } 434 435 struct snp_guest_request_ioctl; 436 437 void setup_ghcb(void); 438 void early_snp_set_memory_private(unsigned long vaddr, unsigned long paddr, 439 unsigned long npages); 440 void early_snp_set_memory_shared(unsigned long vaddr, unsigned long paddr, 441 unsigned long npages); 442 void snp_set_memory_shared(unsigned long vaddr, unsigned long npages); 443 void snp_set_memory_private(unsigned long vaddr, unsigned long npages); 444 void snp_set_wakeup_secondary_cpu(void); 445 bool snp_init(struct boot_params *bp); 446 void __noreturn snp_abort(void); 447 void snp_dmi_setup(void); 448 int snp_issue_guest_request(struct snp_guest_req *req, struct snp_req_data *input, 449 struct snp_guest_request_ioctl *rio); 450 int snp_issue_svsm_attest_req(u64 call_id, struct svsm_call *call, struct svsm_attest_call *input); 451 void snp_accept_memory(phys_addr_t start, phys_addr_t end); 452 u64 snp_get_unsupported_features(u64 status); 453 u64 sev_get_status(void); 454 void sev_show_status(void); 455 void snp_update_svsm_ca(void); 456 int prepare_pte_enc(struct pte_enc_desc *d); 457 void set_pte_enc_mask(pte_t *kpte, unsigned long pfn, pgprot_t new_prot); 458 void snp_kexec_finish(void); 459 void snp_kexec_begin(void); 460 461 #else /* !CONFIG_AMD_MEM_ENCRYPT */ 462 463 #define snp_vmpl 0 464 static inline void sev_es_ist_enter(struct pt_regs *regs) { } 465 static inline void sev_es_ist_exit(void) { } 466 static inline int sev_es_setup_ap_jump_table(struct real_mode_header *rmh) { return 0; } 467 static inline void sev_es_nmi_complete(void) { } 468 static inline int sev_es_efi_map_ghcbs(pgd_t *pgd) { return 0; } 469 static inline void sev_enable(struct boot_params *bp) { } 470 static inline int pvalidate(unsigned long vaddr, bool rmp_psize, bool validate) { return 0; } 471 static inline int rmpadjust(unsigned long vaddr, bool rmp_psize, unsigned long attrs) { return 0; } 472 static inline void setup_ghcb(void) { } 473 static inline void __init 474 early_snp_set_memory_private(unsigned long vaddr, unsigned long paddr, unsigned long npages) { } 475 static inline void __init 476 early_snp_set_memory_shared(unsigned long vaddr, unsigned long paddr, unsigned long npages) { } 477 static inline void snp_set_memory_shared(unsigned long vaddr, unsigned long npages) { } 478 static inline void snp_set_memory_private(unsigned long vaddr, unsigned long npages) { } 479 static inline void snp_set_wakeup_secondary_cpu(void) { } 480 static inline bool snp_init(struct boot_params *bp) { return false; } 481 static inline void snp_abort(void) { } 482 static inline void snp_dmi_setup(void) { } 483 static inline int snp_issue_guest_request(struct snp_guest_req *req, struct snp_req_data *input, 484 struct snp_guest_request_ioctl *rio) 485 { 486 return -ENOTTY; 487 } 488 static inline int snp_issue_svsm_attest_req(u64 call_id, struct svsm_call *call, struct svsm_attest_call *input) 489 { 490 return -ENOTTY; 491 } 492 static inline void snp_accept_memory(phys_addr_t start, phys_addr_t end) { } 493 static inline u64 snp_get_unsupported_features(u64 status) { return 0; } 494 static inline u64 sev_get_status(void) { return 0; } 495 static inline void sev_show_status(void) { } 496 static inline void snp_update_svsm_ca(void) { } 497 static inline int prepare_pte_enc(struct pte_enc_desc *d) { return 0; } 498 static inline void set_pte_enc_mask(pte_t *kpte, unsigned long pfn, pgprot_t new_prot) { } 499 static inline void snp_kexec_finish(void) { } 500 static inline void snp_kexec_begin(void) { } 501 502 #endif /* CONFIG_AMD_MEM_ENCRYPT */ 503 504 #ifdef CONFIG_KVM_AMD_SEV 505 bool snp_probe_rmptable_info(void); 506 int snp_lookup_rmpentry(u64 pfn, bool *assigned, int *level); 507 void snp_dump_hva_rmpentry(unsigned long address); 508 int psmash(u64 pfn); 509 int rmp_make_private(u64 pfn, u64 gpa, enum pg_level level, u32 asid, bool immutable); 510 int rmp_make_shared(u64 pfn, enum pg_level level); 511 void snp_leak_pages(u64 pfn, unsigned int npages); 512 void kdump_sev_callback(void); 513 void snp_fixup_e820_tables(void); 514 #else 515 static inline bool snp_probe_rmptable_info(void) { return false; } 516 static inline int snp_lookup_rmpentry(u64 pfn, bool *assigned, int *level) { return -ENODEV; } 517 static inline void snp_dump_hva_rmpentry(unsigned long address) {} 518 static inline int psmash(u64 pfn) { return -ENODEV; } 519 static inline int rmp_make_private(u64 pfn, u64 gpa, enum pg_level level, u32 asid, 520 bool immutable) 521 { 522 return -ENODEV; 523 } 524 static inline int rmp_make_shared(u64 pfn, enum pg_level level) { return -ENODEV; } 525 static inline void snp_leak_pages(u64 pfn, unsigned int npages) {} 526 static inline void kdump_sev_callback(void) { } 527 static inline void snp_fixup_e820_tables(void) {} 528 #endif 529 530 #endif 531