1 /* SPDX-License-Identifier: GPL-2.0 */ 2 /* 3 * AMD Encrypted Register State Support 4 * 5 * Author: Joerg Roedel <jroedel@suse.de> 6 */ 7 8 #ifndef __ASM_ENCRYPTED_STATE_H 9 #define __ASM_ENCRYPTED_STATE_H 10 11 #include <linux/types.h> 12 #include <linux/sev-guest.h> 13 14 #include <asm/insn.h> 15 #include <asm/sev-common.h> 16 #include <asm/coco.h> 17 18 #define GHCB_PROTOCOL_MIN 1ULL 19 #define GHCB_PROTOCOL_MAX 2ULL 20 #define GHCB_DEFAULT_USAGE 0ULL 21 22 #define VMGEXIT() { asm volatile("rep; vmmcall\n\r"); } 23 24 struct boot_params; 25 26 enum es_result { 27 ES_OK, /* All good */ 28 ES_UNSUPPORTED, /* Requested operation not supported */ 29 ES_VMM_ERROR, /* Unexpected state from the VMM */ 30 ES_DECODE_FAILED, /* Instruction decoding failed */ 31 ES_EXCEPTION, /* Instruction caused exception */ 32 ES_RETRY, /* Retry instruction emulation */ 33 }; 34 35 struct es_fault_info { 36 unsigned long vector; 37 unsigned long error_code; 38 unsigned long cr2; 39 }; 40 41 struct pt_regs; 42 43 /* ES instruction emulation context */ 44 struct es_em_ctxt { 45 struct pt_regs *regs; 46 struct insn insn; 47 struct es_fault_info fi; 48 }; 49 50 /* 51 * AMD SEV Confidential computing blob structure. The structure is 52 * defined in OVMF UEFI firmware header: 53 * https://github.com/tianocore/edk2/blob/master/OvmfPkg/Include/Guid/ConfidentialComputingSevSnpBlob.h 54 */ 55 #define CC_BLOB_SEV_HDR_MAGIC 0x45444d41 56 struct cc_blob_sev_info { 57 u32 magic; 58 u16 version; 59 u16 reserved; 60 u64 secrets_phys; 61 u32 secrets_len; 62 u32 rsvd1; 63 u64 cpuid_phys; 64 u32 cpuid_len; 65 u32 rsvd2; 66 } __packed; 67 68 void do_vc_no_ghcb(struct pt_regs *regs, unsigned long exit_code); 69 70 static inline u64 lower_bits(u64 val, unsigned int bits) 71 { 72 u64 mask = (1ULL << bits) - 1; 73 74 return (val & mask); 75 } 76 77 struct real_mode_header; 78 enum stack_type; 79 80 /* Early IDT entry points for #VC handler */ 81 extern void vc_no_ghcb(void); 82 extern void vc_boot_ghcb(void); 83 extern bool handle_vc_boot_ghcb(struct pt_regs *regs); 84 85 /* PVALIDATE return codes */ 86 #define PVALIDATE_FAIL_SIZEMISMATCH 6 87 88 /* Software defined (when rFlags.CF = 1) */ 89 #define PVALIDATE_FAIL_NOUPDATE 255 90 91 /* RMUPDATE detected 4K page and 2MB page overlap. */ 92 #define RMPUPDATE_FAIL_OVERLAP 4 93 94 /* RMP page size */ 95 #define RMP_PG_SIZE_4K 0 96 #define RMP_PG_SIZE_2M 1 97 #define RMP_TO_PG_LEVEL(level) (((level) == RMP_PG_SIZE_4K) ? PG_LEVEL_4K : PG_LEVEL_2M) 98 #define PG_LEVEL_TO_RMP(level) (((level) == PG_LEVEL_4K) ? RMP_PG_SIZE_4K : RMP_PG_SIZE_2M) 99 100 struct rmp_state { 101 u64 gpa; 102 u8 assigned; 103 u8 pagesize; 104 u8 immutable; 105 u8 rsvd; 106 u32 asid; 107 } __packed; 108 109 #define RMPADJUST_VMSA_PAGE_BIT BIT(16) 110 111 /* SNP Guest message request */ 112 struct snp_req_data { 113 unsigned long req_gpa; 114 unsigned long resp_gpa; 115 unsigned long data_gpa; 116 unsigned int data_npages; 117 }; 118 119 struct sev_guest_platform_data { 120 u64 secrets_gpa; 121 }; 122 123 /* 124 * The secrets page contains 96-bytes of reserved field that can be used by 125 * the guest OS. The guest OS uses the area to save the message sequence 126 * number for each VMPCK. 127 * 128 * See the GHCB spec section Secret page layout for the format for this area. 129 */ 130 struct secrets_os_area { 131 u32 msg_seqno_0; 132 u32 msg_seqno_1; 133 u32 msg_seqno_2; 134 u32 msg_seqno_3; 135 u64 ap_jump_table_pa; 136 u8 rsvd[40]; 137 u8 guest_usage[32]; 138 } __packed; 139 140 #define VMPCK_KEY_LEN 32 141 142 /* See the SNP spec version 0.9 for secrets page format */ 143 struct snp_secrets_page { 144 u32 version; 145 u32 imien : 1, 146 rsvd1 : 31; 147 u32 fms; 148 u32 rsvd2; 149 u8 gosvw[16]; 150 u8 vmpck0[VMPCK_KEY_LEN]; 151 u8 vmpck1[VMPCK_KEY_LEN]; 152 u8 vmpck2[VMPCK_KEY_LEN]; 153 u8 vmpck3[VMPCK_KEY_LEN]; 154 struct secrets_os_area os_area; 155 156 u8 vmsa_tweak_bitmap[64]; 157 158 /* SVSM fields */ 159 u64 svsm_base; 160 u64 svsm_size; 161 u64 svsm_caa; 162 u32 svsm_max_version; 163 u8 svsm_guest_vmpl; 164 u8 rsvd3[3]; 165 166 /* Remainder of page */ 167 u8 rsvd4[3744]; 168 } __packed; 169 170 /* 171 * The SVSM Calling Area (CA) related structures. 172 */ 173 struct svsm_ca { 174 u8 call_pending; 175 u8 mem_available; 176 u8 rsvd1[6]; 177 178 u8 svsm_buffer[PAGE_SIZE - 8]; 179 }; 180 181 #define SVSM_SUCCESS 0 182 #define SVSM_ERR_INCOMPLETE 0x80000000 183 #define SVSM_ERR_UNSUPPORTED_PROTOCOL 0x80000001 184 #define SVSM_ERR_UNSUPPORTED_CALL 0x80000002 185 #define SVSM_ERR_INVALID_ADDRESS 0x80000003 186 #define SVSM_ERR_INVALID_FORMAT 0x80000004 187 #define SVSM_ERR_INVALID_PARAMETER 0x80000005 188 #define SVSM_ERR_INVALID_REQUEST 0x80000006 189 #define SVSM_ERR_BUSY 0x80000007 190 #define SVSM_PVALIDATE_FAIL_SIZEMISMATCH 0x80001006 191 192 /* 193 * The SVSM PVALIDATE related structures 194 */ 195 struct svsm_pvalidate_entry { 196 u64 page_size : 2, 197 action : 1, 198 ignore_cf : 1, 199 rsvd : 8, 200 pfn : 52; 201 }; 202 203 struct svsm_pvalidate_call { 204 u16 num_entries; 205 u16 cur_index; 206 207 u8 rsvd1[4]; 208 209 struct svsm_pvalidate_entry entry[]; 210 }; 211 212 #define SVSM_PVALIDATE_MAX_COUNT ((sizeof_field(struct svsm_ca, svsm_buffer) - \ 213 offsetof(struct svsm_pvalidate_call, entry)) / \ 214 sizeof(struct svsm_pvalidate_entry)) 215 216 /* 217 * The SVSM Attestation related structures 218 */ 219 struct svsm_loc_entry { 220 u64 pa; 221 u32 len; 222 u8 rsvd[4]; 223 }; 224 225 struct svsm_attest_call { 226 struct svsm_loc_entry report_buf; 227 struct svsm_loc_entry nonce; 228 struct svsm_loc_entry manifest_buf; 229 struct svsm_loc_entry certificates_buf; 230 231 /* For attesting a single service */ 232 u8 service_guid[16]; 233 u32 service_manifest_ver; 234 u8 rsvd[4]; 235 }; 236 237 /* 238 * SVSM protocol structure 239 */ 240 struct svsm_call { 241 struct svsm_ca *caa; 242 u64 rax; 243 u64 rcx; 244 u64 rdx; 245 u64 r8; 246 u64 r9; 247 u64 rax_out; 248 u64 rcx_out; 249 u64 rdx_out; 250 u64 r8_out; 251 u64 r9_out; 252 }; 253 254 #define SVSM_CORE_CALL(x) ((0ULL << 32) | (x)) 255 #define SVSM_CORE_REMAP_CA 0 256 #define SVSM_CORE_PVALIDATE 1 257 #define SVSM_CORE_CREATE_VCPU 2 258 #define SVSM_CORE_DELETE_VCPU 3 259 260 #define SVSM_ATTEST_CALL(x) ((1ULL << 32) | (x)) 261 #define SVSM_ATTEST_SERVICES 0 262 #define SVSM_ATTEST_SINGLE_SERVICE 1 263 264 #ifdef CONFIG_AMD_MEM_ENCRYPT 265 266 extern u8 snp_vmpl; 267 268 extern void __sev_es_ist_enter(struct pt_regs *regs); 269 extern void __sev_es_ist_exit(void); 270 static __always_inline void sev_es_ist_enter(struct pt_regs *regs) 271 { 272 if (cc_vendor == CC_VENDOR_AMD && 273 cc_platform_has(CC_ATTR_GUEST_STATE_ENCRYPT)) 274 __sev_es_ist_enter(regs); 275 } 276 static __always_inline void sev_es_ist_exit(void) 277 { 278 if (cc_vendor == CC_VENDOR_AMD && 279 cc_platform_has(CC_ATTR_GUEST_STATE_ENCRYPT)) 280 __sev_es_ist_exit(); 281 } 282 extern int sev_es_setup_ap_jump_table(struct real_mode_header *rmh); 283 extern void __sev_es_nmi_complete(void); 284 static __always_inline void sev_es_nmi_complete(void) 285 { 286 if (cc_vendor == CC_VENDOR_AMD && 287 cc_platform_has(CC_ATTR_GUEST_STATE_ENCRYPT)) 288 __sev_es_nmi_complete(); 289 } 290 extern int __init sev_es_efi_map_ghcbs(pgd_t *pgd); 291 extern void sev_enable(struct boot_params *bp); 292 293 /* 294 * RMPADJUST modifies the RMP permissions of a page of a lesser- 295 * privileged (numerically higher) VMPL. 296 * 297 * If the guest is running at a higher-privilege than the privilege 298 * level the instruction is targeting, the instruction will succeed, 299 * otherwise, it will fail. 300 */ 301 static inline int rmpadjust(unsigned long vaddr, bool rmp_psize, unsigned long attrs) 302 { 303 int rc; 304 305 /* "rmpadjust" mnemonic support in binutils 2.36 and newer */ 306 asm volatile(".byte 0xF3,0x0F,0x01,0xFE\n\t" 307 : "=a"(rc) 308 : "a"(vaddr), "c"(rmp_psize), "d"(attrs) 309 : "memory", "cc"); 310 311 return rc; 312 } 313 static inline int pvalidate(unsigned long vaddr, bool rmp_psize, bool validate) 314 { 315 bool no_rmpupdate; 316 int rc; 317 318 /* "pvalidate" mnemonic support in binutils 2.36 and newer */ 319 asm volatile(".byte 0xF2, 0x0F, 0x01, 0xFF\n\t" 320 CC_SET(c) 321 : CC_OUT(c) (no_rmpupdate), "=a"(rc) 322 : "a"(vaddr), "c"(rmp_psize), "d"(validate) 323 : "memory", "cc"); 324 325 if (no_rmpupdate) 326 return PVALIDATE_FAIL_NOUPDATE; 327 328 return rc; 329 } 330 331 struct snp_guest_request_ioctl; 332 333 void setup_ghcb(void); 334 void early_snp_set_memory_private(unsigned long vaddr, unsigned long paddr, 335 unsigned long npages); 336 void early_snp_set_memory_shared(unsigned long vaddr, unsigned long paddr, 337 unsigned long npages); 338 void snp_set_memory_shared(unsigned long vaddr, unsigned long npages); 339 void snp_set_memory_private(unsigned long vaddr, unsigned long npages); 340 void snp_set_wakeup_secondary_cpu(void); 341 bool snp_init(struct boot_params *bp); 342 void __noreturn snp_abort(void); 343 void snp_dmi_setup(void); 344 int snp_issue_guest_request(u64 exit_code, struct snp_req_data *input, struct snp_guest_request_ioctl *rio); 345 int snp_issue_svsm_attest_req(u64 call_id, struct svsm_call *call, struct svsm_attest_call *input); 346 void snp_accept_memory(phys_addr_t start, phys_addr_t end); 347 u64 snp_get_unsupported_features(u64 status); 348 u64 sev_get_status(void); 349 void sev_show_status(void); 350 void snp_update_svsm_ca(void); 351 352 #else /* !CONFIG_AMD_MEM_ENCRYPT */ 353 354 #define snp_vmpl 0 355 static inline void sev_es_ist_enter(struct pt_regs *regs) { } 356 static inline void sev_es_ist_exit(void) { } 357 static inline int sev_es_setup_ap_jump_table(struct real_mode_header *rmh) { return 0; } 358 static inline void sev_es_nmi_complete(void) { } 359 static inline int sev_es_efi_map_ghcbs(pgd_t *pgd) { return 0; } 360 static inline void sev_enable(struct boot_params *bp) { } 361 static inline int pvalidate(unsigned long vaddr, bool rmp_psize, bool validate) { return 0; } 362 static inline int rmpadjust(unsigned long vaddr, bool rmp_psize, unsigned long attrs) { return 0; } 363 static inline void setup_ghcb(void) { } 364 static inline void __init 365 early_snp_set_memory_private(unsigned long vaddr, unsigned long paddr, unsigned long npages) { } 366 static inline void __init 367 early_snp_set_memory_shared(unsigned long vaddr, unsigned long paddr, unsigned long npages) { } 368 static inline void snp_set_memory_shared(unsigned long vaddr, unsigned long npages) { } 369 static inline void snp_set_memory_private(unsigned long vaddr, unsigned long npages) { } 370 static inline void snp_set_wakeup_secondary_cpu(void) { } 371 static inline bool snp_init(struct boot_params *bp) { return false; } 372 static inline void snp_abort(void) { } 373 static inline void snp_dmi_setup(void) { } 374 static inline int snp_issue_guest_request(u64 exit_code, struct snp_req_data *input, struct snp_guest_request_ioctl *rio) 375 { 376 return -ENOTTY; 377 } 378 static inline int snp_issue_svsm_attest_req(u64 call_id, struct svsm_call *call, struct svsm_attest_call *input) 379 { 380 return -ENOTTY; 381 } 382 static inline void snp_accept_memory(phys_addr_t start, phys_addr_t end) { } 383 static inline u64 snp_get_unsupported_features(u64 status) { return 0; } 384 static inline u64 sev_get_status(void) { return 0; } 385 static inline void sev_show_status(void) { } 386 static inline void snp_update_svsm_ca(void) { } 387 388 #endif /* CONFIG_AMD_MEM_ENCRYPT */ 389 390 #ifdef CONFIG_KVM_AMD_SEV 391 bool snp_probe_rmptable_info(void); 392 int snp_lookup_rmpentry(u64 pfn, bool *assigned, int *level); 393 void snp_dump_hva_rmpentry(unsigned long address); 394 int psmash(u64 pfn); 395 int rmp_make_private(u64 pfn, u64 gpa, enum pg_level level, u32 asid, bool immutable); 396 int rmp_make_shared(u64 pfn, enum pg_level level); 397 void snp_leak_pages(u64 pfn, unsigned int npages); 398 void kdump_sev_callback(void); 399 void snp_fixup_e820_tables(void); 400 #else 401 static inline bool snp_probe_rmptable_info(void) { return false; } 402 static inline int snp_lookup_rmpentry(u64 pfn, bool *assigned, int *level) { return -ENODEV; } 403 static inline void snp_dump_hva_rmpentry(unsigned long address) {} 404 static inline int psmash(u64 pfn) { return -ENODEV; } 405 static inline int rmp_make_private(u64 pfn, u64 gpa, enum pg_level level, u32 asid, 406 bool immutable) 407 { 408 return -ENODEV; 409 } 410 static inline int rmp_make_shared(u64 pfn, enum pg_level level) { return -ENODEV; } 411 static inline void snp_leak_pages(u64 pfn, unsigned int npages) {} 412 static inline void kdump_sev_callback(void) { } 413 static inline void snp_fixup_e820_tables(void) {} 414 #endif 415 416 #endif 417