xref: /linux/arch/x86/mm/mem_encrypt_boot.S (revision 87807f77a03d0271211b75f84b2a8b88f4e8e5d4)
1/* SPDX-License-Identifier: GPL-2.0-only */
2/*
3 * AMD Memory Encryption Support
4 *
5 * Copyright (C) 2016 Advanced Micro Devices, Inc.
6 *
7 * Author: Tom Lendacky <thomas.lendacky@amd.com>
8 */
9
10#include <linux/linkage.h>
11#include <linux/pgtable.h>
12#include <asm/page.h>
13#include <asm/processor-flags.h>
14#include <asm/msr-index.h>
15#include <asm/nospec-branch.h>
16
17	.text
18	.code64
19SYM_FUNC_START(sme_encrypt_execute)
20
21	/*
22	 * Entry parameters:
23	 *   RDI - virtual address for the encrypted mapping
24	 *   RSI - virtual address for the decrypted mapping
25	 *   RDX - length to encrypt
26	 *   RCX - virtual address of the encryption workarea, including:
27	 *     - stack page (PAGE_SIZE)
28	 *     - encryption routine page (PAGE_SIZE)
29	 *     - intermediate copy buffer (PMD_SIZE)
30	 *    R8 - physical address of the pagetables to use for encryption
31	 */
32
33	push	%rbp
34	movq	%rsp, %rbp		/* RBP now has original stack pointer */
35
36	/* Set up a one page stack in the non-encrypted memory area */
37	movq	%rcx, %rax		/* Workarea stack page */
38	leaq	PAGE_SIZE(%rax), %rsp	/* Set new stack pointer */
39	addq	$PAGE_SIZE, %rax	/* Workarea encryption routine */
40
41	push	%r12
42	movq	%rdi, %r10		/* Encrypted area */
43	movq	%rsi, %r11		/* Decrypted area */
44	movq	%rdx, %r12		/* Area length */
45
46	/* Copy encryption routine into the workarea */
47	movq	%rax, %rdi				/* Workarea encryption routine */
48	leaq	__enc_copy(%rip), %rsi			/* Encryption routine */
49	movq	$(.L__enc_copy_end - __enc_copy), %rcx	/* Encryption routine length */
50	rep	movsb
51
52	/* Setup registers for call */
53	movq	%r10, %rdi		/* Encrypted area */
54	movq	%r11, %rsi		/* Decrypted area */
55	movq	%r8, %rdx		/* Pagetables used for encryption */
56	movq	%r12, %rcx		/* Area length */
57	movq	%rax, %r8		/* Workarea encryption routine */
58	addq	$PAGE_SIZE, %r8		/* Workarea intermediate copy buffer */
59
60	ANNOTATE_RETPOLINE_SAFE
61	call	*%rax			/* Call the encryption routine */
62
63	pop	%r12
64
65	movq	%rbp, %rsp		/* Restore original stack pointer */
66	pop	%rbp
67
68	/* Offset to __x86_return_thunk would be wrong here */
69	ANNOTATE_UNRET_SAFE
70	ret
71	int3
72SYM_FUNC_END(sme_encrypt_execute)
73
74SYM_FUNC_START(__enc_copy)
75/*
76 * Routine used to encrypt memory in place.
77 *   This routine must be run outside of the kernel proper since
78 *   the kernel will be encrypted during the process. So this
79 *   routine is defined here and then copied to an area outside
80 *   of the kernel where it will remain and run decrypted
81 *   during execution.
82 *
83 *   On entry the registers must be:
84 *     RDI - virtual address for the encrypted mapping
85 *     RSI - virtual address for the decrypted mapping
86 *     RDX - address of the pagetables to use for encryption
87 *     RCX - length of area
88 *      R8 - intermediate copy buffer
89 *
90 *     RAX - points to this routine
91 *
92 * The area will be encrypted by copying from the non-encrypted
93 * memory space to an intermediate buffer and then copying from the
94 * intermediate buffer back to the encrypted memory space. The physical
95 * addresses of the two mappings are the same which results in the area
96 * being encrypted "in place".
97 */
98	/* Enable the new page tables */
99	mov	%rdx, %cr3
100
101	/* Flush any global TLBs */
102	mov	%cr4, %rdx
103	andq	$~X86_CR4_PGE, %rdx
104	mov	%rdx, %cr4
105	orq	$X86_CR4_PGE, %rdx
106	mov	%rdx, %cr4
107
108	push	%r15
109	push	%r12
110
111	movq	%rcx, %r9		/* Save area length */
112	movq	%rdi, %r10		/* Save encrypted area address */
113	movq	%rsi, %r11		/* Save decrypted area address */
114
115	/* Set the PAT register PA5 entry to write-protect */
116	movl	$MSR_IA32_CR_PAT, %ecx
117	rdmsr
118	mov	%rdx, %r15		/* Save original PAT value */
119	andl	$0xffff00ff, %edx	/* Clear PA5 */
120	orl	$0x00000500, %edx	/* Set PA5 to WP */
121	wrmsr
122
123	wbinvd				/* Invalidate any cache entries */
124
125	/* Copy/encrypt up to 2MB at a time */
126	movq	$PMD_SIZE, %r12
1271:
128	cmpq	%r12, %r9
129	jnb	2f
130	movq	%r9, %r12
131
1322:
133	movq	%r11, %rsi		/* Source - decrypted area */
134	movq	%r8, %rdi		/* Dest   - intermediate copy buffer */
135	movq	%r12, %rcx
136	rep	movsb
137
138	movq	%r8, %rsi		/* Source - intermediate copy buffer */
139	movq	%r10, %rdi		/* Dest   - encrypted area */
140	movq	%r12, %rcx
141	rep	movsb
142
143	addq	%r12, %r11
144	addq	%r12, %r10
145	subq	%r12, %r9		/* Kernel length decrement */
146	jnz	1b			/* Kernel length not zero? */
147
148	/* Restore PAT register */
149	movl	$MSR_IA32_CR_PAT, %ecx
150	rdmsr
151	mov	%r15, %rdx		/* Restore original PAT value */
152	wrmsr
153
154	pop	%r12
155	pop	%r15
156
157	/* Offset to __x86_return_thunk would be wrong here */
158	ANNOTATE_UNRET_SAFE
159	ret
160	int3
161.L__enc_copy_end:
162SYM_FUNC_END(__enc_copy)
163