xref: /linux/arch/x86/include/asm/set_memory.h (revision 7f71507851fc7764b36a3221839607d3a45c2025)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef _ASM_X86_SET_MEMORY_H
3 #define _ASM_X86_SET_MEMORY_H
4 
5 #include <linux/mm.h>
6 #include <asm/page.h>
7 #include <asm-generic/set_memory.h>
8 
9 #define set_memory_rox set_memory_rox
10 int set_memory_rox(unsigned long addr, int numpages);
11 
12 /*
13  * The set_memory_* API can be used to change various attributes of a virtual
14  * address range. The attributes include:
15  * Cacheability  : UnCached, WriteCombining, WriteThrough, WriteBack
16  * Executability : eXecutable, NoteXecutable
17  * Read/Write    : ReadOnly, ReadWrite
18  * Presence      : NotPresent
19  * Encryption    : Encrypted, Decrypted
20  *
21  * Within a category, the attributes are mutually exclusive.
22  *
23  * The implementation of this API will take care of various aspects that
24  * are associated with changing such attributes, such as:
25  * - Flushing TLBs
26  * - Flushing CPU caches
27  * - Making sure aliases of the memory behind the mapping don't violate
28  *   coherency rules as defined by the CPU in the system.
29  *
30  * What this API does not do:
31  * - Provide exclusion between various callers - including callers that
32  *   operation on other mappings of the same physical page
33  * - Restore default attributes when a page is freed
34  * - Guarantee that mappings other than the requested one are
35  *   in any state, other than that these do not violate rules for
36  *   the CPU you have. Do not depend on any effects on other mappings,
37  *   CPUs other than the one you have may have more relaxed rules.
38  * The caller is required to take care of these.
39  */
40 
41 int __set_memory_prot(unsigned long addr, int numpages, pgprot_t prot);
42 int _set_memory_uc(unsigned long addr, int numpages);
43 int _set_memory_wc(unsigned long addr, int numpages);
44 int _set_memory_wt(unsigned long addr, int numpages);
45 int _set_memory_wb(unsigned long addr, int numpages);
46 int set_memory_uc(unsigned long addr, int numpages);
47 int set_memory_wc(unsigned long addr, int numpages);
48 int set_memory_wb(unsigned long addr, int numpages);
49 int set_memory_np(unsigned long addr, int numpages);
50 int set_memory_p(unsigned long addr, int numpages);
51 int set_memory_4k(unsigned long addr, int numpages);
52 
53 bool set_memory_enc_stop_conversion(void);
54 int set_memory_encrypted(unsigned long addr, int numpages);
55 int set_memory_decrypted(unsigned long addr, int numpages);
56 
57 int set_memory_np_noalias(unsigned long addr, int numpages);
58 int set_memory_nonglobal(unsigned long addr, int numpages);
59 int set_memory_global(unsigned long addr, int numpages);
60 
61 int set_pages_array_uc(struct page **pages, int addrinarray);
62 int set_pages_array_wc(struct page **pages, int addrinarray);
63 int set_pages_array_wb(struct page **pages, int addrinarray);
64 
65 /*
66  * For legacy compatibility with the old APIs, a few functions
67  * are provided that work on a "struct page".
68  * These functions operate ONLY on the 1:1 kernel mapping of the
69  * memory that the struct page represents, and internally just
70  * call the set_memory_* function. See the description of the
71  * set_memory_* function for more details on conventions.
72  *
73  * These APIs should be considered *deprecated* and are likely going to
74  * be removed in the future.
75  * The reason for this is the implicit operation on the 1:1 mapping only,
76  * making this not a generally useful API.
77  *
78  * Specifically, many users of the old APIs had a virtual address,
79  * called virt_to_page() or vmalloc_to_page() on that address to
80  * get a struct page* that the old API required.
81  * To convert these cases, use set_memory_*() on the original
82  * virtual address, do not use these functions.
83  */
84 
85 int set_pages_uc(struct page *page, int numpages);
86 int set_pages_wb(struct page *page, int numpages);
87 int set_pages_ro(struct page *page, int numpages);
88 int set_pages_rw(struct page *page, int numpages);
89 
90 int set_direct_map_invalid_noflush(struct page *page);
91 int set_direct_map_default_noflush(struct page *page);
92 int set_direct_map_valid_noflush(struct page *page, unsigned nr, bool valid);
93 bool kernel_page_present(struct page *page);
94 
95 extern int kernel_set_to_readonly;
96 
97 #endif /* _ASM_X86_SET_MEMORY_H */
98