xref: /linux/arch/x86/include/asm/pgtable_32_areas.h (revision 24bce201d79807b668bf9d9e0aca801c5c0d5f78)
1 #ifndef _ASM_X86_PGTABLE_32_AREAS_H
2 #define _ASM_X86_PGTABLE_32_AREAS_H
3 
4 #include <asm/cpu_entry_area.h>
5 
6 /*
7  * Just any arbitrary offset to the start of the vmalloc VM area: the
8  * current 8MB value just means that there will be a 8MB "hole" after the
9  * physical memory until the kernel virtual memory starts.  That means that
10  * any out-of-bounds memory accesses will hopefully be caught.
11  * The vmalloc() routines leaves a hole of 4kB between each vmalloced
12  * area for the same reason. ;)
13  */
14 #define VMALLOC_OFFSET	(8 * 1024 * 1024)
15 
16 #ifndef __ASSEMBLY__
17 extern bool __vmalloc_start_set; /* set once high_memory is set */
18 #endif
19 
20 #define VMALLOC_START	((unsigned long)high_memory + VMALLOC_OFFSET)
21 #ifdef CONFIG_X86_PAE
22 #define LAST_PKMAP 512
23 #else
24 #define LAST_PKMAP 1024
25 #endif
26 
27 #define CPU_ENTRY_AREA_PAGES		(NR_CPUS * DIV_ROUND_UP(sizeof(struct cpu_entry_area), PAGE_SIZE))
28 
29 /* The +1 is for the readonly IDT page: */
30 #define CPU_ENTRY_AREA_BASE	\
31 	((FIXADDR_TOT_START - PAGE_SIZE*(CPU_ENTRY_AREA_PAGES+1)) & PMD_MASK)
32 
33 #define LDT_BASE_ADDR		\
34 	((CPU_ENTRY_AREA_BASE - PAGE_SIZE) & PMD_MASK)
35 
36 #define LDT_END_ADDR		(LDT_BASE_ADDR + PMD_SIZE)
37 
38 #define PKMAP_BASE		\
39 	((LDT_BASE_ADDR - PAGE_SIZE) & PMD_MASK)
40 
41 #ifdef CONFIG_HIGHMEM
42 # define VMALLOC_END	(PKMAP_BASE - 2 * PAGE_SIZE)
43 #else
44 # define VMALLOC_END	(LDT_BASE_ADDR - 2 * PAGE_SIZE)
45 #endif
46 
47 #define MODULES_VADDR	VMALLOC_START
48 #define MODULES_END	VMALLOC_END
49 #define MODULES_LEN	(MODULES_VADDR - MODULES_END)
50 
51 #define MAXMEM	(VMALLOC_END - PAGE_OFFSET - __VMALLOC_RESERVE)
52 
53 #endif /* _ASM_X86_PGTABLE_32_AREAS_H */
54