xref: /linux/arch/x86/include/asm/elf.h (revision 9d9e435f3f2492bfd196acacb61cc9a9212d8170)
1b2441318SGreg Kroah-Hartman /* SPDX-License-Identifier: GPL-2.0 */
21965aae3SH. Peter Anvin #ifndef _ASM_X86_ELF_H
31965aae3SH. Peter Anvin #define _ASM_X86_ELF_H
4bb898558SAl Viro 
5bb898558SAl Viro /*
6bb898558SAl Viro  * ELF register definitions..
7bb898558SAl Viro  */
8dfb09f9bSBorislav Petkov #include <linux/thread_info.h>
9bb898558SAl Viro 
10bb898558SAl Viro #include <asm/ptrace.h>
11bb898558SAl Viro #include <asm/user.h>
12bb898558SAl Viro #include <asm/auxvec.h>
13824eea38SChang S. Bae #include <asm/fsgsbase.h>
14bb898558SAl Viro 
15bb898558SAl Viro typedef unsigned long elf_greg_t;
16bb898558SAl Viro 
17bb898558SAl Viro #define ELF_NGREG (sizeof(struct user_regs_struct) / sizeof(elf_greg_t))
18bb898558SAl Viro typedef elf_greg_t elf_gregset_t[ELF_NGREG];
19bb898558SAl Viro 
20bb898558SAl Viro typedef struct user_i387_struct elf_fpregset_t;
21bb898558SAl Viro 
22bb898558SAl Viro #ifdef __i386__
23bb898558SAl Viro 
24bb898558SAl Viro typedef struct user_fxsr_struct elf_fpxregset_t;
25bb898558SAl Viro 
26bb898558SAl Viro #define R_386_NONE	0
27bb898558SAl Viro #define R_386_32	1
28bb898558SAl Viro #define R_386_PC32	2
29bb898558SAl Viro #define R_386_GOT32	3
30bb898558SAl Viro #define R_386_PLT32	4
31bb898558SAl Viro #define R_386_COPY	5
32bb898558SAl Viro #define R_386_GLOB_DAT	6
33bb898558SAl Viro #define R_386_JMP_SLOT	7
34bb898558SAl Viro #define R_386_RELATIVE	8
35bb898558SAl Viro #define R_386_GOTOFF	9
36bb898558SAl Viro #define R_386_GOTPC	10
37bb898558SAl Viro #define R_386_NUM	11
38bb898558SAl Viro 
39bb898558SAl Viro /*
40bb898558SAl Viro  * These are used to set parameters in the core dumps.
41bb898558SAl Viro  */
42bb898558SAl Viro #define ELF_CLASS	ELFCLASS32
43bb898558SAl Viro #define ELF_DATA	ELFDATA2LSB
44bb898558SAl Viro #define ELF_ARCH	EM_386
45bb898558SAl Viro 
46bb898558SAl Viro #else
47bb898558SAl Viro 
48bb898558SAl Viro /* x86-64 relocation types */
49bb898558SAl Viro #define R_X86_64_NONE		0	/* No reloc */
50bb898558SAl Viro #define R_X86_64_64		1	/* Direct 64 bit  */
51bb898558SAl Viro #define R_X86_64_PC32		2	/* PC relative 32 bit signed */
52bb898558SAl Viro #define R_X86_64_GOT32		3	/* 32 bit GOT entry */
53bb898558SAl Viro #define R_X86_64_PLT32		4	/* 32 bit PLT address */
54bb898558SAl Viro #define R_X86_64_COPY		5	/* Copy symbol at runtime */
55bb898558SAl Viro #define R_X86_64_GLOB_DAT	6	/* Create GOT entry */
56bb898558SAl Viro #define R_X86_64_JUMP_SLOT	7	/* Create PLT entry */
57bb898558SAl Viro #define R_X86_64_RELATIVE	8	/* Adjust by program base */
58bb898558SAl Viro #define R_X86_64_GOTPCREL	9	/* 32 bit signed pc relative
59bb898558SAl Viro 					   offset to GOT */
60bb898558SAl Viro #define R_X86_64_32		10	/* Direct 32 bit zero extended */
61bb898558SAl Viro #define R_X86_64_32S		11	/* Direct 32 bit sign extended */
62bb898558SAl Viro #define R_X86_64_16		12	/* Direct 16 bit zero extended */
63bb898558SAl Viro #define R_X86_64_PC16		13	/* 16 bit sign extended pc relative */
64bb898558SAl Viro #define R_X86_64_8		14	/* Direct 8 bit sign extended  */
65bb898558SAl Viro #define R_X86_64_PC8		15	/* 8 bit sign extended pc relative */
66b40a142bSArd Biesheuvel #define R_X86_64_PC64		24	/* Place relative 64-bit signed */
67bb898558SAl Viro 
68bb898558SAl Viro /*
69bb898558SAl Viro  * These are used to set parameters in the core dumps.
70bb898558SAl Viro  */
71bb898558SAl Viro #define ELF_CLASS	ELFCLASS64
72bb898558SAl Viro #define ELF_DATA	ELFDATA2LSB
73bb898558SAl Viro #define ELF_ARCH	EM_X86_64
74bb898558SAl Viro 
75bb898558SAl Viro #endif
76bb898558SAl Viro 
77bb898558SAl Viro #include <asm/vdso.h>
78bb898558SAl Viro 
793d7ee969SAndy Lutomirski #ifdef CONFIG_X86_64
803d7ee969SAndy Lutomirski extern unsigned int vdso64_enabled;
813d7ee969SAndy Lutomirski #endif
82ab8b82eeSBrian Gerst #if defined(CONFIG_X86_32) || defined(CONFIG_IA32_EMULATION)
833d7ee969SAndy Lutomirski extern unsigned int vdso32_enabled;
843d7ee969SAndy Lutomirski #endif
85bb898558SAl Viro 
86bb898558SAl Viro /*
87bb898558SAl Viro  * This is used to ensure we don't load something for the wrong architecture.
88bb898558SAl Viro  */
89bb898558SAl Viro #define elf_check_arch_ia32(x) \
90bb898558SAl Viro 	(((x)->e_machine == EM_386) || ((x)->e_machine == EM_486))
91bb898558SAl Viro 
92bb898558SAl Viro #include <asm/processor.h>
93bb898558SAl Viro 
94bb898558SAl Viro #ifdef CONFIG_X86_32
95bb898558SAl Viro #include <asm/desc.h>
96bb898558SAl Viro 
97bb898558SAl Viro #define elf_check_arch(x)	elf_check_arch_ia32(x)
98bb898558SAl Viro 
99bb898558SAl Viro /* SVR4/i386 ABI (pages 3-31, 3-32) says that when the program starts %edx
100bb898558SAl Viro    contains a pointer to a function which might be registered using `atexit'.
101bb898558SAl Viro    This provides a mean for the dynamic linker to call DT_FINI functions for
102bb898558SAl Viro    shared libraries that have been loaded before the code runs.
103bb898558SAl Viro 
104bb898558SAl Viro    A value of 0 tells we have no such handler.
105bb898558SAl Viro 
106bb898558SAl Viro    We might as well make sure everything else is cleared too (except for %esp),
107bb898558SAl Viro    just to make things more deterministic.
108bb898558SAl Viro  */
109bb898558SAl Viro #define ELF_PLAT_INIT(_r, load_addr)		\
110bb898558SAl Viro 	do {					\
111bb898558SAl Viro 	_r->bx = 0; _r->cx = 0; _r->dx = 0;	\
112bb898558SAl Viro 	_r->si = 0; _r->di = 0; _r->bp = 0;	\
113bb898558SAl Viro 	_r->ax = 0;				\
114bb898558SAl Viro } while (0)
115bb898558SAl Viro 
116bb898558SAl Viro /*
117bb898558SAl Viro  * regs is struct pt_regs, pr_reg is elf_gregset_t (which is
118bb898558SAl Viro  * now struct_user_regs, they are different)
119bb898558SAl Viro  */
120bb898558SAl Viro 
121ccbeed3aSTejun Heo #define ELF_CORE_COPY_REGS_COMMON(pr_reg, regs)	\
122bb898558SAl Viro do {						\
123bb898558SAl Viro 	pr_reg[0] = regs->bx;			\
124bb898558SAl Viro 	pr_reg[1] = regs->cx;			\
125bb898558SAl Viro 	pr_reg[2] = regs->dx;			\
126bb898558SAl Viro 	pr_reg[3] = regs->si;			\
127bb898558SAl Viro 	pr_reg[4] = regs->di;			\
128bb898558SAl Viro 	pr_reg[5] = regs->bp;			\
129bb898558SAl Viro 	pr_reg[6] = regs->ax;			\
13099504819SAndy Lutomirski 	pr_reg[7] = regs->ds;			\
13199504819SAndy Lutomirski 	pr_reg[8] = regs->es;			\
13299504819SAndy Lutomirski 	pr_reg[9] = regs->fs;			\
133bb898558SAl Viro 	pr_reg[11] = regs->orig_ax;		\
134bb898558SAl Viro 	pr_reg[12] = regs->ip;			\
13599504819SAndy Lutomirski 	pr_reg[13] = regs->cs;			\
136bb898558SAl Viro 	pr_reg[14] = regs->flags;		\
137bb898558SAl Viro 	pr_reg[15] = regs->sp;			\
13899504819SAndy Lutomirski 	pr_reg[16] = regs->ss;			\
139bb898558SAl Viro } while (0);
140bb898558SAl Viro 
141ccbeed3aSTejun Heo #define ELF_CORE_COPY_REGS(pr_reg, regs)	\
142ccbeed3aSTejun Heo do {						\
143ccbeed3aSTejun Heo 	ELF_CORE_COPY_REGS_COMMON(pr_reg, regs);\
144ccbeed3aSTejun Heo 	pr_reg[10] = get_user_gs(regs);		\
145ccbeed3aSTejun Heo } while (0);
146ccbeed3aSTejun Heo 
147ccbeed3aSTejun Heo #define ELF_CORE_COPY_KERNEL_REGS(pr_reg, regs)	\
148ccbeed3aSTejun Heo do {						\
149ccbeed3aSTejun Heo 	ELF_CORE_COPY_REGS_COMMON(pr_reg, regs);\
150ccbeed3aSTejun Heo 	savesegment(gs, pr_reg[10]);		\
151ccbeed3aSTejun Heo } while (0);
152ccbeed3aSTejun Heo 
153bb898558SAl Viro #define ELF_PLATFORM	(utsname()->machine)
154bb898558SAl Viro #define set_personality_64bit()	do { } while (0)
155bb898558SAl Viro 
156bb898558SAl Viro #else /* CONFIG_X86_32 */
157bb898558SAl Viro 
158bb898558SAl Viro /*
159bb898558SAl Viro  * This is used to ensure we don't load something for the wrong architecture.
160bb898558SAl Viro  */
161bb898558SAl Viro #define elf_check_arch(x)			\
162bb898558SAl Viro 	((x)->e_machine == EM_X86_64)
163bb898558SAl Viro 
164d1a797f3SH. Peter Anvin #define compat_elf_check_arch(x)					\
1650e6d3112SBen Hutchings 	(elf_check_arch_ia32(x) ||					\
1660e6d3112SBen Hutchings 	 (IS_ENABLED(CONFIG_X86_X32_ABI) && (x)->e_machine == EM_X86_64))
167d1a797f3SH. Peter Anvin 
168d1a797f3SH. Peter Anvin #if __USER32_DS != __USER_DS
169d1a797f3SH. Peter Anvin # error "The following code assumes __USER32_DS == __USER_DS"
170d1a797f3SH. Peter Anvin #endif
171bb898558SAl Viro 
172bb898558SAl Viro static inline void elf_common_init(struct thread_struct *t,
173bb898558SAl Viro 				   struct pt_regs *regs, const u16 ds)
174bb898558SAl Viro {
1757bcdea4dSAndy Lutomirski 	/* ax gets execve's return value. */
1767bcdea4dSAndy Lutomirski 	/*regs->ax = */ regs->bx = regs->cx = regs->dx = 0;
1777bcdea4dSAndy Lutomirski 	regs->si = regs->di = regs->bp = 0;
178bb898558SAl Viro 	regs->r8 = regs->r9 = regs->r10 = regs->r11 = 0;
1797bcdea4dSAndy Lutomirski 	regs->r12 = regs->r13 = regs->r14 = regs->r15 = 0;
180296f781aSAndy Lutomirski 	t->fsbase = t->gsbase = 0;
181bb898558SAl Viro 	t->fsindex = t->gsindex = 0;
182bb898558SAl Viro 	t->ds = t->es = ds;
183bb898558SAl Viro }
184bb898558SAl Viro 
185bb898558SAl Viro #define ELF_PLAT_INIT(_r, load_addr)			\
18611557b24SOleg Nesterov 	elf_common_init(&current->thread, _r, 0)
187bb898558SAl Viro 
188bb898558SAl Viro #define	COMPAT_ELF_PLAT_INIT(regs, load_addr)		\
189bb898558SAl Viro 	elf_common_init(&current->thread, regs, __USER_DS)
190bb898558SAl Viro 
1917da77078SBrian Gerst void compat_start_thread(struct pt_regs *regs, u32 new_ip, u32 new_sp);
1927da77078SBrian Gerst #define compat_start_thread compat_start_thread
193bb898558SAl Viro 
194d1a797f3SH. Peter Anvin void set_personality_ia32(bool);
195d1a797f3SH. Peter Anvin #define COMPAT_SET_PERSONALITY(ex)			\
196d1a797f3SH. Peter Anvin 	set_personality_ia32((ex).e_machine == EM_X86_64)
197bb898558SAl Viro 
198bb898558SAl Viro #define COMPAT_ELF_PLATFORM			("i686")
199bb898558SAl Viro 
200bb898558SAl Viro /*
201bb898558SAl Viro  * regs is struct pt_regs, pr_reg is elf_gregset_t (which is
202bb898558SAl Viro  * now struct_user_regs, they are different). Assumes current is the process
203bb898558SAl Viro  * getting dumped.
204bb898558SAl Viro  */
205bb898558SAl Viro 
206bb898558SAl Viro #define ELF_CORE_COPY_REGS(pr_reg, regs)			\
207bb898558SAl Viro do {								\
208bb898558SAl Viro 	unsigned v;						\
209bb898558SAl Viro 	(pr_reg)[0] = (regs)->r15;				\
210bb898558SAl Viro 	(pr_reg)[1] = (regs)->r14;				\
211bb898558SAl Viro 	(pr_reg)[2] = (regs)->r13;				\
212bb898558SAl Viro 	(pr_reg)[3] = (regs)->r12;				\
213bb898558SAl Viro 	(pr_reg)[4] = (regs)->bp;				\
214bb898558SAl Viro 	(pr_reg)[5] = (regs)->bx;				\
215bb898558SAl Viro 	(pr_reg)[6] = (regs)->r11;				\
216bb898558SAl Viro 	(pr_reg)[7] = (regs)->r10;				\
217bb898558SAl Viro 	(pr_reg)[8] = (regs)->r9;				\
218bb898558SAl Viro 	(pr_reg)[9] = (regs)->r8;				\
219bb898558SAl Viro 	(pr_reg)[10] = (regs)->ax;				\
220bb898558SAl Viro 	(pr_reg)[11] = (regs)->cx;				\
221bb898558SAl Viro 	(pr_reg)[12] = (regs)->dx;				\
222bb898558SAl Viro 	(pr_reg)[13] = (regs)->si;				\
223bb898558SAl Viro 	(pr_reg)[14] = (regs)->di;				\
224bb898558SAl Viro 	(pr_reg)[15] = (regs)->orig_ax;				\
225bb898558SAl Viro 	(pr_reg)[16] = (regs)->ip;				\
226bb898558SAl Viro 	(pr_reg)[17] = (regs)->cs;				\
227bb898558SAl Viro 	(pr_reg)[18] = (regs)->flags;				\
228bb898558SAl Viro 	(pr_reg)[19] = (regs)->sp;				\
229bb898558SAl Viro 	(pr_reg)[20] = (regs)->ss;				\
230824eea38SChang S. Bae 	(pr_reg)[21] = x86_fsbase_read_cpu();			\
231824eea38SChang S. Bae 	(pr_reg)[22] = x86_gsbase_read_cpu_inactive();		\
232bb898558SAl Viro 	asm("movl %%ds,%0" : "=r" (v)); (pr_reg)[23] = v;	\
233bb898558SAl Viro 	asm("movl %%es,%0" : "=r" (v)); (pr_reg)[24] = v;	\
234bb898558SAl Viro 	asm("movl %%fs,%0" : "=r" (v)); (pr_reg)[25] = v;	\
235bb898558SAl Viro 	asm("movl %%gs,%0" : "=r" (v)); (pr_reg)[26] = v;	\
236bb898558SAl Viro } while (0);
237bb898558SAl Viro 
238bb898558SAl Viro /* I'm not sure if we can use '-' here */
239bb898558SAl Viro #define ELF_PLATFORM       ("x86_64")
240bb898558SAl Viro extern void set_personality_64bit(void);
241bb898558SAl Viro extern unsigned int sysctl_vsyscall32;
242bb898558SAl Viro extern int force_personality32;
243bb898558SAl Viro 
244bb898558SAl Viro #endif /* !CONFIG_X86_32 */
245bb898558SAl Viro 
246bb898558SAl Viro #define CORE_DUMP_USE_REGSET
247bb898558SAl Viro #define ELF_EXEC_PAGESIZE	4096
248bb898558SAl Viro 
249eab09532SKees Cook /*
250eab09532SKees Cook  * This is the base location for PIE (ET_DYN with INTERP) loads. On
251c715b72cSKees Cook  * 64-bit, this is above 4GB to leave the entire 32-bit address
252eab09532SKees Cook  * space open for things that want to use the area for 32-bit pointers.
253eab09532SKees Cook  */
254eab09532SKees Cook #define ELF_ET_DYN_BASE		(mmap_is_ia32() ? 0x000400000UL : \
255be739f4bSKirill A. Shutemov 						  (DEFAULT_MAP_WINDOW / 3 * 2))
256bb898558SAl Viro 
257bb898558SAl Viro /* This yields a mask that user programs can use to figure out what
258bb898558SAl Viro    instruction set this CPU supports.  This could be done in user space,
259bb898558SAl Viro    but it's not easy, and we've already done it here.  */
260bb898558SAl Viro 
26116aaa537SHuaitong Han #define ELF_HWCAP		(boot_cpu_data.x86_capability[CPUID_1_EDX])
262bb898558SAl Viro 
2630274f955SGrzegorz Andrejczuk extern u32 elf_hwcap2;
2640274f955SGrzegorz Andrejczuk 
2650274f955SGrzegorz Andrejczuk /*
2660274f955SGrzegorz Andrejczuk  * HWCAP2 supplies mask with kernel enabled CPU features, so that
2670274f955SGrzegorz Andrejczuk  * the application can discover that it can safely use them.
2680274f955SGrzegorz Andrejczuk  * The bits are defined in uapi/asm/hwcap2.h.
2690274f955SGrzegorz Andrejczuk  */
2700274f955SGrzegorz Andrejczuk #define ELF_HWCAP2		(elf_hwcap2)
2710274f955SGrzegorz Andrejczuk 
272bb898558SAl Viro /* This yields a string that ld.so will use to load implementation
273bb898558SAl Viro    specific libraries for optimization.  This is more specific in
274bb898558SAl Viro    intent than poking at uname or /proc/cpuinfo.
275bb898558SAl Viro 
276bb898558SAl Viro    For the moment, we have only optimizations for the Intel generations,
277bb898558SAl Viro    but that could change... */
278bb898558SAl Viro 
279bb898558SAl Viro #define SET_PERSONALITY(ex) set_personality_64bit()
280bb898558SAl Viro 
281bb898558SAl Viro /*
282bb898558SAl Viro  * An executable for which elf_read_implies_exec() returns TRUE will
283bb898558SAl Viro  * have the READ_IMPLIES_EXEC personality flag set automatically.
284*9d9e435fSKees Cook  *
285*9d9e435fSKees Cook  * The decision process for determining the results are:
286*9d9e435fSKees Cook  *
287*9d9e435fSKees Cook  *                 CPU: | lacks NX*  | has NX, ia32     | has NX, x86_64 |
288*9d9e435fSKees Cook  * ELF:                 |            |                  |                |
289*9d9e435fSKees Cook  * ---------------------|------------|------------------|----------------|
290*9d9e435fSKees Cook  * missing PT_GNU_STACK | exec-all   | exec-all         | exec-all       |
291*9d9e435fSKees Cook  * PT_GNU_STACK == RWX  | exec-all   | exec-all         | exec-all       |
292*9d9e435fSKees Cook  * PT_GNU_STACK == RW   | exec-none  | exec-none        | exec-none      |
293*9d9e435fSKees Cook  *
294*9d9e435fSKees Cook  *  exec-all  : all PROT_READ user mappings are executable, except when
295*9d9e435fSKees Cook  *              backed by files on a noexec-filesystem.
296*9d9e435fSKees Cook  *  exec-none : only PROT_EXEC user mappings are executable.
297*9d9e435fSKees Cook  *
298*9d9e435fSKees Cook  *  *this column has no architectural effect: NX markings are ignored by
299*9d9e435fSKees Cook  *   hardware, but may have behavioral effects when "wants X" collides with
300*9d9e435fSKees Cook  *   "cannot be X" constraints in memory permission flags, as in
301*9d9e435fSKees Cook  *   https://lkml.kernel.org/r/20190418055759.GA3155@mellanox.com
302*9d9e435fSKees Cook  *
303bb898558SAl Viro  */
304bb898558SAl Viro #define elf_read_implies_exec(ex, executable_stack)	\
305bb898558SAl Viro 	(executable_stack != EXSTACK_DISABLE_X)
306bb898558SAl Viro 
307bb898558SAl Viro struct task_struct;
308bb898558SAl Viro 
3093d7ee969SAndy Lutomirski #define	ARCH_DLINFO_IA32						\
310bb898558SAl Viro do {									\
3116fdc6dd9SThomas Gleixner 	if (VDSO_CURRENT_BASE) {					\
312bb898558SAl Viro 		NEW_AUX_ENT(AT_SYSINFO,	VDSO_ENTRY);			\
313bb898558SAl Viro 		NEW_AUX_ENT(AT_SYSINFO_EHDR, VDSO_CURRENT_BASE);	\
314bb898558SAl Viro 	}								\
315bb898558SAl Viro } while (0)
316bb898558SAl Viro 
3178f3e474fSDmitry Safonov /*
3188f3e474fSDmitry Safonov  * True on X86_32 or when emulating IA32 on X86_64
3198f3e474fSDmitry Safonov  */
3208f3e474fSDmitry Safonov static inline int mmap_is_ia32(void)
3218f3e474fSDmitry Safonov {
3228f3e474fSDmitry Safonov 	return IS_ENABLED(CONFIG_X86_32) ||
3238f3e474fSDmitry Safonov 	       (IS_ENABLED(CONFIG_COMPAT) &&
3248f3e474fSDmitry Safonov 		test_thread_flag(TIF_ADDR32));
3258f3e474fSDmitry Safonov }
3268f3e474fSDmitry Safonov 
327e8f01a8dSKirill A. Shutemov extern unsigned long task_size_32bit(void);
328b569bab7SKirill A. Shutemov extern unsigned long task_size_64bit(int full_addr_space);
329e13b73ddSDmitry Safonov extern unsigned long get_mmap_base(int is_legacy);
3301e0f25dbSKirill A. Shutemov extern bool mmap_address_hint_valid(unsigned long addr, unsigned long len);
3311b028f78SDmitry Safonov 
332bb898558SAl Viro #ifdef CONFIG_X86_32
333bb898558SAl Viro 
3348f3e474fSDmitry Safonov #define __STACK_RND_MASK(is32bit) (0x7ff)
33580938332SMichal Hocko #define STACK_RND_MASK (0x7ff)
33680938332SMichal Hocko 
3373d7ee969SAndy Lutomirski #define ARCH_DLINFO		ARCH_DLINFO_IA32
338bb898558SAl Viro 
339bb898558SAl Viro /* update AT_VECTOR_SIZE_ARCH if the number of NEW_AUX_ENT entries changes */
340bb898558SAl Viro 
341bb898558SAl Viro #else /* CONFIG_X86_32 */
342bb898558SAl Viro 
343bb898558SAl Viro /* 1GB for 64bit, 8MB for 32bit */
3448f3e474fSDmitry Safonov #define __STACK_RND_MASK(is32bit) ((is32bit) ? 0x7ff : 0x3fffff)
3458f3e474fSDmitry Safonov #define STACK_RND_MASK __STACK_RND_MASK(mmap_is_ia32())
346bb898558SAl Viro 
347bb898558SAl Viro #define ARCH_DLINFO							\
348bb898558SAl Viro do {									\
3493d7ee969SAndy Lutomirski 	if (vdso64_enabled)						\
350bb898558SAl Viro 		NEW_AUX_ENT(AT_SYSINFO_EHDR,				\
3516f121e54SAndy Lutomirski 			    (unsigned long __force)current->mm->context.vdso); \
352bb898558SAl Viro } while (0)
353bb898558SAl Viro 
3543d7ee969SAndy Lutomirski /* As a historical oddity, the x32 and x86_64 vDSOs are controlled together. */
355d1a797f3SH. Peter Anvin #define ARCH_DLINFO_X32							\
356d1a797f3SH. Peter Anvin do {									\
3573d7ee969SAndy Lutomirski 	if (vdso64_enabled)						\
358d1a797f3SH. Peter Anvin 		NEW_AUX_ENT(AT_SYSINFO_EHDR,				\
3596f121e54SAndy Lutomirski 			    (unsigned long __force)current->mm->context.vdso); \
360d1a797f3SH. Peter Anvin } while (0)
361d1a797f3SH. Peter Anvin 
362bb898558SAl Viro #define AT_SYSINFO		32
363bb898558SAl Viro 
364d1a797f3SH. Peter Anvin #define COMPAT_ARCH_DLINFO						\
365d1a797f3SH. Peter Anvin if (test_thread_flag(TIF_X32))						\
366d1a797f3SH. Peter Anvin 	ARCH_DLINFO_X32;						\
367d1a797f3SH. Peter Anvin else									\
3683d7ee969SAndy Lutomirski 	ARCH_DLINFO_IA32
369bb898558SAl Viro 
370bb898558SAl Viro #define COMPAT_ELF_ET_DYN_BASE	(TASK_UNMAPPED_BASE + 0x1000000)
371bb898558SAl Viro 
372bb898558SAl Viro #endif /* !CONFIG_X86_32 */
373bb898558SAl Viro 
374bb898558SAl Viro #define VDSO_CURRENT_BASE	((unsigned long)current->mm->context.vdso)
375bb898558SAl Viro 
376bb898558SAl Viro #define VDSO_ENTRY							\
3776f121e54SAndy Lutomirski 	((unsigned long)current->mm->context.vdso +			\
3780a6d1fa0SAndy Lutomirski 	 vdso_image_32.sym___kernel_vsyscall)
379bb898558SAl Viro 
380bb898558SAl Viro struct linux_binprm;
381bb898558SAl Viro 
382bb898558SAl Viro #define ARCH_HAS_SETUP_ADDITIONAL_PAGES 1
383bb898558SAl Viro extern int arch_setup_additional_pages(struct linux_binprm *bprm,
384fc5243d9SMartin Schwidefsky 				       int uses_interp);
38518d0a6fdSAndy Lutomirski extern int compat_arch_setup_additional_pages(struct linux_binprm *bprm,
38622e842d4SH. Peter Anvin 					      int uses_interp);
38718d0a6fdSAndy Lutomirski #define compat_arch_setup_additional_pages compat_arch_setup_additional_pages
388bb898558SAl Viro 
389f9902472SMichel Lespinasse /* Do not change the values. See get_align_mask() */
390dfb09f9bSBorislav Petkov enum align_flags {
391dfb09f9bSBorislav Petkov 	ALIGN_VA_32	= BIT(0),
392dfb09f9bSBorislav Petkov 	ALIGN_VA_64	= BIT(1),
393dfb09f9bSBorislav Petkov };
394dfb09f9bSBorislav Petkov 
395dfb09f9bSBorislav Petkov struct va_alignment {
396dfb09f9bSBorislav Petkov 	int flags;
397dfb09f9bSBorislav Petkov 	unsigned long mask;
3984e26d11fSHector Marco-Gisbert 	unsigned long bits;
399dfb09f9bSBorislav Petkov } ____cacheline_aligned;
400dfb09f9bSBorislav Petkov 
401dfb09f9bSBorislav Petkov extern struct va_alignment va_align;
402f9902472SMichel Lespinasse extern unsigned long align_vdso_addr(unsigned long);
4031965aae3SH. Peter Anvin #endif /* _ASM_X86_ELF_H */
404