1 /* SPDX-License-Identifier: GPL-2.0-only */
2 /*
3 * arch/arm/include/asm/ptrace.h
4 *
5 * Copyright (C) 1996-2003 Russell King
6 */
7 #ifndef __ASM_ARM_PTRACE_H
8 #define __ASM_ARM_PTRACE_H
9
10 #include <uapi/asm/ptrace.h>
11
12 #ifndef __ASSEMBLY__
13 #include <linux/bitfield.h>
14 #include <linux/types.h>
15
16 struct pt_regs {
17 unsigned long uregs[18];
18 };
19
20 struct svc_pt_regs {
21 struct pt_regs regs;
22 u32 dacr;
23 u32 ttbcr;
24 };
25
26 #define to_svc_pt_regs(r) container_of(r, struct svc_pt_regs, regs)
27
28 #define user_mode(regs) \
29 (((regs)->ARM_cpsr & 0xf) == 0)
30
31 #ifdef CONFIG_ARM_THUMB
32 #define thumb_mode(regs) \
33 (((regs)->ARM_cpsr & PSR_T_BIT))
34 #else
35 #define thumb_mode(regs) (0)
36 #endif
37
38 #ifndef CONFIG_CPU_V7M
39 #define isa_mode(regs) \
40 (FIELD_GET(PSR_J_BIT, (regs)->ARM_cpsr) << 1 | \
41 FIELD_GET(PSR_T_BIT, (regs)->ARM_cpsr))
42 #else
43 #define isa_mode(regs) 1 /* Thumb */
44 #endif
45
46 #define processor_mode(regs) \
47 ((regs)->ARM_cpsr & MODE_MASK)
48
49 #define interrupts_enabled(regs) \
50 (!((regs)->ARM_cpsr & PSR_I_BIT))
51
52 #define fast_interrupts_enabled(regs) \
53 (!((regs)->ARM_cpsr & PSR_F_BIT))
54
55 /* Are the current registers suitable for user mode?
56 * (used to maintain security in signal handlers)
57 */
valid_user_regs(struct pt_regs * regs)58 static inline int valid_user_regs(struct pt_regs *regs)
59 {
60 #ifndef CONFIG_CPU_V7M
61 unsigned long mode = regs->ARM_cpsr & MODE_MASK;
62
63 /*
64 * Always clear the F (FIQ) and A (delayed abort) bits
65 */
66 regs->ARM_cpsr &= ~(PSR_F_BIT | PSR_A_BIT);
67
68 if ((regs->ARM_cpsr & PSR_I_BIT) == 0) {
69 if (mode == USR_MODE)
70 return 1;
71 if (elf_hwcap & HWCAP_26BIT && mode == USR26_MODE)
72 return 1;
73 }
74
75 /*
76 * Force CPSR to something logical...
77 */
78 regs->ARM_cpsr &= PSR_f | PSR_s | PSR_x | PSR_T_BIT | MODE32_BIT;
79 if (!(elf_hwcap & HWCAP_26BIT))
80 regs->ARM_cpsr |= USR_MODE;
81
82 return 0;
83 #else /* ifndef CONFIG_CPU_V7M */
84 return 1;
85 #endif
86 }
87
regs_return_value(struct pt_regs * regs)88 static inline long regs_return_value(struct pt_regs *regs)
89 {
90 return regs->ARM_r0;
91 }
92
93 #define instruction_pointer(regs) (regs)->ARM_pc
94
95 #ifdef CONFIG_THUMB2_KERNEL
96 #define frame_pointer(regs) (regs)->ARM_r7
97 #else
98 #define frame_pointer(regs) (regs)->ARM_fp
99 #endif
100
instruction_pointer_set(struct pt_regs * regs,unsigned long val)101 static inline void instruction_pointer_set(struct pt_regs *regs,
102 unsigned long val)
103 {
104 instruction_pointer(regs) = val;
105 }
106
107 #ifdef CONFIG_SMP
108 extern unsigned long profile_pc(struct pt_regs *regs);
109 #else
110 #define profile_pc(regs) instruction_pointer(regs)
111 #endif
112
113 #define predicate(x) ((x) & 0xf0000000)
114 #define PREDICATE_ALWAYS 0xe0000000
115
116 /*
117 * True if instr is a 32-bit thumb instruction. This works if instr
118 * is the first or only half-word of a thumb instruction. It also works
119 * when instr holds all 32-bits of a wide thumb instruction if stored
120 * in the form (first_half<<16)|(second_half)
121 */
122 #define is_wide_instruction(instr) ((unsigned)(instr) >= 0xe800)
123
124 /*
125 * kprobe-based event tracer support
126 */
127 #include <linux/compiler.h>
128 #define MAX_REG_OFFSET (offsetof(struct pt_regs, ARM_ORIG_r0))
129
130 extern int regs_query_register_offset(const char *name);
131 extern const char *regs_query_register_name(unsigned int offset);
132 extern bool regs_within_kernel_stack(struct pt_regs *regs, unsigned long addr);
133 extern unsigned long regs_get_kernel_stack_nth(struct pt_regs *regs,
134 unsigned int n);
135
136 /**
137 * regs_get_register() - get register value from its offset
138 * @regs: pt_regs from which register value is gotten
139 * @offset: offset number of the register.
140 *
141 * regs_get_register returns the value of a register whose offset from @regs.
142 * The @offset is the offset of the register in struct pt_regs.
143 * If @offset is bigger than MAX_REG_OFFSET, this returns 0.
144 */
regs_get_register(struct pt_regs * regs,unsigned int offset)145 static inline unsigned long regs_get_register(struct pt_regs *regs,
146 unsigned int offset)
147 {
148 if (unlikely(offset > MAX_REG_OFFSET))
149 return 0;
150 return *(unsigned long *)((unsigned long)regs + offset);
151 }
152
153 /* Valid only for Kernel mode traps. */
kernel_stack_pointer(struct pt_regs * regs)154 static inline unsigned long kernel_stack_pointer(struct pt_regs *regs)
155 {
156 return regs->ARM_sp;
157 }
158
user_stack_pointer(struct pt_regs * regs)159 static inline unsigned long user_stack_pointer(struct pt_regs *regs)
160 {
161 return regs->ARM_sp;
162 }
163
164 #define current_pt_regs(void) ({ (struct pt_regs *) \
165 ((current_stack_pointer | (THREAD_SIZE - 1)) - 7) - 1; \
166 })
167
regs_set_return_value(struct pt_regs * regs,unsigned long rc)168 static inline void regs_set_return_value(struct pt_regs *regs, unsigned long rc)
169 {
170 regs->ARM_r0 = rc;
171 }
172
173 /*
174 * Update ITSTATE after normal execution of an IT block instruction.
175 *
176 * The 8 IT state bits are split into two parts in CPSR:
177 * ITSTATE<1:0> are in CPSR<26:25>
178 * ITSTATE<7:2> are in CPSR<15:10>
179 */
it_advance(unsigned long cpsr)180 static inline unsigned long it_advance(unsigned long cpsr)
181 {
182 if ((cpsr & 0x06000400) == 0) {
183 /* ITSTATE<2:0> == 0 means end of IT block, so clear IT state */
184 cpsr &= ~PSR_IT_MASK;
185 } else {
186 /* We need to shift left ITSTATE<4:0> */
187 const unsigned long mask = 0x06001c00; /* Mask ITSTATE<4:0> */
188 unsigned long it = cpsr & mask;
189 it <<= 1;
190 it |= it >> (27 - 10); /* Carry ITSTATE<2> to correct place */
191 it &= mask;
192 cpsr &= ~mask;
193 cpsr |= it;
194 }
195 return cpsr;
196 }
197
198 int syscall_trace_enter(struct pt_regs *regs);
199 void syscall_trace_exit(struct pt_regs *regs);
200
201 #endif /* __ASSEMBLY__ */
202 #endif
203