xref: /linux/arch/arm/kernel/ptrace.c (revision b7019ac550eb3916f34d79db583e9b7ea2524afa)
1 /*
2  *  linux/arch/arm/kernel/ptrace.c
3  *
4  *  By Ross Biro 1/23/92
5  * edited by Linus Torvalds
6  * ARM modifications Copyright (C) 2000 Russell King
7  *
8  * This program is free software; you can redistribute it and/or modify
9  * it under the terms of the GNU General Public License version 2 as
10  * published by the Free Software Foundation.
11  */
12 #include <linux/kernel.h>
13 #include <linux/sched/signal.h>
14 #include <linux/sched/task_stack.h>
15 #include <linux/mm.h>
16 #include <linux/elf.h>
17 #include <linux/smp.h>
18 #include <linux/ptrace.h>
19 #include <linux/user.h>
20 #include <linux/security.h>
21 #include <linux/init.h>
22 #include <linux/signal.h>
23 #include <linux/uaccess.h>
24 #include <linux/perf_event.h>
25 #include <linux/hw_breakpoint.h>
26 #include <linux/regset.h>
27 #include <linux/audit.h>
28 #include <linux/tracehook.h>
29 #include <linux/unistd.h>
30 
31 #include <asm/pgtable.h>
32 #include <asm/traps.h>
33 
34 #define CREATE_TRACE_POINTS
35 #include <trace/events/syscalls.h>
36 
37 #define REG_PC	15
38 #define REG_PSR	16
39 /*
40  * does not yet catch signals sent when the child dies.
41  * in exit.c or in signal.c.
42  */
43 
44 #if 0
45 /*
46  * Breakpoint SWI instruction: SWI &9F0001
47  */
48 #define BREAKINST_ARM	0xef9f0001
49 #define BREAKINST_THUMB	0xdf00		/* fill this in later */
50 #else
51 /*
52  * New breakpoints - use an undefined instruction.  The ARM architecture
53  * reference manual guarantees that the following instruction space
54  * will produce an undefined instruction exception on all CPUs:
55  *
56  *  ARM:   xxxx 0111 1111 xxxx xxxx xxxx 1111 xxxx
57  *  Thumb: 1101 1110 xxxx xxxx
58  */
59 #define BREAKINST_ARM	0xe7f001f0
60 #define BREAKINST_THUMB	0xde01
61 #endif
62 
63 struct pt_regs_offset {
64 	const char *name;
65 	int offset;
66 };
67 
68 #define REG_OFFSET_NAME(r) \
69 	{.name = #r, .offset = offsetof(struct pt_regs, ARM_##r)}
70 #define REG_OFFSET_END {.name = NULL, .offset = 0}
71 
72 static const struct pt_regs_offset regoffset_table[] = {
73 	REG_OFFSET_NAME(r0),
74 	REG_OFFSET_NAME(r1),
75 	REG_OFFSET_NAME(r2),
76 	REG_OFFSET_NAME(r3),
77 	REG_OFFSET_NAME(r4),
78 	REG_OFFSET_NAME(r5),
79 	REG_OFFSET_NAME(r6),
80 	REG_OFFSET_NAME(r7),
81 	REG_OFFSET_NAME(r8),
82 	REG_OFFSET_NAME(r9),
83 	REG_OFFSET_NAME(r10),
84 	REG_OFFSET_NAME(fp),
85 	REG_OFFSET_NAME(ip),
86 	REG_OFFSET_NAME(sp),
87 	REG_OFFSET_NAME(lr),
88 	REG_OFFSET_NAME(pc),
89 	REG_OFFSET_NAME(cpsr),
90 	REG_OFFSET_NAME(ORIG_r0),
91 	REG_OFFSET_END,
92 };
93 
94 /**
95  * regs_query_register_offset() - query register offset from its name
96  * @name:	the name of a register
97  *
98  * regs_query_register_offset() returns the offset of a register in struct
99  * pt_regs from its name. If the name is invalid, this returns -EINVAL;
100  */
101 int regs_query_register_offset(const char *name)
102 {
103 	const struct pt_regs_offset *roff;
104 	for (roff = regoffset_table; roff->name != NULL; roff++)
105 		if (!strcmp(roff->name, name))
106 			return roff->offset;
107 	return -EINVAL;
108 }
109 
110 /**
111  * regs_query_register_name() - query register name from its offset
112  * @offset:	the offset of a register in struct pt_regs.
113  *
114  * regs_query_register_name() returns the name of a register from its
115  * offset in struct pt_regs. If the @offset is invalid, this returns NULL;
116  */
117 const char *regs_query_register_name(unsigned int offset)
118 {
119 	const struct pt_regs_offset *roff;
120 	for (roff = regoffset_table; roff->name != NULL; roff++)
121 		if (roff->offset == offset)
122 			return roff->name;
123 	return NULL;
124 }
125 
126 /**
127  * regs_within_kernel_stack() - check the address in the stack
128  * @regs:      pt_regs which contains kernel stack pointer.
129  * @addr:      address which is checked.
130  *
131  * regs_within_kernel_stack() checks @addr is within the kernel stack page(s).
132  * If @addr is within the kernel stack, it returns true. If not, returns false.
133  */
134 bool regs_within_kernel_stack(struct pt_regs *regs, unsigned long addr)
135 {
136 	return ((addr & ~(THREAD_SIZE - 1))  ==
137 		(kernel_stack_pointer(regs) & ~(THREAD_SIZE - 1)));
138 }
139 
140 /**
141  * regs_get_kernel_stack_nth() - get Nth entry of the stack
142  * @regs:	pt_regs which contains kernel stack pointer.
143  * @n:		stack entry number.
144  *
145  * regs_get_kernel_stack_nth() returns @n th entry of the kernel stack which
146  * is specified by @regs. If the @n th entry is NOT in the kernel stack,
147  * this returns 0.
148  */
149 unsigned long regs_get_kernel_stack_nth(struct pt_regs *regs, unsigned int n)
150 {
151 	unsigned long *addr = (unsigned long *)kernel_stack_pointer(regs);
152 	addr += n;
153 	if (regs_within_kernel_stack(regs, (unsigned long)addr))
154 		return *addr;
155 	else
156 		return 0;
157 }
158 
159 /*
160  * this routine will get a word off of the processes privileged stack.
161  * the offset is how far from the base addr as stored in the THREAD.
162  * this routine assumes that all the privileged stacks are in our
163  * data space.
164  */
165 static inline long get_user_reg(struct task_struct *task, int offset)
166 {
167 	return task_pt_regs(task)->uregs[offset];
168 }
169 
170 /*
171  * this routine will put a word on the processes privileged stack.
172  * the offset is how far from the base addr as stored in the THREAD.
173  * this routine assumes that all the privileged stacks are in our
174  * data space.
175  */
176 static inline int
177 put_user_reg(struct task_struct *task, int offset, long data)
178 {
179 	struct pt_regs newregs, *regs = task_pt_regs(task);
180 	int ret = -EINVAL;
181 
182 	newregs = *regs;
183 	newregs.uregs[offset] = data;
184 
185 	if (valid_user_regs(&newregs)) {
186 		regs->uregs[offset] = data;
187 		ret = 0;
188 	}
189 
190 	return ret;
191 }
192 
193 /*
194  * Called by kernel/ptrace.c when detaching..
195  */
196 void ptrace_disable(struct task_struct *child)
197 {
198 	/* Nothing to do. */
199 }
200 
201 /*
202  * Handle hitting a breakpoint.
203  */
204 void ptrace_break(struct task_struct *tsk, struct pt_regs *regs)
205 {
206 	force_sig_fault(SIGTRAP, TRAP_BRKPT,
207 			(void __user *)instruction_pointer(regs), tsk);
208 }
209 
210 static int break_trap(struct pt_regs *regs, unsigned int instr)
211 {
212 	ptrace_break(current, regs);
213 	return 0;
214 }
215 
216 static struct undef_hook arm_break_hook = {
217 	.instr_mask	= 0x0fffffff,
218 	.instr_val	= 0x07f001f0,
219 	.cpsr_mask	= PSR_T_BIT,
220 	.cpsr_val	= 0,
221 	.fn		= break_trap,
222 };
223 
224 static struct undef_hook thumb_break_hook = {
225 	.instr_mask	= 0xffff,
226 	.instr_val	= 0xde01,
227 	.cpsr_mask	= PSR_T_BIT,
228 	.cpsr_val	= PSR_T_BIT,
229 	.fn		= break_trap,
230 };
231 
232 static struct undef_hook thumb2_break_hook = {
233 	.instr_mask	= 0xffffffff,
234 	.instr_val	= 0xf7f0a000,
235 	.cpsr_mask	= PSR_T_BIT,
236 	.cpsr_val	= PSR_T_BIT,
237 	.fn		= break_trap,
238 };
239 
240 static int __init ptrace_break_init(void)
241 {
242 	register_undef_hook(&arm_break_hook);
243 	register_undef_hook(&thumb_break_hook);
244 	register_undef_hook(&thumb2_break_hook);
245 	return 0;
246 }
247 
248 core_initcall(ptrace_break_init);
249 
250 /*
251  * Read the word at offset "off" into the "struct user".  We
252  * actually access the pt_regs stored on the kernel stack.
253  */
254 static int ptrace_read_user(struct task_struct *tsk, unsigned long off,
255 			    unsigned long __user *ret)
256 {
257 	unsigned long tmp;
258 
259 	if (off & 3)
260 		return -EIO;
261 
262 	tmp = 0;
263 	if (off == PT_TEXT_ADDR)
264 		tmp = tsk->mm->start_code;
265 	else if (off == PT_DATA_ADDR)
266 		tmp = tsk->mm->start_data;
267 	else if (off == PT_TEXT_END_ADDR)
268 		tmp = tsk->mm->end_code;
269 	else if (off < sizeof(struct pt_regs))
270 		tmp = get_user_reg(tsk, off >> 2);
271 	else if (off >= sizeof(struct user))
272 		return -EIO;
273 
274 	return put_user(tmp, ret);
275 }
276 
277 /*
278  * Write the word at offset "off" into "struct user".  We
279  * actually access the pt_regs stored on the kernel stack.
280  */
281 static int ptrace_write_user(struct task_struct *tsk, unsigned long off,
282 			     unsigned long val)
283 {
284 	if (off & 3 || off >= sizeof(struct user))
285 		return -EIO;
286 
287 	if (off >= sizeof(struct pt_regs))
288 		return 0;
289 
290 	return put_user_reg(tsk, off >> 2, val);
291 }
292 
293 #ifdef CONFIG_IWMMXT
294 
295 /*
296  * Get the child iWMMXt state.
297  */
298 static int ptrace_getwmmxregs(struct task_struct *tsk, void __user *ufp)
299 {
300 	struct thread_info *thread = task_thread_info(tsk);
301 
302 	if (!test_ti_thread_flag(thread, TIF_USING_IWMMXT))
303 		return -ENODATA;
304 	iwmmxt_task_disable(thread);  /* force it to ram */
305 	return copy_to_user(ufp, &thread->fpstate.iwmmxt, IWMMXT_SIZE)
306 		? -EFAULT : 0;
307 }
308 
309 /*
310  * Set the child iWMMXt state.
311  */
312 static int ptrace_setwmmxregs(struct task_struct *tsk, void __user *ufp)
313 {
314 	struct thread_info *thread = task_thread_info(tsk);
315 
316 	if (!test_ti_thread_flag(thread, TIF_USING_IWMMXT))
317 		return -EACCES;
318 	iwmmxt_task_release(thread);  /* force a reload */
319 	return copy_from_user(&thread->fpstate.iwmmxt, ufp, IWMMXT_SIZE)
320 		? -EFAULT : 0;
321 }
322 
323 #endif
324 
325 #ifdef CONFIG_CRUNCH
326 /*
327  * Get the child Crunch state.
328  */
329 static int ptrace_getcrunchregs(struct task_struct *tsk, void __user *ufp)
330 {
331 	struct thread_info *thread = task_thread_info(tsk);
332 
333 	crunch_task_disable(thread);  /* force it to ram */
334 	return copy_to_user(ufp, &thread->crunchstate, CRUNCH_SIZE)
335 		? -EFAULT : 0;
336 }
337 
338 /*
339  * Set the child Crunch state.
340  */
341 static int ptrace_setcrunchregs(struct task_struct *tsk, void __user *ufp)
342 {
343 	struct thread_info *thread = task_thread_info(tsk);
344 
345 	crunch_task_release(thread);  /* force a reload */
346 	return copy_from_user(&thread->crunchstate, ufp, CRUNCH_SIZE)
347 		? -EFAULT : 0;
348 }
349 #endif
350 
351 #ifdef CONFIG_HAVE_HW_BREAKPOINT
352 /*
353  * Convert a virtual register number into an index for a thread_info
354  * breakpoint array. Breakpoints are identified using positive numbers
355  * whilst watchpoints are negative. The registers are laid out as pairs
356  * of (address, control), each pair mapping to a unique hw_breakpoint struct.
357  * Register 0 is reserved for describing resource information.
358  */
359 static int ptrace_hbp_num_to_idx(long num)
360 {
361 	if (num < 0)
362 		num = (ARM_MAX_BRP << 1) - num;
363 	return (num - 1) >> 1;
364 }
365 
366 /*
367  * Returns the virtual register number for the address of the
368  * breakpoint at index idx.
369  */
370 static long ptrace_hbp_idx_to_num(int idx)
371 {
372 	long mid = ARM_MAX_BRP << 1;
373 	long num = (idx << 1) + 1;
374 	return num > mid ? mid - num : num;
375 }
376 
377 /*
378  * Handle hitting a HW-breakpoint.
379  */
380 static void ptrace_hbptriggered(struct perf_event *bp,
381 				     struct perf_sample_data *data,
382 				     struct pt_regs *regs)
383 {
384 	struct arch_hw_breakpoint *bkpt = counter_arch_bp(bp);
385 	long num;
386 	int i;
387 
388 	for (i = 0; i < ARM_MAX_HBP_SLOTS; ++i)
389 		if (current->thread.debug.hbp[i] == bp)
390 			break;
391 
392 	num = (i == ARM_MAX_HBP_SLOTS) ? 0 : ptrace_hbp_idx_to_num(i);
393 
394 	force_sig_ptrace_errno_trap((int)num, (void __user *)(bkpt->trigger));
395 }
396 
397 /*
398  * Set ptrace breakpoint pointers to zero for this task.
399  * This is required in order to prevent child processes from unregistering
400  * breakpoints held by their parent.
401  */
402 void clear_ptrace_hw_breakpoint(struct task_struct *tsk)
403 {
404 	memset(tsk->thread.debug.hbp, 0, sizeof(tsk->thread.debug.hbp));
405 }
406 
407 /*
408  * Unregister breakpoints from this task and reset the pointers in
409  * the thread_struct.
410  */
411 void flush_ptrace_hw_breakpoint(struct task_struct *tsk)
412 {
413 	int i;
414 	struct thread_struct *t = &tsk->thread;
415 
416 	for (i = 0; i < ARM_MAX_HBP_SLOTS; i++) {
417 		if (t->debug.hbp[i]) {
418 			unregister_hw_breakpoint(t->debug.hbp[i]);
419 			t->debug.hbp[i] = NULL;
420 		}
421 	}
422 }
423 
424 static u32 ptrace_get_hbp_resource_info(void)
425 {
426 	u8 num_brps, num_wrps, debug_arch, wp_len;
427 	u32 reg = 0;
428 
429 	num_brps	= hw_breakpoint_slots(TYPE_INST);
430 	num_wrps	= hw_breakpoint_slots(TYPE_DATA);
431 	debug_arch	= arch_get_debug_arch();
432 	wp_len		= arch_get_max_wp_len();
433 
434 	reg		|= debug_arch;
435 	reg		<<= 8;
436 	reg		|= wp_len;
437 	reg		<<= 8;
438 	reg		|= num_wrps;
439 	reg		<<= 8;
440 	reg		|= num_brps;
441 
442 	return reg;
443 }
444 
445 static struct perf_event *ptrace_hbp_create(struct task_struct *tsk, int type)
446 {
447 	struct perf_event_attr attr;
448 
449 	ptrace_breakpoint_init(&attr);
450 
451 	/* Initialise fields to sane defaults. */
452 	attr.bp_addr	= 0;
453 	attr.bp_len	= HW_BREAKPOINT_LEN_4;
454 	attr.bp_type	= type;
455 	attr.disabled	= 1;
456 
457 	return register_user_hw_breakpoint(&attr, ptrace_hbptriggered, NULL,
458 					   tsk);
459 }
460 
461 static int ptrace_gethbpregs(struct task_struct *tsk, long num,
462 			     unsigned long  __user *data)
463 {
464 	u32 reg;
465 	int idx, ret = 0;
466 	struct perf_event *bp;
467 	struct arch_hw_breakpoint_ctrl arch_ctrl;
468 
469 	if (num == 0) {
470 		reg = ptrace_get_hbp_resource_info();
471 	} else {
472 		idx = ptrace_hbp_num_to_idx(num);
473 		if (idx < 0 || idx >= ARM_MAX_HBP_SLOTS) {
474 			ret = -EINVAL;
475 			goto out;
476 		}
477 
478 		bp = tsk->thread.debug.hbp[idx];
479 		if (!bp) {
480 			reg = 0;
481 			goto put;
482 		}
483 
484 		arch_ctrl = counter_arch_bp(bp)->ctrl;
485 
486 		/*
487 		 * Fix up the len because we may have adjusted it
488 		 * to compensate for an unaligned address.
489 		 */
490 		while (!(arch_ctrl.len & 0x1))
491 			arch_ctrl.len >>= 1;
492 
493 		if (num & 0x1)
494 			reg = bp->attr.bp_addr;
495 		else
496 			reg = encode_ctrl_reg(arch_ctrl);
497 	}
498 
499 put:
500 	if (put_user(reg, data))
501 		ret = -EFAULT;
502 
503 out:
504 	return ret;
505 }
506 
507 static int ptrace_sethbpregs(struct task_struct *tsk, long num,
508 			     unsigned long __user *data)
509 {
510 	int idx, gen_len, gen_type, implied_type, ret = 0;
511 	u32 user_val;
512 	struct perf_event *bp;
513 	struct arch_hw_breakpoint_ctrl ctrl;
514 	struct perf_event_attr attr;
515 
516 	if (num == 0)
517 		goto out;
518 	else if (num < 0)
519 		implied_type = HW_BREAKPOINT_RW;
520 	else
521 		implied_type = HW_BREAKPOINT_X;
522 
523 	idx = ptrace_hbp_num_to_idx(num);
524 	if (idx < 0 || idx >= ARM_MAX_HBP_SLOTS) {
525 		ret = -EINVAL;
526 		goto out;
527 	}
528 
529 	if (get_user(user_val, data)) {
530 		ret = -EFAULT;
531 		goto out;
532 	}
533 
534 	bp = tsk->thread.debug.hbp[idx];
535 	if (!bp) {
536 		bp = ptrace_hbp_create(tsk, implied_type);
537 		if (IS_ERR(bp)) {
538 			ret = PTR_ERR(bp);
539 			goto out;
540 		}
541 		tsk->thread.debug.hbp[idx] = bp;
542 	}
543 
544 	attr = bp->attr;
545 
546 	if (num & 0x1) {
547 		/* Address */
548 		attr.bp_addr	= user_val;
549 	} else {
550 		/* Control */
551 		decode_ctrl_reg(user_val, &ctrl);
552 		ret = arch_bp_generic_fields(ctrl, &gen_len, &gen_type);
553 		if (ret)
554 			goto out;
555 
556 		if ((gen_type & implied_type) != gen_type) {
557 			ret = -EINVAL;
558 			goto out;
559 		}
560 
561 		attr.bp_len	= gen_len;
562 		attr.bp_type	= gen_type;
563 		attr.disabled	= !ctrl.enabled;
564 	}
565 
566 	ret = modify_user_hw_breakpoint(bp, &attr);
567 out:
568 	return ret;
569 }
570 #endif
571 
572 /* regset get/set implementations */
573 
574 static int gpr_get(struct task_struct *target,
575 		   const struct user_regset *regset,
576 		   unsigned int pos, unsigned int count,
577 		   void *kbuf, void __user *ubuf)
578 {
579 	struct pt_regs *regs = task_pt_regs(target);
580 
581 	return user_regset_copyout(&pos, &count, &kbuf, &ubuf,
582 				   regs,
583 				   0, sizeof(*regs));
584 }
585 
586 static int gpr_set(struct task_struct *target,
587 		   const struct user_regset *regset,
588 		   unsigned int pos, unsigned int count,
589 		   const void *kbuf, const void __user *ubuf)
590 {
591 	int ret;
592 	struct pt_regs newregs = *task_pt_regs(target);
593 
594 	ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
595 				 &newregs,
596 				 0, sizeof(newregs));
597 	if (ret)
598 		return ret;
599 
600 	if (!valid_user_regs(&newregs))
601 		return -EINVAL;
602 
603 	*task_pt_regs(target) = newregs;
604 	return 0;
605 }
606 
607 static int fpa_get(struct task_struct *target,
608 		   const struct user_regset *regset,
609 		   unsigned int pos, unsigned int count,
610 		   void *kbuf, void __user *ubuf)
611 {
612 	return user_regset_copyout(&pos, &count, &kbuf, &ubuf,
613 				   &task_thread_info(target)->fpstate,
614 				   0, sizeof(struct user_fp));
615 }
616 
617 static int fpa_set(struct task_struct *target,
618 		   const struct user_regset *regset,
619 		   unsigned int pos, unsigned int count,
620 		   const void *kbuf, const void __user *ubuf)
621 {
622 	struct thread_info *thread = task_thread_info(target);
623 
624 	thread->used_cp[1] = thread->used_cp[2] = 1;
625 
626 	return user_regset_copyin(&pos, &count, &kbuf, &ubuf,
627 		&thread->fpstate,
628 		0, sizeof(struct user_fp));
629 }
630 
631 #ifdef CONFIG_VFP
632 /*
633  * VFP register get/set implementations.
634  *
635  * With respect to the kernel, struct user_fp is divided into three chunks:
636  * 16 or 32 real VFP registers (d0-d15 or d0-31)
637  *	These are transferred to/from the real registers in the task's
638  *	vfp_hard_struct.  The number of registers depends on the kernel
639  *	configuration.
640  *
641  * 16 or 0 fake VFP registers (d16-d31 or empty)
642  *	i.e., the user_vfp structure has space for 32 registers even if
643  *	the kernel doesn't have them all.
644  *
645  *	vfp_get() reads this chunk as zero where applicable
646  *	vfp_set() ignores this chunk
647  *
648  * 1 word for the FPSCR
649  *
650  * The bounds-checking logic built into user_regset_copyout and friends
651  * means that we can make a simple sequence of calls to map the relevant data
652  * to/from the specified slice of the user regset structure.
653  */
654 static int vfp_get(struct task_struct *target,
655 		   const struct user_regset *regset,
656 		   unsigned int pos, unsigned int count,
657 		   void *kbuf, void __user *ubuf)
658 {
659 	int ret;
660 	struct thread_info *thread = task_thread_info(target);
661 	struct vfp_hard_struct const *vfp = &thread->vfpstate.hard;
662 	const size_t user_fpregs_offset = offsetof(struct user_vfp, fpregs);
663 	const size_t user_fpscr_offset = offsetof(struct user_vfp, fpscr);
664 
665 	vfp_sync_hwstate(thread);
666 
667 	ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
668 				  &vfp->fpregs,
669 				  user_fpregs_offset,
670 				  user_fpregs_offset + sizeof(vfp->fpregs));
671 	if (ret)
672 		return ret;
673 
674 	ret = user_regset_copyout_zero(&pos, &count, &kbuf, &ubuf,
675 				       user_fpregs_offset + sizeof(vfp->fpregs),
676 				       user_fpscr_offset);
677 	if (ret)
678 		return ret;
679 
680 	return user_regset_copyout(&pos, &count, &kbuf, &ubuf,
681 				   &vfp->fpscr,
682 				   user_fpscr_offset,
683 				   user_fpscr_offset + sizeof(vfp->fpscr));
684 }
685 
686 /*
687  * For vfp_set() a read-modify-write is done on the VFP registers,
688  * in order to avoid writing back a half-modified set of registers on
689  * failure.
690  */
691 static int vfp_set(struct task_struct *target,
692 			  const struct user_regset *regset,
693 			  unsigned int pos, unsigned int count,
694 			  const void *kbuf, const void __user *ubuf)
695 {
696 	int ret;
697 	struct thread_info *thread = task_thread_info(target);
698 	struct vfp_hard_struct new_vfp;
699 	const size_t user_fpregs_offset = offsetof(struct user_vfp, fpregs);
700 	const size_t user_fpscr_offset = offsetof(struct user_vfp, fpscr);
701 
702 	vfp_sync_hwstate(thread);
703 	new_vfp = thread->vfpstate.hard;
704 
705 	ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
706 				  &new_vfp.fpregs,
707 				  user_fpregs_offset,
708 				  user_fpregs_offset + sizeof(new_vfp.fpregs));
709 	if (ret)
710 		return ret;
711 
712 	ret = user_regset_copyin_ignore(&pos, &count, &kbuf, &ubuf,
713 				user_fpregs_offset + sizeof(new_vfp.fpregs),
714 				user_fpscr_offset);
715 	if (ret)
716 		return ret;
717 
718 	ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
719 				 &new_vfp.fpscr,
720 				 user_fpscr_offset,
721 				 user_fpscr_offset + sizeof(new_vfp.fpscr));
722 	if (ret)
723 		return ret;
724 
725 	thread->vfpstate.hard = new_vfp;
726 	vfp_flush_hwstate(thread);
727 
728 	return 0;
729 }
730 #endif /* CONFIG_VFP */
731 
732 enum arm_regset {
733 	REGSET_GPR,
734 	REGSET_FPR,
735 #ifdef CONFIG_VFP
736 	REGSET_VFP,
737 #endif
738 };
739 
740 static const struct user_regset arm_regsets[] = {
741 	[REGSET_GPR] = {
742 		.core_note_type = NT_PRSTATUS,
743 		.n = ELF_NGREG,
744 		.size = sizeof(u32),
745 		.align = sizeof(u32),
746 		.get = gpr_get,
747 		.set = gpr_set
748 	},
749 	[REGSET_FPR] = {
750 		/*
751 		 * For the FPA regs in fpstate, the real fields are a mixture
752 		 * of sizes, so pretend that the registers are word-sized:
753 		 */
754 		.core_note_type = NT_PRFPREG,
755 		.n = sizeof(struct user_fp) / sizeof(u32),
756 		.size = sizeof(u32),
757 		.align = sizeof(u32),
758 		.get = fpa_get,
759 		.set = fpa_set
760 	},
761 #ifdef CONFIG_VFP
762 	[REGSET_VFP] = {
763 		/*
764 		 * Pretend that the VFP regs are word-sized, since the FPSCR is
765 		 * a single word dangling at the end of struct user_vfp:
766 		 */
767 		.core_note_type = NT_ARM_VFP,
768 		.n = ARM_VFPREGS_SIZE / sizeof(u32),
769 		.size = sizeof(u32),
770 		.align = sizeof(u32),
771 		.get = vfp_get,
772 		.set = vfp_set
773 	},
774 #endif /* CONFIG_VFP */
775 };
776 
777 static const struct user_regset_view user_arm_view = {
778 	.name = "arm", .e_machine = ELF_ARCH, .ei_osabi = ELF_OSABI,
779 	.regsets = arm_regsets, .n = ARRAY_SIZE(arm_regsets)
780 };
781 
782 const struct user_regset_view *task_user_regset_view(struct task_struct *task)
783 {
784 	return &user_arm_view;
785 }
786 
787 long arch_ptrace(struct task_struct *child, long request,
788 		 unsigned long addr, unsigned long data)
789 {
790 	int ret;
791 	unsigned long __user *datap = (unsigned long __user *) data;
792 
793 	switch (request) {
794 		case PTRACE_PEEKUSR:
795 			ret = ptrace_read_user(child, addr, datap);
796 			break;
797 
798 		case PTRACE_POKEUSR:
799 			ret = ptrace_write_user(child, addr, data);
800 			break;
801 
802 		case PTRACE_GETREGS:
803 			ret = copy_regset_to_user(child,
804 						  &user_arm_view, REGSET_GPR,
805 						  0, sizeof(struct pt_regs),
806 						  datap);
807 			break;
808 
809 		case PTRACE_SETREGS:
810 			ret = copy_regset_from_user(child,
811 						    &user_arm_view, REGSET_GPR,
812 						    0, sizeof(struct pt_regs),
813 						    datap);
814 			break;
815 
816 		case PTRACE_GETFPREGS:
817 			ret = copy_regset_to_user(child,
818 						  &user_arm_view, REGSET_FPR,
819 						  0, sizeof(union fp_state),
820 						  datap);
821 			break;
822 
823 		case PTRACE_SETFPREGS:
824 			ret = copy_regset_from_user(child,
825 						    &user_arm_view, REGSET_FPR,
826 						    0, sizeof(union fp_state),
827 						    datap);
828 			break;
829 
830 #ifdef CONFIG_IWMMXT
831 		case PTRACE_GETWMMXREGS:
832 			ret = ptrace_getwmmxregs(child, datap);
833 			break;
834 
835 		case PTRACE_SETWMMXREGS:
836 			ret = ptrace_setwmmxregs(child, datap);
837 			break;
838 #endif
839 
840 		case PTRACE_GET_THREAD_AREA:
841 			ret = put_user(task_thread_info(child)->tp_value[0],
842 				       datap);
843 			break;
844 
845 		case PTRACE_SET_SYSCALL:
846 			task_thread_info(child)->syscall = data;
847 			ret = 0;
848 			break;
849 
850 #ifdef CONFIG_CRUNCH
851 		case PTRACE_GETCRUNCHREGS:
852 			ret = ptrace_getcrunchregs(child, datap);
853 			break;
854 
855 		case PTRACE_SETCRUNCHREGS:
856 			ret = ptrace_setcrunchregs(child, datap);
857 			break;
858 #endif
859 
860 #ifdef CONFIG_VFP
861 		case PTRACE_GETVFPREGS:
862 			ret = copy_regset_to_user(child,
863 						  &user_arm_view, REGSET_VFP,
864 						  0, ARM_VFPREGS_SIZE,
865 						  datap);
866 			break;
867 
868 		case PTRACE_SETVFPREGS:
869 			ret = copy_regset_from_user(child,
870 						    &user_arm_view, REGSET_VFP,
871 						    0, ARM_VFPREGS_SIZE,
872 						    datap);
873 			break;
874 #endif
875 
876 #ifdef CONFIG_HAVE_HW_BREAKPOINT
877 		case PTRACE_GETHBPREGS:
878 			ret = ptrace_gethbpregs(child, addr,
879 						(unsigned long __user *)data);
880 			break;
881 		case PTRACE_SETHBPREGS:
882 			ret = ptrace_sethbpregs(child, addr,
883 						(unsigned long __user *)data);
884 			break;
885 #endif
886 
887 		default:
888 			ret = ptrace_request(child, request, addr, data);
889 			break;
890 	}
891 
892 	return ret;
893 }
894 
895 enum ptrace_syscall_dir {
896 	PTRACE_SYSCALL_ENTER = 0,
897 	PTRACE_SYSCALL_EXIT,
898 };
899 
900 static void tracehook_report_syscall(struct pt_regs *regs,
901 				    enum ptrace_syscall_dir dir)
902 {
903 	unsigned long ip;
904 
905 	/*
906 	 * IP is used to denote syscall entry/exit:
907 	 * IP = 0 -> entry, =1 -> exit
908 	 */
909 	ip = regs->ARM_ip;
910 	regs->ARM_ip = dir;
911 
912 	if (dir == PTRACE_SYSCALL_EXIT)
913 		tracehook_report_syscall_exit(regs, 0);
914 	else if (tracehook_report_syscall_entry(regs))
915 		current_thread_info()->syscall = -1;
916 
917 	regs->ARM_ip = ip;
918 }
919 
920 asmlinkage int syscall_trace_enter(struct pt_regs *regs, int scno)
921 {
922 	current_thread_info()->syscall = scno;
923 
924 	if (test_thread_flag(TIF_SYSCALL_TRACE))
925 		tracehook_report_syscall(regs, PTRACE_SYSCALL_ENTER);
926 
927 	/* Do seccomp after ptrace; syscall may have changed. */
928 #ifdef CONFIG_HAVE_ARCH_SECCOMP_FILTER
929 	if (secure_computing(NULL) == -1)
930 		return -1;
931 #else
932 	/* XXX: remove this once OABI gets fixed */
933 	secure_computing_strict(current_thread_info()->syscall);
934 #endif
935 
936 	/* Tracer or seccomp may have changed syscall. */
937 	scno = current_thread_info()->syscall;
938 
939 	if (test_thread_flag(TIF_SYSCALL_TRACEPOINT))
940 		trace_sys_enter(regs, scno);
941 
942 	audit_syscall_entry(scno, regs->ARM_r0, regs->ARM_r1, regs->ARM_r2,
943 			    regs->ARM_r3);
944 
945 	return scno;
946 }
947 
948 asmlinkage void syscall_trace_exit(struct pt_regs *regs)
949 {
950 	/*
951 	 * Audit the syscall before anything else, as a debugger may
952 	 * come in and change the current registers.
953 	 */
954 	audit_syscall_exit(regs);
955 
956 	/*
957 	 * Note that we haven't updated the ->syscall field for the
958 	 * current thread. This isn't a problem because it will have
959 	 * been set on syscall entry and there hasn't been an opportunity
960 	 * for a PTRACE_SET_SYSCALL since then.
961 	 */
962 	if (test_thread_flag(TIF_SYSCALL_TRACEPOINT))
963 		trace_sys_exit(regs, regs_return_value(regs));
964 
965 	if (test_thread_flag(TIF_SYSCALL_TRACE))
966 		tracehook_report_syscall(regs, PTRACE_SYSCALL_EXIT);
967 }
968