zd1201: convert to internal net_device_stats
[linux-2.6/mini2440.git] / arch / powerpc / kernel / ptrace.c
blob3635be61f89995bc43e5dfdbc6e0acc7327f1399
1 /*
2 * PowerPC version
3 * Copyright (C) 1995-1996 Gary Thomas (gdt@linuxppc.org)
5 * Derived from "arch/m68k/kernel/ptrace.c"
6 * Copyright (C) 1994 by Hamish Macdonald
7 * Taken from linux/kernel/ptrace.c and modified for M680x0.
8 * linux/kernel/ptrace.c is by Ross Biro 1/23/92, edited by Linus Torvalds
10 * Modified by Cort Dougan (cort@hq.fsmlabs.com)
11 * and Paul Mackerras (paulus@samba.org).
13 * This file is subject to the terms and conditions of the GNU General
14 * Public License. See the file README.legal in the main directory of
15 * this archive for more details.
18 #include <linux/kernel.h>
19 #include <linux/sched.h>
20 #include <linux/mm.h>
21 #include <linux/smp.h>
22 #include <linux/errno.h>
23 #include <linux/ptrace.h>
24 #include <linux/regset.h>
25 #include <linux/tracehook.h>
26 #include <linux/elf.h>
27 #include <linux/user.h>
28 #include <linux/security.h>
29 #include <linux/signal.h>
30 #include <linux/seccomp.h>
31 #include <linux/audit.h>
32 #ifdef CONFIG_PPC32
33 #include <linux/module.h>
34 #endif
36 #include <asm/uaccess.h>
37 #include <asm/page.h>
38 #include <asm/pgtable.h>
39 #include <asm/system.h>
42 * does not yet catch signals sent when the child dies.
43 * in exit.c or in signal.c.
47 * Set of msr bits that gdb can change on behalf of a process.
49 #if defined(CONFIG_40x) || defined(CONFIG_BOOKE)
50 #define MSR_DEBUGCHANGE 0
51 #else
52 #define MSR_DEBUGCHANGE (MSR_SE | MSR_BE)
53 #endif
56 * Max register writeable via put_reg
58 #ifdef CONFIG_PPC32
59 #define PT_MAX_PUT_REG PT_MQ
60 #else
61 #define PT_MAX_PUT_REG PT_CCR
62 #endif
64 static unsigned long get_user_msr(struct task_struct *task)
66 return task->thread.regs->msr | task->thread.fpexc_mode;
69 static int set_user_msr(struct task_struct *task, unsigned long msr)
71 task->thread.regs->msr &= ~MSR_DEBUGCHANGE;
72 task->thread.regs->msr |= msr & MSR_DEBUGCHANGE;
73 return 0;
77 * We prevent mucking around with the reserved area of trap
78 * which are used internally by the kernel.
80 static int set_user_trap(struct task_struct *task, unsigned long trap)
82 task->thread.regs->trap = trap & 0xfff0;
83 return 0;
87 * Get contents of register REGNO in task TASK.
89 unsigned long ptrace_get_reg(struct task_struct *task, int regno)
91 if (task->thread.regs == NULL)
92 return -EIO;
94 if (regno == PT_MSR)
95 return get_user_msr(task);
97 if (regno < (sizeof(struct pt_regs) / sizeof(unsigned long)))
98 return ((unsigned long *)task->thread.regs)[regno];
100 return -EIO;
104 * Write contents of register REGNO in task TASK.
106 int ptrace_put_reg(struct task_struct *task, int regno, unsigned long data)
108 if (task->thread.regs == NULL)
109 return -EIO;
111 if (regno == PT_MSR)
112 return set_user_msr(task, data);
113 if (regno == PT_TRAP)
114 return set_user_trap(task, data);
116 if (regno <= PT_MAX_PUT_REG) {
117 ((unsigned long *)task->thread.regs)[regno] = data;
118 return 0;
120 return -EIO;
123 static int gpr_get(struct task_struct *target, const struct user_regset *regset,
124 unsigned int pos, unsigned int count,
125 void *kbuf, void __user *ubuf)
127 int ret;
129 if (target->thread.regs == NULL)
130 return -EIO;
132 CHECK_FULL_REGS(target->thread.regs);
134 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
135 target->thread.regs,
136 0, offsetof(struct pt_regs, msr));
137 if (!ret) {
138 unsigned long msr = get_user_msr(target);
139 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf, &msr,
140 offsetof(struct pt_regs, msr),
141 offsetof(struct pt_regs, msr) +
142 sizeof(msr));
145 BUILD_BUG_ON(offsetof(struct pt_regs, orig_gpr3) !=
146 offsetof(struct pt_regs, msr) + sizeof(long));
148 if (!ret)
149 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
150 &target->thread.regs->orig_gpr3,
151 offsetof(struct pt_regs, orig_gpr3),
152 sizeof(struct pt_regs));
153 if (!ret)
154 ret = user_regset_copyout_zero(&pos, &count, &kbuf, &ubuf,
155 sizeof(struct pt_regs), -1);
157 return ret;
160 static int gpr_set(struct task_struct *target, const struct user_regset *regset,
161 unsigned int pos, unsigned int count,
162 const void *kbuf, const void __user *ubuf)
164 unsigned long reg;
165 int ret;
167 if (target->thread.regs == NULL)
168 return -EIO;
170 CHECK_FULL_REGS(target->thread.regs);
172 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
173 target->thread.regs,
174 0, PT_MSR * sizeof(reg));
176 if (!ret && count > 0) {
177 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf, &reg,
178 PT_MSR * sizeof(reg),
179 (PT_MSR + 1) * sizeof(reg));
180 if (!ret)
181 ret = set_user_msr(target, reg);
184 BUILD_BUG_ON(offsetof(struct pt_regs, orig_gpr3) !=
185 offsetof(struct pt_regs, msr) + sizeof(long));
187 if (!ret)
188 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
189 &target->thread.regs->orig_gpr3,
190 PT_ORIG_R3 * sizeof(reg),
191 (PT_MAX_PUT_REG + 1) * sizeof(reg));
193 if (PT_MAX_PUT_REG + 1 < PT_TRAP && !ret)
194 ret = user_regset_copyin_ignore(
195 &pos, &count, &kbuf, &ubuf,
196 (PT_MAX_PUT_REG + 1) * sizeof(reg),
197 PT_TRAP * sizeof(reg));
199 if (!ret && count > 0) {
200 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf, &reg,
201 PT_TRAP * sizeof(reg),
202 (PT_TRAP + 1) * sizeof(reg));
203 if (!ret)
204 ret = set_user_trap(target, reg);
207 if (!ret)
208 ret = user_regset_copyin_ignore(
209 &pos, &count, &kbuf, &ubuf,
210 (PT_TRAP + 1) * sizeof(reg), -1);
212 return ret;
215 static int fpr_get(struct task_struct *target, const struct user_regset *regset,
216 unsigned int pos, unsigned int count,
217 void *kbuf, void __user *ubuf)
219 #ifdef CONFIG_VSX
220 double buf[33];
221 int i;
222 #endif
223 flush_fp_to_thread(target);
225 #ifdef CONFIG_VSX
226 /* copy to local buffer then write that out */
227 for (i = 0; i < 32 ; i++)
228 buf[i] = target->thread.TS_FPR(i);
229 memcpy(&buf[32], &target->thread.fpscr, sizeof(double));
230 return user_regset_copyout(&pos, &count, &kbuf, &ubuf, buf, 0, -1);
232 #else
233 BUILD_BUG_ON(offsetof(struct thread_struct, fpscr) !=
234 offsetof(struct thread_struct, TS_FPR(32)));
236 return user_regset_copyout(&pos, &count, &kbuf, &ubuf,
237 &target->thread.fpr, 0, -1);
238 #endif
241 static int fpr_set(struct task_struct *target, const struct user_regset *regset,
242 unsigned int pos, unsigned int count,
243 const void *kbuf, const void __user *ubuf)
245 #ifdef CONFIG_VSX
246 double buf[33];
247 int i;
248 #endif
249 flush_fp_to_thread(target);
251 #ifdef CONFIG_VSX
252 /* copy to local buffer then write that out */
253 i = user_regset_copyin(&pos, &count, &kbuf, &ubuf, buf, 0, -1);
254 if (i)
255 return i;
256 for (i = 0; i < 32 ; i++)
257 target->thread.TS_FPR(i) = buf[i];
258 memcpy(&target->thread.fpscr, &buf[32], sizeof(double));
259 return 0;
260 #else
261 BUILD_BUG_ON(offsetof(struct thread_struct, fpscr) !=
262 offsetof(struct thread_struct, TS_FPR(32)));
264 return user_regset_copyin(&pos, &count, &kbuf, &ubuf,
265 &target->thread.fpr, 0, -1);
266 #endif
269 #ifdef CONFIG_ALTIVEC
271 * Get/set all the altivec registers vr0..vr31, vscr, vrsave, in one go.
272 * The transfer totals 34 quadword. Quadwords 0-31 contain the
273 * corresponding vector registers. Quadword 32 contains the vscr as the
274 * last word (offset 12) within that quadword. Quadword 33 contains the
275 * vrsave as the first word (offset 0) within the quadword.
277 * This definition of the VMX state is compatible with the current PPC32
278 * ptrace interface. This allows signal handling and ptrace to use the
279 * same structures. This also simplifies the implementation of a bi-arch
280 * (combined (32- and 64-bit) gdb.
283 static int vr_active(struct task_struct *target,
284 const struct user_regset *regset)
286 flush_altivec_to_thread(target);
287 return target->thread.used_vr ? regset->n : 0;
290 static int vr_get(struct task_struct *target, const struct user_regset *regset,
291 unsigned int pos, unsigned int count,
292 void *kbuf, void __user *ubuf)
294 int ret;
296 flush_altivec_to_thread(target);
298 BUILD_BUG_ON(offsetof(struct thread_struct, vscr) !=
299 offsetof(struct thread_struct, vr[32]));
301 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
302 &target->thread.vr, 0,
303 33 * sizeof(vector128));
304 if (!ret) {
306 * Copy out only the low-order word of vrsave.
308 union {
309 elf_vrreg_t reg;
310 u32 word;
311 } vrsave;
312 memset(&vrsave, 0, sizeof(vrsave));
313 vrsave.word = target->thread.vrsave;
314 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf, &vrsave,
315 33 * sizeof(vector128), -1);
318 return ret;
321 static int vr_set(struct task_struct *target, const struct user_regset *regset,
322 unsigned int pos, unsigned int count,
323 const void *kbuf, const void __user *ubuf)
325 int ret;
327 flush_altivec_to_thread(target);
329 BUILD_BUG_ON(offsetof(struct thread_struct, vscr) !=
330 offsetof(struct thread_struct, vr[32]));
332 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
333 &target->thread.vr, 0, 33 * sizeof(vector128));
334 if (!ret && count > 0) {
336 * We use only the first word of vrsave.
338 union {
339 elf_vrreg_t reg;
340 u32 word;
341 } vrsave;
342 memset(&vrsave, 0, sizeof(vrsave));
343 vrsave.word = target->thread.vrsave;
344 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf, &vrsave,
345 33 * sizeof(vector128), -1);
346 if (!ret)
347 target->thread.vrsave = vrsave.word;
350 return ret;
352 #endif /* CONFIG_ALTIVEC */
354 #ifdef CONFIG_VSX
356 * Currently to set and and get all the vsx state, you need to call
357 * the fp and VMX calls aswell. This only get/sets the lower 32
358 * 128bit VSX registers.
361 static int vsr_active(struct task_struct *target,
362 const struct user_regset *regset)
364 flush_vsx_to_thread(target);
365 return target->thread.used_vsr ? regset->n : 0;
368 static int vsr_get(struct task_struct *target, const struct user_regset *regset,
369 unsigned int pos, unsigned int count,
370 void *kbuf, void __user *ubuf)
372 double buf[32];
373 int ret, i;
375 flush_vsx_to_thread(target);
377 for (i = 0; i < 32 ; i++)
378 buf[i] = target->thread.fpr[i][TS_VSRLOWOFFSET];
379 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
380 buf, 0, 32 * sizeof(double));
382 return ret;
385 static int vsr_set(struct task_struct *target, const struct user_regset *regset,
386 unsigned int pos, unsigned int count,
387 const void *kbuf, const void __user *ubuf)
389 double buf[32];
390 int ret,i;
392 flush_vsx_to_thread(target);
394 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
395 buf, 0, 32 * sizeof(double));
396 for (i = 0; i < 32 ; i++)
397 target->thread.fpr[i][TS_VSRLOWOFFSET] = buf[i];
400 return ret;
402 #endif /* CONFIG_VSX */
404 #ifdef CONFIG_SPE
407 * For get_evrregs/set_evrregs functions 'data' has the following layout:
409 * struct {
410 * u32 evr[32];
411 * u64 acc;
412 * u32 spefscr;
416 static int evr_active(struct task_struct *target,
417 const struct user_regset *regset)
419 flush_spe_to_thread(target);
420 return target->thread.used_spe ? regset->n : 0;
423 static int evr_get(struct task_struct *target, const struct user_regset *regset,
424 unsigned int pos, unsigned int count,
425 void *kbuf, void __user *ubuf)
427 int ret;
429 flush_spe_to_thread(target);
431 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
432 &target->thread.evr,
433 0, sizeof(target->thread.evr));
435 BUILD_BUG_ON(offsetof(struct thread_struct, acc) + sizeof(u64) !=
436 offsetof(struct thread_struct, spefscr));
438 if (!ret)
439 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
440 &target->thread.acc,
441 sizeof(target->thread.evr), -1);
443 return ret;
446 static int evr_set(struct task_struct *target, const struct user_regset *regset,
447 unsigned int pos, unsigned int count,
448 const void *kbuf, const void __user *ubuf)
450 int ret;
452 flush_spe_to_thread(target);
454 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
455 &target->thread.evr,
456 0, sizeof(target->thread.evr));
458 BUILD_BUG_ON(offsetof(struct thread_struct, acc) + sizeof(u64) !=
459 offsetof(struct thread_struct, spefscr));
461 if (!ret)
462 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
463 &target->thread.acc,
464 sizeof(target->thread.evr), -1);
466 return ret;
468 #endif /* CONFIG_SPE */
472 * These are our native regset flavors.
474 enum powerpc_regset {
475 REGSET_GPR,
476 REGSET_FPR,
477 #ifdef CONFIG_ALTIVEC
478 REGSET_VMX,
479 #endif
480 #ifdef CONFIG_VSX
481 REGSET_VSX,
482 #endif
483 #ifdef CONFIG_SPE
484 REGSET_SPE,
485 #endif
488 static const struct user_regset native_regsets[] = {
489 [REGSET_GPR] = {
490 .core_note_type = NT_PRSTATUS, .n = ELF_NGREG,
491 .size = sizeof(long), .align = sizeof(long),
492 .get = gpr_get, .set = gpr_set
494 [REGSET_FPR] = {
495 .core_note_type = NT_PRFPREG, .n = ELF_NFPREG,
496 .size = sizeof(double), .align = sizeof(double),
497 .get = fpr_get, .set = fpr_set
499 #ifdef CONFIG_ALTIVEC
500 [REGSET_VMX] = {
501 .core_note_type = NT_PPC_VMX, .n = 34,
502 .size = sizeof(vector128), .align = sizeof(vector128),
503 .active = vr_active, .get = vr_get, .set = vr_set
505 #endif
506 #ifdef CONFIG_VSX
507 [REGSET_VSX] = {
508 .core_note_type = NT_PPC_VSX, .n = 32,
509 .size = sizeof(double), .align = sizeof(double),
510 .active = vsr_active, .get = vsr_get, .set = vsr_set
512 #endif
513 #ifdef CONFIG_SPE
514 [REGSET_SPE] = {
515 .n = 35,
516 .size = sizeof(u32), .align = sizeof(u32),
517 .active = evr_active, .get = evr_get, .set = evr_set
519 #endif
522 static const struct user_regset_view user_ppc_native_view = {
523 .name = UTS_MACHINE, .e_machine = ELF_ARCH, .ei_osabi = ELF_OSABI,
524 .regsets = native_regsets, .n = ARRAY_SIZE(native_regsets)
527 #ifdef CONFIG_PPC64
528 #include <linux/compat.h>
530 static int gpr32_get(struct task_struct *target,
531 const struct user_regset *regset,
532 unsigned int pos, unsigned int count,
533 void *kbuf, void __user *ubuf)
535 const unsigned long *regs = &target->thread.regs->gpr[0];
536 compat_ulong_t *k = kbuf;
537 compat_ulong_t __user *u = ubuf;
538 compat_ulong_t reg;
540 if (target->thread.regs == NULL)
541 return -EIO;
543 CHECK_FULL_REGS(target->thread.regs);
545 pos /= sizeof(reg);
546 count /= sizeof(reg);
548 if (kbuf)
549 for (; count > 0 && pos < PT_MSR; --count)
550 *k++ = regs[pos++];
551 else
552 for (; count > 0 && pos < PT_MSR; --count)
553 if (__put_user((compat_ulong_t) regs[pos++], u++))
554 return -EFAULT;
556 if (count > 0 && pos == PT_MSR) {
557 reg = get_user_msr(target);
558 if (kbuf)
559 *k++ = reg;
560 else if (__put_user(reg, u++))
561 return -EFAULT;
562 ++pos;
563 --count;
566 if (kbuf)
567 for (; count > 0 && pos < PT_REGS_COUNT; --count)
568 *k++ = regs[pos++];
569 else
570 for (; count > 0 && pos < PT_REGS_COUNT; --count)
571 if (__put_user((compat_ulong_t) regs[pos++], u++))
572 return -EFAULT;
574 kbuf = k;
575 ubuf = u;
576 pos *= sizeof(reg);
577 count *= sizeof(reg);
578 return user_regset_copyout_zero(&pos, &count, &kbuf, &ubuf,
579 PT_REGS_COUNT * sizeof(reg), -1);
582 static int gpr32_set(struct task_struct *target,
583 const struct user_regset *regset,
584 unsigned int pos, unsigned int count,
585 const void *kbuf, const void __user *ubuf)
587 unsigned long *regs = &target->thread.regs->gpr[0];
588 const compat_ulong_t *k = kbuf;
589 const compat_ulong_t __user *u = ubuf;
590 compat_ulong_t reg;
592 if (target->thread.regs == NULL)
593 return -EIO;
595 CHECK_FULL_REGS(target->thread.regs);
597 pos /= sizeof(reg);
598 count /= sizeof(reg);
600 if (kbuf)
601 for (; count > 0 && pos < PT_MSR; --count)
602 regs[pos++] = *k++;
603 else
604 for (; count > 0 && pos < PT_MSR; --count) {
605 if (__get_user(reg, u++))
606 return -EFAULT;
607 regs[pos++] = reg;
611 if (count > 0 && pos == PT_MSR) {
612 if (kbuf)
613 reg = *k++;
614 else if (__get_user(reg, u++))
615 return -EFAULT;
616 set_user_msr(target, reg);
617 ++pos;
618 --count;
621 if (kbuf) {
622 for (; count > 0 && pos <= PT_MAX_PUT_REG; --count)
623 regs[pos++] = *k++;
624 for (; count > 0 && pos < PT_TRAP; --count, ++pos)
625 ++k;
626 } else {
627 for (; count > 0 && pos <= PT_MAX_PUT_REG; --count) {
628 if (__get_user(reg, u++))
629 return -EFAULT;
630 regs[pos++] = reg;
632 for (; count > 0 && pos < PT_TRAP; --count, ++pos)
633 if (__get_user(reg, u++))
634 return -EFAULT;
637 if (count > 0 && pos == PT_TRAP) {
638 if (kbuf)
639 reg = *k++;
640 else if (__get_user(reg, u++))
641 return -EFAULT;
642 set_user_trap(target, reg);
643 ++pos;
644 --count;
647 kbuf = k;
648 ubuf = u;
649 pos *= sizeof(reg);
650 count *= sizeof(reg);
651 return user_regset_copyin_ignore(&pos, &count, &kbuf, &ubuf,
652 (PT_TRAP + 1) * sizeof(reg), -1);
656 * These are the regset flavors matching the CONFIG_PPC32 native set.
658 static const struct user_regset compat_regsets[] = {
659 [REGSET_GPR] = {
660 .core_note_type = NT_PRSTATUS, .n = ELF_NGREG,
661 .size = sizeof(compat_long_t), .align = sizeof(compat_long_t),
662 .get = gpr32_get, .set = gpr32_set
664 [REGSET_FPR] = {
665 .core_note_type = NT_PRFPREG, .n = ELF_NFPREG,
666 .size = sizeof(double), .align = sizeof(double),
667 .get = fpr_get, .set = fpr_set
669 #ifdef CONFIG_ALTIVEC
670 [REGSET_VMX] = {
671 .core_note_type = NT_PPC_VMX, .n = 34,
672 .size = sizeof(vector128), .align = sizeof(vector128),
673 .active = vr_active, .get = vr_get, .set = vr_set
675 #endif
676 #ifdef CONFIG_SPE
677 [REGSET_SPE] = {
678 .core_note_type = NT_PPC_SPE, .n = 35,
679 .size = sizeof(u32), .align = sizeof(u32),
680 .active = evr_active, .get = evr_get, .set = evr_set
682 #endif
685 static const struct user_regset_view user_ppc_compat_view = {
686 .name = "ppc", .e_machine = EM_PPC, .ei_osabi = ELF_OSABI,
687 .regsets = compat_regsets, .n = ARRAY_SIZE(compat_regsets)
689 #endif /* CONFIG_PPC64 */
691 const struct user_regset_view *task_user_regset_view(struct task_struct *task)
693 #ifdef CONFIG_PPC64
694 if (test_tsk_thread_flag(task, TIF_32BIT))
695 return &user_ppc_compat_view;
696 #endif
697 return &user_ppc_native_view;
701 void user_enable_single_step(struct task_struct *task)
703 struct pt_regs *regs = task->thread.regs;
705 if (regs != NULL) {
706 #if defined(CONFIG_40x) || defined(CONFIG_BOOKE)
707 task->thread.dbcr0 |= DBCR0_IDM | DBCR0_IC;
708 regs->msr |= MSR_DE;
709 #else
710 regs->msr |= MSR_SE;
711 #endif
713 set_tsk_thread_flag(task, TIF_SINGLESTEP);
716 void user_disable_single_step(struct task_struct *task)
718 struct pt_regs *regs = task->thread.regs;
721 #if defined(CONFIG_BOOKE)
722 /* If DAC then do not single step, skip */
723 if (task->thread.dabr)
724 return;
725 #endif
727 if (regs != NULL) {
728 #if defined(CONFIG_40x) || defined(CONFIG_BOOKE)
729 task->thread.dbcr0 &= ~(DBCR0_IC | DBCR0_IDM);
730 regs->msr &= ~MSR_DE;
731 #else
732 regs->msr &= ~MSR_SE;
733 #endif
735 clear_tsk_thread_flag(task, TIF_SINGLESTEP);
738 int ptrace_set_debugreg(struct task_struct *task, unsigned long addr,
739 unsigned long data)
741 /* For ppc64 we support one DABR and no IABR's at the moment (ppc64).
742 * For embedded processors we support one DAC and no IAC's at the
743 * moment.
745 if (addr > 0)
746 return -EINVAL;
748 /* The bottom 3 bits in dabr are flags */
749 if ((data & ~0x7UL) >= TASK_SIZE)
750 return -EIO;
752 #ifndef CONFIG_BOOKE
754 /* For processors using DABR (i.e. 970), the bottom 3 bits are flags.
755 * It was assumed, on previous implementations, that 3 bits were
756 * passed together with the data address, fitting the design of the
757 * DABR register, as follows:
759 * bit 0: Read flag
760 * bit 1: Write flag
761 * bit 2: Breakpoint translation
763 * Thus, we use them here as so.
766 /* Ensure breakpoint translation bit is set */
767 if (data && !(data & DABR_TRANSLATION))
768 return -EIO;
770 /* Move contents to the DABR register */
771 task->thread.dabr = data;
773 #endif
774 #if defined(CONFIG_BOOKE)
776 /* As described above, it was assumed 3 bits were passed with the data
777 * address, but we will assume only the mode bits will be passed
778 * as to not cause alignment restrictions for DAC-based processors.
781 /* DAC's hold the whole address without any mode flags */
782 task->thread.dabr = data & ~0x3UL;
784 if (task->thread.dabr == 0) {
785 task->thread.dbcr0 &= ~(DBSR_DAC1R | DBSR_DAC1W | DBCR0_IDM);
786 task->thread.regs->msr &= ~MSR_DE;
787 return 0;
790 /* Read or Write bits must be set */
792 if (!(data & 0x3UL))
793 return -EINVAL;
795 /* Set the Internal Debugging flag (IDM bit 1) for the DBCR0
796 register */
797 task->thread.dbcr0 = DBCR0_IDM;
799 /* Check for write and read flags and set DBCR0
800 accordingly */
801 if (data & 0x1UL)
802 task->thread.dbcr0 |= DBSR_DAC1R;
803 if (data & 0x2UL)
804 task->thread.dbcr0 |= DBSR_DAC1W;
806 task->thread.regs->msr |= MSR_DE;
807 #endif
808 return 0;
812 * Called by kernel/ptrace.c when detaching..
814 * Make sure single step bits etc are not set.
816 void ptrace_disable(struct task_struct *child)
818 /* make sure the single step bit is not set. */
819 user_disable_single_step(child);
823 * Here are the old "legacy" powerpc specific getregs/setregs ptrace calls,
824 * we mark them as obsolete now, they will be removed in a future version
826 static long arch_ptrace_old(struct task_struct *child, long request, long addr,
827 long data)
829 switch (request) {
830 case PPC_PTRACE_GETREGS: /* Get GPRs 0 - 31. */
831 return copy_regset_to_user(child, &user_ppc_native_view,
832 REGSET_GPR, 0, 32 * sizeof(long),
833 (void __user *) data);
835 case PPC_PTRACE_SETREGS: /* Set GPRs 0 - 31. */
836 return copy_regset_from_user(child, &user_ppc_native_view,
837 REGSET_GPR, 0, 32 * sizeof(long),
838 (const void __user *) data);
840 case PPC_PTRACE_GETFPREGS: /* Get FPRs 0 - 31. */
841 return copy_regset_to_user(child, &user_ppc_native_view,
842 REGSET_FPR, 0, 32 * sizeof(double),
843 (void __user *) data);
845 case PPC_PTRACE_SETFPREGS: /* Set FPRs 0 - 31. */
846 return copy_regset_from_user(child, &user_ppc_native_view,
847 REGSET_FPR, 0, 32 * sizeof(double),
848 (const void __user *) data);
851 return -EPERM;
854 long arch_ptrace(struct task_struct *child, long request, long addr, long data)
856 int ret = -EPERM;
858 switch (request) {
859 /* read the word at location addr in the USER area. */
860 case PTRACE_PEEKUSR: {
861 unsigned long index, tmp;
863 ret = -EIO;
864 /* convert to index and check */
865 #ifdef CONFIG_PPC32
866 index = (unsigned long) addr >> 2;
867 if ((addr & 3) || (index > PT_FPSCR)
868 || (child->thread.regs == NULL))
869 #else
870 index = (unsigned long) addr >> 3;
871 if ((addr & 7) || (index > PT_FPSCR))
872 #endif
873 break;
875 CHECK_FULL_REGS(child->thread.regs);
876 if (index < PT_FPR0) {
877 tmp = ptrace_get_reg(child, (int) index);
878 } else {
879 flush_fp_to_thread(child);
880 tmp = ((unsigned long *)child->thread.fpr)
881 [TS_FPRWIDTH * (index - PT_FPR0)];
883 ret = put_user(tmp,(unsigned long __user *) data);
884 break;
887 /* write the word at location addr in the USER area */
888 case PTRACE_POKEUSR: {
889 unsigned long index;
891 ret = -EIO;
892 /* convert to index and check */
893 #ifdef CONFIG_PPC32
894 index = (unsigned long) addr >> 2;
895 if ((addr & 3) || (index > PT_FPSCR)
896 || (child->thread.regs == NULL))
897 #else
898 index = (unsigned long) addr >> 3;
899 if ((addr & 7) || (index > PT_FPSCR))
900 #endif
901 break;
903 CHECK_FULL_REGS(child->thread.regs);
904 if (index < PT_FPR0) {
905 ret = ptrace_put_reg(child, index, data);
906 } else {
907 flush_fp_to_thread(child);
908 ((unsigned long *)child->thread.fpr)
909 [TS_FPRWIDTH * (index - PT_FPR0)] = data;
910 ret = 0;
912 break;
915 case PTRACE_GET_DEBUGREG: {
916 ret = -EINVAL;
917 /* We only support one DABR and no IABRS at the moment */
918 if (addr > 0)
919 break;
920 ret = put_user(child->thread.dabr,
921 (unsigned long __user *)data);
922 break;
925 case PTRACE_SET_DEBUGREG:
926 ret = ptrace_set_debugreg(child, addr, data);
927 break;
929 #ifdef CONFIG_PPC64
930 case PTRACE_GETREGS64:
931 #endif
932 case PTRACE_GETREGS: /* Get all pt_regs from the child. */
933 return copy_regset_to_user(child, &user_ppc_native_view,
934 REGSET_GPR,
935 0, sizeof(struct pt_regs),
936 (void __user *) data);
938 #ifdef CONFIG_PPC64
939 case PTRACE_SETREGS64:
940 #endif
941 case PTRACE_SETREGS: /* Set all gp regs in the child. */
942 return copy_regset_from_user(child, &user_ppc_native_view,
943 REGSET_GPR,
944 0, sizeof(struct pt_regs),
945 (const void __user *) data);
947 case PTRACE_GETFPREGS: /* Get the child FPU state (FPR0...31 + FPSCR) */
948 return copy_regset_to_user(child, &user_ppc_native_view,
949 REGSET_FPR,
950 0, sizeof(elf_fpregset_t),
951 (void __user *) data);
953 case PTRACE_SETFPREGS: /* Set the child FPU state (FPR0...31 + FPSCR) */
954 return copy_regset_from_user(child, &user_ppc_native_view,
955 REGSET_FPR,
956 0, sizeof(elf_fpregset_t),
957 (const void __user *) data);
959 #ifdef CONFIG_ALTIVEC
960 case PTRACE_GETVRREGS:
961 return copy_regset_to_user(child, &user_ppc_native_view,
962 REGSET_VMX,
963 0, (33 * sizeof(vector128) +
964 sizeof(u32)),
965 (void __user *) data);
967 case PTRACE_SETVRREGS:
968 return copy_regset_from_user(child, &user_ppc_native_view,
969 REGSET_VMX,
970 0, (33 * sizeof(vector128) +
971 sizeof(u32)),
972 (const void __user *) data);
973 #endif
974 #ifdef CONFIG_VSX
975 case PTRACE_GETVSRREGS:
976 return copy_regset_to_user(child, &user_ppc_native_view,
977 REGSET_VSX,
978 0, 32 * sizeof(double),
979 (void __user *) data);
981 case PTRACE_SETVSRREGS:
982 return copy_regset_from_user(child, &user_ppc_native_view,
983 REGSET_VSX,
984 0, 32 * sizeof(double),
985 (const void __user *) data);
986 #endif
987 #ifdef CONFIG_SPE
988 case PTRACE_GETEVRREGS:
989 /* Get the child spe register state. */
990 return copy_regset_to_user(child, &user_ppc_native_view,
991 REGSET_SPE, 0, 35 * sizeof(u32),
992 (void __user *) data);
994 case PTRACE_SETEVRREGS:
995 /* Set the child spe register state. */
996 return copy_regset_from_user(child, &user_ppc_native_view,
997 REGSET_SPE, 0, 35 * sizeof(u32),
998 (const void __user *) data);
999 #endif
1001 /* Old reverse args ptrace callss */
1002 case PPC_PTRACE_GETREGS: /* Get GPRs 0 - 31. */
1003 case PPC_PTRACE_SETREGS: /* Set GPRs 0 - 31. */
1004 case PPC_PTRACE_GETFPREGS: /* Get FPRs 0 - 31. */
1005 case PPC_PTRACE_SETFPREGS: /* Get FPRs 0 - 31. */
1006 ret = arch_ptrace_old(child, request, addr, data);
1007 break;
1009 default:
1010 ret = ptrace_request(child, request, addr, data);
1011 break;
1013 return ret;
1017 * We must return the syscall number to actually look up in the table.
1018 * This can be -1L to skip running any syscall at all.
1020 long do_syscall_trace_enter(struct pt_regs *regs)
1022 long ret = 0;
1024 secure_computing(regs->gpr[0]);
1026 if (test_thread_flag(TIF_SYSCALL_TRACE) &&
1027 tracehook_report_syscall_entry(regs))
1029 * Tracing decided this syscall should not happen.
1030 * We'll return a bogus call number to get an ENOSYS
1031 * error, but leave the original number in regs->gpr[0].
1033 ret = -1L;
1035 if (unlikely(current->audit_context)) {
1036 #ifdef CONFIG_PPC64
1037 if (!test_thread_flag(TIF_32BIT))
1038 audit_syscall_entry(AUDIT_ARCH_PPC64,
1039 regs->gpr[0],
1040 regs->gpr[3], regs->gpr[4],
1041 regs->gpr[5], regs->gpr[6]);
1042 else
1043 #endif
1044 audit_syscall_entry(AUDIT_ARCH_PPC,
1045 regs->gpr[0],
1046 regs->gpr[3] & 0xffffffff,
1047 regs->gpr[4] & 0xffffffff,
1048 regs->gpr[5] & 0xffffffff,
1049 regs->gpr[6] & 0xffffffff);
1052 return ret ?: regs->gpr[0];
1055 void do_syscall_trace_leave(struct pt_regs *regs)
1057 int step;
1059 if (unlikely(current->audit_context))
1060 audit_syscall_exit((regs->ccr&0x10000000)?AUDITSC_FAILURE:AUDITSC_SUCCESS,
1061 regs->result);
1063 step = test_thread_flag(TIF_SINGLESTEP);
1064 if (step || test_thread_flag(TIF_SYSCALL_TRACE))
1065 tracehook_report_syscall_exit(regs, step);