thermal: remove driver_data direct access of struct device
[linux-2.6/linux-acpi-2.6/ibm-acpi-2.6.git] / arch / powerpc / kernel / ptrace.c
blob9fa2c7dcd05a4501034395e6f587523cffb969d0
1 /*
2 * PowerPC version
3 * Copyright (C) 1995-1996 Gary Thomas (gdt@linuxppc.org)
5 * Derived from "arch/m68k/kernel/ptrace.c"
6 * Copyright (C) 1994 by Hamish Macdonald
7 * Taken from linux/kernel/ptrace.c and modified for M680x0.
8 * linux/kernel/ptrace.c is by Ross Biro 1/23/92, edited by Linus Torvalds
10 * Modified by Cort Dougan (cort@hq.fsmlabs.com)
11 * and Paul Mackerras (paulus@samba.org).
13 * This file is subject to the terms and conditions of the GNU General
14 * Public License. See the file README.legal in the main directory of
15 * this archive for more details.
18 #include <linux/kernel.h>
19 #include <linux/sched.h>
20 #include <linux/mm.h>
21 #include <linux/smp.h>
22 #include <linux/errno.h>
23 #include <linux/ptrace.h>
24 #include <linux/regset.h>
25 #include <linux/tracehook.h>
26 #include <linux/elf.h>
27 #include <linux/user.h>
28 #include <linux/security.h>
29 #include <linux/signal.h>
30 #include <linux/seccomp.h>
31 #include <linux/audit.h>
32 #ifdef CONFIG_PPC32
33 #include <linux/module.h>
34 #endif
36 #include <asm/uaccess.h>
37 #include <asm/page.h>
38 #include <asm/pgtable.h>
39 #include <asm/system.h>
42 * does not yet catch signals sent when the child dies.
43 * in exit.c or in signal.c.
47 * Set of msr bits that gdb can change on behalf of a process.
49 #if defined(CONFIG_40x) || defined(CONFIG_BOOKE)
50 #define MSR_DEBUGCHANGE 0
51 #else
52 #define MSR_DEBUGCHANGE (MSR_SE | MSR_BE)
53 #endif
56 * Max register writeable via put_reg
58 #ifdef CONFIG_PPC32
59 #define PT_MAX_PUT_REG PT_MQ
60 #else
61 #define PT_MAX_PUT_REG PT_CCR
62 #endif
64 static unsigned long get_user_msr(struct task_struct *task)
66 return task->thread.regs->msr | task->thread.fpexc_mode;
69 static int set_user_msr(struct task_struct *task, unsigned long msr)
71 task->thread.regs->msr &= ~MSR_DEBUGCHANGE;
72 task->thread.regs->msr |= msr & MSR_DEBUGCHANGE;
73 return 0;
77 * We prevent mucking around with the reserved area of trap
78 * which are used internally by the kernel.
80 static int set_user_trap(struct task_struct *task, unsigned long trap)
82 task->thread.regs->trap = trap & 0xfff0;
83 return 0;
87 * Get contents of register REGNO in task TASK.
89 unsigned long ptrace_get_reg(struct task_struct *task, int regno)
91 if (task->thread.regs == NULL)
92 return -EIO;
94 if (regno == PT_MSR)
95 return get_user_msr(task);
97 if (regno < (sizeof(struct pt_regs) / sizeof(unsigned long)))
98 return ((unsigned long *)task->thread.regs)[regno];
100 return -EIO;
104 * Write contents of register REGNO in task TASK.
106 int ptrace_put_reg(struct task_struct *task, int regno, unsigned long data)
108 if (task->thread.regs == NULL)
109 return -EIO;
111 if (regno == PT_MSR)
112 return set_user_msr(task, data);
113 if (regno == PT_TRAP)
114 return set_user_trap(task, data);
116 if (regno <= PT_MAX_PUT_REG) {
117 ((unsigned long *)task->thread.regs)[regno] = data;
118 return 0;
120 return -EIO;
123 static int gpr_get(struct task_struct *target, const struct user_regset *regset,
124 unsigned int pos, unsigned int count,
125 void *kbuf, void __user *ubuf)
127 int ret;
129 if (target->thread.regs == NULL)
130 return -EIO;
132 CHECK_FULL_REGS(target->thread.regs);
134 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
135 target->thread.regs,
136 0, offsetof(struct pt_regs, msr));
137 if (!ret) {
138 unsigned long msr = get_user_msr(target);
139 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf, &msr,
140 offsetof(struct pt_regs, msr),
141 offsetof(struct pt_regs, msr) +
142 sizeof(msr));
145 BUILD_BUG_ON(offsetof(struct pt_regs, orig_gpr3) !=
146 offsetof(struct pt_regs, msr) + sizeof(long));
148 if (!ret)
149 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
150 &target->thread.regs->orig_gpr3,
151 offsetof(struct pt_regs, orig_gpr3),
152 sizeof(struct pt_regs));
153 if (!ret)
154 ret = user_regset_copyout_zero(&pos, &count, &kbuf, &ubuf,
155 sizeof(struct pt_regs), -1);
157 return ret;
160 static int gpr_set(struct task_struct *target, const struct user_regset *regset,
161 unsigned int pos, unsigned int count,
162 const void *kbuf, const void __user *ubuf)
164 unsigned long reg;
165 int ret;
167 if (target->thread.regs == NULL)
168 return -EIO;
170 CHECK_FULL_REGS(target->thread.regs);
172 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
173 target->thread.regs,
174 0, PT_MSR * sizeof(reg));
176 if (!ret && count > 0) {
177 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf, &reg,
178 PT_MSR * sizeof(reg),
179 (PT_MSR + 1) * sizeof(reg));
180 if (!ret)
181 ret = set_user_msr(target, reg);
184 BUILD_BUG_ON(offsetof(struct pt_regs, orig_gpr3) !=
185 offsetof(struct pt_regs, msr) + sizeof(long));
187 if (!ret)
188 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
189 &target->thread.regs->orig_gpr3,
190 PT_ORIG_R3 * sizeof(reg),
191 (PT_MAX_PUT_REG + 1) * sizeof(reg));
193 if (PT_MAX_PUT_REG + 1 < PT_TRAP && !ret)
194 ret = user_regset_copyin_ignore(
195 &pos, &count, &kbuf, &ubuf,
196 (PT_MAX_PUT_REG + 1) * sizeof(reg),
197 PT_TRAP * sizeof(reg));
199 if (!ret && count > 0) {
200 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf, &reg,
201 PT_TRAP * sizeof(reg),
202 (PT_TRAP + 1) * sizeof(reg));
203 if (!ret)
204 ret = set_user_trap(target, reg);
207 if (!ret)
208 ret = user_regset_copyin_ignore(
209 &pos, &count, &kbuf, &ubuf,
210 (PT_TRAP + 1) * sizeof(reg), -1);
212 return ret;
215 static int fpr_get(struct task_struct *target, const struct user_regset *regset,
216 unsigned int pos, unsigned int count,
217 void *kbuf, void __user *ubuf)
219 #ifdef CONFIG_VSX
220 double buf[33];
221 int i;
222 #endif
223 flush_fp_to_thread(target);
225 #ifdef CONFIG_VSX
226 /* copy to local buffer then write that out */
227 for (i = 0; i < 32 ; i++)
228 buf[i] = target->thread.TS_FPR(i);
229 memcpy(&buf[32], &target->thread.fpscr, sizeof(double));
230 return user_regset_copyout(&pos, &count, &kbuf, &ubuf, buf, 0, -1);
232 #else
233 BUILD_BUG_ON(offsetof(struct thread_struct, fpscr) !=
234 offsetof(struct thread_struct, TS_FPR(32)));
236 return user_regset_copyout(&pos, &count, &kbuf, &ubuf,
237 &target->thread.fpr, 0, -1);
238 #endif
241 static int fpr_set(struct task_struct *target, const struct user_regset *regset,
242 unsigned int pos, unsigned int count,
243 const void *kbuf, const void __user *ubuf)
245 #ifdef CONFIG_VSX
246 double buf[33];
247 int i;
248 #endif
249 flush_fp_to_thread(target);
251 #ifdef CONFIG_VSX
252 /* copy to local buffer then write that out */
253 i = user_regset_copyin(&pos, &count, &kbuf, &ubuf, buf, 0, -1);
254 if (i)
255 return i;
256 for (i = 0; i < 32 ; i++)
257 target->thread.TS_FPR(i) = buf[i];
258 memcpy(&target->thread.fpscr, &buf[32], sizeof(double));
259 return 0;
260 #else
261 BUILD_BUG_ON(offsetof(struct thread_struct, fpscr) !=
262 offsetof(struct thread_struct, TS_FPR(32)));
264 return user_regset_copyin(&pos, &count, &kbuf, &ubuf,
265 &target->thread.fpr, 0, -1);
266 #endif
269 #ifdef CONFIG_ALTIVEC
271 * Get/set all the altivec registers vr0..vr31, vscr, vrsave, in one go.
272 * The transfer totals 34 quadword. Quadwords 0-31 contain the
273 * corresponding vector registers. Quadword 32 contains the vscr as the
274 * last word (offset 12) within that quadword. Quadword 33 contains the
275 * vrsave as the first word (offset 0) within the quadword.
277 * This definition of the VMX state is compatible with the current PPC32
278 * ptrace interface. This allows signal handling and ptrace to use the
279 * same structures. This also simplifies the implementation of a bi-arch
280 * (combined (32- and 64-bit) gdb.
283 static int vr_active(struct task_struct *target,
284 const struct user_regset *regset)
286 flush_altivec_to_thread(target);
287 return target->thread.used_vr ? regset->n : 0;
290 static int vr_get(struct task_struct *target, const struct user_regset *regset,
291 unsigned int pos, unsigned int count,
292 void *kbuf, void __user *ubuf)
294 int ret;
296 flush_altivec_to_thread(target);
298 BUILD_BUG_ON(offsetof(struct thread_struct, vscr) !=
299 offsetof(struct thread_struct, vr[32]));
301 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
302 &target->thread.vr, 0,
303 33 * sizeof(vector128));
304 if (!ret) {
306 * Copy out only the low-order word of vrsave.
308 union {
309 elf_vrreg_t reg;
310 u32 word;
311 } vrsave;
312 memset(&vrsave, 0, sizeof(vrsave));
313 vrsave.word = target->thread.vrsave;
314 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf, &vrsave,
315 33 * sizeof(vector128), -1);
318 return ret;
321 static int vr_set(struct task_struct *target, const struct user_regset *regset,
322 unsigned int pos, unsigned int count,
323 const void *kbuf, const void __user *ubuf)
325 int ret;
327 flush_altivec_to_thread(target);
329 BUILD_BUG_ON(offsetof(struct thread_struct, vscr) !=
330 offsetof(struct thread_struct, vr[32]));
332 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
333 &target->thread.vr, 0, 33 * sizeof(vector128));
334 if (!ret && count > 0) {
336 * We use only the first word of vrsave.
338 union {
339 elf_vrreg_t reg;
340 u32 word;
341 } vrsave;
342 memset(&vrsave, 0, sizeof(vrsave));
343 vrsave.word = target->thread.vrsave;
344 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf, &vrsave,
345 33 * sizeof(vector128), -1);
346 if (!ret)
347 target->thread.vrsave = vrsave.word;
350 return ret;
352 #endif /* CONFIG_ALTIVEC */
354 #ifdef CONFIG_VSX
356 * Currently to set and and get all the vsx state, you need to call
357 * the fp and VMX calls aswell. This only get/sets the lower 32
358 * 128bit VSX registers.
361 static int vsr_active(struct task_struct *target,
362 const struct user_regset *regset)
364 flush_vsx_to_thread(target);
365 return target->thread.used_vsr ? regset->n : 0;
368 static int vsr_get(struct task_struct *target, const struct user_regset *regset,
369 unsigned int pos, unsigned int count,
370 void *kbuf, void __user *ubuf)
372 double buf[32];
373 int ret, i;
375 flush_vsx_to_thread(target);
377 for (i = 0; i < 32 ; i++)
378 buf[i] = target->thread.fpr[i][TS_VSRLOWOFFSET];
379 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
380 buf, 0, 32 * sizeof(double));
382 return ret;
385 static int vsr_set(struct task_struct *target, const struct user_regset *regset,
386 unsigned int pos, unsigned int count,
387 const void *kbuf, const void __user *ubuf)
389 double buf[32];
390 int ret,i;
392 flush_vsx_to_thread(target);
394 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
395 buf, 0, 32 * sizeof(double));
396 for (i = 0; i < 32 ; i++)
397 target->thread.fpr[i][TS_VSRLOWOFFSET] = buf[i];
400 return ret;
402 #endif /* CONFIG_VSX */
404 #ifdef CONFIG_SPE
407 * For get_evrregs/set_evrregs functions 'data' has the following layout:
409 * struct {
410 * u32 evr[32];
411 * u64 acc;
412 * u32 spefscr;
416 static int evr_active(struct task_struct *target,
417 const struct user_regset *regset)
419 flush_spe_to_thread(target);
420 return target->thread.used_spe ? regset->n : 0;
423 static int evr_get(struct task_struct *target, const struct user_regset *regset,
424 unsigned int pos, unsigned int count,
425 void *kbuf, void __user *ubuf)
427 int ret;
429 flush_spe_to_thread(target);
431 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
432 &target->thread.evr,
433 0, sizeof(target->thread.evr));
435 BUILD_BUG_ON(offsetof(struct thread_struct, acc) + sizeof(u64) !=
436 offsetof(struct thread_struct, spefscr));
438 if (!ret)
439 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
440 &target->thread.acc,
441 sizeof(target->thread.evr), -1);
443 return ret;
446 static int evr_set(struct task_struct *target, const struct user_regset *regset,
447 unsigned int pos, unsigned int count,
448 const void *kbuf, const void __user *ubuf)
450 int ret;
452 flush_spe_to_thread(target);
454 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
455 &target->thread.evr,
456 0, sizeof(target->thread.evr));
458 BUILD_BUG_ON(offsetof(struct thread_struct, acc) + sizeof(u64) !=
459 offsetof(struct thread_struct, spefscr));
461 if (!ret)
462 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
463 &target->thread.acc,
464 sizeof(target->thread.evr), -1);
466 return ret;
468 #endif /* CONFIG_SPE */
472 * These are our native regset flavors.
474 enum powerpc_regset {
475 REGSET_GPR,
476 REGSET_FPR,
477 #ifdef CONFIG_ALTIVEC
478 REGSET_VMX,
479 #endif
480 #ifdef CONFIG_VSX
481 REGSET_VSX,
482 #endif
483 #ifdef CONFIG_SPE
484 REGSET_SPE,
485 #endif
488 static const struct user_regset native_regsets[] = {
489 [REGSET_GPR] = {
490 .core_note_type = NT_PRSTATUS, .n = ELF_NGREG,
491 .size = sizeof(long), .align = sizeof(long),
492 .get = gpr_get, .set = gpr_set
494 [REGSET_FPR] = {
495 .core_note_type = NT_PRFPREG, .n = ELF_NFPREG,
496 .size = sizeof(double), .align = sizeof(double),
497 .get = fpr_get, .set = fpr_set
499 #ifdef CONFIG_ALTIVEC
500 [REGSET_VMX] = {
501 .core_note_type = NT_PPC_VMX, .n = 34,
502 .size = sizeof(vector128), .align = sizeof(vector128),
503 .active = vr_active, .get = vr_get, .set = vr_set
505 #endif
506 #ifdef CONFIG_VSX
507 [REGSET_VSX] = {
508 .core_note_type = NT_PPC_VSX, .n = 32,
509 .size = sizeof(double), .align = sizeof(double),
510 .active = vsr_active, .get = vsr_get, .set = vsr_set
512 #endif
513 #ifdef CONFIG_SPE
514 [REGSET_SPE] = {
515 .n = 35,
516 .size = sizeof(u32), .align = sizeof(u32),
517 .active = evr_active, .get = evr_get, .set = evr_set
519 #endif
522 static const struct user_regset_view user_ppc_native_view = {
523 .name = UTS_MACHINE, .e_machine = ELF_ARCH, .ei_osabi = ELF_OSABI,
524 .regsets = native_regsets, .n = ARRAY_SIZE(native_regsets)
527 #ifdef CONFIG_PPC64
528 #include <linux/compat.h>
530 static int gpr32_get(struct task_struct *target,
531 const struct user_regset *regset,
532 unsigned int pos, unsigned int count,
533 void *kbuf, void __user *ubuf)
535 const unsigned long *regs = &target->thread.regs->gpr[0];
536 compat_ulong_t *k = kbuf;
537 compat_ulong_t __user *u = ubuf;
538 compat_ulong_t reg;
540 if (target->thread.regs == NULL)
541 return -EIO;
543 CHECK_FULL_REGS(target->thread.regs);
545 pos /= sizeof(reg);
546 count /= sizeof(reg);
548 if (kbuf)
549 for (; count > 0 && pos < PT_MSR; --count)
550 *k++ = regs[pos++];
551 else
552 for (; count > 0 && pos < PT_MSR; --count)
553 if (__put_user((compat_ulong_t) regs[pos++], u++))
554 return -EFAULT;
556 if (count > 0 && pos == PT_MSR) {
557 reg = get_user_msr(target);
558 if (kbuf)
559 *k++ = reg;
560 else if (__put_user(reg, u++))
561 return -EFAULT;
562 ++pos;
563 --count;
566 if (kbuf)
567 for (; count > 0 && pos < PT_REGS_COUNT; --count)
568 *k++ = regs[pos++];
569 else
570 for (; count > 0 && pos < PT_REGS_COUNT; --count)
571 if (__put_user((compat_ulong_t) regs[pos++], u++))
572 return -EFAULT;
574 kbuf = k;
575 ubuf = u;
576 pos *= sizeof(reg);
577 count *= sizeof(reg);
578 return user_regset_copyout_zero(&pos, &count, &kbuf, &ubuf,
579 PT_REGS_COUNT * sizeof(reg), -1);
582 static int gpr32_set(struct task_struct *target,
583 const struct user_regset *regset,
584 unsigned int pos, unsigned int count,
585 const void *kbuf, const void __user *ubuf)
587 unsigned long *regs = &target->thread.regs->gpr[0];
588 const compat_ulong_t *k = kbuf;
589 const compat_ulong_t __user *u = ubuf;
590 compat_ulong_t reg;
592 if (target->thread.regs == NULL)
593 return -EIO;
595 CHECK_FULL_REGS(target->thread.regs);
597 pos /= sizeof(reg);
598 count /= sizeof(reg);
600 if (kbuf)
601 for (; count > 0 && pos < PT_MSR; --count)
602 regs[pos++] = *k++;
603 else
604 for (; count > 0 && pos < PT_MSR; --count) {
605 if (__get_user(reg, u++))
606 return -EFAULT;
607 regs[pos++] = reg;
611 if (count > 0 && pos == PT_MSR) {
612 if (kbuf)
613 reg = *k++;
614 else if (__get_user(reg, u++))
615 return -EFAULT;
616 set_user_msr(target, reg);
617 ++pos;
618 --count;
621 if (kbuf) {
622 for (; count > 0 && pos <= PT_MAX_PUT_REG; --count)
623 regs[pos++] = *k++;
624 for (; count > 0 && pos < PT_TRAP; --count, ++pos)
625 ++k;
626 } else {
627 for (; count > 0 && pos <= PT_MAX_PUT_REG; --count) {
628 if (__get_user(reg, u++))
629 return -EFAULT;
630 regs[pos++] = reg;
632 for (; count > 0 && pos < PT_TRAP; --count, ++pos)
633 if (__get_user(reg, u++))
634 return -EFAULT;
637 if (count > 0 && pos == PT_TRAP) {
638 if (kbuf)
639 reg = *k++;
640 else if (__get_user(reg, u++))
641 return -EFAULT;
642 set_user_trap(target, reg);
643 ++pos;
644 --count;
647 kbuf = k;
648 ubuf = u;
649 pos *= sizeof(reg);
650 count *= sizeof(reg);
651 return user_regset_copyin_ignore(&pos, &count, &kbuf, &ubuf,
652 (PT_TRAP + 1) * sizeof(reg), -1);
656 * These are the regset flavors matching the CONFIG_PPC32 native set.
658 static const struct user_regset compat_regsets[] = {
659 [REGSET_GPR] = {
660 .core_note_type = NT_PRSTATUS, .n = ELF_NGREG,
661 .size = sizeof(compat_long_t), .align = sizeof(compat_long_t),
662 .get = gpr32_get, .set = gpr32_set
664 [REGSET_FPR] = {
665 .core_note_type = NT_PRFPREG, .n = ELF_NFPREG,
666 .size = sizeof(double), .align = sizeof(double),
667 .get = fpr_get, .set = fpr_set
669 #ifdef CONFIG_ALTIVEC
670 [REGSET_VMX] = {
671 .core_note_type = NT_PPC_VMX, .n = 34,
672 .size = sizeof(vector128), .align = sizeof(vector128),
673 .active = vr_active, .get = vr_get, .set = vr_set
675 #endif
676 #ifdef CONFIG_SPE
677 [REGSET_SPE] = {
678 .core_note_type = NT_PPC_SPE, .n = 35,
679 .size = sizeof(u32), .align = sizeof(u32),
680 .active = evr_active, .get = evr_get, .set = evr_set
682 #endif
685 static const struct user_regset_view user_ppc_compat_view = {
686 .name = "ppc", .e_machine = EM_PPC, .ei_osabi = ELF_OSABI,
687 .regsets = compat_regsets, .n = ARRAY_SIZE(compat_regsets)
689 #endif /* CONFIG_PPC64 */
691 const struct user_regset_view *task_user_regset_view(struct task_struct *task)
693 #ifdef CONFIG_PPC64
694 if (test_tsk_thread_flag(task, TIF_32BIT))
695 return &user_ppc_compat_view;
696 #endif
697 return &user_ppc_native_view;
701 void user_enable_single_step(struct task_struct *task)
703 struct pt_regs *regs = task->thread.regs;
705 if (regs != NULL) {
706 #if defined(CONFIG_40x) || defined(CONFIG_BOOKE)
707 task->thread.dbcr0 &= ~DBCR0_BT;
708 task->thread.dbcr0 |= DBCR0_IDM | DBCR0_IC;
709 regs->msr |= MSR_DE;
710 #else
711 regs->msr &= ~MSR_BE;
712 regs->msr |= MSR_SE;
713 #endif
715 set_tsk_thread_flag(task, TIF_SINGLESTEP);
718 void user_enable_block_step(struct task_struct *task)
720 struct pt_regs *regs = task->thread.regs;
722 if (regs != NULL) {
723 #if defined(CONFIG_40x) || defined(CONFIG_BOOKE)
724 task->thread.dbcr0 &= ~DBCR0_IC;
725 task->thread.dbcr0 = DBCR0_IDM | DBCR0_BT;
726 regs->msr |= MSR_DE;
727 #else
728 regs->msr &= ~MSR_SE;
729 regs->msr |= MSR_BE;
730 #endif
732 set_tsk_thread_flag(task, TIF_SINGLESTEP);
735 void user_disable_single_step(struct task_struct *task)
737 struct pt_regs *regs = task->thread.regs;
740 #if defined(CONFIG_BOOKE)
741 /* If DAC then do not single step, skip */
742 if (task->thread.dabr)
743 return;
744 #endif
746 if (regs != NULL) {
747 #if defined(CONFIG_40x) || defined(CONFIG_BOOKE)
748 task->thread.dbcr0 &= ~(DBCR0_IC | DBCR0_BT | DBCR0_IDM);
749 regs->msr &= ~MSR_DE;
750 #else
751 regs->msr &= ~(MSR_SE | MSR_BE);
752 #endif
754 clear_tsk_thread_flag(task, TIF_SINGLESTEP);
757 int ptrace_set_debugreg(struct task_struct *task, unsigned long addr,
758 unsigned long data)
760 /* For ppc64 we support one DABR and no IABR's at the moment (ppc64).
761 * For embedded processors we support one DAC and no IAC's at the
762 * moment.
764 if (addr > 0)
765 return -EINVAL;
767 /* The bottom 3 bits in dabr are flags */
768 if ((data & ~0x7UL) >= TASK_SIZE)
769 return -EIO;
771 #ifndef CONFIG_BOOKE
773 /* For processors using DABR (i.e. 970), the bottom 3 bits are flags.
774 * It was assumed, on previous implementations, that 3 bits were
775 * passed together with the data address, fitting the design of the
776 * DABR register, as follows:
778 * bit 0: Read flag
779 * bit 1: Write flag
780 * bit 2: Breakpoint translation
782 * Thus, we use them here as so.
785 /* Ensure breakpoint translation bit is set */
786 if (data && !(data & DABR_TRANSLATION))
787 return -EIO;
789 /* Move contents to the DABR register */
790 task->thread.dabr = data;
792 #endif
793 #if defined(CONFIG_BOOKE)
795 /* As described above, it was assumed 3 bits were passed with the data
796 * address, but we will assume only the mode bits will be passed
797 * as to not cause alignment restrictions for DAC-based processors.
800 /* DAC's hold the whole address without any mode flags */
801 task->thread.dabr = data & ~0x3UL;
803 if (task->thread.dabr == 0) {
804 task->thread.dbcr0 &= ~(DBSR_DAC1R | DBSR_DAC1W | DBCR0_IDM);
805 task->thread.regs->msr &= ~MSR_DE;
806 return 0;
809 /* Read or Write bits must be set */
811 if (!(data & 0x3UL))
812 return -EINVAL;
814 /* Set the Internal Debugging flag (IDM bit 1) for the DBCR0
815 register */
816 task->thread.dbcr0 = DBCR0_IDM;
818 /* Check for write and read flags and set DBCR0
819 accordingly */
820 if (data & 0x1UL)
821 task->thread.dbcr0 |= DBSR_DAC1R;
822 if (data & 0x2UL)
823 task->thread.dbcr0 |= DBSR_DAC1W;
825 task->thread.regs->msr |= MSR_DE;
826 #endif
827 return 0;
831 * Called by kernel/ptrace.c when detaching..
833 * Make sure single step bits etc are not set.
835 void ptrace_disable(struct task_struct *child)
837 /* make sure the single step bit is not set. */
838 user_disable_single_step(child);
842 * Here are the old "legacy" powerpc specific getregs/setregs ptrace calls,
843 * we mark them as obsolete now, they will be removed in a future version
845 static long arch_ptrace_old(struct task_struct *child, long request, long addr,
846 long data)
848 switch (request) {
849 case PPC_PTRACE_GETREGS: /* Get GPRs 0 - 31. */
850 return copy_regset_to_user(child, &user_ppc_native_view,
851 REGSET_GPR, 0, 32 * sizeof(long),
852 (void __user *) data);
854 case PPC_PTRACE_SETREGS: /* Set GPRs 0 - 31. */
855 return copy_regset_from_user(child, &user_ppc_native_view,
856 REGSET_GPR, 0, 32 * sizeof(long),
857 (const void __user *) data);
859 case PPC_PTRACE_GETFPREGS: /* Get FPRs 0 - 31. */
860 return copy_regset_to_user(child, &user_ppc_native_view,
861 REGSET_FPR, 0, 32 * sizeof(double),
862 (void __user *) data);
864 case PPC_PTRACE_SETFPREGS: /* Set FPRs 0 - 31. */
865 return copy_regset_from_user(child, &user_ppc_native_view,
866 REGSET_FPR, 0, 32 * sizeof(double),
867 (const void __user *) data);
870 return -EPERM;
873 long arch_ptrace(struct task_struct *child, long request, long addr, long data)
875 int ret = -EPERM;
877 switch (request) {
878 /* read the word at location addr in the USER area. */
879 case PTRACE_PEEKUSR: {
880 unsigned long index, tmp;
882 ret = -EIO;
883 /* convert to index and check */
884 #ifdef CONFIG_PPC32
885 index = (unsigned long) addr >> 2;
886 if ((addr & 3) || (index > PT_FPSCR)
887 || (child->thread.regs == NULL))
888 #else
889 index = (unsigned long) addr >> 3;
890 if ((addr & 7) || (index > PT_FPSCR))
891 #endif
892 break;
894 CHECK_FULL_REGS(child->thread.regs);
895 if (index < PT_FPR0) {
896 tmp = ptrace_get_reg(child, (int) index);
897 } else {
898 flush_fp_to_thread(child);
899 tmp = ((unsigned long *)child->thread.fpr)
900 [TS_FPRWIDTH * (index - PT_FPR0)];
902 ret = put_user(tmp,(unsigned long __user *) data);
903 break;
906 /* write the word at location addr in the USER area */
907 case PTRACE_POKEUSR: {
908 unsigned long index;
910 ret = -EIO;
911 /* convert to index and check */
912 #ifdef CONFIG_PPC32
913 index = (unsigned long) addr >> 2;
914 if ((addr & 3) || (index > PT_FPSCR)
915 || (child->thread.regs == NULL))
916 #else
917 index = (unsigned long) addr >> 3;
918 if ((addr & 7) || (index > PT_FPSCR))
919 #endif
920 break;
922 CHECK_FULL_REGS(child->thread.regs);
923 if (index < PT_FPR0) {
924 ret = ptrace_put_reg(child, index, data);
925 } else {
926 flush_fp_to_thread(child);
927 ((unsigned long *)child->thread.fpr)
928 [TS_FPRWIDTH * (index - PT_FPR0)] = data;
929 ret = 0;
931 break;
934 case PTRACE_GET_DEBUGREG: {
935 ret = -EINVAL;
936 /* We only support one DABR and no IABRS at the moment */
937 if (addr > 0)
938 break;
939 ret = put_user(child->thread.dabr,
940 (unsigned long __user *)data);
941 break;
944 case PTRACE_SET_DEBUGREG:
945 ret = ptrace_set_debugreg(child, addr, data);
946 break;
948 #ifdef CONFIG_PPC64
949 case PTRACE_GETREGS64:
950 #endif
951 case PTRACE_GETREGS: /* Get all pt_regs from the child. */
952 return copy_regset_to_user(child, &user_ppc_native_view,
953 REGSET_GPR,
954 0, sizeof(struct pt_regs),
955 (void __user *) data);
957 #ifdef CONFIG_PPC64
958 case PTRACE_SETREGS64:
959 #endif
960 case PTRACE_SETREGS: /* Set all gp regs in the child. */
961 return copy_regset_from_user(child, &user_ppc_native_view,
962 REGSET_GPR,
963 0, sizeof(struct pt_regs),
964 (const void __user *) data);
966 case PTRACE_GETFPREGS: /* Get the child FPU state (FPR0...31 + FPSCR) */
967 return copy_regset_to_user(child, &user_ppc_native_view,
968 REGSET_FPR,
969 0, sizeof(elf_fpregset_t),
970 (void __user *) data);
972 case PTRACE_SETFPREGS: /* Set the child FPU state (FPR0...31 + FPSCR) */
973 return copy_regset_from_user(child, &user_ppc_native_view,
974 REGSET_FPR,
975 0, sizeof(elf_fpregset_t),
976 (const void __user *) data);
978 #ifdef CONFIG_ALTIVEC
979 case PTRACE_GETVRREGS:
980 return copy_regset_to_user(child, &user_ppc_native_view,
981 REGSET_VMX,
982 0, (33 * sizeof(vector128) +
983 sizeof(u32)),
984 (void __user *) data);
986 case PTRACE_SETVRREGS:
987 return copy_regset_from_user(child, &user_ppc_native_view,
988 REGSET_VMX,
989 0, (33 * sizeof(vector128) +
990 sizeof(u32)),
991 (const void __user *) data);
992 #endif
993 #ifdef CONFIG_VSX
994 case PTRACE_GETVSRREGS:
995 return copy_regset_to_user(child, &user_ppc_native_view,
996 REGSET_VSX,
997 0, 32 * sizeof(double),
998 (void __user *) data);
1000 case PTRACE_SETVSRREGS:
1001 return copy_regset_from_user(child, &user_ppc_native_view,
1002 REGSET_VSX,
1003 0, 32 * sizeof(double),
1004 (const void __user *) data);
1005 #endif
1006 #ifdef CONFIG_SPE
1007 case PTRACE_GETEVRREGS:
1008 /* Get the child spe register state. */
1009 return copy_regset_to_user(child, &user_ppc_native_view,
1010 REGSET_SPE, 0, 35 * sizeof(u32),
1011 (void __user *) data);
1013 case PTRACE_SETEVRREGS:
1014 /* Set the child spe register state. */
1015 return copy_regset_from_user(child, &user_ppc_native_view,
1016 REGSET_SPE, 0, 35 * sizeof(u32),
1017 (const void __user *) data);
1018 #endif
1020 /* Old reverse args ptrace callss */
1021 case PPC_PTRACE_GETREGS: /* Get GPRs 0 - 31. */
1022 case PPC_PTRACE_SETREGS: /* Set GPRs 0 - 31. */
1023 case PPC_PTRACE_GETFPREGS: /* Get FPRs 0 - 31. */
1024 case PPC_PTRACE_SETFPREGS: /* Get FPRs 0 - 31. */
1025 ret = arch_ptrace_old(child, request, addr, data);
1026 break;
1028 default:
1029 ret = ptrace_request(child, request, addr, data);
1030 break;
1032 return ret;
1036 * We must return the syscall number to actually look up in the table.
1037 * This can be -1L to skip running any syscall at all.
1039 long do_syscall_trace_enter(struct pt_regs *regs)
1041 long ret = 0;
1043 secure_computing(regs->gpr[0]);
1045 if (test_thread_flag(TIF_SYSCALL_TRACE) &&
1046 tracehook_report_syscall_entry(regs))
1048 * Tracing decided this syscall should not happen.
1049 * We'll return a bogus call number to get an ENOSYS
1050 * error, but leave the original number in regs->gpr[0].
1052 ret = -1L;
1054 if (unlikely(current->audit_context)) {
1055 #ifdef CONFIG_PPC64
1056 if (!test_thread_flag(TIF_32BIT))
1057 audit_syscall_entry(AUDIT_ARCH_PPC64,
1058 regs->gpr[0],
1059 regs->gpr[3], regs->gpr[4],
1060 regs->gpr[5], regs->gpr[6]);
1061 else
1062 #endif
1063 audit_syscall_entry(AUDIT_ARCH_PPC,
1064 regs->gpr[0],
1065 regs->gpr[3] & 0xffffffff,
1066 regs->gpr[4] & 0xffffffff,
1067 regs->gpr[5] & 0xffffffff,
1068 regs->gpr[6] & 0xffffffff);
1071 return ret ?: regs->gpr[0];
1074 void do_syscall_trace_leave(struct pt_regs *regs)
1076 int step;
1078 if (unlikely(current->audit_context))
1079 audit_syscall_exit((regs->ccr&0x10000000)?AUDITSC_FAILURE:AUDITSC_SUCCESS,
1080 regs->result);
1082 step = test_thread_flag(TIF_SINGLESTEP);
1083 if (step || test_thread_flag(TIF_SYSCALL_TRACE))
1084 tracehook_report_syscall_exit(regs, step);