2 * Copyright (C) 1998-2004 Hewlett-Packard Co
3 * David Mosberger-Tang <davidm@hpl.hp.com>
4 * Stephane Eranian <eranian@hpl.hp.com>
5 * Copyright (C) 2003 Intel Co
6 * Suresh Siddha <suresh.b.siddha@intel.com>
7 * Fenghua Yu <fenghua.yu@intel.com>
8 * Arun Sharma <arun.sharma@intel.com>
10 * 12/07/98 S. Eranian added pt_regs & switch_stack
11 * 12/21/98 D. Mosberger updated to match latest code
12 * 6/17/99 D. Mosberger added second unat member to "struct switch_stack"
15 #ifndef _ASM_IA64_PTRACE_H
16 #define _ASM_IA64_PTRACE_H
19 #include <asm/asm-offsets.h>
21 #include <uapi/asm/ptrace.h>
24 * Base-2 logarithm of number of pages to allocate per task structure
25 * (including register backing store and memory stack):
27 #if defined(CONFIG_IA64_PAGE_SIZE_4KB)
28 # define KERNEL_STACK_SIZE_ORDER 3
29 #elif defined(CONFIG_IA64_PAGE_SIZE_8KB)
30 # define KERNEL_STACK_SIZE_ORDER 2
31 #elif defined(CONFIG_IA64_PAGE_SIZE_16KB)
32 # define KERNEL_STACK_SIZE_ORDER 1
34 # define KERNEL_STACK_SIZE_ORDER 0
37 #define IA64_RBS_OFFSET ((IA64_TASK_SIZE + IA64_THREAD_INFO_SIZE + 31) & ~31)
38 #define IA64_STK_OFFSET ((1 << KERNEL_STACK_SIZE_ORDER)*PAGE_SIZE)
40 #define KERNEL_STACK_SIZE IA64_STK_OFFSET
44 #include <asm/current.h>
48 * We use the ia64_psr(regs)->ri to determine which of the three
49 * instructions in bundle (16 bytes) took the sample. Generate
50 * the canonical representation by adding to instruction pointer.
52 # define instruction_pointer(regs) ((regs)->cr_iip + ia64_psr(regs)->ri)
54 static inline unsigned long user_stack_pointer(struct pt_regs
*regs
)
56 /* FIXME: should this be bspstore + nr_dirty regs? */
57 return regs
->ar_bspstore
;
60 static inline int is_syscall_success(struct pt_regs
*regs
)
62 return regs
->r10
!= -1;
65 static inline long regs_return_value(struct pt_regs
*regs
)
67 if (is_syscall_success(regs
))
73 /* Conserve space in histogram by encoding slot bits in address
74 * bits 2 and 3 rather than bits 0 and 1.
76 #define profile_pc(regs) \
78 unsigned long __ip = instruction_pointer(regs); \
79 (__ip & ~3UL) + ((__ip & 3UL) << 2); \
82 /* given a pointer to a task_struct, return the user's pt_regs */
83 # define task_pt_regs(t) (((struct pt_regs *) ((char *) (t) + IA64_STK_OFFSET)) - 1)
84 # define ia64_psr(regs) ((struct ia64_psr *) &(regs)->cr_ipsr)
85 # define user_mode(regs) (((struct ia64_psr *) &(regs)->cr_ipsr)->cpl != 0)
86 # define user_stack(task,regs) ((long) regs - (long) task == IA64_STK_OFFSET - sizeof(*regs))
87 # define fsys_mode(task,regs) \
89 struct task_struct *_task = (task); \
90 struct pt_regs *_regs = (regs); \
91 !user_mode(_regs) && user_stack(_task, _regs); \
95 * System call handlers that, upon successful completion, need to return a negative value
96 * should call force_successful_syscall_return() right before returning. On architectures
97 * where the syscall convention provides for a separate error flag (e.g., alpha, ia64,
98 * ppc{,64}, sparc{,64}, possibly others), this macro can be used to ensure that the error
99 * flag will not get set. On architectures which do not support a separate error flag,
100 * the macro is a no-op and the spurious error condition needs to be filtered out by some
101 * other means (e.g., in user-level, by passing an extra argument to the syscall handler,
102 * or something along those lines).
104 * On ia64, we can clear the user's pt_regs->r8 to force a successful syscall.
106 # define force_successful_syscall_return() (task_pt_regs(current)->r8 = 0)
108 struct task_struct
; /* forward decl */
109 struct unw_frame_info
; /* forward decl */
111 extern void ia64_do_show_stack (struct unw_frame_info
*, void *);
112 extern unsigned long ia64_get_user_rbs_end (struct task_struct
*, struct pt_regs
*,
114 extern long ia64_peek (struct task_struct
*, struct switch_stack
*, unsigned long,
115 unsigned long, long *);
116 extern long ia64_poke (struct task_struct
*, struct switch_stack
*, unsigned long,
117 unsigned long, long);
118 extern void ia64_flush_fph (struct task_struct
*);
119 extern void ia64_sync_fph (struct task_struct
*);
120 extern void ia64_sync_krbs(void);
121 extern long ia64_sync_user_rbs (struct task_struct
*, struct switch_stack
*,
122 unsigned long, unsigned long);
124 /* get nat bits for scratch registers such that bit N==1 iff scratch register rN is a NaT */
125 extern unsigned long ia64_get_scratch_nat_bits (struct pt_regs
*pt
, unsigned long scratch_unat
);
126 /* put nat bits for scratch registers such that scratch register rN is a NaT iff bit N==1 */
127 extern unsigned long ia64_put_scratch_nat_bits (struct pt_regs
*pt
, unsigned long nat
);
129 extern void ia64_increment_ip (struct pt_regs
*pt
);
130 extern void ia64_decrement_ip (struct pt_regs
*pt
);
132 extern void ia64_ptrace_stop(void);
133 #define arch_ptrace_stop(code, info) \
135 #define arch_ptrace_stop_needed(code, info) \
136 (!test_thread_flag(TIF_RESTORE_RSE))
138 extern void ptrace_attach_sync_user_rbs (struct task_struct
*);
139 #define arch_ptrace_attach(child) \
140 ptrace_attach_sync_user_rbs(child)
142 #define arch_has_single_step() (1)
143 #define arch_has_block_step() (1)
145 #endif /* !__ASSEMBLY__ */
146 #endif /* _ASM_IA64_PTRACE_H */