i386: move hyperv_vendor_id initialization to x86_cpu_realizefn()
[qemu/ar7.git] / target / arm / debug_helper.c
blob2ff72d47d195d03b7f498e521efc379bbba86e6c
1 /*
2 * ARM debug helpers.
4 * This code is licensed under the GNU GPL v2 or later.
6 * SPDX-License-Identifier: GPL-2.0-or-later
7 */
8 #include "qemu/osdep.h"
9 #include "cpu.h"
10 #include "internals.h"
11 #include "exec/exec-all.h"
12 #include "exec/helper-proto.h"
14 /* Return true if the linked breakpoint entry lbn passes its checks */
15 static bool linked_bp_matches(ARMCPU *cpu, int lbn)
17 CPUARMState *env = &cpu->env;
18 uint64_t bcr = env->cp15.dbgbcr[lbn];
19 int brps = arm_num_brps(cpu);
20 int ctx_cmps = arm_num_ctx_cmps(cpu);
21 int bt;
22 uint32_t contextidr;
23 uint64_t hcr_el2;
26 * Links to unimplemented or non-context aware breakpoints are
27 * CONSTRAINED UNPREDICTABLE: either behave as if disabled, or
28 * as if linked to an UNKNOWN context-aware breakpoint (in which
29 * case DBGWCR<n>_EL1.LBN must indicate that breakpoint).
30 * We choose the former.
32 if (lbn >= brps || lbn < (brps - ctx_cmps)) {
33 return false;
36 bcr = env->cp15.dbgbcr[lbn];
38 if (extract64(bcr, 0, 1) == 0) {
39 /* Linked breakpoint disabled : generate no events */
40 return false;
43 bt = extract64(bcr, 20, 4);
44 hcr_el2 = arm_hcr_el2_eff(env);
46 switch (bt) {
47 case 3: /* linked context ID match */
48 switch (arm_current_el(env)) {
49 default:
50 /* Context matches never fire in AArch64 EL3 */
51 return false;
52 case 2:
53 if (!(hcr_el2 & HCR_E2H)) {
54 /* Context matches never fire in EL2 without E2H enabled. */
55 return false;
57 contextidr = env->cp15.contextidr_el[2];
58 break;
59 case 1:
60 contextidr = env->cp15.contextidr_el[1];
61 break;
62 case 0:
63 if ((hcr_el2 & (HCR_E2H | HCR_TGE)) == (HCR_E2H | HCR_TGE)) {
64 contextidr = env->cp15.contextidr_el[2];
65 } else {
66 contextidr = env->cp15.contextidr_el[1];
68 break;
70 break;
72 case 7: /* linked contextidr_el1 match */
73 contextidr = env->cp15.contextidr_el[1];
74 break;
75 case 13: /* linked contextidr_el2 match */
76 contextidr = env->cp15.contextidr_el[2];
77 break;
79 case 9: /* linked VMID match (reserved if no EL2) */
80 case 11: /* linked context ID and VMID match (reserved if no EL2) */
81 case 15: /* linked full context ID match */
82 default:
84 * Links to Unlinked context breakpoints must generate no
85 * events; we choose to do the same for reserved values too.
87 return false;
91 * We match the whole register even if this is AArch32 using the
92 * short descriptor format (in which case it holds both PROCID and ASID),
93 * since we don't implement the optional v7 context ID masking.
95 return contextidr == (uint32_t)env->cp15.dbgbvr[lbn];
98 static bool bp_wp_matches(ARMCPU *cpu, int n, bool is_wp)
100 CPUARMState *env = &cpu->env;
101 uint64_t cr;
102 int pac, hmc, ssc, wt, lbn;
104 * Note that for watchpoints the check is against the CPU security
105 * state, not the S/NS attribute on the offending data access.
107 bool is_secure = arm_is_secure(env);
108 int access_el = arm_current_el(env);
110 if (is_wp) {
111 CPUWatchpoint *wp = env->cpu_watchpoint[n];
113 if (!wp || !(wp->flags & BP_WATCHPOINT_HIT)) {
114 return false;
116 cr = env->cp15.dbgwcr[n];
117 if (wp->hitattrs.user) {
119 * The LDRT/STRT/LDT/STT "unprivileged access" instructions should
120 * match watchpoints as if they were accesses done at EL0, even if
121 * the CPU is at EL1 or higher.
123 access_el = 0;
125 } else {
126 uint64_t pc = is_a64(env) ? env->pc : env->regs[15];
128 if (!env->cpu_breakpoint[n] || env->cpu_breakpoint[n]->pc != pc) {
129 return false;
131 cr = env->cp15.dbgbcr[n];
134 * The WATCHPOINT_HIT flag guarantees us that the watchpoint is
135 * enabled and that the address and access type match; for breakpoints
136 * we know the address matched; check the remaining fields, including
137 * linked breakpoints. We rely on WCR and BCR having the same layout
138 * for the LBN, SSC, HMC, PAC/PMC and is-linked fields.
139 * Note that some combinations of {PAC, HMC, SSC} are reserved and
140 * must act either like some valid combination or as if the watchpoint
141 * were disabled. We choose the former, and use this together with
142 * the fact that EL3 must always be Secure and EL2 must always be
143 * Non-Secure to simplify the code slightly compared to the full
144 * table in the ARM ARM.
146 pac = extract64(cr, 1, 2);
147 hmc = extract64(cr, 13, 1);
148 ssc = extract64(cr, 14, 2);
150 switch (ssc) {
151 case 0:
152 break;
153 case 1:
154 case 3:
155 if (is_secure) {
156 return false;
158 break;
159 case 2:
160 if (!is_secure) {
161 return false;
163 break;
166 switch (access_el) {
167 case 3:
168 case 2:
169 if (!hmc) {
170 return false;
172 break;
173 case 1:
174 if (extract32(pac, 0, 1) == 0) {
175 return false;
177 break;
178 case 0:
179 if (extract32(pac, 1, 1) == 0) {
180 return false;
182 break;
183 default:
184 g_assert_not_reached();
187 wt = extract64(cr, 20, 1);
188 lbn = extract64(cr, 16, 4);
190 if (wt && !linked_bp_matches(cpu, lbn)) {
191 return false;
194 return true;
197 static bool check_watchpoints(ARMCPU *cpu)
199 CPUARMState *env = &cpu->env;
200 int n;
203 * If watchpoints are disabled globally or we can't take debug
204 * exceptions here then watchpoint firings are ignored.
206 if (extract32(env->cp15.mdscr_el1, 15, 1) == 0
207 || !arm_generate_debug_exceptions(env)) {
208 return false;
211 for (n = 0; n < ARRAY_SIZE(env->cpu_watchpoint); n++) {
212 if (bp_wp_matches(cpu, n, true)) {
213 return true;
216 return false;
219 static bool check_breakpoints(ARMCPU *cpu)
221 CPUARMState *env = &cpu->env;
222 int n;
225 * If breakpoints are disabled globally or we can't take debug
226 * exceptions here then breakpoint firings are ignored.
228 if (extract32(env->cp15.mdscr_el1, 15, 1) == 0
229 || !arm_generate_debug_exceptions(env)) {
230 return false;
233 for (n = 0; n < ARRAY_SIZE(env->cpu_breakpoint); n++) {
234 if (bp_wp_matches(cpu, n, false)) {
235 return true;
238 return false;
241 void HELPER(check_breakpoints)(CPUARMState *env)
243 ARMCPU *cpu = env_archcpu(env);
245 if (check_breakpoints(cpu)) {
246 HELPER(exception_internal(env, EXCP_DEBUG));
250 bool arm_debug_check_watchpoint(CPUState *cs, CPUWatchpoint *wp)
253 * Called by core code when a CPU watchpoint fires; need to check if this
254 * is also an architectural watchpoint match.
256 ARMCPU *cpu = ARM_CPU(cs);
258 return check_watchpoints(cpu);
261 void arm_debug_excp_handler(CPUState *cs)
264 * Called by core code when a watchpoint or breakpoint fires;
265 * need to check which one and raise the appropriate exception.
267 ARMCPU *cpu = ARM_CPU(cs);
268 CPUARMState *env = &cpu->env;
269 CPUWatchpoint *wp_hit = cs->watchpoint_hit;
271 if (wp_hit) {
272 if (wp_hit->flags & BP_CPU) {
273 bool wnr = (wp_hit->flags & BP_WATCHPOINT_HIT_WRITE) != 0;
274 bool same_el = arm_debug_target_el(env) == arm_current_el(env);
276 cs->watchpoint_hit = NULL;
278 env->exception.fsr = arm_debug_exception_fsr(env);
279 env->exception.vaddress = wp_hit->hitaddr;
280 raise_exception(env, EXCP_DATA_ABORT,
281 syn_watchpoint(same_el, 0, wnr),
282 arm_debug_target_el(env));
284 } else {
285 uint64_t pc = is_a64(env) ? env->pc : env->regs[15];
286 bool same_el = (arm_debug_target_el(env) == arm_current_el(env));
289 * (1) GDB breakpoints should be handled first.
290 * (2) Do not raise a CPU exception if no CPU breakpoint has fired,
291 * since singlestep is also done by generating a debug internal
292 * exception.
294 if (cpu_breakpoint_test(cs, pc, BP_GDB)
295 || !cpu_breakpoint_test(cs, pc, BP_CPU)) {
296 return;
299 env->exception.fsr = arm_debug_exception_fsr(env);
301 * FAR is UNKNOWN: clear vaddress to avoid potentially exposing
302 * values to the guest that it shouldn't be able to see at its
303 * exception/security level.
305 env->exception.vaddress = 0;
306 raise_exception(env, EXCP_PREFETCH_ABORT,
307 syn_breakpoint(same_el),
308 arm_debug_target_el(env));
312 #if !defined(CONFIG_USER_ONLY)
314 vaddr arm_adjust_watchpoint_address(CPUState *cs, vaddr addr, int len)
316 ARMCPU *cpu = ARM_CPU(cs);
317 CPUARMState *env = &cpu->env;
320 * In BE32 system mode, target memory is stored byteswapped (on a
321 * little-endian host system), and by the time we reach here (via an
322 * opcode helper) the addresses of subword accesses have been adjusted
323 * to account for that, which means that watchpoints will not match.
324 * Undo the adjustment here.
326 if (arm_sctlr_b(env)) {
327 if (len == 1) {
328 addr ^= 3;
329 } else if (len == 2) {
330 addr ^= 2;
334 return addr;
337 #endif