target-arm: Support fp registers in gdb stub
[qemu.git] / target-lm32 / translate.c
blob6ea0ecd63b14c285a491ce25a7ec4a73f7c9d6a1
1 /*
2 * LatticeMico32 main translation routines.
4 * Copyright (c) 2010 Michael Walle <michael@walle.cc>
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2 of the License, or (at your option) any later version.
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
20 #include "cpu.h"
21 #include "disas/disas.h"
22 #include "helper.h"
23 #include "tcg-op.h"
25 #include "hw/lm32/lm32_pic.h"
27 #define GEN_HELPER 1
28 #include "helper.h"
30 #define DISAS_LM32 1
31 #if DISAS_LM32
32 # define LOG_DIS(...) qemu_log_mask(CPU_LOG_TB_IN_ASM, ## __VA_ARGS__)
33 #else
34 # define LOG_DIS(...) do { } while (0)
35 #endif
37 #define EXTRACT_FIELD(src, start, end) \
38 (((src) >> start) & ((1 << (end - start + 1)) - 1))
40 #define MEM_INDEX 0
42 static TCGv_ptr cpu_env;
43 static TCGv cpu_R[32];
44 static TCGv cpu_pc;
45 static TCGv cpu_ie;
46 static TCGv cpu_icc;
47 static TCGv cpu_dcc;
48 static TCGv cpu_cc;
49 static TCGv cpu_cfg;
50 static TCGv cpu_eba;
51 static TCGv cpu_dc;
52 static TCGv cpu_deba;
53 static TCGv cpu_bp[4];
54 static TCGv cpu_wp[4];
56 #include "exec/gen-icount.h"
58 enum {
59 OP_FMT_RI,
60 OP_FMT_RR,
61 OP_FMT_CR,
62 OP_FMT_I
65 /* This is the state at translation time. */
66 typedef struct DisasContext {
67 CPULM32State *env;
68 target_ulong pc;
70 /* Decoder. */
71 int format;
72 uint32_t ir;
73 uint8_t opcode;
74 uint8_t r0, r1, r2, csr;
75 uint16_t imm5;
76 uint16_t imm16;
77 uint32_t imm26;
79 unsigned int delayed_branch;
80 unsigned int tb_flags, synced_flags; /* tb dependent flags. */
81 int is_jmp;
83 int nr_nops;
84 struct TranslationBlock *tb;
85 int singlestep_enabled;
86 } DisasContext;
88 static const char *regnames[] = {
89 "r0", "r1", "r2", "r3", "r4", "r5", "r6", "r7",
90 "r8", "r9", "r10", "r11", "r12", "r13", "r14", "r15",
91 "r16", "r17", "r18", "r19", "r20", "r21", "r22", "r23",
92 "r24", "r25", "r26/gp", "r27/fp", "r28/sp", "r29/ra",
93 "r30/ea", "r31/ba", "bp0", "bp1", "bp2", "bp3", "wp0",
94 "wp1", "wp2", "wp3"
97 static inline int zero_extend(unsigned int val, int width)
99 return val & ((1 << width) - 1);
102 static inline int sign_extend(unsigned int val, int width)
104 int sval;
106 /* LSL. */
107 val <<= 32 - width;
108 sval = val;
109 /* ASR. */
110 sval >>= 32 - width;
112 return sval;
115 static inline void t_gen_raise_exception(DisasContext *dc, uint32_t index)
117 TCGv_i32 tmp = tcg_const_i32(index);
119 gen_helper_raise_exception(cpu_env, tmp);
120 tcg_temp_free_i32(tmp);
123 static void gen_goto_tb(DisasContext *dc, int n, target_ulong dest)
125 TranslationBlock *tb;
127 tb = dc->tb;
128 if ((tb->pc & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK) &&
129 likely(!dc->singlestep_enabled)) {
130 tcg_gen_goto_tb(n);
131 tcg_gen_movi_tl(cpu_pc, dest);
132 tcg_gen_exit_tb((uintptr_t)tb + n);
133 } else {
134 tcg_gen_movi_tl(cpu_pc, dest);
135 if (dc->singlestep_enabled) {
136 t_gen_raise_exception(dc, EXCP_DEBUG);
138 tcg_gen_exit_tb(0);
142 static void dec_add(DisasContext *dc)
144 if (dc->format == OP_FMT_RI) {
145 if (dc->r0 == R_R0) {
146 if (dc->r1 == R_R0 && dc->imm16 == 0) {
147 LOG_DIS("nop\n");
148 } else {
149 LOG_DIS("mvi r%d, %d\n", dc->r1, sign_extend(dc->imm16, 16));
151 } else {
152 LOG_DIS("addi r%d, r%d, %d\n", dc->r1, dc->r0,
153 sign_extend(dc->imm16, 16));
155 } else {
156 LOG_DIS("add r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
159 if (dc->format == OP_FMT_RI) {
160 tcg_gen_addi_tl(cpu_R[dc->r1], cpu_R[dc->r0],
161 sign_extend(dc->imm16, 16));
162 } else {
163 tcg_gen_add_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
167 static void dec_and(DisasContext *dc)
169 if (dc->format == OP_FMT_RI) {
170 LOG_DIS("andi r%d, r%d, %d\n", dc->r1, dc->r0,
171 zero_extend(dc->imm16, 16));
172 } else {
173 LOG_DIS("and r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
176 if (dc->format == OP_FMT_RI) {
177 tcg_gen_andi_tl(cpu_R[dc->r1], cpu_R[dc->r0],
178 zero_extend(dc->imm16, 16));
179 } else {
180 if (dc->r0 == 0 && dc->r1 == 0 && dc->r2 == 0) {
181 tcg_gen_movi_tl(cpu_pc, dc->pc + 4);
182 gen_helper_hlt(cpu_env);
183 } else {
184 tcg_gen_and_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
189 static void dec_andhi(DisasContext *dc)
191 LOG_DIS("andhi r%d, r%d, %d\n", dc->r2, dc->r0, dc->imm16);
193 tcg_gen_andi_tl(cpu_R[dc->r1], cpu_R[dc->r0], (dc->imm16 << 16));
196 static void dec_b(DisasContext *dc)
198 if (dc->r0 == R_RA) {
199 LOG_DIS("ret\n");
200 } else if (dc->r0 == R_EA) {
201 LOG_DIS("eret\n");
202 } else if (dc->r0 == R_BA) {
203 LOG_DIS("bret\n");
204 } else {
205 LOG_DIS("b r%d\n", dc->r0);
208 /* restore IE.IE in case of an eret */
209 if (dc->r0 == R_EA) {
210 TCGv t0 = tcg_temp_new();
211 int l1 = gen_new_label();
212 tcg_gen_andi_tl(t0, cpu_ie, IE_EIE);
213 tcg_gen_ori_tl(cpu_ie, cpu_ie, IE_IE);
214 tcg_gen_brcondi_tl(TCG_COND_EQ, t0, IE_EIE, l1);
215 tcg_gen_andi_tl(cpu_ie, cpu_ie, ~IE_IE);
216 gen_set_label(l1);
217 tcg_temp_free(t0);
218 } else if (dc->r0 == R_BA) {
219 TCGv t0 = tcg_temp_new();
220 int l1 = gen_new_label();
221 tcg_gen_andi_tl(t0, cpu_ie, IE_BIE);
222 tcg_gen_ori_tl(cpu_ie, cpu_ie, IE_IE);
223 tcg_gen_brcondi_tl(TCG_COND_EQ, t0, IE_BIE, l1);
224 tcg_gen_andi_tl(cpu_ie, cpu_ie, ~IE_IE);
225 gen_set_label(l1);
226 tcg_temp_free(t0);
228 tcg_gen_mov_tl(cpu_pc, cpu_R[dc->r0]);
230 dc->is_jmp = DISAS_JUMP;
233 static void dec_bi(DisasContext *dc)
235 LOG_DIS("bi %d\n", sign_extend(dc->imm26 << 2, 26));
237 gen_goto_tb(dc, 0, dc->pc + (sign_extend(dc->imm26 << 2, 26)));
239 dc->is_jmp = DISAS_TB_JUMP;
242 static inline void gen_cond_branch(DisasContext *dc, int cond)
244 int l1;
246 l1 = gen_new_label();
247 tcg_gen_brcond_tl(cond, cpu_R[dc->r0], cpu_R[dc->r1], l1);
248 gen_goto_tb(dc, 0, dc->pc + 4);
249 gen_set_label(l1);
250 gen_goto_tb(dc, 1, dc->pc + (sign_extend(dc->imm16 << 2, 16)));
251 dc->is_jmp = DISAS_TB_JUMP;
254 static void dec_be(DisasContext *dc)
256 LOG_DIS("be r%d, r%d, %d\n", dc->r0, dc->r1,
257 sign_extend(dc->imm16, 16) * 4);
259 gen_cond_branch(dc, TCG_COND_EQ);
262 static void dec_bg(DisasContext *dc)
264 LOG_DIS("bg r%d, r%d, %d\n", dc->r0, dc->r1,
265 sign_extend(dc->imm16, 16 * 4));
267 gen_cond_branch(dc, TCG_COND_GT);
270 static void dec_bge(DisasContext *dc)
272 LOG_DIS("bge r%d, r%d, %d\n", dc->r0, dc->r1,
273 sign_extend(dc->imm16, 16) * 4);
275 gen_cond_branch(dc, TCG_COND_GE);
278 static void dec_bgeu(DisasContext *dc)
280 LOG_DIS("bgeu r%d, r%d, %d\n", dc->r0, dc->r1,
281 sign_extend(dc->imm16, 16) * 4);
283 gen_cond_branch(dc, TCG_COND_GEU);
286 static void dec_bgu(DisasContext *dc)
288 LOG_DIS("bgu r%d, r%d, %d\n", dc->r0, dc->r1,
289 sign_extend(dc->imm16, 16) * 4);
291 gen_cond_branch(dc, TCG_COND_GTU);
294 static void dec_bne(DisasContext *dc)
296 LOG_DIS("bne r%d, r%d, %d\n", dc->r0, dc->r1,
297 sign_extend(dc->imm16, 16) * 4);
299 gen_cond_branch(dc, TCG_COND_NE);
302 static void dec_call(DisasContext *dc)
304 LOG_DIS("call r%d\n", dc->r0);
306 tcg_gen_movi_tl(cpu_R[R_RA], dc->pc + 4);
307 tcg_gen_mov_tl(cpu_pc, cpu_R[dc->r0]);
309 dc->is_jmp = DISAS_JUMP;
312 static void dec_calli(DisasContext *dc)
314 LOG_DIS("calli %d\n", sign_extend(dc->imm26, 26) * 4);
316 tcg_gen_movi_tl(cpu_R[R_RA], dc->pc + 4);
317 gen_goto_tb(dc, 0, dc->pc + (sign_extend(dc->imm26 << 2, 26)));
319 dc->is_jmp = DISAS_TB_JUMP;
322 static inline void gen_compare(DisasContext *dc, int cond)
324 int rX = (dc->format == OP_FMT_RR) ? dc->r2 : dc->r1;
325 int rY = (dc->format == OP_FMT_RR) ? dc->r0 : dc->r0;
326 int rZ = (dc->format == OP_FMT_RR) ? dc->r1 : -1;
327 int i;
329 if (dc->format == OP_FMT_RI) {
330 switch (cond) {
331 case TCG_COND_GEU:
332 case TCG_COND_GTU:
333 i = zero_extend(dc->imm16, 16);
334 break;
335 default:
336 i = sign_extend(dc->imm16, 16);
337 break;
340 tcg_gen_setcondi_tl(cond, cpu_R[rX], cpu_R[rY], i);
341 } else {
342 tcg_gen_setcond_tl(cond, cpu_R[rX], cpu_R[rY], cpu_R[rZ]);
346 static void dec_cmpe(DisasContext *dc)
348 if (dc->format == OP_FMT_RI) {
349 LOG_DIS("cmpei r%d, r%d, %d\n", dc->r0, dc->r1,
350 sign_extend(dc->imm16, 16));
351 } else {
352 LOG_DIS("cmpe r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
355 gen_compare(dc, TCG_COND_EQ);
358 static void dec_cmpg(DisasContext *dc)
360 if (dc->format == OP_FMT_RI) {
361 LOG_DIS("cmpgi r%d, r%d, %d\n", dc->r0, dc->r1,
362 sign_extend(dc->imm16, 16));
363 } else {
364 LOG_DIS("cmpg r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
367 gen_compare(dc, TCG_COND_GT);
370 static void dec_cmpge(DisasContext *dc)
372 if (dc->format == OP_FMT_RI) {
373 LOG_DIS("cmpgei r%d, r%d, %d\n", dc->r0, dc->r1,
374 sign_extend(dc->imm16, 16));
375 } else {
376 LOG_DIS("cmpge r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
379 gen_compare(dc, TCG_COND_GE);
382 static void dec_cmpgeu(DisasContext *dc)
384 if (dc->format == OP_FMT_RI) {
385 LOG_DIS("cmpgeui r%d, r%d, %d\n", dc->r0, dc->r1,
386 zero_extend(dc->imm16, 16));
387 } else {
388 LOG_DIS("cmpgeu r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
391 gen_compare(dc, TCG_COND_GEU);
394 static void dec_cmpgu(DisasContext *dc)
396 if (dc->format == OP_FMT_RI) {
397 LOG_DIS("cmpgui r%d, r%d, %d\n", dc->r0, dc->r1,
398 zero_extend(dc->imm16, 16));
399 } else {
400 LOG_DIS("cmpgu r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
403 gen_compare(dc, TCG_COND_GTU);
406 static void dec_cmpne(DisasContext *dc)
408 if (dc->format == OP_FMT_RI) {
409 LOG_DIS("cmpnei r%d, r%d, %d\n", dc->r0, dc->r1,
410 sign_extend(dc->imm16, 16));
411 } else {
412 LOG_DIS("cmpne r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
415 gen_compare(dc, TCG_COND_NE);
418 static void dec_divu(DisasContext *dc)
420 int l1;
422 LOG_DIS("divu r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
424 if (!(dc->env->features & LM32_FEATURE_DIVIDE)) {
425 cpu_abort(dc->env, "hardware divider is not available\n");
428 l1 = gen_new_label();
429 tcg_gen_brcondi_tl(TCG_COND_NE, cpu_R[dc->r1], 0, l1);
430 tcg_gen_movi_tl(cpu_pc, dc->pc);
431 t_gen_raise_exception(dc, EXCP_DIVIDE_BY_ZERO);
432 gen_set_label(l1);
433 tcg_gen_divu_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
436 static void dec_lb(DisasContext *dc)
438 TCGv t0;
440 LOG_DIS("lb r%d, (r%d+%d)\n", dc->r1, dc->r0, dc->imm16);
442 t0 = tcg_temp_new();
443 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
444 tcg_gen_qemu_ld8s(cpu_R[dc->r1], t0, MEM_INDEX);
445 tcg_temp_free(t0);
448 static void dec_lbu(DisasContext *dc)
450 TCGv t0;
452 LOG_DIS("lbu r%d, (r%d+%d)\n", dc->r1, dc->r0, dc->imm16);
454 t0 = tcg_temp_new();
455 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
456 tcg_gen_qemu_ld8u(cpu_R[dc->r1], t0, MEM_INDEX);
457 tcg_temp_free(t0);
460 static void dec_lh(DisasContext *dc)
462 TCGv t0;
464 LOG_DIS("lh r%d, (r%d+%d)\n", dc->r1, dc->r0, dc->imm16);
466 t0 = tcg_temp_new();
467 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
468 tcg_gen_qemu_ld16s(cpu_R[dc->r1], t0, MEM_INDEX);
469 tcg_temp_free(t0);
472 static void dec_lhu(DisasContext *dc)
474 TCGv t0;
476 LOG_DIS("lhu r%d, (r%d+%d)\n", dc->r1, dc->r0, dc->imm16);
478 t0 = tcg_temp_new();
479 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
480 tcg_gen_qemu_ld16u(cpu_R[dc->r1], t0, MEM_INDEX);
481 tcg_temp_free(t0);
484 static void dec_lw(DisasContext *dc)
486 TCGv t0;
488 LOG_DIS("lw r%d, (r%d+%d)\n", dc->r1, dc->r0, sign_extend(dc->imm16, 16));
490 t0 = tcg_temp_new();
491 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
492 tcg_gen_qemu_ld32s(cpu_R[dc->r1], t0, MEM_INDEX);
493 tcg_temp_free(t0);
496 static void dec_modu(DisasContext *dc)
498 int l1;
500 LOG_DIS("modu r%d, r%d, %d\n", dc->r2, dc->r0, dc->r1);
502 if (!(dc->env->features & LM32_FEATURE_DIVIDE)) {
503 cpu_abort(dc->env, "hardware divider is not available\n");
506 l1 = gen_new_label();
507 tcg_gen_brcondi_tl(TCG_COND_NE, cpu_R[dc->r1], 0, l1);
508 tcg_gen_movi_tl(cpu_pc, dc->pc);
509 t_gen_raise_exception(dc, EXCP_DIVIDE_BY_ZERO);
510 gen_set_label(l1);
511 tcg_gen_remu_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
514 static void dec_mul(DisasContext *dc)
516 if (dc->format == OP_FMT_RI) {
517 LOG_DIS("muli r%d, r%d, %d\n", dc->r0, dc->r1,
518 sign_extend(dc->imm16, 16));
519 } else {
520 LOG_DIS("mul r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
523 if (!(dc->env->features & LM32_FEATURE_MULTIPLY)) {
524 cpu_abort(dc->env, "hardware multiplier is not available\n");
527 if (dc->format == OP_FMT_RI) {
528 tcg_gen_muli_tl(cpu_R[dc->r1], cpu_R[dc->r0],
529 sign_extend(dc->imm16, 16));
530 } else {
531 tcg_gen_mul_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
535 static void dec_nor(DisasContext *dc)
537 if (dc->format == OP_FMT_RI) {
538 LOG_DIS("nori r%d, r%d, %d\n", dc->r0, dc->r1,
539 zero_extend(dc->imm16, 16));
540 } else {
541 LOG_DIS("nor r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
544 if (dc->format == OP_FMT_RI) {
545 TCGv t0 = tcg_temp_new();
546 tcg_gen_movi_tl(t0, zero_extend(dc->imm16, 16));
547 tcg_gen_nor_tl(cpu_R[dc->r1], cpu_R[dc->r0], t0);
548 tcg_temp_free(t0);
549 } else {
550 tcg_gen_nor_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
554 static void dec_or(DisasContext *dc)
556 if (dc->format == OP_FMT_RI) {
557 LOG_DIS("ori r%d, r%d, %d\n", dc->r1, dc->r0,
558 zero_extend(dc->imm16, 16));
559 } else {
560 if (dc->r1 == R_R0) {
561 LOG_DIS("mv r%d, r%d\n", dc->r2, dc->r0);
562 } else {
563 LOG_DIS("or r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
567 if (dc->format == OP_FMT_RI) {
568 tcg_gen_ori_tl(cpu_R[dc->r1], cpu_R[dc->r0],
569 zero_extend(dc->imm16, 16));
570 } else {
571 tcg_gen_or_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
575 static void dec_orhi(DisasContext *dc)
577 if (dc->r0 == R_R0) {
578 LOG_DIS("mvhi r%d, %d\n", dc->r1, dc->imm16);
579 } else {
580 LOG_DIS("orhi r%d, r%d, %d\n", dc->r1, dc->r0, dc->imm16);
583 tcg_gen_ori_tl(cpu_R[dc->r1], cpu_R[dc->r0], (dc->imm16 << 16));
586 static void dec_scall(DisasContext *dc)
588 if (dc->imm5 == 7) {
589 LOG_DIS("scall\n");
590 } else if (dc->imm5 == 2) {
591 LOG_DIS("break\n");
592 } else {
593 cpu_abort(dc->env, "invalid opcode\n");
596 if (dc->imm5 == 7) {
597 tcg_gen_movi_tl(cpu_pc, dc->pc);
598 t_gen_raise_exception(dc, EXCP_SYSTEMCALL);
599 } else {
600 tcg_gen_movi_tl(cpu_pc, dc->pc);
601 t_gen_raise_exception(dc, EXCP_BREAKPOINT);
605 static void dec_rcsr(DisasContext *dc)
607 LOG_DIS("rcsr r%d, %d\n", dc->r2, dc->csr);
609 switch (dc->csr) {
610 case CSR_IE:
611 tcg_gen_mov_tl(cpu_R[dc->r2], cpu_ie);
612 break;
613 case CSR_IM:
614 gen_helper_rcsr_im(cpu_R[dc->r2], cpu_env);
615 break;
616 case CSR_IP:
617 gen_helper_rcsr_ip(cpu_R[dc->r2], cpu_env);
618 break;
619 case CSR_CC:
620 tcg_gen_mov_tl(cpu_R[dc->r2], cpu_cc);
621 break;
622 case CSR_CFG:
623 tcg_gen_mov_tl(cpu_R[dc->r2], cpu_cfg);
624 break;
625 case CSR_EBA:
626 tcg_gen_mov_tl(cpu_R[dc->r2], cpu_eba);
627 break;
628 case CSR_DC:
629 tcg_gen_mov_tl(cpu_R[dc->r2], cpu_dc);
630 break;
631 case CSR_DEBA:
632 tcg_gen_mov_tl(cpu_R[dc->r2], cpu_deba);
633 break;
634 case CSR_JTX:
635 gen_helper_rcsr_jtx(cpu_R[dc->r2], cpu_env);
636 break;
637 case CSR_JRX:
638 gen_helper_rcsr_jrx(cpu_R[dc->r2], cpu_env);
639 break;
640 case CSR_ICC:
641 case CSR_DCC:
642 case CSR_BP0:
643 case CSR_BP1:
644 case CSR_BP2:
645 case CSR_BP3:
646 case CSR_WP0:
647 case CSR_WP1:
648 case CSR_WP2:
649 case CSR_WP3:
650 cpu_abort(dc->env, "invalid read access csr=%x\n", dc->csr);
651 break;
652 default:
653 cpu_abort(dc->env, "read_csr: unknown csr=%x\n", dc->csr);
654 break;
658 static void dec_sb(DisasContext *dc)
660 TCGv t0;
662 LOG_DIS("sb (r%d+%d), r%d\n", dc->r0, dc->imm16, dc->r1);
664 t0 = tcg_temp_new();
665 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
666 tcg_gen_qemu_st8(cpu_R[dc->r1], t0, MEM_INDEX);
667 tcg_temp_free(t0);
670 static void dec_sextb(DisasContext *dc)
672 LOG_DIS("sextb r%d, r%d\n", dc->r2, dc->r0);
674 if (!(dc->env->features & LM32_FEATURE_SIGN_EXTEND)) {
675 cpu_abort(dc->env, "hardware sign extender is not available\n");
678 tcg_gen_ext8s_tl(cpu_R[dc->r2], cpu_R[dc->r0]);
681 static void dec_sexth(DisasContext *dc)
683 LOG_DIS("sexth r%d, r%d\n", dc->r2, dc->r0);
685 if (!(dc->env->features & LM32_FEATURE_SIGN_EXTEND)) {
686 cpu_abort(dc->env, "hardware sign extender is not available\n");
689 tcg_gen_ext16s_tl(cpu_R[dc->r2], cpu_R[dc->r0]);
692 static void dec_sh(DisasContext *dc)
694 TCGv t0;
696 LOG_DIS("sh (r%d+%d), r%d\n", dc->r0, dc->imm16, dc->r1);
698 t0 = tcg_temp_new();
699 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
700 tcg_gen_qemu_st16(cpu_R[dc->r1], t0, MEM_INDEX);
701 tcg_temp_free(t0);
704 static void dec_sl(DisasContext *dc)
706 if (dc->format == OP_FMT_RI) {
707 LOG_DIS("sli r%d, r%d, %d\n", dc->r1, dc->r0, dc->imm5);
708 } else {
709 LOG_DIS("sl r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
712 if (!(dc->env->features & LM32_FEATURE_SHIFT)) {
713 cpu_abort(dc->env, "hardware shifter is not available\n");
716 if (dc->format == OP_FMT_RI) {
717 tcg_gen_shli_tl(cpu_R[dc->r1], cpu_R[dc->r0], dc->imm5);
718 } else {
719 TCGv t0 = tcg_temp_new();
720 tcg_gen_andi_tl(t0, cpu_R[dc->r1], 0x1f);
721 tcg_gen_shl_tl(cpu_R[dc->r2], cpu_R[dc->r0], t0);
722 tcg_temp_free(t0);
726 static void dec_sr(DisasContext *dc)
728 if (dc->format == OP_FMT_RI) {
729 LOG_DIS("sri r%d, r%d, %d\n", dc->r1, dc->r0, dc->imm5);
730 } else {
731 LOG_DIS("sr r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
734 if (!(dc->env->features & LM32_FEATURE_SHIFT)) {
735 if (dc->format == OP_FMT_RI) {
736 /* TODO: check r1 == 1 during runtime */
737 } else {
738 if (dc->imm5 != 1) {
739 cpu_abort(dc->env, "hardware shifter is not available\n");
744 if (dc->format == OP_FMT_RI) {
745 tcg_gen_sari_tl(cpu_R[dc->r1], cpu_R[dc->r0], dc->imm5);
746 } else {
747 TCGv t0 = tcg_temp_new();
748 tcg_gen_andi_tl(t0, cpu_R[dc->r1], 0x1f);
749 tcg_gen_sar_tl(cpu_R[dc->r2], cpu_R[dc->r0], t0);
750 tcg_temp_free(t0);
754 static void dec_sru(DisasContext *dc)
756 if (dc->format == OP_FMT_RI) {
757 LOG_DIS("srui r%d, r%d, %d\n", dc->r1, dc->r0, dc->imm5);
758 } else {
759 LOG_DIS("sru r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
762 if (!(dc->env->features & LM32_FEATURE_SHIFT)) {
763 if (dc->format == OP_FMT_RI) {
764 /* TODO: check r1 == 1 during runtime */
765 } else {
766 if (dc->imm5 != 1) {
767 cpu_abort(dc->env, "hardware shifter is not available\n");
772 if (dc->format == OP_FMT_RI) {
773 tcg_gen_shri_tl(cpu_R[dc->r1], cpu_R[dc->r0], dc->imm5);
774 } else {
775 TCGv t0 = tcg_temp_new();
776 tcg_gen_andi_tl(t0, cpu_R[dc->r1], 0x1f);
777 tcg_gen_shr_tl(cpu_R[dc->r2], cpu_R[dc->r0], t0);
778 tcg_temp_free(t0);
782 static void dec_sub(DisasContext *dc)
784 LOG_DIS("sub r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
786 tcg_gen_sub_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
789 static void dec_sw(DisasContext *dc)
791 TCGv t0;
793 LOG_DIS("sw (r%d+%d), r%d\n", dc->r0, sign_extend(dc->imm16, 16), dc->r1);
795 t0 = tcg_temp_new();
796 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
797 tcg_gen_qemu_st32(cpu_R[dc->r1], t0, MEM_INDEX);
798 tcg_temp_free(t0);
801 static void dec_user(DisasContext *dc)
803 LOG_DIS("user");
805 cpu_abort(dc->env, "user insn undefined\n");
808 static void dec_wcsr(DisasContext *dc)
810 int no;
812 LOG_DIS("wcsr r%d, %d\n", dc->r1, dc->csr);
814 switch (dc->csr) {
815 case CSR_IE:
816 tcg_gen_mov_tl(cpu_ie, cpu_R[dc->r1]);
817 tcg_gen_movi_tl(cpu_pc, dc->pc + 4);
818 dc->is_jmp = DISAS_UPDATE;
819 break;
820 case CSR_IM:
821 /* mark as an io operation because it could cause an interrupt */
822 if (use_icount) {
823 gen_io_start();
825 gen_helper_wcsr_im(cpu_env, cpu_R[dc->r1]);
826 tcg_gen_movi_tl(cpu_pc, dc->pc + 4);
827 if (use_icount) {
828 gen_io_end();
830 dc->is_jmp = DISAS_UPDATE;
831 break;
832 case CSR_IP:
833 /* mark as an io operation because it could cause an interrupt */
834 if (use_icount) {
835 gen_io_start();
837 gen_helper_wcsr_ip(cpu_env, cpu_R[dc->r1]);
838 tcg_gen_movi_tl(cpu_pc, dc->pc + 4);
839 if (use_icount) {
840 gen_io_end();
842 dc->is_jmp = DISAS_UPDATE;
843 break;
844 case CSR_ICC:
845 /* TODO */
846 break;
847 case CSR_DCC:
848 /* TODO */
849 break;
850 case CSR_EBA:
851 tcg_gen_mov_tl(cpu_eba, cpu_R[dc->r1]);
852 break;
853 case CSR_DEBA:
854 tcg_gen_mov_tl(cpu_deba, cpu_R[dc->r1]);
855 break;
856 case CSR_JTX:
857 gen_helper_wcsr_jtx(cpu_env, cpu_R[dc->r1]);
858 break;
859 case CSR_JRX:
860 gen_helper_wcsr_jrx(cpu_env, cpu_R[dc->r1]);
861 break;
862 case CSR_DC:
863 tcg_gen_mov_tl(cpu_dc, cpu_R[dc->r1]);
864 break;
865 case CSR_BP0:
866 case CSR_BP1:
867 case CSR_BP2:
868 case CSR_BP3:
869 no = dc->csr - CSR_BP0;
870 if (dc->env->num_bps <= no) {
871 cpu_abort(dc->env, "breakpoint #%i is not available\n", no);
873 tcg_gen_mov_tl(cpu_bp[no], cpu_R[dc->r1]);
874 break;
875 case CSR_WP0:
876 case CSR_WP1:
877 case CSR_WP2:
878 case CSR_WP3:
879 no = dc->csr - CSR_WP0;
880 if (dc->env->num_wps <= no) {
881 cpu_abort(dc->env, "watchpoint #%i is not available\n", no);
883 tcg_gen_mov_tl(cpu_wp[no], cpu_R[dc->r1]);
884 break;
885 case CSR_CC:
886 case CSR_CFG:
887 cpu_abort(dc->env, "invalid write access csr=%x\n", dc->csr);
888 break;
889 default:
890 cpu_abort(dc->env, "write_csr unknown csr=%x\n", dc->csr);
891 break;
895 static void dec_xnor(DisasContext *dc)
897 if (dc->format == OP_FMT_RI) {
898 LOG_DIS("xnori r%d, r%d, %d\n", dc->r0, dc->r1,
899 zero_extend(dc->imm16, 16));
900 } else {
901 if (dc->r1 == R_R0) {
902 LOG_DIS("not r%d, r%d\n", dc->r2, dc->r0);
903 } else {
904 LOG_DIS("xnor r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
908 if (dc->format == OP_FMT_RI) {
909 tcg_gen_xori_tl(cpu_R[dc->r1], cpu_R[dc->r0],
910 zero_extend(dc->imm16, 16));
911 tcg_gen_not_tl(cpu_R[dc->r1], cpu_R[dc->r1]);
912 } else {
913 tcg_gen_eqv_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
917 static void dec_xor(DisasContext *dc)
919 if (dc->format == OP_FMT_RI) {
920 LOG_DIS("xori r%d, r%d, %d\n", dc->r0, dc->r1,
921 zero_extend(dc->imm16, 16));
922 } else {
923 LOG_DIS("xor r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
926 if (dc->format == OP_FMT_RI) {
927 tcg_gen_xori_tl(cpu_R[dc->r1], cpu_R[dc->r0],
928 zero_extend(dc->imm16, 16));
929 } else {
930 tcg_gen_xor_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
934 static void dec_ill(DisasContext *dc)
936 cpu_abort(dc->env, "unknown opcode 0x%02x\n", dc->opcode);
939 typedef void (*DecoderInfo)(DisasContext *dc);
940 static const DecoderInfo decinfo[] = {
941 dec_sru, dec_nor, dec_mul, dec_sh, dec_lb, dec_sr, dec_xor, dec_lh,
942 dec_and, dec_xnor, dec_lw, dec_lhu, dec_sb, dec_add, dec_or, dec_sl,
943 dec_lbu, dec_be, dec_bg, dec_bge, dec_bgeu, dec_bgu, dec_sw, dec_bne,
944 dec_andhi, dec_cmpe, dec_cmpg, dec_cmpge, dec_cmpgeu, dec_cmpgu, dec_orhi,
945 dec_cmpne,
946 dec_sru, dec_nor, dec_mul, dec_divu, dec_rcsr, dec_sr, dec_xor, dec_ill,
947 dec_and, dec_xnor, dec_ill, dec_scall, dec_sextb, dec_add, dec_or, dec_sl,
948 dec_b, dec_modu, dec_sub, dec_user, dec_wcsr, dec_ill, dec_call, dec_sexth,
949 dec_bi, dec_cmpe, dec_cmpg, dec_cmpge, dec_cmpgeu, dec_cmpgu, dec_calli,
950 dec_cmpne
953 static inline void decode(DisasContext *dc, uint32_t ir)
955 if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP | CPU_LOG_TB_OP_OPT))) {
956 tcg_gen_debug_insn_start(dc->pc);
959 dc->ir = ir;
960 LOG_DIS("%8.8x\t", dc->ir);
962 /* try guessing 'empty' instruction memory, although it may be a valid
963 * instruction sequence (eg. srui r0, r0, 0) */
964 if (dc->ir) {
965 dc->nr_nops = 0;
966 } else {
967 LOG_DIS("nr_nops=%d\t", dc->nr_nops);
968 dc->nr_nops++;
969 if (dc->nr_nops > 4) {
970 cpu_abort(dc->env, "fetching nop sequence\n");
974 dc->opcode = EXTRACT_FIELD(ir, 26, 31);
976 dc->imm5 = EXTRACT_FIELD(ir, 0, 4);
977 dc->imm16 = EXTRACT_FIELD(ir, 0, 15);
978 dc->imm26 = EXTRACT_FIELD(ir, 0, 25);
980 dc->csr = EXTRACT_FIELD(ir, 21, 25);
981 dc->r0 = EXTRACT_FIELD(ir, 21, 25);
982 dc->r1 = EXTRACT_FIELD(ir, 16, 20);
983 dc->r2 = EXTRACT_FIELD(ir, 11, 15);
985 /* bit 31 seems to indicate insn type. */
986 if (ir & (1 << 31)) {
987 dc->format = OP_FMT_RR;
988 } else {
989 dc->format = OP_FMT_RI;
992 assert(ARRAY_SIZE(decinfo) == 64);
993 assert(dc->opcode < 64);
995 decinfo[dc->opcode](dc);
998 static void check_breakpoint(CPULM32State *env, DisasContext *dc)
1000 CPUBreakpoint *bp;
1002 if (unlikely(!QTAILQ_EMPTY(&env->breakpoints))) {
1003 QTAILQ_FOREACH(bp, &env->breakpoints, entry) {
1004 if (bp->pc == dc->pc) {
1005 tcg_gen_movi_tl(cpu_pc, dc->pc);
1006 t_gen_raise_exception(dc, EXCP_DEBUG);
1007 dc->is_jmp = DISAS_UPDATE;
1013 /* generate intermediate code for basic block 'tb'. */
1014 static inline
1015 void gen_intermediate_code_internal(LM32CPU *cpu,
1016 TranslationBlock *tb, bool search_pc)
1018 CPUState *cs = CPU(cpu);
1019 CPULM32State *env = &cpu->env;
1020 struct DisasContext ctx, *dc = &ctx;
1021 uint16_t *gen_opc_end;
1022 uint32_t pc_start;
1023 int j, lj;
1024 uint32_t next_page_start;
1025 int num_insns;
1026 int max_insns;
1028 pc_start = tb->pc;
1029 dc->env = env;
1030 dc->tb = tb;
1032 gen_opc_end = tcg_ctx.gen_opc_buf + OPC_MAX_SIZE;
1034 dc->is_jmp = DISAS_NEXT;
1035 dc->pc = pc_start;
1036 dc->singlestep_enabled = cs->singlestep_enabled;
1037 dc->nr_nops = 0;
1039 if (pc_start & 3) {
1040 cpu_abort(env, "LM32: unaligned PC=%x\n", pc_start);
1043 next_page_start = (pc_start & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
1044 lj = -1;
1045 num_insns = 0;
1046 max_insns = tb->cflags & CF_COUNT_MASK;
1047 if (max_insns == 0) {
1048 max_insns = CF_COUNT_MASK;
1051 gen_tb_start();
1052 do {
1053 check_breakpoint(env, dc);
1055 if (search_pc) {
1056 j = tcg_ctx.gen_opc_ptr - tcg_ctx.gen_opc_buf;
1057 if (lj < j) {
1058 lj++;
1059 while (lj < j) {
1060 tcg_ctx.gen_opc_instr_start[lj++] = 0;
1063 tcg_ctx.gen_opc_pc[lj] = dc->pc;
1064 tcg_ctx.gen_opc_instr_start[lj] = 1;
1065 tcg_ctx.gen_opc_icount[lj] = num_insns;
1068 /* Pretty disas. */
1069 LOG_DIS("%8.8x:\t", dc->pc);
1071 if (num_insns + 1 == max_insns && (tb->cflags & CF_LAST_IO)) {
1072 gen_io_start();
1075 decode(dc, cpu_ldl_code(env, dc->pc));
1076 dc->pc += 4;
1077 num_insns++;
1079 } while (!dc->is_jmp
1080 && tcg_ctx.gen_opc_ptr < gen_opc_end
1081 && !cs->singlestep_enabled
1082 && !singlestep
1083 && (dc->pc < next_page_start)
1084 && num_insns < max_insns);
1086 if (tb->cflags & CF_LAST_IO) {
1087 gen_io_end();
1090 if (unlikely(cs->singlestep_enabled)) {
1091 if (dc->is_jmp == DISAS_NEXT) {
1092 tcg_gen_movi_tl(cpu_pc, dc->pc);
1094 t_gen_raise_exception(dc, EXCP_DEBUG);
1095 } else {
1096 switch (dc->is_jmp) {
1097 case DISAS_NEXT:
1098 gen_goto_tb(dc, 1, dc->pc);
1099 break;
1100 default:
1101 case DISAS_JUMP:
1102 case DISAS_UPDATE:
1103 /* indicate that the hash table must be used
1104 to find the next TB */
1105 tcg_gen_exit_tb(0);
1106 break;
1107 case DISAS_TB_JUMP:
1108 /* nothing more to generate */
1109 break;
1113 gen_tb_end(tb, num_insns);
1114 *tcg_ctx.gen_opc_ptr = INDEX_op_end;
1115 if (search_pc) {
1116 j = tcg_ctx.gen_opc_ptr - tcg_ctx.gen_opc_buf;
1117 lj++;
1118 while (lj <= j) {
1119 tcg_ctx.gen_opc_instr_start[lj++] = 0;
1121 } else {
1122 tb->size = dc->pc - pc_start;
1123 tb->icount = num_insns;
1126 #ifdef DEBUG_DISAS
1127 if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)) {
1128 qemu_log("\n");
1129 log_target_disas(env, pc_start, dc->pc - pc_start, 0);
1130 qemu_log("\nisize=%d osize=%td\n",
1131 dc->pc - pc_start, tcg_ctx.gen_opc_ptr -
1132 tcg_ctx.gen_opc_buf);
1134 #endif
1137 void gen_intermediate_code(CPULM32State *env, struct TranslationBlock *tb)
1139 gen_intermediate_code_internal(lm32_env_get_cpu(env), tb, false);
1142 void gen_intermediate_code_pc(CPULM32State *env, struct TranslationBlock *tb)
1144 gen_intermediate_code_internal(lm32_env_get_cpu(env), tb, true);
1147 void lm32_cpu_dump_state(CPUState *cs, FILE *f, fprintf_function cpu_fprintf,
1148 int flags)
1150 LM32CPU *cpu = LM32_CPU(cs);
1151 CPULM32State *env = &cpu->env;
1152 int i;
1154 if (!env || !f) {
1155 return;
1158 cpu_fprintf(f, "IN: PC=%x %s\n",
1159 env->pc, lookup_symbol(env->pc));
1161 cpu_fprintf(f, "ie=%8.8x (IE=%x EIE=%x BIE=%x) im=%8.8x ip=%8.8x\n",
1162 env->ie,
1163 (env->ie & IE_IE) ? 1 : 0,
1164 (env->ie & IE_EIE) ? 1 : 0,
1165 (env->ie & IE_BIE) ? 1 : 0,
1166 lm32_pic_get_im(env->pic_state),
1167 lm32_pic_get_ip(env->pic_state));
1168 cpu_fprintf(f, "eba=%8.8x deba=%8.8x\n",
1169 env->eba,
1170 env->deba);
1172 for (i = 0; i < 32; i++) {
1173 cpu_fprintf(f, "r%2.2d=%8.8x ", i, env->regs[i]);
1174 if ((i + 1) % 4 == 0) {
1175 cpu_fprintf(f, "\n");
1178 cpu_fprintf(f, "\n\n");
1181 void restore_state_to_opc(CPULM32State *env, TranslationBlock *tb, int pc_pos)
1183 env->pc = tcg_ctx.gen_opc_pc[pc_pos];
1186 void lm32_translate_init(void)
1188 int i;
1190 cpu_env = tcg_global_reg_new_ptr(TCG_AREG0, "env");
1192 for (i = 0; i < ARRAY_SIZE(cpu_R); i++) {
1193 cpu_R[i] = tcg_global_mem_new(TCG_AREG0,
1194 offsetof(CPULM32State, regs[i]),
1195 regnames[i]);
1198 for (i = 0; i < ARRAY_SIZE(cpu_bp); i++) {
1199 cpu_bp[i] = tcg_global_mem_new(TCG_AREG0,
1200 offsetof(CPULM32State, bp[i]),
1201 regnames[32+i]);
1204 for (i = 0; i < ARRAY_SIZE(cpu_wp); i++) {
1205 cpu_wp[i] = tcg_global_mem_new(TCG_AREG0,
1206 offsetof(CPULM32State, wp[i]),
1207 regnames[36+i]);
1210 cpu_pc = tcg_global_mem_new(TCG_AREG0,
1211 offsetof(CPULM32State, pc),
1212 "pc");
1213 cpu_ie = tcg_global_mem_new(TCG_AREG0,
1214 offsetof(CPULM32State, ie),
1215 "ie");
1216 cpu_icc = tcg_global_mem_new(TCG_AREG0,
1217 offsetof(CPULM32State, icc),
1218 "icc");
1219 cpu_dcc = tcg_global_mem_new(TCG_AREG0,
1220 offsetof(CPULM32State, dcc),
1221 "dcc");
1222 cpu_cc = tcg_global_mem_new(TCG_AREG0,
1223 offsetof(CPULM32State, cc),
1224 "cc");
1225 cpu_cfg = tcg_global_mem_new(TCG_AREG0,
1226 offsetof(CPULM32State, cfg),
1227 "cfg");
1228 cpu_eba = tcg_global_mem_new(TCG_AREG0,
1229 offsetof(CPULM32State, eba),
1230 "eba");
1231 cpu_dc = tcg_global_mem_new(TCG_AREG0,
1232 offsetof(CPULM32State, dc),
1233 "dc");
1234 cpu_deba = tcg_global_mem_new(TCG_AREG0,
1235 offsetof(CPULM32State, deba),
1236 "deba");