Merge remote-tracking branch 'remotes/maxreitz/tags/pull-block-2019-07-02' into staging
[qemu/ar7.git] / target / lm32 / translate.c
blobb9f2f2c4a7ebb66be3f40874405612b964db3aed
1 /*
2 * LatticeMico32 main translation routines.
4 * Copyright (c) 2010 Michael Walle <michael@walle.cc>
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2 of the License, or (at your option) any later version.
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
20 #include "qemu/osdep.h"
21 #include "cpu.h"
22 #include "disas/disas.h"
23 #include "exec/helper-proto.h"
24 #include "exec/exec-all.h"
25 #include "exec/translator.h"
26 #include "tcg-op.h"
27 #include "qemu/qemu-print.h"
29 #include "exec/cpu_ldst.h"
30 #include "hw/lm32/lm32_pic.h"
32 #include "exec/helper-gen.h"
34 #include "trace-tcg.h"
35 #include "exec/log.h"
38 #define DISAS_LM32 0
40 #define LOG_DIS(...) \
41 do { \
42 if (DISAS_LM32) { \
43 qemu_log_mask(CPU_LOG_TB_IN_ASM, ## __VA_ARGS__); \
44 } \
45 } while (0)
47 #define EXTRACT_FIELD(src, start, end) \
48 (((src) >> start) & ((1 << (end - start + 1)) - 1))
50 #define MEM_INDEX 0
52 /* is_jmp field values */
53 #define DISAS_JUMP DISAS_TARGET_0 /* only pc was modified dynamically */
54 #define DISAS_UPDATE DISAS_TARGET_1 /* cpu state was modified dynamically */
55 #define DISAS_TB_JUMP DISAS_TARGET_2 /* only pc was modified statically */
57 static TCGv cpu_R[32];
58 static TCGv cpu_pc;
59 static TCGv cpu_ie;
60 static TCGv cpu_icc;
61 static TCGv cpu_dcc;
62 static TCGv cpu_cc;
63 static TCGv cpu_cfg;
64 static TCGv cpu_eba;
65 static TCGv cpu_dc;
66 static TCGv cpu_deba;
67 static TCGv cpu_bp[4];
68 static TCGv cpu_wp[4];
70 #include "exec/gen-icount.h"
72 enum {
73 OP_FMT_RI,
74 OP_FMT_RR,
75 OP_FMT_CR,
76 OP_FMT_I
79 /* This is the state at translation time. */
80 typedef struct DisasContext {
81 target_ulong pc;
83 /* Decoder. */
84 int format;
85 uint32_t ir;
86 uint8_t opcode;
87 uint8_t r0, r1, r2, csr;
88 uint16_t imm5;
89 uint16_t imm16;
90 uint32_t imm26;
92 unsigned int delayed_branch;
93 unsigned int tb_flags, synced_flags; /* tb dependent flags. */
94 int is_jmp;
96 struct TranslationBlock *tb;
97 int singlestep_enabled;
99 uint32_t features;
100 uint8_t num_breakpoints;
101 uint8_t num_watchpoints;
102 } DisasContext;
104 static const char *regnames[] = {
105 "r0", "r1", "r2", "r3", "r4", "r5", "r6", "r7",
106 "r8", "r9", "r10", "r11", "r12", "r13", "r14", "r15",
107 "r16", "r17", "r18", "r19", "r20", "r21", "r22", "r23",
108 "r24", "r25", "r26/gp", "r27/fp", "r28/sp", "r29/ra",
109 "r30/ea", "r31/ba", "bp0", "bp1", "bp2", "bp3", "wp0",
110 "wp1", "wp2", "wp3"
113 static inline int zero_extend(unsigned int val, int width)
115 return val & ((1 << width) - 1);
118 static inline int sign_extend(unsigned int val, int width)
120 int sval;
122 /* LSL. */
123 val <<= 32 - width;
124 sval = val;
125 /* ASR. */
126 sval >>= 32 - width;
128 return sval;
131 static inline void t_gen_raise_exception(DisasContext *dc, uint32_t index)
133 TCGv_i32 tmp = tcg_const_i32(index);
135 gen_helper_raise_exception(cpu_env, tmp);
136 tcg_temp_free_i32(tmp);
139 static inline void t_gen_illegal_insn(DisasContext *dc)
141 tcg_gen_movi_tl(cpu_pc, dc->pc);
142 gen_helper_ill(cpu_env);
145 static inline bool use_goto_tb(DisasContext *dc, target_ulong dest)
147 if (unlikely(dc->singlestep_enabled)) {
148 return false;
151 #ifndef CONFIG_USER_ONLY
152 return (dc->tb->pc & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK);
153 #else
154 return true;
155 #endif
158 static void gen_goto_tb(DisasContext *dc, int n, target_ulong dest)
160 if (use_goto_tb(dc, dest)) {
161 tcg_gen_goto_tb(n);
162 tcg_gen_movi_tl(cpu_pc, dest);
163 tcg_gen_exit_tb(dc->tb, n);
164 } else {
165 tcg_gen_movi_tl(cpu_pc, dest);
166 if (dc->singlestep_enabled) {
167 t_gen_raise_exception(dc, EXCP_DEBUG);
169 tcg_gen_exit_tb(NULL, 0);
173 static void dec_add(DisasContext *dc)
175 if (dc->format == OP_FMT_RI) {
176 if (dc->r0 == R_R0) {
177 if (dc->r1 == R_R0 && dc->imm16 == 0) {
178 LOG_DIS("nop\n");
179 } else {
180 LOG_DIS("mvi r%d, %d\n", dc->r1, sign_extend(dc->imm16, 16));
182 } else {
183 LOG_DIS("addi r%d, r%d, %d\n", dc->r1, dc->r0,
184 sign_extend(dc->imm16, 16));
186 } else {
187 LOG_DIS("add r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
190 if (dc->format == OP_FMT_RI) {
191 tcg_gen_addi_tl(cpu_R[dc->r1], cpu_R[dc->r0],
192 sign_extend(dc->imm16, 16));
193 } else {
194 tcg_gen_add_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
198 static void dec_and(DisasContext *dc)
200 if (dc->format == OP_FMT_RI) {
201 LOG_DIS("andi r%d, r%d, %d\n", dc->r1, dc->r0,
202 zero_extend(dc->imm16, 16));
203 } else {
204 LOG_DIS("and r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
207 if (dc->format == OP_FMT_RI) {
208 tcg_gen_andi_tl(cpu_R[dc->r1], cpu_R[dc->r0],
209 zero_extend(dc->imm16, 16));
210 } else {
211 if (dc->r0 == 0 && dc->r1 == 0 && dc->r2 == 0) {
212 tcg_gen_movi_tl(cpu_pc, dc->pc + 4);
213 gen_helper_hlt(cpu_env);
214 } else {
215 tcg_gen_and_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
220 static void dec_andhi(DisasContext *dc)
222 LOG_DIS("andhi r%d, r%d, %d\n", dc->r1, dc->r0, dc->imm16);
224 tcg_gen_andi_tl(cpu_R[dc->r1], cpu_R[dc->r0], (dc->imm16 << 16));
227 static void dec_b(DisasContext *dc)
229 if (dc->r0 == R_RA) {
230 LOG_DIS("ret\n");
231 } else if (dc->r0 == R_EA) {
232 LOG_DIS("eret\n");
233 } else if (dc->r0 == R_BA) {
234 LOG_DIS("bret\n");
235 } else {
236 LOG_DIS("b r%d\n", dc->r0);
239 /* restore IE.IE in case of an eret */
240 if (dc->r0 == R_EA) {
241 TCGv t0 = tcg_temp_new();
242 TCGLabel *l1 = gen_new_label();
243 tcg_gen_andi_tl(t0, cpu_ie, IE_EIE);
244 tcg_gen_ori_tl(cpu_ie, cpu_ie, IE_IE);
245 tcg_gen_brcondi_tl(TCG_COND_EQ, t0, IE_EIE, l1);
246 tcg_gen_andi_tl(cpu_ie, cpu_ie, ~IE_IE);
247 gen_set_label(l1);
248 tcg_temp_free(t0);
249 } else if (dc->r0 == R_BA) {
250 TCGv t0 = tcg_temp_new();
251 TCGLabel *l1 = gen_new_label();
252 tcg_gen_andi_tl(t0, cpu_ie, IE_BIE);
253 tcg_gen_ori_tl(cpu_ie, cpu_ie, IE_IE);
254 tcg_gen_brcondi_tl(TCG_COND_EQ, t0, IE_BIE, l1);
255 tcg_gen_andi_tl(cpu_ie, cpu_ie, ~IE_IE);
256 gen_set_label(l1);
257 tcg_temp_free(t0);
259 tcg_gen_mov_tl(cpu_pc, cpu_R[dc->r0]);
261 dc->is_jmp = DISAS_JUMP;
264 static void dec_bi(DisasContext *dc)
266 LOG_DIS("bi %d\n", sign_extend(dc->imm26 << 2, 26));
268 gen_goto_tb(dc, 0, dc->pc + (sign_extend(dc->imm26 << 2, 26)));
270 dc->is_jmp = DISAS_TB_JUMP;
273 static inline void gen_cond_branch(DisasContext *dc, int cond)
275 TCGLabel *l1 = gen_new_label();
276 tcg_gen_brcond_tl(cond, cpu_R[dc->r0], cpu_R[dc->r1], l1);
277 gen_goto_tb(dc, 0, dc->pc + 4);
278 gen_set_label(l1);
279 gen_goto_tb(dc, 1, dc->pc + (sign_extend(dc->imm16 << 2, 16)));
280 dc->is_jmp = DISAS_TB_JUMP;
283 static void dec_be(DisasContext *dc)
285 LOG_DIS("be r%d, r%d, %d\n", dc->r1, dc->r0,
286 sign_extend(dc->imm16, 16) * 4);
288 gen_cond_branch(dc, TCG_COND_EQ);
291 static void dec_bg(DisasContext *dc)
293 LOG_DIS("bg r%d, r%d, %d\n", dc->r1, dc->r0,
294 sign_extend(dc->imm16, 16 * 4));
296 gen_cond_branch(dc, TCG_COND_GT);
299 static void dec_bge(DisasContext *dc)
301 LOG_DIS("bge r%d, r%d, %d\n", dc->r1, dc->r0,
302 sign_extend(dc->imm16, 16) * 4);
304 gen_cond_branch(dc, TCG_COND_GE);
307 static void dec_bgeu(DisasContext *dc)
309 LOG_DIS("bgeu r%d, r%d, %d\n", dc->r1, dc->r0,
310 sign_extend(dc->imm16, 16) * 4);
312 gen_cond_branch(dc, TCG_COND_GEU);
315 static void dec_bgu(DisasContext *dc)
317 LOG_DIS("bgu r%d, r%d, %d\n", dc->r1, dc->r0,
318 sign_extend(dc->imm16, 16) * 4);
320 gen_cond_branch(dc, TCG_COND_GTU);
323 static void dec_bne(DisasContext *dc)
325 LOG_DIS("bne r%d, r%d, %d\n", dc->r1, dc->r0,
326 sign_extend(dc->imm16, 16) * 4);
328 gen_cond_branch(dc, TCG_COND_NE);
331 static void dec_call(DisasContext *dc)
333 LOG_DIS("call r%d\n", dc->r0);
335 tcg_gen_movi_tl(cpu_R[R_RA], dc->pc + 4);
336 tcg_gen_mov_tl(cpu_pc, cpu_R[dc->r0]);
338 dc->is_jmp = DISAS_JUMP;
341 static void dec_calli(DisasContext *dc)
343 LOG_DIS("calli %d\n", sign_extend(dc->imm26, 26) * 4);
345 tcg_gen_movi_tl(cpu_R[R_RA], dc->pc + 4);
346 gen_goto_tb(dc, 0, dc->pc + (sign_extend(dc->imm26 << 2, 26)));
348 dc->is_jmp = DISAS_TB_JUMP;
351 static inline void gen_compare(DisasContext *dc, int cond)
353 int i;
355 if (dc->format == OP_FMT_RI) {
356 switch (cond) {
357 case TCG_COND_GEU:
358 case TCG_COND_GTU:
359 i = zero_extend(dc->imm16, 16);
360 break;
361 default:
362 i = sign_extend(dc->imm16, 16);
363 break;
366 tcg_gen_setcondi_tl(cond, cpu_R[dc->r1], cpu_R[dc->r0], i);
367 } else {
368 tcg_gen_setcond_tl(cond, cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
372 static void dec_cmpe(DisasContext *dc)
374 if (dc->format == OP_FMT_RI) {
375 LOG_DIS("cmpei r%d, r%d, %d\n", dc->r1, dc->r0,
376 sign_extend(dc->imm16, 16));
377 } else {
378 LOG_DIS("cmpe r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
381 gen_compare(dc, TCG_COND_EQ);
384 static void dec_cmpg(DisasContext *dc)
386 if (dc->format == OP_FMT_RI) {
387 LOG_DIS("cmpgi r%d, r%d, %d\n", dc->r1, dc->r0,
388 sign_extend(dc->imm16, 16));
389 } else {
390 LOG_DIS("cmpg r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
393 gen_compare(dc, TCG_COND_GT);
396 static void dec_cmpge(DisasContext *dc)
398 if (dc->format == OP_FMT_RI) {
399 LOG_DIS("cmpgei r%d, r%d, %d\n", dc->r1, dc->r0,
400 sign_extend(dc->imm16, 16));
401 } else {
402 LOG_DIS("cmpge r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
405 gen_compare(dc, TCG_COND_GE);
408 static void dec_cmpgeu(DisasContext *dc)
410 if (dc->format == OP_FMT_RI) {
411 LOG_DIS("cmpgeui r%d, r%d, %d\n", dc->r1, dc->r0,
412 zero_extend(dc->imm16, 16));
413 } else {
414 LOG_DIS("cmpgeu r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
417 gen_compare(dc, TCG_COND_GEU);
420 static void dec_cmpgu(DisasContext *dc)
422 if (dc->format == OP_FMT_RI) {
423 LOG_DIS("cmpgui r%d, r%d, %d\n", dc->r1, dc->r0,
424 zero_extend(dc->imm16, 16));
425 } else {
426 LOG_DIS("cmpgu r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
429 gen_compare(dc, TCG_COND_GTU);
432 static void dec_cmpne(DisasContext *dc)
434 if (dc->format == OP_FMT_RI) {
435 LOG_DIS("cmpnei r%d, r%d, %d\n", dc->r1, dc->r0,
436 sign_extend(dc->imm16, 16));
437 } else {
438 LOG_DIS("cmpne r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
441 gen_compare(dc, TCG_COND_NE);
444 static void dec_divu(DisasContext *dc)
446 TCGLabel *l1;
448 LOG_DIS("divu r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
450 if (!(dc->features & LM32_FEATURE_DIVIDE)) {
451 qemu_log_mask(LOG_GUEST_ERROR, "hardware divider is not available\n");
452 t_gen_illegal_insn(dc);
453 return;
456 l1 = gen_new_label();
457 tcg_gen_brcondi_tl(TCG_COND_NE, cpu_R[dc->r1], 0, l1);
458 tcg_gen_movi_tl(cpu_pc, dc->pc);
459 t_gen_raise_exception(dc, EXCP_DIVIDE_BY_ZERO);
460 gen_set_label(l1);
461 tcg_gen_divu_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
464 static void dec_lb(DisasContext *dc)
466 TCGv t0;
468 LOG_DIS("lb r%d, (r%d+%d)\n", dc->r1, dc->r0, dc->imm16);
470 t0 = tcg_temp_new();
471 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
472 tcg_gen_qemu_ld8s(cpu_R[dc->r1], t0, MEM_INDEX);
473 tcg_temp_free(t0);
476 static void dec_lbu(DisasContext *dc)
478 TCGv t0;
480 LOG_DIS("lbu r%d, (r%d+%d)\n", dc->r1, dc->r0, dc->imm16);
482 t0 = tcg_temp_new();
483 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
484 tcg_gen_qemu_ld8u(cpu_R[dc->r1], t0, MEM_INDEX);
485 tcg_temp_free(t0);
488 static void dec_lh(DisasContext *dc)
490 TCGv t0;
492 LOG_DIS("lh r%d, (r%d+%d)\n", dc->r1, dc->r0, dc->imm16);
494 t0 = tcg_temp_new();
495 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
496 tcg_gen_qemu_ld16s(cpu_R[dc->r1], t0, MEM_INDEX);
497 tcg_temp_free(t0);
500 static void dec_lhu(DisasContext *dc)
502 TCGv t0;
504 LOG_DIS("lhu r%d, (r%d+%d)\n", dc->r1, dc->r0, dc->imm16);
506 t0 = tcg_temp_new();
507 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
508 tcg_gen_qemu_ld16u(cpu_R[dc->r1], t0, MEM_INDEX);
509 tcg_temp_free(t0);
512 static void dec_lw(DisasContext *dc)
514 TCGv t0;
516 LOG_DIS("lw r%d, (r%d+%d)\n", dc->r1, dc->r0, sign_extend(dc->imm16, 16));
518 t0 = tcg_temp_new();
519 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
520 tcg_gen_qemu_ld32s(cpu_R[dc->r1], t0, MEM_INDEX);
521 tcg_temp_free(t0);
524 static void dec_modu(DisasContext *dc)
526 TCGLabel *l1;
528 LOG_DIS("modu r%d, r%d, %d\n", dc->r2, dc->r0, dc->r1);
530 if (!(dc->features & LM32_FEATURE_DIVIDE)) {
531 qemu_log_mask(LOG_GUEST_ERROR, "hardware divider is not available\n");
532 t_gen_illegal_insn(dc);
533 return;
536 l1 = gen_new_label();
537 tcg_gen_brcondi_tl(TCG_COND_NE, cpu_R[dc->r1], 0, l1);
538 tcg_gen_movi_tl(cpu_pc, dc->pc);
539 t_gen_raise_exception(dc, EXCP_DIVIDE_BY_ZERO);
540 gen_set_label(l1);
541 tcg_gen_remu_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
544 static void dec_mul(DisasContext *dc)
546 if (dc->format == OP_FMT_RI) {
547 LOG_DIS("muli r%d, r%d, %d\n", dc->r1, dc->r0,
548 sign_extend(dc->imm16, 16));
549 } else {
550 LOG_DIS("mul r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
553 if (!(dc->features & LM32_FEATURE_MULTIPLY)) {
554 qemu_log_mask(LOG_GUEST_ERROR,
555 "hardware multiplier is not available\n");
556 t_gen_illegal_insn(dc);
557 return;
560 if (dc->format == OP_FMT_RI) {
561 tcg_gen_muli_tl(cpu_R[dc->r1], cpu_R[dc->r0],
562 sign_extend(dc->imm16, 16));
563 } else {
564 tcg_gen_mul_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
568 static void dec_nor(DisasContext *dc)
570 if (dc->format == OP_FMT_RI) {
571 LOG_DIS("nori r%d, r%d, %d\n", dc->r1, dc->r0,
572 zero_extend(dc->imm16, 16));
573 } else {
574 LOG_DIS("nor r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
577 if (dc->format == OP_FMT_RI) {
578 TCGv t0 = tcg_temp_new();
579 tcg_gen_movi_tl(t0, zero_extend(dc->imm16, 16));
580 tcg_gen_nor_tl(cpu_R[dc->r1], cpu_R[dc->r0], t0);
581 tcg_temp_free(t0);
582 } else {
583 tcg_gen_nor_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
587 static void dec_or(DisasContext *dc)
589 if (dc->format == OP_FMT_RI) {
590 LOG_DIS("ori r%d, r%d, %d\n", dc->r1, dc->r0,
591 zero_extend(dc->imm16, 16));
592 } else {
593 if (dc->r1 == R_R0) {
594 LOG_DIS("mv r%d, r%d\n", dc->r2, dc->r0);
595 } else {
596 LOG_DIS("or r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
600 if (dc->format == OP_FMT_RI) {
601 tcg_gen_ori_tl(cpu_R[dc->r1], cpu_R[dc->r0],
602 zero_extend(dc->imm16, 16));
603 } else {
604 tcg_gen_or_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
608 static void dec_orhi(DisasContext *dc)
610 if (dc->r0 == R_R0) {
611 LOG_DIS("mvhi r%d, %d\n", dc->r1, dc->imm16);
612 } else {
613 LOG_DIS("orhi r%d, r%d, %d\n", dc->r1, dc->r0, dc->imm16);
616 tcg_gen_ori_tl(cpu_R[dc->r1], cpu_R[dc->r0], (dc->imm16 << 16));
619 static void dec_scall(DisasContext *dc)
621 switch (dc->imm5) {
622 case 2:
623 LOG_DIS("break\n");
624 tcg_gen_movi_tl(cpu_pc, dc->pc);
625 t_gen_raise_exception(dc, EXCP_BREAKPOINT);
626 break;
627 case 7:
628 LOG_DIS("scall\n");
629 tcg_gen_movi_tl(cpu_pc, dc->pc);
630 t_gen_raise_exception(dc, EXCP_SYSTEMCALL);
631 break;
632 default:
633 qemu_log_mask(LOG_GUEST_ERROR, "invalid opcode @0x%x", dc->pc);
634 t_gen_illegal_insn(dc);
635 break;
639 static void dec_rcsr(DisasContext *dc)
641 LOG_DIS("rcsr r%d, %d\n", dc->r2, dc->csr);
643 switch (dc->csr) {
644 case CSR_IE:
645 tcg_gen_mov_tl(cpu_R[dc->r2], cpu_ie);
646 break;
647 case CSR_IM:
648 gen_helper_rcsr_im(cpu_R[dc->r2], cpu_env);
649 break;
650 case CSR_IP:
651 gen_helper_rcsr_ip(cpu_R[dc->r2], cpu_env);
652 break;
653 case CSR_CC:
654 tcg_gen_mov_tl(cpu_R[dc->r2], cpu_cc);
655 break;
656 case CSR_CFG:
657 tcg_gen_mov_tl(cpu_R[dc->r2], cpu_cfg);
658 break;
659 case CSR_EBA:
660 tcg_gen_mov_tl(cpu_R[dc->r2], cpu_eba);
661 break;
662 case CSR_DC:
663 tcg_gen_mov_tl(cpu_R[dc->r2], cpu_dc);
664 break;
665 case CSR_DEBA:
666 tcg_gen_mov_tl(cpu_R[dc->r2], cpu_deba);
667 break;
668 case CSR_JTX:
669 gen_helper_rcsr_jtx(cpu_R[dc->r2], cpu_env);
670 break;
671 case CSR_JRX:
672 gen_helper_rcsr_jrx(cpu_R[dc->r2], cpu_env);
673 break;
674 case CSR_ICC:
675 case CSR_DCC:
676 case CSR_BP0:
677 case CSR_BP1:
678 case CSR_BP2:
679 case CSR_BP3:
680 case CSR_WP0:
681 case CSR_WP1:
682 case CSR_WP2:
683 case CSR_WP3:
684 qemu_log_mask(LOG_GUEST_ERROR, "invalid read access csr=%x\n", dc->csr);
685 break;
686 default:
687 qemu_log_mask(LOG_GUEST_ERROR, "read_csr: unknown csr=%x\n", dc->csr);
688 break;
692 static void dec_sb(DisasContext *dc)
694 TCGv t0;
696 LOG_DIS("sb (r%d+%d), r%d\n", dc->r0, dc->imm16, dc->r1);
698 t0 = tcg_temp_new();
699 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
700 tcg_gen_qemu_st8(cpu_R[dc->r1], t0, MEM_INDEX);
701 tcg_temp_free(t0);
704 static void dec_sextb(DisasContext *dc)
706 LOG_DIS("sextb r%d, r%d\n", dc->r2, dc->r0);
708 if (!(dc->features & LM32_FEATURE_SIGN_EXTEND)) {
709 qemu_log_mask(LOG_GUEST_ERROR,
710 "hardware sign extender is not available\n");
711 t_gen_illegal_insn(dc);
712 return;
715 tcg_gen_ext8s_tl(cpu_R[dc->r2], cpu_R[dc->r0]);
718 static void dec_sexth(DisasContext *dc)
720 LOG_DIS("sexth r%d, r%d\n", dc->r2, dc->r0);
722 if (!(dc->features & LM32_FEATURE_SIGN_EXTEND)) {
723 qemu_log_mask(LOG_GUEST_ERROR,
724 "hardware sign extender is not available\n");
725 t_gen_illegal_insn(dc);
726 return;
729 tcg_gen_ext16s_tl(cpu_R[dc->r2], cpu_R[dc->r0]);
732 static void dec_sh(DisasContext *dc)
734 TCGv t0;
736 LOG_DIS("sh (r%d+%d), r%d\n", dc->r0, dc->imm16, dc->r1);
738 t0 = tcg_temp_new();
739 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
740 tcg_gen_qemu_st16(cpu_R[dc->r1], t0, MEM_INDEX);
741 tcg_temp_free(t0);
744 static void dec_sl(DisasContext *dc)
746 if (dc->format == OP_FMT_RI) {
747 LOG_DIS("sli r%d, r%d, %d\n", dc->r1, dc->r0, dc->imm5);
748 } else {
749 LOG_DIS("sl r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
752 if (!(dc->features & LM32_FEATURE_SHIFT)) {
753 qemu_log_mask(LOG_GUEST_ERROR, "hardware shifter is not available\n");
754 t_gen_illegal_insn(dc);
755 return;
758 if (dc->format == OP_FMT_RI) {
759 tcg_gen_shli_tl(cpu_R[dc->r1], cpu_R[dc->r0], dc->imm5);
760 } else {
761 TCGv t0 = tcg_temp_new();
762 tcg_gen_andi_tl(t0, cpu_R[dc->r1], 0x1f);
763 tcg_gen_shl_tl(cpu_R[dc->r2], cpu_R[dc->r0], t0);
764 tcg_temp_free(t0);
768 static void dec_sr(DisasContext *dc)
770 if (dc->format == OP_FMT_RI) {
771 LOG_DIS("sri r%d, r%d, %d\n", dc->r1, dc->r0, dc->imm5);
772 } else {
773 LOG_DIS("sr r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
776 /* The real CPU (w/o hardware shifter) only supports right shift by exactly
777 * one bit */
778 if (dc->format == OP_FMT_RI) {
779 if (!(dc->features & LM32_FEATURE_SHIFT) && (dc->imm5 != 1)) {
780 qemu_log_mask(LOG_GUEST_ERROR,
781 "hardware shifter is not available\n");
782 t_gen_illegal_insn(dc);
783 return;
785 tcg_gen_sari_tl(cpu_R[dc->r1], cpu_R[dc->r0], dc->imm5);
786 } else {
787 TCGLabel *l1 = gen_new_label();
788 TCGLabel *l2 = gen_new_label();
789 TCGv t0 = tcg_temp_local_new();
790 tcg_gen_andi_tl(t0, cpu_R[dc->r1], 0x1f);
792 if (!(dc->features & LM32_FEATURE_SHIFT)) {
793 tcg_gen_brcondi_tl(TCG_COND_EQ, t0, 1, l1);
794 t_gen_illegal_insn(dc);
795 tcg_gen_br(l2);
798 gen_set_label(l1);
799 tcg_gen_sar_tl(cpu_R[dc->r2], cpu_R[dc->r0], t0);
800 gen_set_label(l2);
802 tcg_temp_free(t0);
806 static void dec_sru(DisasContext *dc)
808 if (dc->format == OP_FMT_RI) {
809 LOG_DIS("srui r%d, r%d, %d\n", dc->r1, dc->r0, dc->imm5);
810 } else {
811 LOG_DIS("sru r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
814 if (dc->format == OP_FMT_RI) {
815 if (!(dc->features & LM32_FEATURE_SHIFT) && (dc->imm5 != 1)) {
816 qemu_log_mask(LOG_GUEST_ERROR,
817 "hardware shifter is not available\n");
818 t_gen_illegal_insn(dc);
819 return;
821 tcg_gen_shri_tl(cpu_R[dc->r1], cpu_R[dc->r0], dc->imm5);
822 } else {
823 TCGLabel *l1 = gen_new_label();
824 TCGLabel *l2 = gen_new_label();
825 TCGv t0 = tcg_temp_local_new();
826 tcg_gen_andi_tl(t0, cpu_R[dc->r1], 0x1f);
828 if (!(dc->features & LM32_FEATURE_SHIFT)) {
829 tcg_gen_brcondi_tl(TCG_COND_EQ, t0, 1, l1);
830 t_gen_illegal_insn(dc);
831 tcg_gen_br(l2);
834 gen_set_label(l1);
835 tcg_gen_shr_tl(cpu_R[dc->r2], cpu_R[dc->r0], t0);
836 gen_set_label(l2);
838 tcg_temp_free(t0);
842 static void dec_sub(DisasContext *dc)
844 LOG_DIS("sub r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
846 tcg_gen_sub_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
849 static void dec_sw(DisasContext *dc)
851 TCGv t0;
853 LOG_DIS("sw (r%d+%d), r%d\n", dc->r0, sign_extend(dc->imm16, 16), dc->r1);
855 t0 = tcg_temp_new();
856 tcg_gen_addi_tl(t0, cpu_R[dc->r0], sign_extend(dc->imm16, 16));
857 tcg_gen_qemu_st32(cpu_R[dc->r1], t0, MEM_INDEX);
858 tcg_temp_free(t0);
861 static void dec_user(DisasContext *dc)
863 LOG_DIS("user");
865 qemu_log_mask(LOG_GUEST_ERROR, "user instruction undefined\n");
866 t_gen_illegal_insn(dc);
869 static void dec_wcsr(DisasContext *dc)
871 int no;
873 LOG_DIS("wcsr %d, r%d\n", dc->csr, dc->r1);
875 switch (dc->csr) {
876 case CSR_IE:
877 tcg_gen_mov_tl(cpu_ie, cpu_R[dc->r1]);
878 tcg_gen_movi_tl(cpu_pc, dc->pc + 4);
879 dc->is_jmp = DISAS_UPDATE;
880 break;
881 case CSR_IM:
882 /* mark as an io operation because it could cause an interrupt */
883 if (tb_cflags(dc->tb) & CF_USE_ICOUNT) {
884 gen_io_start();
886 gen_helper_wcsr_im(cpu_env, cpu_R[dc->r1]);
887 tcg_gen_movi_tl(cpu_pc, dc->pc + 4);
888 if (tb_cflags(dc->tb) & CF_USE_ICOUNT) {
889 gen_io_end();
891 dc->is_jmp = DISAS_UPDATE;
892 break;
893 case CSR_IP:
894 /* mark as an io operation because it could cause an interrupt */
895 if (tb_cflags(dc->tb) & CF_USE_ICOUNT) {
896 gen_io_start();
898 gen_helper_wcsr_ip(cpu_env, cpu_R[dc->r1]);
899 tcg_gen_movi_tl(cpu_pc, dc->pc + 4);
900 if (tb_cflags(dc->tb) & CF_USE_ICOUNT) {
901 gen_io_end();
903 dc->is_jmp = DISAS_UPDATE;
904 break;
905 case CSR_ICC:
906 /* TODO */
907 break;
908 case CSR_DCC:
909 /* TODO */
910 break;
911 case CSR_EBA:
912 tcg_gen_mov_tl(cpu_eba, cpu_R[dc->r1]);
913 break;
914 case CSR_DEBA:
915 tcg_gen_mov_tl(cpu_deba, cpu_R[dc->r1]);
916 break;
917 case CSR_JTX:
918 gen_helper_wcsr_jtx(cpu_env, cpu_R[dc->r1]);
919 break;
920 case CSR_JRX:
921 gen_helper_wcsr_jrx(cpu_env, cpu_R[dc->r1]);
922 break;
923 case CSR_DC:
924 gen_helper_wcsr_dc(cpu_env, cpu_R[dc->r1]);
925 break;
926 case CSR_BP0:
927 case CSR_BP1:
928 case CSR_BP2:
929 case CSR_BP3:
930 no = dc->csr - CSR_BP0;
931 if (dc->num_breakpoints <= no) {
932 qemu_log_mask(LOG_GUEST_ERROR,
933 "breakpoint #%i is not available\n", no);
934 t_gen_illegal_insn(dc);
935 break;
937 gen_helper_wcsr_bp(cpu_env, cpu_R[dc->r1], tcg_const_i32(no));
938 break;
939 case CSR_WP0:
940 case CSR_WP1:
941 case CSR_WP2:
942 case CSR_WP3:
943 no = dc->csr - CSR_WP0;
944 if (dc->num_watchpoints <= no) {
945 qemu_log_mask(LOG_GUEST_ERROR,
946 "watchpoint #%i is not available\n", no);
947 t_gen_illegal_insn(dc);
948 break;
950 gen_helper_wcsr_wp(cpu_env, cpu_R[dc->r1], tcg_const_i32(no));
951 break;
952 case CSR_CC:
953 case CSR_CFG:
954 qemu_log_mask(LOG_GUEST_ERROR, "invalid write access csr=%x\n",
955 dc->csr);
956 break;
957 default:
958 qemu_log_mask(LOG_GUEST_ERROR, "write_csr: unknown csr=%x\n",
959 dc->csr);
960 break;
964 static void dec_xnor(DisasContext *dc)
966 if (dc->format == OP_FMT_RI) {
967 LOG_DIS("xnori r%d, r%d, %d\n", dc->r1, dc->r0,
968 zero_extend(dc->imm16, 16));
969 } else {
970 if (dc->r1 == R_R0) {
971 LOG_DIS("not r%d, r%d\n", dc->r2, dc->r0);
972 } else {
973 LOG_DIS("xnor r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
977 if (dc->format == OP_FMT_RI) {
978 tcg_gen_xori_tl(cpu_R[dc->r1], cpu_R[dc->r0],
979 zero_extend(dc->imm16, 16));
980 tcg_gen_not_tl(cpu_R[dc->r1], cpu_R[dc->r1]);
981 } else {
982 tcg_gen_eqv_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
986 static void dec_xor(DisasContext *dc)
988 if (dc->format == OP_FMT_RI) {
989 LOG_DIS("xori r%d, r%d, %d\n", dc->r1, dc->r0,
990 zero_extend(dc->imm16, 16));
991 } else {
992 LOG_DIS("xor r%d, r%d, r%d\n", dc->r2, dc->r0, dc->r1);
995 if (dc->format == OP_FMT_RI) {
996 tcg_gen_xori_tl(cpu_R[dc->r1], cpu_R[dc->r0],
997 zero_extend(dc->imm16, 16));
998 } else {
999 tcg_gen_xor_tl(cpu_R[dc->r2], cpu_R[dc->r0], cpu_R[dc->r1]);
1003 static void dec_ill(DisasContext *dc)
1005 qemu_log_mask(LOG_GUEST_ERROR, "invalid opcode 0x%02x\n", dc->opcode);
1006 t_gen_illegal_insn(dc);
1009 typedef void (*DecoderInfo)(DisasContext *dc);
1010 static const DecoderInfo decinfo[] = {
1011 dec_sru, dec_nor, dec_mul, dec_sh, dec_lb, dec_sr, dec_xor, dec_lh,
1012 dec_and, dec_xnor, dec_lw, dec_lhu, dec_sb, dec_add, dec_or, dec_sl,
1013 dec_lbu, dec_be, dec_bg, dec_bge, dec_bgeu, dec_bgu, dec_sw, dec_bne,
1014 dec_andhi, dec_cmpe, dec_cmpg, dec_cmpge, dec_cmpgeu, dec_cmpgu, dec_orhi,
1015 dec_cmpne,
1016 dec_sru, dec_nor, dec_mul, dec_divu, dec_rcsr, dec_sr, dec_xor, dec_ill,
1017 dec_and, dec_xnor, dec_ill, dec_scall, dec_sextb, dec_add, dec_or, dec_sl,
1018 dec_b, dec_modu, dec_sub, dec_user, dec_wcsr, dec_ill, dec_call, dec_sexth,
1019 dec_bi, dec_cmpe, dec_cmpg, dec_cmpge, dec_cmpgeu, dec_cmpgu, dec_calli,
1020 dec_cmpne
1023 static inline void decode(DisasContext *dc, uint32_t ir)
1025 dc->ir = ir;
1026 LOG_DIS("%8.8x\t", dc->ir);
1028 dc->opcode = EXTRACT_FIELD(ir, 26, 31);
1030 dc->imm5 = EXTRACT_FIELD(ir, 0, 4);
1031 dc->imm16 = EXTRACT_FIELD(ir, 0, 15);
1032 dc->imm26 = EXTRACT_FIELD(ir, 0, 25);
1034 dc->csr = EXTRACT_FIELD(ir, 21, 25);
1035 dc->r0 = EXTRACT_FIELD(ir, 21, 25);
1036 dc->r1 = EXTRACT_FIELD(ir, 16, 20);
1037 dc->r2 = EXTRACT_FIELD(ir, 11, 15);
1039 /* bit 31 seems to indicate insn type. */
1040 if (ir & (1 << 31)) {
1041 dc->format = OP_FMT_RR;
1042 } else {
1043 dc->format = OP_FMT_RI;
1046 assert(ARRAY_SIZE(decinfo) == 64);
1047 assert(dc->opcode < 64);
1049 decinfo[dc->opcode](dc);
1052 /* generate intermediate code for basic block 'tb'. */
1053 void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
1055 CPULM32State *env = cs->env_ptr;
1056 LM32CPU *cpu = env_archcpu(env);
1057 struct DisasContext ctx, *dc = &ctx;
1058 uint32_t pc_start;
1059 uint32_t page_start;
1060 int num_insns;
1062 pc_start = tb->pc;
1063 dc->features = cpu->features;
1064 dc->num_breakpoints = cpu->num_breakpoints;
1065 dc->num_watchpoints = cpu->num_watchpoints;
1066 dc->tb = tb;
1068 dc->is_jmp = DISAS_NEXT;
1069 dc->pc = pc_start;
1070 dc->singlestep_enabled = cs->singlestep_enabled;
1072 if (pc_start & 3) {
1073 qemu_log_mask(LOG_GUEST_ERROR,
1074 "unaligned PC=%x. Ignoring lowest bits.\n", pc_start);
1075 pc_start &= ~3;
1078 page_start = pc_start & TARGET_PAGE_MASK;
1079 num_insns = 0;
1081 gen_tb_start(tb);
1082 do {
1083 tcg_gen_insn_start(dc->pc);
1084 num_insns++;
1086 if (unlikely(cpu_breakpoint_test(cs, dc->pc, BP_ANY))) {
1087 tcg_gen_movi_tl(cpu_pc, dc->pc);
1088 t_gen_raise_exception(dc, EXCP_DEBUG);
1089 dc->is_jmp = DISAS_UPDATE;
1090 /* The address covered by the breakpoint must be included in
1091 [tb->pc, tb->pc + tb->size) in order to for it to be
1092 properly cleared -- thus we increment the PC here so that
1093 the logic setting tb->size below does the right thing. */
1094 dc->pc += 4;
1095 break;
1098 /* Pretty disas. */
1099 LOG_DIS("%8.8x:\t", dc->pc);
1101 if (num_insns == max_insns && (tb_cflags(tb) & CF_LAST_IO)) {
1102 gen_io_start();
1105 decode(dc, cpu_ldl_code(env, dc->pc));
1106 dc->pc += 4;
1107 } while (!dc->is_jmp
1108 && !tcg_op_buf_full()
1109 && !cs->singlestep_enabled
1110 && !singlestep
1111 && (dc->pc - page_start < TARGET_PAGE_SIZE)
1112 && num_insns < max_insns);
1114 if (tb_cflags(tb) & CF_LAST_IO) {
1115 gen_io_end();
1118 if (unlikely(cs->singlestep_enabled)) {
1119 if (dc->is_jmp == DISAS_NEXT) {
1120 tcg_gen_movi_tl(cpu_pc, dc->pc);
1122 t_gen_raise_exception(dc, EXCP_DEBUG);
1123 } else {
1124 switch (dc->is_jmp) {
1125 case DISAS_NEXT:
1126 gen_goto_tb(dc, 1, dc->pc);
1127 break;
1128 default:
1129 case DISAS_JUMP:
1130 case DISAS_UPDATE:
1131 /* indicate that the hash table must be used
1132 to find the next TB */
1133 tcg_gen_exit_tb(NULL, 0);
1134 break;
1135 case DISAS_TB_JUMP:
1136 /* nothing more to generate */
1137 break;
1141 gen_tb_end(tb, num_insns);
1143 tb->size = dc->pc - pc_start;
1144 tb->icount = num_insns;
1146 #ifdef DEBUG_DISAS
1147 if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)
1148 && qemu_log_in_addr_range(pc_start)) {
1149 qemu_log_lock();
1150 qemu_log("\n");
1151 log_target_disas(cs, pc_start, dc->pc - pc_start);
1152 qemu_log_unlock();
1154 #endif
1157 void lm32_cpu_dump_state(CPUState *cs, FILE *f, int flags)
1159 LM32CPU *cpu = LM32_CPU(cs);
1160 CPULM32State *env = &cpu->env;
1161 int i;
1163 if (!env) {
1164 return;
1167 qemu_fprintf(f, "IN: PC=%x %s\n",
1168 env->pc, lookup_symbol(env->pc));
1170 qemu_fprintf(f, "ie=%8.8x (IE=%x EIE=%x BIE=%x) im=%8.8x ip=%8.8x\n",
1171 env->ie,
1172 (env->ie & IE_IE) ? 1 : 0,
1173 (env->ie & IE_EIE) ? 1 : 0,
1174 (env->ie & IE_BIE) ? 1 : 0,
1175 lm32_pic_get_im(env->pic_state),
1176 lm32_pic_get_ip(env->pic_state));
1177 qemu_fprintf(f, "eba=%8.8x deba=%8.8x\n",
1178 env->eba,
1179 env->deba);
1181 for (i = 0; i < 32; i++) {
1182 qemu_fprintf(f, "r%2.2d=%8.8x ", i, env->regs[i]);
1183 if ((i + 1) % 4 == 0) {
1184 qemu_fprintf(f, "\n");
1187 qemu_fprintf(f, "\n\n");
1190 void restore_state_to_opc(CPULM32State *env, TranslationBlock *tb,
1191 target_ulong *data)
1193 env->pc = data[0];
1196 void lm32_translate_init(void)
1198 int i;
1200 for (i = 0; i < ARRAY_SIZE(cpu_R); i++) {
1201 cpu_R[i] = tcg_global_mem_new(cpu_env,
1202 offsetof(CPULM32State, regs[i]),
1203 regnames[i]);
1206 for (i = 0; i < ARRAY_SIZE(cpu_bp); i++) {
1207 cpu_bp[i] = tcg_global_mem_new(cpu_env,
1208 offsetof(CPULM32State, bp[i]),
1209 regnames[32+i]);
1212 for (i = 0; i < ARRAY_SIZE(cpu_wp); i++) {
1213 cpu_wp[i] = tcg_global_mem_new(cpu_env,
1214 offsetof(CPULM32State, wp[i]),
1215 regnames[36+i]);
1218 cpu_pc = tcg_global_mem_new(cpu_env,
1219 offsetof(CPULM32State, pc),
1220 "pc");
1221 cpu_ie = tcg_global_mem_new(cpu_env,
1222 offsetof(CPULM32State, ie),
1223 "ie");
1224 cpu_icc = tcg_global_mem_new(cpu_env,
1225 offsetof(CPULM32State, icc),
1226 "icc");
1227 cpu_dcc = tcg_global_mem_new(cpu_env,
1228 offsetof(CPULM32State, dcc),
1229 "dcc");
1230 cpu_cc = tcg_global_mem_new(cpu_env,
1231 offsetof(CPULM32State, cc),
1232 "cc");
1233 cpu_cfg = tcg_global_mem_new(cpu_env,
1234 offsetof(CPULM32State, cfg),
1235 "cfg");
1236 cpu_eba = tcg_global_mem_new(cpu_env,
1237 offsetof(CPULM32State, eba),
1238 "eba");
1239 cpu_dc = tcg_global_mem_new(cpu_env,
1240 offsetof(CPULM32State, dc),
1241 "dc");
1242 cpu_deba = tcg_global_mem_new(cpu_env,
1243 offsetof(CPULM32State, deba),
1244 "deba");