Merge remote-tracking branch 'remotes/armbru/tags/pull-error-2017-01-19' into staging
[qemu/ar7.git] / target / s390x / translate.c
blob01c62176bf702be95c91d30977eac88e42a72be9
1 /*
2 * S/390 translation
4 * Copyright (c) 2009 Ulrich Hecht
5 * Copyright (c) 2010 Alexander Graf
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2 of the License, or (at your option) any later version.
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
21 /* #define DEBUG_INLINE_BRANCHES */
22 #define S390X_DEBUG_DISAS
23 /* #define S390X_DEBUG_DISAS_VERBOSE */
25 #ifdef S390X_DEBUG_DISAS_VERBOSE
26 # define LOG_DISAS(...) qemu_log(__VA_ARGS__)
27 #else
28 # define LOG_DISAS(...) do { } while (0)
29 #endif
31 #include "qemu/osdep.h"
32 #include "cpu.h"
33 #include "disas/disas.h"
34 #include "exec/exec-all.h"
35 #include "tcg-op.h"
36 #include "qemu/log.h"
37 #include "qemu/host-utils.h"
38 #include "exec/cpu_ldst.h"
40 /* global register indexes */
41 static TCGv_env cpu_env;
43 #include "exec/gen-icount.h"
44 #include "exec/helper-proto.h"
45 #include "exec/helper-gen.h"
47 #include "trace-tcg.h"
48 #include "exec/log.h"
51 /* Information that (most) every instruction needs to manipulate. */
52 typedef struct DisasContext DisasContext;
53 typedef struct DisasInsn DisasInsn;
54 typedef struct DisasFields DisasFields;
56 struct DisasContext {
57 struct TranslationBlock *tb;
58 const DisasInsn *insn;
59 DisasFields *fields;
60 uint64_t pc, next_pc;
61 enum cc_op cc_op;
62 bool singlestep_enabled;
65 /* Information carried about a condition to be evaluated. */
66 typedef struct {
67 TCGCond cond:8;
68 bool is_64;
69 bool g1;
70 bool g2;
71 union {
72 struct { TCGv_i64 a, b; } s64;
73 struct { TCGv_i32 a, b; } s32;
74 } u;
75 } DisasCompare;
77 #define DISAS_EXCP 4
79 #ifdef DEBUG_INLINE_BRANCHES
80 static uint64_t inline_branch_hit[CC_OP_MAX];
81 static uint64_t inline_branch_miss[CC_OP_MAX];
82 #endif
84 static uint64_t pc_to_link_info(DisasContext *s, uint64_t pc)
86 if (!(s->tb->flags & FLAG_MASK_64)) {
87 if (s->tb->flags & FLAG_MASK_32) {
88 return pc | 0x80000000;
91 return pc;
94 void s390_cpu_dump_state(CPUState *cs, FILE *f, fprintf_function cpu_fprintf,
95 int flags)
97 S390CPU *cpu = S390_CPU(cs);
98 CPUS390XState *env = &cpu->env;
99 int i;
101 if (env->cc_op > 3) {
102 cpu_fprintf(f, "PSW=mask %016" PRIx64 " addr %016" PRIx64 " cc %15s\n",
103 env->psw.mask, env->psw.addr, cc_name(env->cc_op));
104 } else {
105 cpu_fprintf(f, "PSW=mask %016" PRIx64 " addr %016" PRIx64 " cc %02x\n",
106 env->psw.mask, env->psw.addr, env->cc_op);
109 for (i = 0; i < 16; i++) {
110 cpu_fprintf(f, "R%02d=%016" PRIx64, i, env->regs[i]);
111 if ((i % 4) == 3) {
112 cpu_fprintf(f, "\n");
113 } else {
114 cpu_fprintf(f, " ");
118 for (i = 0; i < 16; i++) {
119 cpu_fprintf(f, "F%02d=%016" PRIx64, i, get_freg(env, i)->ll);
120 if ((i % 4) == 3) {
121 cpu_fprintf(f, "\n");
122 } else {
123 cpu_fprintf(f, " ");
127 for (i = 0; i < 32; i++) {
128 cpu_fprintf(f, "V%02d=%016" PRIx64 "%016" PRIx64, i,
129 env->vregs[i][0].ll, env->vregs[i][1].ll);
130 cpu_fprintf(f, (i % 2) ? "\n" : " ");
133 #ifndef CONFIG_USER_ONLY
134 for (i = 0; i < 16; i++) {
135 cpu_fprintf(f, "C%02d=%016" PRIx64, i, env->cregs[i]);
136 if ((i % 4) == 3) {
137 cpu_fprintf(f, "\n");
138 } else {
139 cpu_fprintf(f, " ");
142 #endif
144 #ifdef DEBUG_INLINE_BRANCHES
145 for (i = 0; i < CC_OP_MAX; i++) {
146 cpu_fprintf(f, " %15s = %10ld\t%10ld\n", cc_name(i),
147 inline_branch_miss[i], inline_branch_hit[i]);
149 #endif
151 cpu_fprintf(f, "\n");
154 static TCGv_i64 psw_addr;
155 static TCGv_i64 psw_mask;
156 static TCGv_i64 gbea;
158 static TCGv_i32 cc_op;
159 static TCGv_i64 cc_src;
160 static TCGv_i64 cc_dst;
161 static TCGv_i64 cc_vr;
163 static char cpu_reg_names[32][4];
164 static TCGv_i64 regs[16];
165 static TCGv_i64 fregs[16];
167 void s390x_translate_init(void)
169 int i;
171 cpu_env = tcg_global_reg_new_ptr(TCG_AREG0, "env");
172 tcg_ctx.tcg_env = cpu_env;
173 psw_addr = tcg_global_mem_new_i64(cpu_env,
174 offsetof(CPUS390XState, psw.addr),
175 "psw_addr");
176 psw_mask = tcg_global_mem_new_i64(cpu_env,
177 offsetof(CPUS390XState, psw.mask),
178 "psw_mask");
179 gbea = tcg_global_mem_new_i64(cpu_env,
180 offsetof(CPUS390XState, gbea),
181 "gbea");
183 cc_op = tcg_global_mem_new_i32(cpu_env, offsetof(CPUS390XState, cc_op),
184 "cc_op");
185 cc_src = tcg_global_mem_new_i64(cpu_env, offsetof(CPUS390XState, cc_src),
186 "cc_src");
187 cc_dst = tcg_global_mem_new_i64(cpu_env, offsetof(CPUS390XState, cc_dst),
188 "cc_dst");
189 cc_vr = tcg_global_mem_new_i64(cpu_env, offsetof(CPUS390XState, cc_vr),
190 "cc_vr");
192 for (i = 0; i < 16; i++) {
193 snprintf(cpu_reg_names[i], sizeof(cpu_reg_names[0]), "r%d", i);
194 regs[i] = tcg_global_mem_new(cpu_env,
195 offsetof(CPUS390XState, regs[i]),
196 cpu_reg_names[i]);
199 for (i = 0; i < 16; i++) {
200 snprintf(cpu_reg_names[i + 16], sizeof(cpu_reg_names[0]), "f%d", i);
201 fregs[i] = tcg_global_mem_new(cpu_env,
202 offsetof(CPUS390XState, vregs[i][0].d),
203 cpu_reg_names[i + 16]);
207 static TCGv_i64 load_reg(int reg)
209 TCGv_i64 r = tcg_temp_new_i64();
210 tcg_gen_mov_i64(r, regs[reg]);
211 return r;
214 static TCGv_i64 load_freg32_i64(int reg)
216 TCGv_i64 r = tcg_temp_new_i64();
217 tcg_gen_shri_i64(r, fregs[reg], 32);
218 return r;
221 static void store_reg(int reg, TCGv_i64 v)
223 tcg_gen_mov_i64(regs[reg], v);
226 static void store_freg(int reg, TCGv_i64 v)
228 tcg_gen_mov_i64(fregs[reg], v);
231 static void store_reg32_i64(int reg, TCGv_i64 v)
233 /* 32 bit register writes keep the upper half */
234 tcg_gen_deposit_i64(regs[reg], regs[reg], v, 0, 32);
237 static void store_reg32h_i64(int reg, TCGv_i64 v)
239 tcg_gen_deposit_i64(regs[reg], regs[reg], v, 32, 32);
242 static void store_freg32_i64(int reg, TCGv_i64 v)
244 tcg_gen_deposit_i64(fregs[reg], fregs[reg], v, 32, 32);
247 static void return_low128(TCGv_i64 dest)
249 tcg_gen_ld_i64(dest, cpu_env, offsetof(CPUS390XState, retxl));
252 static void update_psw_addr(DisasContext *s)
254 /* psw.addr */
255 tcg_gen_movi_i64(psw_addr, s->pc);
258 static void per_branch(DisasContext *s, bool to_next)
260 #ifndef CONFIG_USER_ONLY
261 tcg_gen_movi_i64(gbea, s->pc);
263 if (s->tb->flags & FLAG_MASK_PER) {
264 TCGv_i64 next_pc = to_next ? tcg_const_i64(s->next_pc) : psw_addr;
265 gen_helper_per_branch(cpu_env, gbea, next_pc);
266 if (to_next) {
267 tcg_temp_free_i64(next_pc);
270 #endif
273 static void per_branch_cond(DisasContext *s, TCGCond cond,
274 TCGv_i64 arg1, TCGv_i64 arg2)
276 #ifndef CONFIG_USER_ONLY
277 if (s->tb->flags & FLAG_MASK_PER) {
278 TCGLabel *lab = gen_new_label();
279 tcg_gen_brcond_i64(tcg_invert_cond(cond), arg1, arg2, lab);
281 tcg_gen_movi_i64(gbea, s->pc);
282 gen_helper_per_branch(cpu_env, gbea, psw_addr);
284 gen_set_label(lab);
285 } else {
286 TCGv_i64 pc = tcg_const_i64(s->pc);
287 tcg_gen_movcond_i64(cond, gbea, arg1, arg2, gbea, pc);
288 tcg_temp_free_i64(pc);
290 #endif
293 static void per_breaking_event(DisasContext *s)
295 tcg_gen_movi_i64(gbea, s->pc);
298 static void update_cc_op(DisasContext *s)
300 if (s->cc_op != CC_OP_DYNAMIC && s->cc_op != CC_OP_STATIC) {
301 tcg_gen_movi_i32(cc_op, s->cc_op);
305 static void potential_page_fault(DisasContext *s)
307 update_psw_addr(s);
308 update_cc_op(s);
311 static inline uint64_t ld_code2(CPUS390XState *env, uint64_t pc)
313 return (uint64_t)cpu_lduw_code(env, pc);
316 static inline uint64_t ld_code4(CPUS390XState *env, uint64_t pc)
318 return (uint64_t)(uint32_t)cpu_ldl_code(env, pc);
321 static int get_mem_index(DisasContext *s)
323 switch (s->tb->flags & FLAG_MASK_ASC) {
324 case PSW_ASC_PRIMARY >> 32:
325 return 0;
326 case PSW_ASC_SECONDARY >> 32:
327 return 1;
328 case PSW_ASC_HOME >> 32:
329 return 2;
330 default:
331 tcg_abort();
332 break;
336 static void gen_exception(int excp)
338 TCGv_i32 tmp = tcg_const_i32(excp);
339 gen_helper_exception(cpu_env, tmp);
340 tcg_temp_free_i32(tmp);
343 static void gen_program_exception(DisasContext *s, int code)
345 TCGv_i32 tmp;
347 /* Remember what pgm exeption this was. */
348 tmp = tcg_const_i32(code);
349 tcg_gen_st_i32(tmp, cpu_env, offsetof(CPUS390XState, int_pgm_code));
350 tcg_temp_free_i32(tmp);
352 tmp = tcg_const_i32(s->next_pc - s->pc);
353 tcg_gen_st_i32(tmp, cpu_env, offsetof(CPUS390XState, int_pgm_ilen));
354 tcg_temp_free_i32(tmp);
356 /* Advance past instruction. */
357 s->pc = s->next_pc;
358 update_psw_addr(s);
360 /* Save off cc. */
361 update_cc_op(s);
363 /* Trigger exception. */
364 gen_exception(EXCP_PGM);
367 static inline void gen_illegal_opcode(DisasContext *s)
369 gen_program_exception(s, PGM_OPERATION);
372 static inline void gen_trap(DisasContext *s)
374 TCGv_i32 t;
376 /* Set DXC to 0xff. */
377 t = tcg_temp_new_i32();
378 tcg_gen_ld_i32(t, cpu_env, offsetof(CPUS390XState, fpc));
379 tcg_gen_ori_i32(t, t, 0xff00);
380 tcg_gen_st_i32(t, cpu_env, offsetof(CPUS390XState, fpc));
381 tcg_temp_free_i32(t);
383 gen_program_exception(s, PGM_DATA);
386 #ifndef CONFIG_USER_ONLY
387 static void check_privileged(DisasContext *s)
389 if (s->tb->flags & (PSW_MASK_PSTATE >> 32)) {
390 gen_program_exception(s, PGM_PRIVILEGED);
393 #endif
395 static TCGv_i64 get_address(DisasContext *s, int x2, int b2, int d2)
397 TCGv_i64 tmp = tcg_temp_new_i64();
398 bool need_31 = !(s->tb->flags & FLAG_MASK_64);
400 /* Note that d2 is limited to 20 bits, signed. If we crop negative
401 displacements early we create larger immedate addends. */
403 /* Note that addi optimizes the imm==0 case. */
404 if (b2 && x2) {
405 tcg_gen_add_i64(tmp, regs[b2], regs[x2]);
406 tcg_gen_addi_i64(tmp, tmp, d2);
407 } else if (b2) {
408 tcg_gen_addi_i64(tmp, regs[b2], d2);
409 } else if (x2) {
410 tcg_gen_addi_i64(tmp, regs[x2], d2);
411 } else {
412 if (need_31) {
413 d2 &= 0x7fffffff;
414 need_31 = false;
416 tcg_gen_movi_i64(tmp, d2);
418 if (need_31) {
419 tcg_gen_andi_i64(tmp, tmp, 0x7fffffff);
422 return tmp;
425 static inline bool live_cc_data(DisasContext *s)
427 return (s->cc_op != CC_OP_DYNAMIC
428 && s->cc_op != CC_OP_STATIC
429 && s->cc_op > 3);
432 static inline void gen_op_movi_cc(DisasContext *s, uint32_t val)
434 if (live_cc_data(s)) {
435 tcg_gen_discard_i64(cc_src);
436 tcg_gen_discard_i64(cc_dst);
437 tcg_gen_discard_i64(cc_vr);
439 s->cc_op = CC_OP_CONST0 + val;
442 static void gen_op_update1_cc_i64(DisasContext *s, enum cc_op op, TCGv_i64 dst)
444 if (live_cc_data(s)) {
445 tcg_gen_discard_i64(cc_src);
446 tcg_gen_discard_i64(cc_vr);
448 tcg_gen_mov_i64(cc_dst, dst);
449 s->cc_op = op;
452 static void gen_op_update2_cc_i64(DisasContext *s, enum cc_op op, TCGv_i64 src,
453 TCGv_i64 dst)
455 if (live_cc_data(s)) {
456 tcg_gen_discard_i64(cc_vr);
458 tcg_gen_mov_i64(cc_src, src);
459 tcg_gen_mov_i64(cc_dst, dst);
460 s->cc_op = op;
463 static void gen_op_update3_cc_i64(DisasContext *s, enum cc_op op, TCGv_i64 src,
464 TCGv_i64 dst, TCGv_i64 vr)
466 tcg_gen_mov_i64(cc_src, src);
467 tcg_gen_mov_i64(cc_dst, dst);
468 tcg_gen_mov_i64(cc_vr, vr);
469 s->cc_op = op;
472 static void set_cc_nz_u64(DisasContext *s, TCGv_i64 val)
474 gen_op_update1_cc_i64(s, CC_OP_NZ, val);
477 static void gen_set_cc_nz_f32(DisasContext *s, TCGv_i64 val)
479 gen_op_update1_cc_i64(s, CC_OP_NZ_F32, val);
482 static void gen_set_cc_nz_f64(DisasContext *s, TCGv_i64 val)
484 gen_op_update1_cc_i64(s, CC_OP_NZ_F64, val);
487 static void gen_set_cc_nz_f128(DisasContext *s, TCGv_i64 vh, TCGv_i64 vl)
489 gen_op_update2_cc_i64(s, CC_OP_NZ_F128, vh, vl);
492 /* CC value is in env->cc_op */
493 static void set_cc_static(DisasContext *s)
495 if (live_cc_data(s)) {
496 tcg_gen_discard_i64(cc_src);
497 tcg_gen_discard_i64(cc_dst);
498 tcg_gen_discard_i64(cc_vr);
500 s->cc_op = CC_OP_STATIC;
503 /* calculates cc into cc_op */
504 static void gen_op_calc_cc(DisasContext *s)
506 TCGv_i32 local_cc_op;
507 TCGv_i64 dummy;
509 TCGV_UNUSED_I32(local_cc_op);
510 TCGV_UNUSED_I64(dummy);
511 switch (s->cc_op) {
512 default:
513 dummy = tcg_const_i64(0);
514 /* FALLTHRU */
515 case CC_OP_ADD_64:
516 case CC_OP_ADDU_64:
517 case CC_OP_ADDC_64:
518 case CC_OP_SUB_64:
519 case CC_OP_SUBU_64:
520 case CC_OP_SUBB_64:
521 case CC_OP_ADD_32:
522 case CC_OP_ADDU_32:
523 case CC_OP_ADDC_32:
524 case CC_OP_SUB_32:
525 case CC_OP_SUBU_32:
526 case CC_OP_SUBB_32:
527 local_cc_op = tcg_const_i32(s->cc_op);
528 break;
529 case CC_OP_CONST0:
530 case CC_OP_CONST1:
531 case CC_OP_CONST2:
532 case CC_OP_CONST3:
533 case CC_OP_STATIC:
534 case CC_OP_DYNAMIC:
535 break;
538 switch (s->cc_op) {
539 case CC_OP_CONST0:
540 case CC_OP_CONST1:
541 case CC_OP_CONST2:
542 case CC_OP_CONST3:
543 /* s->cc_op is the cc value */
544 tcg_gen_movi_i32(cc_op, s->cc_op - CC_OP_CONST0);
545 break;
546 case CC_OP_STATIC:
547 /* env->cc_op already is the cc value */
548 break;
549 case CC_OP_NZ:
550 case CC_OP_ABS_64:
551 case CC_OP_NABS_64:
552 case CC_OP_ABS_32:
553 case CC_OP_NABS_32:
554 case CC_OP_LTGT0_32:
555 case CC_OP_LTGT0_64:
556 case CC_OP_COMP_32:
557 case CC_OP_COMP_64:
558 case CC_OP_NZ_F32:
559 case CC_OP_NZ_F64:
560 case CC_OP_FLOGR:
561 /* 1 argument */
562 gen_helper_calc_cc(cc_op, cpu_env, local_cc_op, dummy, cc_dst, dummy);
563 break;
564 case CC_OP_ICM:
565 case CC_OP_LTGT_32:
566 case CC_OP_LTGT_64:
567 case CC_OP_LTUGTU_32:
568 case CC_OP_LTUGTU_64:
569 case CC_OP_TM_32:
570 case CC_OP_TM_64:
571 case CC_OP_SLA_32:
572 case CC_OP_SLA_64:
573 case CC_OP_NZ_F128:
574 /* 2 arguments */
575 gen_helper_calc_cc(cc_op, cpu_env, local_cc_op, cc_src, cc_dst, dummy);
576 break;
577 case CC_OP_ADD_64:
578 case CC_OP_ADDU_64:
579 case CC_OP_ADDC_64:
580 case CC_OP_SUB_64:
581 case CC_OP_SUBU_64:
582 case CC_OP_SUBB_64:
583 case CC_OP_ADD_32:
584 case CC_OP_ADDU_32:
585 case CC_OP_ADDC_32:
586 case CC_OP_SUB_32:
587 case CC_OP_SUBU_32:
588 case CC_OP_SUBB_32:
589 /* 3 arguments */
590 gen_helper_calc_cc(cc_op, cpu_env, local_cc_op, cc_src, cc_dst, cc_vr);
591 break;
592 case CC_OP_DYNAMIC:
593 /* unknown operation - assume 3 arguments and cc_op in env */
594 gen_helper_calc_cc(cc_op, cpu_env, cc_op, cc_src, cc_dst, cc_vr);
595 break;
596 default:
597 tcg_abort();
600 if (!TCGV_IS_UNUSED_I32(local_cc_op)) {
601 tcg_temp_free_i32(local_cc_op);
603 if (!TCGV_IS_UNUSED_I64(dummy)) {
604 tcg_temp_free_i64(dummy);
607 /* We now have cc in cc_op as constant */
608 set_cc_static(s);
611 static int use_goto_tb(DisasContext *s, uint64_t dest)
613 if (unlikely(s->singlestep_enabled) ||
614 (s->tb->cflags & CF_LAST_IO) ||
615 (s->tb->flags & FLAG_MASK_PER)) {
616 return false;
618 #ifndef CONFIG_USER_ONLY
619 return (dest & TARGET_PAGE_MASK) == (s->tb->pc & TARGET_PAGE_MASK) ||
620 (dest & TARGET_PAGE_MASK) == (s->pc & TARGET_PAGE_MASK);
621 #else
622 return true;
623 #endif
626 static void account_noninline_branch(DisasContext *s, int cc_op)
628 #ifdef DEBUG_INLINE_BRANCHES
629 inline_branch_miss[cc_op]++;
630 #endif
633 static void account_inline_branch(DisasContext *s, int cc_op)
635 #ifdef DEBUG_INLINE_BRANCHES
636 inline_branch_hit[cc_op]++;
637 #endif
640 /* Table of mask values to comparison codes, given a comparison as input.
641 For such, CC=3 should not be possible. */
642 static const TCGCond ltgt_cond[16] = {
643 TCG_COND_NEVER, TCG_COND_NEVER, /* | | | x */
644 TCG_COND_GT, TCG_COND_GT, /* | | GT | x */
645 TCG_COND_LT, TCG_COND_LT, /* | LT | | x */
646 TCG_COND_NE, TCG_COND_NE, /* | LT | GT | x */
647 TCG_COND_EQ, TCG_COND_EQ, /* EQ | | | x */
648 TCG_COND_GE, TCG_COND_GE, /* EQ | | GT | x */
649 TCG_COND_LE, TCG_COND_LE, /* EQ | LT | | x */
650 TCG_COND_ALWAYS, TCG_COND_ALWAYS, /* EQ | LT | GT | x */
653 /* Table of mask values to comparison codes, given a logic op as input.
654 For such, only CC=0 and CC=1 should be possible. */
655 static const TCGCond nz_cond[16] = {
656 TCG_COND_NEVER, TCG_COND_NEVER, /* | | x | x */
657 TCG_COND_NEVER, TCG_COND_NEVER,
658 TCG_COND_NE, TCG_COND_NE, /* | NE | x | x */
659 TCG_COND_NE, TCG_COND_NE,
660 TCG_COND_EQ, TCG_COND_EQ, /* EQ | | x | x */
661 TCG_COND_EQ, TCG_COND_EQ,
662 TCG_COND_ALWAYS, TCG_COND_ALWAYS, /* EQ | NE | x | x */
663 TCG_COND_ALWAYS, TCG_COND_ALWAYS,
666 /* Interpret MASK in terms of S->CC_OP, and fill in C with all the
667 details required to generate a TCG comparison. */
668 static void disas_jcc(DisasContext *s, DisasCompare *c, uint32_t mask)
670 TCGCond cond;
671 enum cc_op old_cc_op = s->cc_op;
673 if (mask == 15 || mask == 0) {
674 c->cond = (mask ? TCG_COND_ALWAYS : TCG_COND_NEVER);
675 c->u.s32.a = cc_op;
676 c->u.s32.b = cc_op;
677 c->g1 = c->g2 = true;
678 c->is_64 = false;
679 return;
682 /* Find the TCG condition for the mask + cc op. */
683 switch (old_cc_op) {
684 case CC_OP_LTGT0_32:
685 case CC_OP_LTGT0_64:
686 case CC_OP_LTGT_32:
687 case CC_OP_LTGT_64:
688 cond = ltgt_cond[mask];
689 if (cond == TCG_COND_NEVER) {
690 goto do_dynamic;
692 account_inline_branch(s, old_cc_op);
693 break;
695 case CC_OP_LTUGTU_32:
696 case CC_OP_LTUGTU_64:
697 cond = tcg_unsigned_cond(ltgt_cond[mask]);
698 if (cond == TCG_COND_NEVER) {
699 goto do_dynamic;
701 account_inline_branch(s, old_cc_op);
702 break;
704 case CC_OP_NZ:
705 cond = nz_cond[mask];
706 if (cond == TCG_COND_NEVER) {
707 goto do_dynamic;
709 account_inline_branch(s, old_cc_op);
710 break;
712 case CC_OP_TM_32:
713 case CC_OP_TM_64:
714 switch (mask) {
715 case 8:
716 cond = TCG_COND_EQ;
717 break;
718 case 4 | 2 | 1:
719 cond = TCG_COND_NE;
720 break;
721 default:
722 goto do_dynamic;
724 account_inline_branch(s, old_cc_op);
725 break;
727 case CC_OP_ICM:
728 switch (mask) {
729 case 8:
730 cond = TCG_COND_EQ;
731 break;
732 case 4 | 2 | 1:
733 case 4 | 2:
734 cond = TCG_COND_NE;
735 break;
736 default:
737 goto do_dynamic;
739 account_inline_branch(s, old_cc_op);
740 break;
742 case CC_OP_FLOGR:
743 switch (mask & 0xa) {
744 case 8: /* src == 0 -> no one bit found */
745 cond = TCG_COND_EQ;
746 break;
747 case 2: /* src != 0 -> one bit found */
748 cond = TCG_COND_NE;
749 break;
750 default:
751 goto do_dynamic;
753 account_inline_branch(s, old_cc_op);
754 break;
756 case CC_OP_ADDU_32:
757 case CC_OP_ADDU_64:
758 switch (mask) {
759 case 8 | 2: /* vr == 0 */
760 cond = TCG_COND_EQ;
761 break;
762 case 4 | 1: /* vr != 0 */
763 cond = TCG_COND_NE;
764 break;
765 case 8 | 4: /* no carry -> vr >= src */
766 cond = TCG_COND_GEU;
767 break;
768 case 2 | 1: /* carry -> vr < src */
769 cond = TCG_COND_LTU;
770 break;
771 default:
772 goto do_dynamic;
774 account_inline_branch(s, old_cc_op);
775 break;
777 case CC_OP_SUBU_32:
778 case CC_OP_SUBU_64:
779 /* Note that CC=0 is impossible; treat it as dont-care. */
780 switch (mask & 7) {
781 case 2: /* zero -> op1 == op2 */
782 cond = TCG_COND_EQ;
783 break;
784 case 4 | 1: /* !zero -> op1 != op2 */
785 cond = TCG_COND_NE;
786 break;
787 case 4: /* borrow (!carry) -> op1 < op2 */
788 cond = TCG_COND_LTU;
789 break;
790 case 2 | 1: /* !borrow (carry) -> op1 >= op2 */
791 cond = TCG_COND_GEU;
792 break;
793 default:
794 goto do_dynamic;
796 account_inline_branch(s, old_cc_op);
797 break;
799 default:
800 do_dynamic:
801 /* Calculate cc value. */
802 gen_op_calc_cc(s);
803 /* FALLTHRU */
805 case CC_OP_STATIC:
806 /* Jump based on CC. We'll load up the real cond below;
807 the assignment here merely avoids a compiler warning. */
808 account_noninline_branch(s, old_cc_op);
809 old_cc_op = CC_OP_STATIC;
810 cond = TCG_COND_NEVER;
811 break;
814 /* Load up the arguments of the comparison. */
815 c->is_64 = true;
816 c->g1 = c->g2 = false;
817 switch (old_cc_op) {
818 case CC_OP_LTGT0_32:
819 c->is_64 = false;
820 c->u.s32.a = tcg_temp_new_i32();
821 tcg_gen_extrl_i64_i32(c->u.s32.a, cc_dst);
822 c->u.s32.b = tcg_const_i32(0);
823 break;
824 case CC_OP_LTGT_32:
825 case CC_OP_LTUGTU_32:
826 case CC_OP_SUBU_32:
827 c->is_64 = false;
828 c->u.s32.a = tcg_temp_new_i32();
829 tcg_gen_extrl_i64_i32(c->u.s32.a, cc_src);
830 c->u.s32.b = tcg_temp_new_i32();
831 tcg_gen_extrl_i64_i32(c->u.s32.b, cc_dst);
832 break;
834 case CC_OP_LTGT0_64:
835 case CC_OP_NZ:
836 case CC_OP_FLOGR:
837 c->u.s64.a = cc_dst;
838 c->u.s64.b = tcg_const_i64(0);
839 c->g1 = true;
840 break;
841 case CC_OP_LTGT_64:
842 case CC_OP_LTUGTU_64:
843 case CC_OP_SUBU_64:
844 c->u.s64.a = cc_src;
845 c->u.s64.b = cc_dst;
846 c->g1 = c->g2 = true;
847 break;
849 case CC_OP_TM_32:
850 case CC_OP_TM_64:
851 case CC_OP_ICM:
852 c->u.s64.a = tcg_temp_new_i64();
853 c->u.s64.b = tcg_const_i64(0);
854 tcg_gen_and_i64(c->u.s64.a, cc_src, cc_dst);
855 break;
857 case CC_OP_ADDU_32:
858 c->is_64 = false;
859 c->u.s32.a = tcg_temp_new_i32();
860 c->u.s32.b = tcg_temp_new_i32();
861 tcg_gen_extrl_i64_i32(c->u.s32.a, cc_vr);
862 if (cond == TCG_COND_EQ || cond == TCG_COND_NE) {
863 tcg_gen_movi_i32(c->u.s32.b, 0);
864 } else {
865 tcg_gen_extrl_i64_i32(c->u.s32.b, cc_src);
867 break;
869 case CC_OP_ADDU_64:
870 c->u.s64.a = cc_vr;
871 c->g1 = true;
872 if (cond == TCG_COND_EQ || cond == TCG_COND_NE) {
873 c->u.s64.b = tcg_const_i64(0);
874 } else {
875 c->u.s64.b = cc_src;
876 c->g2 = true;
878 break;
880 case CC_OP_STATIC:
881 c->is_64 = false;
882 c->u.s32.a = cc_op;
883 c->g1 = true;
884 switch (mask) {
885 case 0x8 | 0x4 | 0x2: /* cc != 3 */
886 cond = TCG_COND_NE;
887 c->u.s32.b = tcg_const_i32(3);
888 break;
889 case 0x8 | 0x4 | 0x1: /* cc != 2 */
890 cond = TCG_COND_NE;
891 c->u.s32.b = tcg_const_i32(2);
892 break;
893 case 0x8 | 0x2 | 0x1: /* cc != 1 */
894 cond = TCG_COND_NE;
895 c->u.s32.b = tcg_const_i32(1);
896 break;
897 case 0x8 | 0x2: /* cc == 0 || cc == 2 => (cc & 1) == 0 */
898 cond = TCG_COND_EQ;
899 c->g1 = false;
900 c->u.s32.a = tcg_temp_new_i32();
901 c->u.s32.b = tcg_const_i32(0);
902 tcg_gen_andi_i32(c->u.s32.a, cc_op, 1);
903 break;
904 case 0x8 | 0x4: /* cc < 2 */
905 cond = TCG_COND_LTU;
906 c->u.s32.b = tcg_const_i32(2);
907 break;
908 case 0x8: /* cc == 0 */
909 cond = TCG_COND_EQ;
910 c->u.s32.b = tcg_const_i32(0);
911 break;
912 case 0x4 | 0x2 | 0x1: /* cc != 0 */
913 cond = TCG_COND_NE;
914 c->u.s32.b = tcg_const_i32(0);
915 break;
916 case 0x4 | 0x1: /* cc == 1 || cc == 3 => (cc & 1) != 0 */
917 cond = TCG_COND_NE;
918 c->g1 = false;
919 c->u.s32.a = tcg_temp_new_i32();
920 c->u.s32.b = tcg_const_i32(0);
921 tcg_gen_andi_i32(c->u.s32.a, cc_op, 1);
922 break;
923 case 0x4: /* cc == 1 */
924 cond = TCG_COND_EQ;
925 c->u.s32.b = tcg_const_i32(1);
926 break;
927 case 0x2 | 0x1: /* cc > 1 */
928 cond = TCG_COND_GTU;
929 c->u.s32.b = tcg_const_i32(1);
930 break;
931 case 0x2: /* cc == 2 */
932 cond = TCG_COND_EQ;
933 c->u.s32.b = tcg_const_i32(2);
934 break;
935 case 0x1: /* cc == 3 */
936 cond = TCG_COND_EQ;
937 c->u.s32.b = tcg_const_i32(3);
938 break;
939 default:
940 /* CC is masked by something else: (8 >> cc) & mask. */
941 cond = TCG_COND_NE;
942 c->g1 = false;
943 c->u.s32.a = tcg_const_i32(8);
944 c->u.s32.b = tcg_const_i32(0);
945 tcg_gen_shr_i32(c->u.s32.a, c->u.s32.a, cc_op);
946 tcg_gen_andi_i32(c->u.s32.a, c->u.s32.a, mask);
947 break;
949 break;
951 default:
952 abort();
954 c->cond = cond;
957 static void free_compare(DisasCompare *c)
959 if (!c->g1) {
960 if (c->is_64) {
961 tcg_temp_free_i64(c->u.s64.a);
962 } else {
963 tcg_temp_free_i32(c->u.s32.a);
966 if (!c->g2) {
967 if (c->is_64) {
968 tcg_temp_free_i64(c->u.s64.b);
969 } else {
970 tcg_temp_free_i32(c->u.s32.b);
975 /* ====================================================================== */
976 /* Define the insn format enumeration. */
977 #define F0(N) FMT_##N,
978 #define F1(N, X1) F0(N)
979 #define F2(N, X1, X2) F0(N)
980 #define F3(N, X1, X2, X3) F0(N)
981 #define F4(N, X1, X2, X3, X4) F0(N)
982 #define F5(N, X1, X2, X3, X4, X5) F0(N)
984 typedef enum {
985 #include "insn-format.def"
986 } DisasFormat;
988 #undef F0
989 #undef F1
990 #undef F2
991 #undef F3
992 #undef F4
993 #undef F5
995 /* Define a structure to hold the decoded fields. We'll store each inside
996 an array indexed by an enum. In order to conserve memory, we'll arrange
997 for fields that do not exist at the same time to overlap, thus the "C"
998 for compact. For checking purposes there is an "O" for original index
999 as well that will be applied to availability bitmaps. */
1001 enum DisasFieldIndexO {
1002 FLD_O_r1,
1003 FLD_O_r2,
1004 FLD_O_r3,
1005 FLD_O_m1,
1006 FLD_O_m3,
1007 FLD_O_m4,
1008 FLD_O_b1,
1009 FLD_O_b2,
1010 FLD_O_b4,
1011 FLD_O_d1,
1012 FLD_O_d2,
1013 FLD_O_d4,
1014 FLD_O_x2,
1015 FLD_O_l1,
1016 FLD_O_l2,
1017 FLD_O_i1,
1018 FLD_O_i2,
1019 FLD_O_i3,
1020 FLD_O_i4,
1021 FLD_O_i5
1024 enum DisasFieldIndexC {
1025 FLD_C_r1 = 0,
1026 FLD_C_m1 = 0,
1027 FLD_C_b1 = 0,
1028 FLD_C_i1 = 0,
1030 FLD_C_r2 = 1,
1031 FLD_C_b2 = 1,
1032 FLD_C_i2 = 1,
1034 FLD_C_r3 = 2,
1035 FLD_C_m3 = 2,
1036 FLD_C_i3 = 2,
1038 FLD_C_m4 = 3,
1039 FLD_C_b4 = 3,
1040 FLD_C_i4 = 3,
1041 FLD_C_l1 = 3,
1043 FLD_C_i5 = 4,
1044 FLD_C_d1 = 4,
1046 FLD_C_d2 = 5,
1048 FLD_C_d4 = 6,
1049 FLD_C_x2 = 6,
1050 FLD_C_l2 = 6,
1052 NUM_C_FIELD = 7
1055 struct DisasFields {
1056 uint64_t raw_insn;
1057 unsigned op:8;
1058 unsigned op2:8;
1059 unsigned presentC:16;
1060 unsigned int presentO;
1061 int c[NUM_C_FIELD];
1064 /* This is the way fields are to be accessed out of DisasFields. */
1065 #define have_field(S, F) have_field1((S), FLD_O_##F)
1066 #define get_field(S, F) get_field1((S), FLD_O_##F, FLD_C_##F)
1068 static bool have_field1(const DisasFields *f, enum DisasFieldIndexO c)
1070 return (f->presentO >> c) & 1;
1073 static int get_field1(const DisasFields *f, enum DisasFieldIndexO o,
1074 enum DisasFieldIndexC c)
1076 assert(have_field1(f, o));
1077 return f->c[c];
1080 /* Describe the layout of each field in each format. */
1081 typedef struct DisasField {
1082 unsigned int beg:8;
1083 unsigned int size:8;
1084 unsigned int type:2;
1085 unsigned int indexC:6;
1086 enum DisasFieldIndexO indexO:8;
1087 } DisasField;
1089 typedef struct DisasFormatInfo {
1090 DisasField op[NUM_C_FIELD];
1091 } DisasFormatInfo;
1093 #define R(N, B) { B, 4, 0, FLD_C_r##N, FLD_O_r##N }
1094 #define M(N, B) { B, 4, 0, FLD_C_m##N, FLD_O_m##N }
1095 #define BD(N, BB, BD) { BB, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1096 { BD, 12, 0, FLD_C_d##N, FLD_O_d##N }
1097 #define BXD(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1098 { 12, 4, 0, FLD_C_x##N, FLD_O_x##N }, \
1099 { 20, 12, 0, FLD_C_d##N, FLD_O_d##N }
1100 #define BDL(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1101 { 20, 20, 2, FLD_C_d##N, FLD_O_d##N }
1102 #define BXDL(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1103 { 12, 4, 0, FLD_C_x##N, FLD_O_x##N }, \
1104 { 20, 20, 2, FLD_C_d##N, FLD_O_d##N }
1105 #define I(N, B, S) { B, S, 1, FLD_C_i##N, FLD_O_i##N }
1106 #define L(N, B, S) { B, S, 0, FLD_C_l##N, FLD_O_l##N }
1108 #define F0(N) { { } },
1109 #define F1(N, X1) { { X1 } },
1110 #define F2(N, X1, X2) { { X1, X2 } },
1111 #define F3(N, X1, X2, X3) { { X1, X2, X3 } },
1112 #define F4(N, X1, X2, X3, X4) { { X1, X2, X3, X4 } },
1113 #define F5(N, X1, X2, X3, X4, X5) { { X1, X2, X3, X4, X5 } },
1115 static const DisasFormatInfo format_info[] = {
1116 #include "insn-format.def"
1119 #undef F0
1120 #undef F1
1121 #undef F2
1122 #undef F3
1123 #undef F4
1124 #undef F5
1125 #undef R
1126 #undef M
1127 #undef BD
1128 #undef BXD
1129 #undef BDL
1130 #undef BXDL
1131 #undef I
1132 #undef L
1134 /* Generally, we'll extract operands into this structures, operate upon
1135 them, and store them back. See the "in1", "in2", "prep", "wout" sets
1136 of routines below for more details. */
1137 typedef struct {
1138 bool g_out, g_out2, g_in1, g_in2;
1139 TCGv_i64 out, out2, in1, in2;
1140 TCGv_i64 addr1;
1141 } DisasOps;
1143 /* Instructions can place constraints on their operands, raising specification
1144 exceptions if they are violated. To make this easy to automate, each "in1",
1145 "in2", "prep", "wout" helper will have a SPEC_<name> define that equals one
1146 of the following, or 0. To make this easy to document, we'll put the
1147 SPEC_<name> defines next to <name>. */
1149 #define SPEC_r1_even 1
1150 #define SPEC_r2_even 2
1151 #define SPEC_r3_even 4
1152 #define SPEC_r1_f128 8
1153 #define SPEC_r2_f128 16
1155 /* Return values from translate_one, indicating the state of the TB. */
1156 typedef enum {
1157 /* Continue the TB. */
1158 NO_EXIT,
1159 /* We have emitted one or more goto_tb. No fixup required. */
1160 EXIT_GOTO_TB,
1161 /* We are not using a goto_tb (for whatever reason), but have updated
1162 the PC (for whatever reason), so there's no need to do it again on
1163 exiting the TB. */
1164 EXIT_PC_UPDATED,
1165 /* We are exiting the TB, but have neither emitted a goto_tb, nor
1166 updated the PC for the next instruction to be executed. */
1167 EXIT_PC_STALE,
1168 /* We are ending the TB with a noreturn function call, e.g. longjmp.
1169 No following code will be executed. */
1170 EXIT_NORETURN,
1171 } ExitStatus;
1173 typedef enum DisasFacility {
1174 FAC_Z, /* zarch (default) */
1175 FAC_CASS, /* compare and swap and store */
1176 FAC_CASS2, /* compare and swap and store 2*/
1177 FAC_DFP, /* decimal floating point */
1178 FAC_DFPR, /* decimal floating point rounding */
1179 FAC_DO, /* distinct operands */
1180 FAC_EE, /* execute extensions */
1181 FAC_EI, /* extended immediate */
1182 FAC_FPE, /* floating point extension */
1183 FAC_FPSSH, /* floating point support sign handling */
1184 FAC_FPRGR, /* FPR-GR transfer */
1185 FAC_GIE, /* general instructions extension */
1186 FAC_HFP_MA, /* HFP multiply-and-add/subtract */
1187 FAC_HW, /* high-word */
1188 FAC_IEEEE_SIM, /* IEEE exception sumilation */
1189 FAC_MIE, /* miscellaneous-instruction-extensions */
1190 FAC_LAT, /* load-and-trap */
1191 FAC_LOC, /* load/store on condition */
1192 FAC_LD, /* long displacement */
1193 FAC_PC, /* population count */
1194 FAC_SCF, /* store clock fast */
1195 FAC_SFLE, /* store facility list extended */
1196 FAC_ILA, /* interlocked access facility 1 */
1197 } DisasFacility;
1199 struct DisasInsn {
1200 unsigned opc:16;
1201 DisasFormat fmt:8;
1202 DisasFacility fac:8;
1203 unsigned spec:8;
1205 const char *name;
1207 void (*help_in1)(DisasContext *, DisasFields *, DisasOps *);
1208 void (*help_in2)(DisasContext *, DisasFields *, DisasOps *);
1209 void (*help_prep)(DisasContext *, DisasFields *, DisasOps *);
1210 void (*help_wout)(DisasContext *, DisasFields *, DisasOps *);
1211 void (*help_cout)(DisasContext *, DisasOps *);
1212 ExitStatus (*help_op)(DisasContext *, DisasOps *);
1214 uint64_t data;
1217 /* ====================================================================== */
1218 /* Miscellaneous helpers, used by several operations. */
1220 static void help_l2_shift(DisasContext *s, DisasFields *f,
1221 DisasOps *o, int mask)
1223 int b2 = get_field(f, b2);
1224 int d2 = get_field(f, d2);
1226 if (b2 == 0) {
1227 o->in2 = tcg_const_i64(d2 & mask);
1228 } else {
1229 o->in2 = get_address(s, 0, b2, d2);
1230 tcg_gen_andi_i64(o->in2, o->in2, mask);
1234 static ExitStatus help_goto_direct(DisasContext *s, uint64_t dest)
1236 if (dest == s->next_pc) {
1237 per_branch(s, true);
1238 return NO_EXIT;
1240 if (use_goto_tb(s, dest)) {
1241 update_cc_op(s);
1242 per_breaking_event(s);
1243 tcg_gen_goto_tb(0);
1244 tcg_gen_movi_i64(psw_addr, dest);
1245 tcg_gen_exit_tb((uintptr_t)s->tb);
1246 return EXIT_GOTO_TB;
1247 } else {
1248 tcg_gen_movi_i64(psw_addr, dest);
1249 per_branch(s, false);
1250 return EXIT_PC_UPDATED;
1254 static ExitStatus help_branch(DisasContext *s, DisasCompare *c,
1255 bool is_imm, int imm, TCGv_i64 cdest)
1257 ExitStatus ret;
1258 uint64_t dest = s->pc + 2 * imm;
1259 TCGLabel *lab;
1261 /* Take care of the special cases first. */
1262 if (c->cond == TCG_COND_NEVER) {
1263 ret = NO_EXIT;
1264 goto egress;
1266 if (is_imm) {
1267 if (dest == s->next_pc) {
1268 /* Branch to next. */
1269 per_branch(s, true);
1270 ret = NO_EXIT;
1271 goto egress;
1273 if (c->cond == TCG_COND_ALWAYS) {
1274 ret = help_goto_direct(s, dest);
1275 goto egress;
1277 } else {
1278 if (TCGV_IS_UNUSED_I64(cdest)) {
1279 /* E.g. bcr %r0 -> no branch. */
1280 ret = NO_EXIT;
1281 goto egress;
1283 if (c->cond == TCG_COND_ALWAYS) {
1284 tcg_gen_mov_i64(psw_addr, cdest);
1285 per_branch(s, false);
1286 ret = EXIT_PC_UPDATED;
1287 goto egress;
1291 if (use_goto_tb(s, s->next_pc)) {
1292 if (is_imm && use_goto_tb(s, dest)) {
1293 /* Both exits can use goto_tb. */
1294 update_cc_op(s);
1296 lab = gen_new_label();
1297 if (c->is_64) {
1298 tcg_gen_brcond_i64(c->cond, c->u.s64.a, c->u.s64.b, lab);
1299 } else {
1300 tcg_gen_brcond_i32(c->cond, c->u.s32.a, c->u.s32.b, lab);
1303 /* Branch not taken. */
1304 tcg_gen_goto_tb(0);
1305 tcg_gen_movi_i64(psw_addr, s->next_pc);
1306 tcg_gen_exit_tb((uintptr_t)s->tb + 0);
1308 /* Branch taken. */
1309 gen_set_label(lab);
1310 per_breaking_event(s);
1311 tcg_gen_goto_tb(1);
1312 tcg_gen_movi_i64(psw_addr, dest);
1313 tcg_gen_exit_tb((uintptr_t)s->tb + 1);
1315 ret = EXIT_GOTO_TB;
1316 } else {
1317 /* Fallthru can use goto_tb, but taken branch cannot. */
1318 /* Store taken branch destination before the brcond. This
1319 avoids having to allocate a new local temp to hold it.
1320 We'll overwrite this in the not taken case anyway. */
1321 if (!is_imm) {
1322 tcg_gen_mov_i64(psw_addr, cdest);
1325 lab = gen_new_label();
1326 if (c->is_64) {
1327 tcg_gen_brcond_i64(c->cond, c->u.s64.a, c->u.s64.b, lab);
1328 } else {
1329 tcg_gen_brcond_i32(c->cond, c->u.s32.a, c->u.s32.b, lab);
1332 /* Branch not taken. */
1333 update_cc_op(s);
1334 tcg_gen_goto_tb(0);
1335 tcg_gen_movi_i64(psw_addr, s->next_pc);
1336 tcg_gen_exit_tb((uintptr_t)s->tb + 0);
1338 gen_set_label(lab);
1339 if (is_imm) {
1340 tcg_gen_movi_i64(psw_addr, dest);
1342 per_breaking_event(s);
1343 ret = EXIT_PC_UPDATED;
1345 } else {
1346 /* Fallthru cannot use goto_tb. This by itself is vanishingly rare.
1347 Most commonly we're single-stepping or some other condition that
1348 disables all use of goto_tb. Just update the PC and exit. */
1350 TCGv_i64 next = tcg_const_i64(s->next_pc);
1351 if (is_imm) {
1352 cdest = tcg_const_i64(dest);
1355 if (c->is_64) {
1356 tcg_gen_movcond_i64(c->cond, psw_addr, c->u.s64.a, c->u.s64.b,
1357 cdest, next);
1358 per_branch_cond(s, c->cond, c->u.s64.a, c->u.s64.b);
1359 } else {
1360 TCGv_i32 t0 = tcg_temp_new_i32();
1361 TCGv_i64 t1 = tcg_temp_new_i64();
1362 TCGv_i64 z = tcg_const_i64(0);
1363 tcg_gen_setcond_i32(c->cond, t0, c->u.s32.a, c->u.s32.b);
1364 tcg_gen_extu_i32_i64(t1, t0);
1365 tcg_temp_free_i32(t0);
1366 tcg_gen_movcond_i64(TCG_COND_NE, psw_addr, t1, z, cdest, next);
1367 per_branch_cond(s, TCG_COND_NE, t1, z);
1368 tcg_temp_free_i64(t1);
1369 tcg_temp_free_i64(z);
1372 if (is_imm) {
1373 tcg_temp_free_i64(cdest);
1375 tcg_temp_free_i64(next);
1377 ret = EXIT_PC_UPDATED;
1380 egress:
1381 free_compare(c);
1382 return ret;
1385 /* ====================================================================== */
1386 /* The operations. These perform the bulk of the work for any insn,
1387 usually after the operands have been loaded and output initialized. */
1389 static ExitStatus op_abs(DisasContext *s, DisasOps *o)
1391 TCGv_i64 z, n;
1392 z = tcg_const_i64(0);
1393 n = tcg_temp_new_i64();
1394 tcg_gen_neg_i64(n, o->in2);
1395 tcg_gen_movcond_i64(TCG_COND_LT, o->out, o->in2, z, n, o->in2);
1396 tcg_temp_free_i64(n);
1397 tcg_temp_free_i64(z);
1398 return NO_EXIT;
1401 static ExitStatus op_absf32(DisasContext *s, DisasOps *o)
1403 tcg_gen_andi_i64(o->out, o->in2, 0x7fffffffull);
1404 return NO_EXIT;
1407 static ExitStatus op_absf64(DisasContext *s, DisasOps *o)
1409 tcg_gen_andi_i64(o->out, o->in2, 0x7fffffffffffffffull);
1410 return NO_EXIT;
1413 static ExitStatus op_absf128(DisasContext *s, DisasOps *o)
1415 tcg_gen_andi_i64(o->out, o->in1, 0x7fffffffffffffffull);
1416 tcg_gen_mov_i64(o->out2, o->in2);
1417 return NO_EXIT;
1420 static ExitStatus op_add(DisasContext *s, DisasOps *o)
1422 tcg_gen_add_i64(o->out, o->in1, o->in2);
1423 return NO_EXIT;
1426 static ExitStatus op_addc(DisasContext *s, DisasOps *o)
1428 DisasCompare cmp;
1429 TCGv_i64 carry;
1431 tcg_gen_add_i64(o->out, o->in1, o->in2);
1433 /* The carry flag is the msb of CC, therefore the branch mask that would
1434 create that comparison is 3. Feeding the generated comparison to
1435 setcond produces the carry flag that we desire. */
1436 disas_jcc(s, &cmp, 3);
1437 carry = tcg_temp_new_i64();
1438 if (cmp.is_64) {
1439 tcg_gen_setcond_i64(cmp.cond, carry, cmp.u.s64.a, cmp.u.s64.b);
1440 } else {
1441 TCGv_i32 t = tcg_temp_new_i32();
1442 tcg_gen_setcond_i32(cmp.cond, t, cmp.u.s32.a, cmp.u.s32.b);
1443 tcg_gen_extu_i32_i64(carry, t);
1444 tcg_temp_free_i32(t);
1446 free_compare(&cmp);
1448 tcg_gen_add_i64(o->out, o->out, carry);
1449 tcg_temp_free_i64(carry);
1450 return NO_EXIT;
1453 static ExitStatus op_aeb(DisasContext *s, DisasOps *o)
1455 gen_helper_aeb(o->out, cpu_env, o->in1, o->in2);
1456 return NO_EXIT;
1459 static ExitStatus op_adb(DisasContext *s, DisasOps *o)
1461 gen_helper_adb(o->out, cpu_env, o->in1, o->in2);
1462 return NO_EXIT;
1465 static ExitStatus op_axb(DisasContext *s, DisasOps *o)
1467 gen_helper_axb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
1468 return_low128(o->out2);
1469 return NO_EXIT;
1472 static ExitStatus op_and(DisasContext *s, DisasOps *o)
1474 tcg_gen_and_i64(o->out, o->in1, o->in2);
1475 return NO_EXIT;
1478 static ExitStatus op_andi(DisasContext *s, DisasOps *o)
1480 int shift = s->insn->data & 0xff;
1481 int size = s->insn->data >> 8;
1482 uint64_t mask = ((1ull << size) - 1) << shift;
1484 assert(!o->g_in2);
1485 tcg_gen_shli_i64(o->in2, o->in2, shift);
1486 tcg_gen_ori_i64(o->in2, o->in2, ~mask);
1487 tcg_gen_and_i64(o->out, o->in1, o->in2);
1489 /* Produce the CC from only the bits manipulated. */
1490 tcg_gen_andi_i64(cc_dst, o->out, mask);
1491 set_cc_nz_u64(s, cc_dst);
1492 return NO_EXIT;
1495 static ExitStatus op_bas(DisasContext *s, DisasOps *o)
1497 tcg_gen_movi_i64(o->out, pc_to_link_info(s, s->next_pc));
1498 if (!TCGV_IS_UNUSED_I64(o->in2)) {
1499 tcg_gen_mov_i64(psw_addr, o->in2);
1500 per_branch(s, false);
1501 return EXIT_PC_UPDATED;
1502 } else {
1503 return NO_EXIT;
1507 static ExitStatus op_basi(DisasContext *s, DisasOps *o)
1509 tcg_gen_movi_i64(o->out, pc_to_link_info(s, s->next_pc));
1510 return help_goto_direct(s, s->pc + 2 * get_field(s->fields, i2));
1513 static ExitStatus op_bc(DisasContext *s, DisasOps *o)
1515 int m1 = get_field(s->fields, m1);
1516 bool is_imm = have_field(s->fields, i2);
1517 int imm = is_imm ? get_field(s->fields, i2) : 0;
1518 DisasCompare c;
1520 disas_jcc(s, &c, m1);
1521 return help_branch(s, &c, is_imm, imm, o->in2);
1524 static ExitStatus op_bct32(DisasContext *s, DisasOps *o)
1526 int r1 = get_field(s->fields, r1);
1527 bool is_imm = have_field(s->fields, i2);
1528 int imm = is_imm ? get_field(s->fields, i2) : 0;
1529 DisasCompare c;
1530 TCGv_i64 t;
1532 c.cond = TCG_COND_NE;
1533 c.is_64 = false;
1534 c.g1 = false;
1535 c.g2 = false;
1537 t = tcg_temp_new_i64();
1538 tcg_gen_subi_i64(t, regs[r1], 1);
1539 store_reg32_i64(r1, t);
1540 c.u.s32.a = tcg_temp_new_i32();
1541 c.u.s32.b = tcg_const_i32(0);
1542 tcg_gen_extrl_i64_i32(c.u.s32.a, t);
1543 tcg_temp_free_i64(t);
1545 return help_branch(s, &c, is_imm, imm, o->in2);
1548 static ExitStatus op_bcth(DisasContext *s, DisasOps *o)
1550 int r1 = get_field(s->fields, r1);
1551 int imm = get_field(s->fields, i2);
1552 DisasCompare c;
1553 TCGv_i64 t;
1555 c.cond = TCG_COND_NE;
1556 c.is_64 = false;
1557 c.g1 = false;
1558 c.g2 = false;
1560 t = tcg_temp_new_i64();
1561 tcg_gen_shri_i64(t, regs[r1], 32);
1562 tcg_gen_subi_i64(t, t, 1);
1563 store_reg32h_i64(r1, t);
1564 c.u.s32.a = tcg_temp_new_i32();
1565 c.u.s32.b = tcg_const_i32(0);
1566 tcg_gen_extrl_i64_i32(c.u.s32.a, t);
1567 tcg_temp_free_i64(t);
1569 return help_branch(s, &c, 1, imm, o->in2);
1572 static ExitStatus op_bct64(DisasContext *s, DisasOps *o)
1574 int r1 = get_field(s->fields, r1);
1575 bool is_imm = have_field(s->fields, i2);
1576 int imm = is_imm ? get_field(s->fields, i2) : 0;
1577 DisasCompare c;
1579 c.cond = TCG_COND_NE;
1580 c.is_64 = true;
1581 c.g1 = true;
1582 c.g2 = false;
1584 tcg_gen_subi_i64(regs[r1], regs[r1], 1);
1585 c.u.s64.a = regs[r1];
1586 c.u.s64.b = tcg_const_i64(0);
1588 return help_branch(s, &c, is_imm, imm, o->in2);
1591 static ExitStatus op_bx32(DisasContext *s, DisasOps *o)
1593 int r1 = get_field(s->fields, r1);
1594 int r3 = get_field(s->fields, r3);
1595 bool is_imm = have_field(s->fields, i2);
1596 int imm = is_imm ? get_field(s->fields, i2) : 0;
1597 DisasCompare c;
1598 TCGv_i64 t;
1600 c.cond = (s->insn->data ? TCG_COND_LE : TCG_COND_GT);
1601 c.is_64 = false;
1602 c.g1 = false;
1603 c.g2 = false;
1605 t = tcg_temp_new_i64();
1606 tcg_gen_add_i64(t, regs[r1], regs[r3]);
1607 c.u.s32.a = tcg_temp_new_i32();
1608 c.u.s32.b = tcg_temp_new_i32();
1609 tcg_gen_extrl_i64_i32(c.u.s32.a, t);
1610 tcg_gen_extrl_i64_i32(c.u.s32.b, regs[r3 | 1]);
1611 store_reg32_i64(r1, t);
1612 tcg_temp_free_i64(t);
1614 return help_branch(s, &c, is_imm, imm, o->in2);
1617 static ExitStatus op_bx64(DisasContext *s, DisasOps *o)
1619 int r1 = get_field(s->fields, r1);
1620 int r3 = get_field(s->fields, r3);
1621 bool is_imm = have_field(s->fields, i2);
1622 int imm = is_imm ? get_field(s->fields, i2) : 0;
1623 DisasCompare c;
1625 c.cond = (s->insn->data ? TCG_COND_LE : TCG_COND_GT);
1626 c.is_64 = true;
1628 if (r1 == (r3 | 1)) {
1629 c.u.s64.b = load_reg(r3 | 1);
1630 c.g2 = false;
1631 } else {
1632 c.u.s64.b = regs[r3 | 1];
1633 c.g2 = true;
1636 tcg_gen_add_i64(regs[r1], regs[r1], regs[r3]);
1637 c.u.s64.a = regs[r1];
1638 c.g1 = true;
1640 return help_branch(s, &c, is_imm, imm, o->in2);
1643 static ExitStatus op_cj(DisasContext *s, DisasOps *o)
1645 int imm, m3 = get_field(s->fields, m3);
1646 bool is_imm;
1647 DisasCompare c;
1649 c.cond = ltgt_cond[m3];
1650 if (s->insn->data) {
1651 c.cond = tcg_unsigned_cond(c.cond);
1653 c.is_64 = c.g1 = c.g2 = true;
1654 c.u.s64.a = o->in1;
1655 c.u.s64.b = o->in2;
1657 is_imm = have_field(s->fields, i4);
1658 if (is_imm) {
1659 imm = get_field(s->fields, i4);
1660 } else {
1661 imm = 0;
1662 o->out = get_address(s, 0, get_field(s->fields, b4),
1663 get_field(s->fields, d4));
1666 return help_branch(s, &c, is_imm, imm, o->out);
1669 static ExitStatus op_ceb(DisasContext *s, DisasOps *o)
1671 gen_helper_ceb(cc_op, cpu_env, o->in1, o->in2);
1672 set_cc_static(s);
1673 return NO_EXIT;
1676 static ExitStatus op_cdb(DisasContext *s, DisasOps *o)
1678 gen_helper_cdb(cc_op, cpu_env, o->in1, o->in2);
1679 set_cc_static(s);
1680 return NO_EXIT;
1683 static ExitStatus op_cxb(DisasContext *s, DisasOps *o)
1685 gen_helper_cxb(cc_op, cpu_env, o->out, o->out2, o->in1, o->in2);
1686 set_cc_static(s);
1687 return NO_EXIT;
1690 static ExitStatus op_cfeb(DisasContext *s, DisasOps *o)
1692 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1693 gen_helper_cfeb(o->out, cpu_env, o->in2, m3);
1694 tcg_temp_free_i32(m3);
1695 gen_set_cc_nz_f32(s, o->in2);
1696 return NO_EXIT;
1699 static ExitStatus op_cfdb(DisasContext *s, DisasOps *o)
1701 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1702 gen_helper_cfdb(o->out, cpu_env, o->in2, m3);
1703 tcg_temp_free_i32(m3);
1704 gen_set_cc_nz_f64(s, o->in2);
1705 return NO_EXIT;
1708 static ExitStatus op_cfxb(DisasContext *s, DisasOps *o)
1710 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1711 gen_helper_cfxb(o->out, cpu_env, o->in1, o->in2, m3);
1712 tcg_temp_free_i32(m3);
1713 gen_set_cc_nz_f128(s, o->in1, o->in2);
1714 return NO_EXIT;
1717 static ExitStatus op_cgeb(DisasContext *s, DisasOps *o)
1719 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1720 gen_helper_cgeb(o->out, cpu_env, o->in2, m3);
1721 tcg_temp_free_i32(m3);
1722 gen_set_cc_nz_f32(s, o->in2);
1723 return NO_EXIT;
1726 static ExitStatus op_cgdb(DisasContext *s, DisasOps *o)
1728 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1729 gen_helper_cgdb(o->out, cpu_env, o->in2, m3);
1730 tcg_temp_free_i32(m3);
1731 gen_set_cc_nz_f64(s, o->in2);
1732 return NO_EXIT;
1735 static ExitStatus op_cgxb(DisasContext *s, DisasOps *o)
1737 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1738 gen_helper_cgxb(o->out, cpu_env, o->in1, o->in2, m3);
1739 tcg_temp_free_i32(m3);
1740 gen_set_cc_nz_f128(s, o->in1, o->in2);
1741 return NO_EXIT;
1744 static ExitStatus op_clfeb(DisasContext *s, DisasOps *o)
1746 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1747 gen_helper_clfeb(o->out, cpu_env, o->in2, m3);
1748 tcg_temp_free_i32(m3);
1749 gen_set_cc_nz_f32(s, o->in2);
1750 return NO_EXIT;
1753 static ExitStatus op_clfdb(DisasContext *s, DisasOps *o)
1755 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1756 gen_helper_clfdb(o->out, cpu_env, o->in2, m3);
1757 tcg_temp_free_i32(m3);
1758 gen_set_cc_nz_f64(s, o->in2);
1759 return NO_EXIT;
1762 static ExitStatus op_clfxb(DisasContext *s, DisasOps *o)
1764 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1765 gen_helper_clfxb(o->out, cpu_env, o->in1, o->in2, m3);
1766 tcg_temp_free_i32(m3);
1767 gen_set_cc_nz_f128(s, o->in1, o->in2);
1768 return NO_EXIT;
1771 static ExitStatus op_clgeb(DisasContext *s, DisasOps *o)
1773 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1774 gen_helper_clgeb(o->out, cpu_env, o->in2, m3);
1775 tcg_temp_free_i32(m3);
1776 gen_set_cc_nz_f32(s, o->in2);
1777 return NO_EXIT;
1780 static ExitStatus op_clgdb(DisasContext *s, DisasOps *o)
1782 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1783 gen_helper_clgdb(o->out, cpu_env, o->in2, m3);
1784 tcg_temp_free_i32(m3);
1785 gen_set_cc_nz_f64(s, o->in2);
1786 return NO_EXIT;
1789 static ExitStatus op_clgxb(DisasContext *s, DisasOps *o)
1791 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1792 gen_helper_clgxb(o->out, cpu_env, o->in1, o->in2, m3);
1793 tcg_temp_free_i32(m3);
1794 gen_set_cc_nz_f128(s, o->in1, o->in2);
1795 return NO_EXIT;
1798 static ExitStatus op_cegb(DisasContext *s, DisasOps *o)
1800 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1801 gen_helper_cegb(o->out, cpu_env, o->in2, m3);
1802 tcg_temp_free_i32(m3);
1803 return NO_EXIT;
1806 static ExitStatus op_cdgb(DisasContext *s, DisasOps *o)
1808 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1809 gen_helper_cdgb(o->out, cpu_env, o->in2, m3);
1810 tcg_temp_free_i32(m3);
1811 return NO_EXIT;
1814 static ExitStatus op_cxgb(DisasContext *s, DisasOps *o)
1816 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1817 gen_helper_cxgb(o->out, cpu_env, o->in2, m3);
1818 tcg_temp_free_i32(m3);
1819 return_low128(o->out2);
1820 return NO_EXIT;
1823 static ExitStatus op_celgb(DisasContext *s, DisasOps *o)
1825 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1826 gen_helper_celgb(o->out, cpu_env, o->in2, m3);
1827 tcg_temp_free_i32(m3);
1828 return NO_EXIT;
1831 static ExitStatus op_cdlgb(DisasContext *s, DisasOps *o)
1833 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1834 gen_helper_cdlgb(o->out, cpu_env, o->in2, m3);
1835 tcg_temp_free_i32(m3);
1836 return NO_EXIT;
1839 static ExitStatus op_cxlgb(DisasContext *s, DisasOps *o)
1841 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1842 gen_helper_cxlgb(o->out, cpu_env, o->in2, m3);
1843 tcg_temp_free_i32(m3);
1844 return_low128(o->out2);
1845 return NO_EXIT;
1848 static ExitStatus op_cksm(DisasContext *s, DisasOps *o)
1850 int r2 = get_field(s->fields, r2);
1851 TCGv_i64 len = tcg_temp_new_i64();
1853 potential_page_fault(s);
1854 gen_helper_cksm(len, cpu_env, o->in1, o->in2, regs[r2 + 1]);
1855 set_cc_static(s);
1856 return_low128(o->out);
1858 tcg_gen_add_i64(regs[r2], regs[r2], len);
1859 tcg_gen_sub_i64(regs[r2 + 1], regs[r2 + 1], len);
1860 tcg_temp_free_i64(len);
1862 return NO_EXIT;
1865 static ExitStatus op_clc(DisasContext *s, DisasOps *o)
1867 int l = get_field(s->fields, l1);
1868 TCGv_i32 vl;
1870 switch (l + 1) {
1871 case 1:
1872 tcg_gen_qemu_ld8u(cc_src, o->addr1, get_mem_index(s));
1873 tcg_gen_qemu_ld8u(cc_dst, o->in2, get_mem_index(s));
1874 break;
1875 case 2:
1876 tcg_gen_qemu_ld16u(cc_src, o->addr1, get_mem_index(s));
1877 tcg_gen_qemu_ld16u(cc_dst, o->in2, get_mem_index(s));
1878 break;
1879 case 4:
1880 tcg_gen_qemu_ld32u(cc_src, o->addr1, get_mem_index(s));
1881 tcg_gen_qemu_ld32u(cc_dst, o->in2, get_mem_index(s));
1882 break;
1883 case 8:
1884 tcg_gen_qemu_ld64(cc_src, o->addr1, get_mem_index(s));
1885 tcg_gen_qemu_ld64(cc_dst, o->in2, get_mem_index(s));
1886 break;
1887 default:
1888 potential_page_fault(s);
1889 vl = tcg_const_i32(l);
1890 gen_helper_clc(cc_op, cpu_env, vl, o->addr1, o->in2);
1891 tcg_temp_free_i32(vl);
1892 set_cc_static(s);
1893 return NO_EXIT;
1895 gen_op_update2_cc_i64(s, CC_OP_LTUGTU_64, cc_src, cc_dst);
1896 return NO_EXIT;
1899 static ExitStatus op_clcle(DisasContext *s, DisasOps *o)
1901 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
1902 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
1903 potential_page_fault(s);
1904 gen_helper_clcle(cc_op, cpu_env, r1, o->in2, r3);
1905 tcg_temp_free_i32(r1);
1906 tcg_temp_free_i32(r3);
1907 set_cc_static(s);
1908 return NO_EXIT;
1911 static ExitStatus op_clm(DisasContext *s, DisasOps *o)
1913 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
1914 TCGv_i32 t1 = tcg_temp_new_i32();
1915 tcg_gen_extrl_i64_i32(t1, o->in1);
1916 potential_page_fault(s);
1917 gen_helper_clm(cc_op, cpu_env, t1, m3, o->in2);
1918 set_cc_static(s);
1919 tcg_temp_free_i32(t1);
1920 tcg_temp_free_i32(m3);
1921 return NO_EXIT;
1924 static ExitStatus op_clst(DisasContext *s, DisasOps *o)
1926 potential_page_fault(s);
1927 gen_helper_clst(o->in1, cpu_env, regs[0], o->in1, o->in2);
1928 set_cc_static(s);
1929 return_low128(o->in2);
1930 return NO_EXIT;
1933 static ExitStatus op_cps(DisasContext *s, DisasOps *o)
1935 TCGv_i64 t = tcg_temp_new_i64();
1936 tcg_gen_andi_i64(t, o->in1, 0x8000000000000000ull);
1937 tcg_gen_andi_i64(o->out, o->in2, 0x7fffffffffffffffull);
1938 tcg_gen_or_i64(o->out, o->out, t);
1939 tcg_temp_free_i64(t);
1940 return NO_EXIT;
1943 static ExitStatus op_cs(DisasContext *s, DisasOps *o)
1945 /* FIXME: needs an atomic solution for CONFIG_USER_ONLY. */
1946 int d2 = get_field(s->fields, d2);
1947 int b2 = get_field(s->fields, b2);
1948 int is_64 = s->insn->data;
1949 TCGv_i64 addr, mem, cc, z;
1951 /* Note that in1 = R3 (new value) and
1952 in2 = (zero-extended) R1 (expected value). */
1954 /* Load the memory into the (temporary) output. While the PoO only talks
1955 about moving the memory to R1 on inequality, if we include equality it
1956 means that R1 is equal to the memory in all conditions. */
1957 addr = get_address(s, 0, b2, d2);
1958 if (is_64) {
1959 tcg_gen_qemu_ld64(o->out, addr, get_mem_index(s));
1960 } else {
1961 tcg_gen_qemu_ld32u(o->out, addr, get_mem_index(s));
1964 /* Are the memory and expected values (un)equal? Note that this setcond
1965 produces the output CC value, thus the NE sense of the test. */
1966 cc = tcg_temp_new_i64();
1967 tcg_gen_setcond_i64(TCG_COND_NE, cc, o->in2, o->out);
1969 /* If the memory and expected values are equal (CC==0), copy R3 to MEM.
1970 Recall that we are allowed to unconditionally issue the store (and
1971 thus any possible write trap), so (re-)store the original contents
1972 of MEM in case of inequality. */
1973 z = tcg_const_i64(0);
1974 mem = tcg_temp_new_i64();
1975 tcg_gen_movcond_i64(TCG_COND_EQ, mem, cc, z, o->in1, o->out);
1976 if (is_64) {
1977 tcg_gen_qemu_st64(mem, addr, get_mem_index(s));
1978 } else {
1979 tcg_gen_qemu_st32(mem, addr, get_mem_index(s));
1981 tcg_temp_free_i64(z);
1982 tcg_temp_free_i64(mem);
1983 tcg_temp_free_i64(addr);
1985 /* Store CC back to cc_op. Wait until after the store so that any
1986 exception gets the old cc_op value. */
1987 tcg_gen_extrl_i64_i32(cc_op, cc);
1988 tcg_temp_free_i64(cc);
1989 set_cc_static(s);
1990 return NO_EXIT;
1993 static ExitStatus op_cdsg(DisasContext *s, DisasOps *o)
1995 /* FIXME: needs an atomic solution for CONFIG_USER_ONLY. */
1996 int r1 = get_field(s->fields, r1);
1997 int r3 = get_field(s->fields, r3);
1998 int d2 = get_field(s->fields, d2);
1999 int b2 = get_field(s->fields, b2);
2000 TCGv_i64 addrh, addrl, memh, meml, outh, outl, cc, z;
2002 /* Note that R1:R1+1 = expected value and R3:R3+1 = new value. */
2004 addrh = get_address(s, 0, b2, d2);
2005 addrl = get_address(s, 0, b2, d2 + 8);
2006 outh = tcg_temp_new_i64();
2007 outl = tcg_temp_new_i64();
2009 tcg_gen_qemu_ld64(outh, addrh, get_mem_index(s));
2010 tcg_gen_qemu_ld64(outl, addrl, get_mem_index(s));
2012 /* Fold the double-word compare with arithmetic. */
2013 cc = tcg_temp_new_i64();
2014 z = tcg_temp_new_i64();
2015 tcg_gen_xor_i64(cc, outh, regs[r1]);
2016 tcg_gen_xor_i64(z, outl, regs[r1 + 1]);
2017 tcg_gen_or_i64(cc, cc, z);
2018 tcg_gen_movi_i64(z, 0);
2019 tcg_gen_setcond_i64(TCG_COND_NE, cc, cc, z);
2021 memh = tcg_temp_new_i64();
2022 meml = tcg_temp_new_i64();
2023 tcg_gen_movcond_i64(TCG_COND_EQ, memh, cc, z, regs[r3], outh);
2024 tcg_gen_movcond_i64(TCG_COND_EQ, meml, cc, z, regs[r3 + 1], outl);
2025 tcg_temp_free_i64(z);
2027 tcg_gen_qemu_st64(memh, addrh, get_mem_index(s));
2028 tcg_gen_qemu_st64(meml, addrl, get_mem_index(s));
2029 tcg_temp_free_i64(memh);
2030 tcg_temp_free_i64(meml);
2031 tcg_temp_free_i64(addrh);
2032 tcg_temp_free_i64(addrl);
2034 /* Save back state now that we've passed all exceptions. */
2035 tcg_gen_mov_i64(regs[r1], outh);
2036 tcg_gen_mov_i64(regs[r1 + 1], outl);
2037 tcg_gen_extrl_i64_i32(cc_op, cc);
2038 tcg_temp_free_i64(outh);
2039 tcg_temp_free_i64(outl);
2040 tcg_temp_free_i64(cc);
2041 set_cc_static(s);
2042 return NO_EXIT;
2045 #ifndef CONFIG_USER_ONLY
2046 static ExitStatus op_csp(DisasContext *s, DisasOps *o)
2048 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2049 check_privileged(s);
2050 gen_helper_csp(cc_op, cpu_env, r1, o->in2);
2051 tcg_temp_free_i32(r1);
2052 set_cc_static(s);
2053 return NO_EXIT;
2055 #endif
2057 static ExitStatus op_cvd(DisasContext *s, DisasOps *o)
2059 TCGv_i64 t1 = tcg_temp_new_i64();
2060 TCGv_i32 t2 = tcg_temp_new_i32();
2061 tcg_gen_extrl_i64_i32(t2, o->in1);
2062 gen_helper_cvd(t1, t2);
2063 tcg_temp_free_i32(t2);
2064 tcg_gen_qemu_st64(t1, o->in2, get_mem_index(s));
2065 tcg_temp_free_i64(t1);
2066 return NO_EXIT;
2069 static ExitStatus op_ct(DisasContext *s, DisasOps *o)
2071 int m3 = get_field(s->fields, m3);
2072 TCGLabel *lab = gen_new_label();
2073 TCGCond c;
2075 c = tcg_invert_cond(ltgt_cond[m3]);
2076 if (s->insn->data) {
2077 c = tcg_unsigned_cond(c);
2079 tcg_gen_brcond_i64(c, o->in1, o->in2, lab);
2081 /* Trap. */
2082 gen_trap(s);
2084 gen_set_label(lab);
2085 return NO_EXIT;
2088 #ifndef CONFIG_USER_ONLY
2089 static ExitStatus op_diag(DisasContext *s, DisasOps *o)
2091 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2092 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2093 TCGv_i32 func_code = tcg_const_i32(get_field(s->fields, i2));
2095 check_privileged(s);
2096 update_psw_addr(s);
2097 gen_op_calc_cc(s);
2099 gen_helper_diag(cpu_env, r1, r3, func_code);
2101 tcg_temp_free_i32(func_code);
2102 tcg_temp_free_i32(r3);
2103 tcg_temp_free_i32(r1);
2104 return NO_EXIT;
2106 #endif
2108 static ExitStatus op_divs32(DisasContext *s, DisasOps *o)
2110 gen_helper_divs32(o->out2, cpu_env, o->in1, o->in2);
2111 return_low128(o->out);
2112 return NO_EXIT;
2115 static ExitStatus op_divu32(DisasContext *s, DisasOps *o)
2117 gen_helper_divu32(o->out2, cpu_env, o->in1, o->in2);
2118 return_low128(o->out);
2119 return NO_EXIT;
2122 static ExitStatus op_divs64(DisasContext *s, DisasOps *o)
2124 gen_helper_divs64(o->out2, cpu_env, o->in1, o->in2);
2125 return_low128(o->out);
2126 return NO_EXIT;
2129 static ExitStatus op_divu64(DisasContext *s, DisasOps *o)
2131 gen_helper_divu64(o->out2, cpu_env, o->out, o->out2, o->in2);
2132 return_low128(o->out);
2133 return NO_EXIT;
2136 static ExitStatus op_deb(DisasContext *s, DisasOps *o)
2138 gen_helper_deb(o->out, cpu_env, o->in1, o->in2);
2139 return NO_EXIT;
2142 static ExitStatus op_ddb(DisasContext *s, DisasOps *o)
2144 gen_helper_ddb(o->out, cpu_env, o->in1, o->in2);
2145 return NO_EXIT;
2148 static ExitStatus op_dxb(DisasContext *s, DisasOps *o)
2150 gen_helper_dxb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
2151 return_low128(o->out2);
2152 return NO_EXIT;
2155 static ExitStatus op_ear(DisasContext *s, DisasOps *o)
2157 int r2 = get_field(s->fields, r2);
2158 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, aregs[r2]));
2159 return NO_EXIT;
2162 static ExitStatus op_ecag(DisasContext *s, DisasOps *o)
2164 /* No cache information provided. */
2165 tcg_gen_movi_i64(o->out, -1);
2166 return NO_EXIT;
2169 static ExitStatus op_efpc(DisasContext *s, DisasOps *o)
2171 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, fpc));
2172 return NO_EXIT;
2175 static ExitStatus op_epsw(DisasContext *s, DisasOps *o)
2177 int r1 = get_field(s->fields, r1);
2178 int r2 = get_field(s->fields, r2);
2179 TCGv_i64 t = tcg_temp_new_i64();
2181 /* Note the "subsequently" in the PoO, which implies a defined result
2182 if r1 == r2. Thus we cannot defer these writes to an output hook. */
2183 tcg_gen_shri_i64(t, psw_mask, 32);
2184 store_reg32_i64(r1, t);
2185 if (r2 != 0) {
2186 store_reg32_i64(r2, psw_mask);
2189 tcg_temp_free_i64(t);
2190 return NO_EXIT;
2193 static ExitStatus op_ex(DisasContext *s, DisasOps *o)
2195 /* ??? Perhaps a better way to implement EXECUTE is to set a bit in
2196 tb->flags, (ab)use the tb->cs_base field as the address of
2197 the template in memory, and grab 8 bits of tb->flags/cflags for
2198 the contents of the register. We would then recognize all this
2199 in gen_intermediate_code_internal, generating code for exactly
2200 one instruction. This new TB then gets executed normally.
2202 On the other hand, this seems to be mostly used for modifying
2203 MVC inside of memcpy, which needs a helper call anyway. So
2204 perhaps this doesn't bear thinking about any further. */
2206 TCGv_i64 tmp;
2208 update_psw_addr(s);
2209 gen_op_calc_cc(s);
2211 tmp = tcg_const_i64(s->next_pc);
2212 gen_helper_ex(cc_op, cpu_env, cc_op, o->in1, o->in2, tmp);
2213 tcg_temp_free_i64(tmp);
2215 return NO_EXIT;
2218 static ExitStatus op_fieb(DisasContext *s, DisasOps *o)
2220 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
2221 gen_helper_fieb(o->out, cpu_env, o->in2, m3);
2222 tcg_temp_free_i32(m3);
2223 return NO_EXIT;
2226 static ExitStatus op_fidb(DisasContext *s, DisasOps *o)
2228 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
2229 gen_helper_fidb(o->out, cpu_env, o->in2, m3);
2230 tcg_temp_free_i32(m3);
2231 return NO_EXIT;
2234 static ExitStatus op_fixb(DisasContext *s, DisasOps *o)
2236 TCGv_i32 m3 = tcg_const_i32(get_field(s->fields, m3));
2237 gen_helper_fixb(o->out, cpu_env, o->in1, o->in2, m3);
2238 return_low128(o->out2);
2239 tcg_temp_free_i32(m3);
2240 return NO_EXIT;
2243 static ExitStatus op_flogr(DisasContext *s, DisasOps *o)
2245 /* We'll use the original input for cc computation, since we get to
2246 compare that against 0, which ought to be better than comparing
2247 the real output against 64. It also lets cc_dst be a convenient
2248 temporary during our computation. */
2249 gen_op_update1_cc_i64(s, CC_OP_FLOGR, o->in2);
2251 /* R1 = IN ? CLZ(IN) : 64. */
2252 tcg_gen_clzi_i64(o->out, o->in2, 64);
2254 /* R1+1 = IN & ~(found bit). Note that we may attempt to shift this
2255 value by 64, which is undefined. But since the shift is 64 iff the
2256 input is zero, we still get the correct result after and'ing. */
2257 tcg_gen_movi_i64(o->out2, 0x8000000000000000ull);
2258 tcg_gen_shr_i64(o->out2, o->out2, o->out);
2259 tcg_gen_andc_i64(o->out2, cc_dst, o->out2);
2260 return NO_EXIT;
2263 static ExitStatus op_icm(DisasContext *s, DisasOps *o)
2265 int m3 = get_field(s->fields, m3);
2266 int pos, len, base = s->insn->data;
2267 TCGv_i64 tmp = tcg_temp_new_i64();
2268 uint64_t ccm;
2270 switch (m3) {
2271 case 0xf:
2272 /* Effectively a 32-bit load. */
2273 tcg_gen_qemu_ld32u(tmp, o->in2, get_mem_index(s));
2274 len = 32;
2275 goto one_insert;
2277 case 0xc:
2278 case 0x6:
2279 case 0x3:
2280 /* Effectively a 16-bit load. */
2281 tcg_gen_qemu_ld16u(tmp, o->in2, get_mem_index(s));
2282 len = 16;
2283 goto one_insert;
2285 case 0x8:
2286 case 0x4:
2287 case 0x2:
2288 case 0x1:
2289 /* Effectively an 8-bit load. */
2290 tcg_gen_qemu_ld8u(tmp, o->in2, get_mem_index(s));
2291 len = 8;
2292 goto one_insert;
2294 one_insert:
2295 pos = base + ctz32(m3) * 8;
2296 tcg_gen_deposit_i64(o->out, o->out, tmp, pos, len);
2297 ccm = ((1ull << len) - 1) << pos;
2298 break;
2300 default:
2301 /* This is going to be a sequence of loads and inserts. */
2302 pos = base + 32 - 8;
2303 ccm = 0;
2304 while (m3) {
2305 if (m3 & 0x8) {
2306 tcg_gen_qemu_ld8u(tmp, o->in2, get_mem_index(s));
2307 tcg_gen_addi_i64(o->in2, o->in2, 1);
2308 tcg_gen_deposit_i64(o->out, o->out, tmp, pos, 8);
2309 ccm |= 0xff << pos;
2311 m3 = (m3 << 1) & 0xf;
2312 pos -= 8;
2314 break;
2317 tcg_gen_movi_i64(tmp, ccm);
2318 gen_op_update2_cc_i64(s, CC_OP_ICM, tmp, o->out);
2319 tcg_temp_free_i64(tmp);
2320 return NO_EXIT;
2323 static ExitStatus op_insi(DisasContext *s, DisasOps *o)
2325 int shift = s->insn->data & 0xff;
2326 int size = s->insn->data >> 8;
2327 tcg_gen_deposit_i64(o->out, o->in1, o->in2, shift, size);
2328 return NO_EXIT;
2331 static ExitStatus op_ipm(DisasContext *s, DisasOps *o)
2333 TCGv_i64 t1;
2335 gen_op_calc_cc(s);
2336 tcg_gen_andi_i64(o->out, o->out, ~0xff000000ull);
2338 t1 = tcg_temp_new_i64();
2339 tcg_gen_shli_i64(t1, psw_mask, 20);
2340 tcg_gen_shri_i64(t1, t1, 36);
2341 tcg_gen_or_i64(o->out, o->out, t1);
2343 tcg_gen_extu_i32_i64(t1, cc_op);
2344 tcg_gen_shli_i64(t1, t1, 28);
2345 tcg_gen_or_i64(o->out, o->out, t1);
2346 tcg_temp_free_i64(t1);
2347 return NO_EXIT;
2350 #ifndef CONFIG_USER_ONLY
2351 static ExitStatus op_ipte(DisasContext *s, DisasOps *o)
2353 check_privileged(s);
2354 gen_helper_ipte(cpu_env, o->in1, o->in2);
2355 return NO_EXIT;
2358 static ExitStatus op_iske(DisasContext *s, DisasOps *o)
2360 check_privileged(s);
2361 gen_helper_iske(o->out, cpu_env, o->in2);
2362 return NO_EXIT;
2364 #endif
2366 static ExitStatus op_ldeb(DisasContext *s, DisasOps *o)
2368 gen_helper_ldeb(o->out, cpu_env, o->in2);
2369 return NO_EXIT;
2372 static ExitStatus op_ledb(DisasContext *s, DisasOps *o)
2374 gen_helper_ledb(o->out, cpu_env, o->in2);
2375 return NO_EXIT;
2378 static ExitStatus op_ldxb(DisasContext *s, DisasOps *o)
2380 gen_helper_ldxb(o->out, cpu_env, o->in1, o->in2);
2381 return NO_EXIT;
2384 static ExitStatus op_lexb(DisasContext *s, DisasOps *o)
2386 gen_helper_lexb(o->out, cpu_env, o->in1, o->in2);
2387 return NO_EXIT;
2390 static ExitStatus op_lxdb(DisasContext *s, DisasOps *o)
2392 gen_helper_lxdb(o->out, cpu_env, o->in2);
2393 return_low128(o->out2);
2394 return NO_EXIT;
2397 static ExitStatus op_lxeb(DisasContext *s, DisasOps *o)
2399 gen_helper_lxeb(o->out, cpu_env, o->in2);
2400 return_low128(o->out2);
2401 return NO_EXIT;
2404 static ExitStatus op_llgt(DisasContext *s, DisasOps *o)
2406 tcg_gen_andi_i64(o->out, o->in2, 0x7fffffff);
2407 return NO_EXIT;
2410 static ExitStatus op_ld8s(DisasContext *s, DisasOps *o)
2412 tcg_gen_qemu_ld8s(o->out, o->in2, get_mem_index(s));
2413 return NO_EXIT;
2416 static ExitStatus op_ld8u(DisasContext *s, DisasOps *o)
2418 tcg_gen_qemu_ld8u(o->out, o->in2, get_mem_index(s));
2419 return NO_EXIT;
2422 static ExitStatus op_ld16s(DisasContext *s, DisasOps *o)
2424 tcg_gen_qemu_ld16s(o->out, o->in2, get_mem_index(s));
2425 return NO_EXIT;
2428 static ExitStatus op_ld16u(DisasContext *s, DisasOps *o)
2430 tcg_gen_qemu_ld16u(o->out, o->in2, get_mem_index(s));
2431 return NO_EXIT;
2434 static ExitStatus op_ld32s(DisasContext *s, DisasOps *o)
2436 tcg_gen_qemu_ld32s(o->out, o->in2, get_mem_index(s));
2437 return NO_EXIT;
2440 static ExitStatus op_ld32u(DisasContext *s, DisasOps *o)
2442 tcg_gen_qemu_ld32u(o->out, o->in2, get_mem_index(s));
2443 return NO_EXIT;
2446 static ExitStatus op_ld64(DisasContext *s, DisasOps *o)
2448 tcg_gen_qemu_ld64(o->out, o->in2, get_mem_index(s));
2449 return NO_EXIT;
2452 static ExitStatus op_lat(DisasContext *s, DisasOps *o)
2454 TCGLabel *lab = gen_new_label();
2455 store_reg32_i64(get_field(s->fields, r1), o->in2);
2456 /* The value is stored even in case of trap. */
2457 tcg_gen_brcondi_i64(TCG_COND_NE, o->in2, 0, lab);
2458 gen_trap(s);
2459 gen_set_label(lab);
2460 return NO_EXIT;
2463 static ExitStatus op_lgat(DisasContext *s, DisasOps *o)
2465 TCGLabel *lab = gen_new_label();
2466 tcg_gen_qemu_ld64(o->out, o->in2, get_mem_index(s));
2467 /* The value is stored even in case of trap. */
2468 tcg_gen_brcondi_i64(TCG_COND_NE, o->out, 0, lab);
2469 gen_trap(s);
2470 gen_set_label(lab);
2471 return NO_EXIT;
2474 static ExitStatus op_lfhat(DisasContext *s, DisasOps *o)
2476 TCGLabel *lab = gen_new_label();
2477 store_reg32h_i64(get_field(s->fields, r1), o->in2);
2478 /* The value is stored even in case of trap. */
2479 tcg_gen_brcondi_i64(TCG_COND_NE, o->in2, 0, lab);
2480 gen_trap(s);
2481 gen_set_label(lab);
2482 return NO_EXIT;
2485 static ExitStatus op_llgfat(DisasContext *s, DisasOps *o)
2487 TCGLabel *lab = gen_new_label();
2488 tcg_gen_qemu_ld32u(o->out, o->in2, get_mem_index(s));
2489 /* The value is stored even in case of trap. */
2490 tcg_gen_brcondi_i64(TCG_COND_NE, o->out, 0, lab);
2491 gen_trap(s);
2492 gen_set_label(lab);
2493 return NO_EXIT;
2496 static ExitStatus op_llgtat(DisasContext *s, DisasOps *o)
2498 TCGLabel *lab = gen_new_label();
2499 tcg_gen_andi_i64(o->out, o->in2, 0x7fffffff);
2500 /* The value is stored even in case of trap. */
2501 tcg_gen_brcondi_i64(TCG_COND_NE, o->out, 0, lab);
2502 gen_trap(s);
2503 gen_set_label(lab);
2504 return NO_EXIT;
2507 static ExitStatus op_loc(DisasContext *s, DisasOps *o)
2509 DisasCompare c;
2511 disas_jcc(s, &c, get_field(s->fields, m3));
2513 if (c.is_64) {
2514 tcg_gen_movcond_i64(c.cond, o->out, c.u.s64.a, c.u.s64.b,
2515 o->in2, o->in1);
2516 free_compare(&c);
2517 } else {
2518 TCGv_i32 t32 = tcg_temp_new_i32();
2519 TCGv_i64 t, z;
2521 tcg_gen_setcond_i32(c.cond, t32, c.u.s32.a, c.u.s32.b);
2522 free_compare(&c);
2524 t = tcg_temp_new_i64();
2525 tcg_gen_extu_i32_i64(t, t32);
2526 tcg_temp_free_i32(t32);
2528 z = tcg_const_i64(0);
2529 tcg_gen_movcond_i64(TCG_COND_NE, o->out, t, z, o->in2, o->in1);
2530 tcg_temp_free_i64(t);
2531 tcg_temp_free_i64(z);
2534 return NO_EXIT;
2537 #ifndef CONFIG_USER_ONLY
2538 static ExitStatus op_lctl(DisasContext *s, DisasOps *o)
2540 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2541 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2542 check_privileged(s);
2543 potential_page_fault(s);
2544 gen_helper_lctl(cpu_env, r1, o->in2, r3);
2545 tcg_temp_free_i32(r1);
2546 tcg_temp_free_i32(r3);
2547 return NO_EXIT;
2550 static ExitStatus op_lctlg(DisasContext *s, DisasOps *o)
2552 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2553 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2554 check_privileged(s);
2555 potential_page_fault(s);
2556 gen_helper_lctlg(cpu_env, r1, o->in2, r3);
2557 tcg_temp_free_i32(r1);
2558 tcg_temp_free_i32(r3);
2559 return NO_EXIT;
2561 static ExitStatus op_lra(DisasContext *s, DisasOps *o)
2563 check_privileged(s);
2564 potential_page_fault(s);
2565 gen_helper_lra(o->out, cpu_env, o->in2);
2566 set_cc_static(s);
2567 return NO_EXIT;
2570 static ExitStatus op_lpsw(DisasContext *s, DisasOps *o)
2572 TCGv_i64 t1, t2;
2574 check_privileged(s);
2575 per_breaking_event(s);
2577 t1 = tcg_temp_new_i64();
2578 t2 = tcg_temp_new_i64();
2579 tcg_gen_qemu_ld32u(t1, o->in2, get_mem_index(s));
2580 tcg_gen_addi_i64(o->in2, o->in2, 4);
2581 tcg_gen_qemu_ld32u(t2, o->in2, get_mem_index(s));
2582 /* Convert the 32-bit PSW_MASK into the 64-bit PSW_MASK. */
2583 tcg_gen_shli_i64(t1, t1, 32);
2584 gen_helper_load_psw(cpu_env, t1, t2);
2585 tcg_temp_free_i64(t1);
2586 tcg_temp_free_i64(t2);
2587 return EXIT_NORETURN;
2590 static ExitStatus op_lpswe(DisasContext *s, DisasOps *o)
2592 TCGv_i64 t1, t2;
2594 check_privileged(s);
2595 per_breaking_event(s);
2597 t1 = tcg_temp_new_i64();
2598 t2 = tcg_temp_new_i64();
2599 tcg_gen_qemu_ld64(t1, o->in2, get_mem_index(s));
2600 tcg_gen_addi_i64(o->in2, o->in2, 8);
2601 tcg_gen_qemu_ld64(t2, o->in2, get_mem_index(s));
2602 gen_helper_load_psw(cpu_env, t1, t2);
2603 tcg_temp_free_i64(t1);
2604 tcg_temp_free_i64(t2);
2605 return EXIT_NORETURN;
2607 #endif
2609 static ExitStatus op_lam(DisasContext *s, DisasOps *o)
2611 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2612 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2613 potential_page_fault(s);
2614 gen_helper_lam(cpu_env, r1, o->in2, r3);
2615 tcg_temp_free_i32(r1);
2616 tcg_temp_free_i32(r3);
2617 return NO_EXIT;
2620 static ExitStatus op_lm32(DisasContext *s, DisasOps *o)
2622 int r1 = get_field(s->fields, r1);
2623 int r3 = get_field(s->fields, r3);
2624 TCGv_i64 t1, t2;
2626 /* Only one register to read. */
2627 t1 = tcg_temp_new_i64();
2628 if (unlikely(r1 == r3)) {
2629 tcg_gen_qemu_ld32u(t1, o->in2, get_mem_index(s));
2630 store_reg32_i64(r1, t1);
2631 tcg_temp_free(t1);
2632 return NO_EXIT;
2635 /* First load the values of the first and last registers to trigger
2636 possible page faults. */
2637 t2 = tcg_temp_new_i64();
2638 tcg_gen_qemu_ld32u(t1, o->in2, get_mem_index(s));
2639 tcg_gen_addi_i64(t2, o->in2, 4 * ((r3 - r1) & 15));
2640 tcg_gen_qemu_ld32u(t2, t2, get_mem_index(s));
2641 store_reg32_i64(r1, t1);
2642 store_reg32_i64(r3, t2);
2644 /* Only two registers to read. */
2645 if (((r1 + 1) & 15) == r3) {
2646 tcg_temp_free(t2);
2647 tcg_temp_free(t1);
2648 return NO_EXIT;
2651 /* Then load the remaining registers. Page fault can't occur. */
2652 r3 = (r3 - 1) & 15;
2653 tcg_gen_movi_i64(t2, 4);
2654 while (r1 != r3) {
2655 r1 = (r1 + 1) & 15;
2656 tcg_gen_add_i64(o->in2, o->in2, t2);
2657 tcg_gen_qemu_ld32u(t1, o->in2, get_mem_index(s));
2658 store_reg32_i64(r1, t1);
2660 tcg_temp_free(t2);
2661 tcg_temp_free(t1);
2663 return NO_EXIT;
2666 static ExitStatus op_lmh(DisasContext *s, DisasOps *o)
2668 int r1 = get_field(s->fields, r1);
2669 int r3 = get_field(s->fields, r3);
2670 TCGv_i64 t1, t2;
2672 /* Only one register to read. */
2673 t1 = tcg_temp_new_i64();
2674 if (unlikely(r1 == r3)) {
2675 tcg_gen_qemu_ld32u(t1, o->in2, get_mem_index(s));
2676 store_reg32h_i64(r1, t1);
2677 tcg_temp_free(t1);
2678 return NO_EXIT;
2681 /* First load the values of the first and last registers to trigger
2682 possible page faults. */
2683 t2 = tcg_temp_new_i64();
2684 tcg_gen_qemu_ld32u(t1, o->in2, get_mem_index(s));
2685 tcg_gen_addi_i64(t2, o->in2, 4 * ((r3 - r1) & 15));
2686 tcg_gen_qemu_ld32u(t2, t2, get_mem_index(s));
2687 store_reg32h_i64(r1, t1);
2688 store_reg32h_i64(r3, t2);
2690 /* Only two registers to read. */
2691 if (((r1 + 1) & 15) == r3) {
2692 tcg_temp_free(t2);
2693 tcg_temp_free(t1);
2694 return NO_EXIT;
2697 /* Then load the remaining registers. Page fault can't occur. */
2698 r3 = (r3 - 1) & 15;
2699 tcg_gen_movi_i64(t2, 4);
2700 while (r1 != r3) {
2701 r1 = (r1 + 1) & 15;
2702 tcg_gen_add_i64(o->in2, o->in2, t2);
2703 tcg_gen_qemu_ld32u(t1, o->in2, get_mem_index(s));
2704 store_reg32h_i64(r1, t1);
2706 tcg_temp_free(t2);
2707 tcg_temp_free(t1);
2709 return NO_EXIT;
2712 static ExitStatus op_lm64(DisasContext *s, DisasOps *o)
2714 int r1 = get_field(s->fields, r1);
2715 int r3 = get_field(s->fields, r3);
2716 TCGv_i64 t1, t2;
2718 /* Only one register to read. */
2719 if (unlikely(r1 == r3)) {
2720 tcg_gen_qemu_ld64(regs[r1], o->in2, get_mem_index(s));
2721 return NO_EXIT;
2724 /* First load the values of the first and last registers to trigger
2725 possible page faults. */
2726 t1 = tcg_temp_new_i64();
2727 t2 = tcg_temp_new_i64();
2728 tcg_gen_qemu_ld64(t1, o->in2, get_mem_index(s));
2729 tcg_gen_addi_i64(t2, o->in2, 8 * ((r3 - r1) & 15));
2730 tcg_gen_qemu_ld64(regs[r3], t2, get_mem_index(s));
2731 tcg_gen_mov_i64(regs[r1], t1);
2732 tcg_temp_free(t2);
2734 /* Only two registers to read. */
2735 if (((r1 + 1) & 15) == r3) {
2736 tcg_temp_free(t1);
2737 return NO_EXIT;
2740 /* Then load the remaining registers. Page fault can't occur. */
2741 r3 = (r3 - 1) & 15;
2742 tcg_gen_movi_i64(t1, 8);
2743 while (r1 != r3) {
2744 r1 = (r1 + 1) & 15;
2745 tcg_gen_add_i64(o->in2, o->in2, t1);
2746 tcg_gen_qemu_ld64(regs[r1], o->in2, get_mem_index(s));
2748 tcg_temp_free(t1);
2750 return NO_EXIT;
2753 #ifndef CONFIG_USER_ONLY
2754 static ExitStatus op_lura(DisasContext *s, DisasOps *o)
2756 check_privileged(s);
2757 potential_page_fault(s);
2758 gen_helper_lura(o->out, cpu_env, o->in2);
2759 return NO_EXIT;
2762 static ExitStatus op_lurag(DisasContext *s, DisasOps *o)
2764 check_privileged(s);
2765 potential_page_fault(s);
2766 gen_helper_lurag(o->out, cpu_env, o->in2);
2767 return NO_EXIT;
2769 #endif
2771 static ExitStatus op_mov2(DisasContext *s, DisasOps *o)
2773 o->out = o->in2;
2774 o->g_out = o->g_in2;
2775 TCGV_UNUSED_I64(o->in2);
2776 o->g_in2 = false;
2777 return NO_EXIT;
2780 static ExitStatus op_mov2e(DisasContext *s, DisasOps *o)
2782 int b2 = get_field(s->fields, b2);
2783 TCGv ar1 = tcg_temp_new_i64();
2785 o->out = o->in2;
2786 o->g_out = o->g_in2;
2787 TCGV_UNUSED_I64(o->in2);
2788 o->g_in2 = false;
2790 switch (s->tb->flags & FLAG_MASK_ASC) {
2791 case PSW_ASC_PRIMARY >> 32:
2792 tcg_gen_movi_i64(ar1, 0);
2793 break;
2794 case PSW_ASC_ACCREG >> 32:
2795 tcg_gen_movi_i64(ar1, 1);
2796 break;
2797 case PSW_ASC_SECONDARY >> 32:
2798 if (b2) {
2799 tcg_gen_ld32u_i64(ar1, cpu_env, offsetof(CPUS390XState, aregs[b2]));
2800 } else {
2801 tcg_gen_movi_i64(ar1, 0);
2803 break;
2804 case PSW_ASC_HOME >> 32:
2805 tcg_gen_movi_i64(ar1, 2);
2806 break;
2809 tcg_gen_st32_i64(ar1, cpu_env, offsetof(CPUS390XState, aregs[1]));
2810 tcg_temp_free_i64(ar1);
2812 return NO_EXIT;
2815 static ExitStatus op_movx(DisasContext *s, DisasOps *o)
2817 o->out = o->in1;
2818 o->out2 = o->in2;
2819 o->g_out = o->g_in1;
2820 o->g_out2 = o->g_in2;
2821 TCGV_UNUSED_I64(o->in1);
2822 TCGV_UNUSED_I64(o->in2);
2823 o->g_in1 = o->g_in2 = false;
2824 return NO_EXIT;
2827 static ExitStatus op_mvc(DisasContext *s, DisasOps *o)
2829 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
2830 potential_page_fault(s);
2831 gen_helper_mvc(cpu_env, l, o->addr1, o->in2);
2832 tcg_temp_free_i32(l);
2833 return NO_EXIT;
2836 static ExitStatus op_mvcl(DisasContext *s, DisasOps *o)
2838 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2839 TCGv_i32 r2 = tcg_const_i32(get_field(s->fields, r2));
2840 potential_page_fault(s);
2841 gen_helper_mvcl(cc_op, cpu_env, r1, r2);
2842 tcg_temp_free_i32(r1);
2843 tcg_temp_free_i32(r2);
2844 set_cc_static(s);
2845 return NO_EXIT;
2848 static ExitStatus op_mvcle(DisasContext *s, DisasOps *o)
2850 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
2851 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
2852 potential_page_fault(s);
2853 gen_helper_mvcle(cc_op, cpu_env, r1, o->in2, r3);
2854 tcg_temp_free_i32(r1);
2855 tcg_temp_free_i32(r3);
2856 set_cc_static(s);
2857 return NO_EXIT;
2860 #ifndef CONFIG_USER_ONLY
2861 static ExitStatus op_mvcp(DisasContext *s, DisasOps *o)
2863 int r1 = get_field(s->fields, l1);
2864 check_privileged(s);
2865 potential_page_fault(s);
2866 gen_helper_mvcp(cc_op, cpu_env, regs[r1], o->addr1, o->in2);
2867 set_cc_static(s);
2868 return NO_EXIT;
2871 static ExitStatus op_mvcs(DisasContext *s, DisasOps *o)
2873 int r1 = get_field(s->fields, l1);
2874 check_privileged(s);
2875 potential_page_fault(s);
2876 gen_helper_mvcs(cc_op, cpu_env, regs[r1], o->addr1, o->in2);
2877 set_cc_static(s);
2878 return NO_EXIT;
2880 #endif
2882 static ExitStatus op_mvpg(DisasContext *s, DisasOps *o)
2884 potential_page_fault(s);
2885 gen_helper_mvpg(cpu_env, regs[0], o->in1, o->in2);
2886 set_cc_static(s);
2887 return NO_EXIT;
2890 static ExitStatus op_mvst(DisasContext *s, DisasOps *o)
2892 potential_page_fault(s);
2893 gen_helper_mvst(o->in1, cpu_env, regs[0], o->in1, o->in2);
2894 set_cc_static(s);
2895 return_low128(o->in2);
2896 return NO_EXIT;
2899 static ExitStatus op_mul(DisasContext *s, DisasOps *o)
2901 tcg_gen_mul_i64(o->out, o->in1, o->in2);
2902 return NO_EXIT;
2905 static ExitStatus op_mul128(DisasContext *s, DisasOps *o)
2907 tcg_gen_mulu2_i64(o->out2, o->out, o->in1, o->in2);
2908 return NO_EXIT;
2911 static ExitStatus op_meeb(DisasContext *s, DisasOps *o)
2913 gen_helper_meeb(o->out, cpu_env, o->in1, o->in2);
2914 return NO_EXIT;
2917 static ExitStatus op_mdeb(DisasContext *s, DisasOps *o)
2919 gen_helper_mdeb(o->out, cpu_env, o->in1, o->in2);
2920 return NO_EXIT;
2923 static ExitStatus op_mdb(DisasContext *s, DisasOps *o)
2925 gen_helper_mdb(o->out, cpu_env, o->in1, o->in2);
2926 return NO_EXIT;
2929 static ExitStatus op_mxb(DisasContext *s, DisasOps *o)
2931 gen_helper_mxb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
2932 return_low128(o->out2);
2933 return NO_EXIT;
2936 static ExitStatus op_mxdb(DisasContext *s, DisasOps *o)
2938 gen_helper_mxdb(o->out, cpu_env, o->out, o->out2, o->in2);
2939 return_low128(o->out2);
2940 return NO_EXIT;
2943 static ExitStatus op_maeb(DisasContext *s, DisasOps *o)
2945 TCGv_i64 r3 = load_freg32_i64(get_field(s->fields, r3));
2946 gen_helper_maeb(o->out, cpu_env, o->in1, o->in2, r3);
2947 tcg_temp_free_i64(r3);
2948 return NO_EXIT;
2951 static ExitStatus op_madb(DisasContext *s, DisasOps *o)
2953 int r3 = get_field(s->fields, r3);
2954 gen_helper_madb(o->out, cpu_env, o->in1, o->in2, fregs[r3]);
2955 return NO_EXIT;
2958 static ExitStatus op_mseb(DisasContext *s, DisasOps *o)
2960 TCGv_i64 r3 = load_freg32_i64(get_field(s->fields, r3));
2961 gen_helper_mseb(o->out, cpu_env, o->in1, o->in2, r3);
2962 tcg_temp_free_i64(r3);
2963 return NO_EXIT;
2966 static ExitStatus op_msdb(DisasContext *s, DisasOps *o)
2968 int r3 = get_field(s->fields, r3);
2969 gen_helper_msdb(o->out, cpu_env, o->in1, o->in2, fregs[r3]);
2970 return NO_EXIT;
2973 static ExitStatus op_nabs(DisasContext *s, DisasOps *o)
2975 TCGv_i64 z, n;
2976 z = tcg_const_i64(0);
2977 n = tcg_temp_new_i64();
2978 tcg_gen_neg_i64(n, o->in2);
2979 tcg_gen_movcond_i64(TCG_COND_GE, o->out, o->in2, z, n, o->in2);
2980 tcg_temp_free_i64(n);
2981 tcg_temp_free_i64(z);
2982 return NO_EXIT;
2985 static ExitStatus op_nabsf32(DisasContext *s, DisasOps *o)
2987 tcg_gen_ori_i64(o->out, o->in2, 0x80000000ull);
2988 return NO_EXIT;
2991 static ExitStatus op_nabsf64(DisasContext *s, DisasOps *o)
2993 tcg_gen_ori_i64(o->out, o->in2, 0x8000000000000000ull);
2994 return NO_EXIT;
2997 static ExitStatus op_nabsf128(DisasContext *s, DisasOps *o)
2999 tcg_gen_ori_i64(o->out, o->in1, 0x8000000000000000ull);
3000 tcg_gen_mov_i64(o->out2, o->in2);
3001 return NO_EXIT;
3004 static ExitStatus op_nc(DisasContext *s, DisasOps *o)
3006 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
3007 potential_page_fault(s);
3008 gen_helper_nc(cc_op, cpu_env, l, o->addr1, o->in2);
3009 tcg_temp_free_i32(l);
3010 set_cc_static(s);
3011 return NO_EXIT;
3014 static ExitStatus op_neg(DisasContext *s, DisasOps *o)
3016 tcg_gen_neg_i64(o->out, o->in2);
3017 return NO_EXIT;
3020 static ExitStatus op_negf32(DisasContext *s, DisasOps *o)
3022 tcg_gen_xori_i64(o->out, o->in2, 0x80000000ull);
3023 return NO_EXIT;
3026 static ExitStatus op_negf64(DisasContext *s, DisasOps *o)
3028 tcg_gen_xori_i64(o->out, o->in2, 0x8000000000000000ull);
3029 return NO_EXIT;
3032 static ExitStatus op_negf128(DisasContext *s, DisasOps *o)
3034 tcg_gen_xori_i64(o->out, o->in1, 0x8000000000000000ull);
3035 tcg_gen_mov_i64(o->out2, o->in2);
3036 return NO_EXIT;
3039 static ExitStatus op_oc(DisasContext *s, DisasOps *o)
3041 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
3042 potential_page_fault(s);
3043 gen_helper_oc(cc_op, cpu_env, l, o->addr1, o->in2);
3044 tcg_temp_free_i32(l);
3045 set_cc_static(s);
3046 return NO_EXIT;
3049 static ExitStatus op_or(DisasContext *s, DisasOps *o)
3051 tcg_gen_or_i64(o->out, o->in1, o->in2);
3052 return NO_EXIT;
3055 static ExitStatus op_ori(DisasContext *s, DisasOps *o)
3057 int shift = s->insn->data & 0xff;
3058 int size = s->insn->data >> 8;
3059 uint64_t mask = ((1ull << size) - 1) << shift;
3061 assert(!o->g_in2);
3062 tcg_gen_shli_i64(o->in2, o->in2, shift);
3063 tcg_gen_or_i64(o->out, o->in1, o->in2);
3065 /* Produce the CC from only the bits manipulated. */
3066 tcg_gen_andi_i64(cc_dst, o->out, mask);
3067 set_cc_nz_u64(s, cc_dst);
3068 return NO_EXIT;
3071 static ExitStatus op_popcnt(DisasContext *s, DisasOps *o)
3073 gen_helper_popcnt(o->out, o->in2);
3074 return NO_EXIT;
3077 #ifndef CONFIG_USER_ONLY
3078 static ExitStatus op_ptlb(DisasContext *s, DisasOps *o)
3080 check_privileged(s);
3081 gen_helper_ptlb(cpu_env);
3082 return NO_EXIT;
3084 #endif
3086 static ExitStatus op_risbg(DisasContext *s, DisasOps *o)
3088 int i3 = get_field(s->fields, i3);
3089 int i4 = get_field(s->fields, i4);
3090 int i5 = get_field(s->fields, i5);
3091 int do_zero = i4 & 0x80;
3092 uint64_t mask, imask, pmask;
3093 int pos, len, rot;
3095 /* Adjust the arguments for the specific insn. */
3096 switch (s->fields->op2) {
3097 case 0x55: /* risbg */
3098 i3 &= 63;
3099 i4 &= 63;
3100 pmask = ~0;
3101 break;
3102 case 0x5d: /* risbhg */
3103 i3 &= 31;
3104 i4 &= 31;
3105 pmask = 0xffffffff00000000ull;
3106 break;
3107 case 0x51: /* risblg */
3108 i3 &= 31;
3109 i4 &= 31;
3110 pmask = 0x00000000ffffffffull;
3111 break;
3112 default:
3113 abort();
3116 /* MASK is the set of bits to be inserted from R2.
3117 Take care for I3/I4 wraparound. */
3118 mask = pmask >> i3;
3119 if (i3 <= i4) {
3120 mask ^= pmask >> i4 >> 1;
3121 } else {
3122 mask |= ~(pmask >> i4 >> 1);
3124 mask &= pmask;
3126 /* IMASK is the set of bits to be kept from R1. In the case of the high/low
3127 insns, we need to keep the other half of the register. */
3128 imask = ~mask | ~pmask;
3129 if (do_zero) {
3130 if (s->fields->op2 == 0x55) {
3131 imask = 0;
3132 } else {
3133 imask = ~pmask;
3137 len = i4 - i3 + 1;
3138 pos = 63 - i4;
3139 rot = i5 & 63;
3140 if (s->fields->op2 == 0x5d) {
3141 pos += 32;
3144 /* In some cases we can implement this with extract. */
3145 if (imask == 0 && pos == 0 && len > 0 && rot + len <= 64) {
3146 tcg_gen_extract_i64(o->out, o->in2, rot, len);
3147 return NO_EXIT;
3150 /* In some cases we can implement this with deposit. */
3151 if (len > 0 && (imask == 0 || ~mask == imask)) {
3152 /* Note that we rotate the bits to be inserted to the lsb, not to
3153 the position as described in the PoO. */
3154 rot = (rot - pos) & 63;
3155 } else {
3156 pos = -1;
3159 /* Rotate the input as necessary. */
3160 tcg_gen_rotli_i64(o->in2, o->in2, rot);
3162 /* Insert the selected bits into the output. */
3163 if (pos >= 0) {
3164 if (imask == 0) {
3165 tcg_gen_deposit_z_i64(o->out, o->in2, pos, len);
3166 } else {
3167 tcg_gen_deposit_i64(o->out, o->out, o->in2, pos, len);
3169 } else if (imask == 0) {
3170 tcg_gen_andi_i64(o->out, o->in2, mask);
3171 } else {
3172 tcg_gen_andi_i64(o->in2, o->in2, mask);
3173 tcg_gen_andi_i64(o->out, o->out, imask);
3174 tcg_gen_or_i64(o->out, o->out, o->in2);
3176 return NO_EXIT;
3179 static ExitStatus op_rosbg(DisasContext *s, DisasOps *o)
3181 int i3 = get_field(s->fields, i3);
3182 int i4 = get_field(s->fields, i4);
3183 int i5 = get_field(s->fields, i5);
3184 uint64_t mask;
3186 /* If this is a test-only form, arrange to discard the result. */
3187 if (i3 & 0x80) {
3188 o->out = tcg_temp_new_i64();
3189 o->g_out = false;
3192 i3 &= 63;
3193 i4 &= 63;
3194 i5 &= 63;
3196 /* MASK is the set of bits to be operated on from R2.
3197 Take care for I3/I4 wraparound. */
3198 mask = ~0ull >> i3;
3199 if (i3 <= i4) {
3200 mask ^= ~0ull >> i4 >> 1;
3201 } else {
3202 mask |= ~(~0ull >> i4 >> 1);
3205 /* Rotate the input as necessary. */
3206 tcg_gen_rotli_i64(o->in2, o->in2, i5);
3208 /* Operate. */
3209 switch (s->fields->op2) {
3210 case 0x55: /* AND */
3211 tcg_gen_ori_i64(o->in2, o->in2, ~mask);
3212 tcg_gen_and_i64(o->out, o->out, o->in2);
3213 break;
3214 case 0x56: /* OR */
3215 tcg_gen_andi_i64(o->in2, o->in2, mask);
3216 tcg_gen_or_i64(o->out, o->out, o->in2);
3217 break;
3218 case 0x57: /* XOR */
3219 tcg_gen_andi_i64(o->in2, o->in2, mask);
3220 tcg_gen_xor_i64(o->out, o->out, o->in2);
3221 break;
3222 default:
3223 abort();
3226 /* Set the CC. */
3227 tcg_gen_andi_i64(cc_dst, o->out, mask);
3228 set_cc_nz_u64(s, cc_dst);
3229 return NO_EXIT;
3232 static ExitStatus op_rev16(DisasContext *s, DisasOps *o)
3234 tcg_gen_bswap16_i64(o->out, o->in2);
3235 return NO_EXIT;
3238 static ExitStatus op_rev32(DisasContext *s, DisasOps *o)
3240 tcg_gen_bswap32_i64(o->out, o->in2);
3241 return NO_EXIT;
3244 static ExitStatus op_rev64(DisasContext *s, DisasOps *o)
3246 tcg_gen_bswap64_i64(o->out, o->in2);
3247 return NO_EXIT;
3250 static ExitStatus op_rll32(DisasContext *s, DisasOps *o)
3252 TCGv_i32 t1 = tcg_temp_new_i32();
3253 TCGv_i32 t2 = tcg_temp_new_i32();
3254 TCGv_i32 to = tcg_temp_new_i32();
3255 tcg_gen_extrl_i64_i32(t1, o->in1);
3256 tcg_gen_extrl_i64_i32(t2, o->in2);
3257 tcg_gen_rotl_i32(to, t1, t2);
3258 tcg_gen_extu_i32_i64(o->out, to);
3259 tcg_temp_free_i32(t1);
3260 tcg_temp_free_i32(t2);
3261 tcg_temp_free_i32(to);
3262 return NO_EXIT;
3265 static ExitStatus op_rll64(DisasContext *s, DisasOps *o)
3267 tcg_gen_rotl_i64(o->out, o->in1, o->in2);
3268 return NO_EXIT;
3271 #ifndef CONFIG_USER_ONLY
3272 static ExitStatus op_rrbe(DisasContext *s, DisasOps *o)
3274 check_privileged(s);
3275 gen_helper_rrbe(cc_op, cpu_env, o->in2);
3276 set_cc_static(s);
3277 return NO_EXIT;
3280 static ExitStatus op_sacf(DisasContext *s, DisasOps *o)
3282 check_privileged(s);
3283 gen_helper_sacf(cpu_env, o->in2);
3284 /* Addressing mode has changed, so end the block. */
3285 return EXIT_PC_STALE;
3287 #endif
3289 static ExitStatus op_sam(DisasContext *s, DisasOps *o)
3291 int sam = s->insn->data;
3292 TCGv_i64 tsam;
3293 uint64_t mask;
3295 switch (sam) {
3296 case 0:
3297 mask = 0xffffff;
3298 break;
3299 case 1:
3300 mask = 0x7fffffff;
3301 break;
3302 default:
3303 mask = -1;
3304 break;
3307 /* Bizarre but true, we check the address of the current insn for the
3308 specification exception, not the next to be executed. Thus the PoO
3309 documents that Bad Things Happen two bytes before the end. */
3310 if (s->pc & ~mask) {
3311 gen_program_exception(s, PGM_SPECIFICATION);
3312 return EXIT_NORETURN;
3314 s->next_pc &= mask;
3316 tsam = tcg_const_i64(sam);
3317 tcg_gen_deposit_i64(psw_mask, psw_mask, tsam, 31, 2);
3318 tcg_temp_free_i64(tsam);
3320 /* Always exit the TB, since we (may have) changed execution mode. */
3321 return EXIT_PC_STALE;
3324 static ExitStatus op_sar(DisasContext *s, DisasOps *o)
3326 int r1 = get_field(s->fields, r1);
3327 tcg_gen_st32_i64(o->in2, cpu_env, offsetof(CPUS390XState, aregs[r1]));
3328 return NO_EXIT;
3331 static ExitStatus op_seb(DisasContext *s, DisasOps *o)
3333 gen_helper_seb(o->out, cpu_env, o->in1, o->in2);
3334 return NO_EXIT;
3337 static ExitStatus op_sdb(DisasContext *s, DisasOps *o)
3339 gen_helper_sdb(o->out, cpu_env, o->in1, o->in2);
3340 return NO_EXIT;
3343 static ExitStatus op_sxb(DisasContext *s, DisasOps *o)
3345 gen_helper_sxb(o->out, cpu_env, o->out, o->out2, o->in1, o->in2);
3346 return_low128(o->out2);
3347 return NO_EXIT;
3350 static ExitStatus op_sqeb(DisasContext *s, DisasOps *o)
3352 gen_helper_sqeb(o->out, cpu_env, o->in2);
3353 return NO_EXIT;
3356 static ExitStatus op_sqdb(DisasContext *s, DisasOps *o)
3358 gen_helper_sqdb(o->out, cpu_env, o->in2);
3359 return NO_EXIT;
3362 static ExitStatus op_sqxb(DisasContext *s, DisasOps *o)
3364 gen_helper_sqxb(o->out, cpu_env, o->in1, o->in2);
3365 return_low128(o->out2);
3366 return NO_EXIT;
3369 #ifndef CONFIG_USER_ONLY
3370 static ExitStatus op_servc(DisasContext *s, DisasOps *o)
3372 check_privileged(s);
3373 potential_page_fault(s);
3374 gen_helper_servc(cc_op, cpu_env, o->in2, o->in1);
3375 set_cc_static(s);
3376 return NO_EXIT;
3379 static ExitStatus op_sigp(DisasContext *s, DisasOps *o)
3381 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
3382 check_privileged(s);
3383 potential_page_fault(s);
3384 gen_helper_sigp(cc_op, cpu_env, o->in2, r1, o->in1);
3385 tcg_temp_free_i32(r1);
3386 return NO_EXIT;
3388 #endif
3390 static ExitStatus op_soc(DisasContext *s, DisasOps *o)
3392 DisasCompare c;
3393 TCGv_i64 a;
3394 TCGLabel *lab;
3395 int r1;
3397 disas_jcc(s, &c, get_field(s->fields, m3));
3399 /* We want to store when the condition is fulfilled, so branch
3400 out when it's not */
3401 c.cond = tcg_invert_cond(c.cond);
3403 lab = gen_new_label();
3404 if (c.is_64) {
3405 tcg_gen_brcond_i64(c.cond, c.u.s64.a, c.u.s64.b, lab);
3406 } else {
3407 tcg_gen_brcond_i32(c.cond, c.u.s32.a, c.u.s32.b, lab);
3409 free_compare(&c);
3411 r1 = get_field(s->fields, r1);
3412 a = get_address(s, 0, get_field(s->fields, b2), get_field(s->fields, d2));
3413 if (s->insn->data) {
3414 tcg_gen_qemu_st64(regs[r1], a, get_mem_index(s));
3415 } else {
3416 tcg_gen_qemu_st32(regs[r1], a, get_mem_index(s));
3418 tcg_temp_free_i64(a);
3420 gen_set_label(lab);
3421 return NO_EXIT;
3424 static ExitStatus op_sla(DisasContext *s, DisasOps *o)
3426 uint64_t sign = 1ull << s->insn->data;
3427 enum cc_op cco = s->insn->data == 31 ? CC_OP_SLA_32 : CC_OP_SLA_64;
3428 gen_op_update2_cc_i64(s, cco, o->in1, o->in2);
3429 tcg_gen_shl_i64(o->out, o->in1, o->in2);
3430 /* The arithmetic left shift is curious in that it does not affect
3431 the sign bit. Copy that over from the source unchanged. */
3432 tcg_gen_andi_i64(o->out, o->out, ~sign);
3433 tcg_gen_andi_i64(o->in1, o->in1, sign);
3434 tcg_gen_or_i64(o->out, o->out, o->in1);
3435 return NO_EXIT;
3438 static ExitStatus op_sll(DisasContext *s, DisasOps *o)
3440 tcg_gen_shl_i64(o->out, o->in1, o->in2);
3441 return NO_EXIT;
3444 static ExitStatus op_sra(DisasContext *s, DisasOps *o)
3446 tcg_gen_sar_i64(o->out, o->in1, o->in2);
3447 return NO_EXIT;
3450 static ExitStatus op_srl(DisasContext *s, DisasOps *o)
3452 tcg_gen_shr_i64(o->out, o->in1, o->in2);
3453 return NO_EXIT;
3456 static ExitStatus op_sfpc(DisasContext *s, DisasOps *o)
3458 gen_helper_sfpc(cpu_env, o->in2);
3459 return NO_EXIT;
3462 static ExitStatus op_sfas(DisasContext *s, DisasOps *o)
3464 gen_helper_sfas(cpu_env, o->in2);
3465 return NO_EXIT;
3468 static ExitStatus op_srnm(DisasContext *s, DisasOps *o)
3470 int b2 = get_field(s->fields, b2);
3471 int d2 = get_field(s->fields, d2);
3472 TCGv_i64 t1 = tcg_temp_new_i64();
3473 TCGv_i64 t2 = tcg_temp_new_i64();
3474 int mask, pos, len;
3476 switch (s->fields->op2) {
3477 case 0x99: /* SRNM */
3478 pos = 0, len = 2;
3479 break;
3480 case 0xb8: /* SRNMB */
3481 pos = 0, len = 3;
3482 break;
3483 case 0xb9: /* SRNMT */
3484 pos = 4, len = 3;
3485 break;
3486 default:
3487 tcg_abort();
3489 mask = (1 << len) - 1;
3491 /* Insert the value into the appropriate field of the FPC. */
3492 if (b2 == 0) {
3493 tcg_gen_movi_i64(t1, d2 & mask);
3494 } else {
3495 tcg_gen_addi_i64(t1, regs[b2], d2);
3496 tcg_gen_andi_i64(t1, t1, mask);
3498 tcg_gen_ld32u_i64(t2, cpu_env, offsetof(CPUS390XState, fpc));
3499 tcg_gen_deposit_i64(t2, t2, t1, pos, len);
3500 tcg_temp_free_i64(t1);
3502 /* Then install the new FPC to set the rounding mode in fpu_status. */
3503 gen_helper_sfpc(cpu_env, t2);
3504 tcg_temp_free_i64(t2);
3505 return NO_EXIT;
3508 #ifndef CONFIG_USER_ONLY
3509 static ExitStatus op_spka(DisasContext *s, DisasOps *o)
3511 check_privileged(s);
3512 tcg_gen_shri_i64(o->in2, o->in2, 4);
3513 tcg_gen_deposit_i64(psw_mask, psw_mask, o->in2, PSW_SHIFT_KEY - 4, 4);
3514 return NO_EXIT;
3517 static ExitStatus op_sske(DisasContext *s, DisasOps *o)
3519 check_privileged(s);
3520 gen_helper_sske(cpu_env, o->in1, o->in2);
3521 return NO_EXIT;
3524 static ExitStatus op_ssm(DisasContext *s, DisasOps *o)
3526 check_privileged(s);
3527 tcg_gen_deposit_i64(psw_mask, psw_mask, o->in2, 56, 8);
3528 return NO_EXIT;
3531 static ExitStatus op_stap(DisasContext *s, DisasOps *o)
3533 check_privileged(s);
3534 /* ??? Surely cpu address != cpu number. In any case the previous
3535 version of this stored more than the required half-word, so it
3536 is unlikely this has ever been tested. */
3537 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, cpu_num));
3538 return NO_EXIT;
3541 static ExitStatus op_stck(DisasContext *s, DisasOps *o)
3543 gen_helper_stck(o->out, cpu_env);
3544 /* ??? We don't implement clock states. */
3545 gen_op_movi_cc(s, 0);
3546 return NO_EXIT;
3549 static ExitStatus op_stcke(DisasContext *s, DisasOps *o)
3551 TCGv_i64 c1 = tcg_temp_new_i64();
3552 TCGv_i64 c2 = tcg_temp_new_i64();
3553 gen_helper_stck(c1, cpu_env);
3554 /* Shift the 64-bit value into its place as a zero-extended
3555 104-bit value. Note that "bit positions 64-103 are always
3556 non-zero so that they compare differently to STCK"; we set
3557 the least significant bit to 1. */
3558 tcg_gen_shli_i64(c2, c1, 56);
3559 tcg_gen_shri_i64(c1, c1, 8);
3560 tcg_gen_ori_i64(c2, c2, 0x10000);
3561 tcg_gen_qemu_st64(c1, o->in2, get_mem_index(s));
3562 tcg_gen_addi_i64(o->in2, o->in2, 8);
3563 tcg_gen_qemu_st64(c2, o->in2, get_mem_index(s));
3564 tcg_temp_free_i64(c1);
3565 tcg_temp_free_i64(c2);
3566 /* ??? We don't implement clock states. */
3567 gen_op_movi_cc(s, 0);
3568 return NO_EXIT;
3571 static ExitStatus op_sckc(DisasContext *s, DisasOps *o)
3573 check_privileged(s);
3574 gen_helper_sckc(cpu_env, o->in2);
3575 return NO_EXIT;
3578 static ExitStatus op_stckc(DisasContext *s, DisasOps *o)
3580 check_privileged(s);
3581 gen_helper_stckc(o->out, cpu_env);
3582 return NO_EXIT;
3585 static ExitStatus op_stctg(DisasContext *s, DisasOps *o)
3587 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
3588 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
3589 check_privileged(s);
3590 potential_page_fault(s);
3591 gen_helper_stctg(cpu_env, r1, o->in2, r3);
3592 tcg_temp_free_i32(r1);
3593 tcg_temp_free_i32(r3);
3594 return NO_EXIT;
3597 static ExitStatus op_stctl(DisasContext *s, DisasOps *o)
3599 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
3600 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
3601 check_privileged(s);
3602 potential_page_fault(s);
3603 gen_helper_stctl(cpu_env, r1, o->in2, r3);
3604 tcg_temp_free_i32(r1);
3605 tcg_temp_free_i32(r3);
3606 return NO_EXIT;
3609 static ExitStatus op_stidp(DisasContext *s, DisasOps *o)
3611 TCGv_i64 t1 = tcg_temp_new_i64();
3613 check_privileged(s);
3614 tcg_gen_ld32u_i64(o->out, cpu_env, offsetof(CPUS390XState, cpu_num));
3615 tcg_gen_ld32u_i64(t1, cpu_env, offsetof(CPUS390XState, machine_type));
3616 tcg_gen_deposit_i64(o->out, o->out, t1, 32, 32);
3617 tcg_temp_free_i64(t1);
3619 return NO_EXIT;
3622 static ExitStatus op_spt(DisasContext *s, DisasOps *o)
3624 check_privileged(s);
3625 gen_helper_spt(cpu_env, o->in2);
3626 return NO_EXIT;
3629 static ExitStatus op_stfl(DisasContext *s, DisasOps *o)
3631 TCGv_i64 f, a;
3632 /* We really ought to have more complete indication of facilities
3633 that we implement. Address this when STFLE is implemented. */
3634 check_privileged(s);
3635 f = tcg_const_i64(0xc0000000);
3636 a = tcg_const_i64(200);
3637 tcg_gen_qemu_st32(f, a, get_mem_index(s));
3638 tcg_temp_free_i64(f);
3639 tcg_temp_free_i64(a);
3640 return NO_EXIT;
3643 static ExitStatus op_stpt(DisasContext *s, DisasOps *o)
3645 check_privileged(s);
3646 gen_helper_stpt(o->out, cpu_env);
3647 return NO_EXIT;
3650 static ExitStatus op_stsi(DisasContext *s, DisasOps *o)
3652 check_privileged(s);
3653 potential_page_fault(s);
3654 gen_helper_stsi(cc_op, cpu_env, o->in2, regs[0], regs[1]);
3655 set_cc_static(s);
3656 return NO_EXIT;
3659 static ExitStatus op_spx(DisasContext *s, DisasOps *o)
3661 check_privileged(s);
3662 gen_helper_spx(cpu_env, o->in2);
3663 return NO_EXIT;
3666 static ExitStatus op_xsch(DisasContext *s, DisasOps *o)
3668 check_privileged(s);
3669 potential_page_fault(s);
3670 gen_helper_xsch(cpu_env, regs[1]);
3671 set_cc_static(s);
3672 return NO_EXIT;
3675 static ExitStatus op_csch(DisasContext *s, DisasOps *o)
3677 check_privileged(s);
3678 potential_page_fault(s);
3679 gen_helper_csch(cpu_env, regs[1]);
3680 set_cc_static(s);
3681 return NO_EXIT;
3684 static ExitStatus op_hsch(DisasContext *s, DisasOps *o)
3686 check_privileged(s);
3687 potential_page_fault(s);
3688 gen_helper_hsch(cpu_env, regs[1]);
3689 set_cc_static(s);
3690 return NO_EXIT;
3693 static ExitStatus op_msch(DisasContext *s, DisasOps *o)
3695 check_privileged(s);
3696 potential_page_fault(s);
3697 gen_helper_msch(cpu_env, regs[1], o->in2);
3698 set_cc_static(s);
3699 return NO_EXIT;
3702 static ExitStatus op_rchp(DisasContext *s, DisasOps *o)
3704 check_privileged(s);
3705 potential_page_fault(s);
3706 gen_helper_rchp(cpu_env, regs[1]);
3707 set_cc_static(s);
3708 return NO_EXIT;
3711 static ExitStatus op_rsch(DisasContext *s, DisasOps *o)
3713 check_privileged(s);
3714 potential_page_fault(s);
3715 gen_helper_rsch(cpu_env, regs[1]);
3716 set_cc_static(s);
3717 return NO_EXIT;
3720 static ExitStatus op_ssch(DisasContext *s, DisasOps *o)
3722 check_privileged(s);
3723 potential_page_fault(s);
3724 gen_helper_ssch(cpu_env, regs[1], o->in2);
3725 set_cc_static(s);
3726 return NO_EXIT;
3729 static ExitStatus op_stsch(DisasContext *s, DisasOps *o)
3731 check_privileged(s);
3732 potential_page_fault(s);
3733 gen_helper_stsch(cpu_env, regs[1], o->in2);
3734 set_cc_static(s);
3735 return NO_EXIT;
3738 static ExitStatus op_tsch(DisasContext *s, DisasOps *o)
3740 check_privileged(s);
3741 potential_page_fault(s);
3742 gen_helper_tsch(cpu_env, regs[1], o->in2);
3743 set_cc_static(s);
3744 return NO_EXIT;
3747 static ExitStatus op_chsc(DisasContext *s, DisasOps *o)
3749 check_privileged(s);
3750 potential_page_fault(s);
3751 gen_helper_chsc(cpu_env, o->in2);
3752 set_cc_static(s);
3753 return NO_EXIT;
3756 static ExitStatus op_stpx(DisasContext *s, DisasOps *o)
3758 check_privileged(s);
3759 tcg_gen_ld_i64(o->out, cpu_env, offsetof(CPUS390XState, psa));
3760 tcg_gen_andi_i64(o->out, o->out, 0x7fffe000);
3761 return NO_EXIT;
3764 static ExitStatus op_stnosm(DisasContext *s, DisasOps *o)
3766 uint64_t i2 = get_field(s->fields, i2);
3767 TCGv_i64 t;
3769 check_privileged(s);
3771 /* It is important to do what the instruction name says: STORE THEN.
3772 If we let the output hook perform the store then if we fault and
3773 restart, we'll have the wrong SYSTEM MASK in place. */
3774 t = tcg_temp_new_i64();
3775 tcg_gen_shri_i64(t, psw_mask, 56);
3776 tcg_gen_qemu_st8(t, o->addr1, get_mem_index(s));
3777 tcg_temp_free_i64(t);
3779 if (s->fields->op == 0xac) {
3780 tcg_gen_andi_i64(psw_mask, psw_mask,
3781 (i2 << 56) | 0x00ffffffffffffffull);
3782 } else {
3783 tcg_gen_ori_i64(psw_mask, psw_mask, i2 << 56);
3785 return NO_EXIT;
3788 static ExitStatus op_stura(DisasContext *s, DisasOps *o)
3790 check_privileged(s);
3791 potential_page_fault(s);
3792 gen_helper_stura(cpu_env, o->in2, o->in1);
3793 return NO_EXIT;
3796 static ExitStatus op_sturg(DisasContext *s, DisasOps *o)
3798 check_privileged(s);
3799 potential_page_fault(s);
3800 gen_helper_sturg(cpu_env, o->in2, o->in1);
3801 return NO_EXIT;
3803 #endif
3805 static ExitStatus op_st8(DisasContext *s, DisasOps *o)
3807 tcg_gen_qemu_st8(o->in1, o->in2, get_mem_index(s));
3808 return NO_EXIT;
3811 static ExitStatus op_st16(DisasContext *s, DisasOps *o)
3813 tcg_gen_qemu_st16(o->in1, o->in2, get_mem_index(s));
3814 return NO_EXIT;
3817 static ExitStatus op_st32(DisasContext *s, DisasOps *o)
3819 tcg_gen_qemu_st32(o->in1, o->in2, get_mem_index(s));
3820 return NO_EXIT;
3823 static ExitStatus op_st64(DisasContext *s, DisasOps *o)
3825 tcg_gen_qemu_st64(o->in1, o->in2, get_mem_index(s));
3826 return NO_EXIT;
3829 static ExitStatus op_stam(DisasContext *s, DisasOps *o)
3831 TCGv_i32 r1 = tcg_const_i32(get_field(s->fields, r1));
3832 TCGv_i32 r3 = tcg_const_i32(get_field(s->fields, r3));
3833 potential_page_fault(s);
3834 gen_helper_stam(cpu_env, r1, o->in2, r3);
3835 tcg_temp_free_i32(r1);
3836 tcg_temp_free_i32(r3);
3837 return NO_EXIT;
3840 static ExitStatus op_stcm(DisasContext *s, DisasOps *o)
3842 int m3 = get_field(s->fields, m3);
3843 int pos, base = s->insn->data;
3844 TCGv_i64 tmp = tcg_temp_new_i64();
3846 pos = base + ctz32(m3) * 8;
3847 switch (m3) {
3848 case 0xf:
3849 /* Effectively a 32-bit store. */
3850 tcg_gen_shri_i64(tmp, o->in1, pos);
3851 tcg_gen_qemu_st32(tmp, o->in2, get_mem_index(s));
3852 break;
3854 case 0xc:
3855 case 0x6:
3856 case 0x3:
3857 /* Effectively a 16-bit store. */
3858 tcg_gen_shri_i64(tmp, o->in1, pos);
3859 tcg_gen_qemu_st16(tmp, o->in2, get_mem_index(s));
3860 break;
3862 case 0x8:
3863 case 0x4:
3864 case 0x2:
3865 case 0x1:
3866 /* Effectively an 8-bit store. */
3867 tcg_gen_shri_i64(tmp, o->in1, pos);
3868 tcg_gen_qemu_st8(tmp, o->in2, get_mem_index(s));
3869 break;
3871 default:
3872 /* This is going to be a sequence of shifts and stores. */
3873 pos = base + 32 - 8;
3874 while (m3) {
3875 if (m3 & 0x8) {
3876 tcg_gen_shri_i64(tmp, o->in1, pos);
3877 tcg_gen_qemu_st8(tmp, o->in2, get_mem_index(s));
3878 tcg_gen_addi_i64(o->in2, o->in2, 1);
3880 m3 = (m3 << 1) & 0xf;
3881 pos -= 8;
3883 break;
3885 tcg_temp_free_i64(tmp);
3886 return NO_EXIT;
3889 static ExitStatus op_stm(DisasContext *s, DisasOps *o)
3891 int r1 = get_field(s->fields, r1);
3892 int r3 = get_field(s->fields, r3);
3893 int size = s->insn->data;
3894 TCGv_i64 tsize = tcg_const_i64(size);
3896 while (1) {
3897 if (size == 8) {
3898 tcg_gen_qemu_st64(regs[r1], o->in2, get_mem_index(s));
3899 } else {
3900 tcg_gen_qemu_st32(regs[r1], o->in2, get_mem_index(s));
3902 if (r1 == r3) {
3903 break;
3905 tcg_gen_add_i64(o->in2, o->in2, tsize);
3906 r1 = (r1 + 1) & 15;
3909 tcg_temp_free_i64(tsize);
3910 return NO_EXIT;
3913 static ExitStatus op_stmh(DisasContext *s, DisasOps *o)
3915 int r1 = get_field(s->fields, r1);
3916 int r3 = get_field(s->fields, r3);
3917 TCGv_i64 t = tcg_temp_new_i64();
3918 TCGv_i64 t4 = tcg_const_i64(4);
3919 TCGv_i64 t32 = tcg_const_i64(32);
3921 while (1) {
3922 tcg_gen_shl_i64(t, regs[r1], t32);
3923 tcg_gen_qemu_st32(t, o->in2, get_mem_index(s));
3924 if (r1 == r3) {
3925 break;
3927 tcg_gen_add_i64(o->in2, o->in2, t4);
3928 r1 = (r1 + 1) & 15;
3931 tcg_temp_free_i64(t);
3932 tcg_temp_free_i64(t4);
3933 tcg_temp_free_i64(t32);
3934 return NO_EXIT;
3937 static ExitStatus op_srst(DisasContext *s, DisasOps *o)
3939 potential_page_fault(s);
3940 gen_helper_srst(o->in1, cpu_env, regs[0], o->in1, o->in2);
3941 set_cc_static(s);
3942 return_low128(o->in2);
3943 return NO_EXIT;
3946 static ExitStatus op_sub(DisasContext *s, DisasOps *o)
3948 tcg_gen_sub_i64(o->out, o->in1, o->in2);
3949 return NO_EXIT;
3952 static ExitStatus op_subb(DisasContext *s, DisasOps *o)
3954 DisasCompare cmp;
3955 TCGv_i64 borrow;
3957 tcg_gen_sub_i64(o->out, o->in1, o->in2);
3959 /* The !borrow flag is the msb of CC. Since we want the inverse of
3960 that, we ask for a comparison of CC=0 | CC=1 -> mask of 8 | 4. */
3961 disas_jcc(s, &cmp, 8 | 4);
3962 borrow = tcg_temp_new_i64();
3963 if (cmp.is_64) {
3964 tcg_gen_setcond_i64(cmp.cond, borrow, cmp.u.s64.a, cmp.u.s64.b);
3965 } else {
3966 TCGv_i32 t = tcg_temp_new_i32();
3967 tcg_gen_setcond_i32(cmp.cond, t, cmp.u.s32.a, cmp.u.s32.b);
3968 tcg_gen_extu_i32_i64(borrow, t);
3969 tcg_temp_free_i32(t);
3971 free_compare(&cmp);
3973 tcg_gen_sub_i64(o->out, o->out, borrow);
3974 tcg_temp_free_i64(borrow);
3975 return NO_EXIT;
3978 static ExitStatus op_svc(DisasContext *s, DisasOps *o)
3980 TCGv_i32 t;
3982 update_psw_addr(s);
3983 update_cc_op(s);
3985 t = tcg_const_i32(get_field(s->fields, i1) & 0xff);
3986 tcg_gen_st_i32(t, cpu_env, offsetof(CPUS390XState, int_svc_code));
3987 tcg_temp_free_i32(t);
3989 t = tcg_const_i32(s->next_pc - s->pc);
3990 tcg_gen_st_i32(t, cpu_env, offsetof(CPUS390XState, int_svc_ilen));
3991 tcg_temp_free_i32(t);
3993 gen_exception(EXCP_SVC);
3994 return EXIT_NORETURN;
3997 static ExitStatus op_tceb(DisasContext *s, DisasOps *o)
3999 gen_helper_tceb(cc_op, cpu_env, o->in1, o->in2);
4000 set_cc_static(s);
4001 return NO_EXIT;
4004 static ExitStatus op_tcdb(DisasContext *s, DisasOps *o)
4006 gen_helper_tcdb(cc_op, cpu_env, o->in1, o->in2);
4007 set_cc_static(s);
4008 return NO_EXIT;
4011 static ExitStatus op_tcxb(DisasContext *s, DisasOps *o)
4013 gen_helper_tcxb(cc_op, cpu_env, o->out, o->out2, o->in2);
4014 set_cc_static(s);
4015 return NO_EXIT;
4018 #ifndef CONFIG_USER_ONLY
4019 static ExitStatus op_tprot(DisasContext *s, DisasOps *o)
4021 potential_page_fault(s);
4022 gen_helper_tprot(cc_op, o->addr1, o->in2);
4023 set_cc_static(s);
4024 return NO_EXIT;
4026 #endif
4028 static ExitStatus op_tr(DisasContext *s, DisasOps *o)
4030 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
4031 potential_page_fault(s);
4032 gen_helper_tr(cpu_env, l, o->addr1, o->in2);
4033 tcg_temp_free_i32(l);
4034 set_cc_static(s);
4035 return NO_EXIT;
4038 static ExitStatus op_tre(DisasContext *s, DisasOps *o)
4040 potential_page_fault(s);
4041 gen_helper_tre(o->out, cpu_env, o->out, o->out2, o->in2);
4042 return_low128(o->out2);
4043 set_cc_static(s);
4044 return NO_EXIT;
4047 static ExitStatus op_trt(DisasContext *s, DisasOps *o)
4049 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
4050 potential_page_fault(s);
4051 gen_helper_trt(cc_op, cpu_env, l, o->addr1, o->in2);
4052 tcg_temp_free_i32(l);
4053 set_cc_static(s);
4054 return NO_EXIT;
4057 static ExitStatus op_unpk(DisasContext *s, DisasOps *o)
4059 TCGv_i32 l = tcg_const_i32(get_field(s->fields, l1));
4060 potential_page_fault(s);
4061 gen_helper_unpk(cpu_env, l, o->addr1, o->in2);
4062 tcg_temp_free_i32(l);
4063 return NO_EXIT;
4066 static ExitStatus op_xc(DisasContext *s, DisasOps *o)
4068 int d1 = get_field(s->fields, d1);
4069 int d2 = get_field(s->fields, d2);
4070 int b1 = get_field(s->fields, b1);
4071 int b2 = get_field(s->fields, b2);
4072 int l = get_field(s->fields, l1);
4073 TCGv_i32 t32;
4075 o->addr1 = get_address(s, 0, b1, d1);
4077 /* If the addresses are identical, this is a store/memset of zero. */
4078 if (b1 == b2 && d1 == d2 && (l + 1) <= 32) {
4079 o->in2 = tcg_const_i64(0);
4081 l++;
4082 while (l >= 8) {
4083 tcg_gen_qemu_st64(o->in2, o->addr1, get_mem_index(s));
4084 l -= 8;
4085 if (l > 0) {
4086 tcg_gen_addi_i64(o->addr1, o->addr1, 8);
4089 if (l >= 4) {
4090 tcg_gen_qemu_st32(o->in2, o->addr1, get_mem_index(s));
4091 l -= 4;
4092 if (l > 0) {
4093 tcg_gen_addi_i64(o->addr1, o->addr1, 4);
4096 if (l >= 2) {
4097 tcg_gen_qemu_st16(o->in2, o->addr1, get_mem_index(s));
4098 l -= 2;
4099 if (l > 0) {
4100 tcg_gen_addi_i64(o->addr1, o->addr1, 2);
4103 if (l) {
4104 tcg_gen_qemu_st8(o->in2, o->addr1, get_mem_index(s));
4106 gen_op_movi_cc(s, 0);
4107 return NO_EXIT;
4110 /* But in general we'll defer to a helper. */
4111 o->in2 = get_address(s, 0, b2, d2);
4112 t32 = tcg_const_i32(l);
4113 potential_page_fault(s);
4114 gen_helper_xc(cc_op, cpu_env, t32, o->addr1, o->in2);
4115 tcg_temp_free_i32(t32);
4116 set_cc_static(s);
4117 return NO_EXIT;
4120 static ExitStatus op_xor(DisasContext *s, DisasOps *o)
4122 tcg_gen_xor_i64(o->out, o->in1, o->in2);
4123 return NO_EXIT;
4126 static ExitStatus op_xori(DisasContext *s, DisasOps *o)
4128 int shift = s->insn->data & 0xff;
4129 int size = s->insn->data >> 8;
4130 uint64_t mask = ((1ull << size) - 1) << shift;
4132 assert(!o->g_in2);
4133 tcg_gen_shli_i64(o->in2, o->in2, shift);
4134 tcg_gen_xor_i64(o->out, o->in1, o->in2);
4136 /* Produce the CC from only the bits manipulated. */
4137 tcg_gen_andi_i64(cc_dst, o->out, mask);
4138 set_cc_nz_u64(s, cc_dst);
4139 return NO_EXIT;
4142 static ExitStatus op_zero(DisasContext *s, DisasOps *o)
4144 o->out = tcg_const_i64(0);
4145 return NO_EXIT;
4148 static ExitStatus op_zero2(DisasContext *s, DisasOps *o)
4150 o->out = tcg_const_i64(0);
4151 o->out2 = o->out;
4152 o->g_out2 = true;
4153 return NO_EXIT;
4156 /* ====================================================================== */
4157 /* The "Cc OUTput" generators. Given the generated output (and in some cases
4158 the original inputs), update the various cc data structures in order to
4159 be able to compute the new condition code. */
4161 static void cout_abs32(DisasContext *s, DisasOps *o)
4163 gen_op_update1_cc_i64(s, CC_OP_ABS_32, o->out);
4166 static void cout_abs64(DisasContext *s, DisasOps *o)
4168 gen_op_update1_cc_i64(s, CC_OP_ABS_64, o->out);
4171 static void cout_adds32(DisasContext *s, DisasOps *o)
4173 gen_op_update3_cc_i64(s, CC_OP_ADD_32, o->in1, o->in2, o->out);
4176 static void cout_adds64(DisasContext *s, DisasOps *o)
4178 gen_op_update3_cc_i64(s, CC_OP_ADD_64, o->in1, o->in2, o->out);
4181 static void cout_addu32(DisasContext *s, DisasOps *o)
4183 gen_op_update3_cc_i64(s, CC_OP_ADDU_32, o->in1, o->in2, o->out);
4186 static void cout_addu64(DisasContext *s, DisasOps *o)
4188 gen_op_update3_cc_i64(s, CC_OP_ADDU_64, o->in1, o->in2, o->out);
4191 static void cout_addc32(DisasContext *s, DisasOps *o)
4193 gen_op_update3_cc_i64(s, CC_OP_ADDC_32, o->in1, o->in2, o->out);
4196 static void cout_addc64(DisasContext *s, DisasOps *o)
4198 gen_op_update3_cc_i64(s, CC_OP_ADDC_64, o->in1, o->in2, o->out);
4201 static void cout_cmps32(DisasContext *s, DisasOps *o)
4203 gen_op_update2_cc_i64(s, CC_OP_LTGT_32, o->in1, o->in2);
4206 static void cout_cmps64(DisasContext *s, DisasOps *o)
4208 gen_op_update2_cc_i64(s, CC_OP_LTGT_64, o->in1, o->in2);
4211 static void cout_cmpu32(DisasContext *s, DisasOps *o)
4213 gen_op_update2_cc_i64(s, CC_OP_LTUGTU_32, o->in1, o->in2);
4216 static void cout_cmpu64(DisasContext *s, DisasOps *o)
4218 gen_op_update2_cc_i64(s, CC_OP_LTUGTU_64, o->in1, o->in2);
4221 static void cout_f32(DisasContext *s, DisasOps *o)
4223 gen_op_update1_cc_i64(s, CC_OP_NZ_F32, o->out);
4226 static void cout_f64(DisasContext *s, DisasOps *o)
4228 gen_op_update1_cc_i64(s, CC_OP_NZ_F64, o->out);
4231 static void cout_f128(DisasContext *s, DisasOps *o)
4233 gen_op_update2_cc_i64(s, CC_OP_NZ_F128, o->out, o->out2);
4236 static void cout_nabs32(DisasContext *s, DisasOps *o)
4238 gen_op_update1_cc_i64(s, CC_OP_NABS_32, o->out);
4241 static void cout_nabs64(DisasContext *s, DisasOps *o)
4243 gen_op_update1_cc_i64(s, CC_OP_NABS_64, o->out);
4246 static void cout_neg32(DisasContext *s, DisasOps *o)
4248 gen_op_update1_cc_i64(s, CC_OP_COMP_32, o->out);
4251 static void cout_neg64(DisasContext *s, DisasOps *o)
4253 gen_op_update1_cc_i64(s, CC_OP_COMP_64, o->out);
4256 static void cout_nz32(DisasContext *s, DisasOps *o)
4258 tcg_gen_ext32u_i64(cc_dst, o->out);
4259 gen_op_update1_cc_i64(s, CC_OP_NZ, cc_dst);
4262 static void cout_nz64(DisasContext *s, DisasOps *o)
4264 gen_op_update1_cc_i64(s, CC_OP_NZ, o->out);
4267 static void cout_s32(DisasContext *s, DisasOps *o)
4269 gen_op_update1_cc_i64(s, CC_OP_LTGT0_32, o->out);
4272 static void cout_s64(DisasContext *s, DisasOps *o)
4274 gen_op_update1_cc_i64(s, CC_OP_LTGT0_64, o->out);
4277 static void cout_subs32(DisasContext *s, DisasOps *o)
4279 gen_op_update3_cc_i64(s, CC_OP_SUB_32, o->in1, o->in2, o->out);
4282 static void cout_subs64(DisasContext *s, DisasOps *o)
4284 gen_op_update3_cc_i64(s, CC_OP_SUB_64, o->in1, o->in2, o->out);
4287 static void cout_subu32(DisasContext *s, DisasOps *o)
4289 gen_op_update3_cc_i64(s, CC_OP_SUBU_32, o->in1, o->in2, o->out);
4292 static void cout_subu64(DisasContext *s, DisasOps *o)
4294 gen_op_update3_cc_i64(s, CC_OP_SUBU_64, o->in1, o->in2, o->out);
4297 static void cout_subb32(DisasContext *s, DisasOps *o)
4299 gen_op_update3_cc_i64(s, CC_OP_SUBB_32, o->in1, o->in2, o->out);
4302 static void cout_subb64(DisasContext *s, DisasOps *o)
4304 gen_op_update3_cc_i64(s, CC_OP_SUBB_64, o->in1, o->in2, o->out);
4307 static void cout_tm32(DisasContext *s, DisasOps *o)
4309 gen_op_update2_cc_i64(s, CC_OP_TM_32, o->in1, o->in2);
4312 static void cout_tm64(DisasContext *s, DisasOps *o)
4314 gen_op_update2_cc_i64(s, CC_OP_TM_64, o->in1, o->in2);
4317 /* ====================================================================== */
4318 /* The "PREParation" generators. These initialize the DisasOps.OUT fields
4319 with the TCG register to which we will write. Used in combination with
4320 the "wout" generators, in some cases we need a new temporary, and in
4321 some cases we can write to a TCG global. */
4323 static void prep_new(DisasContext *s, DisasFields *f, DisasOps *o)
4325 o->out = tcg_temp_new_i64();
4327 #define SPEC_prep_new 0
4329 static void prep_new_P(DisasContext *s, DisasFields *f, DisasOps *o)
4331 o->out = tcg_temp_new_i64();
4332 o->out2 = tcg_temp_new_i64();
4334 #define SPEC_prep_new_P 0
4336 static void prep_r1(DisasContext *s, DisasFields *f, DisasOps *o)
4338 o->out = regs[get_field(f, r1)];
4339 o->g_out = true;
4341 #define SPEC_prep_r1 0
4343 static void prep_r1_P(DisasContext *s, DisasFields *f, DisasOps *o)
4345 int r1 = get_field(f, r1);
4346 o->out = regs[r1];
4347 o->out2 = regs[r1 + 1];
4348 o->g_out = o->g_out2 = true;
4350 #define SPEC_prep_r1_P SPEC_r1_even
4352 static void prep_f1(DisasContext *s, DisasFields *f, DisasOps *o)
4354 o->out = fregs[get_field(f, r1)];
4355 o->g_out = true;
4357 #define SPEC_prep_f1 0
4359 static void prep_x1(DisasContext *s, DisasFields *f, DisasOps *o)
4361 int r1 = get_field(f, r1);
4362 o->out = fregs[r1];
4363 o->out2 = fregs[r1 + 2];
4364 o->g_out = o->g_out2 = true;
4366 #define SPEC_prep_x1 SPEC_r1_f128
4368 /* ====================================================================== */
4369 /* The "Write OUTput" generators. These generally perform some non-trivial
4370 copy of data to TCG globals, or to main memory. The trivial cases are
4371 generally handled by having a "prep" generator install the TCG global
4372 as the destination of the operation. */
4374 static void wout_r1(DisasContext *s, DisasFields *f, DisasOps *o)
4376 store_reg(get_field(f, r1), o->out);
4378 #define SPEC_wout_r1 0
4380 static void wout_r1_8(DisasContext *s, DisasFields *f, DisasOps *o)
4382 int r1 = get_field(f, r1);
4383 tcg_gen_deposit_i64(regs[r1], regs[r1], o->out, 0, 8);
4385 #define SPEC_wout_r1_8 0
4387 static void wout_r1_16(DisasContext *s, DisasFields *f, DisasOps *o)
4389 int r1 = get_field(f, r1);
4390 tcg_gen_deposit_i64(regs[r1], regs[r1], o->out, 0, 16);
4392 #define SPEC_wout_r1_16 0
4394 static void wout_r1_32(DisasContext *s, DisasFields *f, DisasOps *o)
4396 store_reg32_i64(get_field(f, r1), o->out);
4398 #define SPEC_wout_r1_32 0
4400 static void wout_r1_32h(DisasContext *s, DisasFields *f, DisasOps *o)
4402 store_reg32h_i64(get_field(f, r1), o->out);
4404 #define SPEC_wout_r1_32h 0
4406 static void wout_r1_P32(DisasContext *s, DisasFields *f, DisasOps *o)
4408 int r1 = get_field(f, r1);
4409 store_reg32_i64(r1, o->out);
4410 store_reg32_i64(r1 + 1, o->out2);
4412 #define SPEC_wout_r1_P32 SPEC_r1_even
4414 static void wout_r1_D32(DisasContext *s, DisasFields *f, DisasOps *o)
4416 int r1 = get_field(f, r1);
4417 store_reg32_i64(r1 + 1, o->out);
4418 tcg_gen_shri_i64(o->out, o->out, 32);
4419 store_reg32_i64(r1, o->out);
4421 #define SPEC_wout_r1_D32 SPEC_r1_even
4423 static void wout_e1(DisasContext *s, DisasFields *f, DisasOps *o)
4425 store_freg32_i64(get_field(f, r1), o->out);
4427 #define SPEC_wout_e1 0
4429 static void wout_f1(DisasContext *s, DisasFields *f, DisasOps *o)
4431 store_freg(get_field(f, r1), o->out);
4433 #define SPEC_wout_f1 0
4435 static void wout_x1(DisasContext *s, DisasFields *f, DisasOps *o)
4437 int f1 = get_field(s->fields, r1);
4438 store_freg(f1, o->out);
4439 store_freg(f1 + 2, o->out2);
4441 #define SPEC_wout_x1 SPEC_r1_f128
4443 static void wout_cond_r1r2_32(DisasContext *s, DisasFields *f, DisasOps *o)
4445 if (get_field(f, r1) != get_field(f, r2)) {
4446 store_reg32_i64(get_field(f, r1), o->out);
4449 #define SPEC_wout_cond_r1r2_32 0
4451 static void wout_cond_e1e2(DisasContext *s, DisasFields *f, DisasOps *o)
4453 if (get_field(f, r1) != get_field(f, r2)) {
4454 store_freg32_i64(get_field(f, r1), o->out);
4457 #define SPEC_wout_cond_e1e2 0
4459 static void wout_m1_8(DisasContext *s, DisasFields *f, DisasOps *o)
4461 tcg_gen_qemu_st8(o->out, o->addr1, get_mem_index(s));
4463 #define SPEC_wout_m1_8 0
4465 static void wout_m1_16(DisasContext *s, DisasFields *f, DisasOps *o)
4467 tcg_gen_qemu_st16(o->out, o->addr1, get_mem_index(s));
4469 #define SPEC_wout_m1_16 0
4471 static void wout_m1_32(DisasContext *s, DisasFields *f, DisasOps *o)
4473 tcg_gen_qemu_st32(o->out, o->addr1, get_mem_index(s));
4475 #define SPEC_wout_m1_32 0
4477 static void wout_m1_64(DisasContext *s, DisasFields *f, DisasOps *o)
4479 tcg_gen_qemu_st64(o->out, o->addr1, get_mem_index(s));
4481 #define SPEC_wout_m1_64 0
4483 static void wout_m2_32(DisasContext *s, DisasFields *f, DisasOps *o)
4485 tcg_gen_qemu_st32(o->out, o->in2, get_mem_index(s));
4487 #define SPEC_wout_m2_32 0
4489 static void wout_m2_32_r1_atomic(DisasContext *s, DisasFields *f, DisasOps *o)
4491 /* XXX release reservation */
4492 tcg_gen_qemu_st32(o->out, o->addr1, get_mem_index(s));
4493 store_reg32_i64(get_field(f, r1), o->in2);
4495 #define SPEC_wout_m2_32_r1_atomic 0
4497 static void wout_m2_64_r1_atomic(DisasContext *s, DisasFields *f, DisasOps *o)
4499 /* XXX release reservation */
4500 tcg_gen_qemu_st64(o->out, o->addr1, get_mem_index(s));
4501 store_reg(get_field(f, r1), o->in2);
4503 #define SPEC_wout_m2_64_r1_atomic 0
4505 /* ====================================================================== */
4506 /* The "INput 1" generators. These load the first operand to an insn. */
4508 static void in1_r1(DisasContext *s, DisasFields *f, DisasOps *o)
4510 o->in1 = load_reg(get_field(f, r1));
4512 #define SPEC_in1_r1 0
4514 static void in1_r1_o(DisasContext *s, DisasFields *f, DisasOps *o)
4516 o->in1 = regs[get_field(f, r1)];
4517 o->g_in1 = true;
4519 #define SPEC_in1_r1_o 0
4521 static void in1_r1_32s(DisasContext *s, DisasFields *f, DisasOps *o)
4523 o->in1 = tcg_temp_new_i64();
4524 tcg_gen_ext32s_i64(o->in1, regs[get_field(f, r1)]);
4526 #define SPEC_in1_r1_32s 0
4528 static void in1_r1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
4530 o->in1 = tcg_temp_new_i64();
4531 tcg_gen_ext32u_i64(o->in1, regs[get_field(f, r1)]);
4533 #define SPEC_in1_r1_32u 0
4535 static void in1_r1_sr32(DisasContext *s, DisasFields *f, DisasOps *o)
4537 o->in1 = tcg_temp_new_i64();
4538 tcg_gen_shri_i64(o->in1, regs[get_field(f, r1)], 32);
4540 #define SPEC_in1_r1_sr32 0
4542 static void in1_r1p1(DisasContext *s, DisasFields *f, DisasOps *o)
4544 o->in1 = load_reg(get_field(f, r1) + 1);
4546 #define SPEC_in1_r1p1 SPEC_r1_even
4548 static void in1_r1p1_32s(DisasContext *s, DisasFields *f, DisasOps *o)
4550 o->in1 = tcg_temp_new_i64();
4551 tcg_gen_ext32s_i64(o->in1, regs[get_field(f, r1) + 1]);
4553 #define SPEC_in1_r1p1_32s SPEC_r1_even
4555 static void in1_r1p1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
4557 o->in1 = tcg_temp_new_i64();
4558 tcg_gen_ext32u_i64(o->in1, regs[get_field(f, r1) + 1]);
4560 #define SPEC_in1_r1p1_32u SPEC_r1_even
4562 static void in1_r1_D32(DisasContext *s, DisasFields *f, DisasOps *o)
4564 int r1 = get_field(f, r1);
4565 o->in1 = tcg_temp_new_i64();
4566 tcg_gen_concat32_i64(o->in1, regs[r1 + 1], regs[r1]);
4568 #define SPEC_in1_r1_D32 SPEC_r1_even
4570 static void in1_r2(DisasContext *s, DisasFields *f, DisasOps *o)
4572 o->in1 = load_reg(get_field(f, r2));
4574 #define SPEC_in1_r2 0
4576 static void in1_r2_sr32(DisasContext *s, DisasFields *f, DisasOps *o)
4578 o->in1 = tcg_temp_new_i64();
4579 tcg_gen_shri_i64(o->in1, regs[get_field(f, r2)], 32);
4581 #define SPEC_in1_r2_sr32 0
4583 static void in1_r3(DisasContext *s, DisasFields *f, DisasOps *o)
4585 o->in1 = load_reg(get_field(f, r3));
4587 #define SPEC_in1_r3 0
4589 static void in1_r3_o(DisasContext *s, DisasFields *f, DisasOps *o)
4591 o->in1 = regs[get_field(f, r3)];
4592 o->g_in1 = true;
4594 #define SPEC_in1_r3_o 0
4596 static void in1_r3_32s(DisasContext *s, DisasFields *f, DisasOps *o)
4598 o->in1 = tcg_temp_new_i64();
4599 tcg_gen_ext32s_i64(o->in1, regs[get_field(f, r3)]);
4601 #define SPEC_in1_r3_32s 0
4603 static void in1_r3_32u(DisasContext *s, DisasFields *f, DisasOps *o)
4605 o->in1 = tcg_temp_new_i64();
4606 tcg_gen_ext32u_i64(o->in1, regs[get_field(f, r3)]);
4608 #define SPEC_in1_r3_32u 0
4610 static void in1_r3_D32(DisasContext *s, DisasFields *f, DisasOps *o)
4612 int r3 = get_field(f, r3);
4613 o->in1 = tcg_temp_new_i64();
4614 tcg_gen_concat32_i64(o->in1, regs[r3 + 1], regs[r3]);
4616 #define SPEC_in1_r3_D32 SPEC_r3_even
4618 static void in1_e1(DisasContext *s, DisasFields *f, DisasOps *o)
4620 o->in1 = load_freg32_i64(get_field(f, r1));
4622 #define SPEC_in1_e1 0
4624 static void in1_f1_o(DisasContext *s, DisasFields *f, DisasOps *o)
4626 o->in1 = fregs[get_field(f, r1)];
4627 o->g_in1 = true;
4629 #define SPEC_in1_f1_o 0
4631 static void in1_x1_o(DisasContext *s, DisasFields *f, DisasOps *o)
4633 int r1 = get_field(f, r1);
4634 o->out = fregs[r1];
4635 o->out2 = fregs[r1 + 2];
4636 o->g_out = o->g_out2 = true;
4638 #define SPEC_in1_x1_o SPEC_r1_f128
4640 static void in1_f3_o(DisasContext *s, DisasFields *f, DisasOps *o)
4642 o->in1 = fregs[get_field(f, r3)];
4643 o->g_in1 = true;
4645 #define SPEC_in1_f3_o 0
4647 static void in1_la1(DisasContext *s, DisasFields *f, DisasOps *o)
4649 o->addr1 = get_address(s, 0, get_field(f, b1), get_field(f, d1));
4651 #define SPEC_in1_la1 0
4653 static void in1_la2(DisasContext *s, DisasFields *f, DisasOps *o)
4655 int x2 = have_field(f, x2) ? get_field(f, x2) : 0;
4656 o->addr1 = get_address(s, x2, get_field(f, b2), get_field(f, d2));
4658 #define SPEC_in1_la2 0
4660 static void in1_m1_8u(DisasContext *s, DisasFields *f, DisasOps *o)
4662 in1_la1(s, f, o);
4663 o->in1 = tcg_temp_new_i64();
4664 tcg_gen_qemu_ld8u(o->in1, o->addr1, get_mem_index(s));
4666 #define SPEC_in1_m1_8u 0
4668 static void in1_m1_16s(DisasContext *s, DisasFields *f, DisasOps *o)
4670 in1_la1(s, f, o);
4671 o->in1 = tcg_temp_new_i64();
4672 tcg_gen_qemu_ld16s(o->in1, o->addr1, get_mem_index(s));
4674 #define SPEC_in1_m1_16s 0
4676 static void in1_m1_16u(DisasContext *s, DisasFields *f, DisasOps *o)
4678 in1_la1(s, f, o);
4679 o->in1 = tcg_temp_new_i64();
4680 tcg_gen_qemu_ld16u(o->in1, o->addr1, get_mem_index(s));
4682 #define SPEC_in1_m1_16u 0
4684 static void in1_m1_32s(DisasContext *s, DisasFields *f, DisasOps *o)
4686 in1_la1(s, f, o);
4687 o->in1 = tcg_temp_new_i64();
4688 tcg_gen_qemu_ld32s(o->in1, o->addr1, get_mem_index(s));
4690 #define SPEC_in1_m1_32s 0
4692 static void in1_m1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
4694 in1_la1(s, f, o);
4695 o->in1 = tcg_temp_new_i64();
4696 tcg_gen_qemu_ld32u(o->in1, o->addr1, get_mem_index(s));
4698 #define SPEC_in1_m1_32u 0
4700 static void in1_m1_64(DisasContext *s, DisasFields *f, DisasOps *o)
4702 in1_la1(s, f, o);
4703 o->in1 = tcg_temp_new_i64();
4704 tcg_gen_qemu_ld64(o->in1, o->addr1, get_mem_index(s));
4706 #define SPEC_in1_m1_64 0
4708 /* ====================================================================== */
4709 /* The "INput 2" generators. These load the second operand to an insn. */
4711 static void in2_r1_o(DisasContext *s, DisasFields *f, DisasOps *o)
4713 o->in2 = regs[get_field(f, r1)];
4714 o->g_in2 = true;
4716 #define SPEC_in2_r1_o 0
4718 static void in2_r1_16u(DisasContext *s, DisasFields *f, DisasOps *o)
4720 o->in2 = tcg_temp_new_i64();
4721 tcg_gen_ext16u_i64(o->in2, regs[get_field(f, r1)]);
4723 #define SPEC_in2_r1_16u 0
4725 static void in2_r1_32u(DisasContext *s, DisasFields *f, DisasOps *o)
4727 o->in2 = tcg_temp_new_i64();
4728 tcg_gen_ext32u_i64(o->in2, regs[get_field(f, r1)]);
4730 #define SPEC_in2_r1_32u 0
4732 static void in2_r1_D32(DisasContext *s, DisasFields *f, DisasOps *o)
4734 int r1 = get_field(f, r1);
4735 o->in2 = tcg_temp_new_i64();
4736 tcg_gen_concat32_i64(o->in2, regs[r1 + 1], regs[r1]);
4738 #define SPEC_in2_r1_D32 SPEC_r1_even
4740 static void in2_r2(DisasContext *s, DisasFields *f, DisasOps *o)
4742 o->in2 = load_reg(get_field(f, r2));
4744 #define SPEC_in2_r2 0
4746 static void in2_r2_o(DisasContext *s, DisasFields *f, DisasOps *o)
4748 o->in2 = regs[get_field(f, r2)];
4749 o->g_in2 = true;
4751 #define SPEC_in2_r2_o 0
4753 static void in2_r2_nz(DisasContext *s, DisasFields *f, DisasOps *o)
4755 int r2 = get_field(f, r2);
4756 if (r2 != 0) {
4757 o->in2 = load_reg(r2);
4760 #define SPEC_in2_r2_nz 0
4762 static void in2_r2_8s(DisasContext *s, DisasFields *f, DisasOps *o)
4764 o->in2 = tcg_temp_new_i64();
4765 tcg_gen_ext8s_i64(o->in2, regs[get_field(f, r2)]);
4767 #define SPEC_in2_r2_8s 0
4769 static void in2_r2_8u(DisasContext *s, DisasFields *f, DisasOps *o)
4771 o->in2 = tcg_temp_new_i64();
4772 tcg_gen_ext8u_i64(o->in2, regs[get_field(f, r2)]);
4774 #define SPEC_in2_r2_8u 0
4776 static void in2_r2_16s(DisasContext *s, DisasFields *f, DisasOps *o)
4778 o->in2 = tcg_temp_new_i64();
4779 tcg_gen_ext16s_i64(o->in2, regs[get_field(f, r2)]);
4781 #define SPEC_in2_r2_16s 0
4783 static void in2_r2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
4785 o->in2 = tcg_temp_new_i64();
4786 tcg_gen_ext16u_i64(o->in2, regs[get_field(f, r2)]);
4788 #define SPEC_in2_r2_16u 0
4790 static void in2_r3(DisasContext *s, DisasFields *f, DisasOps *o)
4792 o->in2 = load_reg(get_field(f, r3));
4794 #define SPEC_in2_r3 0
4796 static void in2_r3_sr32(DisasContext *s, DisasFields *f, DisasOps *o)
4798 o->in2 = tcg_temp_new_i64();
4799 tcg_gen_shri_i64(o->in2, regs[get_field(f, r3)], 32);
4801 #define SPEC_in2_r3_sr32 0
4803 static void in2_r2_32s(DisasContext *s, DisasFields *f, DisasOps *o)
4805 o->in2 = tcg_temp_new_i64();
4806 tcg_gen_ext32s_i64(o->in2, regs[get_field(f, r2)]);
4808 #define SPEC_in2_r2_32s 0
4810 static void in2_r2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
4812 o->in2 = tcg_temp_new_i64();
4813 tcg_gen_ext32u_i64(o->in2, regs[get_field(f, r2)]);
4815 #define SPEC_in2_r2_32u 0
4817 static void in2_r2_sr32(DisasContext *s, DisasFields *f, DisasOps *o)
4819 o->in2 = tcg_temp_new_i64();
4820 tcg_gen_shri_i64(o->in2, regs[get_field(f, r2)], 32);
4822 #define SPEC_in2_r2_sr32 0
4824 static void in2_e2(DisasContext *s, DisasFields *f, DisasOps *o)
4826 o->in2 = load_freg32_i64(get_field(f, r2));
4828 #define SPEC_in2_e2 0
4830 static void in2_f2_o(DisasContext *s, DisasFields *f, DisasOps *o)
4832 o->in2 = fregs[get_field(f, r2)];
4833 o->g_in2 = true;
4835 #define SPEC_in2_f2_o 0
4837 static void in2_x2_o(DisasContext *s, DisasFields *f, DisasOps *o)
4839 int r2 = get_field(f, r2);
4840 o->in1 = fregs[r2];
4841 o->in2 = fregs[r2 + 2];
4842 o->g_in1 = o->g_in2 = true;
4844 #define SPEC_in2_x2_o SPEC_r2_f128
4846 static void in2_ra2(DisasContext *s, DisasFields *f, DisasOps *o)
4848 o->in2 = get_address(s, 0, get_field(f, r2), 0);
4850 #define SPEC_in2_ra2 0
4852 static void in2_a2(DisasContext *s, DisasFields *f, DisasOps *o)
4854 int x2 = have_field(f, x2) ? get_field(f, x2) : 0;
4855 o->in2 = get_address(s, x2, get_field(f, b2), get_field(f, d2));
4857 #define SPEC_in2_a2 0
4859 static void in2_ri2(DisasContext *s, DisasFields *f, DisasOps *o)
4861 o->in2 = tcg_const_i64(s->pc + (int64_t)get_field(f, i2) * 2);
4863 #define SPEC_in2_ri2 0
4865 static void in2_sh32(DisasContext *s, DisasFields *f, DisasOps *o)
4867 help_l2_shift(s, f, o, 31);
4869 #define SPEC_in2_sh32 0
4871 static void in2_sh64(DisasContext *s, DisasFields *f, DisasOps *o)
4873 help_l2_shift(s, f, o, 63);
4875 #define SPEC_in2_sh64 0
4877 static void in2_m2_8u(DisasContext *s, DisasFields *f, DisasOps *o)
4879 in2_a2(s, f, o);
4880 tcg_gen_qemu_ld8u(o->in2, o->in2, get_mem_index(s));
4882 #define SPEC_in2_m2_8u 0
4884 static void in2_m2_16s(DisasContext *s, DisasFields *f, DisasOps *o)
4886 in2_a2(s, f, o);
4887 tcg_gen_qemu_ld16s(o->in2, o->in2, get_mem_index(s));
4889 #define SPEC_in2_m2_16s 0
4891 static void in2_m2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
4893 in2_a2(s, f, o);
4894 tcg_gen_qemu_ld16u(o->in2, o->in2, get_mem_index(s));
4896 #define SPEC_in2_m2_16u 0
4898 static void in2_m2_32s(DisasContext *s, DisasFields *f, DisasOps *o)
4900 in2_a2(s, f, o);
4901 tcg_gen_qemu_ld32s(o->in2, o->in2, get_mem_index(s));
4903 #define SPEC_in2_m2_32s 0
4905 static void in2_m2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
4907 in2_a2(s, f, o);
4908 tcg_gen_qemu_ld32u(o->in2, o->in2, get_mem_index(s));
4910 #define SPEC_in2_m2_32u 0
4912 static void in2_m2_64(DisasContext *s, DisasFields *f, DisasOps *o)
4914 in2_a2(s, f, o);
4915 tcg_gen_qemu_ld64(o->in2, o->in2, get_mem_index(s));
4917 #define SPEC_in2_m2_64 0
4919 static void in2_mri2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
4921 in2_ri2(s, f, o);
4922 tcg_gen_qemu_ld16u(o->in2, o->in2, get_mem_index(s));
4924 #define SPEC_in2_mri2_16u 0
4926 static void in2_mri2_32s(DisasContext *s, DisasFields *f, DisasOps *o)
4928 in2_ri2(s, f, o);
4929 tcg_gen_qemu_ld32s(o->in2, o->in2, get_mem_index(s));
4931 #define SPEC_in2_mri2_32s 0
4933 static void in2_mri2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
4935 in2_ri2(s, f, o);
4936 tcg_gen_qemu_ld32u(o->in2, o->in2, get_mem_index(s));
4938 #define SPEC_in2_mri2_32u 0
4940 static void in2_mri2_64(DisasContext *s, DisasFields *f, DisasOps *o)
4942 in2_ri2(s, f, o);
4943 tcg_gen_qemu_ld64(o->in2, o->in2, get_mem_index(s));
4945 #define SPEC_in2_mri2_64 0
4947 static void in2_m2_32s_atomic(DisasContext *s, DisasFields *f, DisasOps *o)
4949 /* XXX should reserve the address */
4950 in1_la2(s, f, o);
4951 o->in2 = tcg_temp_new_i64();
4952 tcg_gen_qemu_ld32s(o->in2, o->addr1, get_mem_index(s));
4954 #define SPEC_in2_m2_32s_atomic 0
4956 static void in2_m2_64_atomic(DisasContext *s, DisasFields *f, DisasOps *o)
4958 /* XXX should reserve the address */
4959 in1_la2(s, f, o);
4960 o->in2 = tcg_temp_new_i64();
4961 tcg_gen_qemu_ld64(o->in2, o->addr1, get_mem_index(s));
4963 #define SPEC_in2_m2_64_atomic 0
4965 static void in2_i2(DisasContext *s, DisasFields *f, DisasOps *o)
4967 o->in2 = tcg_const_i64(get_field(f, i2));
4969 #define SPEC_in2_i2 0
4971 static void in2_i2_8u(DisasContext *s, DisasFields *f, DisasOps *o)
4973 o->in2 = tcg_const_i64((uint8_t)get_field(f, i2));
4975 #define SPEC_in2_i2_8u 0
4977 static void in2_i2_16u(DisasContext *s, DisasFields *f, DisasOps *o)
4979 o->in2 = tcg_const_i64((uint16_t)get_field(f, i2));
4981 #define SPEC_in2_i2_16u 0
4983 static void in2_i2_32u(DisasContext *s, DisasFields *f, DisasOps *o)
4985 o->in2 = tcg_const_i64((uint32_t)get_field(f, i2));
4987 #define SPEC_in2_i2_32u 0
4989 static void in2_i2_16u_shl(DisasContext *s, DisasFields *f, DisasOps *o)
4991 uint64_t i2 = (uint16_t)get_field(f, i2);
4992 o->in2 = tcg_const_i64(i2 << s->insn->data);
4994 #define SPEC_in2_i2_16u_shl 0
4996 static void in2_i2_32u_shl(DisasContext *s, DisasFields *f, DisasOps *o)
4998 uint64_t i2 = (uint32_t)get_field(f, i2);
4999 o->in2 = tcg_const_i64(i2 << s->insn->data);
5001 #define SPEC_in2_i2_32u_shl 0
5003 #ifndef CONFIG_USER_ONLY
5004 static void in2_insn(DisasContext *s, DisasFields *f, DisasOps *o)
5006 o->in2 = tcg_const_i64(s->fields->raw_insn);
5008 #define SPEC_in2_insn 0
5009 #endif
5011 /* ====================================================================== */
5013 /* Find opc within the table of insns. This is formulated as a switch
5014 statement so that (1) we get compile-time notice of cut-paste errors
5015 for duplicated opcodes, and (2) the compiler generates the binary
5016 search tree, rather than us having to post-process the table. */
5018 #define C(OPC, NM, FT, FC, I1, I2, P, W, OP, CC) \
5019 D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, 0)
5021 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) insn_ ## NM,
5023 enum DisasInsnEnum {
5024 #include "insn-data.def"
5027 #undef D
5028 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) { \
5029 .opc = OPC, \
5030 .fmt = FMT_##FT, \
5031 .fac = FAC_##FC, \
5032 .spec = SPEC_in1_##I1 | SPEC_in2_##I2 | SPEC_prep_##P | SPEC_wout_##W, \
5033 .name = #NM, \
5034 .help_in1 = in1_##I1, \
5035 .help_in2 = in2_##I2, \
5036 .help_prep = prep_##P, \
5037 .help_wout = wout_##W, \
5038 .help_cout = cout_##CC, \
5039 .help_op = op_##OP, \
5040 .data = D \
5043 /* Allow 0 to be used for NULL in the table below. */
5044 #define in1_0 NULL
5045 #define in2_0 NULL
5046 #define prep_0 NULL
5047 #define wout_0 NULL
5048 #define cout_0 NULL
5049 #define op_0 NULL
5051 #define SPEC_in1_0 0
5052 #define SPEC_in2_0 0
5053 #define SPEC_prep_0 0
5054 #define SPEC_wout_0 0
5056 static const DisasInsn insn_info[] = {
5057 #include "insn-data.def"
5060 #undef D
5061 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) \
5062 case OPC: return &insn_info[insn_ ## NM];
5064 static const DisasInsn *lookup_opc(uint16_t opc)
5066 switch (opc) {
5067 #include "insn-data.def"
5068 default:
5069 return NULL;
5073 #undef D
5074 #undef C
5076 /* Extract a field from the insn. The INSN should be left-aligned in
5077 the uint64_t so that we can more easily utilize the big-bit-endian
5078 definitions we extract from the Principals of Operation. */
5080 static void extract_field(DisasFields *o, const DisasField *f, uint64_t insn)
5082 uint32_t r, m;
5084 if (f->size == 0) {
5085 return;
5088 /* Zero extract the field from the insn. */
5089 r = (insn << f->beg) >> (64 - f->size);
5091 /* Sign-extend, or un-swap the field as necessary. */
5092 switch (f->type) {
5093 case 0: /* unsigned */
5094 break;
5095 case 1: /* signed */
5096 assert(f->size <= 32);
5097 m = 1u << (f->size - 1);
5098 r = (r ^ m) - m;
5099 break;
5100 case 2: /* dl+dh split, signed 20 bit. */
5101 r = ((int8_t)r << 12) | (r >> 8);
5102 break;
5103 default:
5104 abort();
5107 /* Validate that the "compressed" encoding we selected above is valid.
5108 I.e. we havn't make two different original fields overlap. */
5109 assert(((o->presentC >> f->indexC) & 1) == 0);
5110 o->presentC |= 1 << f->indexC;
5111 o->presentO |= 1 << f->indexO;
5113 o->c[f->indexC] = r;
5116 /* Lookup the insn at the current PC, extracting the operands into O and
5117 returning the info struct for the insn. Returns NULL for invalid insn. */
5119 static const DisasInsn *extract_insn(CPUS390XState *env, DisasContext *s,
5120 DisasFields *f)
5122 uint64_t insn, pc = s->pc;
5123 int op, op2, ilen;
5124 const DisasInsn *info;
5126 insn = ld_code2(env, pc);
5127 op = (insn >> 8) & 0xff;
5128 ilen = get_ilen(op);
5129 s->next_pc = s->pc + ilen;
5131 switch (ilen) {
5132 case 2:
5133 insn = insn << 48;
5134 break;
5135 case 4:
5136 insn = ld_code4(env, pc) << 32;
5137 break;
5138 case 6:
5139 insn = (insn << 48) | (ld_code4(env, pc + 2) << 16);
5140 break;
5141 default:
5142 abort();
5145 /* We can't actually determine the insn format until we've looked up
5146 the full insn opcode. Which we can't do without locating the
5147 secondary opcode. Assume by default that OP2 is at bit 40; for
5148 those smaller insns that don't actually have a secondary opcode
5149 this will correctly result in OP2 = 0. */
5150 switch (op) {
5151 case 0x01: /* E */
5152 case 0x80: /* S */
5153 case 0x82: /* S */
5154 case 0x93: /* S */
5155 case 0xb2: /* S, RRF, RRE */
5156 case 0xb3: /* RRE, RRD, RRF */
5157 case 0xb9: /* RRE, RRF */
5158 case 0xe5: /* SSE, SIL */
5159 op2 = (insn << 8) >> 56;
5160 break;
5161 case 0xa5: /* RI */
5162 case 0xa7: /* RI */
5163 case 0xc0: /* RIL */
5164 case 0xc2: /* RIL */
5165 case 0xc4: /* RIL */
5166 case 0xc6: /* RIL */
5167 case 0xc8: /* SSF */
5168 case 0xcc: /* RIL */
5169 op2 = (insn << 12) >> 60;
5170 break;
5171 case 0xd0 ... 0xdf: /* SS */
5172 case 0xe1: /* SS */
5173 case 0xe2: /* SS */
5174 case 0xe8: /* SS */
5175 case 0xe9: /* SS */
5176 case 0xea: /* SS */
5177 case 0xee ... 0xf3: /* SS */
5178 case 0xf8 ... 0xfd: /* SS */
5179 op2 = 0;
5180 break;
5181 default:
5182 op2 = (insn << 40) >> 56;
5183 break;
5186 memset(f, 0, sizeof(*f));
5187 f->raw_insn = insn;
5188 f->op = op;
5189 f->op2 = op2;
5191 /* Lookup the instruction. */
5192 info = lookup_opc(op << 8 | op2);
5194 /* If we found it, extract the operands. */
5195 if (info != NULL) {
5196 DisasFormat fmt = info->fmt;
5197 int i;
5199 for (i = 0; i < NUM_C_FIELD; ++i) {
5200 extract_field(f, &format_info[fmt].op[i], insn);
5203 return info;
5206 static ExitStatus translate_one(CPUS390XState *env, DisasContext *s)
5208 const DisasInsn *insn;
5209 ExitStatus ret = NO_EXIT;
5210 DisasFields f;
5211 DisasOps o;
5213 /* Search for the insn in the table. */
5214 insn = extract_insn(env, s, &f);
5216 /* Not found means unimplemented/illegal opcode. */
5217 if (insn == NULL) {
5218 qemu_log_mask(LOG_UNIMP, "unimplemented opcode 0x%02x%02x\n",
5219 f.op, f.op2);
5220 gen_illegal_opcode(s);
5221 return EXIT_NORETURN;
5224 #ifndef CONFIG_USER_ONLY
5225 if (s->tb->flags & FLAG_MASK_PER) {
5226 TCGv_i64 addr = tcg_const_i64(s->pc);
5227 gen_helper_per_ifetch(cpu_env, addr);
5228 tcg_temp_free_i64(addr);
5230 #endif
5232 /* Check for insn specification exceptions. */
5233 if (insn->spec) {
5234 int spec = insn->spec, excp = 0, r;
5236 if (spec & SPEC_r1_even) {
5237 r = get_field(&f, r1);
5238 if (r & 1) {
5239 excp = PGM_SPECIFICATION;
5242 if (spec & SPEC_r2_even) {
5243 r = get_field(&f, r2);
5244 if (r & 1) {
5245 excp = PGM_SPECIFICATION;
5248 if (spec & SPEC_r3_even) {
5249 r = get_field(&f, r3);
5250 if (r & 1) {
5251 excp = PGM_SPECIFICATION;
5254 if (spec & SPEC_r1_f128) {
5255 r = get_field(&f, r1);
5256 if (r > 13) {
5257 excp = PGM_SPECIFICATION;
5260 if (spec & SPEC_r2_f128) {
5261 r = get_field(&f, r2);
5262 if (r > 13) {
5263 excp = PGM_SPECIFICATION;
5266 if (excp) {
5267 gen_program_exception(s, excp);
5268 return EXIT_NORETURN;
5272 /* Set up the strutures we use to communicate with the helpers. */
5273 s->insn = insn;
5274 s->fields = &f;
5275 o.g_out = o.g_out2 = o.g_in1 = o.g_in2 = false;
5276 TCGV_UNUSED_I64(o.out);
5277 TCGV_UNUSED_I64(o.out2);
5278 TCGV_UNUSED_I64(o.in1);
5279 TCGV_UNUSED_I64(o.in2);
5280 TCGV_UNUSED_I64(o.addr1);
5282 /* Implement the instruction. */
5283 if (insn->help_in1) {
5284 insn->help_in1(s, &f, &o);
5286 if (insn->help_in2) {
5287 insn->help_in2(s, &f, &o);
5289 if (insn->help_prep) {
5290 insn->help_prep(s, &f, &o);
5292 if (insn->help_op) {
5293 ret = insn->help_op(s, &o);
5295 if (insn->help_wout) {
5296 insn->help_wout(s, &f, &o);
5298 if (insn->help_cout) {
5299 insn->help_cout(s, &o);
5302 /* Free any temporaries created by the helpers. */
5303 if (!TCGV_IS_UNUSED_I64(o.out) && !o.g_out) {
5304 tcg_temp_free_i64(o.out);
5306 if (!TCGV_IS_UNUSED_I64(o.out2) && !o.g_out2) {
5307 tcg_temp_free_i64(o.out2);
5309 if (!TCGV_IS_UNUSED_I64(o.in1) && !o.g_in1) {
5310 tcg_temp_free_i64(o.in1);
5312 if (!TCGV_IS_UNUSED_I64(o.in2) && !o.g_in2) {
5313 tcg_temp_free_i64(o.in2);
5315 if (!TCGV_IS_UNUSED_I64(o.addr1)) {
5316 tcg_temp_free_i64(o.addr1);
5319 #ifndef CONFIG_USER_ONLY
5320 if (s->tb->flags & FLAG_MASK_PER) {
5321 /* An exception might be triggered, save PSW if not already done. */
5322 if (ret == NO_EXIT || ret == EXIT_PC_STALE) {
5323 tcg_gen_movi_i64(psw_addr, s->next_pc);
5326 /* Save off cc. */
5327 update_cc_op(s);
5329 /* Call the helper to check for a possible PER exception. */
5330 gen_helper_per_check_exception(cpu_env);
5332 #endif
5334 /* Advance to the next instruction. */
5335 s->pc = s->next_pc;
5336 return ret;
5339 void gen_intermediate_code(CPUS390XState *env, struct TranslationBlock *tb)
5341 S390CPU *cpu = s390_env_get_cpu(env);
5342 CPUState *cs = CPU(cpu);
5343 DisasContext dc;
5344 target_ulong pc_start;
5345 uint64_t next_page_start;
5346 int num_insns, max_insns;
5347 ExitStatus status;
5348 bool do_debug;
5350 pc_start = tb->pc;
5352 /* 31-bit mode */
5353 if (!(tb->flags & FLAG_MASK_64)) {
5354 pc_start &= 0x7fffffff;
5357 dc.tb = tb;
5358 dc.pc = pc_start;
5359 dc.cc_op = CC_OP_DYNAMIC;
5360 do_debug = dc.singlestep_enabled = cs->singlestep_enabled;
5362 next_page_start = (pc_start & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
5364 num_insns = 0;
5365 max_insns = tb->cflags & CF_COUNT_MASK;
5366 if (max_insns == 0) {
5367 max_insns = CF_COUNT_MASK;
5369 if (max_insns > TCG_MAX_INSNS) {
5370 max_insns = TCG_MAX_INSNS;
5373 gen_tb_start(tb);
5375 do {
5376 tcg_gen_insn_start(dc.pc, dc.cc_op);
5377 num_insns++;
5379 if (unlikely(cpu_breakpoint_test(cs, dc.pc, BP_ANY))) {
5380 status = EXIT_PC_STALE;
5381 do_debug = true;
5382 /* The address covered by the breakpoint must be included in
5383 [tb->pc, tb->pc + tb->size) in order to for it to be
5384 properly cleared -- thus we increment the PC here so that
5385 the logic setting tb->size below does the right thing. */
5386 dc.pc += 2;
5387 break;
5390 if (num_insns == max_insns && (tb->cflags & CF_LAST_IO)) {
5391 gen_io_start();
5394 status = NO_EXIT;
5395 if (status == NO_EXIT) {
5396 status = translate_one(env, &dc);
5399 /* If we reach a page boundary, are single stepping,
5400 or exhaust instruction count, stop generation. */
5401 if (status == NO_EXIT
5402 && (dc.pc >= next_page_start
5403 || tcg_op_buf_full()
5404 || num_insns >= max_insns
5405 || singlestep
5406 || cs->singlestep_enabled)) {
5407 status = EXIT_PC_STALE;
5409 } while (status == NO_EXIT);
5411 if (tb->cflags & CF_LAST_IO) {
5412 gen_io_end();
5415 switch (status) {
5416 case EXIT_GOTO_TB:
5417 case EXIT_NORETURN:
5418 break;
5419 case EXIT_PC_STALE:
5420 update_psw_addr(&dc);
5421 /* FALLTHRU */
5422 case EXIT_PC_UPDATED:
5423 /* Next TB starts off with CC_OP_DYNAMIC, so make sure the
5424 cc op type is in env */
5425 update_cc_op(&dc);
5426 /* Exit the TB, either by raising a debug exception or by return. */
5427 if (do_debug) {
5428 gen_exception(EXCP_DEBUG);
5429 } else {
5430 tcg_gen_exit_tb(0);
5432 break;
5433 default:
5434 abort();
5437 gen_tb_end(tb, num_insns);
5439 tb->size = dc.pc - pc_start;
5440 tb->icount = num_insns;
5442 #if defined(S390X_DEBUG_DISAS)
5443 if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)
5444 && qemu_log_in_addr_range(pc_start)) {
5445 qemu_log_lock();
5446 qemu_log("IN: %s\n", lookup_symbol(pc_start));
5447 log_target_disas(cs, pc_start, dc.pc - pc_start, 1);
5448 qemu_log("\n");
5449 qemu_log_unlock();
5451 #endif
5454 void restore_state_to_opc(CPUS390XState *env, TranslationBlock *tb,
5455 target_ulong *data)
5457 int cc_op = data[1];
5458 env->psw.addr = data[0];
5459 if ((cc_op != CC_OP_DYNAMIC) && (cc_op != CC_OP_STATIC)) {
5460 env->cc_op = cc_op;