4 * Copyright (c) 2009 Ulrich Hecht
5 * Copyright (c) 2010 Alexander Graf
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2 of the License, or (at your option) any later version.
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
21 /* #define DEBUG_INLINE_BRANCHES */
22 #define S390X_DEBUG_DISAS
23 /* #define S390X_DEBUG_DISAS_VERBOSE */
25 #ifdef S390X_DEBUG_DISAS_VERBOSE
26 # define LOG_DISAS(...) qemu_log(__VA_ARGS__)
28 # define LOG_DISAS(...) do { } while (0)
32 #include "disas/disas.h"
35 #include "qemu/host-utils.h"
36 #include "exec/cpu_ldst.h"
38 /* global register indexes */
39 static TCGv_ptr cpu_env
;
41 #include "exec/gen-icount.h"
42 #include "exec/helper-proto.h"
43 #include "exec/helper-gen.h"
45 #include "trace-tcg.h"
48 /* Information that (most) every instruction needs to manipulate. */
49 typedef struct DisasContext DisasContext
;
50 typedef struct DisasInsn DisasInsn
;
51 typedef struct DisasFields DisasFields
;
54 struct TranslationBlock
*tb
;
55 const DisasInsn
*insn
;
59 bool singlestep_enabled
;
62 /* Information carried about a condition to be evaluated. */
69 struct { TCGv_i64 a
, b
; } s64
;
70 struct { TCGv_i32 a
, b
; } s32
;
76 #ifdef DEBUG_INLINE_BRANCHES
77 static uint64_t inline_branch_hit
[CC_OP_MAX
];
78 static uint64_t inline_branch_miss
[CC_OP_MAX
];
81 static uint64_t pc_to_link_info(DisasContext
*s
, uint64_t pc
)
83 if (!(s
->tb
->flags
& FLAG_MASK_64
)) {
84 if (s
->tb
->flags
& FLAG_MASK_32
) {
85 return pc
| 0x80000000;
91 void s390_cpu_dump_state(CPUState
*cs
, FILE *f
, fprintf_function cpu_fprintf
,
94 S390CPU
*cpu
= S390_CPU(cs
);
95 CPUS390XState
*env
= &cpu
->env
;
99 cpu_fprintf(f
, "PSW=mask %016" PRIx64
" addr %016" PRIx64
" cc %15s\n",
100 env
->psw
.mask
, env
->psw
.addr
, cc_name(env
->cc_op
));
102 cpu_fprintf(f
, "PSW=mask %016" PRIx64
" addr %016" PRIx64
" cc %02x\n",
103 env
->psw
.mask
, env
->psw
.addr
, env
->cc_op
);
106 for (i
= 0; i
< 16; i
++) {
107 cpu_fprintf(f
, "R%02d=%016" PRIx64
, i
, env
->regs
[i
]);
109 cpu_fprintf(f
, "\n");
115 for (i
= 0; i
< 16; i
++) {
116 cpu_fprintf(f
, "F%02d=%016" PRIx64
, i
, get_freg(env
, i
)->ll
);
118 cpu_fprintf(f
, "\n");
124 for (i
= 0; i
< 32; i
++) {
125 cpu_fprintf(f
, "V%02d=%016" PRIx64
"%016" PRIx64
, i
,
126 env
->vregs
[i
][0].ll
, env
->vregs
[i
][1].ll
);
127 cpu_fprintf(f
, (i
% 2) ? " " : "\n");
130 #ifndef CONFIG_USER_ONLY
131 for (i
= 0; i
< 16; i
++) {
132 cpu_fprintf(f
, "C%02d=%016" PRIx64
, i
, env
->cregs
[i
]);
134 cpu_fprintf(f
, "\n");
141 #ifdef DEBUG_INLINE_BRANCHES
142 for (i
= 0; i
< CC_OP_MAX
; i
++) {
143 cpu_fprintf(f
, " %15s = %10ld\t%10ld\n", cc_name(i
),
144 inline_branch_miss
[i
], inline_branch_hit
[i
]);
148 cpu_fprintf(f
, "\n");
151 static TCGv_i64 psw_addr
;
152 static TCGv_i64 psw_mask
;
154 static TCGv_i32 cc_op
;
155 static TCGv_i64 cc_src
;
156 static TCGv_i64 cc_dst
;
157 static TCGv_i64 cc_vr
;
159 static char cpu_reg_names
[32][4];
160 static TCGv_i64 regs
[16];
161 static TCGv_i64 fregs
[16];
163 static uint8_t gen_opc_cc_op
[OPC_BUF_SIZE
];
165 void s390x_translate_init(void)
169 cpu_env
= tcg_global_reg_new_ptr(TCG_AREG0
, "env");
170 psw_addr
= tcg_global_mem_new_i64(TCG_AREG0
,
171 offsetof(CPUS390XState
, psw
.addr
),
173 psw_mask
= tcg_global_mem_new_i64(TCG_AREG0
,
174 offsetof(CPUS390XState
, psw
.mask
),
177 cc_op
= tcg_global_mem_new_i32(TCG_AREG0
, offsetof(CPUS390XState
, cc_op
),
179 cc_src
= tcg_global_mem_new_i64(TCG_AREG0
, offsetof(CPUS390XState
, cc_src
),
181 cc_dst
= tcg_global_mem_new_i64(TCG_AREG0
, offsetof(CPUS390XState
, cc_dst
),
183 cc_vr
= tcg_global_mem_new_i64(TCG_AREG0
, offsetof(CPUS390XState
, cc_vr
),
186 for (i
= 0; i
< 16; i
++) {
187 snprintf(cpu_reg_names
[i
], sizeof(cpu_reg_names
[0]), "r%d", i
);
188 regs
[i
] = tcg_global_mem_new(TCG_AREG0
,
189 offsetof(CPUS390XState
, regs
[i
]),
193 for (i
= 0; i
< 16; i
++) {
194 snprintf(cpu_reg_names
[i
+ 16], sizeof(cpu_reg_names
[0]), "f%d", i
);
195 fregs
[i
] = tcg_global_mem_new(TCG_AREG0
,
196 offsetof(CPUS390XState
, vregs
[i
][0].d
),
197 cpu_reg_names
[i
+ 16]);
201 static TCGv_i64
load_reg(int reg
)
203 TCGv_i64 r
= tcg_temp_new_i64();
204 tcg_gen_mov_i64(r
, regs
[reg
]);
208 static TCGv_i64
load_freg32_i64(int reg
)
210 TCGv_i64 r
= tcg_temp_new_i64();
211 tcg_gen_shri_i64(r
, fregs
[reg
], 32);
215 static void store_reg(int reg
, TCGv_i64 v
)
217 tcg_gen_mov_i64(regs
[reg
], v
);
220 static void store_freg(int reg
, TCGv_i64 v
)
222 tcg_gen_mov_i64(fregs
[reg
], v
);
225 static void store_reg32_i64(int reg
, TCGv_i64 v
)
227 /* 32 bit register writes keep the upper half */
228 tcg_gen_deposit_i64(regs
[reg
], regs
[reg
], v
, 0, 32);
231 static void store_reg32h_i64(int reg
, TCGv_i64 v
)
233 tcg_gen_deposit_i64(regs
[reg
], regs
[reg
], v
, 32, 32);
236 static void store_freg32_i64(int reg
, TCGv_i64 v
)
238 tcg_gen_deposit_i64(fregs
[reg
], fregs
[reg
], v
, 32, 32);
241 static void return_low128(TCGv_i64 dest
)
243 tcg_gen_ld_i64(dest
, cpu_env
, offsetof(CPUS390XState
, retxl
));
246 static void update_psw_addr(DisasContext
*s
)
249 tcg_gen_movi_i64(psw_addr
, s
->pc
);
252 static void update_cc_op(DisasContext
*s
)
254 if (s
->cc_op
!= CC_OP_DYNAMIC
&& s
->cc_op
!= CC_OP_STATIC
) {
255 tcg_gen_movi_i32(cc_op
, s
->cc_op
);
259 static void potential_page_fault(DisasContext
*s
)
265 static inline uint64_t ld_code2(CPUS390XState
*env
, uint64_t pc
)
267 return (uint64_t)cpu_lduw_code(env
, pc
);
270 static inline uint64_t ld_code4(CPUS390XState
*env
, uint64_t pc
)
272 return (uint64_t)(uint32_t)cpu_ldl_code(env
, pc
);
275 static int get_mem_index(DisasContext
*s
)
277 switch (s
->tb
->flags
& FLAG_MASK_ASC
) {
278 case PSW_ASC_PRIMARY
>> 32:
280 case PSW_ASC_SECONDARY
>> 32:
282 case PSW_ASC_HOME
>> 32:
290 static void gen_exception(int excp
)
292 TCGv_i32 tmp
= tcg_const_i32(excp
);
293 gen_helper_exception(cpu_env
, tmp
);
294 tcg_temp_free_i32(tmp
);
297 static void gen_program_exception(DisasContext
*s
, int code
)
301 /* Remember what pgm exeption this was. */
302 tmp
= tcg_const_i32(code
);
303 tcg_gen_st_i32(tmp
, cpu_env
, offsetof(CPUS390XState
, int_pgm_code
));
304 tcg_temp_free_i32(tmp
);
306 tmp
= tcg_const_i32(s
->next_pc
- s
->pc
);
307 tcg_gen_st_i32(tmp
, cpu_env
, offsetof(CPUS390XState
, int_pgm_ilen
));
308 tcg_temp_free_i32(tmp
);
310 /* Advance past instruction. */
317 /* Trigger exception. */
318 gen_exception(EXCP_PGM
);
321 static inline void gen_illegal_opcode(DisasContext
*s
)
323 gen_program_exception(s
, PGM_SPECIFICATION
);
326 #ifndef CONFIG_USER_ONLY
327 static void check_privileged(DisasContext
*s
)
329 if (s
->tb
->flags
& (PSW_MASK_PSTATE
>> 32)) {
330 gen_program_exception(s
, PGM_PRIVILEGED
);
335 static TCGv_i64
get_address(DisasContext
*s
, int x2
, int b2
, int d2
)
337 TCGv_i64 tmp
= tcg_temp_new_i64();
338 bool need_31
= !(s
->tb
->flags
& FLAG_MASK_64
);
340 /* Note that d2 is limited to 20 bits, signed. If we crop negative
341 displacements early we create larger immedate addends. */
343 /* Note that addi optimizes the imm==0 case. */
345 tcg_gen_add_i64(tmp
, regs
[b2
], regs
[x2
]);
346 tcg_gen_addi_i64(tmp
, tmp
, d2
);
348 tcg_gen_addi_i64(tmp
, regs
[b2
], d2
);
350 tcg_gen_addi_i64(tmp
, regs
[x2
], d2
);
356 tcg_gen_movi_i64(tmp
, d2
);
359 tcg_gen_andi_i64(tmp
, tmp
, 0x7fffffff);
365 static inline bool live_cc_data(DisasContext
*s
)
367 return (s
->cc_op
!= CC_OP_DYNAMIC
368 && s
->cc_op
!= CC_OP_STATIC
372 static inline void gen_op_movi_cc(DisasContext
*s
, uint32_t val
)
374 if (live_cc_data(s
)) {
375 tcg_gen_discard_i64(cc_src
);
376 tcg_gen_discard_i64(cc_dst
);
377 tcg_gen_discard_i64(cc_vr
);
379 s
->cc_op
= CC_OP_CONST0
+ val
;
382 static void gen_op_update1_cc_i64(DisasContext
*s
, enum cc_op op
, TCGv_i64 dst
)
384 if (live_cc_data(s
)) {
385 tcg_gen_discard_i64(cc_src
);
386 tcg_gen_discard_i64(cc_vr
);
388 tcg_gen_mov_i64(cc_dst
, dst
);
392 static void gen_op_update2_cc_i64(DisasContext
*s
, enum cc_op op
, TCGv_i64 src
,
395 if (live_cc_data(s
)) {
396 tcg_gen_discard_i64(cc_vr
);
398 tcg_gen_mov_i64(cc_src
, src
);
399 tcg_gen_mov_i64(cc_dst
, dst
);
403 static void gen_op_update3_cc_i64(DisasContext
*s
, enum cc_op op
, TCGv_i64 src
,
404 TCGv_i64 dst
, TCGv_i64 vr
)
406 tcg_gen_mov_i64(cc_src
, src
);
407 tcg_gen_mov_i64(cc_dst
, dst
);
408 tcg_gen_mov_i64(cc_vr
, vr
);
412 static void set_cc_nz_u64(DisasContext
*s
, TCGv_i64 val
)
414 gen_op_update1_cc_i64(s
, CC_OP_NZ
, val
);
417 static void gen_set_cc_nz_f32(DisasContext
*s
, TCGv_i64 val
)
419 gen_op_update1_cc_i64(s
, CC_OP_NZ_F32
, val
);
422 static void gen_set_cc_nz_f64(DisasContext
*s
, TCGv_i64 val
)
424 gen_op_update1_cc_i64(s
, CC_OP_NZ_F64
, val
);
427 static void gen_set_cc_nz_f128(DisasContext
*s
, TCGv_i64 vh
, TCGv_i64 vl
)
429 gen_op_update2_cc_i64(s
, CC_OP_NZ_F128
, vh
, vl
);
432 /* CC value is in env->cc_op */
433 static void set_cc_static(DisasContext
*s
)
435 if (live_cc_data(s
)) {
436 tcg_gen_discard_i64(cc_src
);
437 tcg_gen_discard_i64(cc_dst
);
438 tcg_gen_discard_i64(cc_vr
);
440 s
->cc_op
= CC_OP_STATIC
;
443 /* calculates cc into cc_op */
444 static void gen_op_calc_cc(DisasContext
*s
)
446 TCGv_i32 local_cc_op
;
449 TCGV_UNUSED_I32(local_cc_op
);
450 TCGV_UNUSED_I64(dummy
);
453 dummy
= tcg_const_i64(0);
467 local_cc_op
= tcg_const_i32(s
->cc_op
);
483 /* s->cc_op is the cc value */
484 tcg_gen_movi_i32(cc_op
, s
->cc_op
- CC_OP_CONST0
);
487 /* env->cc_op already is the cc value */
502 gen_helper_calc_cc(cc_op
, cpu_env
, local_cc_op
, dummy
, cc_dst
, dummy
);
507 case CC_OP_LTUGTU_32
:
508 case CC_OP_LTUGTU_64
:
515 gen_helper_calc_cc(cc_op
, cpu_env
, local_cc_op
, cc_src
, cc_dst
, dummy
);
530 gen_helper_calc_cc(cc_op
, cpu_env
, local_cc_op
, cc_src
, cc_dst
, cc_vr
);
533 /* unknown operation - assume 3 arguments and cc_op in env */
534 gen_helper_calc_cc(cc_op
, cpu_env
, cc_op
, cc_src
, cc_dst
, cc_vr
);
540 if (!TCGV_IS_UNUSED_I32(local_cc_op
)) {
541 tcg_temp_free_i32(local_cc_op
);
543 if (!TCGV_IS_UNUSED_I64(dummy
)) {
544 tcg_temp_free_i64(dummy
);
547 /* We now have cc in cc_op as constant */
551 static int use_goto_tb(DisasContext
*s
, uint64_t dest
)
553 /* NOTE: we handle the case where the TB spans two pages here */
554 return (((dest
& TARGET_PAGE_MASK
) == (s
->tb
->pc
& TARGET_PAGE_MASK
)
555 || (dest
& TARGET_PAGE_MASK
) == ((s
->pc
- 1) & TARGET_PAGE_MASK
))
556 && !s
->singlestep_enabled
557 && !(s
->tb
->cflags
& CF_LAST_IO
));
560 static void account_noninline_branch(DisasContext
*s
, int cc_op
)
562 #ifdef DEBUG_INLINE_BRANCHES
563 inline_branch_miss
[cc_op
]++;
567 static void account_inline_branch(DisasContext
*s
, int cc_op
)
569 #ifdef DEBUG_INLINE_BRANCHES
570 inline_branch_hit
[cc_op
]++;
574 /* Table of mask values to comparison codes, given a comparison as input.
575 For such, CC=3 should not be possible. */
576 static const TCGCond ltgt_cond
[16] = {
577 TCG_COND_NEVER
, TCG_COND_NEVER
, /* | | | x */
578 TCG_COND_GT
, TCG_COND_GT
, /* | | GT | x */
579 TCG_COND_LT
, TCG_COND_LT
, /* | LT | | x */
580 TCG_COND_NE
, TCG_COND_NE
, /* | LT | GT | x */
581 TCG_COND_EQ
, TCG_COND_EQ
, /* EQ | | | x */
582 TCG_COND_GE
, TCG_COND_GE
, /* EQ | | GT | x */
583 TCG_COND_LE
, TCG_COND_LE
, /* EQ | LT | | x */
584 TCG_COND_ALWAYS
, TCG_COND_ALWAYS
, /* EQ | LT | GT | x */
587 /* Table of mask values to comparison codes, given a logic op as input.
588 For such, only CC=0 and CC=1 should be possible. */
589 static const TCGCond nz_cond
[16] = {
590 TCG_COND_NEVER
, TCG_COND_NEVER
, /* | | x | x */
591 TCG_COND_NEVER
, TCG_COND_NEVER
,
592 TCG_COND_NE
, TCG_COND_NE
, /* | NE | x | x */
593 TCG_COND_NE
, TCG_COND_NE
,
594 TCG_COND_EQ
, TCG_COND_EQ
, /* EQ | | x | x */
595 TCG_COND_EQ
, TCG_COND_EQ
,
596 TCG_COND_ALWAYS
, TCG_COND_ALWAYS
, /* EQ | NE | x | x */
597 TCG_COND_ALWAYS
, TCG_COND_ALWAYS
,
600 /* Interpret MASK in terms of S->CC_OP, and fill in C with all the
601 details required to generate a TCG comparison. */
602 static void disas_jcc(DisasContext
*s
, DisasCompare
*c
, uint32_t mask
)
605 enum cc_op old_cc_op
= s
->cc_op
;
607 if (mask
== 15 || mask
== 0) {
608 c
->cond
= (mask
? TCG_COND_ALWAYS
: TCG_COND_NEVER
);
611 c
->g1
= c
->g2
= true;
616 /* Find the TCG condition for the mask + cc op. */
622 cond
= ltgt_cond
[mask
];
623 if (cond
== TCG_COND_NEVER
) {
626 account_inline_branch(s
, old_cc_op
);
629 case CC_OP_LTUGTU_32
:
630 case CC_OP_LTUGTU_64
:
631 cond
= tcg_unsigned_cond(ltgt_cond
[mask
]);
632 if (cond
== TCG_COND_NEVER
) {
635 account_inline_branch(s
, old_cc_op
);
639 cond
= nz_cond
[mask
];
640 if (cond
== TCG_COND_NEVER
) {
643 account_inline_branch(s
, old_cc_op
);
658 account_inline_branch(s
, old_cc_op
);
673 account_inline_branch(s
, old_cc_op
);
677 switch (mask
& 0xa) {
678 case 8: /* src == 0 -> no one bit found */
681 case 2: /* src != 0 -> one bit found */
687 account_inline_branch(s
, old_cc_op
);
693 case 8 | 2: /* vr == 0 */
696 case 4 | 1: /* vr != 0 */
699 case 8 | 4: /* no carry -> vr >= src */
702 case 2 | 1: /* carry -> vr < src */
708 account_inline_branch(s
, old_cc_op
);
713 /* Note that CC=0 is impossible; treat it as dont-care. */
715 case 2: /* zero -> op1 == op2 */
718 case 4 | 1: /* !zero -> op1 != op2 */
721 case 4: /* borrow (!carry) -> op1 < op2 */
724 case 2 | 1: /* !borrow (carry) -> op1 >= op2 */
730 account_inline_branch(s
, old_cc_op
);
735 /* Calculate cc value. */
740 /* Jump based on CC. We'll load up the real cond below;
741 the assignment here merely avoids a compiler warning. */
742 account_noninline_branch(s
, old_cc_op
);
743 old_cc_op
= CC_OP_STATIC
;
744 cond
= TCG_COND_NEVER
;
748 /* Load up the arguments of the comparison. */
750 c
->g1
= c
->g2
= false;
754 c
->u
.s32
.a
= tcg_temp_new_i32();
755 tcg_gen_trunc_i64_i32(c
->u
.s32
.a
, cc_dst
);
756 c
->u
.s32
.b
= tcg_const_i32(0);
759 case CC_OP_LTUGTU_32
:
762 c
->u
.s32
.a
= tcg_temp_new_i32();
763 tcg_gen_trunc_i64_i32(c
->u
.s32
.a
, cc_src
);
764 c
->u
.s32
.b
= tcg_temp_new_i32();
765 tcg_gen_trunc_i64_i32(c
->u
.s32
.b
, cc_dst
);
772 c
->u
.s64
.b
= tcg_const_i64(0);
776 case CC_OP_LTUGTU_64
:
780 c
->g1
= c
->g2
= true;
786 c
->u
.s64
.a
= tcg_temp_new_i64();
787 c
->u
.s64
.b
= tcg_const_i64(0);
788 tcg_gen_and_i64(c
->u
.s64
.a
, cc_src
, cc_dst
);
793 c
->u
.s32
.a
= tcg_temp_new_i32();
794 c
->u
.s32
.b
= tcg_temp_new_i32();
795 tcg_gen_trunc_i64_i32(c
->u
.s32
.a
, cc_vr
);
796 if (cond
== TCG_COND_EQ
|| cond
== TCG_COND_NE
) {
797 tcg_gen_movi_i32(c
->u
.s32
.b
, 0);
799 tcg_gen_trunc_i64_i32(c
->u
.s32
.b
, cc_src
);
806 if (cond
== TCG_COND_EQ
|| cond
== TCG_COND_NE
) {
807 c
->u
.s64
.b
= tcg_const_i64(0);
819 case 0x8 | 0x4 | 0x2: /* cc != 3 */
821 c
->u
.s32
.b
= tcg_const_i32(3);
823 case 0x8 | 0x4 | 0x1: /* cc != 2 */
825 c
->u
.s32
.b
= tcg_const_i32(2);
827 case 0x8 | 0x2 | 0x1: /* cc != 1 */
829 c
->u
.s32
.b
= tcg_const_i32(1);
831 case 0x8 | 0x2: /* cc == 0 ||Â cc == 2 => (cc & 1) == 0 */
834 c
->u
.s32
.a
= tcg_temp_new_i32();
835 c
->u
.s32
.b
= tcg_const_i32(0);
836 tcg_gen_andi_i32(c
->u
.s32
.a
, cc_op
, 1);
838 case 0x8 | 0x4: /* cc < 2 */
840 c
->u
.s32
.b
= tcg_const_i32(2);
842 case 0x8: /* cc == 0 */
844 c
->u
.s32
.b
= tcg_const_i32(0);
846 case 0x4 | 0x2 | 0x1: /* cc != 0 */
848 c
->u
.s32
.b
= tcg_const_i32(0);
850 case 0x4 | 0x1: /* cc == 1 ||Â cc == 3 => (cc & 1) != 0 */
853 c
->u
.s32
.a
= tcg_temp_new_i32();
854 c
->u
.s32
.b
= tcg_const_i32(0);
855 tcg_gen_andi_i32(c
->u
.s32
.a
, cc_op
, 1);
857 case 0x4: /* cc == 1 */
859 c
->u
.s32
.b
= tcg_const_i32(1);
861 case 0x2 | 0x1: /* cc > 1 */
863 c
->u
.s32
.b
= tcg_const_i32(1);
865 case 0x2: /* cc == 2 */
867 c
->u
.s32
.b
= tcg_const_i32(2);
869 case 0x1: /* cc == 3 */
871 c
->u
.s32
.b
= tcg_const_i32(3);
874 /* CC is masked by something else: (8 >> cc) & mask. */
877 c
->u
.s32
.a
= tcg_const_i32(8);
878 c
->u
.s32
.b
= tcg_const_i32(0);
879 tcg_gen_shr_i32(c
->u
.s32
.a
, c
->u
.s32
.a
, cc_op
);
880 tcg_gen_andi_i32(c
->u
.s32
.a
, c
->u
.s32
.a
, mask
);
891 static void free_compare(DisasCompare
*c
)
895 tcg_temp_free_i64(c
->u
.s64
.a
);
897 tcg_temp_free_i32(c
->u
.s32
.a
);
902 tcg_temp_free_i64(c
->u
.s64
.b
);
904 tcg_temp_free_i32(c
->u
.s32
.b
);
909 /* ====================================================================== */
910 /* Define the insn format enumeration. */
911 #define F0(N) FMT_##N,
912 #define F1(N, X1) F0(N)
913 #define F2(N, X1, X2) F0(N)
914 #define F3(N, X1, X2, X3) F0(N)
915 #define F4(N, X1, X2, X3, X4) F0(N)
916 #define F5(N, X1, X2, X3, X4, X5) F0(N)
919 #include "insn-format.def"
929 /* Define a structure to hold the decoded fields. We'll store each inside
930 an array indexed by an enum. In order to conserve memory, we'll arrange
931 for fields that do not exist at the same time to overlap, thus the "C"
932 for compact. For checking purposes there is an "O" for original index
933 as well that will be applied to availability bitmaps. */
935 enum DisasFieldIndexO
{
958 enum DisasFieldIndexC
{
992 unsigned presentC
:16;
993 unsigned int presentO
;
997 /* This is the way fields are to be accessed out of DisasFields. */
998 #define have_field(S, F) have_field1((S), FLD_O_##F)
999 #define get_field(S, F) get_field1((S), FLD_O_##F, FLD_C_##F)
1001 static bool have_field1(const DisasFields
*f
, enum DisasFieldIndexO c
)
1003 return (f
->presentO
>> c
) & 1;
1006 static int get_field1(const DisasFields
*f
, enum DisasFieldIndexO o
,
1007 enum DisasFieldIndexC c
)
1009 assert(have_field1(f
, o
));
1013 /* Describe the layout of each field in each format. */
1014 typedef struct DisasField
{
1016 unsigned int size
:8;
1017 unsigned int type
:2;
1018 unsigned int indexC
:6;
1019 enum DisasFieldIndexO indexO
:8;
1022 typedef struct DisasFormatInfo
{
1023 DisasField op
[NUM_C_FIELD
];
1026 #define R(N, B) { B, 4, 0, FLD_C_r##N, FLD_O_r##N }
1027 #define M(N, B) { B, 4, 0, FLD_C_m##N, FLD_O_m##N }
1028 #define BD(N, BB, BD) { BB, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1029 { BD, 12, 0, FLD_C_d##N, FLD_O_d##N }
1030 #define BXD(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1031 { 12, 4, 0, FLD_C_x##N, FLD_O_x##N }, \
1032 { 20, 12, 0, FLD_C_d##N, FLD_O_d##N }
1033 #define BDL(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1034 { 20, 20, 2, FLD_C_d##N, FLD_O_d##N }
1035 #define BXDL(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1036 { 12, 4, 0, FLD_C_x##N, FLD_O_x##N }, \
1037 { 20, 20, 2, FLD_C_d##N, FLD_O_d##N }
1038 #define I(N, B, S) { B, S, 1, FLD_C_i##N, FLD_O_i##N }
1039 #define L(N, B, S) { B, S, 0, FLD_C_l##N, FLD_O_l##N }
1041 #define F0(N) { { } },
1042 #define F1(N, X1) { { X1 } },
1043 #define F2(N, X1, X2) { { X1, X2 } },
1044 #define F3(N, X1, X2, X3) { { X1, X2, X3 } },
1045 #define F4(N, X1, X2, X3, X4) { { X1, X2, X3, X4 } },
1046 #define F5(N, X1, X2, X3, X4, X5) { { X1, X2, X3, X4, X5 } },
1048 static const DisasFormatInfo format_info
[] = {
1049 #include "insn-format.def"
1067 /* Generally, we'll extract operands into this structures, operate upon
1068 them, and store them back. See the "in1", "in2", "prep", "wout" sets
1069 of routines below for more details. */
1071 bool g_out
, g_out2
, g_in1
, g_in2
;
1072 TCGv_i64 out
, out2
, in1
, in2
;
1076 /* Instructions can place constraints on their operands, raising specification
1077 exceptions if they are violated. To make this easy to automate, each "in1",
1078 "in2", "prep", "wout" helper will have a SPEC_<name> define that equals one
1079 of the following, or 0. To make this easy to document, we'll put the
1080 SPEC_<name> defines next to <name>. */
1082 #define SPEC_r1_even 1
1083 #define SPEC_r2_even 2
1084 #define SPEC_r3_even 4
1085 #define SPEC_r1_f128 8
1086 #define SPEC_r2_f128 16
1088 /* Return values from translate_one, indicating the state of the TB. */
1090 /* Continue the TB. */
1092 /* We have emitted one or more goto_tb. No fixup required. */
1094 /* We are not using a goto_tb (for whatever reason), but have updated
1095 the PC (for whatever reason), so there's no need to do it again on
1098 /* We are exiting the TB, but have neither emitted a goto_tb, nor
1099 updated the PC for the next instruction to be executed. */
1101 /* We are ending the TB with a noreturn function call, e.g. longjmp.
1102 No following code will be executed. */
1106 typedef enum DisasFacility
{
1107 FAC_Z
, /* zarch (default) */
1108 FAC_CASS
, /* compare and swap and store */
1109 FAC_CASS2
, /* compare and swap and store 2*/
1110 FAC_DFP
, /* decimal floating point */
1111 FAC_DFPR
, /* decimal floating point rounding */
1112 FAC_DO
, /* distinct operands */
1113 FAC_EE
, /* execute extensions */
1114 FAC_EI
, /* extended immediate */
1115 FAC_FPE
, /* floating point extension */
1116 FAC_FPSSH
, /* floating point support sign handling */
1117 FAC_FPRGR
, /* FPR-GR transfer */
1118 FAC_GIE
, /* general instructions extension */
1119 FAC_HFP_MA
, /* HFP multiply-and-add/subtract */
1120 FAC_HW
, /* high-word */
1121 FAC_IEEEE_SIM
, /* IEEE exception sumilation */
1122 FAC_LOC
, /* load/store on condition */
1123 FAC_LD
, /* long displacement */
1124 FAC_PC
, /* population count */
1125 FAC_SCF
, /* store clock fast */
1126 FAC_SFLE
, /* store facility list extended */
1127 FAC_ILA
, /* interlocked access facility 1 */
1133 DisasFacility fac
:8;
1138 void (*help_in1
)(DisasContext
*, DisasFields
*, DisasOps
*);
1139 void (*help_in2
)(DisasContext
*, DisasFields
*, DisasOps
*);
1140 void (*help_prep
)(DisasContext
*, DisasFields
*, DisasOps
*);
1141 void (*help_wout
)(DisasContext
*, DisasFields
*, DisasOps
*);
1142 void (*help_cout
)(DisasContext
*, DisasOps
*);
1143 ExitStatus (*help_op
)(DisasContext
*, DisasOps
*);
1148 /* ====================================================================== */
1149 /* Miscellaneous helpers, used by several operations. */
1151 static void help_l2_shift(DisasContext
*s
, DisasFields
*f
,
1152 DisasOps
*o
, int mask
)
1154 int b2
= get_field(f
, b2
);
1155 int d2
= get_field(f
, d2
);
1158 o
->in2
= tcg_const_i64(d2
& mask
);
1160 o
->in2
= get_address(s
, 0, b2
, d2
);
1161 tcg_gen_andi_i64(o
->in2
, o
->in2
, mask
);
1165 static ExitStatus
help_goto_direct(DisasContext
*s
, uint64_t dest
)
1167 if (dest
== s
->next_pc
) {
1170 if (use_goto_tb(s
, dest
)) {
1173 tcg_gen_movi_i64(psw_addr
, dest
);
1174 tcg_gen_exit_tb((uintptr_t)s
->tb
);
1175 return EXIT_GOTO_TB
;
1177 tcg_gen_movi_i64(psw_addr
, dest
);
1178 return EXIT_PC_UPDATED
;
1182 static ExitStatus
help_branch(DisasContext
*s
, DisasCompare
*c
,
1183 bool is_imm
, int imm
, TCGv_i64 cdest
)
1186 uint64_t dest
= s
->pc
+ 2 * imm
;
1189 /* Take care of the special cases first. */
1190 if (c
->cond
== TCG_COND_NEVER
) {
1195 if (dest
== s
->next_pc
) {
1196 /* Branch to next. */
1200 if (c
->cond
== TCG_COND_ALWAYS
) {
1201 ret
= help_goto_direct(s
, dest
);
1205 if (TCGV_IS_UNUSED_I64(cdest
)) {
1206 /* E.g. bcr %r0 -> no branch. */
1210 if (c
->cond
== TCG_COND_ALWAYS
) {
1211 tcg_gen_mov_i64(psw_addr
, cdest
);
1212 ret
= EXIT_PC_UPDATED
;
1217 if (use_goto_tb(s
, s
->next_pc
)) {
1218 if (is_imm
&& use_goto_tb(s
, dest
)) {
1219 /* Both exits can use goto_tb. */
1222 lab
= gen_new_label();
1224 tcg_gen_brcond_i64(c
->cond
, c
->u
.s64
.a
, c
->u
.s64
.b
, lab
);
1226 tcg_gen_brcond_i32(c
->cond
, c
->u
.s32
.a
, c
->u
.s32
.b
, lab
);
1229 /* Branch not taken. */
1231 tcg_gen_movi_i64(psw_addr
, s
->next_pc
);
1232 tcg_gen_exit_tb((uintptr_t)s
->tb
+ 0);
1237 tcg_gen_movi_i64(psw_addr
, dest
);
1238 tcg_gen_exit_tb((uintptr_t)s
->tb
+ 1);
1242 /* Fallthru can use goto_tb, but taken branch cannot. */
1243 /* Store taken branch destination before the brcond. This
1244 avoids having to allocate a new local temp to hold it.
1245 We'll overwrite this in the not taken case anyway. */
1247 tcg_gen_mov_i64(psw_addr
, cdest
);
1250 lab
= gen_new_label();
1252 tcg_gen_brcond_i64(c
->cond
, c
->u
.s64
.a
, c
->u
.s64
.b
, lab
);
1254 tcg_gen_brcond_i32(c
->cond
, c
->u
.s32
.a
, c
->u
.s32
.b
, lab
);
1257 /* Branch not taken. */
1260 tcg_gen_movi_i64(psw_addr
, s
->next_pc
);
1261 tcg_gen_exit_tb((uintptr_t)s
->tb
+ 0);
1265 tcg_gen_movi_i64(psw_addr
, dest
);
1267 ret
= EXIT_PC_UPDATED
;
1270 /* Fallthru cannot use goto_tb. This by itself is vanishingly rare.
1271 Most commonly we're single-stepping or some other condition that
1272 disables all use of goto_tb. Just update the PC and exit. */
1274 TCGv_i64 next
= tcg_const_i64(s
->next_pc
);
1276 cdest
= tcg_const_i64(dest
);
1280 tcg_gen_movcond_i64(c
->cond
, psw_addr
, c
->u
.s64
.a
, c
->u
.s64
.b
,
1283 TCGv_i32 t0
= tcg_temp_new_i32();
1284 TCGv_i64 t1
= tcg_temp_new_i64();
1285 TCGv_i64 z
= tcg_const_i64(0);
1286 tcg_gen_setcond_i32(c
->cond
, t0
, c
->u
.s32
.a
, c
->u
.s32
.b
);
1287 tcg_gen_extu_i32_i64(t1
, t0
);
1288 tcg_temp_free_i32(t0
);
1289 tcg_gen_movcond_i64(TCG_COND_NE
, psw_addr
, t1
, z
, cdest
, next
);
1290 tcg_temp_free_i64(t1
);
1291 tcg_temp_free_i64(z
);
1295 tcg_temp_free_i64(cdest
);
1297 tcg_temp_free_i64(next
);
1299 ret
= EXIT_PC_UPDATED
;
1307 /* ====================================================================== */
1308 /* The operations. These perform the bulk of the work for any insn,
1309 usually after the operands have been loaded and output initialized. */
1311 static ExitStatus
op_abs(DisasContext
*s
, DisasOps
*o
)
1313 gen_helper_abs_i64(o
->out
, o
->in2
);
1317 static ExitStatus
op_absf32(DisasContext
*s
, DisasOps
*o
)
1319 tcg_gen_andi_i64(o
->out
, o
->in2
, 0x7fffffffull
);
1323 static ExitStatus
op_absf64(DisasContext
*s
, DisasOps
*o
)
1325 tcg_gen_andi_i64(o
->out
, o
->in2
, 0x7fffffffffffffffull
);
1329 static ExitStatus
op_absf128(DisasContext
*s
, DisasOps
*o
)
1331 tcg_gen_andi_i64(o
->out
, o
->in1
, 0x7fffffffffffffffull
);
1332 tcg_gen_mov_i64(o
->out2
, o
->in2
);
1336 static ExitStatus
op_add(DisasContext
*s
, DisasOps
*o
)
1338 tcg_gen_add_i64(o
->out
, o
->in1
, o
->in2
);
1342 static ExitStatus
op_addc(DisasContext
*s
, DisasOps
*o
)
1347 tcg_gen_add_i64(o
->out
, o
->in1
, o
->in2
);
1349 /* The carry flag is the msb of CC, therefore the branch mask that would
1350 create that comparison is 3. Feeding the generated comparison to
1351 setcond produces the carry flag that we desire. */
1352 disas_jcc(s
, &cmp
, 3);
1353 carry
= tcg_temp_new_i64();
1355 tcg_gen_setcond_i64(cmp
.cond
, carry
, cmp
.u
.s64
.a
, cmp
.u
.s64
.b
);
1357 TCGv_i32 t
= tcg_temp_new_i32();
1358 tcg_gen_setcond_i32(cmp
.cond
, t
, cmp
.u
.s32
.a
, cmp
.u
.s32
.b
);
1359 tcg_gen_extu_i32_i64(carry
, t
);
1360 tcg_temp_free_i32(t
);
1364 tcg_gen_add_i64(o
->out
, o
->out
, carry
);
1365 tcg_temp_free_i64(carry
);
1369 static ExitStatus
op_aeb(DisasContext
*s
, DisasOps
*o
)
1371 gen_helper_aeb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
1375 static ExitStatus
op_adb(DisasContext
*s
, DisasOps
*o
)
1377 gen_helper_adb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
1381 static ExitStatus
op_axb(DisasContext
*s
, DisasOps
*o
)
1383 gen_helper_axb(o
->out
, cpu_env
, o
->out
, o
->out2
, o
->in1
, o
->in2
);
1384 return_low128(o
->out2
);
1388 static ExitStatus
op_and(DisasContext
*s
, DisasOps
*o
)
1390 tcg_gen_and_i64(o
->out
, o
->in1
, o
->in2
);
1394 static ExitStatus
op_andi(DisasContext
*s
, DisasOps
*o
)
1396 int shift
= s
->insn
->data
& 0xff;
1397 int size
= s
->insn
->data
>> 8;
1398 uint64_t mask
= ((1ull << size
) - 1) << shift
;
1401 tcg_gen_shli_i64(o
->in2
, o
->in2
, shift
);
1402 tcg_gen_ori_i64(o
->in2
, o
->in2
, ~mask
);
1403 tcg_gen_and_i64(o
->out
, o
->in1
, o
->in2
);
1405 /* Produce the CC from only the bits manipulated. */
1406 tcg_gen_andi_i64(cc_dst
, o
->out
, mask
);
1407 set_cc_nz_u64(s
, cc_dst
);
1411 static ExitStatus
op_bas(DisasContext
*s
, DisasOps
*o
)
1413 tcg_gen_movi_i64(o
->out
, pc_to_link_info(s
, s
->next_pc
));
1414 if (!TCGV_IS_UNUSED_I64(o
->in2
)) {
1415 tcg_gen_mov_i64(psw_addr
, o
->in2
);
1416 return EXIT_PC_UPDATED
;
1422 static ExitStatus
op_basi(DisasContext
*s
, DisasOps
*o
)
1424 tcg_gen_movi_i64(o
->out
, pc_to_link_info(s
, s
->next_pc
));
1425 return help_goto_direct(s
, s
->pc
+ 2 * get_field(s
->fields
, i2
));
1428 static ExitStatus
op_bc(DisasContext
*s
, DisasOps
*o
)
1430 int m1
= get_field(s
->fields
, m1
);
1431 bool is_imm
= have_field(s
->fields
, i2
);
1432 int imm
= is_imm
? get_field(s
->fields
, i2
) : 0;
1435 disas_jcc(s
, &c
, m1
);
1436 return help_branch(s
, &c
, is_imm
, imm
, o
->in2
);
1439 static ExitStatus
op_bct32(DisasContext
*s
, DisasOps
*o
)
1441 int r1
= get_field(s
->fields
, r1
);
1442 bool is_imm
= have_field(s
->fields
, i2
);
1443 int imm
= is_imm
? get_field(s
->fields
, i2
) : 0;
1447 c
.cond
= TCG_COND_NE
;
1452 t
= tcg_temp_new_i64();
1453 tcg_gen_subi_i64(t
, regs
[r1
], 1);
1454 store_reg32_i64(r1
, t
);
1455 c
.u
.s32
.a
= tcg_temp_new_i32();
1456 c
.u
.s32
.b
= tcg_const_i32(0);
1457 tcg_gen_trunc_i64_i32(c
.u
.s32
.a
, t
);
1458 tcg_temp_free_i64(t
);
1460 return help_branch(s
, &c
, is_imm
, imm
, o
->in2
);
1463 static ExitStatus
op_bct64(DisasContext
*s
, DisasOps
*o
)
1465 int r1
= get_field(s
->fields
, r1
);
1466 bool is_imm
= have_field(s
->fields
, i2
);
1467 int imm
= is_imm
? get_field(s
->fields
, i2
) : 0;
1470 c
.cond
= TCG_COND_NE
;
1475 tcg_gen_subi_i64(regs
[r1
], regs
[r1
], 1);
1476 c
.u
.s64
.a
= regs
[r1
];
1477 c
.u
.s64
.b
= tcg_const_i64(0);
1479 return help_branch(s
, &c
, is_imm
, imm
, o
->in2
);
1482 static ExitStatus
op_bx32(DisasContext
*s
, DisasOps
*o
)
1484 int r1
= get_field(s
->fields
, r1
);
1485 int r3
= get_field(s
->fields
, r3
);
1486 bool is_imm
= have_field(s
->fields
, i2
);
1487 int imm
= is_imm
? get_field(s
->fields
, i2
) : 0;
1491 c
.cond
= (s
->insn
->data
? TCG_COND_LE
: TCG_COND_GT
);
1496 t
= tcg_temp_new_i64();
1497 tcg_gen_add_i64(t
, regs
[r1
], regs
[r3
]);
1498 c
.u
.s32
.a
= tcg_temp_new_i32();
1499 c
.u
.s32
.b
= tcg_temp_new_i32();
1500 tcg_gen_trunc_i64_i32(c
.u
.s32
.a
, t
);
1501 tcg_gen_trunc_i64_i32(c
.u
.s32
.b
, regs
[r3
| 1]);
1502 store_reg32_i64(r1
, t
);
1503 tcg_temp_free_i64(t
);
1505 return help_branch(s
, &c
, is_imm
, imm
, o
->in2
);
1508 static ExitStatus
op_bx64(DisasContext
*s
, DisasOps
*o
)
1510 int r1
= get_field(s
->fields
, r1
);
1511 int r3
= get_field(s
->fields
, r3
);
1512 bool is_imm
= have_field(s
->fields
, i2
);
1513 int imm
= is_imm
? get_field(s
->fields
, i2
) : 0;
1516 c
.cond
= (s
->insn
->data
? TCG_COND_LE
: TCG_COND_GT
);
1519 if (r1
== (r3
| 1)) {
1520 c
.u
.s64
.b
= load_reg(r3
| 1);
1523 c
.u
.s64
.b
= regs
[r3
| 1];
1527 tcg_gen_add_i64(regs
[r1
], regs
[r1
], regs
[r3
]);
1528 c
.u
.s64
.a
= regs
[r1
];
1531 return help_branch(s
, &c
, is_imm
, imm
, o
->in2
);
1534 static ExitStatus
op_cj(DisasContext
*s
, DisasOps
*o
)
1536 int imm
, m3
= get_field(s
->fields
, m3
);
1540 c
.cond
= ltgt_cond
[m3
];
1541 if (s
->insn
->data
) {
1542 c
.cond
= tcg_unsigned_cond(c
.cond
);
1544 c
.is_64
= c
.g1
= c
.g2
= true;
1548 is_imm
= have_field(s
->fields
, i4
);
1550 imm
= get_field(s
->fields
, i4
);
1553 o
->out
= get_address(s
, 0, get_field(s
->fields
, b4
),
1554 get_field(s
->fields
, d4
));
1557 return help_branch(s
, &c
, is_imm
, imm
, o
->out
);
1560 static ExitStatus
op_ceb(DisasContext
*s
, DisasOps
*o
)
1562 gen_helper_ceb(cc_op
, cpu_env
, o
->in1
, o
->in2
);
1567 static ExitStatus
op_cdb(DisasContext
*s
, DisasOps
*o
)
1569 gen_helper_cdb(cc_op
, cpu_env
, o
->in1
, o
->in2
);
1574 static ExitStatus
op_cxb(DisasContext
*s
, DisasOps
*o
)
1576 gen_helper_cxb(cc_op
, cpu_env
, o
->out
, o
->out2
, o
->in1
, o
->in2
);
1581 static ExitStatus
op_cfeb(DisasContext
*s
, DisasOps
*o
)
1583 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1584 gen_helper_cfeb(o
->out
, cpu_env
, o
->in2
, m3
);
1585 tcg_temp_free_i32(m3
);
1586 gen_set_cc_nz_f32(s
, o
->in2
);
1590 static ExitStatus
op_cfdb(DisasContext
*s
, DisasOps
*o
)
1592 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1593 gen_helper_cfdb(o
->out
, cpu_env
, o
->in2
, m3
);
1594 tcg_temp_free_i32(m3
);
1595 gen_set_cc_nz_f64(s
, o
->in2
);
1599 static ExitStatus
op_cfxb(DisasContext
*s
, DisasOps
*o
)
1601 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1602 gen_helper_cfxb(o
->out
, cpu_env
, o
->in1
, o
->in2
, m3
);
1603 tcg_temp_free_i32(m3
);
1604 gen_set_cc_nz_f128(s
, o
->in1
, o
->in2
);
1608 static ExitStatus
op_cgeb(DisasContext
*s
, DisasOps
*o
)
1610 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1611 gen_helper_cgeb(o
->out
, cpu_env
, o
->in2
, m3
);
1612 tcg_temp_free_i32(m3
);
1613 gen_set_cc_nz_f32(s
, o
->in2
);
1617 static ExitStatus
op_cgdb(DisasContext
*s
, DisasOps
*o
)
1619 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1620 gen_helper_cgdb(o
->out
, cpu_env
, o
->in2
, m3
);
1621 tcg_temp_free_i32(m3
);
1622 gen_set_cc_nz_f64(s
, o
->in2
);
1626 static ExitStatus
op_cgxb(DisasContext
*s
, DisasOps
*o
)
1628 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1629 gen_helper_cgxb(o
->out
, cpu_env
, o
->in1
, o
->in2
, m3
);
1630 tcg_temp_free_i32(m3
);
1631 gen_set_cc_nz_f128(s
, o
->in1
, o
->in2
);
1635 static ExitStatus
op_clfeb(DisasContext
*s
, DisasOps
*o
)
1637 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1638 gen_helper_clfeb(o
->out
, cpu_env
, o
->in2
, m3
);
1639 tcg_temp_free_i32(m3
);
1640 gen_set_cc_nz_f32(s
, o
->in2
);
1644 static ExitStatus
op_clfdb(DisasContext
*s
, DisasOps
*o
)
1646 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1647 gen_helper_clfdb(o
->out
, cpu_env
, o
->in2
, m3
);
1648 tcg_temp_free_i32(m3
);
1649 gen_set_cc_nz_f64(s
, o
->in2
);
1653 static ExitStatus
op_clfxb(DisasContext
*s
, DisasOps
*o
)
1655 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1656 gen_helper_clfxb(o
->out
, cpu_env
, o
->in1
, o
->in2
, m3
);
1657 tcg_temp_free_i32(m3
);
1658 gen_set_cc_nz_f128(s
, o
->in1
, o
->in2
);
1662 static ExitStatus
op_clgeb(DisasContext
*s
, DisasOps
*o
)
1664 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1665 gen_helper_clgeb(o
->out
, cpu_env
, o
->in2
, m3
);
1666 tcg_temp_free_i32(m3
);
1667 gen_set_cc_nz_f32(s
, o
->in2
);
1671 static ExitStatus
op_clgdb(DisasContext
*s
, DisasOps
*o
)
1673 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1674 gen_helper_clgdb(o
->out
, cpu_env
, o
->in2
, m3
);
1675 tcg_temp_free_i32(m3
);
1676 gen_set_cc_nz_f64(s
, o
->in2
);
1680 static ExitStatus
op_clgxb(DisasContext
*s
, DisasOps
*o
)
1682 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1683 gen_helper_clgxb(o
->out
, cpu_env
, o
->in1
, o
->in2
, m3
);
1684 tcg_temp_free_i32(m3
);
1685 gen_set_cc_nz_f128(s
, o
->in1
, o
->in2
);
1689 static ExitStatus
op_cegb(DisasContext
*s
, DisasOps
*o
)
1691 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1692 gen_helper_cegb(o
->out
, cpu_env
, o
->in2
, m3
);
1693 tcg_temp_free_i32(m3
);
1697 static ExitStatus
op_cdgb(DisasContext
*s
, DisasOps
*o
)
1699 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1700 gen_helper_cdgb(o
->out
, cpu_env
, o
->in2
, m3
);
1701 tcg_temp_free_i32(m3
);
1705 static ExitStatus
op_cxgb(DisasContext
*s
, DisasOps
*o
)
1707 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1708 gen_helper_cxgb(o
->out
, cpu_env
, o
->in2
, m3
);
1709 tcg_temp_free_i32(m3
);
1710 return_low128(o
->out2
);
1714 static ExitStatus
op_celgb(DisasContext
*s
, DisasOps
*o
)
1716 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1717 gen_helper_celgb(o
->out
, cpu_env
, o
->in2
, m3
);
1718 tcg_temp_free_i32(m3
);
1722 static ExitStatus
op_cdlgb(DisasContext
*s
, DisasOps
*o
)
1724 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1725 gen_helper_cdlgb(o
->out
, cpu_env
, o
->in2
, m3
);
1726 tcg_temp_free_i32(m3
);
1730 static ExitStatus
op_cxlgb(DisasContext
*s
, DisasOps
*o
)
1732 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1733 gen_helper_cxlgb(o
->out
, cpu_env
, o
->in2
, m3
);
1734 tcg_temp_free_i32(m3
);
1735 return_low128(o
->out2
);
1739 static ExitStatus
op_cksm(DisasContext
*s
, DisasOps
*o
)
1741 int r2
= get_field(s
->fields
, r2
);
1742 TCGv_i64 len
= tcg_temp_new_i64();
1744 potential_page_fault(s
);
1745 gen_helper_cksm(len
, cpu_env
, o
->in1
, o
->in2
, regs
[r2
+ 1]);
1747 return_low128(o
->out
);
1749 tcg_gen_add_i64(regs
[r2
], regs
[r2
], len
);
1750 tcg_gen_sub_i64(regs
[r2
+ 1], regs
[r2
+ 1], len
);
1751 tcg_temp_free_i64(len
);
1756 static ExitStatus
op_clc(DisasContext
*s
, DisasOps
*o
)
1758 int l
= get_field(s
->fields
, l1
);
1763 tcg_gen_qemu_ld8u(cc_src
, o
->addr1
, get_mem_index(s
));
1764 tcg_gen_qemu_ld8u(cc_dst
, o
->in2
, get_mem_index(s
));
1767 tcg_gen_qemu_ld16u(cc_src
, o
->addr1
, get_mem_index(s
));
1768 tcg_gen_qemu_ld16u(cc_dst
, o
->in2
, get_mem_index(s
));
1771 tcg_gen_qemu_ld32u(cc_src
, o
->addr1
, get_mem_index(s
));
1772 tcg_gen_qemu_ld32u(cc_dst
, o
->in2
, get_mem_index(s
));
1775 tcg_gen_qemu_ld64(cc_src
, o
->addr1
, get_mem_index(s
));
1776 tcg_gen_qemu_ld64(cc_dst
, o
->in2
, get_mem_index(s
));
1779 potential_page_fault(s
);
1780 vl
= tcg_const_i32(l
);
1781 gen_helper_clc(cc_op
, cpu_env
, vl
, o
->addr1
, o
->in2
);
1782 tcg_temp_free_i32(vl
);
1786 gen_op_update2_cc_i64(s
, CC_OP_LTUGTU_64
, cc_src
, cc_dst
);
1790 static ExitStatus
op_clcle(DisasContext
*s
, DisasOps
*o
)
1792 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
1793 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
1794 potential_page_fault(s
);
1795 gen_helper_clcle(cc_op
, cpu_env
, r1
, o
->in2
, r3
);
1796 tcg_temp_free_i32(r1
);
1797 tcg_temp_free_i32(r3
);
1802 static ExitStatus
op_clm(DisasContext
*s
, DisasOps
*o
)
1804 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1805 TCGv_i32 t1
= tcg_temp_new_i32();
1806 tcg_gen_trunc_i64_i32(t1
, o
->in1
);
1807 potential_page_fault(s
);
1808 gen_helper_clm(cc_op
, cpu_env
, t1
, m3
, o
->in2
);
1810 tcg_temp_free_i32(t1
);
1811 tcg_temp_free_i32(m3
);
1815 static ExitStatus
op_clst(DisasContext
*s
, DisasOps
*o
)
1817 potential_page_fault(s
);
1818 gen_helper_clst(o
->in1
, cpu_env
, regs
[0], o
->in1
, o
->in2
);
1820 return_low128(o
->in2
);
1824 static ExitStatus
op_cps(DisasContext
*s
, DisasOps
*o
)
1826 TCGv_i64 t
= tcg_temp_new_i64();
1827 tcg_gen_andi_i64(t
, o
->in1
, 0x8000000000000000ull
);
1828 tcg_gen_andi_i64(o
->out
, o
->in2
, 0x7fffffffffffffffull
);
1829 tcg_gen_or_i64(o
->out
, o
->out
, t
);
1830 tcg_temp_free_i64(t
);
1834 static ExitStatus
op_cs(DisasContext
*s
, DisasOps
*o
)
1836 /* FIXME: needs an atomic solution for CONFIG_USER_ONLY. */
1837 int d2
= get_field(s
->fields
, d2
);
1838 int b2
= get_field(s
->fields
, b2
);
1839 int is_64
= s
->insn
->data
;
1840 TCGv_i64 addr
, mem
, cc
, z
;
1842 /* Note that in1 = R3 (new value) and
1843 in2 = (zero-extended) R1 (expected value). */
1845 /* Load the memory into the (temporary) output. While the PoO only talks
1846 about moving the memory to R1 on inequality, if we include equality it
1847 means that R1 is equal to the memory in all conditions. */
1848 addr
= get_address(s
, 0, b2
, d2
);
1850 tcg_gen_qemu_ld64(o
->out
, addr
, get_mem_index(s
));
1852 tcg_gen_qemu_ld32u(o
->out
, addr
, get_mem_index(s
));
1855 /* Are the memory and expected values (un)equal? Note that this setcond
1856 produces the output CC value, thus the NE sense of the test. */
1857 cc
= tcg_temp_new_i64();
1858 tcg_gen_setcond_i64(TCG_COND_NE
, cc
, o
->in2
, o
->out
);
1860 /* If the memory and expected values are equal (CC==0), copy R3 to MEM.
1861 Recall that we are allowed to unconditionally issue the store (and
1862 thus any possible write trap), so (re-)store the original contents
1863 of MEM in case of inequality. */
1864 z
= tcg_const_i64(0);
1865 mem
= tcg_temp_new_i64();
1866 tcg_gen_movcond_i64(TCG_COND_EQ
, mem
, cc
, z
, o
->in1
, o
->out
);
1868 tcg_gen_qemu_st64(mem
, addr
, get_mem_index(s
));
1870 tcg_gen_qemu_st32(mem
, addr
, get_mem_index(s
));
1872 tcg_temp_free_i64(z
);
1873 tcg_temp_free_i64(mem
);
1874 tcg_temp_free_i64(addr
);
1876 /* Store CC back to cc_op. Wait until after the store so that any
1877 exception gets the old cc_op value. */
1878 tcg_gen_trunc_i64_i32(cc_op
, cc
);
1879 tcg_temp_free_i64(cc
);
1884 static ExitStatus
op_cdsg(DisasContext
*s
, DisasOps
*o
)
1886 /* FIXME: needs an atomic solution for CONFIG_USER_ONLY. */
1887 int r1
= get_field(s
->fields
, r1
);
1888 int r3
= get_field(s
->fields
, r3
);
1889 int d2
= get_field(s
->fields
, d2
);
1890 int b2
= get_field(s
->fields
, b2
);
1891 TCGv_i64 addrh
, addrl
, memh
, meml
, outh
, outl
, cc
, z
;
1893 /* Note that R1:R1+1 = expected value and R3:R3+1 = new value. */
1895 addrh
= get_address(s
, 0, b2
, d2
);
1896 addrl
= get_address(s
, 0, b2
, d2
+ 8);
1897 outh
= tcg_temp_new_i64();
1898 outl
= tcg_temp_new_i64();
1900 tcg_gen_qemu_ld64(outh
, addrh
, get_mem_index(s
));
1901 tcg_gen_qemu_ld64(outl
, addrl
, get_mem_index(s
));
1903 /* Fold the double-word compare with arithmetic. */
1904 cc
= tcg_temp_new_i64();
1905 z
= tcg_temp_new_i64();
1906 tcg_gen_xor_i64(cc
, outh
, regs
[r1
]);
1907 tcg_gen_xor_i64(z
, outl
, regs
[r1
+ 1]);
1908 tcg_gen_or_i64(cc
, cc
, z
);
1909 tcg_gen_movi_i64(z
, 0);
1910 tcg_gen_setcond_i64(TCG_COND_NE
, cc
, cc
, z
);
1912 memh
= tcg_temp_new_i64();
1913 meml
= tcg_temp_new_i64();
1914 tcg_gen_movcond_i64(TCG_COND_EQ
, memh
, cc
, z
, regs
[r3
], outh
);
1915 tcg_gen_movcond_i64(TCG_COND_EQ
, meml
, cc
, z
, regs
[r3
+ 1], outl
);
1916 tcg_temp_free_i64(z
);
1918 tcg_gen_qemu_st64(memh
, addrh
, get_mem_index(s
));
1919 tcg_gen_qemu_st64(meml
, addrl
, get_mem_index(s
));
1920 tcg_temp_free_i64(memh
);
1921 tcg_temp_free_i64(meml
);
1922 tcg_temp_free_i64(addrh
);
1923 tcg_temp_free_i64(addrl
);
1925 /* Save back state now that we've passed all exceptions. */
1926 tcg_gen_mov_i64(regs
[r1
], outh
);
1927 tcg_gen_mov_i64(regs
[r1
+ 1], outl
);
1928 tcg_gen_trunc_i64_i32(cc_op
, cc
);
1929 tcg_temp_free_i64(outh
);
1930 tcg_temp_free_i64(outl
);
1931 tcg_temp_free_i64(cc
);
1936 #ifndef CONFIG_USER_ONLY
1937 static ExitStatus
op_csp(DisasContext
*s
, DisasOps
*o
)
1939 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
1940 check_privileged(s
);
1941 gen_helper_csp(cc_op
, cpu_env
, r1
, o
->in2
);
1942 tcg_temp_free_i32(r1
);
1948 static ExitStatus
op_cvd(DisasContext
*s
, DisasOps
*o
)
1950 TCGv_i64 t1
= tcg_temp_new_i64();
1951 TCGv_i32 t2
= tcg_temp_new_i32();
1952 tcg_gen_trunc_i64_i32(t2
, o
->in1
);
1953 gen_helper_cvd(t1
, t2
);
1954 tcg_temp_free_i32(t2
);
1955 tcg_gen_qemu_st64(t1
, o
->in2
, get_mem_index(s
));
1956 tcg_temp_free_i64(t1
);
1960 static ExitStatus
op_ct(DisasContext
*s
, DisasOps
*o
)
1962 int m3
= get_field(s
->fields
, m3
);
1963 TCGLabel
*lab
= gen_new_label();
1967 c
= tcg_invert_cond(ltgt_cond
[m3
]);
1968 if (s
->insn
->data
) {
1969 c
= tcg_unsigned_cond(c
);
1971 tcg_gen_brcond_i64(c
, o
->in1
, o
->in2
, lab
);
1973 /* Set DXC to 0xff. */
1974 t
= tcg_temp_new_i32();
1975 tcg_gen_ld_i32(t
, cpu_env
, offsetof(CPUS390XState
, fpc
));
1976 tcg_gen_ori_i32(t
, t
, 0xff00);
1977 tcg_gen_st_i32(t
, cpu_env
, offsetof(CPUS390XState
, fpc
));
1978 tcg_temp_free_i32(t
);
1981 gen_program_exception(s
, PGM_DATA
);
1987 #ifndef CONFIG_USER_ONLY
1988 static ExitStatus
op_diag(DisasContext
*s
, DisasOps
*o
)
1992 check_privileged(s
);
1993 potential_page_fault(s
);
1995 /* We pretend the format is RX_a so that D2 is the field we want. */
1996 tmp
= tcg_const_i32(get_field(s
->fields
, d2
) & 0xfff);
1997 gen_helper_diag(regs
[2], cpu_env
, tmp
, regs
[2], regs
[1]);
1998 tcg_temp_free_i32(tmp
);
2003 static ExitStatus
op_divs32(DisasContext
*s
, DisasOps
*o
)
2005 gen_helper_divs32(o
->out2
, cpu_env
, o
->in1
, o
->in2
);
2006 return_low128(o
->out
);
2010 static ExitStatus
op_divu32(DisasContext
*s
, DisasOps
*o
)
2012 gen_helper_divu32(o
->out2
, cpu_env
, o
->in1
, o
->in2
);
2013 return_low128(o
->out
);
2017 static ExitStatus
op_divs64(DisasContext
*s
, DisasOps
*o
)
2019 gen_helper_divs64(o
->out2
, cpu_env
, o
->in1
, o
->in2
);
2020 return_low128(o
->out
);
2024 static ExitStatus
op_divu64(DisasContext
*s
, DisasOps
*o
)
2026 gen_helper_divu64(o
->out2
, cpu_env
, o
->out
, o
->out2
, o
->in2
);
2027 return_low128(o
->out
);
2031 static ExitStatus
op_deb(DisasContext
*s
, DisasOps
*o
)
2033 gen_helper_deb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2037 static ExitStatus
op_ddb(DisasContext
*s
, DisasOps
*o
)
2039 gen_helper_ddb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2043 static ExitStatus
op_dxb(DisasContext
*s
, DisasOps
*o
)
2045 gen_helper_dxb(o
->out
, cpu_env
, o
->out
, o
->out2
, o
->in1
, o
->in2
);
2046 return_low128(o
->out2
);
2050 static ExitStatus
op_ear(DisasContext
*s
, DisasOps
*o
)
2052 int r2
= get_field(s
->fields
, r2
);
2053 tcg_gen_ld32u_i64(o
->out
, cpu_env
, offsetof(CPUS390XState
, aregs
[r2
]));
2057 static ExitStatus
op_ecag(DisasContext
*s
, DisasOps
*o
)
2059 /* No cache information provided. */
2060 tcg_gen_movi_i64(o
->out
, -1);
2064 static ExitStatus
op_efpc(DisasContext
*s
, DisasOps
*o
)
2066 tcg_gen_ld32u_i64(o
->out
, cpu_env
, offsetof(CPUS390XState
, fpc
));
2070 static ExitStatus
op_epsw(DisasContext
*s
, DisasOps
*o
)
2072 int r1
= get_field(s
->fields
, r1
);
2073 int r2
= get_field(s
->fields
, r2
);
2074 TCGv_i64 t
= tcg_temp_new_i64();
2076 /* Note the "subsequently" in the PoO, which implies a defined result
2077 if r1 == r2. Thus we cannot defer these writes to an output hook. */
2078 tcg_gen_shri_i64(t
, psw_mask
, 32);
2079 store_reg32_i64(r1
, t
);
2081 store_reg32_i64(r2
, psw_mask
);
2084 tcg_temp_free_i64(t
);
2088 static ExitStatus
op_ex(DisasContext
*s
, DisasOps
*o
)
2090 /* ??? Perhaps a better way to implement EXECUTE is to set a bit in
2091 tb->flags, (ab)use the tb->cs_base field as the address of
2092 the template in memory, and grab 8 bits of tb->flags/cflags for
2093 the contents of the register. We would then recognize all this
2094 in gen_intermediate_code_internal, generating code for exactly
2095 one instruction. This new TB then gets executed normally.
2097 On the other hand, this seems to be mostly used for modifying
2098 MVC inside of memcpy, which needs a helper call anyway. So
2099 perhaps this doesn't bear thinking about any further. */
2106 tmp
= tcg_const_i64(s
->next_pc
);
2107 gen_helper_ex(cc_op
, cpu_env
, cc_op
, o
->in1
, o
->in2
, tmp
);
2108 tcg_temp_free_i64(tmp
);
2113 static ExitStatus
op_flogr(DisasContext
*s
, DisasOps
*o
)
2115 /* We'll use the original input for cc computation, since we get to
2116 compare that against 0, which ought to be better than comparing
2117 the real output against 64. It also lets cc_dst be a convenient
2118 temporary during our computation. */
2119 gen_op_update1_cc_i64(s
, CC_OP_FLOGR
, o
->in2
);
2121 /* R1 = IN ? CLZ(IN) : 64. */
2122 gen_helper_clz(o
->out
, o
->in2
);
2124 /* R1+1 = IN & ~(found bit). Note that we may attempt to shift this
2125 value by 64, which is undefined. But since the shift is 64 iff the
2126 input is zero, we still get the correct result after and'ing. */
2127 tcg_gen_movi_i64(o
->out2
, 0x8000000000000000ull
);
2128 tcg_gen_shr_i64(o
->out2
, o
->out2
, o
->out
);
2129 tcg_gen_andc_i64(o
->out2
, cc_dst
, o
->out2
);
2133 static ExitStatus
op_icm(DisasContext
*s
, DisasOps
*o
)
2135 int m3
= get_field(s
->fields
, m3
);
2136 int pos
, len
, base
= s
->insn
->data
;
2137 TCGv_i64 tmp
= tcg_temp_new_i64();
2142 /* Effectively a 32-bit load. */
2143 tcg_gen_qemu_ld32u(tmp
, o
->in2
, get_mem_index(s
));
2150 /* Effectively a 16-bit load. */
2151 tcg_gen_qemu_ld16u(tmp
, o
->in2
, get_mem_index(s
));
2159 /* Effectively an 8-bit load. */
2160 tcg_gen_qemu_ld8u(tmp
, o
->in2
, get_mem_index(s
));
2165 pos
= base
+ ctz32(m3
) * 8;
2166 tcg_gen_deposit_i64(o
->out
, o
->out
, tmp
, pos
, len
);
2167 ccm
= ((1ull << len
) - 1) << pos
;
2171 /* This is going to be a sequence of loads and inserts. */
2172 pos
= base
+ 32 - 8;
2176 tcg_gen_qemu_ld8u(tmp
, o
->in2
, get_mem_index(s
));
2177 tcg_gen_addi_i64(o
->in2
, o
->in2
, 1);
2178 tcg_gen_deposit_i64(o
->out
, o
->out
, tmp
, pos
, 8);
2181 m3
= (m3
<< 1) & 0xf;
2187 tcg_gen_movi_i64(tmp
, ccm
);
2188 gen_op_update2_cc_i64(s
, CC_OP_ICM
, tmp
, o
->out
);
2189 tcg_temp_free_i64(tmp
);
2193 static ExitStatus
op_insi(DisasContext
*s
, DisasOps
*o
)
2195 int shift
= s
->insn
->data
& 0xff;
2196 int size
= s
->insn
->data
>> 8;
2197 tcg_gen_deposit_i64(o
->out
, o
->in1
, o
->in2
, shift
, size
);
2201 static ExitStatus
op_ipm(DisasContext
*s
, DisasOps
*o
)
2206 tcg_gen_andi_i64(o
->out
, o
->out
, ~0xff000000ull
);
2208 t1
= tcg_temp_new_i64();
2209 tcg_gen_shli_i64(t1
, psw_mask
, 20);
2210 tcg_gen_shri_i64(t1
, t1
, 36);
2211 tcg_gen_or_i64(o
->out
, o
->out
, t1
);
2213 tcg_gen_extu_i32_i64(t1
, cc_op
);
2214 tcg_gen_shli_i64(t1
, t1
, 28);
2215 tcg_gen_or_i64(o
->out
, o
->out
, t1
);
2216 tcg_temp_free_i64(t1
);
2220 #ifndef CONFIG_USER_ONLY
2221 static ExitStatus
op_ipte(DisasContext
*s
, DisasOps
*o
)
2223 check_privileged(s
);
2224 gen_helper_ipte(cpu_env
, o
->in1
, o
->in2
);
2228 static ExitStatus
op_iske(DisasContext
*s
, DisasOps
*o
)
2230 check_privileged(s
);
2231 gen_helper_iske(o
->out
, cpu_env
, o
->in2
);
2236 static ExitStatus
op_ldeb(DisasContext
*s
, DisasOps
*o
)
2238 gen_helper_ldeb(o
->out
, cpu_env
, o
->in2
);
2242 static ExitStatus
op_ledb(DisasContext
*s
, DisasOps
*o
)
2244 gen_helper_ledb(o
->out
, cpu_env
, o
->in2
);
2248 static ExitStatus
op_ldxb(DisasContext
*s
, DisasOps
*o
)
2250 gen_helper_ldxb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2254 static ExitStatus
op_lexb(DisasContext
*s
, DisasOps
*o
)
2256 gen_helper_lexb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2260 static ExitStatus
op_lxdb(DisasContext
*s
, DisasOps
*o
)
2262 gen_helper_lxdb(o
->out
, cpu_env
, o
->in2
);
2263 return_low128(o
->out2
);
2267 static ExitStatus
op_lxeb(DisasContext
*s
, DisasOps
*o
)
2269 gen_helper_lxeb(o
->out
, cpu_env
, o
->in2
);
2270 return_low128(o
->out2
);
2274 static ExitStatus
op_llgt(DisasContext
*s
, DisasOps
*o
)
2276 tcg_gen_andi_i64(o
->out
, o
->in2
, 0x7fffffff);
2280 static ExitStatus
op_ld8s(DisasContext
*s
, DisasOps
*o
)
2282 tcg_gen_qemu_ld8s(o
->out
, o
->in2
, get_mem_index(s
));
2286 static ExitStatus
op_ld8u(DisasContext
*s
, DisasOps
*o
)
2288 tcg_gen_qemu_ld8u(o
->out
, o
->in2
, get_mem_index(s
));
2292 static ExitStatus
op_ld16s(DisasContext
*s
, DisasOps
*o
)
2294 tcg_gen_qemu_ld16s(o
->out
, o
->in2
, get_mem_index(s
));
2298 static ExitStatus
op_ld16u(DisasContext
*s
, DisasOps
*o
)
2300 tcg_gen_qemu_ld16u(o
->out
, o
->in2
, get_mem_index(s
));
2304 static ExitStatus
op_ld32s(DisasContext
*s
, DisasOps
*o
)
2306 tcg_gen_qemu_ld32s(o
->out
, o
->in2
, get_mem_index(s
));
2310 static ExitStatus
op_ld32u(DisasContext
*s
, DisasOps
*o
)
2312 tcg_gen_qemu_ld32u(o
->out
, o
->in2
, get_mem_index(s
));
2316 static ExitStatus
op_ld64(DisasContext
*s
, DisasOps
*o
)
2318 tcg_gen_qemu_ld64(o
->out
, o
->in2
, get_mem_index(s
));
2322 static ExitStatus
op_loc(DisasContext
*s
, DisasOps
*o
)
2326 disas_jcc(s
, &c
, get_field(s
->fields
, m3
));
2329 tcg_gen_movcond_i64(c
.cond
, o
->out
, c
.u
.s64
.a
, c
.u
.s64
.b
,
2333 TCGv_i32 t32
= tcg_temp_new_i32();
2336 tcg_gen_setcond_i32(c
.cond
, t32
, c
.u
.s32
.a
, c
.u
.s32
.b
);
2339 t
= tcg_temp_new_i64();
2340 tcg_gen_extu_i32_i64(t
, t32
);
2341 tcg_temp_free_i32(t32
);
2343 z
= tcg_const_i64(0);
2344 tcg_gen_movcond_i64(TCG_COND_NE
, o
->out
, t
, z
, o
->in2
, o
->in1
);
2345 tcg_temp_free_i64(t
);
2346 tcg_temp_free_i64(z
);
2352 #ifndef CONFIG_USER_ONLY
2353 static ExitStatus
op_lctl(DisasContext
*s
, DisasOps
*o
)
2355 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
2356 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
2357 check_privileged(s
);
2358 potential_page_fault(s
);
2359 gen_helper_lctl(cpu_env
, r1
, o
->in2
, r3
);
2360 tcg_temp_free_i32(r1
);
2361 tcg_temp_free_i32(r3
);
2365 static ExitStatus
op_lctlg(DisasContext
*s
, DisasOps
*o
)
2367 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
2368 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
2369 check_privileged(s
);
2370 potential_page_fault(s
);
2371 gen_helper_lctlg(cpu_env
, r1
, o
->in2
, r3
);
2372 tcg_temp_free_i32(r1
);
2373 tcg_temp_free_i32(r3
);
2376 static ExitStatus
op_lra(DisasContext
*s
, DisasOps
*o
)
2378 check_privileged(s
);
2379 potential_page_fault(s
);
2380 gen_helper_lra(o
->out
, cpu_env
, o
->in2
);
2385 static ExitStatus
op_lpsw(DisasContext
*s
, DisasOps
*o
)
2389 check_privileged(s
);
2391 t1
= tcg_temp_new_i64();
2392 t2
= tcg_temp_new_i64();
2393 tcg_gen_qemu_ld32u(t1
, o
->in2
, get_mem_index(s
));
2394 tcg_gen_addi_i64(o
->in2
, o
->in2
, 4);
2395 tcg_gen_qemu_ld32u(t2
, o
->in2
, get_mem_index(s
));
2396 /* Convert the 32-bit PSW_MASK into the 64-bit PSW_MASK. */
2397 tcg_gen_shli_i64(t1
, t1
, 32);
2398 gen_helper_load_psw(cpu_env
, t1
, t2
);
2399 tcg_temp_free_i64(t1
);
2400 tcg_temp_free_i64(t2
);
2401 return EXIT_NORETURN
;
2404 static ExitStatus
op_lpswe(DisasContext
*s
, DisasOps
*o
)
2408 check_privileged(s
);
2410 t1
= tcg_temp_new_i64();
2411 t2
= tcg_temp_new_i64();
2412 tcg_gen_qemu_ld64(t1
, o
->in2
, get_mem_index(s
));
2413 tcg_gen_addi_i64(o
->in2
, o
->in2
, 8);
2414 tcg_gen_qemu_ld64(t2
, o
->in2
, get_mem_index(s
));
2415 gen_helper_load_psw(cpu_env
, t1
, t2
);
2416 tcg_temp_free_i64(t1
);
2417 tcg_temp_free_i64(t2
);
2418 return EXIT_NORETURN
;
2422 static ExitStatus
op_lam(DisasContext
*s
, DisasOps
*o
)
2424 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
2425 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
2426 potential_page_fault(s
);
2427 gen_helper_lam(cpu_env
, r1
, o
->in2
, r3
);
2428 tcg_temp_free_i32(r1
);
2429 tcg_temp_free_i32(r3
);
2433 static ExitStatus
op_lm32(DisasContext
*s
, DisasOps
*o
)
2435 int r1
= get_field(s
->fields
, r1
);
2436 int r3
= get_field(s
->fields
, r3
);
2437 TCGv_i64 t
= tcg_temp_new_i64();
2438 TCGv_i64 t4
= tcg_const_i64(4);
2441 tcg_gen_qemu_ld32u(t
, o
->in2
, get_mem_index(s
));
2442 store_reg32_i64(r1
, t
);
2446 tcg_gen_add_i64(o
->in2
, o
->in2
, t4
);
2450 tcg_temp_free_i64(t
);
2451 tcg_temp_free_i64(t4
);
2455 static ExitStatus
op_lmh(DisasContext
*s
, DisasOps
*o
)
2457 int r1
= get_field(s
->fields
, r1
);
2458 int r3
= get_field(s
->fields
, r3
);
2459 TCGv_i64 t
= tcg_temp_new_i64();
2460 TCGv_i64 t4
= tcg_const_i64(4);
2463 tcg_gen_qemu_ld32u(t
, o
->in2
, get_mem_index(s
));
2464 store_reg32h_i64(r1
, t
);
2468 tcg_gen_add_i64(o
->in2
, o
->in2
, t4
);
2472 tcg_temp_free_i64(t
);
2473 tcg_temp_free_i64(t4
);
2477 static ExitStatus
op_lm64(DisasContext
*s
, DisasOps
*o
)
2479 int r1
= get_field(s
->fields
, r1
);
2480 int r3
= get_field(s
->fields
, r3
);
2481 TCGv_i64 t8
= tcg_const_i64(8);
2484 tcg_gen_qemu_ld64(regs
[r1
], o
->in2
, get_mem_index(s
));
2488 tcg_gen_add_i64(o
->in2
, o
->in2
, t8
);
2492 tcg_temp_free_i64(t8
);
2496 #ifndef CONFIG_USER_ONLY
2497 static ExitStatus
op_lura(DisasContext
*s
, DisasOps
*o
)
2499 check_privileged(s
);
2500 potential_page_fault(s
);
2501 gen_helper_lura(o
->out
, cpu_env
, o
->in2
);
2505 static ExitStatus
op_lurag(DisasContext
*s
, DisasOps
*o
)
2507 check_privileged(s
);
2508 potential_page_fault(s
);
2509 gen_helper_lurag(o
->out
, cpu_env
, o
->in2
);
2514 static ExitStatus
op_mov2(DisasContext
*s
, DisasOps
*o
)
2517 o
->g_out
= o
->g_in2
;
2518 TCGV_UNUSED_I64(o
->in2
);
2523 static ExitStatus
op_movx(DisasContext
*s
, DisasOps
*o
)
2527 o
->g_out
= o
->g_in1
;
2528 o
->g_out2
= o
->g_in2
;
2529 TCGV_UNUSED_I64(o
->in1
);
2530 TCGV_UNUSED_I64(o
->in2
);
2531 o
->g_in1
= o
->g_in2
= false;
2535 static ExitStatus
op_mvc(DisasContext
*s
, DisasOps
*o
)
2537 TCGv_i32 l
= tcg_const_i32(get_field(s
->fields
, l1
));
2538 potential_page_fault(s
);
2539 gen_helper_mvc(cpu_env
, l
, o
->addr1
, o
->in2
);
2540 tcg_temp_free_i32(l
);
2544 static ExitStatus
op_mvcl(DisasContext
*s
, DisasOps
*o
)
2546 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
2547 TCGv_i32 r2
= tcg_const_i32(get_field(s
->fields
, r2
));
2548 potential_page_fault(s
);
2549 gen_helper_mvcl(cc_op
, cpu_env
, r1
, r2
);
2550 tcg_temp_free_i32(r1
);
2551 tcg_temp_free_i32(r2
);
2556 static ExitStatus
op_mvcle(DisasContext
*s
, DisasOps
*o
)
2558 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
2559 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
2560 potential_page_fault(s
);
2561 gen_helper_mvcle(cc_op
, cpu_env
, r1
, o
->in2
, r3
);
2562 tcg_temp_free_i32(r1
);
2563 tcg_temp_free_i32(r3
);
2568 #ifndef CONFIG_USER_ONLY
2569 static ExitStatus
op_mvcp(DisasContext
*s
, DisasOps
*o
)
2571 int r1
= get_field(s
->fields
, l1
);
2572 check_privileged(s
);
2573 potential_page_fault(s
);
2574 gen_helper_mvcp(cc_op
, cpu_env
, regs
[r1
], o
->addr1
, o
->in2
);
2579 static ExitStatus
op_mvcs(DisasContext
*s
, DisasOps
*o
)
2581 int r1
= get_field(s
->fields
, l1
);
2582 check_privileged(s
);
2583 potential_page_fault(s
);
2584 gen_helper_mvcs(cc_op
, cpu_env
, regs
[r1
], o
->addr1
, o
->in2
);
2590 static ExitStatus
op_mvpg(DisasContext
*s
, DisasOps
*o
)
2592 potential_page_fault(s
);
2593 gen_helper_mvpg(cpu_env
, regs
[0], o
->in1
, o
->in2
);
2598 static ExitStatus
op_mvst(DisasContext
*s
, DisasOps
*o
)
2600 potential_page_fault(s
);
2601 gen_helper_mvst(o
->in1
, cpu_env
, regs
[0], o
->in1
, o
->in2
);
2603 return_low128(o
->in2
);
2607 static ExitStatus
op_mul(DisasContext
*s
, DisasOps
*o
)
2609 tcg_gen_mul_i64(o
->out
, o
->in1
, o
->in2
);
2613 static ExitStatus
op_mul128(DisasContext
*s
, DisasOps
*o
)
2615 tcg_gen_mulu2_i64(o
->out2
, o
->out
, o
->in1
, o
->in2
);
2619 static ExitStatus
op_meeb(DisasContext
*s
, DisasOps
*o
)
2621 gen_helper_meeb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2625 static ExitStatus
op_mdeb(DisasContext
*s
, DisasOps
*o
)
2627 gen_helper_mdeb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2631 static ExitStatus
op_mdb(DisasContext
*s
, DisasOps
*o
)
2633 gen_helper_mdb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2637 static ExitStatus
op_mxb(DisasContext
*s
, DisasOps
*o
)
2639 gen_helper_mxb(o
->out
, cpu_env
, o
->out
, o
->out2
, o
->in1
, o
->in2
);
2640 return_low128(o
->out2
);
2644 static ExitStatus
op_mxdb(DisasContext
*s
, DisasOps
*o
)
2646 gen_helper_mxdb(o
->out
, cpu_env
, o
->out
, o
->out2
, o
->in2
);
2647 return_low128(o
->out2
);
2651 static ExitStatus
op_maeb(DisasContext
*s
, DisasOps
*o
)
2653 TCGv_i64 r3
= load_freg32_i64(get_field(s
->fields
, r3
));
2654 gen_helper_maeb(o
->out
, cpu_env
, o
->in1
, o
->in2
, r3
);
2655 tcg_temp_free_i64(r3
);
2659 static ExitStatus
op_madb(DisasContext
*s
, DisasOps
*o
)
2661 int r3
= get_field(s
->fields
, r3
);
2662 gen_helper_madb(o
->out
, cpu_env
, o
->in1
, o
->in2
, fregs
[r3
]);
2666 static ExitStatus
op_mseb(DisasContext
*s
, DisasOps
*o
)
2668 TCGv_i64 r3
= load_freg32_i64(get_field(s
->fields
, r3
));
2669 gen_helper_mseb(o
->out
, cpu_env
, o
->in1
, o
->in2
, r3
);
2670 tcg_temp_free_i64(r3
);
2674 static ExitStatus
op_msdb(DisasContext
*s
, DisasOps
*o
)
2676 int r3
= get_field(s
->fields
, r3
);
2677 gen_helper_msdb(o
->out
, cpu_env
, o
->in1
, o
->in2
, fregs
[r3
]);
2681 static ExitStatus
op_nabs(DisasContext
*s
, DisasOps
*o
)
2683 gen_helper_nabs_i64(o
->out
, o
->in2
);
2687 static ExitStatus
op_nabsf32(DisasContext
*s
, DisasOps
*o
)
2689 tcg_gen_ori_i64(o
->out
, o
->in2
, 0x80000000ull
);
2693 static ExitStatus
op_nabsf64(DisasContext
*s
, DisasOps
*o
)
2695 tcg_gen_ori_i64(o
->out
, o
->in2
, 0x8000000000000000ull
);
2699 static ExitStatus
op_nabsf128(DisasContext
*s
, DisasOps
*o
)
2701 tcg_gen_ori_i64(o
->out
, o
->in1
, 0x8000000000000000ull
);
2702 tcg_gen_mov_i64(o
->out2
, o
->in2
);
2706 static ExitStatus
op_nc(DisasContext
*s
, DisasOps
*o
)
2708 TCGv_i32 l
= tcg_const_i32(get_field(s
->fields
, l1
));
2709 potential_page_fault(s
);
2710 gen_helper_nc(cc_op
, cpu_env
, l
, o
->addr1
, o
->in2
);
2711 tcg_temp_free_i32(l
);
2716 static ExitStatus
op_neg(DisasContext
*s
, DisasOps
*o
)
2718 tcg_gen_neg_i64(o
->out
, o
->in2
);
2722 static ExitStatus
op_negf32(DisasContext
*s
, DisasOps
*o
)
2724 tcg_gen_xori_i64(o
->out
, o
->in2
, 0x80000000ull
);
2728 static ExitStatus
op_negf64(DisasContext
*s
, DisasOps
*o
)
2730 tcg_gen_xori_i64(o
->out
, o
->in2
, 0x8000000000000000ull
);
2734 static ExitStatus
op_negf128(DisasContext
*s
, DisasOps
*o
)
2736 tcg_gen_xori_i64(o
->out
, o
->in1
, 0x8000000000000000ull
);
2737 tcg_gen_mov_i64(o
->out2
, o
->in2
);
2741 static ExitStatus
op_oc(DisasContext
*s
, DisasOps
*o
)
2743 TCGv_i32 l
= tcg_const_i32(get_field(s
->fields
, l1
));
2744 potential_page_fault(s
);
2745 gen_helper_oc(cc_op
, cpu_env
, l
, o
->addr1
, o
->in2
);
2746 tcg_temp_free_i32(l
);
2751 static ExitStatus
op_or(DisasContext
*s
, DisasOps
*o
)
2753 tcg_gen_or_i64(o
->out
, o
->in1
, o
->in2
);
2757 static ExitStatus
op_ori(DisasContext
*s
, DisasOps
*o
)
2759 int shift
= s
->insn
->data
& 0xff;
2760 int size
= s
->insn
->data
>> 8;
2761 uint64_t mask
= ((1ull << size
) - 1) << shift
;
2764 tcg_gen_shli_i64(o
->in2
, o
->in2
, shift
);
2765 tcg_gen_or_i64(o
->out
, o
->in1
, o
->in2
);
2767 /* Produce the CC from only the bits manipulated. */
2768 tcg_gen_andi_i64(cc_dst
, o
->out
, mask
);
2769 set_cc_nz_u64(s
, cc_dst
);
2773 static ExitStatus
op_popcnt(DisasContext
*s
, DisasOps
*o
)
2775 gen_helper_popcnt(o
->out
, o
->in2
);
2779 #ifndef CONFIG_USER_ONLY
2780 static ExitStatus
op_ptlb(DisasContext
*s
, DisasOps
*o
)
2782 check_privileged(s
);
2783 gen_helper_ptlb(cpu_env
);
2788 static ExitStatus
op_risbg(DisasContext
*s
, DisasOps
*o
)
2790 int i3
= get_field(s
->fields
, i3
);
2791 int i4
= get_field(s
->fields
, i4
);
2792 int i5
= get_field(s
->fields
, i5
);
2793 int do_zero
= i4
& 0x80;
2794 uint64_t mask
, imask
, pmask
;
2797 /* Adjust the arguments for the specific insn. */
2798 switch (s
->fields
->op2
) {
2799 case 0x55: /* risbg */
2804 case 0x5d: /* risbhg */
2807 pmask
= 0xffffffff00000000ull
;
2809 case 0x51: /* risblg */
2812 pmask
= 0x00000000ffffffffull
;
2818 /* MASK is the set of bits to be inserted from R2.
2819 Take care for I3/I4 wraparound. */
2822 mask
^= pmask
>> i4
>> 1;
2824 mask
|= ~(pmask
>> i4
>> 1);
2828 /* IMASK is the set of bits to be kept from R1. In the case of the high/low
2829 insns, we need to keep the other half of the register. */
2830 imask
= ~mask
| ~pmask
;
2832 if (s
->fields
->op2
== 0x55) {
2839 /* In some cases we can implement this with deposit, which can be more
2840 efficient on some hosts. */
2841 if (~mask
== imask
&& i3
<= i4
) {
2842 if (s
->fields
->op2
== 0x5d) {
2845 /* Note that we rotate the bits to be inserted to the lsb, not to
2846 the position as described in the PoO. */
2849 rot
= (i5
- pos
) & 63;
2855 /* Rotate the input as necessary. */
2856 tcg_gen_rotli_i64(o
->in2
, o
->in2
, rot
);
2858 /* Insert the selected bits into the output. */
2860 tcg_gen_deposit_i64(o
->out
, o
->out
, o
->in2
, pos
, len
);
2861 } else if (imask
== 0) {
2862 tcg_gen_andi_i64(o
->out
, o
->in2
, mask
);
2864 tcg_gen_andi_i64(o
->in2
, o
->in2
, mask
);
2865 tcg_gen_andi_i64(o
->out
, o
->out
, imask
);
2866 tcg_gen_or_i64(o
->out
, o
->out
, o
->in2
);
2871 static ExitStatus
op_rosbg(DisasContext
*s
, DisasOps
*o
)
2873 int i3
= get_field(s
->fields
, i3
);
2874 int i4
= get_field(s
->fields
, i4
);
2875 int i5
= get_field(s
->fields
, i5
);
2878 /* If this is a test-only form, arrange to discard the result. */
2880 o
->out
= tcg_temp_new_i64();
2888 /* MASK is the set of bits to be operated on from R2.
2889 Take care for I3/I4 wraparound. */
2892 mask
^= ~0ull >> i4
>> 1;
2894 mask
|= ~(~0ull >> i4
>> 1);
2897 /* Rotate the input as necessary. */
2898 tcg_gen_rotli_i64(o
->in2
, o
->in2
, i5
);
2901 switch (s
->fields
->op2
) {
2902 case 0x55: /* AND */
2903 tcg_gen_ori_i64(o
->in2
, o
->in2
, ~mask
);
2904 tcg_gen_and_i64(o
->out
, o
->out
, o
->in2
);
2907 tcg_gen_andi_i64(o
->in2
, o
->in2
, mask
);
2908 tcg_gen_or_i64(o
->out
, o
->out
, o
->in2
);
2910 case 0x57: /* XOR */
2911 tcg_gen_andi_i64(o
->in2
, o
->in2
, mask
);
2912 tcg_gen_xor_i64(o
->out
, o
->out
, o
->in2
);
2919 tcg_gen_andi_i64(cc_dst
, o
->out
, mask
);
2920 set_cc_nz_u64(s
, cc_dst
);
2924 static ExitStatus
op_rev16(DisasContext
*s
, DisasOps
*o
)
2926 tcg_gen_bswap16_i64(o
->out
, o
->in2
);
2930 static ExitStatus
op_rev32(DisasContext
*s
, DisasOps
*o
)
2932 tcg_gen_bswap32_i64(o
->out
, o
->in2
);
2936 static ExitStatus
op_rev64(DisasContext
*s
, DisasOps
*o
)
2938 tcg_gen_bswap64_i64(o
->out
, o
->in2
);
2942 static ExitStatus
op_rll32(DisasContext
*s
, DisasOps
*o
)
2944 TCGv_i32 t1
= tcg_temp_new_i32();
2945 TCGv_i32 t2
= tcg_temp_new_i32();
2946 TCGv_i32 to
= tcg_temp_new_i32();
2947 tcg_gen_trunc_i64_i32(t1
, o
->in1
);
2948 tcg_gen_trunc_i64_i32(t2
, o
->in2
);
2949 tcg_gen_rotl_i32(to
, t1
, t2
);
2950 tcg_gen_extu_i32_i64(o
->out
, to
);
2951 tcg_temp_free_i32(t1
);
2952 tcg_temp_free_i32(t2
);
2953 tcg_temp_free_i32(to
);
2957 static ExitStatus
op_rll64(DisasContext
*s
, DisasOps
*o
)
2959 tcg_gen_rotl_i64(o
->out
, o
->in1
, o
->in2
);
2963 #ifndef CONFIG_USER_ONLY
2964 static ExitStatus
op_rrbe(DisasContext
*s
, DisasOps
*o
)
2966 check_privileged(s
);
2967 gen_helper_rrbe(cc_op
, cpu_env
, o
->in2
);
2972 static ExitStatus
op_sacf(DisasContext
*s
, DisasOps
*o
)
2974 check_privileged(s
);
2975 gen_helper_sacf(cpu_env
, o
->in2
);
2976 /* Addressing mode has changed, so end the block. */
2977 return EXIT_PC_STALE
;
2981 static ExitStatus
op_sam(DisasContext
*s
, DisasOps
*o
)
2983 int sam
= s
->insn
->data
;
2999 /* Bizarre but true, we check the address of the current insn for the
3000 specification exception, not the next to be executed. Thus the PoO
3001 documents that Bad Things Happen two bytes before the end. */
3002 if (s
->pc
& ~mask
) {
3003 gen_program_exception(s
, PGM_SPECIFICATION
);
3004 return EXIT_NORETURN
;
3008 tsam
= tcg_const_i64(sam
);
3009 tcg_gen_deposit_i64(psw_mask
, psw_mask
, tsam
, 31, 2);
3010 tcg_temp_free_i64(tsam
);
3012 /* Always exit the TB, since we (may have) changed execution mode. */
3013 return EXIT_PC_STALE
;
3016 static ExitStatus
op_sar(DisasContext
*s
, DisasOps
*o
)
3018 int r1
= get_field(s
->fields
, r1
);
3019 tcg_gen_st32_i64(o
->in2
, cpu_env
, offsetof(CPUS390XState
, aregs
[r1
]));
3023 static ExitStatus
op_seb(DisasContext
*s
, DisasOps
*o
)
3025 gen_helper_seb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
3029 static ExitStatus
op_sdb(DisasContext
*s
, DisasOps
*o
)
3031 gen_helper_sdb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
3035 static ExitStatus
op_sxb(DisasContext
*s
, DisasOps
*o
)
3037 gen_helper_sxb(o
->out
, cpu_env
, o
->out
, o
->out2
, o
->in1
, o
->in2
);
3038 return_low128(o
->out2
);
3042 static ExitStatus
op_sqeb(DisasContext
*s
, DisasOps
*o
)
3044 gen_helper_sqeb(o
->out
, cpu_env
, o
->in2
);
3048 static ExitStatus
op_sqdb(DisasContext
*s
, DisasOps
*o
)
3050 gen_helper_sqdb(o
->out
, cpu_env
, o
->in2
);
3054 static ExitStatus
op_sqxb(DisasContext
*s
, DisasOps
*o
)
3056 gen_helper_sqxb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
3057 return_low128(o
->out2
);
3061 #ifndef CONFIG_USER_ONLY
3062 static ExitStatus
op_servc(DisasContext
*s
, DisasOps
*o
)
3064 check_privileged(s
);
3065 potential_page_fault(s
);
3066 gen_helper_servc(cc_op
, cpu_env
, o
->in2
, o
->in1
);
3071 static ExitStatus
op_sigp(DisasContext
*s
, DisasOps
*o
)
3073 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
3074 check_privileged(s
);
3075 potential_page_fault(s
);
3076 gen_helper_sigp(cc_op
, cpu_env
, o
->in2
, r1
, o
->in1
);
3077 tcg_temp_free_i32(r1
);
3082 static ExitStatus
op_soc(DisasContext
*s
, DisasOps
*o
)
3089 disas_jcc(s
, &c
, get_field(s
->fields
, m3
));
3091 /* We want to store when the condition is fulfilled, so branch
3092 out when it's not */
3093 c
.cond
= tcg_invert_cond(c
.cond
);
3095 lab
= gen_new_label();
3097 tcg_gen_brcond_i64(c
.cond
, c
.u
.s64
.a
, c
.u
.s64
.b
, lab
);
3099 tcg_gen_brcond_i32(c
.cond
, c
.u
.s32
.a
, c
.u
.s32
.b
, lab
);
3103 r1
= get_field(s
->fields
, r1
);
3104 a
= get_address(s
, 0, get_field(s
->fields
, b2
), get_field(s
->fields
, d2
));
3105 if (s
->insn
->data
) {
3106 tcg_gen_qemu_st64(regs
[r1
], a
, get_mem_index(s
));
3108 tcg_gen_qemu_st32(regs
[r1
], a
, get_mem_index(s
));
3110 tcg_temp_free_i64(a
);
3116 static ExitStatus
op_sla(DisasContext
*s
, DisasOps
*o
)
3118 uint64_t sign
= 1ull << s
->insn
->data
;
3119 enum cc_op cco
= s
->insn
->data
== 31 ? CC_OP_SLA_32
: CC_OP_SLA_64
;
3120 gen_op_update2_cc_i64(s
, cco
, o
->in1
, o
->in2
);
3121 tcg_gen_shl_i64(o
->out
, o
->in1
, o
->in2
);
3122 /* The arithmetic left shift is curious in that it does not affect
3123 the sign bit. Copy that over from the source unchanged. */
3124 tcg_gen_andi_i64(o
->out
, o
->out
, ~sign
);
3125 tcg_gen_andi_i64(o
->in1
, o
->in1
, sign
);
3126 tcg_gen_or_i64(o
->out
, o
->out
, o
->in1
);
3130 static ExitStatus
op_sll(DisasContext
*s
, DisasOps
*o
)
3132 tcg_gen_shl_i64(o
->out
, o
->in1
, o
->in2
);
3136 static ExitStatus
op_sra(DisasContext
*s
, DisasOps
*o
)
3138 tcg_gen_sar_i64(o
->out
, o
->in1
, o
->in2
);
3142 static ExitStatus
op_srl(DisasContext
*s
, DisasOps
*o
)
3144 tcg_gen_shr_i64(o
->out
, o
->in1
, o
->in2
);
3148 static ExitStatus
op_sfpc(DisasContext
*s
, DisasOps
*o
)
3150 gen_helper_sfpc(cpu_env
, o
->in2
);
3154 static ExitStatus
op_sfas(DisasContext
*s
, DisasOps
*o
)
3156 gen_helper_sfas(cpu_env
, o
->in2
);
3160 static ExitStatus
op_srnm(DisasContext
*s
, DisasOps
*o
)
3162 int b2
= get_field(s
->fields
, b2
);
3163 int d2
= get_field(s
->fields
, d2
);
3164 TCGv_i64 t1
= tcg_temp_new_i64();
3165 TCGv_i64 t2
= tcg_temp_new_i64();
3168 switch (s
->fields
->op2
) {
3169 case 0x99: /* SRNM */
3172 case 0xb8: /* SRNMB */
3175 case 0xb9: /* SRNMT */
3181 mask
= (1 << len
) - 1;
3183 /* Insert the value into the appropriate field of the FPC. */
3185 tcg_gen_movi_i64(t1
, d2
& mask
);
3187 tcg_gen_addi_i64(t1
, regs
[b2
], d2
);
3188 tcg_gen_andi_i64(t1
, t1
, mask
);
3190 tcg_gen_ld32u_i64(t2
, cpu_env
, offsetof(CPUS390XState
, fpc
));
3191 tcg_gen_deposit_i64(t2
, t2
, t1
, pos
, len
);
3192 tcg_temp_free_i64(t1
);
3194 /* Then install the new FPC to set the rounding mode in fpu_status. */
3195 gen_helper_sfpc(cpu_env
, t2
);
3196 tcg_temp_free_i64(t2
);
3200 #ifndef CONFIG_USER_ONLY
3201 static ExitStatus
op_spka(DisasContext
*s
, DisasOps
*o
)
3203 check_privileged(s
);
3204 tcg_gen_shri_i64(o
->in2
, o
->in2
, 4);
3205 tcg_gen_deposit_i64(psw_mask
, psw_mask
, o
->in2
, PSW_SHIFT_KEY
- 4, 4);
3209 static ExitStatus
op_sske(DisasContext
*s
, DisasOps
*o
)
3211 check_privileged(s
);
3212 gen_helper_sske(cpu_env
, o
->in1
, o
->in2
);
3216 static ExitStatus
op_ssm(DisasContext
*s
, DisasOps
*o
)
3218 check_privileged(s
);
3219 tcg_gen_deposit_i64(psw_mask
, psw_mask
, o
->in2
, 56, 8);
3223 static ExitStatus
op_stap(DisasContext
*s
, DisasOps
*o
)
3225 check_privileged(s
);
3226 /* ??? Surely cpu address != cpu number. In any case the previous
3227 version of this stored more than the required half-word, so it
3228 is unlikely this has ever been tested. */
3229 tcg_gen_ld32u_i64(o
->out
, cpu_env
, offsetof(CPUS390XState
, cpu_num
));
3233 static ExitStatus
op_stck(DisasContext
*s
, DisasOps
*o
)
3235 gen_helper_stck(o
->out
, cpu_env
);
3236 /* ??? We don't implement clock states. */
3237 gen_op_movi_cc(s
, 0);
3241 static ExitStatus
op_stcke(DisasContext
*s
, DisasOps
*o
)
3243 TCGv_i64 c1
= tcg_temp_new_i64();
3244 TCGv_i64 c2
= tcg_temp_new_i64();
3245 gen_helper_stck(c1
, cpu_env
);
3246 /* Shift the 64-bit value into its place as a zero-extended
3247 104-bit value. Note that "bit positions 64-103 are always
3248 non-zero so that they compare differently to STCK"; we set
3249 the least significant bit to 1. */
3250 tcg_gen_shli_i64(c2
, c1
, 56);
3251 tcg_gen_shri_i64(c1
, c1
, 8);
3252 tcg_gen_ori_i64(c2
, c2
, 0x10000);
3253 tcg_gen_qemu_st64(c1
, o
->in2
, get_mem_index(s
));
3254 tcg_gen_addi_i64(o
->in2
, o
->in2
, 8);
3255 tcg_gen_qemu_st64(c2
, o
->in2
, get_mem_index(s
));
3256 tcg_temp_free_i64(c1
);
3257 tcg_temp_free_i64(c2
);
3258 /* ??? We don't implement clock states. */
3259 gen_op_movi_cc(s
, 0);
3263 static ExitStatus
op_sckc(DisasContext
*s
, DisasOps
*o
)
3265 check_privileged(s
);
3266 gen_helper_sckc(cpu_env
, o
->in2
);
3270 static ExitStatus
op_stckc(DisasContext
*s
, DisasOps
*o
)
3272 check_privileged(s
);
3273 gen_helper_stckc(o
->out
, cpu_env
);
3277 static ExitStatus
op_stctg(DisasContext
*s
, DisasOps
*o
)
3279 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
3280 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
3281 check_privileged(s
);
3282 potential_page_fault(s
);
3283 gen_helper_stctg(cpu_env
, r1
, o
->in2
, r3
);
3284 tcg_temp_free_i32(r1
);
3285 tcg_temp_free_i32(r3
);
3289 static ExitStatus
op_stctl(DisasContext
*s
, DisasOps
*o
)
3291 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
3292 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
3293 check_privileged(s
);
3294 potential_page_fault(s
);
3295 gen_helper_stctl(cpu_env
, r1
, o
->in2
, r3
);
3296 tcg_temp_free_i32(r1
);
3297 tcg_temp_free_i32(r3
);
3301 static ExitStatus
op_stidp(DisasContext
*s
, DisasOps
*o
)
3303 TCGv_i64 t1
= tcg_temp_new_i64();
3305 check_privileged(s
);
3306 tcg_gen_ld32u_i64(o
->out
, cpu_env
, offsetof(CPUS390XState
, cpu_num
));
3307 tcg_gen_ld32u_i64(t1
, cpu_env
, offsetof(CPUS390XState
, machine_type
));
3308 tcg_gen_deposit_i64(o
->out
, o
->out
, t1
, 32, 32);
3309 tcg_temp_free_i64(t1
);
3314 static ExitStatus
op_spt(DisasContext
*s
, DisasOps
*o
)
3316 check_privileged(s
);
3317 gen_helper_spt(cpu_env
, o
->in2
);
3321 static ExitStatus
op_stfl(DisasContext
*s
, DisasOps
*o
)
3324 /* We really ought to have more complete indication of facilities
3325 that we implement. Address this when STFLE is implemented. */
3326 check_privileged(s
);
3327 f
= tcg_const_i64(0xc0000000);
3328 a
= tcg_const_i64(200);
3329 tcg_gen_qemu_st32(f
, a
, get_mem_index(s
));
3330 tcg_temp_free_i64(f
);
3331 tcg_temp_free_i64(a
);
3335 static ExitStatus
op_stpt(DisasContext
*s
, DisasOps
*o
)
3337 check_privileged(s
);
3338 gen_helper_stpt(o
->out
, cpu_env
);
3342 static ExitStatus
op_stsi(DisasContext
*s
, DisasOps
*o
)
3344 check_privileged(s
);
3345 potential_page_fault(s
);
3346 gen_helper_stsi(cc_op
, cpu_env
, o
->in2
, regs
[0], regs
[1]);
3351 static ExitStatus
op_spx(DisasContext
*s
, DisasOps
*o
)
3353 check_privileged(s
);
3354 gen_helper_spx(cpu_env
, o
->in2
);
3358 static ExitStatus
op_subchannel(DisasContext
*s
, DisasOps
*o
)
3360 check_privileged(s
);
3361 /* Not operational. */
3362 gen_op_movi_cc(s
, 3);
3366 static ExitStatus
op_stpx(DisasContext
*s
, DisasOps
*o
)
3368 check_privileged(s
);
3369 tcg_gen_ld_i64(o
->out
, cpu_env
, offsetof(CPUS390XState
, psa
));
3370 tcg_gen_andi_i64(o
->out
, o
->out
, 0x7fffe000);
3374 static ExitStatus
op_stnosm(DisasContext
*s
, DisasOps
*o
)
3376 uint64_t i2
= get_field(s
->fields
, i2
);
3379 check_privileged(s
);
3381 /* It is important to do what the instruction name says: STORE THEN.
3382 If we let the output hook perform the store then if we fault and
3383 restart, we'll have the wrong SYSTEM MASK in place. */
3384 t
= tcg_temp_new_i64();
3385 tcg_gen_shri_i64(t
, psw_mask
, 56);
3386 tcg_gen_qemu_st8(t
, o
->addr1
, get_mem_index(s
));
3387 tcg_temp_free_i64(t
);
3389 if (s
->fields
->op
== 0xac) {
3390 tcg_gen_andi_i64(psw_mask
, psw_mask
,
3391 (i2
<< 56) | 0x00ffffffffffffffull
);
3393 tcg_gen_ori_i64(psw_mask
, psw_mask
, i2
<< 56);
3398 static ExitStatus
op_stura(DisasContext
*s
, DisasOps
*o
)
3400 check_privileged(s
);
3401 potential_page_fault(s
);
3402 gen_helper_stura(cpu_env
, o
->in2
, o
->in1
);
3406 static ExitStatus
op_sturg(DisasContext
*s
, DisasOps
*o
)
3408 check_privileged(s
);
3409 potential_page_fault(s
);
3410 gen_helper_sturg(cpu_env
, o
->in2
, o
->in1
);
3415 static ExitStatus
op_st8(DisasContext
*s
, DisasOps
*o
)
3417 tcg_gen_qemu_st8(o
->in1
, o
->in2
, get_mem_index(s
));
3421 static ExitStatus
op_st16(DisasContext
*s
, DisasOps
*o
)
3423 tcg_gen_qemu_st16(o
->in1
, o
->in2
, get_mem_index(s
));
3427 static ExitStatus
op_st32(DisasContext
*s
, DisasOps
*o
)
3429 tcg_gen_qemu_st32(o
->in1
, o
->in2
, get_mem_index(s
));
3433 static ExitStatus
op_st64(DisasContext
*s
, DisasOps
*o
)
3435 tcg_gen_qemu_st64(o
->in1
, o
->in2
, get_mem_index(s
));
3439 static ExitStatus
op_stam(DisasContext
*s
, DisasOps
*o
)
3441 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
3442 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
3443 potential_page_fault(s
);
3444 gen_helper_stam(cpu_env
, r1
, o
->in2
, r3
);
3445 tcg_temp_free_i32(r1
);
3446 tcg_temp_free_i32(r3
);
3450 static ExitStatus
op_stcm(DisasContext
*s
, DisasOps
*o
)
3452 int m3
= get_field(s
->fields
, m3
);
3453 int pos
, base
= s
->insn
->data
;
3454 TCGv_i64 tmp
= tcg_temp_new_i64();
3456 pos
= base
+ ctz32(m3
) * 8;
3459 /* Effectively a 32-bit store. */
3460 tcg_gen_shri_i64(tmp
, o
->in1
, pos
);
3461 tcg_gen_qemu_st32(tmp
, o
->in2
, get_mem_index(s
));
3467 /* Effectively a 16-bit store. */
3468 tcg_gen_shri_i64(tmp
, o
->in1
, pos
);
3469 tcg_gen_qemu_st16(tmp
, o
->in2
, get_mem_index(s
));
3476 /* Effectively an 8-bit store. */
3477 tcg_gen_shri_i64(tmp
, o
->in1
, pos
);
3478 tcg_gen_qemu_st8(tmp
, o
->in2
, get_mem_index(s
));
3482 /* This is going to be a sequence of shifts and stores. */
3483 pos
= base
+ 32 - 8;
3486 tcg_gen_shri_i64(tmp
, o
->in1
, pos
);
3487 tcg_gen_qemu_st8(tmp
, o
->in2
, get_mem_index(s
));
3488 tcg_gen_addi_i64(o
->in2
, o
->in2
, 1);
3490 m3
= (m3
<< 1) & 0xf;
3495 tcg_temp_free_i64(tmp
);
3499 static ExitStatus
op_stm(DisasContext
*s
, DisasOps
*o
)
3501 int r1
= get_field(s
->fields
, r1
);
3502 int r3
= get_field(s
->fields
, r3
);
3503 int size
= s
->insn
->data
;
3504 TCGv_i64 tsize
= tcg_const_i64(size
);
3508 tcg_gen_qemu_st64(regs
[r1
], o
->in2
, get_mem_index(s
));
3510 tcg_gen_qemu_st32(regs
[r1
], o
->in2
, get_mem_index(s
));
3515 tcg_gen_add_i64(o
->in2
, o
->in2
, tsize
);
3519 tcg_temp_free_i64(tsize
);
3523 static ExitStatus
op_stmh(DisasContext
*s
, DisasOps
*o
)
3525 int r1
= get_field(s
->fields
, r1
);
3526 int r3
= get_field(s
->fields
, r3
);
3527 TCGv_i64 t
= tcg_temp_new_i64();
3528 TCGv_i64 t4
= tcg_const_i64(4);
3529 TCGv_i64 t32
= tcg_const_i64(32);
3532 tcg_gen_shl_i64(t
, regs
[r1
], t32
);
3533 tcg_gen_qemu_st32(t
, o
->in2
, get_mem_index(s
));
3537 tcg_gen_add_i64(o
->in2
, o
->in2
, t4
);
3541 tcg_temp_free_i64(t
);
3542 tcg_temp_free_i64(t4
);
3543 tcg_temp_free_i64(t32
);
3547 static ExitStatus
op_srst(DisasContext
*s
, DisasOps
*o
)
3549 potential_page_fault(s
);
3550 gen_helper_srst(o
->in1
, cpu_env
, regs
[0], o
->in1
, o
->in2
);
3552 return_low128(o
->in2
);
3556 static ExitStatus
op_sub(DisasContext
*s
, DisasOps
*o
)
3558 tcg_gen_sub_i64(o
->out
, o
->in1
, o
->in2
);
3562 static ExitStatus
op_subb(DisasContext
*s
, DisasOps
*o
)
3567 tcg_gen_sub_i64(o
->out
, o
->in1
, o
->in2
);
3569 /* The !borrow flag is the msb of CC. Since we want the inverse of
3570 that, we ask for a comparison of CC=0 | CC=1 -> mask of 8 | 4. */
3571 disas_jcc(s
, &cmp
, 8 | 4);
3572 borrow
= tcg_temp_new_i64();
3574 tcg_gen_setcond_i64(cmp
.cond
, borrow
, cmp
.u
.s64
.a
, cmp
.u
.s64
.b
);
3576 TCGv_i32 t
= tcg_temp_new_i32();
3577 tcg_gen_setcond_i32(cmp
.cond
, t
, cmp
.u
.s32
.a
, cmp
.u
.s32
.b
);
3578 tcg_gen_extu_i32_i64(borrow
, t
);
3579 tcg_temp_free_i32(t
);
3583 tcg_gen_sub_i64(o
->out
, o
->out
, borrow
);
3584 tcg_temp_free_i64(borrow
);
3588 static ExitStatus
op_svc(DisasContext
*s
, DisasOps
*o
)
3595 t
= tcg_const_i32(get_field(s
->fields
, i1
) & 0xff);
3596 tcg_gen_st_i32(t
, cpu_env
, offsetof(CPUS390XState
, int_svc_code
));
3597 tcg_temp_free_i32(t
);
3599 t
= tcg_const_i32(s
->next_pc
- s
->pc
);
3600 tcg_gen_st_i32(t
, cpu_env
, offsetof(CPUS390XState
, int_svc_ilen
));
3601 tcg_temp_free_i32(t
);
3603 gen_exception(EXCP_SVC
);
3604 return EXIT_NORETURN
;
3607 static ExitStatus
op_tceb(DisasContext
*s
, DisasOps
*o
)
3609 gen_helper_tceb(cc_op
, o
->in1
, o
->in2
);
3614 static ExitStatus
op_tcdb(DisasContext
*s
, DisasOps
*o
)
3616 gen_helper_tcdb(cc_op
, o
->in1
, o
->in2
);
3621 static ExitStatus
op_tcxb(DisasContext
*s
, DisasOps
*o
)
3623 gen_helper_tcxb(cc_op
, o
->out
, o
->out2
, o
->in2
);
3628 #ifndef CONFIG_USER_ONLY
3629 static ExitStatus
op_tprot(DisasContext
*s
, DisasOps
*o
)
3631 potential_page_fault(s
);
3632 gen_helper_tprot(cc_op
, o
->addr1
, o
->in2
);
3638 static ExitStatus
op_tr(DisasContext
*s
, DisasOps
*o
)
3640 TCGv_i32 l
= tcg_const_i32(get_field(s
->fields
, l1
));
3641 potential_page_fault(s
);
3642 gen_helper_tr(cpu_env
, l
, o
->addr1
, o
->in2
);
3643 tcg_temp_free_i32(l
);
3648 static ExitStatus
op_unpk(DisasContext
*s
, DisasOps
*o
)
3650 TCGv_i32 l
= tcg_const_i32(get_field(s
->fields
, l1
));
3651 potential_page_fault(s
);
3652 gen_helper_unpk(cpu_env
, l
, o
->addr1
, o
->in2
);
3653 tcg_temp_free_i32(l
);
3657 static ExitStatus
op_xc(DisasContext
*s
, DisasOps
*o
)
3659 int d1
= get_field(s
->fields
, d1
);
3660 int d2
= get_field(s
->fields
, d2
);
3661 int b1
= get_field(s
->fields
, b1
);
3662 int b2
= get_field(s
->fields
, b2
);
3663 int l
= get_field(s
->fields
, l1
);
3666 o
->addr1
= get_address(s
, 0, b1
, d1
);
3668 /* If the addresses are identical, this is a store/memset of zero. */
3669 if (b1
== b2
&& d1
== d2
&& (l
+ 1) <= 32) {
3670 o
->in2
= tcg_const_i64(0);
3674 tcg_gen_qemu_st64(o
->in2
, o
->addr1
, get_mem_index(s
));
3677 tcg_gen_addi_i64(o
->addr1
, o
->addr1
, 8);
3681 tcg_gen_qemu_st32(o
->in2
, o
->addr1
, get_mem_index(s
));
3684 tcg_gen_addi_i64(o
->addr1
, o
->addr1
, 4);
3688 tcg_gen_qemu_st16(o
->in2
, o
->addr1
, get_mem_index(s
));
3691 tcg_gen_addi_i64(o
->addr1
, o
->addr1
, 2);
3695 tcg_gen_qemu_st8(o
->in2
, o
->addr1
, get_mem_index(s
));
3697 gen_op_movi_cc(s
, 0);
3701 /* But in general we'll defer to a helper. */
3702 o
->in2
= get_address(s
, 0, b2
, d2
);
3703 t32
= tcg_const_i32(l
);
3704 potential_page_fault(s
);
3705 gen_helper_xc(cc_op
, cpu_env
, t32
, o
->addr1
, o
->in2
);
3706 tcg_temp_free_i32(t32
);
3711 static ExitStatus
op_xor(DisasContext
*s
, DisasOps
*o
)
3713 tcg_gen_xor_i64(o
->out
, o
->in1
, o
->in2
);
3717 static ExitStatus
op_xori(DisasContext
*s
, DisasOps
*o
)
3719 int shift
= s
->insn
->data
& 0xff;
3720 int size
= s
->insn
->data
>> 8;
3721 uint64_t mask
= ((1ull << size
) - 1) << shift
;
3724 tcg_gen_shli_i64(o
->in2
, o
->in2
, shift
);
3725 tcg_gen_xor_i64(o
->out
, o
->in1
, o
->in2
);
3727 /* Produce the CC from only the bits manipulated. */
3728 tcg_gen_andi_i64(cc_dst
, o
->out
, mask
);
3729 set_cc_nz_u64(s
, cc_dst
);
3733 static ExitStatus
op_zero(DisasContext
*s
, DisasOps
*o
)
3735 o
->out
= tcg_const_i64(0);
3739 static ExitStatus
op_zero2(DisasContext
*s
, DisasOps
*o
)
3741 o
->out
= tcg_const_i64(0);
3747 /* ====================================================================== */
3748 /* The "Cc OUTput" generators. Given the generated output (and in some cases
3749 the original inputs), update the various cc data structures in order to
3750 be able to compute the new condition code. */
3752 static void cout_abs32(DisasContext
*s
, DisasOps
*o
)
3754 gen_op_update1_cc_i64(s
, CC_OP_ABS_32
, o
->out
);
3757 static void cout_abs64(DisasContext
*s
, DisasOps
*o
)
3759 gen_op_update1_cc_i64(s
, CC_OP_ABS_64
, o
->out
);
3762 static void cout_adds32(DisasContext
*s
, DisasOps
*o
)
3764 gen_op_update3_cc_i64(s
, CC_OP_ADD_32
, o
->in1
, o
->in2
, o
->out
);
3767 static void cout_adds64(DisasContext
*s
, DisasOps
*o
)
3769 gen_op_update3_cc_i64(s
, CC_OP_ADD_64
, o
->in1
, o
->in2
, o
->out
);
3772 static void cout_addu32(DisasContext
*s
, DisasOps
*o
)
3774 gen_op_update3_cc_i64(s
, CC_OP_ADDU_32
, o
->in1
, o
->in2
, o
->out
);
3777 static void cout_addu64(DisasContext
*s
, DisasOps
*o
)
3779 gen_op_update3_cc_i64(s
, CC_OP_ADDU_64
, o
->in1
, o
->in2
, o
->out
);
3782 static void cout_addc32(DisasContext
*s
, DisasOps
*o
)
3784 gen_op_update3_cc_i64(s
, CC_OP_ADDC_32
, o
->in1
, o
->in2
, o
->out
);
3787 static void cout_addc64(DisasContext
*s
, DisasOps
*o
)
3789 gen_op_update3_cc_i64(s
, CC_OP_ADDC_64
, o
->in1
, o
->in2
, o
->out
);
3792 static void cout_cmps32(DisasContext
*s
, DisasOps
*o
)
3794 gen_op_update2_cc_i64(s
, CC_OP_LTGT_32
, o
->in1
, o
->in2
);
3797 static void cout_cmps64(DisasContext
*s
, DisasOps
*o
)
3799 gen_op_update2_cc_i64(s
, CC_OP_LTGT_64
, o
->in1
, o
->in2
);
3802 static void cout_cmpu32(DisasContext
*s
, DisasOps
*o
)
3804 gen_op_update2_cc_i64(s
, CC_OP_LTUGTU_32
, o
->in1
, o
->in2
);
3807 static void cout_cmpu64(DisasContext
*s
, DisasOps
*o
)
3809 gen_op_update2_cc_i64(s
, CC_OP_LTUGTU_64
, o
->in1
, o
->in2
);
3812 static void cout_f32(DisasContext
*s
, DisasOps
*o
)
3814 gen_op_update1_cc_i64(s
, CC_OP_NZ_F32
, o
->out
);
3817 static void cout_f64(DisasContext
*s
, DisasOps
*o
)
3819 gen_op_update1_cc_i64(s
, CC_OP_NZ_F64
, o
->out
);
3822 static void cout_f128(DisasContext
*s
, DisasOps
*o
)
3824 gen_op_update2_cc_i64(s
, CC_OP_NZ_F128
, o
->out
, o
->out2
);
3827 static void cout_nabs32(DisasContext
*s
, DisasOps
*o
)
3829 gen_op_update1_cc_i64(s
, CC_OP_NABS_32
, o
->out
);
3832 static void cout_nabs64(DisasContext
*s
, DisasOps
*o
)
3834 gen_op_update1_cc_i64(s
, CC_OP_NABS_64
, o
->out
);
3837 static void cout_neg32(DisasContext
*s
, DisasOps
*o
)
3839 gen_op_update1_cc_i64(s
, CC_OP_COMP_32
, o
->out
);
3842 static void cout_neg64(DisasContext
*s
, DisasOps
*o
)
3844 gen_op_update1_cc_i64(s
, CC_OP_COMP_64
, o
->out
);
3847 static void cout_nz32(DisasContext
*s
, DisasOps
*o
)
3849 tcg_gen_ext32u_i64(cc_dst
, o
->out
);
3850 gen_op_update1_cc_i64(s
, CC_OP_NZ
, cc_dst
);
3853 static void cout_nz64(DisasContext
*s
, DisasOps
*o
)
3855 gen_op_update1_cc_i64(s
, CC_OP_NZ
, o
->out
);
3858 static void cout_s32(DisasContext
*s
, DisasOps
*o
)
3860 gen_op_update1_cc_i64(s
, CC_OP_LTGT0_32
, o
->out
);
3863 static void cout_s64(DisasContext
*s
, DisasOps
*o
)
3865 gen_op_update1_cc_i64(s
, CC_OP_LTGT0_64
, o
->out
);
3868 static void cout_subs32(DisasContext
*s
, DisasOps
*o
)
3870 gen_op_update3_cc_i64(s
, CC_OP_SUB_32
, o
->in1
, o
->in2
, o
->out
);
3873 static void cout_subs64(DisasContext
*s
, DisasOps
*o
)
3875 gen_op_update3_cc_i64(s
, CC_OP_SUB_64
, o
->in1
, o
->in2
, o
->out
);
3878 static void cout_subu32(DisasContext
*s
, DisasOps
*o
)
3880 gen_op_update3_cc_i64(s
, CC_OP_SUBU_32
, o
->in1
, o
->in2
, o
->out
);
3883 static void cout_subu64(DisasContext
*s
, DisasOps
*o
)
3885 gen_op_update3_cc_i64(s
, CC_OP_SUBU_64
, o
->in1
, o
->in2
, o
->out
);
3888 static void cout_subb32(DisasContext
*s
, DisasOps
*o
)
3890 gen_op_update3_cc_i64(s
, CC_OP_SUBB_32
, o
->in1
, o
->in2
, o
->out
);
3893 static void cout_subb64(DisasContext
*s
, DisasOps
*o
)
3895 gen_op_update3_cc_i64(s
, CC_OP_SUBB_64
, o
->in1
, o
->in2
, o
->out
);
3898 static void cout_tm32(DisasContext
*s
, DisasOps
*o
)
3900 gen_op_update2_cc_i64(s
, CC_OP_TM_32
, o
->in1
, o
->in2
);
3903 static void cout_tm64(DisasContext
*s
, DisasOps
*o
)
3905 gen_op_update2_cc_i64(s
, CC_OP_TM_64
, o
->in1
, o
->in2
);
3908 /* ====================================================================== */
3909 /* The "PREParation" generators. These initialize the DisasOps.OUT fields
3910 with the TCG register to which we will write. Used in combination with
3911 the "wout" generators, in some cases we need a new temporary, and in
3912 some cases we can write to a TCG global. */
3914 static void prep_new(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3916 o
->out
= tcg_temp_new_i64();
3918 #define SPEC_prep_new 0
3920 static void prep_new_P(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3922 o
->out
= tcg_temp_new_i64();
3923 o
->out2
= tcg_temp_new_i64();
3925 #define SPEC_prep_new_P 0
3927 static void prep_r1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3929 o
->out
= regs
[get_field(f
, r1
)];
3932 #define SPEC_prep_r1 0
3934 static void prep_r1_P(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3936 int r1
= get_field(f
, r1
);
3938 o
->out2
= regs
[r1
+ 1];
3939 o
->g_out
= o
->g_out2
= true;
3941 #define SPEC_prep_r1_P SPEC_r1_even
3943 static void prep_f1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3945 o
->out
= fregs
[get_field(f
, r1
)];
3948 #define SPEC_prep_f1 0
3950 static void prep_x1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3952 int r1
= get_field(f
, r1
);
3954 o
->out2
= fregs
[r1
+ 2];
3955 o
->g_out
= o
->g_out2
= true;
3957 #define SPEC_prep_x1 SPEC_r1_f128
3959 /* ====================================================================== */
3960 /* The "Write OUTput" generators. These generally perform some non-trivial
3961 copy of data to TCG globals, or to main memory. The trivial cases are
3962 generally handled by having a "prep" generator install the TCG global
3963 as the destination of the operation. */
3965 static void wout_r1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3967 store_reg(get_field(f
, r1
), o
->out
);
3969 #define SPEC_wout_r1 0
3971 static void wout_r1_8(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3973 int r1
= get_field(f
, r1
);
3974 tcg_gen_deposit_i64(regs
[r1
], regs
[r1
], o
->out
, 0, 8);
3976 #define SPEC_wout_r1_8 0
3978 static void wout_r1_16(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3980 int r1
= get_field(f
, r1
);
3981 tcg_gen_deposit_i64(regs
[r1
], regs
[r1
], o
->out
, 0, 16);
3983 #define SPEC_wout_r1_16 0
3985 static void wout_r1_32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3987 store_reg32_i64(get_field(f
, r1
), o
->out
);
3989 #define SPEC_wout_r1_32 0
3991 static void wout_r1_P32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3993 int r1
= get_field(f
, r1
);
3994 store_reg32_i64(r1
, o
->out
);
3995 store_reg32_i64(r1
+ 1, o
->out2
);
3997 #define SPEC_wout_r1_P32 SPEC_r1_even
3999 static void wout_r1_D32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4001 int r1
= get_field(f
, r1
);
4002 store_reg32_i64(r1
+ 1, o
->out
);
4003 tcg_gen_shri_i64(o
->out
, o
->out
, 32);
4004 store_reg32_i64(r1
, o
->out
);
4006 #define SPEC_wout_r1_D32 SPEC_r1_even
4008 static void wout_e1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4010 store_freg32_i64(get_field(f
, r1
), o
->out
);
4012 #define SPEC_wout_e1 0
4014 static void wout_f1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4016 store_freg(get_field(f
, r1
), o
->out
);
4018 #define SPEC_wout_f1 0
4020 static void wout_x1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4022 int f1
= get_field(s
->fields
, r1
);
4023 store_freg(f1
, o
->out
);
4024 store_freg(f1
+ 2, o
->out2
);
4026 #define SPEC_wout_x1 SPEC_r1_f128
4028 static void wout_cond_r1r2_32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4030 if (get_field(f
, r1
) != get_field(f
, r2
)) {
4031 store_reg32_i64(get_field(f
, r1
), o
->out
);
4034 #define SPEC_wout_cond_r1r2_32 0
4036 static void wout_cond_e1e2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4038 if (get_field(f
, r1
) != get_field(f
, r2
)) {
4039 store_freg32_i64(get_field(f
, r1
), o
->out
);
4042 #define SPEC_wout_cond_e1e2 0
4044 static void wout_m1_8(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4046 tcg_gen_qemu_st8(o
->out
, o
->addr1
, get_mem_index(s
));
4048 #define SPEC_wout_m1_8 0
4050 static void wout_m1_16(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4052 tcg_gen_qemu_st16(o
->out
, o
->addr1
, get_mem_index(s
));
4054 #define SPEC_wout_m1_16 0
4056 static void wout_m1_32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4058 tcg_gen_qemu_st32(o
->out
, o
->addr1
, get_mem_index(s
));
4060 #define SPEC_wout_m1_32 0
4062 static void wout_m1_64(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4064 tcg_gen_qemu_st64(o
->out
, o
->addr1
, get_mem_index(s
));
4066 #define SPEC_wout_m1_64 0
4068 static void wout_m2_32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4070 tcg_gen_qemu_st32(o
->out
, o
->in2
, get_mem_index(s
));
4072 #define SPEC_wout_m2_32 0
4074 static void wout_m2_32_r1_atomic(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4076 /* XXX release reservation */
4077 tcg_gen_qemu_st32(o
->out
, o
->addr1
, get_mem_index(s
));
4078 store_reg32_i64(get_field(f
, r1
), o
->in2
);
4080 #define SPEC_wout_m2_32_r1_atomic 0
4082 static void wout_m2_64_r1_atomic(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4084 /* XXX release reservation */
4085 tcg_gen_qemu_st64(o
->out
, o
->addr1
, get_mem_index(s
));
4086 store_reg(get_field(f
, r1
), o
->in2
);
4088 #define SPEC_wout_m2_64_r1_atomic 0
4090 /* ====================================================================== */
4091 /* The "INput 1" generators. These load the first operand to an insn. */
4093 static void in1_r1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4095 o
->in1
= load_reg(get_field(f
, r1
));
4097 #define SPEC_in1_r1 0
4099 static void in1_r1_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4101 o
->in1
= regs
[get_field(f
, r1
)];
4104 #define SPEC_in1_r1_o 0
4106 static void in1_r1_32s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4108 o
->in1
= tcg_temp_new_i64();
4109 tcg_gen_ext32s_i64(o
->in1
, regs
[get_field(f
, r1
)]);
4111 #define SPEC_in1_r1_32s 0
4113 static void in1_r1_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4115 o
->in1
= tcg_temp_new_i64();
4116 tcg_gen_ext32u_i64(o
->in1
, regs
[get_field(f
, r1
)]);
4118 #define SPEC_in1_r1_32u 0
4120 static void in1_r1_sr32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4122 o
->in1
= tcg_temp_new_i64();
4123 tcg_gen_shri_i64(o
->in1
, regs
[get_field(f
, r1
)], 32);
4125 #define SPEC_in1_r1_sr32 0
4127 static void in1_r1p1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4129 o
->in1
= load_reg(get_field(f
, r1
) + 1);
4131 #define SPEC_in1_r1p1 SPEC_r1_even
4133 static void in1_r1p1_32s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4135 o
->in1
= tcg_temp_new_i64();
4136 tcg_gen_ext32s_i64(o
->in1
, regs
[get_field(f
, r1
) + 1]);
4138 #define SPEC_in1_r1p1_32s SPEC_r1_even
4140 static void in1_r1p1_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4142 o
->in1
= tcg_temp_new_i64();
4143 tcg_gen_ext32u_i64(o
->in1
, regs
[get_field(f
, r1
) + 1]);
4145 #define SPEC_in1_r1p1_32u SPEC_r1_even
4147 static void in1_r1_D32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4149 int r1
= get_field(f
, r1
);
4150 o
->in1
= tcg_temp_new_i64();
4151 tcg_gen_concat32_i64(o
->in1
, regs
[r1
+ 1], regs
[r1
]);
4153 #define SPEC_in1_r1_D32 SPEC_r1_even
4155 static void in1_r2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4157 o
->in1
= load_reg(get_field(f
, r2
));
4159 #define SPEC_in1_r2 0
4161 static void in1_r3(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4163 o
->in1
= load_reg(get_field(f
, r3
));
4165 #define SPEC_in1_r3 0
4167 static void in1_r3_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4169 o
->in1
= regs
[get_field(f
, r3
)];
4172 #define SPEC_in1_r3_o 0
4174 static void in1_r3_32s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4176 o
->in1
= tcg_temp_new_i64();
4177 tcg_gen_ext32s_i64(o
->in1
, regs
[get_field(f
, r3
)]);
4179 #define SPEC_in1_r3_32s 0
4181 static void in1_r3_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4183 o
->in1
= tcg_temp_new_i64();
4184 tcg_gen_ext32u_i64(o
->in1
, regs
[get_field(f
, r3
)]);
4186 #define SPEC_in1_r3_32u 0
4188 static void in1_r3_D32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4190 int r3
= get_field(f
, r3
);
4191 o
->in1
= tcg_temp_new_i64();
4192 tcg_gen_concat32_i64(o
->in1
, regs
[r3
+ 1], regs
[r3
]);
4194 #define SPEC_in1_r3_D32 SPEC_r3_even
4196 static void in1_e1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4198 o
->in1
= load_freg32_i64(get_field(f
, r1
));
4200 #define SPEC_in1_e1 0
4202 static void in1_f1_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4204 o
->in1
= fregs
[get_field(f
, r1
)];
4207 #define SPEC_in1_f1_o 0
4209 static void in1_x1_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4211 int r1
= get_field(f
, r1
);
4213 o
->out2
= fregs
[r1
+ 2];
4214 o
->g_out
= o
->g_out2
= true;
4216 #define SPEC_in1_x1_o SPEC_r1_f128
4218 static void in1_f3_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4220 o
->in1
= fregs
[get_field(f
, r3
)];
4223 #define SPEC_in1_f3_o 0
4225 static void in1_la1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4227 o
->addr1
= get_address(s
, 0, get_field(f
, b1
), get_field(f
, d1
));
4229 #define SPEC_in1_la1 0
4231 static void in1_la2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4233 int x2
= have_field(f
, x2
) ? get_field(f
, x2
) : 0;
4234 o
->addr1
= get_address(s
, x2
, get_field(f
, b2
), get_field(f
, d2
));
4236 #define SPEC_in1_la2 0
4238 static void in1_m1_8u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4241 o
->in1
= tcg_temp_new_i64();
4242 tcg_gen_qemu_ld8u(o
->in1
, o
->addr1
, get_mem_index(s
));
4244 #define SPEC_in1_m1_8u 0
4246 static void in1_m1_16s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4249 o
->in1
= tcg_temp_new_i64();
4250 tcg_gen_qemu_ld16s(o
->in1
, o
->addr1
, get_mem_index(s
));
4252 #define SPEC_in1_m1_16s 0
4254 static void in1_m1_16u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4257 o
->in1
= tcg_temp_new_i64();
4258 tcg_gen_qemu_ld16u(o
->in1
, o
->addr1
, get_mem_index(s
));
4260 #define SPEC_in1_m1_16u 0
4262 static void in1_m1_32s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4265 o
->in1
= tcg_temp_new_i64();
4266 tcg_gen_qemu_ld32s(o
->in1
, o
->addr1
, get_mem_index(s
));
4268 #define SPEC_in1_m1_32s 0
4270 static void in1_m1_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4273 o
->in1
= tcg_temp_new_i64();
4274 tcg_gen_qemu_ld32u(o
->in1
, o
->addr1
, get_mem_index(s
));
4276 #define SPEC_in1_m1_32u 0
4278 static void in1_m1_64(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4281 o
->in1
= tcg_temp_new_i64();
4282 tcg_gen_qemu_ld64(o
->in1
, o
->addr1
, get_mem_index(s
));
4284 #define SPEC_in1_m1_64 0
4286 /* ====================================================================== */
4287 /* The "INput 2" generators. These load the second operand to an insn. */
4289 static void in2_r1_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4291 o
->in2
= regs
[get_field(f
, r1
)];
4294 #define SPEC_in2_r1_o 0
4296 static void in2_r1_16u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4298 o
->in2
= tcg_temp_new_i64();
4299 tcg_gen_ext16u_i64(o
->in2
, regs
[get_field(f
, r1
)]);
4301 #define SPEC_in2_r1_16u 0
4303 static void in2_r1_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4305 o
->in2
= tcg_temp_new_i64();
4306 tcg_gen_ext32u_i64(o
->in2
, regs
[get_field(f
, r1
)]);
4308 #define SPEC_in2_r1_32u 0
4310 static void in2_r1_D32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4312 int r1
= get_field(f
, r1
);
4313 o
->in2
= tcg_temp_new_i64();
4314 tcg_gen_concat32_i64(o
->in2
, regs
[r1
+ 1], regs
[r1
]);
4316 #define SPEC_in2_r1_D32 SPEC_r1_even
4318 static void in2_r2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4320 o
->in2
= load_reg(get_field(f
, r2
));
4322 #define SPEC_in2_r2 0
4324 static void in2_r2_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4326 o
->in2
= regs
[get_field(f
, r2
)];
4329 #define SPEC_in2_r2_o 0
4331 static void in2_r2_nz(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4333 int r2
= get_field(f
, r2
);
4335 o
->in2
= load_reg(r2
);
4338 #define SPEC_in2_r2_nz 0
4340 static void in2_r2_8s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4342 o
->in2
= tcg_temp_new_i64();
4343 tcg_gen_ext8s_i64(o
->in2
, regs
[get_field(f
, r2
)]);
4345 #define SPEC_in2_r2_8s 0
4347 static void in2_r2_8u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4349 o
->in2
= tcg_temp_new_i64();
4350 tcg_gen_ext8u_i64(o
->in2
, regs
[get_field(f
, r2
)]);
4352 #define SPEC_in2_r2_8u 0
4354 static void in2_r2_16s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4356 o
->in2
= tcg_temp_new_i64();
4357 tcg_gen_ext16s_i64(o
->in2
, regs
[get_field(f
, r2
)]);
4359 #define SPEC_in2_r2_16s 0
4361 static void in2_r2_16u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4363 o
->in2
= tcg_temp_new_i64();
4364 tcg_gen_ext16u_i64(o
->in2
, regs
[get_field(f
, r2
)]);
4366 #define SPEC_in2_r2_16u 0
4368 static void in2_r3(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4370 o
->in2
= load_reg(get_field(f
, r3
));
4372 #define SPEC_in2_r3 0
4374 static void in2_r2_32s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4376 o
->in2
= tcg_temp_new_i64();
4377 tcg_gen_ext32s_i64(o
->in2
, regs
[get_field(f
, r2
)]);
4379 #define SPEC_in2_r2_32s 0
4381 static void in2_r2_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4383 o
->in2
= tcg_temp_new_i64();
4384 tcg_gen_ext32u_i64(o
->in2
, regs
[get_field(f
, r2
)]);
4386 #define SPEC_in2_r2_32u 0
4388 static void in2_e2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4390 o
->in2
= load_freg32_i64(get_field(f
, r2
));
4392 #define SPEC_in2_e2 0
4394 static void in2_f2_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4396 o
->in2
= fregs
[get_field(f
, r2
)];
4399 #define SPEC_in2_f2_o 0
4401 static void in2_x2_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4403 int r2
= get_field(f
, r2
);
4405 o
->in2
= fregs
[r2
+ 2];
4406 o
->g_in1
= o
->g_in2
= true;
4408 #define SPEC_in2_x2_o SPEC_r2_f128
4410 static void in2_ra2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4412 o
->in2
= get_address(s
, 0, get_field(f
, r2
), 0);
4414 #define SPEC_in2_ra2 0
4416 static void in2_a2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4418 int x2
= have_field(f
, x2
) ? get_field(f
, x2
) : 0;
4419 o
->in2
= get_address(s
, x2
, get_field(f
, b2
), get_field(f
, d2
));
4421 #define SPEC_in2_a2 0
4423 static void in2_ri2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4425 o
->in2
= tcg_const_i64(s
->pc
+ (int64_t)get_field(f
, i2
) * 2);
4427 #define SPEC_in2_ri2 0
4429 static void in2_sh32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4431 help_l2_shift(s
, f
, o
, 31);
4433 #define SPEC_in2_sh32 0
4435 static void in2_sh64(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4437 help_l2_shift(s
, f
, o
, 63);
4439 #define SPEC_in2_sh64 0
4441 static void in2_m2_8u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4444 tcg_gen_qemu_ld8u(o
->in2
, o
->in2
, get_mem_index(s
));
4446 #define SPEC_in2_m2_8u 0
4448 static void in2_m2_16s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4451 tcg_gen_qemu_ld16s(o
->in2
, o
->in2
, get_mem_index(s
));
4453 #define SPEC_in2_m2_16s 0
4455 static void in2_m2_16u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4458 tcg_gen_qemu_ld16u(o
->in2
, o
->in2
, get_mem_index(s
));
4460 #define SPEC_in2_m2_16u 0
4462 static void in2_m2_32s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4465 tcg_gen_qemu_ld32s(o
->in2
, o
->in2
, get_mem_index(s
));
4467 #define SPEC_in2_m2_32s 0
4469 static void in2_m2_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4472 tcg_gen_qemu_ld32u(o
->in2
, o
->in2
, get_mem_index(s
));
4474 #define SPEC_in2_m2_32u 0
4476 static void in2_m2_64(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4479 tcg_gen_qemu_ld64(o
->in2
, o
->in2
, get_mem_index(s
));
4481 #define SPEC_in2_m2_64 0
4483 static void in2_mri2_16u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4486 tcg_gen_qemu_ld16u(o
->in2
, o
->in2
, get_mem_index(s
));
4488 #define SPEC_in2_mri2_16u 0
4490 static void in2_mri2_32s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4493 tcg_gen_qemu_ld32s(o
->in2
, o
->in2
, get_mem_index(s
));
4495 #define SPEC_in2_mri2_32s 0
4497 static void in2_mri2_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4500 tcg_gen_qemu_ld32u(o
->in2
, o
->in2
, get_mem_index(s
));
4502 #define SPEC_in2_mri2_32u 0
4504 static void in2_mri2_64(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4507 tcg_gen_qemu_ld64(o
->in2
, o
->in2
, get_mem_index(s
));
4509 #define SPEC_in2_mri2_64 0
4511 static void in2_m2_32s_atomic(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4513 /* XXX should reserve the address */
4515 o
->in2
= tcg_temp_new_i64();
4516 tcg_gen_qemu_ld32s(o
->in2
, o
->addr1
, get_mem_index(s
));
4518 #define SPEC_in2_m2_32s_atomic 0
4520 static void in2_m2_64_atomic(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4522 /* XXX should reserve the address */
4524 o
->in2
= tcg_temp_new_i64();
4525 tcg_gen_qemu_ld64(o
->in2
, o
->addr1
, get_mem_index(s
));
4527 #define SPEC_in2_m2_64_atomic 0
4529 static void in2_i2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4531 o
->in2
= tcg_const_i64(get_field(f
, i2
));
4533 #define SPEC_in2_i2 0
4535 static void in2_i2_8u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4537 o
->in2
= tcg_const_i64((uint8_t)get_field(f
, i2
));
4539 #define SPEC_in2_i2_8u 0
4541 static void in2_i2_16u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4543 o
->in2
= tcg_const_i64((uint16_t)get_field(f
, i2
));
4545 #define SPEC_in2_i2_16u 0
4547 static void in2_i2_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4549 o
->in2
= tcg_const_i64((uint32_t)get_field(f
, i2
));
4551 #define SPEC_in2_i2_32u 0
4553 static void in2_i2_16u_shl(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4555 uint64_t i2
= (uint16_t)get_field(f
, i2
);
4556 o
->in2
= tcg_const_i64(i2
<< s
->insn
->data
);
4558 #define SPEC_in2_i2_16u_shl 0
4560 static void in2_i2_32u_shl(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
4562 uint64_t i2
= (uint32_t)get_field(f
, i2
);
4563 o
->in2
= tcg_const_i64(i2
<< s
->insn
->data
);
4565 #define SPEC_in2_i2_32u_shl 0
4567 /* ====================================================================== */
4569 /* Find opc within the table of insns. This is formulated as a switch
4570 statement so that (1) we get compile-time notice of cut-paste errors
4571 for duplicated opcodes, and (2) the compiler generates the binary
4572 search tree, rather than us having to post-process the table. */
4574 #define C(OPC, NM, FT, FC, I1, I2, P, W, OP, CC) \
4575 D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, 0)
4577 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) insn_ ## NM,
4579 enum DisasInsnEnum
{
4580 #include "insn-data.def"
4584 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) { \
4588 .spec = SPEC_in1_##I1 | SPEC_in2_##I2 | SPEC_prep_##P | SPEC_wout_##W, \
4590 .help_in1 = in1_##I1, \
4591 .help_in2 = in2_##I2, \
4592 .help_prep = prep_##P, \
4593 .help_wout = wout_##W, \
4594 .help_cout = cout_##CC, \
4595 .help_op = op_##OP, \
4599 /* Allow 0 to be used for NULL in the table below. */
4607 #define SPEC_in1_0 0
4608 #define SPEC_in2_0 0
4609 #define SPEC_prep_0 0
4610 #define SPEC_wout_0 0
4612 static const DisasInsn insn_info
[] = {
4613 #include "insn-data.def"
4617 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) \
4618 case OPC: return &insn_info[insn_ ## NM];
4620 static const DisasInsn
*lookup_opc(uint16_t opc
)
4623 #include "insn-data.def"
4632 /* Extract a field from the insn. The INSN should be left-aligned in
4633 the uint64_t so that we can more easily utilize the big-bit-endian
4634 definitions we extract from the Principals of Operation. */
4636 static void extract_field(DisasFields
*o
, const DisasField
*f
, uint64_t insn
)
4644 /* Zero extract the field from the insn. */
4645 r
= (insn
<< f
->beg
) >> (64 - f
->size
);
4647 /* Sign-extend, or un-swap the field as necessary. */
4649 case 0: /* unsigned */
4651 case 1: /* signed */
4652 assert(f
->size
<= 32);
4653 m
= 1u << (f
->size
- 1);
4656 case 2: /* dl+dh split, signed 20 bit. */
4657 r
= ((int8_t)r
<< 12) | (r
>> 8);
4663 /* Validate that the "compressed" encoding we selected above is valid.
4664 I.e. we havn't make two different original fields overlap. */
4665 assert(((o
->presentC
>> f
->indexC
) & 1) == 0);
4666 o
->presentC
|= 1 << f
->indexC
;
4667 o
->presentO
|= 1 << f
->indexO
;
4669 o
->c
[f
->indexC
] = r
;
4672 /* Lookup the insn at the current PC, extracting the operands into O and
4673 returning the info struct for the insn. Returns NULL for invalid insn. */
4675 static const DisasInsn
*extract_insn(CPUS390XState
*env
, DisasContext
*s
,
4678 uint64_t insn
, pc
= s
->pc
;
4680 const DisasInsn
*info
;
4682 insn
= ld_code2(env
, pc
);
4683 op
= (insn
>> 8) & 0xff;
4684 ilen
= get_ilen(op
);
4685 s
->next_pc
= s
->pc
+ ilen
;
4692 insn
= ld_code4(env
, pc
) << 32;
4695 insn
= (insn
<< 48) | (ld_code4(env
, pc
+ 2) << 16);
4701 /* We can't actually determine the insn format until we've looked up
4702 the full insn opcode. Which we can't do without locating the
4703 secondary opcode. Assume by default that OP2 is at bit 40; for
4704 those smaller insns that don't actually have a secondary opcode
4705 this will correctly result in OP2 = 0. */
4711 case 0xb2: /* S, RRF, RRE */
4712 case 0xb3: /* RRE, RRD, RRF */
4713 case 0xb9: /* RRE, RRF */
4714 case 0xe5: /* SSE, SIL */
4715 op2
= (insn
<< 8) >> 56;
4719 case 0xc0: /* RIL */
4720 case 0xc2: /* RIL */
4721 case 0xc4: /* RIL */
4722 case 0xc6: /* RIL */
4723 case 0xc8: /* SSF */
4724 case 0xcc: /* RIL */
4725 op2
= (insn
<< 12) >> 60;
4727 case 0xd0 ... 0xdf: /* SS */
4733 case 0xee ... 0xf3: /* SS */
4734 case 0xf8 ... 0xfd: /* SS */
4738 op2
= (insn
<< 40) >> 56;
4742 memset(f
, 0, sizeof(*f
));
4746 /* Lookup the instruction. */
4747 info
= lookup_opc(op
<< 8 | op2
);
4749 /* If we found it, extract the operands. */
4751 DisasFormat fmt
= info
->fmt
;
4754 for (i
= 0; i
< NUM_C_FIELD
; ++i
) {
4755 extract_field(f
, &format_info
[fmt
].op
[i
], insn
);
4761 static ExitStatus
translate_one(CPUS390XState
*env
, DisasContext
*s
)
4763 const DisasInsn
*insn
;
4764 ExitStatus ret
= NO_EXIT
;
4768 /* Search for the insn in the table. */
4769 insn
= extract_insn(env
, s
, &f
);
4771 /* Not found means unimplemented/illegal opcode. */
4773 qemu_log_mask(LOG_UNIMP
, "unimplemented opcode 0x%02x%02x\n",
4775 gen_illegal_opcode(s
);
4776 return EXIT_NORETURN
;
4779 /* Check for insn specification exceptions. */
4781 int spec
= insn
->spec
, excp
= 0, r
;
4783 if (spec
& SPEC_r1_even
) {
4784 r
= get_field(&f
, r1
);
4786 excp
= PGM_SPECIFICATION
;
4789 if (spec
& SPEC_r2_even
) {
4790 r
= get_field(&f
, r2
);
4792 excp
= PGM_SPECIFICATION
;
4795 if (spec
& SPEC_r3_even
) {
4796 r
= get_field(&f
, r3
);
4798 excp
= PGM_SPECIFICATION
;
4801 if (spec
& SPEC_r1_f128
) {
4802 r
= get_field(&f
, r1
);
4804 excp
= PGM_SPECIFICATION
;
4807 if (spec
& SPEC_r2_f128
) {
4808 r
= get_field(&f
, r2
);
4810 excp
= PGM_SPECIFICATION
;
4814 gen_program_exception(s
, excp
);
4815 return EXIT_NORETURN
;
4819 /* Set up the strutures we use to communicate with the helpers. */
4822 o
.g_out
= o
.g_out2
= o
.g_in1
= o
.g_in2
= false;
4823 TCGV_UNUSED_I64(o
.out
);
4824 TCGV_UNUSED_I64(o
.out2
);
4825 TCGV_UNUSED_I64(o
.in1
);
4826 TCGV_UNUSED_I64(o
.in2
);
4827 TCGV_UNUSED_I64(o
.addr1
);
4829 /* Implement the instruction. */
4830 if (insn
->help_in1
) {
4831 insn
->help_in1(s
, &f
, &o
);
4833 if (insn
->help_in2
) {
4834 insn
->help_in2(s
, &f
, &o
);
4836 if (insn
->help_prep
) {
4837 insn
->help_prep(s
, &f
, &o
);
4839 if (insn
->help_op
) {
4840 ret
= insn
->help_op(s
, &o
);
4842 if (insn
->help_wout
) {
4843 insn
->help_wout(s
, &f
, &o
);
4845 if (insn
->help_cout
) {
4846 insn
->help_cout(s
, &o
);
4849 /* Free any temporaries created by the helpers. */
4850 if (!TCGV_IS_UNUSED_I64(o
.out
) && !o
.g_out
) {
4851 tcg_temp_free_i64(o
.out
);
4853 if (!TCGV_IS_UNUSED_I64(o
.out2
) && !o
.g_out2
) {
4854 tcg_temp_free_i64(o
.out2
);
4856 if (!TCGV_IS_UNUSED_I64(o
.in1
) && !o
.g_in1
) {
4857 tcg_temp_free_i64(o
.in1
);
4859 if (!TCGV_IS_UNUSED_I64(o
.in2
) && !o
.g_in2
) {
4860 tcg_temp_free_i64(o
.in2
);
4862 if (!TCGV_IS_UNUSED_I64(o
.addr1
)) {
4863 tcg_temp_free_i64(o
.addr1
);
4866 /* Advance to the next instruction. */
4871 static inline void gen_intermediate_code_internal(S390CPU
*cpu
,
4872 TranslationBlock
*tb
,
4875 CPUState
*cs
= CPU(cpu
);
4876 CPUS390XState
*env
= &cpu
->env
;
4878 target_ulong pc_start
;
4879 uint64_t next_page_start
;
4881 int num_insns
, max_insns
;
4889 if (!(tb
->flags
& FLAG_MASK_64
)) {
4890 pc_start
&= 0x7fffffff;
4895 dc
.cc_op
= CC_OP_DYNAMIC
;
4896 do_debug
= dc
.singlestep_enabled
= cs
->singlestep_enabled
;
4898 next_page_start
= (pc_start
& TARGET_PAGE_MASK
) + TARGET_PAGE_SIZE
;
4901 max_insns
= tb
->cflags
& CF_COUNT_MASK
;
4902 if (max_insns
== 0) {
4903 max_insns
= CF_COUNT_MASK
;
4910 j
= tcg_op_buf_count();
4914 tcg_ctx
.gen_opc_instr_start
[lj
++] = 0;
4917 tcg_ctx
.gen_opc_pc
[lj
] = dc
.pc
;
4918 gen_opc_cc_op
[lj
] = dc
.cc_op
;
4919 tcg_ctx
.gen_opc_instr_start
[lj
] = 1;
4920 tcg_ctx
.gen_opc_icount
[lj
] = num_insns
;
4922 if (++num_insns
== max_insns
&& (tb
->cflags
& CF_LAST_IO
)) {
4926 if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP
| CPU_LOG_TB_OP_OPT
))) {
4927 tcg_gen_debug_insn_start(dc
.pc
);
4931 if (unlikely(!QTAILQ_EMPTY(&cs
->breakpoints
))) {
4932 QTAILQ_FOREACH(bp
, &cs
->breakpoints
, entry
) {
4933 if (bp
->pc
== dc
.pc
) {
4934 status
= EXIT_PC_STALE
;
4940 if (status
== NO_EXIT
) {
4941 status
= translate_one(env
, &dc
);
4944 /* If we reach a page boundary, are single stepping,
4945 or exhaust instruction count, stop generation. */
4946 if (status
== NO_EXIT
4947 && (dc
.pc
>= next_page_start
4948 || tcg_op_buf_full()
4949 || num_insns
>= max_insns
4951 || cs
->singlestep_enabled
)) {
4952 status
= EXIT_PC_STALE
;
4954 } while (status
== NO_EXIT
);
4956 if (tb
->cflags
& CF_LAST_IO
) {
4965 update_psw_addr(&dc
);
4967 case EXIT_PC_UPDATED
:
4968 /* Next TB starts off with CC_OP_DYNAMIC, so make sure the
4969 cc op type is in env */
4971 /* Exit the TB, either by raising a debug exception or by return. */
4973 gen_exception(EXCP_DEBUG
);
4982 gen_tb_end(tb
, num_insns
);
4985 j
= tcg_op_buf_count();
4988 tcg_ctx
.gen_opc_instr_start
[lj
++] = 0;
4991 tb
->size
= dc
.pc
- pc_start
;
4992 tb
->icount
= num_insns
;
4995 #if defined(S390X_DEBUG_DISAS)
4996 if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM
)) {
4997 qemu_log("IN: %s\n", lookup_symbol(pc_start
));
4998 log_target_disas(env
, pc_start
, dc
.pc
- pc_start
, 1);
5004 void gen_intermediate_code (CPUS390XState
*env
, struct TranslationBlock
*tb
)
5006 gen_intermediate_code_internal(s390_env_get_cpu(env
), tb
, false);
5009 void gen_intermediate_code_pc (CPUS390XState
*env
, struct TranslationBlock
*tb
)
5011 gen_intermediate_code_internal(s390_env_get_cpu(env
), tb
, true);
5014 void restore_state_to_opc(CPUS390XState
*env
, TranslationBlock
*tb
, int pc_pos
)
5017 env
->psw
.addr
= tcg_ctx
.gen_opc_pc
[pc_pos
];
5018 cc_op
= gen_opc_cc_op
[pc_pos
];
5019 if ((cc_op
!= CC_OP_DYNAMIC
) && (cc_op
!= CC_OP_STATIC
)) {