4 * Copyright (c) 2003 Fabrice Bellard
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2 of the License, or (at your option) any later version.
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
26 #include "qemu/host-utils.h"
28 #include "disas/disas.h"
30 #include "exec/cpu_ldst.h"
32 #include "exec/helper-proto.h"
33 #include "exec/helper-gen.h"
35 #include "trace-tcg.h"
38 #define PREFIX_REPZ 0x01
39 #define PREFIX_REPNZ 0x02
40 #define PREFIX_LOCK 0x04
41 #define PREFIX_DATA 0x08
42 #define PREFIX_ADR 0x10
43 #define PREFIX_VEX 0x20
46 #define CODE64(s) ((s)->code64)
47 #define REX_X(s) ((s)->rex_x)
48 #define REX_B(s) ((s)->rex_b)
63 //#define MACRO_TEST 1
65 /* global register indexes */
66 static TCGv_ptr cpu_env
;
68 static TCGv cpu_cc_dst
, cpu_cc_src
, cpu_cc_src2
, cpu_cc_srcT
;
69 static TCGv_i32 cpu_cc_op
;
70 static TCGv cpu_regs
[CPU_NB_REGS
];
73 /* local register indexes (only used inside old micro ops) */
74 static TCGv cpu_tmp0
, cpu_tmp4
;
75 static TCGv_ptr cpu_ptr0
, cpu_ptr1
;
76 static TCGv_i32 cpu_tmp2_i32
, cpu_tmp3_i32
;
77 static TCGv_i64 cpu_tmp1_i64
;
79 static uint8_t gen_opc_cc_op
[OPC_BUF_SIZE
];
81 #include "exec/gen-icount.h"
84 static int x86_64_hregs
;
87 typedef struct DisasContext
{
88 /* current insn context */
89 int override
; /* -1 if no override */
93 target_ulong pc
; /* pc = eip + cs_base */
94 int is_jmp
; /* 1 = means jump (stop translation), 2 means CPU
95 static state change (stop translation) */
96 /* current block context */
97 target_ulong cs_base
; /* base of CS segment */
98 int pe
; /* protected mode */
99 int code32
; /* 32 bit code segment */
101 int lma
; /* long mode active */
102 int code64
; /* 64 bit code segment */
105 int vex_l
; /* vex vector length */
106 int vex_v
; /* vex vvvv register, without 1's compliment. */
107 int ss32
; /* 32 bit stack segment */
108 CCOp cc_op
; /* current CC operation */
110 int addseg
; /* non zero if either DS/ES/SS have a non zero base */
111 int f_st
; /* currently unused */
112 int vm86
; /* vm86 mode */
115 int tf
; /* TF cpu flag */
116 int singlestep_enabled
; /* "hardware" single step enabled */
117 int jmp_opt
; /* use direct block chaining for direct jumps */
118 int repz_opt
; /* optimize jumps within repz instructions */
119 int mem_index
; /* select memory access functions */
120 uint64_t flags
; /* all execution flags */
121 struct TranslationBlock
*tb
;
122 int popl_esp_hack
; /* for correct popl with esp base handling */
123 int rip_offset
; /* only used in x86_64, but left for simplicity */
125 int cpuid_ext_features
;
126 int cpuid_ext2_features
;
127 int cpuid_ext3_features
;
128 int cpuid_7_0_ebx_features
;
131 static void gen_eob(DisasContext
*s
);
132 static void gen_jmp(DisasContext
*s
, target_ulong eip
);
133 static void gen_jmp_tb(DisasContext
*s
, target_ulong eip
, int tb_num
);
134 static void gen_op(DisasContext
*s1
, int op
, TCGMemOp ot
, int d
);
136 /* i386 arith/logic operations */
156 OP_SHL1
, /* undocumented */
172 /* I386 int registers */
173 OR_EAX
, /* MUST be even numbered */
182 OR_TMP0
= 16, /* temporary operand register */
184 OR_A0
, /* temporary register used when doing address evaluation */
194 /* Bit set if the global variable is live after setting CC_OP to X. */
195 static const uint8_t cc_op_live
[CC_OP_NB
] = {
196 [CC_OP_DYNAMIC
] = USES_CC_DST
| USES_CC_SRC
| USES_CC_SRC2
,
197 [CC_OP_EFLAGS
] = USES_CC_SRC
,
198 [CC_OP_MULB
... CC_OP_MULQ
] = USES_CC_DST
| USES_CC_SRC
,
199 [CC_OP_ADDB
... CC_OP_ADDQ
] = USES_CC_DST
| USES_CC_SRC
,
200 [CC_OP_ADCB
... CC_OP_ADCQ
] = USES_CC_DST
| USES_CC_SRC
| USES_CC_SRC2
,
201 [CC_OP_SUBB
... CC_OP_SUBQ
] = USES_CC_DST
| USES_CC_SRC
| USES_CC_SRCT
,
202 [CC_OP_SBBB
... CC_OP_SBBQ
] = USES_CC_DST
| USES_CC_SRC
| USES_CC_SRC2
,
203 [CC_OP_LOGICB
... CC_OP_LOGICQ
] = USES_CC_DST
,
204 [CC_OP_INCB
... CC_OP_INCQ
] = USES_CC_DST
| USES_CC_SRC
,
205 [CC_OP_DECB
... CC_OP_DECQ
] = USES_CC_DST
| USES_CC_SRC
,
206 [CC_OP_SHLB
... CC_OP_SHLQ
] = USES_CC_DST
| USES_CC_SRC
,
207 [CC_OP_SARB
... CC_OP_SARQ
] = USES_CC_DST
| USES_CC_SRC
,
208 [CC_OP_BMILGB
... CC_OP_BMILGQ
] = USES_CC_DST
| USES_CC_SRC
,
209 [CC_OP_ADCX
] = USES_CC_DST
| USES_CC_SRC
,
210 [CC_OP_ADOX
] = USES_CC_SRC
| USES_CC_SRC2
,
211 [CC_OP_ADCOX
] = USES_CC_DST
| USES_CC_SRC
| USES_CC_SRC2
,
215 static void set_cc_op(DisasContext
*s
, CCOp op
)
219 if (s
->cc_op
== op
) {
223 /* Discard CC computation that will no longer be used. */
224 dead
= cc_op_live
[s
->cc_op
] & ~cc_op_live
[op
];
225 if (dead
& USES_CC_DST
) {
226 tcg_gen_discard_tl(cpu_cc_dst
);
228 if (dead
& USES_CC_SRC
) {
229 tcg_gen_discard_tl(cpu_cc_src
);
231 if (dead
& USES_CC_SRC2
) {
232 tcg_gen_discard_tl(cpu_cc_src2
);
234 if (dead
& USES_CC_SRCT
) {
235 tcg_gen_discard_tl(cpu_cc_srcT
);
238 if (op
== CC_OP_DYNAMIC
) {
239 /* The DYNAMIC setting is translator only, and should never be
240 stored. Thus we always consider it clean. */
241 s
->cc_op_dirty
= false;
243 /* Discard any computed CC_OP value (see shifts). */
244 if (s
->cc_op
== CC_OP_DYNAMIC
) {
245 tcg_gen_discard_i32(cpu_cc_op
);
247 s
->cc_op_dirty
= true;
252 static void gen_update_cc_op(DisasContext
*s
)
254 if (s
->cc_op_dirty
) {
255 tcg_gen_movi_i32(cpu_cc_op
, s
->cc_op
);
256 s
->cc_op_dirty
= false;
262 #define NB_OP_SIZES 4
264 #else /* !TARGET_X86_64 */
266 #define NB_OP_SIZES 3
268 #endif /* !TARGET_X86_64 */
270 #if defined(HOST_WORDS_BIGENDIAN)
271 #define REG_B_OFFSET (sizeof(target_ulong) - 1)
272 #define REG_H_OFFSET (sizeof(target_ulong) - 2)
273 #define REG_W_OFFSET (sizeof(target_ulong) - 2)
274 #define REG_L_OFFSET (sizeof(target_ulong) - 4)
275 #define REG_LH_OFFSET (sizeof(target_ulong) - 8)
277 #define REG_B_OFFSET 0
278 #define REG_H_OFFSET 1
279 #define REG_W_OFFSET 0
280 #define REG_L_OFFSET 0
281 #define REG_LH_OFFSET 4
284 /* In instruction encodings for byte register accesses the
285 * register number usually indicates "low 8 bits of register N";
286 * however there are some special cases where N 4..7 indicates
287 * [AH, CH, DH, BH], ie "bits 15..8 of register N-4". Return
288 * true for this special case, false otherwise.
290 static inline bool byte_reg_is_xH(int reg
)
296 if (reg
>= 8 || x86_64_hregs
) {
303 /* Select the size of a push/pop operation. */
304 static inline TCGMemOp
mo_pushpop(DisasContext
*s
, TCGMemOp ot
)
307 return ot
== MO_16
? MO_16
: MO_64
;
313 /* Select only size 64 else 32. Used for SSE operand sizes. */
314 static inline TCGMemOp
mo_64_32(TCGMemOp ot
)
317 return ot
== MO_64
? MO_64
: MO_32
;
323 /* Select size 8 if lsb of B is clear, else OT. Used for decoding
324 byte vs word opcodes. */
325 static inline TCGMemOp
mo_b_d(int b
, TCGMemOp ot
)
327 return b
& 1 ? ot
: MO_8
;
330 /* Select size 8 if lsb of B is clear, else OT capped at 32.
331 Used for decoding operand size of port opcodes. */
332 static inline TCGMemOp
mo_b_d32(int b
, TCGMemOp ot
)
334 return b
& 1 ? (ot
== MO_16
? MO_16
: MO_32
) : MO_8
;
337 static void gen_op_mov_reg_v(TCGMemOp ot
, int reg
, TCGv t0
)
341 if (!byte_reg_is_xH(reg
)) {
342 tcg_gen_deposit_tl(cpu_regs
[reg
], cpu_regs
[reg
], t0
, 0, 8);
344 tcg_gen_deposit_tl(cpu_regs
[reg
- 4], cpu_regs
[reg
- 4], t0
, 8, 8);
348 tcg_gen_deposit_tl(cpu_regs
[reg
], cpu_regs
[reg
], t0
, 0, 16);
351 /* For x86_64, this sets the higher half of register to zero.
352 For i386, this is equivalent to a mov. */
353 tcg_gen_ext32u_tl(cpu_regs
[reg
], t0
);
357 tcg_gen_mov_tl(cpu_regs
[reg
], t0
);
365 static inline void gen_op_mov_v_reg(TCGMemOp ot
, TCGv t0
, int reg
)
367 if (ot
== MO_8
&& byte_reg_is_xH(reg
)) {
368 tcg_gen_shri_tl(t0
, cpu_regs
[reg
- 4], 8);
369 tcg_gen_ext8u_tl(t0
, t0
);
371 tcg_gen_mov_tl(t0
, cpu_regs
[reg
]);
375 static inline void gen_op_movl_A0_reg(int reg
)
377 tcg_gen_mov_tl(cpu_A0
, cpu_regs
[reg
]);
380 static inline void gen_op_addl_A0_im(int32_t val
)
382 tcg_gen_addi_tl(cpu_A0
, cpu_A0
, val
);
384 tcg_gen_andi_tl(cpu_A0
, cpu_A0
, 0xffffffff);
389 static inline void gen_op_addq_A0_im(int64_t val
)
391 tcg_gen_addi_tl(cpu_A0
, cpu_A0
, val
);
395 static void gen_add_A0_im(DisasContext
*s
, int val
)
399 gen_op_addq_A0_im(val
);
402 gen_op_addl_A0_im(val
);
405 static inline void gen_op_jmp_v(TCGv dest
)
407 tcg_gen_st_tl(dest
, cpu_env
, offsetof(CPUX86State
, eip
));
410 static inline void gen_op_add_reg_im(TCGMemOp size
, int reg
, int32_t val
)
412 tcg_gen_addi_tl(cpu_tmp0
, cpu_regs
[reg
], val
);
413 gen_op_mov_reg_v(size
, reg
, cpu_tmp0
);
416 static inline void gen_op_add_reg_T0(TCGMemOp size
, int reg
)
418 tcg_gen_add_tl(cpu_tmp0
, cpu_regs
[reg
], cpu_T
[0]);
419 gen_op_mov_reg_v(size
, reg
, cpu_tmp0
);
422 static inline void gen_op_addl_A0_reg_sN(int shift
, int reg
)
424 tcg_gen_mov_tl(cpu_tmp0
, cpu_regs
[reg
]);
426 tcg_gen_shli_tl(cpu_tmp0
, cpu_tmp0
, shift
);
427 tcg_gen_add_tl(cpu_A0
, cpu_A0
, cpu_tmp0
);
428 /* For x86_64, this sets the higher half of register to zero.
429 For i386, this is equivalent to a nop. */
430 tcg_gen_ext32u_tl(cpu_A0
, cpu_A0
);
433 static inline void gen_op_movl_A0_seg(int reg
)
435 tcg_gen_ld32u_tl(cpu_A0
, cpu_env
, offsetof(CPUX86State
, segs
[reg
].base
) + REG_L_OFFSET
);
438 static inline void gen_op_addl_A0_seg(DisasContext
*s
, int reg
)
440 tcg_gen_ld_tl(cpu_tmp0
, cpu_env
, offsetof(CPUX86State
, segs
[reg
].base
));
443 tcg_gen_andi_tl(cpu_A0
, cpu_A0
, 0xffffffff);
444 tcg_gen_add_tl(cpu_A0
, cpu_A0
, cpu_tmp0
);
446 tcg_gen_add_tl(cpu_A0
, cpu_A0
, cpu_tmp0
);
447 tcg_gen_andi_tl(cpu_A0
, cpu_A0
, 0xffffffff);
450 tcg_gen_add_tl(cpu_A0
, cpu_A0
, cpu_tmp0
);
455 static inline void gen_op_movq_A0_seg(int reg
)
457 tcg_gen_ld_tl(cpu_A0
, cpu_env
, offsetof(CPUX86State
, segs
[reg
].base
));
460 static inline void gen_op_addq_A0_seg(int reg
)
462 tcg_gen_ld_tl(cpu_tmp0
, cpu_env
, offsetof(CPUX86State
, segs
[reg
].base
));
463 tcg_gen_add_tl(cpu_A0
, cpu_A0
, cpu_tmp0
);
466 static inline void gen_op_movq_A0_reg(int reg
)
468 tcg_gen_mov_tl(cpu_A0
, cpu_regs
[reg
]);
471 static inline void gen_op_addq_A0_reg_sN(int shift
, int reg
)
473 tcg_gen_mov_tl(cpu_tmp0
, cpu_regs
[reg
]);
475 tcg_gen_shli_tl(cpu_tmp0
, cpu_tmp0
, shift
);
476 tcg_gen_add_tl(cpu_A0
, cpu_A0
, cpu_tmp0
);
480 static inline void gen_op_ld_v(DisasContext
*s
, int idx
, TCGv t0
, TCGv a0
)
482 tcg_gen_qemu_ld_tl(t0
, a0
, s
->mem_index
, idx
| MO_LE
);
485 static inline void gen_op_st_v(DisasContext
*s
, int idx
, TCGv t0
, TCGv a0
)
487 tcg_gen_qemu_st_tl(t0
, a0
, s
->mem_index
, idx
| MO_LE
);
490 static inline void gen_op_st_rm_T0_A0(DisasContext
*s
, int idx
, int d
)
493 gen_op_st_v(s
, idx
, cpu_T
[0], cpu_A0
);
495 gen_op_mov_reg_v(idx
, d
, cpu_T
[0]);
499 static inline void gen_jmp_im(target_ulong pc
)
501 tcg_gen_movi_tl(cpu_tmp0
, pc
);
502 gen_op_jmp_v(cpu_tmp0
);
505 static inline void gen_string_movl_A0_ESI(DisasContext
*s
)
509 override
= s
->override
;
514 gen_op_movq_A0_seg(override
);
515 gen_op_addq_A0_reg_sN(0, R_ESI
);
517 gen_op_movq_A0_reg(R_ESI
);
523 if (s
->addseg
&& override
< 0)
526 gen_op_movl_A0_seg(override
);
527 gen_op_addl_A0_reg_sN(0, R_ESI
);
529 gen_op_movl_A0_reg(R_ESI
);
533 /* 16 address, always override */
536 tcg_gen_ext16u_tl(cpu_A0
, cpu_regs
[R_ESI
]);
537 gen_op_addl_A0_seg(s
, override
);
544 static inline void gen_string_movl_A0_EDI(DisasContext
*s
)
549 gen_op_movq_A0_reg(R_EDI
);
554 gen_op_movl_A0_seg(R_ES
);
555 gen_op_addl_A0_reg_sN(0, R_EDI
);
557 gen_op_movl_A0_reg(R_EDI
);
561 tcg_gen_ext16u_tl(cpu_A0
, cpu_regs
[R_EDI
]);
562 gen_op_addl_A0_seg(s
, R_ES
);
569 static inline void gen_op_movl_T0_Dshift(TCGMemOp ot
)
571 tcg_gen_ld32s_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
, df
));
572 tcg_gen_shli_tl(cpu_T
[0], cpu_T
[0], ot
);
575 static TCGv
gen_ext_tl(TCGv dst
, TCGv src
, TCGMemOp size
, bool sign
)
580 tcg_gen_ext8s_tl(dst
, src
);
582 tcg_gen_ext8u_tl(dst
, src
);
587 tcg_gen_ext16s_tl(dst
, src
);
589 tcg_gen_ext16u_tl(dst
, src
);
595 tcg_gen_ext32s_tl(dst
, src
);
597 tcg_gen_ext32u_tl(dst
, src
);
606 static void gen_extu(TCGMemOp ot
, TCGv reg
)
608 gen_ext_tl(reg
, reg
, ot
, false);
611 static void gen_exts(TCGMemOp ot
, TCGv reg
)
613 gen_ext_tl(reg
, reg
, ot
, true);
616 static inline void gen_op_jnz_ecx(TCGMemOp size
, int label1
)
618 tcg_gen_mov_tl(cpu_tmp0
, cpu_regs
[R_ECX
]);
619 gen_extu(size
, cpu_tmp0
);
620 tcg_gen_brcondi_tl(TCG_COND_NE
, cpu_tmp0
, 0, label1
);
623 static inline void gen_op_jz_ecx(TCGMemOp size
, int label1
)
625 tcg_gen_mov_tl(cpu_tmp0
, cpu_regs
[R_ECX
]);
626 gen_extu(size
, cpu_tmp0
);
627 tcg_gen_brcondi_tl(TCG_COND_EQ
, cpu_tmp0
, 0, label1
);
630 static void gen_helper_in_func(TCGMemOp ot
, TCGv v
, TCGv_i32 n
)
634 gen_helper_inb(v
, n
);
637 gen_helper_inw(v
, n
);
640 gen_helper_inl(v
, n
);
647 static void gen_helper_out_func(TCGMemOp ot
, TCGv_i32 v
, TCGv_i32 n
)
651 gen_helper_outb(v
, n
);
654 gen_helper_outw(v
, n
);
657 gen_helper_outl(v
, n
);
664 static void gen_check_io(DisasContext
*s
, TCGMemOp ot
, target_ulong cur_eip
,
668 target_ulong next_eip
;
671 if (s
->pe
&& (s
->cpl
> s
->iopl
|| s
->vm86
)) {
675 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
678 gen_helper_check_iob(cpu_env
, cpu_tmp2_i32
);
681 gen_helper_check_iow(cpu_env
, cpu_tmp2_i32
);
684 gen_helper_check_iol(cpu_env
, cpu_tmp2_i32
);
690 if(s
->flags
& HF_SVMI_MASK
) {
695 svm_flags
|= (1 << (4 + ot
));
696 next_eip
= s
->pc
- s
->cs_base
;
697 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
698 gen_helper_svm_check_io(cpu_env
, cpu_tmp2_i32
,
699 tcg_const_i32(svm_flags
),
700 tcg_const_i32(next_eip
- cur_eip
));
704 static inline void gen_movs(DisasContext
*s
, TCGMemOp ot
)
706 gen_string_movl_A0_ESI(s
);
707 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
708 gen_string_movl_A0_EDI(s
);
709 gen_op_st_v(s
, ot
, cpu_T
[0], cpu_A0
);
710 gen_op_movl_T0_Dshift(ot
);
711 gen_op_add_reg_T0(s
->aflag
, R_ESI
);
712 gen_op_add_reg_T0(s
->aflag
, R_EDI
);
715 static void gen_op_update1_cc(void)
717 tcg_gen_mov_tl(cpu_cc_dst
, cpu_T
[0]);
720 static void gen_op_update2_cc(void)
722 tcg_gen_mov_tl(cpu_cc_src
, cpu_T
[1]);
723 tcg_gen_mov_tl(cpu_cc_dst
, cpu_T
[0]);
726 static void gen_op_update3_cc(TCGv reg
)
728 tcg_gen_mov_tl(cpu_cc_src2
, reg
);
729 tcg_gen_mov_tl(cpu_cc_src
, cpu_T
[1]);
730 tcg_gen_mov_tl(cpu_cc_dst
, cpu_T
[0]);
733 static inline void gen_op_testl_T0_T1_cc(void)
735 tcg_gen_and_tl(cpu_cc_dst
, cpu_T
[0], cpu_T
[1]);
738 static void gen_op_update_neg_cc(void)
740 tcg_gen_mov_tl(cpu_cc_dst
, cpu_T
[0]);
741 tcg_gen_neg_tl(cpu_cc_src
, cpu_T
[0]);
742 tcg_gen_movi_tl(cpu_cc_srcT
, 0);
745 /* compute all eflags to cc_src */
746 static void gen_compute_eflags(DisasContext
*s
)
748 TCGv zero
, dst
, src1
, src2
;
751 if (s
->cc_op
== CC_OP_EFLAGS
) {
754 if (s
->cc_op
== CC_OP_CLR
) {
755 tcg_gen_movi_tl(cpu_cc_src
, CC_Z
| CC_P
);
756 set_cc_op(s
, CC_OP_EFLAGS
);
765 /* Take care to not read values that are not live. */
766 live
= cc_op_live
[s
->cc_op
] & ~USES_CC_SRCT
;
767 dead
= live
^ (USES_CC_DST
| USES_CC_SRC
| USES_CC_SRC2
);
769 zero
= tcg_const_tl(0);
770 if (dead
& USES_CC_DST
) {
773 if (dead
& USES_CC_SRC
) {
776 if (dead
& USES_CC_SRC2
) {
782 gen_helper_cc_compute_all(cpu_cc_src
, dst
, src1
, src2
, cpu_cc_op
);
783 set_cc_op(s
, CC_OP_EFLAGS
);
790 typedef struct CCPrepare
{
800 /* compute eflags.C to reg */
801 static CCPrepare
gen_prepare_eflags_c(DisasContext
*s
, TCGv reg
)
807 case CC_OP_SUBB
... CC_OP_SUBQ
:
808 /* (DATA_TYPE)CC_SRCT < (DATA_TYPE)CC_SRC */
809 size
= s
->cc_op
- CC_OP_SUBB
;
810 t1
= gen_ext_tl(cpu_tmp0
, cpu_cc_src
, size
, false);
811 /* If no temporary was used, be careful not to alias t1 and t0. */
812 t0
= TCGV_EQUAL(t1
, cpu_cc_src
) ? cpu_tmp0
: reg
;
813 tcg_gen_mov_tl(t0
, cpu_cc_srcT
);
817 case CC_OP_ADDB
... CC_OP_ADDQ
:
818 /* (DATA_TYPE)CC_DST < (DATA_TYPE)CC_SRC */
819 size
= s
->cc_op
- CC_OP_ADDB
;
820 t1
= gen_ext_tl(cpu_tmp0
, cpu_cc_src
, size
, false);
821 t0
= gen_ext_tl(reg
, cpu_cc_dst
, size
, false);
823 return (CCPrepare
) { .cond
= TCG_COND_LTU
, .reg
= t0
,
824 .reg2
= t1
, .mask
= -1, .use_reg2
= true };
826 case CC_OP_LOGICB
... CC_OP_LOGICQ
:
828 return (CCPrepare
) { .cond
= TCG_COND_NEVER
, .mask
= -1 };
830 case CC_OP_INCB
... CC_OP_INCQ
:
831 case CC_OP_DECB
... CC_OP_DECQ
:
832 return (CCPrepare
) { .cond
= TCG_COND_NE
, .reg
= cpu_cc_src
,
833 .mask
= -1, .no_setcond
= true };
835 case CC_OP_SHLB
... CC_OP_SHLQ
:
836 /* (CC_SRC >> (DATA_BITS - 1)) & 1 */
837 size
= s
->cc_op
- CC_OP_SHLB
;
838 shift
= (8 << size
) - 1;
839 return (CCPrepare
) { .cond
= TCG_COND_NE
, .reg
= cpu_cc_src
,
840 .mask
= (target_ulong
)1 << shift
};
842 case CC_OP_MULB
... CC_OP_MULQ
:
843 return (CCPrepare
) { .cond
= TCG_COND_NE
,
844 .reg
= cpu_cc_src
, .mask
= -1 };
846 case CC_OP_BMILGB
... CC_OP_BMILGQ
:
847 size
= s
->cc_op
- CC_OP_BMILGB
;
848 t0
= gen_ext_tl(reg
, cpu_cc_src
, size
, false);
849 return (CCPrepare
) { .cond
= TCG_COND_EQ
, .reg
= t0
, .mask
= -1 };
853 return (CCPrepare
) { .cond
= TCG_COND_NE
, .reg
= cpu_cc_dst
,
854 .mask
= -1, .no_setcond
= true };
857 case CC_OP_SARB
... CC_OP_SARQ
:
859 return (CCPrepare
) { .cond
= TCG_COND_NE
,
860 .reg
= cpu_cc_src
, .mask
= CC_C
};
863 /* The need to compute only C from CC_OP_DYNAMIC is important
864 in efficiently implementing e.g. INC at the start of a TB. */
866 gen_helper_cc_compute_c(reg
, cpu_cc_dst
, cpu_cc_src
,
867 cpu_cc_src2
, cpu_cc_op
);
868 return (CCPrepare
) { .cond
= TCG_COND_NE
, .reg
= reg
,
869 .mask
= -1, .no_setcond
= true };
873 /* compute eflags.P to reg */
874 static CCPrepare
gen_prepare_eflags_p(DisasContext
*s
, TCGv reg
)
876 gen_compute_eflags(s
);
877 return (CCPrepare
) { .cond
= TCG_COND_NE
, .reg
= cpu_cc_src
,
881 /* compute eflags.S to reg */
882 static CCPrepare
gen_prepare_eflags_s(DisasContext
*s
, TCGv reg
)
886 gen_compute_eflags(s
);
892 return (CCPrepare
) { .cond
= TCG_COND_NE
, .reg
= cpu_cc_src
,
895 return (CCPrepare
) { .cond
= TCG_COND_NEVER
, .mask
= -1 };
898 TCGMemOp size
= (s
->cc_op
- CC_OP_ADDB
) & 3;
899 TCGv t0
= gen_ext_tl(reg
, cpu_cc_dst
, size
, true);
900 return (CCPrepare
) { .cond
= TCG_COND_LT
, .reg
= t0
, .mask
= -1 };
905 /* compute eflags.O to reg */
906 static CCPrepare
gen_prepare_eflags_o(DisasContext
*s
, TCGv reg
)
911 return (CCPrepare
) { .cond
= TCG_COND_NE
, .reg
= cpu_cc_src2
,
912 .mask
= -1, .no_setcond
= true };
914 return (CCPrepare
) { .cond
= TCG_COND_NEVER
, .mask
= -1 };
916 gen_compute_eflags(s
);
917 return (CCPrepare
) { .cond
= TCG_COND_NE
, .reg
= cpu_cc_src
,
922 /* compute eflags.Z to reg */
923 static CCPrepare
gen_prepare_eflags_z(DisasContext
*s
, TCGv reg
)
927 gen_compute_eflags(s
);
933 return (CCPrepare
) { .cond
= TCG_COND_NE
, .reg
= cpu_cc_src
,
936 return (CCPrepare
) { .cond
= TCG_COND_ALWAYS
, .mask
= -1 };
939 TCGMemOp size
= (s
->cc_op
- CC_OP_ADDB
) & 3;
940 TCGv t0
= gen_ext_tl(reg
, cpu_cc_dst
, size
, false);
941 return (CCPrepare
) { .cond
= TCG_COND_EQ
, .reg
= t0
, .mask
= -1 };
946 /* perform a conditional store into register 'reg' according to jump opcode
947 value 'b'. In the fast case, T0 is guaranted not to be used. */
948 static CCPrepare
gen_prepare_cc(DisasContext
*s
, int b
, TCGv reg
)
950 int inv
, jcc_op
, cond
;
956 jcc_op
= (b
>> 1) & 7;
959 case CC_OP_SUBB
... CC_OP_SUBQ
:
960 /* We optimize relational operators for the cmp/jcc case. */
961 size
= s
->cc_op
- CC_OP_SUBB
;
964 tcg_gen_mov_tl(cpu_tmp4
, cpu_cc_srcT
);
965 gen_extu(size
, cpu_tmp4
);
966 t0
= gen_ext_tl(cpu_tmp0
, cpu_cc_src
, size
, false);
967 cc
= (CCPrepare
) { .cond
= TCG_COND_LEU
, .reg
= cpu_tmp4
,
968 .reg2
= t0
, .mask
= -1, .use_reg2
= true };
977 tcg_gen_mov_tl(cpu_tmp4
, cpu_cc_srcT
);
978 gen_exts(size
, cpu_tmp4
);
979 t0
= gen_ext_tl(cpu_tmp0
, cpu_cc_src
, size
, true);
980 cc
= (CCPrepare
) { .cond
= cond
, .reg
= cpu_tmp4
,
981 .reg2
= t0
, .mask
= -1, .use_reg2
= true };
991 /* This actually generates good code for JC, JZ and JS. */
994 cc
= gen_prepare_eflags_o(s
, reg
);
997 cc
= gen_prepare_eflags_c(s
, reg
);
1000 cc
= gen_prepare_eflags_z(s
, reg
);
1003 gen_compute_eflags(s
);
1004 cc
= (CCPrepare
) { .cond
= TCG_COND_NE
, .reg
= cpu_cc_src
,
1005 .mask
= CC_Z
| CC_C
};
1008 cc
= gen_prepare_eflags_s(s
, reg
);
1011 cc
= gen_prepare_eflags_p(s
, reg
);
1014 gen_compute_eflags(s
);
1015 if (TCGV_EQUAL(reg
, cpu_cc_src
)) {
1018 tcg_gen_shri_tl(reg
, cpu_cc_src
, 4); /* CC_O -> CC_S */
1019 tcg_gen_xor_tl(reg
, reg
, cpu_cc_src
);
1020 cc
= (CCPrepare
) { .cond
= TCG_COND_NE
, .reg
= reg
,
1025 gen_compute_eflags(s
);
1026 if (TCGV_EQUAL(reg
, cpu_cc_src
)) {
1029 tcg_gen_shri_tl(reg
, cpu_cc_src
, 4); /* CC_O -> CC_S */
1030 tcg_gen_xor_tl(reg
, reg
, cpu_cc_src
);
1031 cc
= (CCPrepare
) { .cond
= TCG_COND_NE
, .reg
= reg
,
1032 .mask
= CC_S
| CC_Z
};
1039 cc
.cond
= tcg_invert_cond(cc
.cond
);
1044 static void gen_setcc1(DisasContext
*s
, int b
, TCGv reg
)
1046 CCPrepare cc
= gen_prepare_cc(s
, b
, reg
);
1048 if (cc
.no_setcond
) {
1049 if (cc
.cond
== TCG_COND_EQ
) {
1050 tcg_gen_xori_tl(reg
, cc
.reg
, 1);
1052 tcg_gen_mov_tl(reg
, cc
.reg
);
1057 if (cc
.cond
== TCG_COND_NE
&& !cc
.use_reg2
&& cc
.imm
== 0 &&
1058 cc
.mask
!= 0 && (cc
.mask
& (cc
.mask
- 1)) == 0) {
1059 tcg_gen_shri_tl(reg
, cc
.reg
, ctztl(cc
.mask
));
1060 tcg_gen_andi_tl(reg
, reg
, 1);
1063 if (cc
.mask
!= -1) {
1064 tcg_gen_andi_tl(reg
, cc
.reg
, cc
.mask
);
1068 tcg_gen_setcond_tl(cc
.cond
, reg
, cc
.reg
, cc
.reg2
);
1070 tcg_gen_setcondi_tl(cc
.cond
, reg
, cc
.reg
, cc
.imm
);
1074 static inline void gen_compute_eflags_c(DisasContext
*s
, TCGv reg
)
1076 gen_setcc1(s
, JCC_B
<< 1, reg
);
1079 /* generate a conditional jump to label 'l1' according to jump opcode
1080 value 'b'. In the fast case, T0 is guaranted not to be used. */
1081 static inline void gen_jcc1_noeob(DisasContext
*s
, int b
, int l1
)
1083 CCPrepare cc
= gen_prepare_cc(s
, b
, cpu_T
[0]);
1085 if (cc
.mask
!= -1) {
1086 tcg_gen_andi_tl(cpu_T
[0], cc
.reg
, cc
.mask
);
1090 tcg_gen_brcond_tl(cc
.cond
, cc
.reg
, cc
.reg2
, l1
);
1092 tcg_gen_brcondi_tl(cc
.cond
, cc
.reg
, cc
.imm
, l1
);
1096 /* Generate a conditional jump to label 'l1' according to jump opcode
1097 value 'b'. In the fast case, T0 is guaranted not to be used.
1098 A translation block must end soon. */
1099 static inline void gen_jcc1(DisasContext
*s
, int b
, int l1
)
1101 CCPrepare cc
= gen_prepare_cc(s
, b
, cpu_T
[0]);
1103 gen_update_cc_op(s
);
1104 if (cc
.mask
!= -1) {
1105 tcg_gen_andi_tl(cpu_T
[0], cc
.reg
, cc
.mask
);
1108 set_cc_op(s
, CC_OP_DYNAMIC
);
1110 tcg_gen_brcond_tl(cc
.cond
, cc
.reg
, cc
.reg2
, l1
);
1112 tcg_gen_brcondi_tl(cc
.cond
, cc
.reg
, cc
.imm
, l1
);
1116 /* XXX: does not work with gdbstub "ice" single step - not a
1118 static int gen_jz_ecx_string(DisasContext
*s
, target_ulong next_eip
)
1122 l1
= gen_new_label();
1123 l2
= gen_new_label();
1124 gen_op_jnz_ecx(s
->aflag
, l1
);
1126 gen_jmp_tb(s
, next_eip
, 1);
1131 static inline void gen_stos(DisasContext
*s
, TCGMemOp ot
)
1133 gen_op_mov_v_reg(MO_32
, cpu_T
[0], R_EAX
);
1134 gen_string_movl_A0_EDI(s
);
1135 gen_op_st_v(s
, ot
, cpu_T
[0], cpu_A0
);
1136 gen_op_movl_T0_Dshift(ot
);
1137 gen_op_add_reg_T0(s
->aflag
, R_EDI
);
1140 static inline void gen_lods(DisasContext
*s
, TCGMemOp ot
)
1142 gen_string_movl_A0_ESI(s
);
1143 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
1144 gen_op_mov_reg_v(ot
, R_EAX
, cpu_T
[0]);
1145 gen_op_movl_T0_Dshift(ot
);
1146 gen_op_add_reg_T0(s
->aflag
, R_ESI
);
1149 static inline void gen_scas(DisasContext
*s
, TCGMemOp ot
)
1151 gen_string_movl_A0_EDI(s
);
1152 gen_op_ld_v(s
, ot
, cpu_T
[1], cpu_A0
);
1153 gen_op(s
, OP_CMPL
, ot
, R_EAX
);
1154 gen_op_movl_T0_Dshift(ot
);
1155 gen_op_add_reg_T0(s
->aflag
, R_EDI
);
1158 static inline void gen_cmps(DisasContext
*s
, TCGMemOp ot
)
1160 gen_string_movl_A0_EDI(s
);
1161 gen_op_ld_v(s
, ot
, cpu_T
[1], cpu_A0
);
1162 gen_string_movl_A0_ESI(s
);
1163 gen_op(s
, OP_CMPL
, ot
, OR_TMP0
);
1164 gen_op_movl_T0_Dshift(ot
);
1165 gen_op_add_reg_T0(s
->aflag
, R_ESI
);
1166 gen_op_add_reg_T0(s
->aflag
, R_EDI
);
1169 static inline void gen_ins(DisasContext
*s
, TCGMemOp ot
)
1171 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
1174 gen_string_movl_A0_EDI(s
);
1175 /* Note: we must do this dummy write first to be restartable in
1176 case of page fault. */
1177 tcg_gen_movi_tl(cpu_T
[0], 0);
1178 gen_op_st_v(s
, ot
, cpu_T
[0], cpu_A0
);
1179 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_regs
[R_EDX
]);
1180 tcg_gen_andi_i32(cpu_tmp2_i32
, cpu_tmp2_i32
, 0xffff);
1181 gen_helper_in_func(ot
, cpu_T
[0], cpu_tmp2_i32
);
1182 gen_op_st_v(s
, ot
, cpu_T
[0], cpu_A0
);
1183 gen_op_movl_T0_Dshift(ot
);
1184 gen_op_add_reg_T0(s
->aflag
, R_EDI
);
1185 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
1190 static inline void gen_outs(DisasContext
*s
, TCGMemOp ot
)
1192 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
1195 gen_string_movl_A0_ESI(s
);
1196 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
1198 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_regs
[R_EDX
]);
1199 tcg_gen_andi_i32(cpu_tmp2_i32
, cpu_tmp2_i32
, 0xffff);
1200 tcg_gen_trunc_tl_i32(cpu_tmp3_i32
, cpu_T
[0]);
1201 gen_helper_out_func(ot
, cpu_tmp2_i32
, cpu_tmp3_i32
);
1203 gen_op_movl_T0_Dshift(ot
);
1204 gen_op_add_reg_T0(s
->aflag
, R_ESI
);
1205 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
1210 /* same method as Valgrind : we generate jumps to current or next
1212 #define GEN_REPZ(op) \
1213 static inline void gen_repz_ ## op(DisasContext *s, TCGMemOp ot, \
1214 target_ulong cur_eip, target_ulong next_eip) \
1217 gen_update_cc_op(s); \
1218 l2 = gen_jz_ecx_string(s, next_eip); \
1219 gen_ ## op(s, ot); \
1220 gen_op_add_reg_im(s->aflag, R_ECX, -1); \
1221 /* a loop would cause two single step exceptions if ECX = 1 \
1222 before rep string_insn */ \
1224 gen_op_jz_ecx(s->aflag, l2); \
1225 gen_jmp(s, cur_eip); \
1228 #define GEN_REPZ2(op) \
1229 static inline void gen_repz_ ## op(DisasContext *s, TCGMemOp ot, \
1230 target_ulong cur_eip, \
1231 target_ulong next_eip, \
1235 gen_update_cc_op(s); \
1236 l2 = gen_jz_ecx_string(s, next_eip); \
1237 gen_ ## op(s, ot); \
1238 gen_op_add_reg_im(s->aflag, R_ECX, -1); \
1239 gen_update_cc_op(s); \
1240 gen_jcc1(s, (JCC_Z << 1) | (nz ^ 1), l2); \
1242 gen_op_jz_ecx(s->aflag, l2); \
1243 gen_jmp(s, cur_eip); \
1254 static void gen_helper_fp_arith_ST0_FT0(int op
)
1258 gen_helper_fadd_ST0_FT0(cpu_env
);
1261 gen_helper_fmul_ST0_FT0(cpu_env
);
1264 gen_helper_fcom_ST0_FT0(cpu_env
);
1267 gen_helper_fcom_ST0_FT0(cpu_env
);
1270 gen_helper_fsub_ST0_FT0(cpu_env
);
1273 gen_helper_fsubr_ST0_FT0(cpu_env
);
1276 gen_helper_fdiv_ST0_FT0(cpu_env
);
1279 gen_helper_fdivr_ST0_FT0(cpu_env
);
1284 /* NOTE the exception in "r" op ordering */
1285 static void gen_helper_fp_arith_STN_ST0(int op
, int opreg
)
1287 TCGv_i32 tmp
= tcg_const_i32(opreg
);
1290 gen_helper_fadd_STN_ST0(cpu_env
, tmp
);
1293 gen_helper_fmul_STN_ST0(cpu_env
, tmp
);
1296 gen_helper_fsubr_STN_ST0(cpu_env
, tmp
);
1299 gen_helper_fsub_STN_ST0(cpu_env
, tmp
);
1302 gen_helper_fdivr_STN_ST0(cpu_env
, tmp
);
1305 gen_helper_fdiv_STN_ST0(cpu_env
, tmp
);
1310 /* if d == OR_TMP0, it means memory operand (address in A0) */
1311 static void gen_op(DisasContext
*s1
, int op
, TCGMemOp ot
, int d
)
1314 gen_op_mov_v_reg(ot
, cpu_T
[0], d
);
1316 gen_op_ld_v(s1
, ot
, cpu_T
[0], cpu_A0
);
1320 gen_compute_eflags_c(s1
, cpu_tmp4
);
1321 tcg_gen_add_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
1322 tcg_gen_add_tl(cpu_T
[0], cpu_T
[0], cpu_tmp4
);
1323 gen_op_st_rm_T0_A0(s1
, ot
, d
);
1324 gen_op_update3_cc(cpu_tmp4
);
1325 set_cc_op(s1
, CC_OP_ADCB
+ ot
);
1328 gen_compute_eflags_c(s1
, cpu_tmp4
);
1329 tcg_gen_sub_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
1330 tcg_gen_sub_tl(cpu_T
[0], cpu_T
[0], cpu_tmp4
);
1331 gen_op_st_rm_T0_A0(s1
, ot
, d
);
1332 gen_op_update3_cc(cpu_tmp4
);
1333 set_cc_op(s1
, CC_OP_SBBB
+ ot
);
1336 tcg_gen_add_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
1337 gen_op_st_rm_T0_A0(s1
, ot
, d
);
1338 gen_op_update2_cc();
1339 set_cc_op(s1
, CC_OP_ADDB
+ ot
);
1342 tcg_gen_mov_tl(cpu_cc_srcT
, cpu_T
[0]);
1343 tcg_gen_sub_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
1344 gen_op_st_rm_T0_A0(s1
, ot
, d
);
1345 gen_op_update2_cc();
1346 set_cc_op(s1
, CC_OP_SUBB
+ ot
);
1350 tcg_gen_and_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
1351 gen_op_st_rm_T0_A0(s1
, ot
, d
);
1352 gen_op_update1_cc();
1353 set_cc_op(s1
, CC_OP_LOGICB
+ ot
);
1356 tcg_gen_or_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
1357 gen_op_st_rm_T0_A0(s1
, ot
, d
);
1358 gen_op_update1_cc();
1359 set_cc_op(s1
, CC_OP_LOGICB
+ ot
);
1362 tcg_gen_xor_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
1363 gen_op_st_rm_T0_A0(s1
, ot
, d
);
1364 gen_op_update1_cc();
1365 set_cc_op(s1
, CC_OP_LOGICB
+ ot
);
1368 tcg_gen_mov_tl(cpu_cc_src
, cpu_T
[1]);
1369 tcg_gen_mov_tl(cpu_cc_srcT
, cpu_T
[0]);
1370 tcg_gen_sub_tl(cpu_cc_dst
, cpu_T
[0], cpu_T
[1]);
1371 set_cc_op(s1
, CC_OP_SUBB
+ ot
);
1376 /* if d == OR_TMP0, it means memory operand (address in A0) */
1377 static void gen_inc(DisasContext
*s1
, TCGMemOp ot
, int d
, int c
)
1380 gen_op_mov_v_reg(ot
, cpu_T
[0], d
);
1382 gen_op_ld_v(s1
, ot
, cpu_T
[0], cpu_A0
);
1384 gen_compute_eflags_c(s1
, cpu_cc_src
);
1386 tcg_gen_addi_tl(cpu_T
[0], cpu_T
[0], 1);
1387 set_cc_op(s1
, CC_OP_INCB
+ ot
);
1389 tcg_gen_addi_tl(cpu_T
[0], cpu_T
[0], -1);
1390 set_cc_op(s1
, CC_OP_DECB
+ ot
);
1392 gen_op_st_rm_T0_A0(s1
, ot
, d
);
1393 tcg_gen_mov_tl(cpu_cc_dst
, cpu_T
[0]);
1396 static void gen_shift_flags(DisasContext
*s
, TCGMemOp ot
, TCGv result
,
1397 TCGv shm1
, TCGv count
, bool is_right
)
1399 TCGv_i32 z32
, s32
, oldop
;
1402 /* Store the results into the CC variables. If we know that the
1403 variable must be dead, store unconditionally. Otherwise we'll
1404 need to not disrupt the current contents. */
1405 z_tl
= tcg_const_tl(0);
1406 if (cc_op_live
[s
->cc_op
] & USES_CC_DST
) {
1407 tcg_gen_movcond_tl(TCG_COND_NE
, cpu_cc_dst
, count
, z_tl
,
1408 result
, cpu_cc_dst
);
1410 tcg_gen_mov_tl(cpu_cc_dst
, result
);
1412 if (cc_op_live
[s
->cc_op
] & USES_CC_SRC
) {
1413 tcg_gen_movcond_tl(TCG_COND_NE
, cpu_cc_src
, count
, z_tl
,
1416 tcg_gen_mov_tl(cpu_cc_src
, shm1
);
1418 tcg_temp_free(z_tl
);
1420 /* Get the two potential CC_OP values into temporaries. */
1421 tcg_gen_movi_i32(cpu_tmp2_i32
, (is_right
? CC_OP_SARB
: CC_OP_SHLB
) + ot
);
1422 if (s
->cc_op
== CC_OP_DYNAMIC
) {
1425 tcg_gen_movi_i32(cpu_tmp3_i32
, s
->cc_op
);
1426 oldop
= cpu_tmp3_i32
;
1429 /* Conditionally store the CC_OP value. */
1430 z32
= tcg_const_i32(0);
1431 s32
= tcg_temp_new_i32();
1432 tcg_gen_trunc_tl_i32(s32
, count
);
1433 tcg_gen_movcond_i32(TCG_COND_NE
, cpu_cc_op
, s32
, z32
, cpu_tmp2_i32
, oldop
);
1434 tcg_temp_free_i32(z32
);
1435 tcg_temp_free_i32(s32
);
1437 /* The CC_OP value is no longer predictable. */
1438 set_cc_op(s
, CC_OP_DYNAMIC
);
1441 static void gen_shift_rm_T1(DisasContext
*s
, TCGMemOp ot
, int op1
,
1442 int is_right
, int is_arith
)
1444 target_ulong mask
= (ot
== MO_64
? 0x3f : 0x1f);
1447 if (op1
== OR_TMP0
) {
1448 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
1450 gen_op_mov_v_reg(ot
, cpu_T
[0], op1
);
1453 tcg_gen_andi_tl(cpu_T
[1], cpu_T
[1], mask
);
1454 tcg_gen_subi_tl(cpu_tmp0
, cpu_T
[1], 1);
1458 gen_exts(ot
, cpu_T
[0]);
1459 tcg_gen_sar_tl(cpu_tmp0
, cpu_T
[0], cpu_tmp0
);
1460 tcg_gen_sar_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
1462 gen_extu(ot
, cpu_T
[0]);
1463 tcg_gen_shr_tl(cpu_tmp0
, cpu_T
[0], cpu_tmp0
);
1464 tcg_gen_shr_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
1467 tcg_gen_shl_tl(cpu_tmp0
, cpu_T
[0], cpu_tmp0
);
1468 tcg_gen_shl_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
1472 gen_op_st_rm_T0_A0(s
, ot
, op1
);
1474 gen_shift_flags(s
, ot
, cpu_T
[0], cpu_tmp0
, cpu_T
[1], is_right
);
1477 static void gen_shift_rm_im(DisasContext
*s
, TCGMemOp ot
, int op1
, int op2
,
1478 int is_right
, int is_arith
)
1480 int mask
= (ot
== MO_64
? 0x3f : 0x1f);
1484 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
1486 gen_op_mov_v_reg(ot
, cpu_T
[0], op1
);
1492 gen_exts(ot
, cpu_T
[0]);
1493 tcg_gen_sari_tl(cpu_tmp4
, cpu_T
[0], op2
- 1);
1494 tcg_gen_sari_tl(cpu_T
[0], cpu_T
[0], op2
);
1496 gen_extu(ot
, cpu_T
[0]);
1497 tcg_gen_shri_tl(cpu_tmp4
, cpu_T
[0], op2
- 1);
1498 tcg_gen_shri_tl(cpu_T
[0], cpu_T
[0], op2
);
1501 tcg_gen_shli_tl(cpu_tmp4
, cpu_T
[0], op2
- 1);
1502 tcg_gen_shli_tl(cpu_T
[0], cpu_T
[0], op2
);
1507 gen_op_st_rm_T0_A0(s
, ot
, op1
);
1509 /* update eflags if non zero shift */
1511 tcg_gen_mov_tl(cpu_cc_src
, cpu_tmp4
);
1512 tcg_gen_mov_tl(cpu_cc_dst
, cpu_T
[0]);
1513 set_cc_op(s
, (is_right
? CC_OP_SARB
: CC_OP_SHLB
) + ot
);
1517 static void gen_rot_rm_T1(DisasContext
*s
, TCGMemOp ot
, int op1
, int is_right
)
1519 target_ulong mask
= (ot
== MO_64
? 0x3f : 0x1f);
1523 if (op1
== OR_TMP0
) {
1524 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
1526 gen_op_mov_v_reg(ot
, cpu_T
[0], op1
);
1529 tcg_gen_andi_tl(cpu_T
[1], cpu_T
[1], mask
);
1533 /* Replicate the 8-bit input so that a 32-bit rotate works. */
1534 tcg_gen_ext8u_tl(cpu_T
[0], cpu_T
[0]);
1535 tcg_gen_muli_tl(cpu_T
[0], cpu_T
[0], 0x01010101);
1538 /* Replicate the 16-bit input so that a 32-bit rotate works. */
1539 tcg_gen_deposit_tl(cpu_T
[0], cpu_T
[0], cpu_T
[0], 16, 16);
1542 #ifdef TARGET_X86_64
1544 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
1545 tcg_gen_trunc_tl_i32(cpu_tmp3_i32
, cpu_T
[1]);
1547 tcg_gen_rotr_i32(cpu_tmp2_i32
, cpu_tmp2_i32
, cpu_tmp3_i32
);
1549 tcg_gen_rotl_i32(cpu_tmp2_i32
, cpu_tmp2_i32
, cpu_tmp3_i32
);
1551 tcg_gen_extu_i32_tl(cpu_T
[0], cpu_tmp2_i32
);
1556 tcg_gen_rotr_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
1558 tcg_gen_rotl_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
1564 gen_op_st_rm_T0_A0(s
, ot
, op1
);
1566 /* We'll need the flags computed into CC_SRC. */
1567 gen_compute_eflags(s
);
1569 /* The value that was "rotated out" is now present at the other end
1570 of the word. Compute C into CC_DST and O into CC_SRC2. Note that
1571 since we've computed the flags into CC_SRC, these variables are
1574 tcg_gen_shri_tl(cpu_cc_src2
, cpu_T
[0], mask
- 1);
1575 tcg_gen_shri_tl(cpu_cc_dst
, cpu_T
[0], mask
);
1576 tcg_gen_andi_tl(cpu_cc_dst
, cpu_cc_dst
, 1);
1578 tcg_gen_shri_tl(cpu_cc_src2
, cpu_T
[0], mask
);
1579 tcg_gen_andi_tl(cpu_cc_dst
, cpu_T
[0], 1);
1581 tcg_gen_andi_tl(cpu_cc_src2
, cpu_cc_src2
, 1);
1582 tcg_gen_xor_tl(cpu_cc_src2
, cpu_cc_src2
, cpu_cc_dst
);
1584 /* Now conditionally store the new CC_OP value. If the shift count
1585 is 0 we keep the CC_OP_EFLAGS setting so that only CC_SRC is live.
1586 Otherwise reuse CC_OP_ADCOX which have the C and O flags split out
1587 exactly as we computed above. */
1588 t0
= tcg_const_i32(0);
1589 t1
= tcg_temp_new_i32();
1590 tcg_gen_trunc_tl_i32(t1
, cpu_T
[1]);
1591 tcg_gen_movi_i32(cpu_tmp2_i32
, CC_OP_ADCOX
);
1592 tcg_gen_movi_i32(cpu_tmp3_i32
, CC_OP_EFLAGS
);
1593 tcg_gen_movcond_i32(TCG_COND_NE
, cpu_cc_op
, t1
, t0
,
1594 cpu_tmp2_i32
, cpu_tmp3_i32
);
1595 tcg_temp_free_i32(t0
);
1596 tcg_temp_free_i32(t1
);
1598 /* The CC_OP value is no longer predictable. */
1599 set_cc_op(s
, CC_OP_DYNAMIC
);
1602 static void gen_rot_rm_im(DisasContext
*s
, TCGMemOp ot
, int op1
, int op2
,
1605 int mask
= (ot
== MO_64
? 0x3f : 0x1f);
1609 if (op1
== OR_TMP0
) {
1610 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
1612 gen_op_mov_v_reg(ot
, cpu_T
[0], op1
);
1618 #ifdef TARGET_X86_64
1620 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
1622 tcg_gen_rotri_i32(cpu_tmp2_i32
, cpu_tmp2_i32
, op2
);
1624 tcg_gen_rotli_i32(cpu_tmp2_i32
, cpu_tmp2_i32
, op2
);
1626 tcg_gen_extu_i32_tl(cpu_T
[0], cpu_tmp2_i32
);
1631 tcg_gen_rotri_tl(cpu_T
[0], cpu_T
[0], op2
);
1633 tcg_gen_rotli_tl(cpu_T
[0], cpu_T
[0], op2
);
1644 shift
= mask
+ 1 - shift
;
1646 gen_extu(ot
, cpu_T
[0]);
1647 tcg_gen_shli_tl(cpu_tmp0
, cpu_T
[0], shift
);
1648 tcg_gen_shri_tl(cpu_T
[0], cpu_T
[0], mask
+ 1 - shift
);
1649 tcg_gen_or_tl(cpu_T
[0], cpu_T
[0], cpu_tmp0
);
1655 gen_op_st_rm_T0_A0(s
, ot
, op1
);
1658 /* Compute the flags into CC_SRC. */
1659 gen_compute_eflags(s
);
1661 /* The value that was "rotated out" is now present at the other end
1662 of the word. Compute C into CC_DST and O into CC_SRC2. Note that
1663 since we've computed the flags into CC_SRC, these variables are
1666 tcg_gen_shri_tl(cpu_cc_src2
, cpu_T
[0], mask
- 1);
1667 tcg_gen_shri_tl(cpu_cc_dst
, cpu_T
[0], mask
);
1668 tcg_gen_andi_tl(cpu_cc_dst
, cpu_cc_dst
, 1);
1670 tcg_gen_shri_tl(cpu_cc_src2
, cpu_T
[0], mask
);
1671 tcg_gen_andi_tl(cpu_cc_dst
, cpu_T
[0], 1);
1673 tcg_gen_andi_tl(cpu_cc_src2
, cpu_cc_src2
, 1);
1674 tcg_gen_xor_tl(cpu_cc_src2
, cpu_cc_src2
, cpu_cc_dst
);
1675 set_cc_op(s
, CC_OP_ADCOX
);
1679 /* XXX: add faster immediate = 1 case */
1680 static void gen_rotc_rm_T1(DisasContext
*s
, TCGMemOp ot
, int op1
,
1683 gen_compute_eflags(s
);
1684 assert(s
->cc_op
== CC_OP_EFLAGS
);
1688 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
1690 gen_op_mov_v_reg(ot
, cpu_T
[0], op1
);
1695 gen_helper_rcrb(cpu_T
[0], cpu_env
, cpu_T
[0], cpu_T
[1]);
1698 gen_helper_rcrw(cpu_T
[0], cpu_env
, cpu_T
[0], cpu_T
[1]);
1701 gen_helper_rcrl(cpu_T
[0], cpu_env
, cpu_T
[0], cpu_T
[1]);
1703 #ifdef TARGET_X86_64
1705 gen_helper_rcrq(cpu_T
[0], cpu_env
, cpu_T
[0], cpu_T
[1]);
1714 gen_helper_rclb(cpu_T
[0], cpu_env
, cpu_T
[0], cpu_T
[1]);
1717 gen_helper_rclw(cpu_T
[0], cpu_env
, cpu_T
[0], cpu_T
[1]);
1720 gen_helper_rcll(cpu_T
[0], cpu_env
, cpu_T
[0], cpu_T
[1]);
1722 #ifdef TARGET_X86_64
1724 gen_helper_rclq(cpu_T
[0], cpu_env
, cpu_T
[0], cpu_T
[1]);
1732 gen_op_st_rm_T0_A0(s
, ot
, op1
);
1735 /* XXX: add faster immediate case */
1736 static void gen_shiftd_rm_T1(DisasContext
*s
, TCGMemOp ot
, int op1
,
1737 bool is_right
, TCGv count_in
)
1739 target_ulong mask
= (ot
== MO_64
? 63 : 31);
1743 if (op1
== OR_TMP0
) {
1744 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
1746 gen_op_mov_v_reg(ot
, cpu_T
[0], op1
);
1749 count
= tcg_temp_new();
1750 tcg_gen_andi_tl(count
, count_in
, mask
);
1754 /* Note: we implement the Intel behaviour for shift count > 16.
1755 This means "shrdw C, B, A" shifts A:B:A >> C. Build the B:A
1756 portion by constructing it as a 32-bit value. */
1758 tcg_gen_deposit_tl(cpu_tmp0
, cpu_T
[0], cpu_T
[1], 16, 16);
1759 tcg_gen_mov_tl(cpu_T
[1], cpu_T
[0]);
1760 tcg_gen_mov_tl(cpu_T
[0], cpu_tmp0
);
1762 tcg_gen_deposit_tl(cpu_T
[1], cpu_T
[0], cpu_T
[1], 16, 16);
1765 #ifdef TARGET_X86_64
1767 /* Concatenate the two 32-bit values and use a 64-bit shift. */
1768 tcg_gen_subi_tl(cpu_tmp0
, count
, 1);
1770 tcg_gen_concat_tl_i64(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
1771 tcg_gen_shr_i64(cpu_tmp0
, cpu_T
[0], cpu_tmp0
);
1772 tcg_gen_shr_i64(cpu_T
[0], cpu_T
[0], count
);
1774 tcg_gen_concat_tl_i64(cpu_T
[0], cpu_T
[1], cpu_T
[0]);
1775 tcg_gen_shl_i64(cpu_tmp0
, cpu_T
[0], cpu_tmp0
);
1776 tcg_gen_shl_i64(cpu_T
[0], cpu_T
[0], count
);
1777 tcg_gen_shri_i64(cpu_tmp0
, cpu_tmp0
, 32);
1778 tcg_gen_shri_i64(cpu_T
[0], cpu_T
[0], 32);
1783 tcg_gen_subi_tl(cpu_tmp0
, count
, 1);
1785 tcg_gen_shr_tl(cpu_tmp0
, cpu_T
[0], cpu_tmp0
);
1787 tcg_gen_subfi_tl(cpu_tmp4
, mask
+ 1, count
);
1788 tcg_gen_shr_tl(cpu_T
[0], cpu_T
[0], count
);
1789 tcg_gen_shl_tl(cpu_T
[1], cpu_T
[1], cpu_tmp4
);
1791 tcg_gen_shl_tl(cpu_tmp0
, cpu_T
[0], cpu_tmp0
);
1793 /* Only needed if count > 16, for Intel behaviour. */
1794 tcg_gen_subfi_tl(cpu_tmp4
, 33, count
);
1795 tcg_gen_shr_tl(cpu_tmp4
, cpu_T
[1], cpu_tmp4
);
1796 tcg_gen_or_tl(cpu_tmp0
, cpu_tmp0
, cpu_tmp4
);
1799 tcg_gen_subfi_tl(cpu_tmp4
, mask
+ 1, count
);
1800 tcg_gen_shl_tl(cpu_T
[0], cpu_T
[0], count
);
1801 tcg_gen_shr_tl(cpu_T
[1], cpu_T
[1], cpu_tmp4
);
1803 tcg_gen_movi_tl(cpu_tmp4
, 0);
1804 tcg_gen_movcond_tl(TCG_COND_EQ
, cpu_T
[1], count
, cpu_tmp4
,
1805 cpu_tmp4
, cpu_T
[1]);
1806 tcg_gen_or_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
1811 gen_op_st_rm_T0_A0(s
, ot
, op1
);
1813 gen_shift_flags(s
, ot
, cpu_T
[0], cpu_tmp0
, count
, is_right
);
1814 tcg_temp_free(count
);
1817 static void gen_shift(DisasContext
*s1
, int op
, TCGMemOp ot
, int d
, int s
)
1820 gen_op_mov_v_reg(ot
, cpu_T
[1], s
);
1823 gen_rot_rm_T1(s1
, ot
, d
, 0);
1826 gen_rot_rm_T1(s1
, ot
, d
, 1);
1830 gen_shift_rm_T1(s1
, ot
, d
, 0, 0);
1833 gen_shift_rm_T1(s1
, ot
, d
, 1, 0);
1836 gen_shift_rm_T1(s1
, ot
, d
, 1, 1);
1839 gen_rotc_rm_T1(s1
, ot
, d
, 0);
1842 gen_rotc_rm_T1(s1
, ot
, d
, 1);
1847 static void gen_shifti(DisasContext
*s1
, int op
, TCGMemOp ot
, int d
, int c
)
1851 gen_rot_rm_im(s1
, ot
, d
, c
, 0);
1854 gen_rot_rm_im(s1
, ot
, d
, c
, 1);
1858 gen_shift_rm_im(s1
, ot
, d
, c
, 0, 0);
1861 gen_shift_rm_im(s1
, ot
, d
, c
, 1, 0);
1864 gen_shift_rm_im(s1
, ot
, d
, c
, 1, 1);
1867 /* currently not optimized */
1868 tcg_gen_movi_tl(cpu_T
[1], c
);
1869 gen_shift(s1
, op
, ot
, d
, OR_TMP1
);
1874 static void gen_lea_modrm(CPUX86State
*env
, DisasContext
*s
, int modrm
)
1881 int mod
, rm
, code
, override
, must_add_seg
;
1884 override
= s
->override
;
1885 must_add_seg
= s
->addseg
;
1888 mod
= (modrm
>> 6) & 3;
1901 code
= cpu_ldub_code(env
, s
->pc
++);
1902 scale
= (code
>> 6) & 3;
1903 index
= ((code
>> 3) & 7) | REX_X(s
);
1905 index
= -1; /* no index */
1913 if ((base
& 7) == 5) {
1915 disp
= (int32_t)cpu_ldl_code(env
, s
->pc
);
1917 if (CODE64(s
) && !havesib
) {
1918 disp
+= s
->pc
+ s
->rip_offset
;
1925 disp
= (int8_t)cpu_ldub_code(env
, s
->pc
++);
1929 disp
= (int32_t)cpu_ldl_code(env
, s
->pc
);
1934 /* For correct popl handling with esp. */
1935 if (base
== R_ESP
&& s
->popl_esp_hack
) {
1936 disp
+= s
->popl_esp_hack
;
1939 /* Compute the address, with a minimum number of TCG ops. */
1943 sum
= cpu_regs
[index
];
1945 tcg_gen_shli_tl(cpu_A0
, cpu_regs
[index
], scale
);
1949 tcg_gen_add_tl(cpu_A0
, sum
, cpu_regs
[base
]);
1952 } else if (base
>= 0) {
1953 sum
= cpu_regs
[base
];
1955 if (TCGV_IS_UNUSED(sum
)) {
1956 tcg_gen_movi_tl(cpu_A0
, disp
);
1958 tcg_gen_addi_tl(cpu_A0
, sum
, disp
);
1963 if (base
== R_EBP
|| base
== R_ESP
) {
1970 tcg_gen_ld_tl(cpu_tmp0
, cpu_env
,
1971 offsetof(CPUX86State
, segs
[override
].base
));
1973 if (s
->aflag
== MO_32
) {
1974 tcg_gen_ext32u_tl(cpu_A0
, cpu_A0
);
1976 tcg_gen_add_tl(cpu_A0
, cpu_A0
, cpu_tmp0
);
1980 tcg_gen_add_tl(cpu_A0
, cpu_A0
, cpu_tmp0
);
1983 if (s
->aflag
== MO_32
) {
1984 tcg_gen_ext32u_tl(cpu_A0
, cpu_A0
);
1992 disp
= cpu_lduw_code(env
, s
->pc
);
1994 tcg_gen_movi_tl(cpu_A0
, disp
);
1995 rm
= 0; /* avoid SS override */
2002 disp
= (int8_t)cpu_ldub_code(env
, s
->pc
++);
2006 disp
= (int16_t)cpu_lduw_code(env
, s
->pc
);
2014 tcg_gen_add_tl(cpu_A0
, cpu_regs
[R_EBX
], cpu_regs
[R_ESI
]);
2017 tcg_gen_add_tl(cpu_A0
, cpu_regs
[R_EBX
], cpu_regs
[R_EDI
]);
2020 tcg_gen_add_tl(cpu_A0
, cpu_regs
[R_EBP
], cpu_regs
[R_ESI
]);
2023 tcg_gen_add_tl(cpu_A0
, cpu_regs
[R_EBP
], cpu_regs
[R_EDI
]);
2026 sum
= cpu_regs
[R_ESI
];
2029 sum
= cpu_regs
[R_EDI
];
2032 sum
= cpu_regs
[R_EBP
];
2036 sum
= cpu_regs
[R_EBX
];
2039 tcg_gen_addi_tl(cpu_A0
, sum
, disp
);
2040 tcg_gen_ext16u_tl(cpu_A0
, cpu_A0
);
2044 if (rm
== 2 || rm
== 3 || rm
== 6) {
2050 gen_op_addl_A0_seg(s
, override
);
2059 static void gen_nop_modrm(CPUX86State
*env
, DisasContext
*s
, int modrm
)
2061 int mod
, rm
, base
, code
;
2063 mod
= (modrm
>> 6) & 3;
2074 code
= cpu_ldub_code(env
, s
->pc
++);
2116 /* used for LEA and MOV AX, mem */
2117 static void gen_add_A0_ds_seg(DisasContext
*s
)
2119 int override
, must_add_seg
;
2120 must_add_seg
= s
->addseg
;
2122 if (s
->override
>= 0) {
2123 override
= s
->override
;
2127 #ifdef TARGET_X86_64
2129 gen_op_addq_A0_seg(override
);
2133 gen_op_addl_A0_seg(s
, override
);
2138 /* generate modrm memory load or store of 'reg'. TMP0 is used if reg ==
2140 static void gen_ldst_modrm(CPUX86State
*env
, DisasContext
*s
, int modrm
,
2141 TCGMemOp ot
, int reg
, int is_store
)
2145 mod
= (modrm
>> 6) & 3;
2146 rm
= (modrm
& 7) | REX_B(s
);
2150 gen_op_mov_v_reg(ot
, cpu_T
[0], reg
);
2151 gen_op_mov_reg_v(ot
, rm
, cpu_T
[0]);
2153 gen_op_mov_v_reg(ot
, cpu_T
[0], rm
);
2155 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
2158 gen_lea_modrm(env
, s
, modrm
);
2161 gen_op_mov_v_reg(ot
, cpu_T
[0], reg
);
2162 gen_op_st_v(s
, ot
, cpu_T
[0], cpu_A0
);
2164 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
2166 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
2171 static inline uint32_t insn_get(CPUX86State
*env
, DisasContext
*s
, TCGMemOp ot
)
2177 ret
= cpu_ldub_code(env
, s
->pc
);
2181 ret
= cpu_lduw_code(env
, s
->pc
);
2185 #ifdef TARGET_X86_64
2188 ret
= cpu_ldl_code(env
, s
->pc
);
2197 static inline int insn_const_size(TCGMemOp ot
)
2206 static inline void gen_goto_tb(DisasContext
*s
, int tb_num
, target_ulong eip
)
2208 TranslationBlock
*tb
;
2211 pc
= s
->cs_base
+ eip
;
2213 /* NOTE: we handle the case where the TB spans two pages here */
2214 if ((pc
& TARGET_PAGE_MASK
) == (tb
->pc
& TARGET_PAGE_MASK
) ||
2215 (pc
& TARGET_PAGE_MASK
) == ((s
->pc
- 1) & TARGET_PAGE_MASK
)) {
2216 /* jump to same page: we can use a direct jump */
2217 tcg_gen_goto_tb(tb_num
);
2219 tcg_gen_exit_tb((uintptr_t)tb
+ tb_num
);
2221 /* jump to another page: currently not optimized */
2227 static inline void gen_jcc(DisasContext
*s
, int b
,
2228 target_ulong val
, target_ulong next_eip
)
2233 l1
= gen_new_label();
2236 gen_goto_tb(s
, 0, next_eip
);
2239 gen_goto_tb(s
, 1, val
);
2240 s
->is_jmp
= DISAS_TB_JUMP
;
2242 l1
= gen_new_label();
2243 l2
= gen_new_label();
2246 gen_jmp_im(next_eip
);
2256 static void gen_cmovcc1(CPUX86State
*env
, DisasContext
*s
, TCGMemOp ot
, int b
,
2261 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
2263 cc
= gen_prepare_cc(s
, b
, cpu_T
[1]);
2264 if (cc
.mask
!= -1) {
2265 TCGv t0
= tcg_temp_new();
2266 tcg_gen_andi_tl(t0
, cc
.reg
, cc
.mask
);
2270 cc
.reg2
= tcg_const_tl(cc
.imm
);
2273 tcg_gen_movcond_tl(cc
.cond
, cpu_T
[0], cc
.reg
, cc
.reg2
,
2274 cpu_T
[0], cpu_regs
[reg
]);
2275 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
2277 if (cc
.mask
!= -1) {
2278 tcg_temp_free(cc
.reg
);
2281 tcg_temp_free(cc
.reg2
);
2285 static inline void gen_op_movl_T0_seg(int seg_reg
)
2287 tcg_gen_ld32u_tl(cpu_T
[0], cpu_env
,
2288 offsetof(CPUX86State
,segs
[seg_reg
].selector
));
2291 static inline void gen_op_movl_seg_T0_vm(int seg_reg
)
2293 tcg_gen_andi_tl(cpu_T
[0], cpu_T
[0], 0xffff);
2294 tcg_gen_st32_tl(cpu_T
[0], cpu_env
,
2295 offsetof(CPUX86State
,segs
[seg_reg
].selector
));
2296 tcg_gen_shli_tl(cpu_T
[0], cpu_T
[0], 4);
2297 tcg_gen_st_tl(cpu_T
[0], cpu_env
,
2298 offsetof(CPUX86State
,segs
[seg_reg
].base
));
2301 /* move T0 to seg_reg and compute if the CPU state may change. Never
2302 call this function with seg_reg == R_CS */
2303 static void gen_movl_seg_T0(DisasContext
*s
, int seg_reg
, target_ulong cur_eip
)
2305 if (s
->pe
&& !s
->vm86
) {
2306 /* XXX: optimize by finding processor state dynamically */
2307 gen_update_cc_op(s
);
2308 gen_jmp_im(cur_eip
);
2309 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
2310 gen_helper_load_seg(cpu_env
, tcg_const_i32(seg_reg
), cpu_tmp2_i32
);
2311 /* abort translation because the addseg value may change or
2312 because ss32 may change. For R_SS, translation must always
2313 stop as a special handling must be done to disable hardware
2314 interrupts for the next instruction */
2315 if (seg_reg
== R_SS
|| (s
->code32
&& seg_reg
< R_FS
))
2316 s
->is_jmp
= DISAS_TB_JUMP
;
2318 gen_op_movl_seg_T0_vm(seg_reg
);
2319 if (seg_reg
== R_SS
)
2320 s
->is_jmp
= DISAS_TB_JUMP
;
2324 static inline int svm_is_rep(int prefixes
)
2326 return ((prefixes
& (PREFIX_REPZ
| PREFIX_REPNZ
)) ? 8 : 0);
2330 gen_svm_check_intercept_param(DisasContext
*s
, target_ulong pc_start
,
2331 uint32_t type
, uint64_t param
)
2333 /* no SVM activated; fast case */
2334 if (likely(!(s
->flags
& HF_SVMI_MASK
)))
2336 gen_update_cc_op(s
);
2337 gen_jmp_im(pc_start
- s
->cs_base
);
2338 gen_helper_svm_check_intercept_param(cpu_env
, tcg_const_i32(type
),
2339 tcg_const_i64(param
));
2343 gen_svm_check_intercept(DisasContext
*s
, target_ulong pc_start
, uint64_t type
)
2345 gen_svm_check_intercept_param(s
, pc_start
, type
, 0);
2348 static inline void gen_stack_update(DisasContext
*s
, int addend
)
2350 #ifdef TARGET_X86_64
2352 gen_op_add_reg_im(MO_64
, R_ESP
, addend
);
2356 gen_op_add_reg_im(MO_32
, R_ESP
, addend
);
2358 gen_op_add_reg_im(MO_16
, R_ESP
, addend
);
2362 /* Generate a push. It depends on ss32, addseg and dflag. */
2363 static void gen_push_v(DisasContext
*s
, TCGv val
)
2365 TCGMemOp a_ot
, d_ot
= mo_pushpop(s
, s
->dflag
);
2366 int size
= 1 << d_ot
;
2367 TCGv new_esp
= cpu_A0
;
2369 tcg_gen_subi_tl(cpu_A0
, cpu_regs
[R_ESP
], size
);
2373 } else if (s
->ss32
) {
2377 tcg_gen_mov_tl(new_esp
, cpu_A0
);
2378 gen_op_addl_A0_seg(s
, R_SS
);
2380 tcg_gen_ext32u_tl(cpu_A0
, cpu_A0
);
2385 tcg_gen_ext16u_tl(cpu_A0
, cpu_A0
);
2386 tcg_gen_mov_tl(new_esp
, cpu_A0
);
2387 gen_op_addl_A0_seg(s
, R_SS
);
2390 gen_op_st_v(s
, d_ot
, val
, cpu_A0
);
2391 gen_op_mov_reg_v(a_ot
, R_ESP
, new_esp
);
2394 /* two step pop is necessary for precise exceptions */
2395 static TCGMemOp
gen_pop_T0(DisasContext
*s
)
2397 TCGMemOp d_ot
= mo_pushpop(s
, s
->dflag
);
2401 addr
= cpu_regs
[R_ESP
];
2402 } else if (!s
->ss32
) {
2403 tcg_gen_ext16u_tl(cpu_A0
, cpu_regs
[R_ESP
]);
2404 gen_op_addl_A0_seg(s
, R_SS
);
2405 } else if (s
->addseg
) {
2406 tcg_gen_mov_tl(cpu_A0
, cpu_regs
[R_ESP
]);
2407 gen_op_addl_A0_seg(s
, R_SS
);
2409 tcg_gen_ext32u_tl(cpu_A0
, cpu_regs
[R_ESP
]);
2412 gen_op_ld_v(s
, d_ot
, cpu_T
[0], addr
);
2416 static void gen_pop_update(DisasContext
*s
, TCGMemOp ot
)
2418 gen_stack_update(s
, 1 << ot
);
2421 static void gen_stack_A0(DisasContext
*s
)
2423 gen_op_movl_A0_reg(R_ESP
);
2425 tcg_gen_ext16u_tl(cpu_A0
, cpu_A0
);
2426 tcg_gen_mov_tl(cpu_T
[1], cpu_A0
);
2428 gen_op_addl_A0_seg(s
, R_SS
);
2431 /* NOTE: wrap around in 16 bit not fully handled */
2432 static void gen_pusha(DisasContext
*s
)
2435 gen_op_movl_A0_reg(R_ESP
);
2436 gen_op_addl_A0_im(-8 << s
->dflag
);
2438 tcg_gen_ext16u_tl(cpu_A0
, cpu_A0
);
2439 tcg_gen_mov_tl(cpu_T
[1], cpu_A0
);
2441 gen_op_addl_A0_seg(s
, R_SS
);
2442 for(i
= 0;i
< 8; i
++) {
2443 gen_op_mov_v_reg(MO_32
, cpu_T
[0], 7 - i
);
2444 gen_op_st_v(s
, s
->dflag
, cpu_T
[0], cpu_A0
);
2445 gen_op_addl_A0_im(1 << s
->dflag
);
2447 gen_op_mov_reg_v(MO_16
+ s
->ss32
, R_ESP
, cpu_T
[1]);
2450 /* NOTE: wrap around in 16 bit not fully handled */
2451 static void gen_popa(DisasContext
*s
)
2454 gen_op_movl_A0_reg(R_ESP
);
2456 tcg_gen_ext16u_tl(cpu_A0
, cpu_A0
);
2457 tcg_gen_mov_tl(cpu_T
[1], cpu_A0
);
2458 tcg_gen_addi_tl(cpu_T
[1], cpu_T
[1], 8 << s
->dflag
);
2460 gen_op_addl_A0_seg(s
, R_SS
);
2461 for(i
= 0;i
< 8; i
++) {
2462 /* ESP is not reloaded */
2464 gen_op_ld_v(s
, s
->dflag
, cpu_T
[0], cpu_A0
);
2465 gen_op_mov_reg_v(s
->dflag
, 7 - i
, cpu_T
[0]);
2467 gen_op_addl_A0_im(1 << s
->dflag
);
2469 gen_op_mov_reg_v(MO_16
+ s
->ss32
, R_ESP
, cpu_T
[1]);
2472 static void gen_enter(DisasContext
*s
, int esp_addend
, int level
)
2474 TCGMemOp ot
= mo_pushpop(s
, s
->dflag
);
2475 int opsize
= 1 << ot
;
2478 #ifdef TARGET_X86_64
2480 gen_op_movl_A0_reg(R_ESP
);
2481 gen_op_addq_A0_im(-opsize
);
2482 tcg_gen_mov_tl(cpu_T
[1], cpu_A0
);
2485 gen_op_mov_v_reg(MO_32
, cpu_T
[0], R_EBP
);
2486 gen_op_st_v(s
, ot
, cpu_T
[0], cpu_A0
);
2488 /* XXX: must save state */
2489 gen_helper_enter64_level(cpu_env
, tcg_const_i32(level
),
2490 tcg_const_i32((ot
== MO_64
)),
2493 gen_op_mov_reg_v(ot
, R_EBP
, cpu_T
[1]);
2494 tcg_gen_addi_tl(cpu_T
[1], cpu_T
[1], -esp_addend
+ (-opsize
* level
));
2495 gen_op_mov_reg_v(MO_64
, R_ESP
, cpu_T
[1]);
2499 gen_op_movl_A0_reg(R_ESP
);
2500 gen_op_addl_A0_im(-opsize
);
2502 tcg_gen_ext16u_tl(cpu_A0
, cpu_A0
);
2503 tcg_gen_mov_tl(cpu_T
[1], cpu_A0
);
2505 gen_op_addl_A0_seg(s
, R_SS
);
2507 gen_op_mov_v_reg(MO_32
, cpu_T
[0], R_EBP
);
2508 gen_op_st_v(s
, ot
, cpu_T
[0], cpu_A0
);
2510 /* XXX: must save state */
2511 gen_helper_enter_level(cpu_env
, tcg_const_i32(level
),
2512 tcg_const_i32(s
->dflag
- 1),
2515 gen_op_mov_reg_v(ot
, R_EBP
, cpu_T
[1]);
2516 tcg_gen_addi_tl(cpu_T
[1], cpu_T
[1], -esp_addend
+ (-opsize
* level
));
2517 gen_op_mov_reg_v(MO_16
+ s
->ss32
, R_ESP
, cpu_T
[1]);
2521 static void gen_exception(DisasContext
*s
, int trapno
, target_ulong cur_eip
)
2523 gen_update_cc_op(s
);
2524 gen_jmp_im(cur_eip
);
2525 gen_helper_raise_exception(cpu_env
, tcg_const_i32(trapno
));
2526 s
->is_jmp
= DISAS_TB_JUMP
;
2529 /* an interrupt is different from an exception because of the
2531 static void gen_interrupt(DisasContext
*s
, int intno
,
2532 target_ulong cur_eip
, target_ulong next_eip
)
2534 gen_update_cc_op(s
);
2535 gen_jmp_im(cur_eip
);
2536 gen_helper_raise_interrupt(cpu_env
, tcg_const_i32(intno
),
2537 tcg_const_i32(next_eip
- cur_eip
));
2538 s
->is_jmp
= DISAS_TB_JUMP
;
2541 static void gen_debug(DisasContext
*s
, target_ulong cur_eip
)
2543 gen_update_cc_op(s
);
2544 gen_jmp_im(cur_eip
);
2545 gen_helper_debug(cpu_env
);
2546 s
->is_jmp
= DISAS_TB_JUMP
;
2549 /* generate a generic end of block. Trace exception is also generated
2551 static void gen_eob(DisasContext
*s
)
2553 gen_update_cc_op(s
);
2554 if (s
->tb
->flags
& HF_INHIBIT_IRQ_MASK
) {
2555 gen_helper_reset_inhibit_irq(cpu_env
);
2557 if (s
->tb
->flags
& HF_RF_MASK
) {
2558 gen_helper_reset_rf(cpu_env
);
2560 if (s
->singlestep_enabled
) {
2561 gen_helper_debug(cpu_env
);
2563 gen_helper_single_step(cpu_env
);
2567 s
->is_jmp
= DISAS_TB_JUMP
;
2570 /* generate a jump to eip. No segment change must happen before as a
2571 direct call to the next block may occur */
2572 static void gen_jmp_tb(DisasContext
*s
, target_ulong eip
, int tb_num
)
2574 gen_update_cc_op(s
);
2575 set_cc_op(s
, CC_OP_DYNAMIC
);
2577 gen_goto_tb(s
, tb_num
, eip
);
2578 s
->is_jmp
= DISAS_TB_JUMP
;
2585 static void gen_jmp(DisasContext
*s
, target_ulong eip
)
2587 gen_jmp_tb(s
, eip
, 0);
2590 static inline void gen_ldq_env_A0(DisasContext
*s
, int offset
)
2592 tcg_gen_qemu_ld_i64(cpu_tmp1_i64
, cpu_A0
, s
->mem_index
, MO_LEQ
);
2593 tcg_gen_st_i64(cpu_tmp1_i64
, cpu_env
, offset
);
2596 static inline void gen_stq_env_A0(DisasContext
*s
, int offset
)
2598 tcg_gen_ld_i64(cpu_tmp1_i64
, cpu_env
, offset
);
2599 tcg_gen_qemu_st_i64(cpu_tmp1_i64
, cpu_A0
, s
->mem_index
, MO_LEQ
);
2602 static inline void gen_ldo_env_A0(DisasContext
*s
, int offset
)
2604 int mem_index
= s
->mem_index
;
2605 tcg_gen_qemu_ld_i64(cpu_tmp1_i64
, cpu_A0
, mem_index
, MO_LEQ
);
2606 tcg_gen_st_i64(cpu_tmp1_i64
, cpu_env
, offset
+ offsetof(XMMReg
, XMM_Q(0)));
2607 tcg_gen_addi_tl(cpu_tmp0
, cpu_A0
, 8);
2608 tcg_gen_qemu_ld_i64(cpu_tmp1_i64
, cpu_tmp0
, mem_index
, MO_LEQ
);
2609 tcg_gen_st_i64(cpu_tmp1_i64
, cpu_env
, offset
+ offsetof(XMMReg
, XMM_Q(1)));
2612 static inline void gen_sto_env_A0(DisasContext
*s
, int offset
)
2614 int mem_index
= s
->mem_index
;
2615 tcg_gen_ld_i64(cpu_tmp1_i64
, cpu_env
, offset
+ offsetof(XMMReg
, XMM_Q(0)));
2616 tcg_gen_qemu_st_i64(cpu_tmp1_i64
, cpu_A0
, mem_index
, MO_LEQ
);
2617 tcg_gen_addi_tl(cpu_tmp0
, cpu_A0
, 8);
2618 tcg_gen_ld_i64(cpu_tmp1_i64
, cpu_env
, offset
+ offsetof(XMMReg
, XMM_Q(1)));
2619 tcg_gen_qemu_st_i64(cpu_tmp1_i64
, cpu_tmp0
, mem_index
, MO_LEQ
);
2622 static inline void gen_op_movo(int d_offset
, int s_offset
)
2624 tcg_gen_ld_i64(cpu_tmp1_i64
, cpu_env
, s_offset
+ offsetof(XMMReg
, XMM_Q(0)));
2625 tcg_gen_st_i64(cpu_tmp1_i64
, cpu_env
, d_offset
+ offsetof(XMMReg
, XMM_Q(0)));
2626 tcg_gen_ld_i64(cpu_tmp1_i64
, cpu_env
, s_offset
+ offsetof(XMMReg
, XMM_Q(1)));
2627 tcg_gen_st_i64(cpu_tmp1_i64
, cpu_env
, d_offset
+ offsetof(XMMReg
, XMM_Q(1)));
2630 static inline void gen_op_movq(int d_offset
, int s_offset
)
2632 tcg_gen_ld_i64(cpu_tmp1_i64
, cpu_env
, s_offset
);
2633 tcg_gen_st_i64(cpu_tmp1_i64
, cpu_env
, d_offset
);
2636 static inline void gen_op_movl(int d_offset
, int s_offset
)
2638 tcg_gen_ld_i32(cpu_tmp2_i32
, cpu_env
, s_offset
);
2639 tcg_gen_st_i32(cpu_tmp2_i32
, cpu_env
, d_offset
);
2642 static inline void gen_op_movq_env_0(int d_offset
)
2644 tcg_gen_movi_i64(cpu_tmp1_i64
, 0);
2645 tcg_gen_st_i64(cpu_tmp1_i64
, cpu_env
, d_offset
);
2648 typedef void (*SSEFunc_i_ep
)(TCGv_i32 val
, TCGv_ptr env
, TCGv_ptr reg
);
2649 typedef void (*SSEFunc_l_ep
)(TCGv_i64 val
, TCGv_ptr env
, TCGv_ptr reg
);
2650 typedef void (*SSEFunc_0_epi
)(TCGv_ptr env
, TCGv_ptr reg
, TCGv_i32 val
);
2651 typedef void (*SSEFunc_0_epl
)(TCGv_ptr env
, TCGv_ptr reg
, TCGv_i64 val
);
2652 typedef void (*SSEFunc_0_epp
)(TCGv_ptr env
, TCGv_ptr reg_a
, TCGv_ptr reg_b
);
2653 typedef void (*SSEFunc_0_eppi
)(TCGv_ptr env
, TCGv_ptr reg_a
, TCGv_ptr reg_b
,
2655 typedef void (*SSEFunc_0_ppi
)(TCGv_ptr reg_a
, TCGv_ptr reg_b
, TCGv_i32 val
);
2656 typedef void (*SSEFunc_0_eppt
)(TCGv_ptr env
, TCGv_ptr reg_a
, TCGv_ptr reg_b
,
2659 #define SSE_SPECIAL ((void *)1)
2660 #define SSE_DUMMY ((void *)2)
2662 #define MMX_OP2(x) { gen_helper_ ## x ## _mmx, gen_helper_ ## x ## _xmm }
2663 #define SSE_FOP(x) { gen_helper_ ## x ## ps, gen_helper_ ## x ## pd, \
2664 gen_helper_ ## x ## ss, gen_helper_ ## x ## sd, }
2666 static const SSEFunc_0_epp sse_op_table1
[256][4] = {
2667 /* 3DNow! extensions */
2668 [0x0e] = { SSE_DUMMY
}, /* femms */
2669 [0x0f] = { SSE_DUMMY
}, /* pf... */
2670 /* pure SSE operations */
2671 [0x10] = { SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
}, /* movups, movupd, movss, movsd */
2672 [0x11] = { SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
}, /* movups, movupd, movss, movsd */
2673 [0x12] = { SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
}, /* movlps, movlpd, movsldup, movddup */
2674 [0x13] = { SSE_SPECIAL
, SSE_SPECIAL
}, /* movlps, movlpd */
2675 [0x14] = { gen_helper_punpckldq_xmm
, gen_helper_punpcklqdq_xmm
},
2676 [0x15] = { gen_helper_punpckhdq_xmm
, gen_helper_punpckhqdq_xmm
},
2677 [0x16] = { SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
}, /* movhps, movhpd, movshdup */
2678 [0x17] = { SSE_SPECIAL
, SSE_SPECIAL
}, /* movhps, movhpd */
2680 [0x28] = { SSE_SPECIAL
, SSE_SPECIAL
}, /* movaps, movapd */
2681 [0x29] = { SSE_SPECIAL
, SSE_SPECIAL
}, /* movaps, movapd */
2682 [0x2a] = { SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
}, /* cvtpi2ps, cvtpi2pd, cvtsi2ss, cvtsi2sd */
2683 [0x2b] = { SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
}, /* movntps, movntpd, movntss, movntsd */
2684 [0x2c] = { SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
}, /* cvttps2pi, cvttpd2pi, cvttsd2si, cvttss2si */
2685 [0x2d] = { SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
}, /* cvtps2pi, cvtpd2pi, cvtsd2si, cvtss2si */
2686 [0x2e] = { gen_helper_ucomiss
, gen_helper_ucomisd
},
2687 [0x2f] = { gen_helper_comiss
, gen_helper_comisd
},
2688 [0x50] = { SSE_SPECIAL
, SSE_SPECIAL
}, /* movmskps, movmskpd */
2689 [0x51] = SSE_FOP(sqrt
),
2690 [0x52] = { gen_helper_rsqrtps
, NULL
, gen_helper_rsqrtss
, NULL
},
2691 [0x53] = { gen_helper_rcpps
, NULL
, gen_helper_rcpss
, NULL
},
2692 [0x54] = { gen_helper_pand_xmm
, gen_helper_pand_xmm
}, /* andps, andpd */
2693 [0x55] = { gen_helper_pandn_xmm
, gen_helper_pandn_xmm
}, /* andnps, andnpd */
2694 [0x56] = { gen_helper_por_xmm
, gen_helper_por_xmm
}, /* orps, orpd */
2695 [0x57] = { gen_helper_pxor_xmm
, gen_helper_pxor_xmm
}, /* xorps, xorpd */
2696 [0x58] = SSE_FOP(add
),
2697 [0x59] = SSE_FOP(mul
),
2698 [0x5a] = { gen_helper_cvtps2pd
, gen_helper_cvtpd2ps
,
2699 gen_helper_cvtss2sd
, gen_helper_cvtsd2ss
},
2700 [0x5b] = { gen_helper_cvtdq2ps
, gen_helper_cvtps2dq
, gen_helper_cvttps2dq
},
2701 [0x5c] = SSE_FOP(sub
),
2702 [0x5d] = SSE_FOP(min
),
2703 [0x5e] = SSE_FOP(div
),
2704 [0x5f] = SSE_FOP(max
),
2706 [0xc2] = SSE_FOP(cmpeq
),
2707 [0xc6] = { (SSEFunc_0_epp
)gen_helper_shufps
,
2708 (SSEFunc_0_epp
)gen_helper_shufpd
}, /* XXX: casts */
2710 /* SSSE3, SSE4, MOVBE, CRC32, BMI1, BMI2, ADX. */
2711 [0x38] = { SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
},
2712 [0x3a] = { SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
},
2714 /* MMX ops and their SSE extensions */
2715 [0x60] = MMX_OP2(punpcklbw
),
2716 [0x61] = MMX_OP2(punpcklwd
),
2717 [0x62] = MMX_OP2(punpckldq
),
2718 [0x63] = MMX_OP2(packsswb
),
2719 [0x64] = MMX_OP2(pcmpgtb
),
2720 [0x65] = MMX_OP2(pcmpgtw
),
2721 [0x66] = MMX_OP2(pcmpgtl
),
2722 [0x67] = MMX_OP2(packuswb
),
2723 [0x68] = MMX_OP2(punpckhbw
),
2724 [0x69] = MMX_OP2(punpckhwd
),
2725 [0x6a] = MMX_OP2(punpckhdq
),
2726 [0x6b] = MMX_OP2(packssdw
),
2727 [0x6c] = { NULL
, gen_helper_punpcklqdq_xmm
},
2728 [0x6d] = { NULL
, gen_helper_punpckhqdq_xmm
},
2729 [0x6e] = { SSE_SPECIAL
, SSE_SPECIAL
}, /* movd mm, ea */
2730 [0x6f] = { SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
}, /* movq, movdqa, , movqdu */
2731 [0x70] = { (SSEFunc_0_epp
)gen_helper_pshufw_mmx
,
2732 (SSEFunc_0_epp
)gen_helper_pshufd_xmm
,
2733 (SSEFunc_0_epp
)gen_helper_pshufhw_xmm
,
2734 (SSEFunc_0_epp
)gen_helper_pshuflw_xmm
}, /* XXX: casts */
2735 [0x71] = { SSE_SPECIAL
, SSE_SPECIAL
}, /* shiftw */
2736 [0x72] = { SSE_SPECIAL
, SSE_SPECIAL
}, /* shiftd */
2737 [0x73] = { SSE_SPECIAL
, SSE_SPECIAL
}, /* shiftq */
2738 [0x74] = MMX_OP2(pcmpeqb
),
2739 [0x75] = MMX_OP2(pcmpeqw
),
2740 [0x76] = MMX_OP2(pcmpeql
),
2741 [0x77] = { SSE_DUMMY
}, /* emms */
2742 [0x78] = { NULL
, SSE_SPECIAL
, NULL
, SSE_SPECIAL
}, /* extrq_i, insertq_i */
2743 [0x79] = { NULL
, gen_helper_extrq_r
, NULL
, gen_helper_insertq_r
},
2744 [0x7c] = { NULL
, gen_helper_haddpd
, NULL
, gen_helper_haddps
},
2745 [0x7d] = { NULL
, gen_helper_hsubpd
, NULL
, gen_helper_hsubps
},
2746 [0x7e] = { SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
}, /* movd, movd, , movq */
2747 [0x7f] = { SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
}, /* movq, movdqa, movdqu */
2748 [0xc4] = { SSE_SPECIAL
, SSE_SPECIAL
}, /* pinsrw */
2749 [0xc5] = { SSE_SPECIAL
, SSE_SPECIAL
}, /* pextrw */
2750 [0xd0] = { NULL
, gen_helper_addsubpd
, NULL
, gen_helper_addsubps
},
2751 [0xd1] = MMX_OP2(psrlw
),
2752 [0xd2] = MMX_OP2(psrld
),
2753 [0xd3] = MMX_OP2(psrlq
),
2754 [0xd4] = MMX_OP2(paddq
),
2755 [0xd5] = MMX_OP2(pmullw
),
2756 [0xd6] = { NULL
, SSE_SPECIAL
, SSE_SPECIAL
, SSE_SPECIAL
},
2757 [0xd7] = { SSE_SPECIAL
, SSE_SPECIAL
}, /* pmovmskb */
2758 [0xd8] = MMX_OP2(psubusb
),
2759 [0xd9] = MMX_OP2(psubusw
),
2760 [0xda] = MMX_OP2(pminub
),
2761 [0xdb] = MMX_OP2(pand
),
2762 [0xdc] = MMX_OP2(paddusb
),
2763 [0xdd] = MMX_OP2(paddusw
),
2764 [0xde] = MMX_OP2(pmaxub
),
2765 [0xdf] = MMX_OP2(pandn
),
2766 [0xe0] = MMX_OP2(pavgb
),
2767 [0xe1] = MMX_OP2(psraw
),
2768 [0xe2] = MMX_OP2(psrad
),
2769 [0xe3] = MMX_OP2(pavgw
),
2770 [0xe4] = MMX_OP2(pmulhuw
),
2771 [0xe5] = MMX_OP2(pmulhw
),
2772 [0xe6] = { NULL
, gen_helper_cvttpd2dq
, gen_helper_cvtdq2pd
, gen_helper_cvtpd2dq
},
2773 [0xe7] = { SSE_SPECIAL
, SSE_SPECIAL
}, /* movntq, movntq */
2774 [0xe8] = MMX_OP2(psubsb
),
2775 [0xe9] = MMX_OP2(psubsw
),
2776 [0xea] = MMX_OP2(pminsw
),
2777 [0xeb] = MMX_OP2(por
),
2778 [0xec] = MMX_OP2(paddsb
),
2779 [0xed] = MMX_OP2(paddsw
),
2780 [0xee] = MMX_OP2(pmaxsw
),
2781 [0xef] = MMX_OP2(pxor
),
2782 [0xf0] = { NULL
, NULL
, NULL
, SSE_SPECIAL
}, /* lddqu */
2783 [0xf1] = MMX_OP2(psllw
),
2784 [0xf2] = MMX_OP2(pslld
),
2785 [0xf3] = MMX_OP2(psllq
),
2786 [0xf4] = MMX_OP2(pmuludq
),
2787 [0xf5] = MMX_OP2(pmaddwd
),
2788 [0xf6] = MMX_OP2(psadbw
),
2789 [0xf7] = { (SSEFunc_0_epp
)gen_helper_maskmov_mmx
,
2790 (SSEFunc_0_epp
)gen_helper_maskmov_xmm
}, /* XXX: casts */
2791 [0xf8] = MMX_OP2(psubb
),
2792 [0xf9] = MMX_OP2(psubw
),
2793 [0xfa] = MMX_OP2(psubl
),
2794 [0xfb] = MMX_OP2(psubq
),
2795 [0xfc] = MMX_OP2(paddb
),
2796 [0xfd] = MMX_OP2(paddw
),
2797 [0xfe] = MMX_OP2(paddl
),
2800 static const SSEFunc_0_epp sse_op_table2
[3 * 8][2] = {
2801 [0 + 2] = MMX_OP2(psrlw
),
2802 [0 + 4] = MMX_OP2(psraw
),
2803 [0 + 6] = MMX_OP2(psllw
),
2804 [8 + 2] = MMX_OP2(psrld
),
2805 [8 + 4] = MMX_OP2(psrad
),
2806 [8 + 6] = MMX_OP2(pslld
),
2807 [16 + 2] = MMX_OP2(psrlq
),
2808 [16 + 3] = { NULL
, gen_helper_psrldq_xmm
},
2809 [16 + 6] = MMX_OP2(psllq
),
2810 [16 + 7] = { NULL
, gen_helper_pslldq_xmm
},
2813 static const SSEFunc_0_epi sse_op_table3ai
[] = {
2814 gen_helper_cvtsi2ss
,
2818 #ifdef TARGET_X86_64
2819 static const SSEFunc_0_epl sse_op_table3aq
[] = {
2820 gen_helper_cvtsq2ss
,
2825 static const SSEFunc_i_ep sse_op_table3bi
[] = {
2826 gen_helper_cvttss2si
,
2827 gen_helper_cvtss2si
,
2828 gen_helper_cvttsd2si
,
2832 #ifdef TARGET_X86_64
2833 static const SSEFunc_l_ep sse_op_table3bq
[] = {
2834 gen_helper_cvttss2sq
,
2835 gen_helper_cvtss2sq
,
2836 gen_helper_cvttsd2sq
,
2841 static const SSEFunc_0_epp sse_op_table4
[8][4] = {
2852 static const SSEFunc_0_epp sse_op_table5
[256] = {
2853 [0x0c] = gen_helper_pi2fw
,
2854 [0x0d] = gen_helper_pi2fd
,
2855 [0x1c] = gen_helper_pf2iw
,
2856 [0x1d] = gen_helper_pf2id
,
2857 [0x8a] = gen_helper_pfnacc
,
2858 [0x8e] = gen_helper_pfpnacc
,
2859 [0x90] = gen_helper_pfcmpge
,
2860 [0x94] = gen_helper_pfmin
,
2861 [0x96] = gen_helper_pfrcp
,
2862 [0x97] = gen_helper_pfrsqrt
,
2863 [0x9a] = gen_helper_pfsub
,
2864 [0x9e] = gen_helper_pfadd
,
2865 [0xa0] = gen_helper_pfcmpgt
,
2866 [0xa4] = gen_helper_pfmax
,
2867 [0xa6] = gen_helper_movq
, /* pfrcpit1; no need to actually increase precision */
2868 [0xa7] = gen_helper_movq
, /* pfrsqit1 */
2869 [0xaa] = gen_helper_pfsubr
,
2870 [0xae] = gen_helper_pfacc
,
2871 [0xb0] = gen_helper_pfcmpeq
,
2872 [0xb4] = gen_helper_pfmul
,
2873 [0xb6] = gen_helper_movq
, /* pfrcpit2 */
2874 [0xb7] = gen_helper_pmulhrw_mmx
,
2875 [0xbb] = gen_helper_pswapd
,
2876 [0xbf] = gen_helper_pavgb_mmx
/* pavgusb */
2879 struct SSEOpHelper_epp
{
2880 SSEFunc_0_epp op
[2];
2884 struct SSEOpHelper_eppi
{
2885 SSEFunc_0_eppi op
[2];
2889 #define SSSE3_OP(x) { MMX_OP2(x), CPUID_EXT_SSSE3 }
2890 #define SSE41_OP(x) { { NULL, gen_helper_ ## x ## _xmm }, CPUID_EXT_SSE41 }
2891 #define SSE42_OP(x) { { NULL, gen_helper_ ## x ## _xmm }, CPUID_EXT_SSE42 }
2892 #define SSE41_SPECIAL { { NULL, SSE_SPECIAL }, CPUID_EXT_SSE41 }
2893 #define PCLMULQDQ_OP(x) { { NULL, gen_helper_ ## x ## _xmm }, \
2894 CPUID_EXT_PCLMULQDQ }
2895 #define AESNI_OP(x) { { NULL, gen_helper_ ## x ## _xmm }, CPUID_EXT_AES }
2897 static const struct SSEOpHelper_epp sse_op_table6
[256] = {
2898 [0x00] = SSSE3_OP(pshufb
),
2899 [0x01] = SSSE3_OP(phaddw
),
2900 [0x02] = SSSE3_OP(phaddd
),
2901 [0x03] = SSSE3_OP(phaddsw
),
2902 [0x04] = SSSE3_OP(pmaddubsw
),
2903 [0x05] = SSSE3_OP(phsubw
),
2904 [0x06] = SSSE3_OP(phsubd
),
2905 [0x07] = SSSE3_OP(phsubsw
),
2906 [0x08] = SSSE3_OP(psignb
),
2907 [0x09] = SSSE3_OP(psignw
),
2908 [0x0a] = SSSE3_OP(psignd
),
2909 [0x0b] = SSSE3_OP(pmulhrsw
),
2910 [0x10] = SSE41_OP(pblendvb
),
2911 [0x14] = SSE41_OP(blendvps
),
2912 [0x15] = SSE41_OP(blendvpd
),
2913 [0x17] = SSE41_OP(ptest
),
2914 [0x1c] = SSSE3_OP(pabsb
),
2915 [0x1d] = SSSE3_OP(pabsw
),
2916 [0x1e] = SSSE3_OP(pabsd
),
2917 [0x20] = SSE41_OP(pmovsxbw
),
2918 [0x21] = SSE41_OP(pmovsxbd
),
2919 [0x22] = SSE41_OP(pmovsxbq
),
2920 [0x23] = SSE41_OP(pmovsxwd
),
2921 [0x24] = SSE41_OP(pmovsxwq
),
2922 [0x25] = SSE41_OP(pmovsxdq
),
2923 [0x28] = SSE41_OP(pmuldq
),
2924 [0x29] = SSE41_OP(pcmpeqq
),
2925 [0x2a] = SSE41_SPECIAL
, /* movntqda */
2926 [0x2b] = SSE41_OP(packusdw
),
2927 [0x30] = SSE41_OP(pmovzxbw
),
2928 [0x31] = SSE41_OP(pmovzxbd
),
2929 [0x32] = SSE41_OP(pmovzxbq
),
2930 [0x33] = SSE41_OP(pmovzxwd
),
2931 [0x34] = SSE41_OP(pmovzxwq
),
2932 [0x35] = SSE41_OP(pmovzxdq
),
2933 [0x37] = SSE42_OP(pcmpgtq
),
2934 [0x38] = SSE41_OP(pminsb
),
2935 [0x39] = SSE41_OP(pminsd
),
2936 [0x3a] = SSE41_OP(pminuw
),
2937 [0x3b] = SSE41_OP(pminud
),
2938 [0x3c] = SSE41_OP(pmaxsb
),
2939 [0x3d] = SSE41_OP(pmaxsd
),
2940 [0x3e] = SSE41_OP(pmaxuw
),
2941 [0x3f] = SSE41_OP(pmaxud
),
2942 [0x40] = SSE41_OP(pmulld
),
2943 [0x41] = SSE41_OP(phminposuw
),
2944 [0xdb] = AESNI_OP(aesimc
),
2945 [0xdc] = AESNI_OP(aesenc
),
2946 [0xdd] = AESNI_OP(aesenclast
),
2947 [0xde] = AESNI_OP(aesdec
),
2948 [0xdf] = AESNI_OP(aesdeclast
),
2951 static const struct SSEOpHelper_eppi sse_op_table7
[256] = {
2952 [0x08] = SSE41_OP(roundps
),
2953 [0x09] = SSE41_OP(roundpd
),
2954 [0x0a] = SSE41_OP(roundss
),
2955 [0x0b] = SSE41_OP(roundsd
),
2956 [0x0c] = SSE41_OP(blendps
),
2957 [0x0d] = SSE41_OP(blendpd
),
2958 [0x0e] = SSE41_OP(pblendw
),
2959 [0x0f] = SSSE3_OP(palignr
),
2960 [0x14] = SSE41_SPECIAL
, /* pextrb */
2961 [0x15] = SSE41_SPECIAL
, /* pextrw */
2962 [0x16] = SSE41_SPECIAL
, /* pextrd/pextrq */
2963 [0x17] = SSE41_SPECIAL
, /* extractps */
2964 [0x20] = SSE41_SPECIAL
, /* pinsrb */
2965 [0x21] = SSE41_SPECIAL
, /* insertps */
2966 [0x22] = SSE41_SPECIAL
, /* pinsrd/pinsrq */
2967 [0x40] = SSE41_OP(dpps
),
2968 [0x41] = SSE41_OP(dppd
),
2969 [0x42] = SSE41_OP(mpsadbw
),
2970 [0x44] = PCLMULQDQ_OP(pclmulqdq
),
2971 [0x60] = SSE42_OP(pcmpestrm
),
2972 [0x61] = SSE42_OP(pcmpestri
),
2973 [0x62] = SSE42_OP(pcmpistrm
),
2974 [0x63] = SSE42_OP(pcmpistri
),
2975 [0xdf] = AESNI_OP(aeskeygenassist
),
2978 static void gen_sse(CPUX86State
*env
, DisasContext
*s
, int b
,
2979 target_ulong pc_start
, int rex_r
)
2981 int b1
, op1_offset
, op2_offset
, is_xmm
, val
;
2982 int modrm
, mod
, rm
, reg
;
2983 SSEFunc_0_epp sse_fn_epp
;
2984 SSEFunc_0_eppi sse_fn_eppi
;
2985 SSEFunc_0_ppi sse_fn_ppi
;
2986 SSEFunc_0_eppt sse_fn_eppt
;
2990 if (s
->prefix
& PREFIX_DATA
)
2992 else if (s
->prefix
& PREFIX_REPZ
)
2994 else if (s
->prefix
& PREFIX_REPNZ
)
2998 sse_fn_epp
= sse_op_table1
[b
][b1
];
3002 if ((b
<= 0x5f && b
>= 0x10) || b
== 0xc6 || b
== 0xc2) {
3012 /* simple MMX/SSE operation */
3013 if (s
->flags
& HF_TS_MASK
) {
3014 gen_exception(s
, EXCP07_PREX
, pc_start
- s
->cs_base
);
3017 if (s
->flags
& HF_EM_MASK
) {
3019 gen_exception(s
, EXCP06_ILLOP
, pc_start
- s
->cs_base
);
3022 if (is_xmm
&& !(s
->flags
& HF_OSFXSR_MASK
))
3023 if ((b
!= 0x38 && b
!= 0x3a) || (s
->prefix
& PREFIX_DATA
))
3026 if (!(s
->cpuid_ext2_features
& CPUID_EXT2_3DNOW
))
3029 gen_helper_emms(cpu_env
);
3034 gen_helper_emms(cpu_env
);
3037 /* prepare MMX state (XXX: optimize by storing fptt and fptags in
3038 the static cpu state) */
3040 gen_helper_enter_mmx(cpu_env
);
3043 modrm
= cpu_ldub_code(env
, s
->pc
++);
3044 reg
= ((modrm
>> 3) & 7);
3047 mod
= (modrm
>> 6) & 3;
3048 if (sse_fn_epp
== SSE_SPECIAL
) {
3051 case 0x0e7: /* movntq */
3054 gen_lea_modrm(env
, s
, modrm
);
3055 gen_stq_env_A0(s
, offsetof(CPUX86State
, fpregs
[reg
].mmx
));
3057 case 0x1e7: /* movntdq */
3058 case 0x02b: /* movntps */
3059 case 0x12b: /* movntps */
3062 gen_lea_modrm(env
, s
, modrm
);
3063 gen_sto_env_A0(s
, offsetof(CPUX86State
, xmm_regs
[reg
]));
3065 case 0x3f0: /* lddqu */
3068 gen_lea_modrm(env
, s
, modrm
);
3069 gen_ldo_env_A0(s
, offsetof(CPUX86State
, xmm_regs
[reg
]));
3071 case 0x22b: /* movntss */
3072 case 0x32b: /* movntsd */
3075 gen_lea_modrm(env
, s
, modrm
);
3077 gen_stq_env_A0(s
, offsetof(CPUX86State
,
3078 xmm_regs
[reg
].XMM_Q(0)));
3080 tcg_gen_ld32u_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,
3081 xmm_regs
[reg
].XMM_L(0)));
3082 gen_op_st_v(s
, MO_32
, cpu_T
[0], cpu_A0
);
3085 case 0x6e: /* movd mm, ea */
3086 #ifdef TARGET_X86_64
3087 if (s
->dflag
== MO_64
) {
3088 gen_ldst_modrm(env
, s
, modrm
, MO_64
, OR_TMP0
, 0);
3089 tcg_gen_st_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,fpregs
[reg
].mmx
));
3093 gen_ldst_modrm(env
, s
, modrm
, MO_32
, OR_TMP0
, 0);
3094 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
,
3095 offsetof(CPUX86State
,fpregs
[reg
].mmx
));
3096 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
3097 gen_helper_movl_mm_T0_mmx(cpu_ptr0
, cpu_tmp2_i32
);
3100 case 0x16e: /* movd xmm, ea */
3101 #ifdef TARGET_X86_64
3102 if (s
->dflag
== MO_64
) {
3103 gen_ldst_modrm(env
, s
, modrm
, MO_64
, OR_TMP0
, 0);
3104 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
,
3105 offsetof(CPUX86State
,xmm_regs
[reg
]));
3106 gen_helper_movq_mm_T0_xmm(cpu_ptr0
, cpu_T
[0]);
3110 gen_ldst_modrm(env
, s
, modrm
, MO_32
, OR_TMP0
, 0);
3111 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
,
3112 offsetof(CPUX86State
,xmm_regs
[reg
]));
3113 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
3114 gen_helper_movl_mm_T0_xmm(cpu_ptr0
, cpu_tmp2_i32
);
3117 case 0x6f: /* movq mm, ea */
3119 gen_lea_modrm(env
, s
, modrm
);
3120 gen_ldq_env_A0(s
, offsetof(CPUX86State
, fpregs
[reg
].mmx
));
3123 tcg_gen_ld_i64(cpu_tmp1_i64
, cpu_env
,
3124 offsetof(CPUX86State
,fpregs
[rm
].mmx
));
3125 tcg_gen_st_i64(cpu_tmp1_i64
, cpu_env
,
3126 offsetof(CPUX86State
,fpregs
[reg
].mmx
));
3129 case 0x010: /* movups */
3130 case 0x110: /* movupd */
3131 case 0x028: /* movaps */
3132 case 0x128: /* movapd */
3133 case 0x16f: /* movdqa xmm, ea */
3134 case 0x26f: /* movdqu xmm, ea */
3136 gen_lea_modrm(env
, s
, modrm
);
3137 gen_ldo_env_A0(s
, offsetof(CPUX86State
, xmm_regs
[reg
]));
3139 rm
= (modrm
& 7) | REX_B(s
);
3140 gen_op_movo(offsetof(CPUX86State
,xmm_regs
[reg
]),
3141 offsetof(CPUX86State
,xmm_regs
[rm
]));
3144 case 0x210: /* movss xmm, ea */
3146 gen_lea_modrm(env
, s
, modrm
);
3147 gen_op_ld_v(s
, MO_32
, cpu_T
[0], cpu_A0
);
3148 tcg_gen_st32_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(0)));
3149 tcg_gen_movi_tl(cpu_T
[0], 0);
3150 tcg_gen_st32_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(1)));
3151 tcg_gen_st32_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(2)));
3152 tcg_gen_st32_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(3)));
3154 rm
= (modrm
& 7) | REX_B(s
);
3155 gen_op_movl(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(0)),
3156 offsetof(CPUX86State
,xmm_regs
[rm
].XMM_L(0)));
3159 case 0x310: /* movsd xmm, ea */
3161 gen_lea_modrm(env
, s
, modrm
);
3162 gen_ldq_env_A0(s
, offsetof(CPUX86State
,
3163 xmm_regs
[reg
].XMM_Q(0)));
3164 tcg_gen_movi_tl(cpu_T
[0], 0);
3165 tcg_gen_st32_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(2)));
3166 tcg_gen_st32_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(3)));
3168 rm
= (modrm
& 7) | REX_B(s
);
3169 gen_op_movq(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_Q(0)),
3170 offsetof(CPUX86State
,xmm_regs
[rm
].XMM_Q(0)));
3173 case 0x012: /* movlps */
3174 case 0x112: /* movlpd */
3176 gen_lea_modrm(env
, s
, modrm
);
3177 gen_ldq_env_A0(s
, offsetof(CPUX86State
,
3178 xmm_regs
[reg
].XMM_Q(0)));
3181 rm
= (modrm
& 7) | REX_B(s
);
3182 gen_op_movq(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_Q(0)),
3183 offsetof(CPUX86State
,xmm_regs
[rm
].XMM_Q(1)));
3186 case 0x212: /* movsldup */
3188 gen_lea_modrm(env
, s
, modrm
);
3189 gen_ldo_env_A0(s
, offsetof(CPUX86State
, xmm_regs
[reg
]));
3191 rm
= (modrm
& 7) | REX_B(s
);
3192 gen_op_movl(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(0)),
3193 offsetof(CPUX86State
,xmm_regs
[rm
].XMM_L(0)));
3194 gen_op_movl(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(2)),
3195 offsetof(CPUX86State
,xmm_regs
[rm
].XMM_L(2)));
3197 gen_op_movl(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(1)),
3198 offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(0)));
3199 gen_op_movl(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(3)),
3200 offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(2)));
3202 case 0x312: /* movddup */
3204 gen_lea_modrm(env
, s
, modrm
);
3205 gen_ldq_env_A0(s
, offsetof(CPUX86State
,
3206 xmm_regs
[reg
].XMM_Q(0)));
3208 rm
= (modrm
& 7) | REX_B(s
);
3209 gen_op_movq(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_Q(0)),
3210 offsetof(CPUX86State
,xmm_regs
[rm
].XMM_Q(0)));
3212 gen_op_movq(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_Q(1)),
3213 offsetof(CPUX86State
,xmm_regs
[reg
].XMM_Q(0)));
3215 case 0x016: /* movhps */
3216 case 0x116: /* movhpd */
3218 gen_lea_modrm(env
, s
, modrm
);
3219 gen_ldq_env_A0(s
, offsetof(CPUX86State
,
3220 xmm_regs
[reg
].XMM_Q(1)));
3223 rm
= (modrm
& 7) | REX_B(s
);
3224 gen_op_movq(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_Q(1)),
3225 offsetof(CPUX86State
,xmm_regs
[rm
].XMM_Q(0)));
3228 case 0x216: /* movshdup */
3230 gen_lea_modrm(env
, s
, modrm
);
3231 gen_ldo_env_A0(s
, offsetof(CPUX86State
, xmm_regs
[reg
]));
3233 rm
= (modrm
& 7) | REX_B(s
);
3234 gen_op_movl(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(1)),
3235 offsetof(CPUX86State
,xmm_regs
[rm
].XMM_L(1)));
3236 gen_op_movl(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(3)),
3237 offsetof(CPUX86State
,xmm_regs
[rm
].XMM_L(3)));
3239 gen_op_movl(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(0)),
3240 offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(1)));
3241 gen_op_movl(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(2)),
3242 offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(3)));
3247 int bit_index
, field_length
;
3249 if (b1
== 1 && reg
!= 0)
3251 field_length
= cpu_ldub_code(env
, s
->pc
++) & 0x3F;
3252 bit_index
= cpu_ldub_code(env
, s
->pc
++) & 0x3F;
3253 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
,
3254 offsetof(CPUX86State
,xmm_regs
[reg
]));
3256 gen_helper_extrq_i(cpu_env
, cpu_ptr0
,
3257 tcg_const_i32(bit_index
),
3258 tcg_const_i32(field_length
));
3260 gen_helper_insertq_i(cpu_env
, cpu_ptr0
,
3261 tcg_const_i32(bit_index
),
3262 tcg_const_i32(field_length
));
3265 case 0x7e: /* movd ea, mm */
3266 #ifdef TARGET_X86_64
3267 if (s
->dflag
== MO_64
) {
3268 tcg_gen_ld_i64(cpu_T
[0], cpu_env
,
3269 offsetof(CPUX86State
,fpregs
[reg
].mmx
));
3270 gen_ldst_modrm(env
, s
, modrm
, MO_64
, OR_TMP0
, 1);
3274 tcg_gen_ld32u_tl(cpu_T
[0], cpu_env
,
3275 offsetof(CPUX86State
,fpregs
[reg
].mmx
.MMX_L(0)));
3276 gen_ldst_modrm(env
, s
, modrm
, MO_32
, OR_TMP0
, 1);
3279 case 0x17e: /* movd ea, xmm */
3280 #ifdef TARGET_X86_64
3281 if (s
->dflag
== MO_64
) {
3282 tcg_gen_ld_i64(cpu_T
[0], cpu_env
,
3283 offsetof(CPUX86State
,xmm_regs
[reg
].XMM_Q(0)));
3284 gen_ldst_modrm(env
, s
, modrm
, MO_64
, OR_TMP0
, 1);
3288 tcg_gen_ld32u_tl(cpu_T
[0], cpu_env
,
3289 offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(0)));
3290 gen_ldst_modrm(env
, s
, modrm
, MO_32
, OR_TMP0
, 1);
3293 case 0x27e: /* movq xmm, ea */
3295 gen_lea_modrm(env
, s
, modrm
);
3296 gen_ldq_env_A0(s
, offsetof(CPUX86State
,
3297 xmm_regs
[reg
].XMM_Q(0)));
3299 rm
= (modrm
& 7) | REX_B(s
);
3300 gen_op_movq(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_Q(0)),
3301 offsetof(CPUX86State
,xmm_regs
[rm
].XMM_Q(0)));
3303 gen_op_movq_env_0(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_Q(1)));
3305 case 0x7f: /* movq ea, mm */
3307 gen_lea_modrm(env
, s
, modrm
);
3308 gen_stq_env_A0(s
, offsetof(CPUX86State
, fpregs
[reg
].mmx
));
3311 gen_op_movq(offsetof(CPUX86State
,fpregs
[rm
].mmx
),
3312 offsetof(CPUX86State
,fpregs
[reg
].mmx
));
3315 case 0x011: /* movups */
3316 case 0x111: /* movupd */
3317 case 0x029: /* movaps */
3318 case 0x129: /* movapd */
3319 case 0x17f: /* movdqa ea, xmm */
3320 case 0x27f: /* movdqu ea, xmm */
3322 gen_lea_modrm(env
, s
, modrm
);
3323 gen_sto_env_A0(s
, offsetof(CPUX86State
, xmm_regs
[reg
]));
3325 rm
= (modrm
& 7) | REX_B(s
);
3326 gen_op_movo(offsetof(CPUX86State
,xmm_regs
[rm
]),
3327 offsetof(CPUX86State
,xmm_regs
[reg
]));
3330 case 0x211: /* movss ea, xmm */
3332 gen_lea_modrm(env
, s
, modrm
);
3333 tcg_gen_ld32u_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(0)));
3334 gen_op_st_v(s
, MO_32
, cpu_T
[0], cpu_A0
);
3336 rm
= (modrm
& 7) | REX_B(s
);
3337 gen_op_movl(offsetof(CPUX86State
,xmm_regs
[rm
].XMM_L(0)),
3338 offsetof(CPUX86State
,xmm_regs
[reg
].XMM_L(0)));
3341 case 0x311: /* movsd ea, xmm */
3343 gen_lea_modrm(env
, s
, modrm
);
3344 gen_stq_env_A0(s
, offsetof(CPUX86State
,
3345 xmm_regs
[reg
].XMM_Q(0)));
3347 rm
= (modrm
& 7) | REX_B(s
);
3348 gen_op_movq(offsetof(CPUX86State
,xmm_regs
[rm
].XMM_Q(0)),
3349 offsetof(CPUX86State
,xmm_regs
[reg
].XMM_Q(0)));
3352 case 0x013: /* movlps */
3353 case 0x113: /* movlpd */
3355 gen_lea_modrm(env
, s
, modrm
);
3356 gen_stq_env_A0(s
, offsetof(CPUX86State
,
3357 xmm_regs
[reg
].XMM_Q(0)));
3362 case 0x017: /* movhps */
3363 case 0x117: /* movhpd */
3365 gen_lea_modrm(env
, s
, modrm
);
3366 gen_stq_env_A0(s
, offsetof(CPUX86State
,
3367 xmm_regs
[reg
].XMM_Q(1)));
3372 case 0x71: /* shift mm, im */
3375 case 0x171: /* shift xmm, im */
3381 val
= cpu_ldub_code(env
, s
->pc
++);
3383 tcg_gen_movi_tl(cpu_T
[0], val
);
3384 tcg_gen_st32_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,xmm_t0
.XMM_L(0)));
3385 tcg_gen_movi_tl(cpu_T
[0], 0);
3386 tcg_gen_st32_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,xmm_t0
.XMM_L(1)));
3387 op1_offset
= offsetof(CPUX86State
,xmm_t0
);
3389 tcg_gen_movi_tl(cpu_T
[0], val
);
3390 tcg_gen_st32_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,mmx_t0
.MMX_L(0)));
3391 tcg_gen_movi_tl(cpu_T
[0], 0);
3392 tcg_gen_st32_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,mmx_t0
.MMX_L(1)));
3393 op1_offset
= offsetof(CPUX86State
,mmx_t0
);
3395 sse_fn_epp
= sse_op_table2
[((b
- 1) & 3) * 8 +
3396 (((modrm
>> 3)) & 7)][b1
];
3401 rm
= (modrm
& 7) | REX_B(s
);
3402 op2_offset
= offsetof(CPUX86State
,xmm_regs
[rm
]);
3405 op2_offset
= offsetof(CPUX86State
,fpregs
[rm
].mmx
);
3407 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
, op2_offset
);
3408 tcg_gen_addi_ptr(cpu_ptr1
, cpu_env
, op1_offset
);
3409 sse_fn_epp(cpu_env
, cpu_ptr0
, cpu_ptr1
);
3411 case 0x050: /* movmskps */
3412 rm
= (modrm
& 7) | REX_B(s
);
3413 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
,
3414 offsetof(CPUX86State
,xmm_regs
[rm
]));
3415 gen_helper_movmskps(cpu_tmp2_i32
, cpu_env
, cpu_ptr0
);
3416 tcg_gen_extu_i32_tl(cpu_regs
[reg
], cpu_tmp2_i32
);
3418 case 0x150: /* movmskpd */
3419 rm
= (modrm
& 7) | REX_B(s
);
3420 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
,
3421 offsetof(CPUX86State
,xmm_regs
[rm
]));
3422 gen_helper_movmskpd(cpu_tmp2_i32
, cpu_env
, cpu_ptr0
);
3423 tcg_gen_extu_i32_tl(cpu_regs
[reg
], cpu_tmp2_i32
);
3425 case 0x02a: /* cvtpi2ps */
3426 case 0x12a: /* cvtpi2pd */
3427 gen_helper_enter_mmx(cpu_env
);
3429 gen_lea_modrm(env
, s
, modrm
);
3430 op2_offset
= offsetof(CPUX86State
,mmx_t0
);
3431 gen_ldq_env_A0(s
, op2_offset
);
3434 op2_offset
= offsetof(CPUX86State
,fpregs
[rm
].mmx
);
3436 op1_offset
= offsetof(CPUX86State
,xmm_regs
[reg
]);
3437 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
, op1_offset
);
3438 tcg_gen_addi_ptr(cpu_ptr1
, cpu_env
, op2_offset
);
3441 gen_helper_cvtpi2ps(cpu_env
, cpu_ptr0
, cpu_ptr1
);
3445 gen_helper_cvtpi2pd(cpu_env
, cpu_ptr0
, cpu_ptr1
);
3449 case 0x22a: /* cvtsi2ss */
3450 case 0x32a: /* cvtsi2sd */
3451 ot
= mo_64_32(s
->dflag
);
3452 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
3453 op1_offset
= offsetof(CPUX86State
,xmm_regs
[reg
]);
3454 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
, op1_offset
);
3456 SSEFunc_0_epi sse_fn_epi
= sse_op_table3ai
[(b
>> 8) & 1];
3457 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
3458 sse_fn_epi(cpu_env
, cpu_ptr0
, cpu_tmp2_i32
);
3460 #ifdef TARGET_X86_64
3461 SSEFunc_0_epl sse_fn_epl
= sse_op_table3aq
[(b
>> 8) & 1];
3462 sse_fn_epl(cpu_env
, cpu_ptr0
, cpu_T
[0]);
3468 case 0x02c: /* cvttps2pi */
3469 case 0x12c: /* cvttpd2pi */
3470 case 0x02d: /* cvtps2pi */
3471 case 0x12d: /* cvtpd2pi */
3472 gen_helper_enter_mmx(cpu_env
);
3474 gen_lea_modrm(env
, s
, modrm
);
3475 op2_offset
= offsetof(CPUX86State
,xmm_t0
);
3476 gen_ldo_env_A0(s
, op2_offset
);
3478 rm
= (modrm
& 7) | REX_B(s
);
3479 op2_offset
= offsetof(CPUX86State
,xmm_regs
[rm
]);
3481 op1_offset
= offsetof(CPUX86State
,fpregs
[reg
& 7].mmx
);
3482 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
, op1_offset
);
3483 tcg_gen_addi_ptr(cpu_ptr1
, cpu_env
, op2_offset
);
3486 gen_helper_cvttps2pi(cpu_env
, cpu_ptr0
, cpu_ptr1
);
3489 gen_helper_cvttpd2pi(cpu_env
, cpu_ptr0
, cpu_ptr1
);
3492 gen_helper_cvtps2pi(cpu_env
, cpu_ptr0
, cpu_ptr1
);
3495 gen_helper_cvtpd2pi(cpu_env
, cpu_ptr0
, cpu_ptr1
);
3499 case 0x22c: /* cvttss2si */
3500 case 0x32c: /* cvttsd2si */
3501 case 0x22d: /* cvtss2si */
3502 case 0x32d: /* cvtsd2si */
3503 ot
= mo_64_32(s
->dflag
);
3505 gen_lea_modrm(env
, s
, modrm
);
3507 gen_ldq_env_A0(s
, offsetof(CPUX86State
, xmm_t0
.XMM_Q(0)));
3509 gen_op_ld_v(s
, MO_32
, cpu_T
[0], cpu_A0
);
3510 tcg_gen_st32_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,xmm_t0
.XMM_L(0)));
3512 op2_offset
= offsetof(CPUX86State
,xmm_t0
);
3514 rm
= (modrm
& 7) | REX_B(s
);
3515 op2_offset
= offsetof(CPUX86State
,xmm_regs
[rm
]);
3517 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
, op2_offset
);
3519 SSEFunc_i_ep sse_fn_i_ep
=
3520 sse_op_table3bi
[((b
>> 7) & 2) | (b
& 1)];
3521 sse_fn_i_ep(cpu_tmp2_i32
, cpu_env
, cpu_ptr0
);
3522 tcg_gen_extu_i32_tl(cpu_T
[0], cpu_tmp2_i32
);
3524 #ifdef TARGET_X86_64
3525 SSEFunc_l_ep sse_fn_l_ep
=
3526 sse_op_table3bq
[((b
>> 7) & 2) | (b
& 1)];
3527 sse_fn_l_ep(cpu_T
[0], cpu_env
, cpu_ptr0
);
3532 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
3534 case 0xc4: /* pinsrw */
3537 gen_ldst_modrm(env
, s
, modrm
, MO_16
, OR_TMP0
, 0);
3538 val
= cpu_ldub_code(env
, s
->pc
++);
3541 tcg_gen_st16_tl(cpu_T
[0], cpu_env
,
3542 offsetof(CPUX86State
,xmm_regs
[reg
].XMM_W(val
)));
3545 tcg_gen_st16_tl(cpu_T
[0], cpu_env
,
3546 offsetof(CPUX86State
,fpregs
[reg
].mmx
.MMX_W(val
)));
3549 case 0xc5: /* pextrw */
3553 ot
= mo_64_32(s
->dflag
);
3554 val
= cpu_ldub_code(env
, s
->pc
++);
3557 rm
= (modrm
& 7) | REX_B(s
);
3558 tcg_gen_ld16u_tl(cpu_T
[0], cpu_env
,
3559 offsetof(CPUX86State
,xmm_regs
[rm
].XMM_W(val
)));
3563 tcg_gen_ld16u_tl(cpu_T
[0], cpu_env
,
3564 offsetof(CPUX86State
,fpregs
[rm
].mmx
.MMX_W(val
)));
3566 reg
= ((modrm
>> 3) & 7) | rex_r
;
3567 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
3569 case 0x1d6: /* movq ea, xmm */
3571 gen_lea_modrm(env
, s
, modrm
);
3572 gen_stq_env_A0(s
, offsetof(CPUX86State
,
3573 xmm_regs
[reg
].XMM_Q(0)));
3575 rm
= (modrm
& 7) | REX_B(s
);
3576 gen_op_movq(offsetof(CPUX86State
,xmm_regs
[rm
].XMM_Q(0)),
3577 offsetof(CPUX86State
,xmm_regs
[reg
].XMM_Q(0)));
3578 gen_op_movq_env_0(offsetof(CPUX86State
,xmm_regs
[rm
].XMM_Q(1)));
3581 case 0x2d6: /* movq2dq */
3582 gen_helper_enter_mmx(cpu_env
);
3584 gen_op_movq(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_Q(0)),
3585 offsetof(CPUX86State
,fpregs
[rm
].mmx
));
3586 gen_op_movq_env_0(offsetof(CPUX86State
,xmm_regs
[reg
].XMM_Q(1)));
3588 case 0x3d6: /* movdq2q */
3589 gen_helper_enter_mmx(cpu_env
);
3590 rm
= (modrm
& 7) | REX_B(s
);
3591 gen_op_movq(offsetof(CPUX86State
,fpregs
[reg
& 7].mmx
),
3592 offsetof(CPUX86State
,xmm_regs
[rm
].XMM_Q(0)));
3594 case 0xd7: /* pmovmskb */
3599 rm
= (modrm
& 7) | REX_B(s
);
3600 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
, offsetof(CPUX86State
,xmm_regs
[rm
]));
3601 gen_helper_pmovmskb_xmm(cpu_tmp2_i32
, cpu_env
, cpu_ptr0
);
3604 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
, offsetof(CPUX86State
,fpregs
[rm
].mmx
));
3605 gen_helper_pmovmskb_mmx(cpu_tmp2_i32
, cpu_env
, cpu_ptr0
);
3607 reg
= ((modrm
>> 3) & 7) | rex_r
;
3608 tcg_gen_extu_i32_tl(cpu_regs
[reg
], cpu_tmp2_i32
);
3614 if ((b
& 0xf0) == 0xf0) {
3617 modrm
= cpu_ldub_code(env
, s
->pc
++);
3619 reg
= ((modrm
>> 3) & 7) | rex_r
;
3620 mod
= (modrm
>> 6) & 3;
3625 sse_fn_epp
= sse_op_table6
[b
].op
[b1
];
3629 if (!(s
->cpuid_ext_features
& sse_op_table6
[b
].ext_mask
))
3633 op1_offset
= offsetof(CPUX86State
,xmm_regs
[reg
]);
3635 op2_offset
= offsetof(CPUX86State
,xmm_regs
[rm
| REX_B(s
)]);
3637 op2_offset
= offsetof(CPUX86State
,xmm_t0
);
3638 gen_lea_modrm(env
, s
, modrm
);
3640 case 0x20: case 0x30: /* pmovsxbw, pmovzxbw */
3641 case 0x23: case 0x33: /* pmovsxwd, pmovzxwd */
3642 case 0x25: case 0x35: /* pmovsxdq, pmovzxdq */
3643 gen_ldq_env_A0(s
, op2_offset
+
3644 offsetof(XMMReg
, XMM_Q(0)));
3646 case 0x21: case 0x31: /* pmovsxbd, pmovzxbd */
3647 case 0x24: case 0x34: /* pmovsxwq, pmovzxwq */
3648 tcg_gen_qemu_ld_i32(cpu_tmp2_i32
, cpu_A0
,
3649 s
->mem_index
, MO_LEUL
);
3650 tcg_gen_st_i32(cpu_tmp2_i32
, cpu_env
, op2_offset
+
3651 offsetof(XMMReg
, XMM_L(0)));
3653 case 0x22: case 0x32: /* pmovsxbq, pmovzxbq */
3654 tcg_gen_qemu_ld_tl(cpu_tmp0
, cpu_A0
,
3655 s
->mem_index
, MO_LEUW
);
3656 tcg_gen_st16_tl(cpu_tmp0
, cpu_env
, op2_offset
+
3657 offsetof(XMMReg
, XMM_W(0)));
3659 case 0x2a: /* movntqda */
3660 gen_ldo_env_A0(s
, op1_offset
);
3663 gen_ldo_env_A0(s
, op2_offset
);
3667 op1_offset
= offsetof(CPUX86State
,fpregs
[reg
].mmx
);
3669 op2_offset
= offsetof(CPUX86State
,fpregs
[rm
].mmx
);
3671 op2_offset
= offsetof(CPUX86State
,mmx_t0
);
3672 gen_lea_modrm(env
, s
, modrm
);
3673 gen_ldq_env_A0(s
, op2_offset
);
3676 if (sse_fn_epp
== SSE_SPECIAL
) {
3680 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
, op1_offset
);
3681 tcg_gen_addi_ptr(cpu_ptr1
, cpu_env
, op2_offset
);
3682 sse_fn_epp(cpu_env
, cpu_ptr0
, cpu_ptr1
);
3685 set_cc_op(s
, CC_OP_EFLAGS
);
3692 /* Various integer extensions at 0f 38 f[0-f]. */
3693 b
= modrm
| (b1
<< 8);
3694 modrm
= cpu_ldub_code(env
, s
->pc
++);
3695 reg
= ((modrm
>> 3) & 7) | rex_r
;
3698 case 0x3f0: /* crc32 Gd,Eb */
3699 case 0x3f1: /* crc32 Gd,Ey */
3701 if (!(s
->cpuid_ext_features
& CPUID_EXT_SSE42
)) {
3704 if ((b
& 0xff) == 0xf0) {
3706 } else if (s
->dflag
!= MO_64
) {
3707 ot
= (s
->prefix
& PREFIX_DATA
? MO_16
: MO_32
);
3712 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_regs
[reg
]);
3713 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
3714 gen_helper_crc32(cpu_T
[0], cpu_tmp2_i32
,
3715 cpu_T
[0], tcg_const_i32(8 << ot
));
3717 ot
= mo_64_32(s
->dflag
);
3718 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
3721 case 0x1f0: /* crc32 or movbe */
3723 /* For these insns, the f3 prefix is supposed to have priority
3724 over the 66 prefix, but that's not what we implement above
3726 if (s
->prefix
& PREFIX_REPNZ
) {
3730 case 0x0f0: /* movbe Gy,My */
3731 case 0x0f1: /* movbe My,Gy */
3732 if (!(s
->cpuid_ext_features
& CPUID_EXT_MOVBE
)) {
3735 if (s
->dflag
!= MO_64
) {
3736 ot
= (s
->prefix
& PREFIX_DATA
? MO_16
: MO_32
);
3741 gen_lea_modrm(env
, s
, modrm
);
3743 tcg_gen_qemu_ld_tl(cpu_T
[0], cpu_A0
,
3744 s
->mem_index
, ot
| MO_BE
);
3745 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
3747 tcg_gen_qemu_st_tl(cpu_regs
[reg
], cpu_A0
,
3748 s
->mem_index
, ot
| MO_BE
);
3752 case 0x0f2: /* andn Gy, By, Ey */
3753 if (!(s
->cpuid_7_0_ebx_features
& CPUID_7_0_EBX_BMI1
)
3754 || !(s
->prefix
& PREFIX_VEX
)
3758 ot
= mo_64_32(s
->dflag
);
3759 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
3760 tcg_gen_andc_tl(cpu_T
[0], cpu_regs
[s
->vex_v
], cpu_T
[0]);
3761 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
3762 gen_op_update1_cc();
3763 set_cc_op(s
, CC_OP_LOGICB
+ ot
);
3766 case 0x0f7: /* bextr Gy, Ey, By */
3767 if (!(s
->cpuid_7_0_ebx_features
& CPUID_7_0_EBX_BMI1
)
3768 || !(s
->prefix
& PREFIX_VEX
)
3772 ot
= mo_64_32(s
->dflag
);
3776 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
3777 /* Extract START, and shift the operand.
3778 Shifts larger than operand size get zeros. */
3779 tcg_gen_ext8u_tl(cpu_A0
, cpu_regs
[s
->vex_v
]);
3780 tcg_gen_shr_tl(cpu_T
[0], cpu_T
[0], cpu_A0
);
3782 bound
= tcg_const_tl(ot
== MO_64
? 63 : 31);
3783 zero
= tcg_const_tl(0);
3784 tcg_gen_movcond_tl(TCG_COND_LEU
, cpu_T
[0], cpu_A0
, bound
,
3786 tcg_temp_free(zero
);
3788 /* Extract the LEN into a mask. Lengths larger than
3789 operand size get all ones. */
3790 tcg_gen_shri_tl(cpu_A0
, cpu_regs
[s
->vex_v
], 8);
3791 tcg_gen_ext8u_tl(cpu_A0
, cpu_A0
);
3792 tcg_gen_movcond_tl(TCG_COND_LEU
, cpu_A0
, cpu_A0
, bound
,
3794 tcg_temp_free(bound
);
3795 tcg_gen_movi_tl(cpu_T
[1], 1);
3796 tcg_gen_shl_tl(cpu_T
[1], cpu_T
[1], cpu_A0
);
3797 tcg_gen_subi_tl(cpu_T
[1], cpu_T
[1], 1);
3798 tcg_gen_and_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
3800 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
3801 gen_op_update1_cc();
3802 set_cc_op(s
, CC_OP_LOGICB
+ ot
);
3806 case 0x0f5: /* bzhi Gy, Ey, By */
3807 if (!(s
->cpuid_7_0_ebx_features
& CPUID_7_0_EBX_BMI2
)
3808 || !(s
->prefix
& PREFIX_VEX
)
3812 ot
= mo_64_32(s
->dflag
);
3813 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
3814 tcg_gen_ext8u_tl(cpu_T
[1], cpu_regs
[s
->vex_v
]);
3816 TCGv bound
= tcg_const_tl(ot
== MO_64
? 63 : 31);
3817 /* Note that since we're using BMILG (in order to get O
3818 cleared) we need to store the inverse into C. */
3819 tcg_gen_setcond_tl(TCG_COND_LT
, cpu_cc_src
,
3821 tcg_gen_movcond_tl(TCG_COND_GT
, cpu_T
[1], cpu_T
[1],
3822 bound
, bound
, cpu_T
[1]);
3823 tcg_temp_free(bound
);
3825 tcg_gen_movi_tl(cpu_A0
, -1);
3826 tcg_gen_shl_tl(cpu_A0
, cpu_A0
, cpu_T
[1]);
3827 tcg_gen_andc_tl(cpu_T
[0], cpu_T
[0], cpu_A0
);
3828 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
3829 gen_op_update1_cc();
3830 set_cc_op(s
, CC_OP_BMILGB
+ ot
);
3833 case 0x3f6: /* mulx By, Gy, rdx, Ey */
3834 if (!(s
->cpuid_7_0_ebx_features
& CPUID_7_0_EBX_BMI2
)
3835 || !(s
->prefix
& PREFIX_VEX
)
3839 ot
= mo_64_32(s
->dflag
);
3840 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
3843 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
3844 tcg_gen_trunc_tl_i32(cpu_tmp3_i32
, cpu_regs
[R_EDX
]);
3845 tcg_gen_mulu2_i32(cpu_tmp2_i32
, cpu_tmp3_i32
,
3846 cpu_tmp2_i32
, cpu_tmp3_i32
);
3847 tcg_gen_extu_i32_tl(cpu_regs
[s
->vex_v
], cpu_tmp2_i32
);
3848 tcg_gen_extu_i32_tl(cpu_regs
[reg
], cpu_tmp3_i32
);
3850 #ifdef TARGET_X86_64
3852 tcg_gen_mulu2_i64(cpu_regs
[s
->vex_v
], cpu_regs
[reg
],
3853 cpu_T
[0], cpu_regs
[R_EDX
]);
3859 case 0x3f5: /* pdep Gy, By, Ey */
3860 if (!(s
->cpuid_7_0_ebx_features
& CPUID_7_0_EBX_BMI2
)
3861 || !(s
->prefix
& PREFIX_VEX
)
3865 ot
= mo_64_32(s
->dflag
);
3866 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
3867 /* Note that by zero-extending the mask operand, we
3868 automatically handle zero-extending the result. */
3870 tcg_gen_mov_tl(cpu_T
[1], cpu_regs
[s
->vex_v
]);
3872 tcg_gen_ext32u_tl(cpu_T
[1], cpu_regs
[s
->vex_v
]);
3874 gen_helper_pdep(cpu_regs
[reg
], cpu_T
[0], cpu_T
[1]);
3877 case 0x2f5: /* pext Gy, By, Ey */
3878 if (!(s
->cpuid_7_0_ebx_features
& CPUID_7_0_EBX_BMI2
)
3879 || !(s
->prefix
& PREFIX_VEX
)
3883 ot
= mo_64_32(s
->dflag
);
3884 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
3885 /* Note that by zero-extending the mask operand, we
3886 automatically handle zero-extending the result. */
3888 tcg_gen_mov_tl(cpu_T
[1], cpu_regs
[s
->vex_v
]);
3890 tcg_gen_ext32u_tl(cpu_T
[1], cpu_regs
[s
->vex_v
]);
3892 gen_helper_pext(cpu_regs
[reg
], cpu_T
[0], cpu_T
[1]);
3895 case 0x1f6: /* adcx Gy, Ey */
3896 case 0x2f6: /* adox Gy, Ey */
3897 if (!(s
->cpuid_7_0_ebx_features
& CPUID_7_0_EBX_ADX
)) {
3900 TCGv carry_in
, carry_out
, zero
;
3903 ot
= mo_64_32(s
->dflag
);
3904 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
3906 /* Re-use the carry-out from a previous round. */
3907 TCGV_UNUSED(carry_in
);
3908 carry_out
= (b
== 0x1f6 ? cpu_cc_dst
: cpu_cc_src2
);
3912 carry_in
= cpu_cc_dst
;
3913 end_op
= CC_OP_ADCX
;
3915 end_op
= CC_OP_ADCOX
;
3920 end_op
= CC_OP_ADCOX
;
3922 carry_in
= cpu_cc_src2
;
3923 end_op
= CC_OP_ADOX
;
3927 end_op
= CC_OP_ADCOX
;
3928 carry_in
= carry_out
;
3931 end_op
= (b
== 0x1f6 ? CC_OP_ADCX
: CC_OP_ADOX
);
3934 /* If we can't reuse carry-out, get it out of EFLAGS. */
3935 if (TCGV_IS_UNUSED(carry_in
)) {
3936 if (s
->cc_op
!= CC_OP_ADCX
&& s
->cc_op
!= CC_OP_ADOX
) {
3937 gen_compute_eflags(s
);
3939 carry_in
= cpu_tmp0
;
3940 tcg_gen_shri_tl(carry_in
, cpu_cc_src
,
3941 ctz32(b
== 0x1f6 ? CC_C
: CC_O
));
3942 tcg_gen_andi_tl(carry_in
, carry_in
, 1);
3946 #ifdef TARGET_X86_64
3948 /* If we know TL is 64-bit, and we want a 32-bit
3949 result, just do everything in 64-bit arithmetic. */
3950 tcg_gen_ext32u_i64(cpu_regs
[reg
], cpu_regs
[reg
]);
3951 tcg_gen_ext32u_i64(cpu_T
[0], cpu_T
[0]);
3952 tcg_gen_add_i64(cpu_T
[0], cpu_T
[0], cpu_regs
[reg
]);
3953 tcg_gen_add_i64(cpu_T
[0], cpu_T
[0], carry_in
);
3954 tcg_gen_ext32u_i64(cpu_regs
[reg
], cpu_T
[0]);
3955 tcg_gen_shri_i64(carry_out
, cpu_T
[0], 32);
3959 /* Otherwise compute the carry-out in two steps. */
3960 zero
= tcg_const_tl(0);
3961 tcg_gen_add2_tl(cpu_T
[0], carry_out
,
3964 tcg_gen_add2_tl(cpu_regs
[reg
], carry_out
,
3965 cpu_regs
[reg
], carry_out
,
3967 tcg_temp_free(zero
);
3970 set_cc_op(s
, end_op
);
3974 case 0x1f7: /* shlx Gy, Ey, By */
3975 case 0x2f7: /* sarx Gy, Ey, By */
3976 case 0x3f7: /* shrx Gy, Ey, By */
3977 if (!(s
->cpuid_7_0_ebx_features
& CPUID_7_0_EBX_BMI2
)
3978 || !(s
->prefix
& PREFIX_VEX
)
3982 ot
= mo_64_32(s
->dflag
);
3983 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
3985 tcg_gen_andi_tl(cpu_T
[1], cpu_regs
[s
->vex_v
], 63);
3987 tcg_gen_andi_tl(cpu_T
[1], cpu_regs
[s
->vex_v
], 31);
3990 tcg_gen_shl_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
3991 } else if (b
== 0x2f7) {
3993 tcg_gen_ext32s_tl(cpu_T
[0], cpu_T
[0]);
3995 tcg_gen_sar_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
3998 tcg_gen_ext32u_tl(cpu_T
[0], cpu_T
[0]);
4000 tcg_gen_shr_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
4002 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
4008 case 0x3f3: /* Group 17 */
4009 if (!(s
->cpuid_7_0_ebx_features
& CPUID_7_0_EBX_BMI1
)
4010 || !(s
->prefix
& PREFIX_VEX
)
4014 ot
= mo_64_32(s
->dflag
);
4015 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
4018 case 1: /* blsr By,Ey */
4019 tcg_gen_neg_tl(cpu_T
[1], cpu_T
[0]);
4020 tcg_gen_and_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
4021 gen_op_mov_reg_v(ot
, s
->vex_v
, cpu_T
[0]);
4022 gen_op_update2_cc();
4023 set_cc_op(s
, CC_OP_BMILGB
+ ot
);
4026 case 2: /* blsmsk By,Ey */
4027 tcg_gen_mov_tl(cpu_cc_src
, cpu_T
[0]);
4028 tcg_gen_subi_tl(cpu_T
[0], cpu_T
[0], 1);
4029 tcg_gen_xor_tl(cpu_T
[0], cpu_T
[0], cpu_cc_src
);
4030 tcg_gen_mov_tl(cpu_cc_dst
, cpu_T
[0]);
4031 set_cc_op(s
, CC_OP_BMILGB
+ ot
);
4034 case 3: /* blsi By, Ey */
4035 tcg_gen_mov_tl(cpu_cc_src
, cpu_T
[0]);
4036 tcg_gen_subi_tl(cpu_T
[0], cpu_T
[0], 1);
4037 tcg_gen_and_tl(cpu_T
[0], cpu_T
[0], cpu_cc_src
);
4038 tcg_gen_mov_tl(cpu_cc_dst
, cpu_T
[0]);
4039 set_cc_op(s
, CC_OP_BMILGB
+ ot
);
4055 modrm
= cpu_ldub_code(env
, s
->pc
++);
4057 reg
= ((modrm
>> 3) & 7) | rex_r
;
4058 mod
= (modrm
>> 6) & 3;
4063 sse_fn_eppi
= sse_op_table7
[b
].op
[b1
];
4067 if (!(s
->cpuid_ext_features
& sse_op_table7
[b
].ext_mask
))
4070 if (sse_fn_eppi
== SSE_SPECIAL
) {
4071 ot
= mo_64_32(s
->dflag
);
4072 rm
= (modrm
& 7) | REX_B(s
);
4074 gen_lea_modrm(env
, s
, modrm
);
4075 reg
= ((modrm
>> 3) & 7) | rex_r
;
4076 val
= cpu_ldub_code(env
, s
->pc
++);
4078 case 0x14: /* pextrb */
4079 tcg_gen_ld8u_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,
4080 xmm_regs
[reg
].XMM_B(val
& 15)));
4082 gen_op_mov_reg_v(ot
, rm
, cpu_T
[0]);
4084 tcg_gen_qemu_st_tl(cpu_T
[0], cpu_A0
,
4085 s
->mem_index
, MO_UB
);
4088 case 0x15: /* pextrw */
4089 tcg_gen_ld16u_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,
4090 xmm_regs
[reg
].XMM_W(val
& 7)));
4092 gen_op_mov_reg_v(ot
, rm
, cpu_T
[0]);
4094 tcg_gen_qemu_st_tl(cpu_T
[0], cpu_A0
,
4095 s
->mem_index
, MO_LEUW
);
4099 if (ot
== MO_32
) { /* pextrd */
4100 tcg_gen_ld_i32(cpu_tmp2_i32
, cpu_env
,
4101 offsetof(CPUX86State
,
4102 xmm_regs
[reg
].XMM_L(val
& 3)));
4104 tcg_gen_extu_i32_tl(cpu_regs
[rm
], cpu_tmp2_i32
);
4106 tcg_gen_qemu_st_i32(cpu_tmp2_i32
, cpu_A0
,
4107 s
->mem_index
, MO_LEUL
);
4109 } else { /* pextrq */
4110 #ifdef TARGET_X86_64
4111 tcg_gen_ld_i64(cpu_tmp1_i64
, cpu_env
,
4112 offsetof(CPUX86State
,
4113 xmm_regs
[reg
].XMM_Q(val
& 1)));
4115 tcg_gen_mov_i64(cpu_regs
[rm
], cpu_tmp1_i64
);
4117 tcg_gen_qemu_st_i64(cpu_tmp1_i64
, cpu_A0
,
4118 s
->mem_index
, MO_LEQ
);
4125 case 0x17: /* extractps */
4126 tcg_gen_ld32u_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,
4127 xmm_regs
[reg
].XMM_L(val
& 3)));
4129 gen_op_mov_reg_v(ot
, rm
, cpu_T
[0]);
4131 tcg_gen_qemu_st_tl(cpu_T
[0], cpu_A0
,
4132 s
->mem_index
, MO_LEUL
);
4135 case 0x20: /* pinsrb */
4137 gen_op_mov_v_reg(MO_32
, cpu_T
[0], rm
);
4139 tcg_gen_qemu_ld_tl(cpu_T
[0], cpu_A0
,
4140 s
->mem_index
, MO_UB
);
4142 tcg_gen_st8_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,
4143 xmm_regs
[reg
].XMM_B(val
& 15)));
4145 case 0x21: /* insertps */
4147 tcg_gen_ld_i32(cpu_tmp2_i32
, cpu_env
,
4148 offsetof(CPUX86State
,xmm_regs
[rm
]
4149 .XMM_L((val
>> 6) & 3)));
4151 tcg_gen_qemu_ld_i32(cpu_tmp2_i32
, cpu_A0
,
4152 s
->mem_index
, MO_LEUL
);
4154 tcg_gen_st_i32(cpu_tmp2_i32
, cpu_env
,
4155 offsetof(CPUX86State
,xmm_regs
[reg
]
4156 .XMM_L((val
>> 4) & 3)));
4158 tcg_gen_st_i32(tcg_const_i32(0 /*float32_zero*/),
4159 cpu_env
, offsetof(CPUX86State
,
4160 xmm_regs
[reg
].XMM_L(0)));
4162 tcg_gen_st_i32(tcg_const_i32(0 /*float32_zero*/),
4163 cpu_env
, offsetof(CPUX86State
,
4164 xmm_regs
[reg
].XMM_L(1)));
4166 tcg_gen_st_i32(tcg_const_i32(0 /*float32_zero*/),
4167 cpu_env
, offsetof(CPUX86State
,
4168 xmm_regs
[reg
].XMM_L(2)));
4170 tcg_gen_st_i32(tcg_const_i32(0 /*float32_zero*/),
4171 cpu_env
, offsetof(CPUX86State
,
4172 xmm_regs
[reg
].XMM_L(3)));
4175 if (ot
== MO_32
) { /* pinsrd */
4177 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_regs
[rm
]);
4179 tcg_gen_qemu_ld_i32(cpu_tmp2_i32
, cpu_A0
,
4180 s
->mem_index
, MO_LEUL
);
4182 tcg_gen_st_i32(cpu_tmp2_i32
, cpu_env
,
4183 offsetof(CPUX86State
,
4184 xmm_regs
[reg
].XMM_L(val
& 3)));
4185 } else { /* pinsrq */
4186 #ifdef TARGET_X86_64
4188 gen_op_mov_v_reg(ot
, cpu_tmp1_i64
, rm
);
4190 tcg_gen_qemu_ld_i64(cpu_tmp1_i64
, cpu_A0
,
4191 s
->mem_index
, MO_LEQ
);
4193 tcg_gen_st_i64(cpu_tmp1_i64
, cpu_env
,
4194 offsetof(CPUX86State
,
4195 xmm_regs
[reg
].XMM_Q(val
& 1)));
4206 op1_offset
= offsetof(CPUX86State
,xmm_regs
[reg
]);
4208 op2_offset
= offsetof(CPUX86State
,xmm_regs
[rm
| REX_B(s
)]);
4210 op2_offset
= offsetof(CPUX86State
,xmm_t0
);
4211 gen_lea_modrm(env
, s
, modrm
);
4212 gen_ldo_env_A0(s
, op2_offset
);
4215 op1_offset
= offsetof(CPUX86State
,fpregs
[reg
].mmx
);
4217 op2_offset
= offsetof(CPUX86State
,fpregs
[rm
].mmx
);
4219 op2_offset
= offsetof(CPUX86State
,mmx_t0
);
4220 gen_lea_modrm(env
, s
, modrm
);
4221 gen_ldq_env_A0(s
, op2_offset
);
4224 val
= cpu_ldub_code(env
, s
->pc
++);
4226 if ((b
& 0xfc) == 0x60) { /* pcmpXstrX */
4227 set_cc_op(s
, CC_OP_EFLAGS
);
4229 if (s
->dflag
== MO_64
) {
4230 /* The helper must use entire 64-bit gp registers */
4235 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
, op1_offset
);
4236 tcg_gen_addi_ptr(cpu_ptr1
, cpu_env
, op2_offset
);
4237 sse_fn_eppi(cpu_env
, cpu_ptr0
, cpu_ptr1
, tcg_const_i32(val
));
4241 /* Various integer extensions at 0f 3a f[0-f]. */
4242 b
= modrm
| (b1
<< 8);
4243 modrm
= cpu_ldub_code(env
, s
->pc
++);
4244 reg
= ((modrm
>> 3) & 7) | rex_r
;
4247 case 0x3f0: /* rorx Gy,Ey, Ib */
4248 if (!(s
->cpuid_7_0_ebx_features
& CPUID_7_0_EBX_BMI2
)
4249 || !(s
->prefix
& PREFIX_VEX
)
4253 ot
= mo_64_32(s
->dflag
);
4254 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
4255 b
= cpu_ldub_code(env
, s
->pc
++);
4257 tcg_gen_rotri_tl(cpu_T
[0], cpu_T
[0], b
& 63);
4259 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
4260 tcg_gen_rotri_i32(cpu_tmp2_i32
, cpu_tmp2_i32
, b
& 31);
4261 tcg_gen_extu_i32_tl(cpu_T
[0], cpu_tmp2_i32
);
4263 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
4275 /* generic MMX or SSE operation */
4277 case 0x70: /* pshufx insn */
4278 case 0xc6: /* pshufx insn */
4279 case 0xc2: /* compare insns */
4286 op1_offset
= offsetof(CPUX86State
,xmm_regs
[reg
]);
4290 gen_lea_modrm(env
, s
, modrm
);
4291 op2_offset
= offsetof(CPUX86State
,xmm_t0
);
4297 /* Most sse scalar operations. */
4300 } else if (b1
== 3) {
4305 case 0x2e: /* ucomis[sd] */
4306 case 0x2f: /* comis[sd] */
4318 gen_op_ld_v(s
, MO_32
, cpu_T
[0], cpu_A0
);
4319 tcg_gen_st32_tl(cpu_T
[0], cpu_env
,
4320 offsetof(CPUX86State
,xmm_t0
.XMM_L(0)));
4324 gen_ldq_env_A0(s
, offsetof(CPUX86State
, xmm_t0
.XMM_D(0)));
4327 /* 128 bit access */
4328 gen_ldo_env_A0(s
, op2_offset
);
4332 rm
= (modrm
& 7) | REX_B(s
);
4333 op2_offset
= offsetof(CPUX86State
,xmm_regs
[rm
]);
4336 op1_offset
= offsetof(CPUX86State
,fpregs
[reg
].mmx
);
4338 gen_lea_modrm(env
, s
, modrm
);
4339 op2_offset
= offsetof(CPUX86State
,mmx_t0
);
4340 gen_ldq_env_A0(s
, op2_offset
);
4343 op2_offset
= offsetof(CPUX86State
,fpregs
[rm
].mmx
);
4347 case 0x0f: /* 3DNow! data insns */
4348 if (!(s
->cpuid_ext2_features
& CPUID_EXT2_3DNOW
))
4350 val
= cpu_ldub_code(env
, s
->pc
++);
4351 sse_fn_epp
= sse_op_table5
[val
];
4355 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
, op1_offset
);
4356 tcg_gen_addi_ptr(cpu_ptr1
, cpu_env
, op2_offset
);
4357 sse_fn_epp(cpu_env
, cpu_ptr0
, cpu_ptr1
);
4359 case 0x70: /* pshufx insn */
4360 case 0xc6: /* pshufx insn */
4361 val
= cpu_ldub_code(env
, s
->pc
++);
4362 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
, op1_offset
);
4363 tcg_gen_addi_ptr(cpu_ptr1
, cpu_env
, op2_offset
);
4364 /* XXX: introduce a new table? */
4365 sse_fn_ppi
= (SSEFunc_0_ppi
)sse_fn_epp
;
4366 sse_fn_ppi(cpu_ptr0
, cpu_ptr1
, tcg_const_i32(val
));
4370 val
= cpu_ldub_code(env
, s
->pc
++);
4373 sse_fn_epp
= sse_op_table4
[val
][b1
];
4375 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
, op1_offset
);
4376 tcg_gen_addi_ptr(cpu_ptr1
, cpu_env
, op2_offset
);
4377 sse_fn_epp(cpu_env
, cpu_ptr0
, cpu_ptr1
);
4380 /* maskmov : we must prepare A0 */
4383 tcg_gen_mov_tl(cpu_A0
, cpu_regs
[R_EDI
]);
4384 gen_extu(s
->aflag
, cpu_A0
);
4385 gen_add_A0_ds_seg(s
);
4387 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
, op1_offset
);
4388 tcg_gen_addi_ptr(cpu_ptr1
, cpu_env
, op2_offset
);
4389 /* XXX: introduce a new table? */
4390 sse_fn_eppt
= (SSEFunc_0_eppt
)sse_fn_epp
;
4391 sse_fn_eppt(cpu_env
, cpu_ptr0
, cpu_ptr1
, cpu_A0
);
4394 tcg_gen_addi_ptr(cpu_ptr0
, cpu_env
, op1_offset
);
4395 tcg_gen_addi_ptr(cpu_ptr1
, cpu_env
, op2_offset
);
4396 sse_fn_epp(cpu_env
, cpu_ptr0
, cpu_ptr1
);
4399 if (b
== 0x2e || b
== 0x2f) {
4400 set_cc_op(s
, CC_OP_EFLAGS
);
4405 /* convert one instruction. s->is_jmp is set if the translation must
4406 be stopped. Return the next pc value */
4407 static target_ulong
disas_insn(CPUX86State
*env
, DisasContext
*s
,
4408 target_ulong pc_start
)
4412 TCGMemOp ot
, aflag
, dflag
;
4413 int modrm
, reg
, rm
, mod
, op
, opreg
, val
;
4414 target_ulong next_eip
, tval
;
4417 if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP
| CPU_LOG_TB_OP_OPT
))) {
4418 tcg_gen_debug_insn_start(pc_start
);
4425 #ifdef TARGET_X86_64
4430 s
->rip_offset
= 0; /* for relative ip address */
4434 b
= cpu_ldub_code(env
, s
->pc
);
4436 /* Collect prefixes. */
4439 prefixes
|= PREFIX_REPZ
;
4442 prefixes
|= PREFIX_REPNZ
;
4445 prefixes
|= PREFIX_LOCK
;
4466 prefixes
|= PREFIX_DATA
;
4469 prefixes
|= PREFIX_ADR
;
4471 #ifdef TARGET_X86_64
4475 rex_w
= (b
>> 3) & 1;
4476 rex_r
= (b
& 0x4) << 1;
4477 s
->rex_x
= (b
& 0x2) << 2;
4478 REX_B(s
) = (b
& 0x1) << 3;
4479 x86_64_hregs
= 1; /* select uniform byte register addressing */
4484 case 0xc5: /* 2-byte VEX */
4485 case 0xc4: /* 3-byte VEX */
4486 /* VEX prefixes cannot be used except in 32-bit mode.
4487 Otherwise the instruction is LES or LDS. */
4488 if (s
->code32
&& !s
->vm86
) {
4489 static const int pp_prefix
[4] = {
4490 0, PREFIX_DATA
, PREFIX_REPZ
, PREFIX_REPNZ
4492 int vex3
, vex2
= cpu_ldub_code(env
, s
->pc
);
4494 if (!CODE64(s
) && (vex2
& 0xc0) != 0xc0) {
4495 /* 4.1.4.6: In 32-bit mode, bits [7:6] must be 11b,
4496 otherwise the instruction is LES or LDS. */
4501 /* 4.1.1-4.1.3: No preceding lock, 66, f2, f3, or rex prefixes. */
4502 if (prefixes
& (PREFIX_REPZ
| PREFIX_REPNZ
4503 | PREFIX_LOCK
| PREFIX_DATA
)) {
4506 #ifdef TARGET_X86_64
4511 rex_r
= (~vex2
>> 4) & 8;
4514 b
= cpu_ldub_code(env
, s
->pc
++);
4516 #ifdef TARGET_X86_64
4517 s
->rex_x
= (~vex2
>> 3) & 8;
4518 s
->rex_b
= (~vex2
>> 2) & 8;
4520 vex3
= cpu_ldub_code(env
, s
->pc
++);
4521 rex_w
= (vex3
>> 7) & 1;
4522 switch (vex2
& 0x1f) {
4523 case 0x01: /* Implied 0f leading opcode bytes. */
4524 b
= cpu_ldub_code(env
, s
->pc
++) | 0x100;
4526 case 0x02: /* Implied 0f 38 leading opcode bytes. */
4529 case 0x03: /* Implied 0f 3a leading opcode bytes. */
4532 default: /* Reserved for future use. */
4536 s
->vex_v
= (~vex3
>> 3) & 0xf;
4537 s
->vex_l
= (vex3
>> 2) & 1;
4538 prefixes
|= pp_prefix
[vex3
& 3] | PREFIX_VEX
;
4543 /* Post-process prefixes. */
4545 /* In 64-bit mode, the default data size is 32-bit. Select 64-bit
4546 data with rex_w, and 16-bit data with 0x66; rex_w takes precedence
4547 over 0x66 if both are present. */
4548 dflag
= (rex_w
> 0 ? MO_64
: prefixes
& PREFIX_DATA
? MO_16
: MO_32
);
4549 /* In 64-bit mode, 0x67 selects 32-bit addressing. */
4550 aflag
= (prefixes
& PREFIX_ADR
? MO_32
: MO_64
);
4552 /* In 16/32-bit mode, 0x66 selects the opposite data size. */
4553 if (s
->code32
^ ((prefixes
& PREFIX_DATA
) != 0)) {
4558 /* In 16/32-bit mode, 0x67 selects the opposite addressing. */
4559 if (s
->code32
^ ((prefixes
& PREFIX_ADR
) != 0)) {
4566 s
->prefix
= prefixes
;
4570 /* lock generation */
4571 if (prefixes
& PREFIX_LOCK
)
4574 /* now check op code */
4578 /**************************/
4579 /* extended op code */
4580 b
= cpu_ldub_code(env
, s
->pc
++) | 0x100;
4583 /**************************/
4598 ot
= mo_b_d(b
, dflag
);
4601 case 0: /* OP Ev, Gv */
4602 modrm
= cpu_ldub_code(env
, s
->pc
++);
4603 reg
= ((modrm
>> 3) & 7) | rex_r
;
4604 mod
= (modrm
>> 6) & 3;
4605 rm
= (modrm
& 7) | REX_B(s
);
4607 gen_lea_modrm(env
, s
, modrm
);
4609 } else if (op
== OP_XORL
&& rm
== reg
) {
4611 /* xor reg, reg optimisation */
4612 set_cc_op(s
, CC_OP_CLR
);
4613 tcg_gen_movi_tl(cpu_T
[0], 0);
4614 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
4619 gen_op_mov_v_reg(ot
, cpu_T
[1], reg
);
4620 gen_op(s
, op
, ot
, opreg
);
4622 case 1: /* OP Gv, Ev */
4623 modrm
= cpu_ldub_code(env
, s
->pc
++);
4624 mod
= (modrm
>> 6) & 3;
4625 reg
= ((modrm
>> 3) & 7) | rex_r
;
4626 rm
= (modrm
& 7) | REX_B(s
);
4628 gen_lea_modrm(env
, s
, modrm
);
4629 gen_op_ld_v(s
, ot
, cpu_T
[1], cpu_A0
);
4630 } else if (op
== OP_XORL
&& rm
== reg
) {
4633 gen_op_mov_v_reg(ot
, cpu_T
[1], rm
);
4635 gen_op(s
, op
, ot
, reg
);
4637 case 2: /* OP A, Iv */
4638 val
= insn_get(env
, s
, ot
);
4639 tcg_gen_movi_tl(cpu_T
[1], val
);
4640 gen_op(s
, op
, ot
, OR_EAX
);
4649 case 0x80: /* GRP1 */
4655 ot
= mo_b_d(b
, dflag
);
4657 modrm
= cpu_ldub_code(env
, s
->pc
++);
4658 mod
= (modrm
>> 6) & 3;
4659 rm
= (modrm
& 7) | REX_B(s
);
4660 op
= (modrm
>> 3) & 7;
4666 s
->rip_offset
= insn_const_size(ot
);
4667 gen_lea_modrm(env
, s
, modrm
);
4678 val
= insn_get(env
, s
, ot
);
4681 val
= (int8_t)insn_get(env
, s
, MO_8
);
4684 tcg_gen_movi_tl(cpu_T
[1], val
);
4685 gen_op(s
, op
, ot
, opreg
);
4689 /**************************/
4690 /* inc, dec, and other misc arith */
4691 case 0x40 ... 0x47: /* inc Gv */
4693 gen_inc(s
, ot
, OR_EAX
+ (b
& 7), 1);
4695 case 0x48 ... 0x4f: /* dec Gv */
4697 gen_inc(s
, ot
, OR_EAX
+ (b
& 7), -1);
4699 case 0xf6: /* GRP3 */
4701 ot
= mo_b_d(b
, dflag
);
4703 modrm
= cpu_ldub_code(env
, s
->pc
++);
4704 mod
= (modrm
>> 6) & 3;
4705 rm
= (modrm
& 7) | REX_B(s
);
4706 op
= (modrm
>> 3) & 7;
4709 s
->rip_offset
= insn_const_size(ot
);
4710 gen_lea_modrm(env
, s
, modrm
);
4711 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
4713 gen_op_mov_v_reg(ot
, cpu_T
[0], rm
);
4718 val
= insn_get(env
, s
, ot
);
4719 tcg_gen_movi_tl(cpu_T
[1], val
);
4720 gen_op_testl_T0_T1_cc();
4721 set_cc_op(s
, CC_OP_LOGICB
+ ot
);
4724 tcg_gen_not_tl(cpu_T
[0], cpu_T
[0]);
4726 gen_op_st_v(s
, ot
, cpu_T
[0], cpu_A0
);
4728 gen_op_mov_reg_v(ot
, rm
, cpu_T
[0]);
4732 tcg_gen_neg_tl(cpu_T
[0], cpu_T
[0]);
4734 gen_op_st_v(s
, ot
, cpu_T
[0], cpu_A0
);
4736 gen_op_mov_reg_v(ot
, rm
, cpu_T
[0]);
4738 gen_op_update_neg_cc();
4739 set_cc_op(s
, CC_OP_SUBB
+ ot
);
4744 gen_op_mov_v_reg(MO_8
, cpu_T
[1], R_EAX
);
4745 tcg_gen_ext8u_tl(cpu_T
[0], cpu_T
[0]);
4746 tcg_gen_ext8u_tl(cpu_T
[1], cpu_T
[1]);
4747 /* XXX: use 32 bit mul which could be faster */
4748 tcg_gen_mul_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
4749 gen_op_mov_reg_v(MO_16
, R_EAX
, cpu_T
[0]);
4750 tcg_gen_mov_tl(cpu_cc_dst
, cpu_T
[0]);
4751 tcg_gen_andi_tl(cpu_cc_src
, cpu_T
[0], 0xff00);
4752 set_cc_op(s
, CC_OP_MULB
);
4755 gen_op_mov_v_reg(MO_16
, cpu_T
[1], R_EAX
);
4756 tcg_gen_ext16u_tl(cpu_T
[0], cpu_T
[0]);
4757 tcg_gen_ext16u_tl(cpu_T
[1], cpu_T
[1]);
4758 /* XXX: use 32 bit mul which could be faster */
4759 tcg_gen_mul_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
4760 gen_op_mov_reg_v(MO_16
, R_EAX
, cpu_T
[0]);
4761 tcg_gen_mov_tl(cpu_cc_dst
, cpu_T
[0]);
4762 tcg_gen_shri_tl(cpu_T
[0], cpu_T
[0], 16);
4763 gen_op_mov_reg_v(MO_16
, R_EDX
, cpu_T
[0]);
4764 tcg_gen_mov_tl(cpu_cc_src
, cpu_T
[0]);
4765 set_cc_op(s
, CC_OP_MULW
);
4769 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
4770 tcg_gen_trunc_tl_i32(cpu_tmp3_i32
, cpu_regs
[R_EAX
]);
4771 tcg_gen_mulu2_i32(cpu_tmp2_i32
, cpu_tmp3_i32
,
4772 cpu_tmp2_i32
, cpu_tmp3_i32
);
4773 tcg_gen_extu_i32_tl(cpu_regs
[R_EAX
], cpu_tmp2_i32
);
4774 tcg_gen_extu_i32_tl(cpu_regs
[R_EDX
], cpu_tmp3_i32
);
4775 tcg_gen_mov_tl(cpu_cc_dst
, cpu_regs
[R_EAX
]);
4776 tcg_gen_mov_tl(cpu_cc_src
, cpu_regs
[R_EDX
]);
4777 set_cc_op(s
, CC_OP_MULL
);
4779 #ifdef TARGET_X86_64
4781 tcg_gen_mulu2_i64(cpu_regs
[R_EAX
], cpu_regs
[R_EDX
],
4782 cpu_T
[0], cpu_regs
[R_EAX
]);
4783 tcg_gen_mov_tl(cpu_cc_dst
, cpu_regs
[R_EAX
]);
4784 tcg_gen_mov_tl(cpu_cc_src
, cpu_regs
[R_EDX
]);
4785 set_cc_op(s
, CC_OP_MULQ
);
4793 gen_op_mov_v_reg(MO_8
, cpu_T
[1], R_EAX
);
4794 tcg_gen_ext8s_tl(cpu_T
[0], cpu_T
[0]);
4795 tcg_gen_ext8s_tl(cpu_T
[1], cpu_T
[1]);
4796 /* XXX: use 32 bit mul which could be faster */
4797 tcg_gen_mul_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
4798 gen_op_mov_reg_v(MO_16
, R_EAX
, cpu_T
[0]);
4799 tcg_gen_mov_tl(cpu_cc_dst
, cpu_T
[0]);
4800 tcg_gen_ext8s_tl(cpu_tmp0
, cpu_T
[0]);
4801 tcg_gen_sub_tl(cpu_cc_src
, cpu_T
[0], cpu_tmp0
);
4802 set_cc_op(s
, CC_OP_MULB
);
4805 gen_op_mov_v_reg(MO_16
, cpu_T
[1], R_EAX
);
4806 tcg_gen_ext16s_tl(cpu_T
[0], cpu_T
[0]);
4807 tcg_gen_ext16s_tl(cpu_T
[1], cpu_T
[1]);
4808 /* XXX: use 32 bit mul which could be faster */
4809 tcg_gen_mul_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
4810 gen_op_mov_reg_v(MO_16
, R_EAX
, cpu_T
[0]);
4811 tcg_gen_mov_tl(cpu_cc_dst
, cpu_T
[0]);
4812 tcg_gen_ext16s_tl(cpu_tmp0
, cpu_T
[0]);
4813 tcg_gen_sub_tl(cpu_cc_src
, cpu_T
[0], cpu_tmp0
);
4814 tcg_gen_shri_tl(cpu_T
[0], cpu_T
[0], 16);
4815 gen_op_mov_reg_v(MO_16
, R_EDX
, cpu_T
[0]);
4816 set_cc_op(s
, CC_OP_MULW
);
4820 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
4821 tcg_gen_trunc_tl_i32(cpu_tmp3_i32
, cpu_regs
[R_EAX
]);
4822 tcg_gen_muls2_i32(cpu_tmp2_i32
, cpu_tmp3_i32
,
4823 cpu_tmp2_i32
, cpu_tmp3_i32
);
4824 tcg_gen_extu_i32_tl(cpu_regs
[R_EAX
], cpu_tmp2_i32
);
4825 tcg_gen_extu_i32_tl(cpu_regs
[R_EDX
], cpu_tmp3_i32
);
4826 tcg_gen_sari_i32(cpu_tmp2_i32
, cpu_tmp2_i32
, 31);
4827 tcg_gen_mov_tl(cpu_cc_dst
, cpu_regs
[R_EAX
]);
4828 tcg_gen_sub_i32(cpu_tmp2_i32
, cpu_tmp2_i32
, cpu_tmp3_i32
);
4829 tcg_gen_extu_i32_tl(cpu_cc_src
, cpu_tmp2_i32
);
4830 set_cc_op(s
, CC_OP_MULL
);
4832 #ifdef TARGET_X86_64
4834 tcg_gen_muls2_i64(cpu_regs
[R_EAX
], cpu_regs
[R_EDX
],
4835 cpu_T
[0], cpu_regs
[R_EAX
]);
4836 tcg_gen_mov_tl(cpu_cc_dst
, cpu_regs
[R_EAX
]);
4837 tcg_gen_sari_tl(cpu_cc_src
, cpu_regs
[R_EAX
], 63);
4838 tcg_gen_sub_tl(cpu_cc_src
, cpu_cc_src
, cpu_regs
[R_EDX
]);
4839 set_cc_op(s
, CC_OP_MULQ
);
4847 gen_jmp_im(pc_start
- s
->cs_base
);
4848 gen_helper_divb_AL(cpu_env
, cpu_T
[0]);
4851 gen_jmp_im(pc_start
- s
->cs_base
);
4852 gen_helper_divw_AX(cpu_env
, cpu_T
[0]);
4856 gen_jmp_im(pc_start
- s
->cs_base
);
4857 gen_helper_divl_EAX(cpu_env
, cpu_T
[0]);
4859 #ifdef TARGET_X86_64
4861 gen_jmp_im(pc_start
- s
->cs_base
);
4862 gen_helper_divq_EAX(cpu_env
, cpu_T
[0]);
4870 gen_jmp_im(pc_start
- s
->cs_base
);
4871 gen_helper_idivb_AL(cpu_env
, cpu_T
[0]);
4874 gen_jmp_im(pc_start
- s
->cs_base
);
4875 gen_helper_idivw_AX(cpu_env
, cpu_T
[0]);
4879 gen_jmp_im(pc_start
- s
->cs_base
);
4880 gen_helper_idivl_EAX(cpu_env
, cpu_T
[0]);
4882 #ifdef TARGET_X86_64
4884 gen_jmp_im(pc_start
- s
->cs_base
);
4885 gen_helper_idivq_EAX(cpu_env
, cpu_T
[0]);
4895 case 0xfe: /* GRP4 */
4896 case 0xff: /* GRP5 */
4897 ot
= mo_b_d(b
, dflag
);
4899 modrm
= cpu_ldub_code(env
, s
->pc
++);
4900 mod
= (modrm
>> 6) & 3;
4901 rm
= (modrm
& 7) | REX_B(s
);
4902 op
= (modrm
>> 3) & 7;
4903 if (op
>= 2 && b
== 0xfe) {
4907 if (op
== 2 || op
== 4) {
4908 /* operand size for jumps is 64 bit */
4910 } else if (op
== 3 || op
== 5) {
4911 ot
= dflag
!= MO_16
? MO_32
+ (rex_w
== 1) : MO_16
;
4912 } else if (op
== 6) {
4913 /* default push size is 64 bit */
4914 ot
= mo_pushpop(s
, dflag
);
4918 gen_lea_modrm(env
, s
, modrm
);
4919 if (op
>= 2 && op
!= 3 && op
!= 5)
4920 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
4922 gen_op_mov_v_reg(ot
, cpu_T
[0], rm
);
4926 case 0: /* inc Ev */
4931 gen_inc(s
, ot
, opreg
, 1);
4933 case 1: /* dec Ev */
4938 gen_inc(s
, ot
, opreg
, -1);
4940 case 2: /* call Ev */
4941 /* XXX: optimize if memory (no 'and' is necessary) */
4942 if (dflag
== MO_16
) {
4943 tcg_gen_ext16u_tl(cpu_T
[0], cpu_T
[0]);
4945 next_eip
= s
->pc
- s
->cs_base
;
4946 tcg_gen_movi_tl(cpu_T
[1], next_eip
);
4947 gen_push_v(s
, cpu_T
[1]);
4948 gen_op_jmp_v(cpu_T
[0]);
4951 case 3: /* lcall Ev */
4952 gen_op_ld_v(s
, ot
, cpu_T
[1], cpu_A0
);
4953 gen_add_A0_im(s
, 1 << ot
);
4954 gen_op_ld_v(s
, MO_16
, cpu_T
[0], cpu_A0
);
4956 if (s
->pe
&& !s
->vm86
) {
4957 gen_update_cc_op(s
);
4958 gen_jmp_im(pc_start
- s
->cs_base
);
4959 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
4960 gen_helper_lcall_protected(cpu_env
, cpu_tmp2_i32
, cpu_T
[1],
4961 tcg_const_i32(dflag
- 1),
4962 tcg_const_i32(s
->pc
- pc_start
));
4964 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
4965 gen_helper_lcall_real(cpu_env
, cpu_tmp2_i32
, cpu_T
[1],
4966 tcg_const_i32(dflag
- 1),
4967 tcg_const_i32(s
->pc
- s
->cs_base
));
4971 case 4: /* jmp Ev */
4972 if (dflag
== MO_16
) {
4973 tcg_gen_ext16u_tl(cpu_T
[0], cpu_T
[0]);
4975 gen_op_jmp_v(cpu_T
[0]);
4978 case 5: /* ljmp Ev */
4979 gen_op_ld_v(s
, ot
, cpu_T
[1], cpu_A0
);
4980 gen_add_A0_im(s
, 1 << ot
);
4981 gen_op_ld_v(s
, MO_16
, cpu_T
[0], cpu_A0
);
4983 if (s
->pe
&& !s
->vm86
) {
4984 gen_update_cc_op(s
);
4985 gen_jmp_im(pc_start
- s
->cs_base
);
4986 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
4987 gen_helper_ljmp_protected(cpu_env
, cpu_tmp2_i32
, cpu_T
[1],
4988 tcg_const_i32(s
->pc
- pc_start
));
4990 gen_op_movl_seg_T0_vm(R_CS
);
4991 gen_op_jmp_v(cpu_T
[1]);
4995 case 6: /* push Ev */
4996 gen_push_v(s
, cpu_T
[0]);
5003 case 0x84: /* test Ev, Gv */
5005 ot
= mo_b_d(b
, dflag
);
5007 modrm
= cpu_ldub_code(env
, s
->pc
++);
5008 reg
= ((modrm
>> 3) & 7) | rex_r
;
5010 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
5011 gen_op_mov_v_reg(ot
, cpu_T
[1], reg
);
5012 gen_op_testl_T0_T1_cc();
5013 set_cc_op(s
, CC_OP_LOGICB
+ ot
);
5016 case 0xa8: /* test eAX, Iv */
5018 ot
= mo_b_d(b
, dflag
);
5019 val
= insn_get(env
, s
, ot
);
5021 gen_op_mov_v_reg(ot
, cpu_T
[0], OR_EAX
);
5022 tcg_gen_movi_tl(cpu_T
[1], val
);
5023 gen_op_testl_T0_T1_cc();
5024 set_cc_op(s
, CC_OP_LOGICB
+ ot
);
5027 case 0x98: /* CWDE/CBW */
5029 #ifdef TARGET_X86_64
5031 gen_op_mov_v_reg(MO_32
, cpu_T
[0], R_EAX
);
5032 tcg_gen_ext32s_tl(cpu_T
[0], cpu_T
[0]);
5033 gen_op_mov_reg_v(MO_64
, R_EAX
, cpu_T
[0]);
5037 gen_op_mov_v_reg(MO_16
, cpu_T
[0], R_EAX
);
5038 tcg_gen_ext16s_tl(cpu_T
[0], cpu_T
[0]);
5039 gen_op_mov_reg_v(MO_32
, R_EAX
, cpu_T
[0]);
5042 gen_op_mov_v_reg(MO_8
, cpu_T
[0], R_EAX
);
5043 tcg_gen_ext8s_tl(cpu_T
[0], cpu_T
[0]);
5044 gen_op_mov_reg_v(MO_16
, R_EAX
, cpu_T
[0]);
5050 case 0x99: /* CDQ/CWD */
5052 #ifdef TARGET_X86_64
5054 gen_op_mov_v_reg(MO_64
, cpu_T
[0], R_EAX
);
5055 tcg_gen_sari_tl(cpu_T
[0], cpu_T
[0], 63);
5056 gen_op_mov_reg_v(MO_64
, R_EDX
, cpu_T
[0]);
5060 gen_op_mov_v_reg(MO_32
, cpu_T
[0], R_EAX
);
5061 tcg_gen_ext32s_tl(cpu_T
[0], cpu_T
[0]);
5062 tcg_gen_sari_tl(cpu_T
[0], cpu_T
[0], 31);
5063 gen_op_mov_reg_v(MO_32
, R_EDX
, cpu_T
[0]);
5066 gen_op_mov_v_reg(MO_16
, cpu_T
[0], R_EAX
);
5067 tcg_gen_ext16s_tl(cpu_T
[0], cpu_T
[0]);
5068 tcg_gen_sari_tl(cpu_T
[0], cpu_T
[0], 15);
5069 gen_op_mov_reg_v(MO_16
, R_EDX
, cpu_T
[0]);
5075 case 0x1af: /* imul Gv, Ev */
5076 case 0x69: /* imul Gv, Ev, I */
5079 modrm
= cpu_ldub_code(env
, s
->pc
++);
5080 reg
= ((modrm
>> 3) & 7) | rex_r
;
5082 s
->rip_offset
= insn_const_size(ot
);
5085 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
5087 val
= insn_get(env
, s
, ot
);
5088 tcg_gen_movi_tl(cpu_T
[1], val
);
5089 } else if (b
== 0x6b) {
5090 val
= (int8_t)insn_get(env
, s
, MO_8
);
5091 tcg_gen_movi_tl(cpu_T
[1], val
);
5093 gen_op_mov_v_reg(ot
, cpu_T
[1], reg
);
5096 #ifdef TARGET_X86_64
5098 tcg_gen_muls2_i64(cpu_regs
[reg
], cpu_T
[1], cpu_T
[0], cpu_T
[1]);
5099 tcg_gen_mov_tl(cpu_cc_dst
, cpu_regs
[reg
]);
5100 tcg_gen_sari_tl(cpu_cc_src
, cpu_cc_dst
, 63);
5101 tcg_gen_sub_tl(cpu_cc_src
, cpu_cc_src
, cpu_T
[1]);
5105 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
5106 tcg_gen_trunc_tl_i32(cpu_tmp3_i32
, cpu_T
[1]);
5107 tcg_gen_muls2_i32(cpu_tmp2_i32
, cpu_tmp3_i32
,
5108 cpu_tmp2_i32
, cpu_tmp3_i32
);
5109 tcg_gen_extu_i32_tl(cpu_regs
[reg
], cpu_tmp2_i32
);
5110 tcg_gen_sari_i32(cpu_tmp2_i32
, cpu_tmp2_i32
, 31);
5111 tcg_gen_mov_tl(cpu_cc_dst
, cpu_regs
[reg
]);
5112 tcg_gen_sub_i32(cpu_tmp2_i32
, cpu_tmp2_i32
, cpu_tmp3_i32
);
5113 tcg_gen_extu_i32_tl(cpu_cc_src
, cpu_tmp2_i32
);
5116 tcg_gen_ext16s_tl(cpu_T
[0], cpu_T
[0]);
5117 tcg_gen_ext16s_tl(cpu_T
[1], cpu_T
[1]);
5118 /* XXX: use 32 bit mul which could be faster */
5119 tcg_gen_mul_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
5120 tcg_gen_mov_tl(cpu_cc_dst
, cpu_T
[0]);
5121 tcg_gen_ext16s_tl(cpu_tmp0
, cpu_T
[0]);
5122 tcg_gen_sub_tl(cpu_cc_src
, cpu_T
[0], cpu_tmp0
);
5123 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
5126 set_cc_op(s
, CC_OP_MULB
+ ot
);
5129 case 0x1c1: /* xadd Ev, Gv */
5130 ot
= mo_b_d(b
, dflag
);
5131 modrm
= cpu_ldub_code(env
, s
->pc
++);
5132 reg
= ((modrm
>> 3) & 7) | rex_r
;
5133 mod
= (modrm
>> 6) & 3;
5135 rm
= (modrm
& 7) | REX_B(s
);
5136 gen_op_mov_v_reg(ot
, cpu_T
[0], reg
);
5137 gen_op_mov_v_reg(ot
, cpu_T
[1], rm
);
5138 tcg_gen_add_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
5139 gen_op_mov_reg_v(ot
, reg
, cpu_T
[1]);
5140 gen_op_mov_reg_v(ot
, rm
, cpu_T
[0]);
5142 gen_lea_modrm(env
, s
, modrm
);
5143 gen_op_mov_v_reg(ot
, cpu_T
[0], reg
);
5144 gen_op_ld_v(s
, ot
, cpu_T
[1], cpu_A0
);
5145 tcg_gen_add_tl(cpu_T
[0], cpu_T
[0], cpu_T
[1]);
5146 gen_op_st_v(s
, ot
, cpu_T
[0], cpu_A0
);
5147 gen_op_mov_reg_v(ot
, reg
, cpu_T
[1]);
5149 gen_op_update2_cc();
5150 set_cc_op(s
, CC_OP_ADDB
+ ot
);
5153 case 0x1b1: /* cmpxchg Ev, Gv */
5156 TCGv t0
, t1
, t2
, a0
;
5158 ot
= mo_b_d(b
, dflag
);
5159 modrm
= cpu_ldub_code(env
, s
->pc
++);
5160 reg
= ((modrm
>> 3) & 7) | rex_r
;
5161 mod
= (modrm
>> 6) & 3;
5162 t0
= tcg_temp_local_new();
5163 t1
= tcg_temp_local_new();
5164 t2
= tcg_temp_local_new();
5165 a0
= tcg_temp_local_new();
5166 gen_op_mov_v_reg(ot
, t1
, reg
);
5168 rm
= (modrm
& 7) | REX_B(s
);
5169 gen_op_mov_v_reg(ot
, t0
, rm
);
5171 gen_lea_modrm(env
, s
, modrm
);
5172 tcg_gen_mov_tl(a0
, cpu_A0
);
5173 gen_op_ld_v(s
, ot
, t0
, a0
);
5174 rm
= 0; /* avoid warning */
5176 label1
= gen_new_label();
5177 tcg_gen_mov_tl(t2
, cpu_regs
[R_EAX
]);
5180 tcg_gen_brcond_tl(TCG_COND_EQ
, t2
, t0
, label1
);
5181 label2
= gen_new_label();
5183 gen_op_mov_reg_v(ot
, R_EAX
, t0
);
5185 gen_set_label(label1
);
5186 gen_op_mov_reg_v(ot
, rm
, t1
);
5188 /* perform no-op store cycle like physical cpu; must be
5189 before changing accumulator to ensure idempotency if
5190 the store faults and the instruction is restarted */
5191 gen_op_st_v(s
, ot
, t0
, a0
);
5192 gen_op_mov_reg_v(ot
, R_EAX
, t0
);
5194 gen_set_label(label1
);
5195 gen_op_st_v(s
, ot
, t1
, a0
);
5197 gen_set_label(label2
);
5198 tcg_gen_mov_tl(cpu_cc_src
, t0
);
5199 tcg_gen_mov_tl(cpu_cc_srcT
, t2
);
5200 tcg_gen_sub_tl(cpu_cc_dst
, t2
, t0
);
5201 set_cc_op(s
, CC_OP_SUBB
+ ot
);
5208 case 0x1c7: /* cmpxchg8b */
5209 modrm
= cpu_ldub_code(env
, s
->pc
++);
5210 mod
= (modrm
>> 6) & 3;
5211 if ((mod
== 3) || ((modrm
& 0x38) != 0x8))
5213 #ifdef TARGET_X86_64
5214 if (dflag
== MO_64
) {
5215 if (!(s
->cpuid_ext_features
& CPUID_EXT_CX16
))
5217 gen_jmp_im(pc_start
- s
->cs_base
);
5218 gen_update_cc_op(s
);
5219 gen_lea_modrm(env
, s
, modrm
);
5220 gen_helper_cmpxchg16b(cpu_env
, cpu_A0
);
5224 if (!(s
->cpuid_features
& CPUID_CX8
))
5226 gen_jmp_im(pc_start
- s
->cs_base
);
5227 gen_update_cc_op(s
);
5228 gen_lea_modrm(env
, s
, modrm
);
5229 gen_helper_cmpxchg8b(cpu_env
, cpu_A0
);
5231 set_cc_op(s
, CC_OP_EFLAGS
);
5234 /**************************/
5236 case 0x50 ... 0x57: /* push */
5237 gen_op_mov_v_reg(MO_32
, cpu_T
[0], (b
& 7) | REX_B(s
));
5238 gen_push_v(s
, cpu_T
[0]);
5240 case 0x58 ... 0x5f: /* pop */
5242 /* NOTE: order is important for pop %sp */
5243 gen_pop_update(s
, ot
);
5244 gen_op_mov_reg_v(ot
, (b
& 7) | REX_B(s
), cpu_T
[0]);
5246 case 0x60: /* pusha */
5251 case 0x61: /* popa */
5256 case 0x68: /* push Iv */
5258 ot
= mo_pushpop(s
, dflag
);
5260 val
= insn_get(env
, s
, ot
);
5262 val
= (int8_t)insn_get(env
, s
, MO_8
);
5263 tcg_gen_movi_tl(cpu_T
[0], val
);
5264 gen_push_v(s
, cpu_T
[0]);
5266 case 0x8f: /* pop Ev */
5267 modrm
= cpu_ldub_code(env
, s
->pc
++);
5268 mod
= (modrm
>> 6) & 3;
5271 /* NOTE: order is important for pop %sp */
5272 gen_pop_update(s
, ot
);
5273 rm
= (modrm
& 7) | REX_B(s
);
5274 gen_op_mov_reg_v(ot
, rm
, cpu_T
[0]);
5276 /* NOTE: order is important too for MMU exceptions */
5277 s
->popl_esp_hack
= 1 << ot
;
5278 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 1);
5279 s
->popl_esp_hack
= 0;
5280 gen_pop_update(s
, ot
);
5283 case 0xc8: /* enter */
5286 val
= cpu_lduw_code(env
, s
->pc
);
5288 level
= cpu_ldub_code(env
, s
->pc
++);
5289 gen_enter(s
, val
, level
);
5292 case 0xc9: /* leave */
5293 /* XXX: exception not precise (ESP is updated before potential exception) */
5295 gen_op_mov_v_reg(MO_64
, cpu_T
[0], R_EBP
);
5296 gen_op_mov_reg_v(MO_64
, R_ESP
, cpu_T
[0]);
5297 } else if (s
->ss32
) {
5298 gen_op_mov_v_reg(MO_32
, cpu_T
[0], R_EBP
);
5299 gen_op_mov_reg_v(MO_32
, R_ESP
, cpu_T
[0]);
5301 gen_op_mov_v_reg(MO_16
, cpu_T
[0], R_EBP
);
5302 gen_op_mov_reg_v(MO_16
, R_ESP
, cpu_T
[0]);
5305 gen_op_mov_reg_v(ot
, R_EBP
, cpu_T
[0]);
5306 gen_pop_update(s
, ot
);
5308 case 0x06: /* push es */
5309 case 0x0e: /* push cs */
5310 case 0x16: /* push ss */
5311 case 0x1e: /* push ds */
5314 gen_op_movl_T0_seg(b
>> 3);
5315 gen_push_v(s
, cpu_T
[0]);
5317 case 0x1a0: /* push fs */
5318 case 0x1a8: /* push gs */
5319 gen_op_movl_T0_seg((b
>> 3) & 7);
5320 gen_push_v(s
, cpu_T
[0]);
5322 case 0x07: /* pop es */
5323 case 0x17: /* pop ss */
5324 case 0x1f: /* pop ds */
5329 gen_movl_seg_T0(s
, reg
, pc_start
- s
->cs_base
);
5330 gen_pop_update(s
, ot
);
5332 /* if reg == SS, inhibit interrupts/trace. */
5333 /* If several instructions disable interrupts, only the
5335 if (!(s
->tb
->flags
& HF_INHIBIT_IRQ_MASK
))
5336 gen_helper_set_inhibit_irq(cpu_env
);
5340 gen_jmp_im(s
->pc
- s
->cs_base
);
5344 case 0x1a1: /* pop fs */
5345 case 0x1a9: /* pop gs */
5347 gen_movl_seg_T0(s
, (b
>> 3) & 7, pc_start
- s
->cs_base
);
5348 gen_pop_update(s
, ot
);
5350 gen_jmp_im(s
->pc
- s
->cs_base
);
5355 /**************************/
5358 case 0x89: /* mov Gv, Ev */
5359 ot
= mo_b_d(b
, dflag
);
5360 modrm
= cpu_ldub_code(env
, s
->pc
++);
5361 reg
= ((modrm
>> 3) & 7) | rex_r
;
5363 /* generate a generic store */
5364 gen_ldst_modrm(env
, s
, modrm
, ot
, reg
, 1);
5367 case 0xc7: /* mov Ev, Iv */
5368 ot
= mo_b_d(b
, dflag
);
5369 modrm
= cpu_ldub_code(env
, s
->pc
++);
5370 mod
= (modrm
>> 6) & 3;
5372 s
->rip_offset
= insn_const_size(ot
);
5373 gen_lea_modrm(env
, s
, modrm
);
5375 val
= insn_get(env
, s
, ot
);
5376 tcg_gen_movi_tl(cpu_T
[0], val
);
5378 gen_op_st_v(s
, ot
, cpu_T
[0], cpu_A0
);
5380 gen_op_mov_reg_v(ot
, (modrm
& 7) | REX_B(s
), cpu_T
[0]);
5384 case 0x8b: /* mov Ev, Gv */
5385 ot
= mo_b_d(b
, dflag
);
5386 modrm
= cpu_ldub_code(env
, s
->pc
++);
5387 reg
= ((modrm
>> 3) & 7) | rex_r
;
5389 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
5390 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
5392 case 0x8e: /* mov seg, Gv */
5393 modrm
= cpu_ldub_code(env
, s
->pc
++);
5394 reg
= (modrm
>> 3) & 7;
5395 if (reg
>= 6 || reg
== R_CS
)
5397 gen_ldst_modrm(env
, s
, modrm
, MO_16
, OR_TMP0
, 0);
5398 gen_movl_seg_T0(s
, reg
, pc_start
- s
->cs_base
);
5400 /* if reg == SS, inhibit interrupts/trace */
5401 /* If several instructions disable interrupts, only the
5403 if (!(s
->tb
->flags
& HF_INHIBIT_IRQ_MASK
))
5404 gen_helper_set_inhibit_irq(cpu_env
);
5408 gen_jmp_im(s
->pc
- s
->cs_base
);
5412 case 0x8c: /* mov Gv, seg */
5413 modrm
= cpu_ldub_code(env
, s
->pc
++);
5414 reg
= (modrm
>> 3) & 7;
5415 mod
= (modrm
>> 6) & 3;
5418 gen_op_movl_T0_seg(reg
);
5419 ot
= mod
== 3 ? dflag
: MO_16
;
5420 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 1);
5423 case 0x1b6: /* movzbS Gv, Eb */
5424 case 0x1b7: /* movzwS Gv, Eb */
5425 case 0x1be: /* movsbS Gv, Eb */
5426 case 0x1bf: /* movswS Gv, Eb */
5431 /* d_ot is the size of destination */
5433 /* ot is the size of source */
5434 ot
= (b
& 1) + MO_8
;
5435 /* s_ot is the sign+size of source */
5436 s_ot
= b
& 8 ? MO_SIGN
| ot
: ot
;
5438 modrm
= cpu_ldub_code(env
, s
->pc
++);
5439 reg
= ((modrm
>> 3) & 7) | rex_r
;
5440 mod
= (modrm
>> 6) & 3;
5441 rm
= (modrm
& 7) | REX_B(s
);
5444 gen_op_mov_v_reg(ot
, cpu_T
[0], rm
);
5447 tcg_gen_ext8u_tl(cpu_T
[0], cpu_T
[0]);
5450 tcg_gen_ext8s_tl(cpu_T
[0], cpu_T
[0]);
5453 tcg_gen_ext16u_tl(cpu_T
[0], cpu_T
[0]);
5457 tcg_gen_ext16s_tl(cpu_T
[0], cpu_T
[0]);
5460 gen_op_mov_reg_v(d_ot
, reg
, cpu_T
[0]);
5462 gen_lea_modrm(env
, s
, modrm
);
5463 gen_op_ld_v(s
, s_ot
, cpu_T
[0], cpu_A0
);
5464 gen_op_mov_reg_v(d_ot
, reg
, cpu_T
[0]);
5469 case 0x8d: /* lea */
5471 modrm
= cpu_ldub_code(env
, s
->pc
++);
5472 mod
= (modrm
>> 6) & 3;
5475 reg
= ((modrm
>> 3) & 7) | rex_r
;
5476 /* we must ensure that no segment is added */
5480 gen_lea_modrm(env
, s
, modrm
);
5482 gen_op_mov_reg_v(ot
, reg
, cpu_A0
);
5485 case 0xa0: /* mov EAX, Ov */
5487 case 0xa2: /* mov Ov, EAX */
5490 target_ulong offset_addr
;
5492 ot
= mo_b_d(b
, dflag
);
5494 #ifdef TARGET_X86_64
5496 offset_addr
= cpu_ldq_code(env
, s
->pc
);
5501 offset_addr
= insn_get(env
, s
, s
->aflag
);
5504 tcg_gen_movi_tl(cpu_A0
, offset_addr
);
5505 gen_add_A0_ds_seg(s
);
5507 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
5508 gen_op_mov_reg_v(ot
, R_EAX
, cpu_T
[0]);
5510 gen_op_mov_v_reg(ot
, cpu_T
[0], R_EAX
);
5511 gen_op_st_v(s
, ot
, cpu_T
[0], cpu_A0
);
5515 case 0xd7: /* xlat */
5516 tcg_gen_mov_tl(cpu_A0
, cpu_regs
[R_EBX
]);
5517 tcg_gen_ext8u_tl(cpu_T
[0], cpu_regs
[R_EAX
]);
5518 tcg_gen_add_tl(cpu_A0
, cpu_A0
, cpu_T
[0]);
5519 gen_extu(s
->aflag
, cpu_A0
);
5520 gen_add_A0_ds_seg(s
);
5521 gen_op_ld_v(s
, MO_8
, cpu_T
[0], cpu_A0
);
5522 gen_op_mov_reg_v(MO_8
, R_EAX
, cpu_T
[0]);
5524 case 0xb0 ... 0xb7: /* mov R, Ib */
5525 val
= insn_get(env
, s
, MO_8
);
5526 tcg_gen_movi_tl(cpu_T
[0], val
);
5527 gen_op_mov_reg_v(MO_8
, (b
& 7) | REX_B(s
), cpu_T
[0]);
5529 case 0xb8 ... 0xbf: /* mov R, Iv */
5530 #ifdef TARGET_X86_64
5531 if (dflag
== MO_64
) {
5534 tmp
= cpu_ldq_code(env
, s
->pc
);
5536 reg
= (b
& 7) | REX_B(s
);
5537 tcg_gen_movi_tl(cpu_T
[0], tmp
);
5538 gen_op_mov_reg_v(MO_64
, reg
, cpu_T
[0]);
5543 val
= insn_get(env
, s
, ot
);
5544 reg
= (b
& 7) | REX_B(s
);
5545 tcg_gen_movi_tl(cpu_T
[0], val
);
5546 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
5550 case 0x91 ... 0x97: /* xchg R, EAX */
5553 reg
= (b
& 7) | REX_B(s
);
5557 case 0x87: /* xchg Ev, Gv */
5558 ot
= mo_b_d(b
, dflag
);
5559 modrm
= cpu_ldub_code(env
, s
->pc
++);
5560 reg
= ((modrm
>> 3) & 7) | rex_r
;
5561 mod
= (modrm
>> 6) & 3;
5563 rm
= (modrm
& 7) | REX_B(s
);
5565 gen_op_mov_v_reg(ot
, cpu_T
[0], reg
);
5566 gen_op_mov_v_reg(ot
, cpu_T
[1], rm
);
5567 gen_op_mov_reg_v(ot
, rm
, cpu_T
[0]);
5568 gen_op_mov_reg_v(ot
, reg
, cpu_T
[1]);
5570 gen_lea_modrm(env
, s
, modrm
);
5571 gen_op_mov_v_reg(ot
, cpu_T
[0], reg
);
5572 /* for xchg, lock is implicit */
5573 if (!(prefixes
& PREFIX_LOCK
))
5575 gen_op_ld_v(s
, ot
, cpu_T
[1], cpu_A0
);
5576 gen_op_st_v(s
, ot
, cpu_T
[0], cpu_A0
);
5577 if (!(prefixes
& PREFIX_LOCK
))
5578 gen_helper_unlock();
5579 gen_op_mov_reg_v(ot
, reg
, cpu_T
[1]);
5582 case 0xc4: /* les Gv */
5583 /* In CODE64 this is VEX3; see above. */
5586 case 0xc5: /* lds Gv */
5587 /* In CODE64 this is VEX2; see above. */
5590 case 0x1b2: /* lss Gv */
5593 case 0x1b4: /* lfs Gv */
5596 case 0x1b5: /* lgs Gv */
5599 ot
= dflag
!= MO_16
? MO_32
: MO_16
;
5600 modrm
= cpu_ldub_code(env
, s
->pc
++);
5601 reg
= ((modrm
>> 3) & 7) | rex_r
;
5602 mod
= (modrm
>> 6) & 3;
5605 gen_lea_modrm(env
, s
, modrm
);
5606 gen_op_ld_v(s
, ot
, cpu_T
[1], cpu_A0
);
5607 gen_add_A0_im(s
, 1 << ot
);
5608 /* load the segment first to handle exceptions properly */
5609 gen_op_ld_v(s
, MO_16
, cpu_T
[0], cpu_A0
);
5610 gen_movl_seg_T0(s
, op
, pc_start
- s
->cs_base
);
5611 /* then put the data */
5612 gen_op_mov_reg_v(ot
, reg
, cpu_T
[1]);
5614 gen_jmp_im(s
->pc
- s
->cs_base
);
5619 /************************/
5627 ot
= mo_b_d(b
, dflag
);
5628 modrm
= cpu_ldub_code(env
, s
->pc
++);
5629 mod
= (modrm
>> 6) & 3;
5630 op
= (modrm
>> 3) & 7;
5636 gen_lea_modrm(env
, s
, modrm
);
5639 opreg
= (modrm
& 7) | REX_B(s
);
5644 gen_shift(s
, op
, ot
, opreg
, OR_ECX
);
5647 shift
= cpu_ldub_code(env
, s
->pc
++);
5649 gen_shifti(s
, op
, ot
, opreg
, shift
);
5664 case 0x1a4: /* shld imm */
5668 case 0x1a5: /* shld cl */
5672 case 0x1ac: /* shrd imm */
5676 case 0x1ad: /* shrd cl */
5681 modrm
= cpu_ldub_code(env
, s
->pc
++);
5682 mod
= (modrm
>> 6) & 3;
5683 rm
= (modrm
& 7) | REX_B(s
);
5684 reg
= ((modrm
>> 3) & 7) | rex_r
;
5686 gen_lea_modrm(env
, s
, modrm
);
5691 gen_op_mov_v_reg(ot
, cpu_T
[1], reg
);
5694 TCGv imm
= tcg_const_tl(cpu_ldub_code(env
, s
->pc
++));
5695 gen_shiftd_rm_T1(s
, ot
, opreg
, op
, imm
);
5698 gen_shiftd_rm_T1(s
, ot
, opreg
, op
, cpu_regs
[R_ECX
]);
5702 /************************/
5705 if (s
->flags
& (HF_EM_MASK
| HF_TS_MASK
)) {
5706 /* if CR0.EM or CR0.TS are set, generate an FPU exception */
5707 /* XXX: what to do if illegal op ? */
5708 gen_exception(s
, EXCP07_PREX
, pc_start
- s
->cs_base
);
5711 modrm
= cpu_ldub_code(env
, s
->pc
++);
5712 mod
= (modrm
>> 6) & 3;
5714 op
= ((b
& 7) << 3) | ((modrm
>> 3) & 7);
5717 gen_lea_modrm(env
, s
, modrm
);
5719 case 0x00 ... 0x07: /* fxxxs */
5720 case 0x10 ... 0x17: /* fixxxl */
5721 case 0x20 ... 0x27: /* fxxxl */
5722 case 0x30 ... 0x37: /* fixxx */
5729 tcg_gen_qemu_ld_i32(cpu_tmp2_i32
, cpu_A0
,
5730 s
->mem_index
, MO_LEUL
);
5731 gen_helper_flds_FT0(cpu_env
, cpu_tmp2_i32
);
5734 tcg_gen_qemu_ld_i32(cpu_tmp2_i32
, cpu_A0
,
5735 s
->mem_index
, MO_LEUL
);
5736 gen_helper_fildl_FT0(cpu_env
, cpu_tmp2_i32
);
5739 tcg_gen_qemu_ld_i64(cpu_tmp1_i64
, cpu_A0
,
5740 s
->mem_index
, MO_LEQ
);
5741 gen_helper_fldl_FT0(cpu_env
, cpu_tmp1_i64
);
5745 tcg_gen_qemu_ld_i32(cpu_tmp2_i32
, cpu_A0
,
5746 s
->mem_index
, MO_LESW
);
5747 gen_helper_fildl_FT0(cpu_env
, cpu_tmp2_i32
);
5751 gen_helper_fp_arith_ST0_FT0(op1
);
5753 /* fcomp needs pop */
5754 gen_helper_fpop(cpu_env
);
5758 case 0x08: /* flds */
5759 case 0x0a: /* fsts */
5760 case 0x0b: /* fstps */
5761 case 0x18 ... 0x1b: /* fildl, fisttpl, fistl, fistpl */
5762 case 0x28 ... 0x2b: /* fldl, fisttpll, fstl, fstpl */
5763 case 0x38 ... 0x3b: /* filds, fisttps, fists, fistps */
5768 tcg_gen_qemu_ld_i32(cpu_tmp2_i32
, cpu_A0
,
5769 s
->mem_index
, MO_LEUL
);
5770 gen_helper_flds_ST0(cpu_env
, cpu_tmp2_i32
);
5773 tcg_gen_qemu_ld_i32(cpu_tmp2_i32
, cpu_A0
,
5774 s
->mem_index
, MO_LEUL
);
5775 gen_helper_fildl_ST0(cpu_env
, cpu_tmp2_i32
);
5778 tcg_gen_qemu_ld_i64(cpu_tmp1_i64
, cpu_A0
,
5779 s
->mem_index
, MO_LEQ
);
5780 gen_helper_fldl_ST0(cpu_env
, cpu_tmp1_i64
);
5784 tcg_gen_qemu_ld_i32(cpu_tmp2_i32
, cpu_A0
,
5785 s
->mem_index
, MO_LESW
);
5786 gen_helper_fildl_ST0(cpu_env
, cpu_tmp2_i32
);
5791 /* XXX: the corresponding CPUID bit must be tested ! */
5794 gen_helper_fisttl_ST0(cpu_tmp2_i32
, cpu_env
);
5795 tcg_gen_qemu_st_i32(cpu_tmp2_i32
, cpu_A0
,
5796 s
->mem_index
, MO_LEUL
);
5799 gen_helper_fisttll_ST0(cpu_tmp1_i64
, cpu_env
);
5800 tcg_gen_qemu_st_i64(cpu_tmp1_i64
, cpu_A0
,
5801 s
->mem_index
, MO_LEQ
);
5805 gen_helper_fistt_ST0(cpu_tmp2_i32
, cpu_env
);
5806 tcg_gen_qemu_st_i32(cpu_tmp2_i32
, cpu_A0
,
5807 s
->mem_index
, MO_LEUW
);
5810 gen_helper_fpop(cpu_env
);
5815 gen_helper_fsts_ST0(cpu_tmp2_i32
, cpu_env
);
5816 tcg_gen_qemu_st_i32(cpu_tmp2_i32
, cpu_A0
,
5817 s
->mem_index
, MO_LEUL
);
5820 gen_helper_fistl_ST0(cpu_tmp2_i32
, cpu_env
);
5821 tcg_gen_qemu_st_i32(cpu_tmp2_i32
, cpu_A0
,
5822 s
->mem_index
, MO_LEUL
);
5825 gen_helper_fstl_ST0(cpu_tmp1_i64
, cpu_env
);
5826 tcg_gen_qemu_st_i64(cpu_tmp1_i64
, cpu_A0
,
5827 s
->mem_index
, MO_LEQ
);
5831 gen_helper_fist_ST0(cpu_tmp2_i32
, cpu_env
);
5832 tcg_gen_qemu_st_i32(cpu_tmp2_i32
, cpu_A0
,
5833 s
->mem_index
, MO_LEUW
);
5837 gen_helper_fpop(cpu_env
);
5841 case 0x0c: /* fldenv mem */
5842 gen_update_cc_op(s
);
5843 gen_jmp_im(pc_start
- s
->cs_base
);
5844 gen_helper_fldenv(cpu_env
, cpu_A0
, tcg_const_i32(dflag
- 1));
5846 case 0x0d: /* fldcw mem */
5847 tcg_gen_qemu_ld_i32(cpu_tmp2_i32
, cpu_A0
,
5848 s
->mem_index
, MO_LEUW
);
5849 gen_helper_fldcw(cpu_env
, cpu_tmp2_i32
);
5851 case 0x0e: /* fnstenv mem */
5852 gen_update_cc_op(s
);
5853 gen_jmp_im(pc_start
- s
->cs_base
);
5854 gen_helper_fstenv(cpu_env
, cpu_A0
, tcg_const_i32(dflag
- 1));
5856 case 0x0f: /* fnstcw mem */
5857 gen_helper_fnstcw(cpu_tmp2_i32
, cpu_env
);
5858 tcg_gen_qemu_st_i32(cpu_tmp2_i32
, cpu_A0
,
5859 s
->mem_index
, MO_LEUW
);
5861 case 0x1d: /* fldt mem */
5862 gen_update_cc_op(s
);
5863 gen_jmp_im(pc_start
- s
->cs_base
);
5864 gen_helper_fldt_ST0(cpu_env
, cpu_A0
);
5866 case 0x1f: /* fstpt mem */
5867 gen_update_cc_op(s
);
5868 gen_jmp_im(pc_start
- s
->cs_base
);
5869 gen_helper_fstt_ST0(cpu_env
, cpu_A0
);
5870 gen_helper_fpop(cpu_env
);
5872 case 0x2c: /* frstor mem */
5873 gen_update_cc_op(s
);
5874 gen_jmp_im(pc_start
- s
->cs_base
);
5875 gen_helper_frstor(cpu_env
, cpu_A0
, tcg_const_i32(dflag
- 1));
5877 case 0x2e: /* fnsave mem */
5878 gen_update_cc_op(s
);
5879 gen_jmp_im(pc_start
- s
->cs_base
);
5880 gen_helper_fsave(cpu_env
, cpu_A0
, tcg_const_i32(dflag
- 1));
5882 case 0x2f: /* fnstsw mem */
5883 gen_helper_fnstsw(cpu_tmp2_i32
, cpu_env
);
5884 tcg_gen_qemu_st_i32(cpu_tmp2_i32
, cpu_A0
,
5885 s
->mem_index
, MO_LEUW
);
5887 case 0x3c: /* fbld */
5888 gen_update_cc_op(s
);
5889 gen_jmp_im(pc_start
- s
->cs_base
);
5890 gen_helper_fbld_ST0(cpu_env
, cpu_A0
);
5892 case 0x3e: /* fbstp */
5893 gen_update_cc_op(s
);
5894 gen_jmp_im(pc_start
- s
->cs_base
);
5895 gen_helper_fbst_ST0(cpu_env
, cpu_A0
);
5896 gen_helper_fpop(cpu_env
);
5898 case 0x3d: /* fildll */
5899 tcg_gen_qemu_ld_i64(cpu_tmp1_i64
, cpu_A0
, s
->mem_index
, MO_LEQ
);
5900 gen_helper_fildll_ST0(cpu_env
, cpu_tmp1_i64
);
5902 case 0x3f: /* fistpll */
5903 gen_helper_fistll_ST0(cpu_tmp1_i64
, cpu_env
);
5904 tcg_gen_qemu_st_i64(cpu_tmp1_i64
, cpu_A0
, s
->mem_index
, MO_LEQ
);
5905 gen_helper_fpop(cpu_env
);
5911 /* register float ops */
5915 case 0x08: /* fld sti */
5916 gen_helper_fpush(cpu_env
);
5917 gen_helper_fmov_ST0_STN(cpu_env
,
5918 tcg_const_i32((opreg
+ 1) & 7));
5920 case 0x09: /* fxchg sti */
5921 case 0x29: /* fxchg4 sti, undocumented op */
5922 case 0x39: /* fxchg7 sti, undocumented op */
5923 gen_helper_fxchg_ST0_STN(cpu_env
, tcg_const_i32(opreg
));
5925 case 0x0a: /* grp d9/2 */
5928 /* check exceptions (FreeBSD FPU probe) */
5929 gen_update_cc_op(s
);
5930 gen_jmp_im(pc_start
- s
->cs_base
);
5931 gen_helper_fwait(cpu_env
);
5937 case 0x0c: /* grp d9/4 */
5940 gen_helper_fchs_ST0(cpu_env
);
5943 gen_helper_fabs_ST0(cpu_env
);
5946 gen_helper_fldz_FT0(cpu_env
);
5947 gen_helper_fcom_ST0_FT0(cpu_env
);
5950 gen_helper_fxam_ST0(cpu_env
);
5956 case 0x0d: /* grp d9/5 */
5960 gen_helper_fpush(cpu_env
);
5961 gen_helper_fld1_ST0(cpu_env
);
5964 gen_helper_fpush(cpu_env
);
5965 gen_helper_fldl2t_ST0(cpu_env
);
5968 gen_helper_fpush(cpu_env
);
5969 gen_helper_fldl2e_ST0(cpu_env
);
5972 gen_helper_fpush(cpu_env
);
5973 gen_helper_fldpi_ST0(cpu_env
);
5976 gen_helper_fpush(cpu_env
);
5977 gen_helper_fldlg2_ST0(cpu_env
);
5980 gen_helper_fpush(cpu_env
);
5981 gen_helper_fldln2_ST0(cpu_env
);
5984 gen_helper_fpush(cpu_env
);
5985 gen_helper_fldz_ST0(cpu_env
);
5992 case 0x0e: /* grp d9/6 */
5995 gen_helper_f2xm1(cpu_env
);
5998 gen_helper_fyl2x(cpu_env
);
6001 gen_helper_fptan(cpu_env
);
6003 case 3: /* fpatan */
6004 gen_helper_fpatan(cpu_env
);
6006 case 4: /* fxtract */
6007 gen_helper_fxtract(cpu_env
);
6009 case 5: /* fprem1 */
6010 gen_helper_fprem1(cpu_env
);
6012 case 6: /* fdecstp */
6013 gen_helper_fdecstp(cpu_env
);
6016 case 7: /* fincstp */
6017 gen_helper_fincstp(cpu_env
);
6021 case 0x0f: /* grp d9/7 */
6024 gen_helper_fprem(cpu_env
);
6026 case 1: /* fyl2xp1 */
6027 gen_helper_fyl2xp1(cpu_env
);
6030 gen_helper_fsqrt(cpu_env
);
6032 case 3: /* fsincos */
6033 gen_helper_fsincos(cpu_env
);
6035 case 5: /* fscale */
6036 gen_helper_fscale(cpu_env
);
6038 case 4: /* frndint */
6039 gen_helper_frndint(cpu_env
);
6042 gen_helper_fsin(cpu_env
);
6046 gen_helper_fcos(cpu_env
);
6050 case 0x00: case 0x01: case 0x04 ... 0x07: /* fxxx st, sti */
6051 case 0x20: case 0x21: case 0x24 ... 0x27: /* fxxx sti, st */
6052 case 0x30: case 0x31: case 0x34 ... 0x37: /* fxxxp sti, st */
6058 gen_helper_fp_arith_STN_ST0(op1
, opreg
);
6060 gen_helper_fpop(cpu_env
);
6062 gen_helper_fmov_FT0_STN(cpu_env
, tcg_const_i32(opreg
));
6063 gen_helper_fp_arith_ST0_FT0(op1
);
6067 case 0x02: /* fcom */
6068 case 0x22: /* fcom2, undocumented op */
6069 gen_helper_fmov_FT0_STN(cpu_env
, tcg_const_i32(opreg
));
6070 gen_helper_fcom_ST0_FT0(cpu_env
);
6072 case 0x03: /* fcomp */
6073 case 0x23: /* fcomp3, undocumented op */
6074 case 0x32: /* fcomp5, undocumented op */
6075 gen_helper_fmov_FT0_STN(cpu_env
, tcg_const_i32(opreg
));
6076 gen_helper_fcom_ST0_FT0(cpu_env
);
6077 gen_helper_fpop(cpu_env
);
6079 case 0x15: /* da/5 */
6081 case 1: /* fucompp */
6082 gen_helper_fmov_FT0_STN(cpu_env
, tcg_const_i32(1));
6083 gen_helper_fucom_ST0_FT0(cpu_env
);
6084 gen_helper_fpop(cpu_env
);
6085 gen_helper_fpop(cpu_env
);
6093 case 0: /* feni (287 only, just do nop here) */
6095 case 1: /* fdisi (287 only, just do nop here) */
6098 gen_helper_fclex(cpu_env
);
6100 case 3: /* fninit */
6101 gen_helper_fninit(cpu_env
);
6103 case 4: /* fsetpm (287 only, just do nop here) */
6109 case 0x1d: /* fucomi */
6110 if (!(s
->cpuid_features
& CPUID_CMOV
)) {
6113 gen_update_cc_op(s
);
6114 gen_helper_fmov_FT0_STN(cpu_env
, tcg_const_i32(opreg
));
6115 gen_helper_fucomi_ST0_FT0(cpu_env
);
6116 set_cc_op(s
, CC_OP_EFLAGS
);
6118 case 0x1e: /* fcomi */
6119 if (!(s
->cpuid_features
& CPUID_CMOV
)) {
6122 gen_update_cc_op(s
);
6123 gen_helper_fmov_FT0_STN(cpu_env
, tcg_const_i32(opreg
));
6124 gen_helper_fcomi_ST0_FT0(cpu_env
);
6125 set_cc_op(s
, CC_OP_EFLAGS
);
6127 case 0x28: /* ffree sti */
6128 gen_helper_ffree_STN(cpu_env
, tcg_const_i32(opreg
));
6130 case 0x2a: /* fst sti */
6131 gen_helper_fmov_STN_ST0(cpu_env
, tcg_const_i32(opreg
));
6133 case 0x2b: /* fstp sti */
6134 case 0x0b: /* fstp1 sti, undocumented op */
6135 case 0x3a: /* fstp8 sti, undocumented op */
6136 case 0x3b: /* fstp9 sti, undocumented op */
6137 gen_helper_fmov_STN_ST0(cpu_env
, tcg_const_i32(opreg
));
6138 gen_helper_fpop(cpu_env
);
6140 case 0x2c: /* fucom st(i) */
6141 gen_helper_fmov_FT0_STN(cpu_env
, tcg_const_i32(opreg
));
6142 gen_helper_fucom_ST0_FT0(cpu_env
);
6144 case 0x2d: /* fucomp st(i) */
6145 gen_helper_fmov_FT0_STN(cpu_env
, tcg_const_i32(opreg
));
6146 gen_helper_fucom_ST0_FT0(cpu_env
);
6147 gen_helper_fpop(cpu_env
);
6149 case 0x33: /* de/3 */
6151 case 1: /* fcompp */
6152 gen_helper_fmov_FT0_STN(cpu_env
, tcg_const_i32(1));
6153 gen_helper_fcom_ST0_FT0(cpu_env
);
6154 gen_helper_fpop(cpu_env
);
6155 gen_helper_fpop(cpu_env
);
6161 case 0x38: /* ffreep sti, undocumented op */
6162 gen_helper_ffree_STN(cpu_env
, tcg_const_i32(opreg
));
6163 gen_helper_fpop(cpu_env
);
6165 case 0x3c: /* df/4 */
6168 gen_helper_fnstsw(cpu_tmp2_i32
, cpu_env
);
6169 tcg_gen_extu_i32_tl(cpu_T
[0], cpu_tmp2_i32
);
6170 gen_op_mov_reg_v(MO_16
, R_EAX
, cpu_T
[0]);
6176 case 0x3d: /* fucomip */
6177 if (!(s
->cpuid_features
& CPUID_CMOV
)) {
6180 gen_update_cc_op(s
);
6181 gen_helper_fmov_FT0_STN(cpu_env
, tcg_const_i32(opreg
));
6182 gen_helper_fucomi_ST0_FT0(cpu_env
);
6183 gen_helper_fpop(cpu_env
);
6184 set_cc_op(s
, CC_OP_EFLAGS
);
6186 case 0x3e: /* fcomip */
6187 if (!(s
->cpuid_features
& CPUID_CMOV
)) {
6190 gen_update_cc_op(s
);
6191 gen_helper_fmov_FT0_STN(cpu_env
, tcg_const_i32(opreg
));
6192 gen_helper_fcomi_ST0_FT0(cpu_env
);
6193 gen_helper_fpop(cpu_env
);
6194 set_cc_op(s
, CC_OP_EFLAGS
);
6196 case 0x10 ... 0x13: /* fcmovxx */
6200 static const uint8_t fcmov_cc
[8] = {
6207 if (!(s
->cpuid_features
& CPUID_CMOV
)) {
6210 op1
= fcmov_cc
[op
& 3] | (((op
>> 3) & 1) ^ 1);
6211 l1
= gen_new_label();
6212 gen_jcc1_noeob(s
, op1
, l1
);
6213 gen_helper_fmov_ST0_STN(cpu_env
, tcg_const_i32(opreg
));
6222 /************************/
6225 case 0xa4: /* movsS */
6227 ot
= mo_b_d(b
, dflag
);
6228 if (prefixes
& (PREFIX_REPZ
| PREFIX_REPNZ
)) {
6229 gen_repz_movs(s
, ot
, pc_start
- s
->cs_base
, s
->pc
- s
->cs_base
);
6235 case 0xaa: /* stosS */
6237 ot
= mo_b_d(b
, dflag
);
6238 if (prefixes
& (PREFIX_REPZ
| PREFIX_REPNZ
)) {
6239 gen_repz_stos(s
, ot
, pc_start
- s
->cs_base
, s
->pc
- s
->cs_base
);
6244 case 0xac: /* lodsS */
6246 ot
= mo_b_d(b
, dflag
);
6247 if (prefixes
& (PREFIX_REPZ
| PREFIX_REPNZ
)) {
6248 gen_repz_lods(s
, ot
, pc_start
- s
->cs_base
, s
->pc
- s
->cs_base
);
6253 case 0xae: /* scasS */
6255 ot
= mo_b_d(b
, dflag
);
6256 if (prefixes
& PREFIX_REPNZ
) {
6257 gen_repz_scas(s
, ot
, pc_start
- s
->cs_base
, s
->pc
- s
->cs_base
, 1);
6258 } else if (prefixes
& PREFIX_REPZ
) {
6259 gen_repz_scas(s
, ot
, pc_start
- s
->cs_base
, s
->pc
- s
->cs_base
, 0);
6265 case 0xa6: /* cmpsS */
6267 ot
= mo_b_d(b
, dflag
);
6268 if (prefixes
& PREFIX_REPNZ
) {
6269 gen_repz_cmps(s
, ot
, pc_start
- s
->cs_base
, s
->pc
- s
->cs_base
, 1);
6270 } else if (prefixes
& PREFIX_REPZ
) {
6271 gen_repz_cmps(s
, ot
, pc_start
- s
->cs_base
, s
->pc
- s
->cs_base
, 0);
6276 case 0x6c: /* insS */
6278 ot
= mo_b_d32(b
, dflag
);
6279 tcg_gen_ext16u_tl(cpu_T
[0], cpu_regs
[R_EDX
]);
6280 gen_check_io(s
, ot
, pc_start
- s
->cs_base
,
6281 SVM_IOIO_TYPE_MASK
| svm_is_rep(prefixes
) | 4);
6282 if (prefixes
& (PREFIX_REPZ
| PREFIX_REPNZ
)) {
6283 gen_repz_ins(s
, ot
, pc_start
- s
->cs_base
, s
->pc
- s
->cs_base
);
6286 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
6287 gen_jmp(s
, s
->pc
- s
->cs_base
);
6291 case 0x6e: /* outsS */
6293 ot
= mo_b_d32(b
, dflag
);
6294 tcg_gen_ext16u_tl(cpu_T
[0], cpu_regs
[R_EDX
]);
6295 gen_check_io(s
, ot
, pc_start
- s
->cs_base
,
6296 svm_is_rep(prefixes
) | 4);
6297 if (prefixes
& (PREFIX_REPZ
| PREFIX_REPNZ
)) {
6298 gen_repz_outs(s
, ot
, pc_start
- s
->cs_base
, s
->pc
- s
->cs_base
);
6301 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
6302 gen_jmp(s
, s
->pc
- s
->cs_base
);
6307 /************************/
6312 ot
= mo_b_d32(b
, dflag
);
6313 val
= cpu_ldub_code(env
, s
->pc
++);
6314 tcg_gen_movi_tl(cpu_T
[0], val
);
6315 gen_check_io(s
, ot
, pc_start
- s
->cs_base
,
6316 SVM_IOIO_TYPE_MASK
| svm_is_rep(prefixes
));
6317 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
6320 tcg_gen_movi_i32(cpu_tmp2_i32
, val
);
6321 gen_helper_in_func(ot
, cpu_T
[1], cpu_tmp2_i32
);
6322 gen_op_mov_reg_v(ot
, R_EAX
, cpu_T
[1]);
6323 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
6325 gen_jmp(s
, s
->pc
- s
->cs_base
);
6330 ot
= mo_b_d32(b
, dflag
);
6331 val
= cpu_ldub_code(env
, s
->pc
++);
6332 tcg_gen_movi_tl(cpu_T
[0], val
);
6333 gen_check_io(s
, ot
, pc_start
- s
->cs_base
,
6334 svm_is_rep(prefixes
));
6335 gen_op_mov_v_reg(ot
, cpu_T
[1], R_EAX
);
6337 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
6340 tcg_gen_movi_i32(cpu_tmp2_i32
, val
);
6341 tcg_gen_trunc_tl_i32(cpu_tmp3_i32
, cpu_T
[1]);
6342 gen_helper_out_func(ot
, cpu_tmp2_i32
, cpu_tmp3_i32
);
6343 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
6345 gen_jmp(s
, s
->pc
- s
->cs_base
);
6350 ot
= mo_b_d32(b
, dflag
);
6351 tcg_gen_ext16u_tl(cpu_T
[0], cpu_regs
[R_EDX
]);
6352 gen_check_io(s
, ot
, pc_start
- s
->cs_base
,
6353 SVM_IOIO_TYPE_MASK
| svm_is_rep(prefixes
));
6354 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
6357 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
6358 gen_helper_in_func(ot
, cpu_T
[1], cpu_tmp2_i32
);
6359 gen_op_mov_reg_v(ot
, R_EAX
, cpu_T
[1]);
6360 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
6362 gen_jmp(s
, s
->pc
- s
->cs_base
);
6367 ot
= mo_b_d32(b
, dflag
);
6368 tcg_gen_ext16u_tl(cpu_T
[0], cpu_regs
[R_EDX
]);
6369 gen_check_io(s
, ot
, pc_start
- s
->cs_base
,
6370 svm_is_rep(prefixes
));
6371 gen_op_mov_v_reg(ot
, cpu_T
[1], R_EAX
);
6373 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
6376 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
6377 tcg_gen_trunc_tl_i32(cpu_tmp3_i32
, cpu_T
[1]);
6378 gen_helper_out_func(ot
, cpu_tmp2_i32
, cpu_tmp3_i32
);
6379 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
6381 gen_jmp(s
, s
->pc
- s
->cs_base
);
6385 /************************/
6387 case 0xc2: /* ret im */
6388 val
= cpu_ldsw_code(env
, s
->pc
);
6391 gen_stack_update(s
, val
+ (1 << ot
));
6392 /* Note that gen_pop_T0 uses a zero-extending load. */
6393 gen_op_jmp_v(cpu_T
[0]);
6396 case 0xc3: /* ret */
6398 gen_pop_update(s
, ot
);
6399 /* Note that gen_pop_T0 uses a zero-extending load. */
6400 gen_op_jmp_v(cpu_T
[0]);
6403 case 0xca: /* lret im */
6404 val
= cpu_ldsw_code(env
, s
->pc
);
6407 if (s
->pe
&& !s
->vm86
) {
6408 gen_update_cc_op(s
);
6409 gen_jmp_im(pc_start
- s
->cs_base
);
6410 gen_helper_lret_protected(cpu_env
, tcg_const_i32(dflag
- 1),
6411 tcg_const_i32(val
));
6415 gen_op_ld_v(s
, dflag
, cpu_T
[0], cpu_A0
);
6416 /* NOTE: keeping EIP updated is not a problem in case of
6418 gen_op_jmp_v(cpu_T
[0]);
6420 gen_op_addl_A0_im(1 << dflag
);
6421 gen_op_ld_v(s
, dflag
, cpu_T
[0], cpu_A0
);
6422 gen_op_movl_seg_T0_vm(R_CS
);
6423 /* add stack offset */
6424 gen_stack_update(s
, val
+ (2 << dflag
));
6428 case 0xcb: /* lret */
6431 case 0xcf: /* iret */
6432 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_IRET
);
6435 gen_helper_iret_real(cpu_env
, tcg_const_i32(dflag
- 1));
6436 set_cc_op(s
, CC_OP_EFLAGS
);
6437 } else if (s
->vm86
) {
6439 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
6441 gen_helper_iret_real(cpu_env
, tcg_const_i32(dflag
- 1));
6442 set_cc_op(s
, CC_OP_EFLAGS
);
6445 gen_update_cc_op(s
);
6446 gen_jmp_im(pc_start
- s
->cs_base
);
6447 gen_helper_iret_protected(cpu_env
, tcg_const_i32(dflag
- 1),
6448 tcg_const_i32(s
->pc
- s
->cs_base
));
6449 set_cc_op(s
, CC_OP_EFLAGS
);
6453 case 0xe8: /* call im */
6455 if (dflag
!= MO_16
) {
6456 tval
= (int32_t)insn_get(env
, s
, MO_32
);
6458 tval
= (int16_t)insn_get(env
, s
, MO_16
);
6460 next_eip
= s
->pc
- s
->cs_base
;
6462 if (dflag
== MO_16
) {
6464 } else if (!CODE64(s
)) {
6467 tcg_gen_movi_tl(cpu_T
[0], next_eip
);
6468 gen_push_v(s
, cpu_T
[0]);
6472 case 0x9a: /* lcall im */
6474 unsigned int selector
, offset
;
6479 offset
= insn_get(env
, s
, ot
);
6480 selector
= insn_get(env
, s
, MO_16
);
6482 tcg_gen_movi_tl(cpu_T
[0], selector
);
6483 tcg_gen_movi_tl(cpu_T
[1], offset
);
6486 case 0xe9: /* jmp im */
6487 if (dflag
!= MO_16
) {
6488 tval
= (int32_t)insn_get(env
, s
, MO_32
);
6490 tval
= (int16_t)insn_get(env
, s
, MO_16
);
6492 tval
+= s
->pc
- s
->cs_base
;
6493 if (dflag
== MO_16
) {
6495 } else if (!CODE64(s
)) {
6500 case 0xea: /* ljmp im */
6502 unsigned int selector
, offset
;
6507 offset
= insn_get(env
, s
, ot
);
6508 selector
= insn_get(env
, s
, MO_16
);
6510 tcg_gen_movi_tl(cpu_T
[0], selector
);
6511 tcg_gen_movi_tl(cpu_T
[1], offset
);
6514 case 0xeb: /* jmp Jb */
6515 tval
= (int8_t)insn_get(env
, s
, MO_8
);
6516 tval
+= s
->pc
- s
->cs_base
;
6517 if (dflag
== MO_16
) {
6522 case 0x70 ... 0x7f: /* jcc Jb */
6523 tval
= (int8_t)insn_get(env
, s
, MO_8
);
6525 case 0x180 ... 0x18f: /* jcc Jv */
6526 if (dflag
!= MO_16
) {
6527 tval
= (int32_t)insn_get(env
, s
, MO_32
);
6529 tval
= (int16_t)insn_get(env
, s
, MO_16
);
6532 next_eip
= s
->pc
- s
->cs_base
;
6534 if (dflag
== MO_16
) {
6537 gen_jcc(s
, b
, tval
, next_eip
);
6540 case 0x190 ... 0x19f: /* setcc Gv */
6541 modrm
= cpu_ldub_code(env
, s
->pc
++);
6542 gen_setcc1(s
, b
, cpu_T
[0]);
6543 gen_ldst_modrm(env
, s
, modrm
, MO_8
, OR_TMP0
, 1);
6545 case 0x140 ... 0x14f: /* cmov Gv, Ev */
6546 if (!(s
->cpuid_features
& CPUID_CMOV
)) {
6550 modrm
= cpu_ldub_code(env
, s
->pc
++);
6551 reg
= ((modrm
>> 3) & 7) | rex_r
;
6552 gen_cmovcc1(env
, s
, ot
, b
, modrm
, reg
);
6555 /************************/
6557 case 0x9c: /* pushf */
6558 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_PUSHF
);
6559 if (s
->vm86
&& s
->iopl
!= 3) {
6560 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
6562 gen_update_cc_op(s
);
6563 gen_helper_read_eflags(cpu_T
[0], cpu_env
);
6564 gen_push_v(s
, cpu_T
[0]);
6567 case 0x9d: /* popf */
6568 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_POPF
);
6569 if (s
->vm86
&& s
->iopl
!= 3) {
6570 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
6574 if (dflag
!= MO_16
) {
6575 gen_helper_write_eflags(cpu_env
, cpu_T
[0],
6576 tcg_const_i32((TF_MASK
| AC_MASK
|
6581 gen_helper_write_eflags(cpu_env
, cpu_T
[0],
6582 tcg_const_i32((TF_MASK
| AC_MASK
|
6584 IF_MASK
| IOPL_MASK
)
6588 if (s
->cpl
<= s
->iopl
) {
6589 if (dflag
!= MO_16
) {
6590 gen_helper_write_eflags(cpu_env
, cpu_T
[0],
6591 tcg_const_i32((TF_MASK
|
6597 gen_helper_write_eflags(cpu_env
, cpu_T
[0],
6598 tcg_const_i32((TF_MASK
|
6606 if (dflag
!= MO_16
) {
6607 gen_helper_write_eflags(cpu_env
, cpu_T
[0],
6608 tcg_const_i32((TF_MASK
| AC_MASK
|
6609 ID_MASK
| NT_MASK
)));
6611 gen_helper_write_eflags(cpu_env
, cpu_T
[0],
6612 tcg_const_i32((TF_MASK
| AC_MASK
|
6618 gen_pop_update(s
, ot
);
6619 set_cc_op(s
, CC_OP_EFLAGS
);
6620 /* abort translation because TF/AC flag may change */
6621 gen_jmp_im(s
->pc
- s
->cs_base
);
6625 case 0x9e: /* sahf */
6626 if (CODE64(s
) && !(s
->cpuid_ext3_features
& CPUID_EXT3_LAHF_LM
))
6628 gen_op_mov_v_reg(MO_8
, cpu_T
[0], R_AH
);
6629 gen_compute_eflags(s
);
6630 tcg_gen_andi_tl(cpu_cc_src
, cpu_cc_src
, CC_O
);
6631 tcg_gen_andi_tl(cpu_T
[0], cpu_T
[0], CC_S
| CC_Z
| CC_A
| CC_P
| CC_C
);
6632 tcg_gen_or_tl(cpu_cc_src
, cpu_cc_src
, cpu_T
[0]);
6634 case 0x9f: /* lahf */
6635 if (CODE64(s
) && !(s
->cpuid_ext3_features
& CPUID_EXT3_LAHF_LM
))
6637 gen_compute_eflags(s
);
6638 /* Note: gen_compute_eflags() only gives the condition codes */
6639 tcg_gen_ori_tl(cpu_T
[0], cpu_cc_src
, 0x02);
6640 gen_op_mov_reg_v(MO_8
, R_AH
, cpu_T
[0]);
6642 case 0xf5: /* cmc */
6643 gen_compute_eflags(s
);
6644 tcg_gen_xori_tl(cpu_cc_src
, cpu_cc_src
, CC_C
);
6646 case 0xf8: /* clc */
6647 gen_compute_eflags(s
);
6648 tcg_gen_andi_tl(cpu_cc_src
, cpu_cc_src
, ~CC_C
);
6650 case 0xf9: /* stc */
6651 gen_compute_eflags(s
);
6652 tcg_gen_ori_tl(cpu_cc_src
, cpu_cc_src
, CC_C
);
6654 case 0xfc: /* cld */
6655 tcg_gen_movi_i32(cpu_tmp2_i32
, 1);
6656 tcg_gen_st_i32(cpu_tmp2_i32
, cpu_env
, offsetof(CPUX86State
, df
));
6658 case 0xfd: /* std */
6659 tcg_gen_movi_i32(cpu_tmp2_i32
, -1);
6660 tcg_gen_st_i32(cpu_tmp2_i32
, cpu_env
, offsetof(CPUX86State
, df
));
6663 /************************/
6664 /* bit operations */
6665 case 0x1ba: /* bt/bts/btr/btc Gv, im */
6667 modrm
= cpu_ldub_code(env
, s
->pc
++);
6668 op
= (modrm
>> 3) & 7;
6669 mod
= (modrm
>> 6) & 3;
6670 rm
= (modrm
& 7) | REX_B(s
);
6673 gen_lea_modrm(env
, s
, modrm
);
6674 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
6676 gen_op_mov_v_reg(ot
, cpu_T
[0], rm
);
6679 val
= cpu_ldub_code(env
, s
->pc
++);
6680 tcg_gen_movi_tl(cpu_T
[1], val
);
6685 case 0x1a3: /* bt Gv, Ev */
6688 case 0x1ab: /* bts */
6691 case 0x1b3: /* btr */
6694 case 0x1bb: /* btc */
6698 modrm
= cpu_ldub_code(env
, s
->pc
++);
6699 reg
= ((modrm
>> 3) & 7) | rex_r
;
6700 mod
= (modrm
>> 6) & 3;
6701 rm
= (modrm
& 7) | REX_B(s
);
6702 gen_op_mov_v_reg(MO_32
, cpu_T
[1], reg
);
6704 gen_lea_modrm(env
, s
, modrm
);
6705 /* specific case: we need to add a displacement */
6706 gen_exts(ot
, cpu_T
[1]);
6707 tcg_gen_sari_tl(cpu_tmp0
, cpu_T
[1], 3 + ot
);
6708 tcg_gen_shli_tl(cpu_tmp0
, cpu_tmp0
, ot
);
6709 tcg_gen_add_tl(cpu_A0
, cpu_A0
, cpu_tmp0
);
6710 gen_op_ld_v(s
, ot
, cpu_T
[0], cpu_A0
);
6712 gen_op_mov_v_reg(ot
, cpu_T
[0], rm
);
6715 tcg_gen_andi_tl(cpu_T
[1], cpu_T
[1], (1 << (3 + ot
)) - 1);
6716 tcg_gen_shr_tl(cpu_tmp4
, cpu_T
[0], cpu_T
[1]);
6721 tcg_gen_movi_tl(cpu_tmp0
, 1);
6722 tcg_gen_shl_tl(cpu_tmp0
, cpu_tmp0
, cpu_T
[1]);
6723 tcg_gen_or_tl(cpu_T
[0], cpu_T
[0], cpu_tmp0
);
6726 tcg_gen_movi_tl(cpu_tmp0
, 1);
6727 tcg_gen_shl_tl(cpu_tmp0
, cpu_tmp0
, cpu_T
[1]);
6728 tcg_gen_andc_tl(cpu_T
[0], cpu_T
[0], cpu_tmp0
);
6732 tcg_gen_movi_tl(cpu_tmp0
, 1);
6733 tcg_gen_shl_tl(cpu_tmp0
, cpu_tmp0
, cpu_T
[1]);
6734 tcg_gen_xor_tl(cpu_T
[0], cpu_T
[0], cpu_tmp0
);
6739 gen_op_st_v(s
, ot
, cpu_T
[0], cpu_A0
);
6741 gen_op_mov_reg_v(ot
, rm
, cpu_T
[0]);
6745 /* Delay all CC updates until after the store above. Note that
6746 C is the result of the test, Z is unchanged, and the others
6747 are all undefined. */
6749 case CC_OP_MULB
... CC_OP_MULQ
:
6750 case CC_OP_ADDB
... CC_OP_ADDQ
:
6751 case CC_OP_ADCB
... CC_OP_ADCQ
:
6752 case CC_OP_SUBB
... CC_OP_SUBQ
:
6753 case CC_OP_SBBB
... CC_OP_SBBQ
:
6754 case CC_OP_LOGICB
... CC_OP_LOGICQ
:
6755 case CC_OP_INCB
... CC_OP_INCQ
:
6756 case CC_OP_DECB
... CC_OP_DECQ
:
6757 case CC_OP_SHLB
... CC_OP_SHLQ
:
6758 case CC_OP_SARB
... CC_OP_SARQ
:
6759 case CC_OP_BMILGB
... CC_OP_BMILGQ
:
6760 /* Z was going to be computed from the non-zero status of CC_DST.
6761 We can get that same Z value (and the new C value) by leaving
6762 CC_DST alone, setting CC_SRC, and using a CC_OP_SAR of the
6764 tcg_gen_mov_tl(cpu_cc_src
, cpu_tmp4
);
6765 set_cc_op(s
, ((s
->cc_op
- CC_OP_MULB
) & 3) + CC_OP_SARB
);
6768 /* Otherwise, generate EFLAGS and replace the C bit. */
6769 gen_compute_eflags(s
);
6770 tcg_gen_deposit_tl(cpu_cc_src
, cpu_cc_src
, cpu_tmp4
,
6775 case 0x1bc: /* bsf / tzcnt */
6776 case 0x1bd: /* bsr / lzcnt */
6778 modrm
= cpu_ldub_code(env
, s
->pc
++);
6779 reg
= ((modrm
>> 3) & 7) | rex_r
;
6780 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
6781 gen_extu(ot
, cpu_T
[0]);
6783 /* Note that lzcnt and tzcnt are in different extensions. */
6784 if ((prefixes
& PREFIX_REPZ
)
6786 ? s
->cpuid_ext3_features
& CPUID_EXT3_ABM
6787 : s
->cpuid_7_0_ebx_features
& CPUID_7_0_EBX_BMI1
)) {
6789 tcg_gen_mov_tl(cpu_cc_src
, cpu_T
[0]);
6791 /* For lzcnt, reduce the target_ulong result by the
6792 number of zeros that we expect to find at the top. */
6793 gen_helper_clz(cpu_T
[0], cpu_T
[0]);
6794 tcg_gen_subi_tl(cpu_T
[0], cpu_T
[0], TARGET_LONG_BITS
- size
);
6796 /* For tzcnt, a zero input must return the operand size:
6797 force all bits outside the operand size to 1. */
6798 target_ulong mask
= (target_ulong
)-2 << (size
- 1);
6799 tcg_gen_ori_tl(cpu_T
[0], cpu_T
[0], mask
);
6800 gen_helper_ctz(cpu_T
[0], cpu_T
[0]);
6802 /* For lzcnt/tzcnt, C and Z bits are defined and are
6803 related to the result. */
6804 gen_op_update1_cc();
6805 set_cc_op(s
, CC_OP_BMILGB
+ ot
);
6807 /* For bsr/bsf, only the Z bit is defined and it is related
6808 to the input and not the result. */
6809 tcg_gen_mov_tl(cpu_cc_dst
, cpu_T
[0]);
6810 set_cc_op(s
, CC_OP_LOGICB
+ ot
);
6812 /* For bsr, return the bit index of the first 1 bit,
6813 not the count of leading zeros. */
6814 gen_helper_clz(cpu_T
[0], cpu_T
[0]);
6815 tcg_gen_xori_tl(cpu_T
[0], cpu_T
[0], TARGET_LONG_BITS
- 1);
6817 gen_helper_ctz(cpu_T
[0], cpu_T
[0]);
6819 /* ??? The manual says that the output is undefined when the
6820 input is zero, but real hardware leaves it unchanged, and
6821 real programs appear to depend on that. */
6822 tcg_gen_movi_tl(cpu_tmp0
, 0);
6823 tcg_gen_movcond_tl(TCG_COND_EQ
, cpu_T
[0], cpu_cc_dst
, cpu_tmp0
,
6824 cpu_regs
[reg
], cpu_T
[0]);
6826 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
6828 /************************/
6830 case 0x27: /* daa */
6833 gen_update_cc_op(s
);
6834 gen_helper_daa(cpu_env
);
6835 set_cc_op(s
, CC_OP_EFLAGS
);
6837 case 0x2f: /* das */
6840 gen_update_cc_op(s
);
6841 gen_helper_das(cpu_env
);
6842 set_cc_op(s
, CC_OP_EFLAGS
);
6844 case 0x37: /* aaa */
6847 gen_update_cc_op(s
);
6848 gen_helper_aaa(cpu_env
);
6849 set_cc_op(s
, CC_OP_EFLAGS
);
6851 case 0x3f: /* aas */
6854 gen_update_cc_op(s
);
6855 gen_helper_aas(cpu_env
);
6856 set_cc_op(s
, CC_OP_EFLAGS
);
6858 case 0xd4: /* aam */
6861 val
= cpu_ldub_code(env
, s
->pc
++);
6863 gen_exception(s
, EXCP00_DIVZ
, pc_start
- s
->cs_base
);
6865 gen_helper_aam(cpu_env
, tcg_const_i32(val
));
6866 set_cc_op(s
, CC_OP_LOGICB
);
6869 case 0xd5: /* aad */
6872 val
= cpu_ldub_code(env
, s
->pc
++);
6873 gen_helper_aad(cpu_env
, tcg_const_i32(val
));
6874 set_cc_op(s
, CC_OP_LOGICB
);
6876 /************************/
6878 case 0x90: /* nop */
6879 /* XXX: correct lock test for all insn */
6880 if (prefixes
& PREFIX_LOCK
) {
6883 /* If REX_B is set, then this is xchg eax, r8d, not a nop. */
6885 goto do_xchg_reg_eax
;
6887 if (prefixes
& PREFIX_REPZ
) {
6888 gen_update_cc_op(s
);
6889 gen_jmp_im(pc_start
- s
->cs_base
);
6890 gen_helper_pause(cpu_env
, tcg_const_i32(s
->pc
- pc_start
));
6891 s
->is_jmp
= DISAS_TB_JUMP
;
6894 case 0x9b: /* fwait */
6895 if ((s
->flags
& (HF_MP_MASK
| HF_TS_MASK
)) ==
6896 (HF_MP_MASK
| HF_TS_MASK
)) {
6897 gen_exception(s
, EXCP07_PREX
, pc_start
- s
->cs_base
);
6899 gen_update_cc_op(s
);
6900 gen_jmp_im(pc_start
- s
->cs_base
);
6901 gen_helper_fwait(cpu_env
);
6904 case 0xcc: /* int3 */
6905 gen_interrupt(s
, EXCP03_INT3
, pc_start
- s
->cs_base
, s
->pc
- s
->cs_base
);
6907 case 0xcd: /* int N */
6908 val
= cpu_ldub_code(env
, s
->pc
++);
6909 if (s
->vm86
&& s
->iopl
!= 3) {
6910 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
6912 gen_interrupt(s
, val
, pc_start
- s
->cs_base
, s
->pc
- s
->cs_base
);
6915 case 0xce: /* into */
6918 gen_update_cc_op(s
);
6919 gen_jmp_im(pc_start
- s
->cs_base
);
6920 gen_helper_into(cpu_env
, tcg_const_i32(s
->pc
- pc_start
));
6923 case 0xf1: /* icebp (undocumented, exits to external debugger) */
6924 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_ICEBP
);
6926 gen_debug(s
, pc_start
- s
->cs_base
);
6930 qemu_set_log(CPU_LOG_INT
| CPU_LOG_TB_IN_ASM
);
6934 case 0xfa: /* cli */
6936 if (s
->cpl
<= s
->iopl
) {
6937 gen_helper_cli(cpu_env
);
6939 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
6943 gen_helper_cli(cpu_env
);
6945 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
6949 case 0xfb: /* sti */
6951 if (s
->cpl
<= s
->iopl
) {
6953 gen_helper_sti(cpu_env
);
6954 /* interruptions are enabled only the first insn after sti */
6955 /* If several instructions disable interrupts, only the
6957 if (!(s
->tb
->flags
& HF_INHIBIT_IRQ_MASK
))
6958 gen_helper_set_inhibit_irq(cpu_env
);
6959 /* give a chance to handle pending irqs */
6960 gen_jmp_im(s
->pc
- s
->cs_base
);
6963 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
6969 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
6973 case 0x62: /* bound */
6977 modrm
= cpu_ldub_code(env
, s
->pc
++);
6978 reg
= (modrm
>> 3) & 7;
6979 mod
= (modrm
>> 6) & 3;
6982 gen_op_mov_v_reg(ot
, cpu_T
[0], reg
);
6983 gen_lea_modrm(env
, s
, modrm
);
6984 gen_jmp_im(pc_start
- s
->cs_base
);
6985 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
6987 gen_helper_boundw(cpu_env
, cpu_A0
, cpu_tmp2_i32
);
6989 gen_helper_boundl(cpu_env
, cpu_A0
, cpu_tmp2_i32
);
6992 case 0x1c8 ... 0x1cf: /* bswap reg */
6993 reg
= (b
& 7) | REX_B(s
);
6994 #ifdef TARGET_X86_64
6995 if (dflag
== MO_64
) {
6996 gen_op_mov_v_reg(MO_64
, cpu_T
[0], reg
);
6997 tcg_gen_bswap64_i64(cpu_T
[0], cpu_T
[0]);
6998 gen_op_mov_reg_v(MO_64
, reg
, cpu_T
[0]);
7002 gen_op_mov_v_reg(MO_32
, cpu_T
[0], reg
);
7003 tcg_gen_ext32u_tl(cpu_T
[0], cpu_T
[0]);
7004 tcg_gen_bswap32_tl(cpu_T
[0], cpu_T
[0]);
7005 gen_op_mov_reg_v(MO_32
, reg
, cpu_T
[0]);
7008 case 0xd6: /* salc */
7011 gen_compute_eflags_c(s
, cpu_T
[0]);
7012 tcg_gen_neg_tl(cpu_T
[0], cpu_T
[0]);
7013 gen_op_mov_reg_v(MO_8
, R_EAX
, cpu_T
[0]);
7015 case 0xe0: /* loopnz */
7016 case 0xe1: /* loopz */
7017 case 0xe2: /* loop */
7018 case 0xe3: /* jecxz */
7022 tval
= (int8_t)insn_get(env
, s
, MO_8
);
7023 next_eip
= s
->pc
- s
->cs_base
;
7025 if (dflag
== MO_16
) {
7029 l1
= gen_new_label();
7030 l2
= gen_new_label();
7031 l3
= gen_new_label();
7034 case 0: /* loopnz */
7036 gen_op_add_reg_im(s
->aflag
, R_ECX
, -1);
7037 gen_op_jz_ecx(s
->aflag
, l3
);
7038 gen_jcc1(s
, (JCC_Z
<< 1) | (b
^ 1), l1
);
7041 gen_op_add_reg_im(s
->aflag
, R_ECX
, -1);
7042 gen_op_jnz_ecx(s
->aflag
, l1
);
7046 gen_op_jz_ecx(s
->aflag
, l1
);
7051 gen_jmp_im(next_eip
);
7060 case 0x130: /* wrmsr */
7061 case 0x132: /* rdmsr */
7063 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7065 gen_update_cc_op(s
);
7066 gen_jmp_im(pc_start
- s
->cs_base
);
7068 gen_helper_rdmsr(cpu_env
);
7070 gen_helper_wrmsr(cpu_env
);
7074 case 0x131: /* rdtsc */
7075 gen_update_cc_op(s
);
7076 gen_jmp_im(pc_start
- s
->cs_base
);
7077 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
7080 gen_helper_rdtsc(cpu_env
);
7081 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
7083 gen_jmp(s
, s
->pc
- s
->cs_base
);
7086 case 0x133: /* rdpmc */
7087 gen_update_cc_op(s
);
7088 gen_jmp_im(pc_start
- s
->cs_base
);
7089 gen_helper_rdpmc(cpu_env
);
7091 case 0x134: /* sysenter */
7092 /* For Intel SYSENTER is valid on 64-bit */
7093 if (CODE64(s
) && env
->cpuid_vendor1
!= CPUID_VENDOR_INTEL_1
)
7096 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7098 gen_update_cc_op(s
);
7099 gen_jmp_im(pc_start
- s
->cs_base
);
7100 gen_helper_sysenter(cpu_env
);
7104 case 0x135: /* sysexit */
7105 /* For Intel SYSEXIT is valid on 64-bit */
7106 if (CODE64(s
) && env
->cpuid_vendor1
!= CPUID_VENDOR_INTEL_1
)
7109 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7111 gen_update_cc_op(s
);
7112 gen_jmp_im(pc_start
- s
->cs_base
);
7113 gen_helper_sysexit(cpu_env
, tcg_const_i32(dflag
- 1));
7117 #ifdef TARGET_X86_64
7118 case 0x105: /* syscall */
7119 /* XXX: is it usable in real mode ? */
7120 gen_update_cc_op(s
);
7121 gen_jmp_im(pc_start
- s
->cs_base
);
7122 gen_helper_syscall(cpu_env
, tcg_const_i32(s
->pc
- pc_start
));
7125 case 0x107: /* sysret */
7127 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7129 gen_update_cc_op(s
);
7130 gen_jmp_im(pc_start
- s
->cs_base
);
7131 gen_helper_sysret(cpu_env
, tcg_const_i32(dflag
- 1));
7132 /* condition codes are modified only in long mode */
7134 set_cc_op(s
, CC_OP_EFLAGS
);
7140 case 0x1a2: /* cpuid */
7141 gen_update_cc_op(s
);
7142 gen_jmp_im(pc_start
- s
->cs_base
);
7143 gen_helper_cpuid(cpu_env
);
7145 case 0xf4: /* hlt */
7147 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7149 gen_update_cc_op(s
);
7150 gen_jmp_im(pc_start
- s
->cs_base
);
7151 gen_helper_hlt(cpu_env
, tcg_const_i32(s
->pc
- pc_start
));
7152 s
->is_jmp
= DISAS_TB_JUMP
;
7156 modrm
= cpu_ldub_code(env
, s
->pc
++);
7157 mod
= (modrm
>> 6) & 3;
7158 op
= (modrm
>> 3) & 7;
7161 if (!s
->pe
|| s
->vm86
)
7163 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_LDTR_READ
);
7164 tcg_gen_ld32u_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,ldt
.selector
));
7165 ot
= mod
== 3 ? dflag
: MO_16
;
7166 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 1);
7169 if (!s
->pe
|| s
->vm86
)
7172 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7174 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_LDTR_WRITE
);
7175 gen_ldst_modrm(env
, s
, modrm
, MO_16
, OR_TMP0
, 0);
7176 gen_jmp_im(pc_start
- s
->cs_base
);
7177 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
7178 gen_helper_lldt(cpu_env
, cpu_tmp2_i32
);
7182 if (!s
->pe
|| s
->vm86
)
7184 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_TR_READ
);
7185 tcg_gen_ld32u_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,tr
.selector
));
7186 ot
= mod
== 3 ? dflag
: MO_16
;
7187 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 1);
7190 if (!s
->pe
|| s
->vm86
)
7193 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7195 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_TR_WRITE
);
7196 gen_ldst_modrm(env
, s
, modrm
, MO_16
, OR_TMP0
, 0);
7197 gen_jmp_im(pc_start
- s
->cs_base
);
7198 tcg_gen_trunc_tl_i32(cpu_tmp2_i32
, cpu_T
[0]);
7199 gen_helper_ltr(cpu_env
, cpu_tmp2_i32
);
7204 if (!s
->pe
|| s
->vm86
)
7206 gen_ldst_modrm(env
, s
, modrm
, MO_16
, OR_TMP0
, 0);
7207 gen_update_cc_op(s
);
7209 gen_helper_verr(cpu_env
, cpu_T
[0]);
7211 gen_helper_verw(cpu_env
, cpu_T
[0]);
7213 set_cc_op(s
, CC_OP_EFLAGS
);
7220 modrm
= cpu_ldub_code(env
, s
->pc
++);
7221 mod
= (modrm
>> 6) & 3;
7222 op
= (modrm
>> 3) & 7;
7228 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_GDTR_READ
);
7229 gen_lea_modrm(env
, s
, modrm
);
7230 tcg_gen_ld32u_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
, gdt
.limit
));
7231 gen_op_st_v(s
, MO_16
, cpu_T
[0], cpu_A0
);
7232 gen_add_A0_im(s
, 2);
7233 tcg_gen_ld_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
, gdt
.base
));
7234 if (dflag
== MO_16
) {
7235 tcg_gen_andi_tl(cpu_T
[0], cpu_T
[0], 0xffffff);
7237 gen_op_st_v(s
, CODE64(s
) + MO_32
, cpu_T
[0], cpu_A0
);
7242 case 0: /* monitor */
7243 if (!(s
->cpuid_ext_features
& CPUID_EXT_MONITOR
) ||
7246 gen_update_cc_op(s
);
7247 gen_jmp_im(pc_start
- s
->cs_base
);
7248 tcg_gen_mov_tl(cpu_A0
, cpu_regs
[R_EAX
]);
7249 gen_extu(s
->aflag
, cpu_A0
);
7250 gen_add_A0_ds_seg(s
);
7251 gen_helper_monitor(cpu_env
, cpu_A0
);
7254 if (!(s
->cpuid_ext_features
& CPUID_EXT_MONITOR
) ||
7257 gen_update_cc_op(s
);
7258 gen_jmp_im(pc_start
- s
->cs_base
);
7259 gen_helper_mwait(cpu_env
, tcg_const_i32(s
->pc
- pc_start
));
7263 if (!(s
->cpuid_7_0_ebx_features
& CPUID_7_0_EBX_SMAP
) ||
7267 gen_helper_clac(cpu_env
);
7268 gen_jmp_im(s
->pc
- s
->cs_base
);
7272 if (!(s
->cpuid_7_0_ebx_features
& CPUID_7_0_EBX_SMAP
) ||
7276 gen_helper_stac(cpu_env
);
7277 gen_jmp_im(s
->pc
- s
->cs_base
);
7284 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_IDTR_READ
);
7285 gen_lea_modrm(env
, s
, modrm
);
7286 tcg_gen_ld32u_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
, idt
.limit
));
7287 gen_op_st_v(s
, MO_16
, cpu_T
[0], cpu_A0
);
7288 gen_add_A0_im(s
, 2);
7289 tcg_gen_ld_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
, idt
.base
));
7290 if (dflag
== MO_16
) {
7291 tcg_gen_andi_tl(cpu_T
[0], cpu_T
[0], 0xffffff);
7293 gen_op_st_v(s
, CODE64(s
) + MO_32
, cpu_T
[0], cpu_A0
);
7299 gen_update_cc_op(s
);
7300 gen_jmp_im(pc_start
- s
->cs_base
);
7303 if (!(s
->flags
& HF_SVME_MASK
) || !s
->pe
)
7306 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7309 gen_helper_vmrun(cpu_env
, tcg_const_i32(s
->aflag
- 1),
7310 tcg_const_i32(s
->pc
- pc_start
));
7312 s
->is_jmp
= DISAS_TB_JUMP
;
7315 case 1: /* VMMCALL */
7316 if (!(s
->flags
& HF_SVME_MASK
))
7318 gen_helper_vmmcall(cpu_env
);
7320 case 2: /* VMLOAD */
7321 if (!(s
->flags
& HF_SVME_MASK
) || !s
->pe
)
7324 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7327 gen_helper_vmload(cpu_env
, tcg_const_i32(s
->aflag
- 1));
7330 case 3: /* VMSAVE */
7331 if (!(s
->flags
& HF_SVME_MASK
) || !s
->pe
)
7334 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7337 gen_helper_vmsave(cpu_env
, tcg_const_i32(s
->aflag
- 1));
7341 if ((!(s
->flags
& HF_SVME_MASK
) &&
7342 !(s
->cpuid_ext3_features
& CPUID_EXT3_SKINIT
)) ||
7346 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7349 gen_helper_stgi(cpu_env
);
7353 if (!(s
->flags
& HF_SVME_MASK
) || !s
->pe
)
7356 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7359 gen_helper_clgi(cpu_env
);
7362 case 6: /* SKINIT */
7363 if ((!(s
->flags
& HF_SVME_MASK
) &&
7364 !(s
->cpuid_ext3_features
& CPUID_EXT3_SKINIT
)) ||
7367 gen_helper_skinit(cpu_env
);
7369 case 7: /* INVLPGA */
7370 if (!(s
->flags
& HF_SVME_MASK
) || !s
->pe
)
7373 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7376 gen_helper_invlpga(cpu_env
,
7377 tcg_const_i32(s
->aflag
- 1));
7383 } else if (s
->cpl
!= 0) {
7384 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7386 gen_svm_check_intercept(s
, pc_start
,
7387 op
==2 ? SVM_EXIT_GDTR_WRITE
: SVM_EXIT_IDTR_WRITE
);
7388 gen_lea_modrm(env
, s
, modrm
);
7389 gen_op_ld_v(s
, MO_16
, cpu_T
[1], cpu_A0
);
7390 gen_add_A0_im(s
, 2);
7391 gen_op_ld_v(s
, CODE64(s
) + MO_32
, cpu_T
[0], cpu_A0
);
7392 if (dflag
== MO_16
) {
7393 tcg_gen_andi_tl(cpu_T
[0], cpu_T
[0], 0xffffff);
7396 tcg_gen_st_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,gdt
.base
));
7397 tcg_gen_st32_tl(cpu_T
[1], cpu_env
, offsetof(CPUX86State
,gdt
.limit
));
7399 tcg_gen_st_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,idt
.base
));
7400 tcg_gen_st32_tl(cpu_T
[1], cpu_env
, offsetof(CPUX86State
,idt
.limit
));
7405 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_READ_CR0
);
7406 #if defined TARGET_X86_64 && defined HOST_WORDS_BIGENDIAN
7407 tcg_gen_ld32u_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,cr
[0]) + 4);
7409 tcg_gen_ld32u_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,cr
[0]));
7411 gen_ldst_modrm(env
, s
, modrm
, MO_16
, OR_TMP0
, 1);
7415 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7417 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_WRITE_CR0
);
7418 gen_ldst_modrm(env
, s
, modrm
, MO_16
, OR_TMP0
, 0);
7419 gen_helper_lmsw(cpu_env
, cpu_T
[0]);
7420 gen_jmp_im(s
->pc
- s
->cs_base
);
7425 if (mod
!= 3) { /* invlpg */
7427 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7429 gen_update_cc_op(s
);
7430 gen_jmp_im(pc_start
- s
->cs_base
);
7431 gen_lea_modrm(env
, s
, modrm
);
7432 gen_helper_invlpg(cpu_env
, cpu_A0
);
7433 gen_jmp_im(s
->pc
- s
->cs_base
);
7438 case 0: /* swapgs */
7439 #ifdef TARGET_X86_64
7442 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7444 tcg_gen_ld_tl(cpu_T
[0], cpu_env
,
7445 offsetof(CPUX86State
,segs
[R_GS
].base
));
7446 tcg_gen_ld_tl(cpu_T
[1], cpu_env
,
7447 offsetof(CPUX86State
,kernelgsbase
));
7448 tcg_gen_st_tl(cpu_T
[1], cpu_env
,
7449 offsetof(CPUX86State
,segs
[R_GS
].base
));
7450 tcg_gen_st_tl(cpu_T
[0], cpu_env
,
7451 offsetof(CPUX86State
,kernelgsbase
));
7459 case 1: /* rdtscp */
7460 if (!(s
->cpuid_ext2_features
& CPUID_EXT2_RDTSCP
))
7462 gen_update_cc_op(s
);
7463 gen_jmp_im(pc_start
- s
->cs_base
);
7464 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
7467 gen_helper_rdtscp(cpu_env
);
7468 if (s
->tb
->cflags
& CF_USE_ICOUNT
) {
7470 gen_jmp(s
, s
->pc
- s
->cs_base
);
7482 case 0x108: /* invd */
7483 case 0x109: /* wbinvd */
7485 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7487 gen_svm_check_intercept(s
, pc_start
, (b
& 2) ? SVM_EXIT_INVD
: SVM_EXIT_WBINVD
);
7491 case 0x63: /* arpl or movslS (x86_64) */
7492 #ifdef TARGET_X86_64
7495 /* d_ot is the size of destination */
7498 modrm
= cpu_ldub_code(env
, s
->pc
++);
7499 reg
= ((modrm
>> 3) & 7) | rex_r
;
7500 mod
= (modrm
>> 6) & 3;
7501 rm
= (modrm
& 7) | REX_B(s
);
7504 gen_op_mov_v_reg(MO_32
, cpu_T
[0], rm
);
7506 if (d_ot
== MO_64
) {
7507 tcg_gen_ext32s_tl(cpu_T
[0], cpu_T
[0]);
7509 gen_op_mov_reg_v(d_ot
, reg
, cpu_T
[0]);
7511 gen_lea_modrm(env
, s
, modrm
);
7512 gen_op_ld_v(s
, MO_32
| MO_SIGN
, cpu_T
[0], cpu_A0
);
7513 gen_op_mov_reg_v(d_ot
, reg
, cpu_T
[0]);
7519 TCGv t0
, t1
, t2
, a0
;
7521 if (!s
->pe
|| s
->vm86
)
7523 t0
= tcg_temp_local_new();
7524 t1
= tcg_temp_local_new();
7525 t2
= tcg_temp_local_new();
7527 modrm
= cpu_ldub_code(env
, s
->pc
++);
7528 reg
= (modrm
>> 3) & 7;
7529 mod
= (modrm
>> 6) & 3;
7532 gen_lea_modrm(env
, s
, modrm
);
7533 gen_op_ld_v(s
, ot
, t0
, cpu_A0
);
7534 a0
= tcg_temp_local_new();
7535 tcg_gen_mov_tl(a0
, cpu_A0
);
7537 gen_op_mov_v_reg(ot
, t0
, rm
);
7540 gen_op_mov_v_reg(ot
, t1
, reg
);
7541 tcg_gen_andi_tl(cpu_tmp0
, t0
, 3);
7542 tcg_gen_andi_tl(t1
, t1
, 3);
7543 tcg_gen_movi_tl(t2
, 0);
7544 label1
= gen_new_label();
7545 tcg_gen_brcond_tl(TCG_COND_GE
, cpu_tmp0
, t1
, label1
);
7546 tcg_gen_andi_tl(t0
, t0
, ~3);
7547 tcg_gen_or_tl(t0
, t0
, t1
);
7548 tcg_gen_movi_tl(t2
, CC_Z
);
7549 gen_set_label(label1
);
7551 gen_op_st_v(s
, ot
, t0
, a0
);
7554 gen_op_mov_reg_v(ot
, rm
, t0
);
7556 gen_compute_eflags(s
);
7557 tcg_gen_andi_tl(cpu_cc_src
, cpu_cc_src
, ~CC_Z
);
7558 tcg_gen_or_tl(cpu_cc_src
, cpu_cc_src
, t2
);
7564 case 0x102: /* lar */
7565 case 0x103: /* lsl */
7569 if (!s
->pe
|| s
->vm86
)
7571 ot
= dflag
!= MO_16
? MO_32
: MO_16
;
7572 modrm
= cpu_ldub_code(env
, s
->pc
++);
7573 reg
= ((modrm
>> 3) & 7) | rex_r
;
7574 gen_ldst_modrm(env
, s
, modrm
, MO_16
, OR_TMP0
, 0);
7575 t0
= tcg_temp_local_new();
7576 gen_update_cc_op(s
);
7578 gen_helper_lar(t0
, cpu_env
, cpu_T
[0]);
7580 gen_helper_lsl(t0
, cpu_env
, cpu_T
[0]);
7582 tcg_gen_andi_tl(cpu_tmp0
, cpu_cc_src
, CC_Z
);
7583 label1
= gen_new_label();
7584 tcg_gen_brcondi_tl(TCG_COND_EQ
, cpu_tmp0
, 0, label1
);
7585 gen_op_mov_reg_v(ot
, reg
, t0
);
7586 gen_set_label(label1
);
7587 set_cc_op(s
, CC_OP_EFLAGS
);
7592 modrm
= cpu_ldub_code(env
, s
->pc
++);
7593 mod
= (modrm
>> 6) & 3;
7594 op
= (modrm
>> 3) & 7;
7596 case 0: /* prefetchnta */
7597 case 1: /* prefetchnt0 */
7598 case 2: /* prefetchnt0 */
7599 case 3: /* prefetchnt0 */
7602 gen_lea_modrm(env
, s
, modrm
);
7603 /* nothing more to do */
7605 default: /* nop (multi byte) */
7606 gen_nop_modrm(env
, s
, modrm
);
7610 case 0x119 ... 0x11f: /* nop (multi byte) */
7611 modrm
= cpu_ldub_code(env
, s
->pc
++);
7612 gen_nop_modrm(env
, s
, modrm
);
7614 case 0x120: /* mov reg, crN */
7615 case 0x122: /* mov crN, reg */
7617 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7619 modrm
= cpu_ldub_code(env
, s
->pc
++);
7620 /* Ignore the mod bits (assume (modrm&0xc0)==0xc0).
7621 * AMD documentation (24594.pdf) and testing of
7622 * intel 386 and 486 processors all show that the mod bits
7623 * are assumed to be 1's, regardless of actual values.
7625 rm
= (modrm
& 7) | REX_B(s
);
7626 reg
= ((modrm
>> 3) & 7) | rex_r
;
7631 if ((prefixes
& PREFIX_LOCK
) && (reg
== 0) &&
7632 (s
->cpuid_ext3_features
& CPUID_EXT3_CR8LEG
)) {
7641 gen_update_cc_op(s
);
7642 gen_jmp_im(pc_start
- s
->cs_base
);
7644 gen_op_mov_v_reg(ot
, cpu_T
[0], rm
);
7645 gen_helper_write_crN(cpu_env
, tcg_const_i32(reg
),
7647 gen_jmp_im(s
->pc
- s
->cs_base
);
7650 gen_helper_read_crN(cpu_T
[0], cpu_env
, tcg_const_i32(reg
));
7651 gen_op_mov_reg_v(ot
, rm
, cpu_T
[0]);
7659 case 0x121: /* mov reg, drN */
7660 case 0x123: /* mov drN, reg */
7662 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7664 modrm
= cpu_ldub_code(env
, s
->pc
++);
7665 /* Ignore the mod bits (assume (modrm&0xc0)==0xc0).
7666 * AMD documentation (24594.pdf) and testing of
7667 * intel 386 and 486 processors all show that the mod bits
7668 * are assumed to be 1's, regardless of actual values.
7670 rm
= (modrm
& 7) | REX_B(s
);
7671 reg
= ((modrm
>> 3) & 7) | rex_r
;
7676 /* XXX: do it dynamically with CR4.DE bit */
7677 if (reg
== 4 || reg
== 5 || reg
>= 8)
7680 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_WRITE_DR0
+ reg
);
7681 gen_op_mov_v_reg(ot
, cpu_T
[0], rm
);
7682 gen_helper_movl_drN_T0(cpu_env
, tcg_const_i32(reg
), cpu_T
[0]);
7683 gen_jmp_im(s
->pc
- s
->cs_base
);
7686 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_READ_DR0
+ reg
);
7687 tcg_gen_ld_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
,dr
[reg
]));
7688 gen_op_mov_reg_v(ot
, rm
, cpu_T
[0]);
7692 case 0x106: /* clts */
7694 gen_exception(s
, EXCP0D_GPF
, pc_start
- s
->cs_base
);
7696 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_WRITE_CR0
);
7697 gen_helper_clts(cpu_env
);
7698 /* abort block because static cpu state changed */
7699 gen_jmp_im(s
->pc
- s
->cs_base
);
7703 /* MMX/3DNow!/SSE/SSE2/SSE3/SSSE3/SSE4 support */
7704 case 0x1c3: /* MOVNTI reg, mem */
7705 if (!(s
->cpuid_features
& CPUID_SSE2
))
7707 ot
= mo_64_32(dflag
);
7708 modrm
= cpu_ldub_code(env
, s
->pc
++);
7709 mod
= (modrm
>> 6) & 3;
7712 reg
= ((modrm
>> 3) & 7) | rex_r
;
7713 /* generate a generic store */
7714 gen_ldst_modrm(env
, s
, modrm
, ot
, reg
, 1);
7717 modrm
= cpu_ldub_code(env
, s
->pc
++);
7718 mod
= (modrm
>> 6) & 3;
7719 op
= (modrm
>> 3) & 7;
7721 case 0: /* fxsave */
7722 if (mod
== 3 || !(s
->cpuid_features
& CPUID_FXSR
) ||
7723 (s
->prefix
& PREFIX_LOCK
))
7725 if ((s
->flags
& HF_EM_MASK
) || (s
->flags
& HF_TS_MASK
)) {
7726 gen_exception(s
, EXCP07_PREX
, pc_start
- s
->cs_base
);
7729 gen_lea_modrm(env
, s
, modrm
);
7730 gen_update_cc_op(s
);
7731 gen_jmp_im(pc_start
- s
->cs_base
);
7732 gen_helper_fxsave(cpu_env
, cpu_A0
, tcg_const_i32(dflag
== MO_64
));
7734 case 1: /* fxrstor */
7735 if (mod
== 3 || !(s
->cpuid_features
& CPUID_FXSR
) ||
7736 (s
->prefix
& PREFIX_LOCK
))
7738 if ((s
->flags
& HF_EM_MASK
) || (s
->flags
& HF_TS_MASK
)) {
7739 gen_exception(s
, EXCP07_PREX
, pc_start
- s
->cs_base
);
7742 gen_lea_modrm(env
, s
, modrm
);
7743 gen_update_cc_op(s
);
7744 gen_jmp_im(pc_start
- s
->cs_base
);
7745 gen_helper_fxrstor(cpu_env
, cpu_A0
, tcg_const_i32(dflag
== MO_64
));
7747 case 2: /* ldmxcsr */
7748 case 3: /* stmxcsr */
7749 if (s
->flags
& HF_TS_MASK
) {
7750 gen_exception(s
, EXCP07_PREX
, pc_start
- s
->cs_base
);
7753 if ((s
->flags
& HF_EM_MASK
) || !(s
->flags
& HF_OSFXSR_MASK
) ||
7756 gen_lea_modrm(env
, s
, modrm
);
7758 tcg_gen_qemu_ld_i32(cpu_tmp2_i32
, cpu_A0
,
7759 s
->mem_index
, MO_LEUL
);
7760 gen_helper_ldmxcsr(cpu_env
, cpu_tmp2_i32
);
7762 tcg_gen_ld32u_tl(cpu_T
[0], cpu_env
, offsetof(CPUX86State
, mxcsr
));
7763 gen_op_st_v(s
, MO_32
, cpu_T
[0], cpu_A0
);
7766 case 5: /* lfence */
7767 case 6: /* mfence */
7768 if ((modrm
& 0xc7) != 0xc0 || !(s
->cpuid_features
& CPUID_SSE2
))
7771 case 7: /* sfence / clflush */
7772 if ((modrm
& 0xc7) == 0xc0) {
7774 /* XXX: also check for cpuid_ext2_features & CPUID_EXT2_EMMX */
7775 if (!(s
->cpuid_features
& CPUID_SSE
))
7779 if (!(s
->cpuid_features
& CPUID_CLFLUSH
))
7781 gen_lea_modrm(env
, s
, modrm
);
7788 case 0x10d: /* 3DNow! prefetch(w) */
7789 modrm
= cpu_ldub_code(env
, s
->pc
++);
7790 mod
= (modrm
>> 6) & 3;
7793 gen_lea_modrm(env
, s
, modrm
);
7794 /* ignore for now */
7796 case 0x1aa: /* rsm */
7797 gen_svm_check_intercept(s
, pc_start
, SVM_EXIT_RSM
);
7798 if (!(s
->flags
& HF_SMM_MASK
))
7800 gen_update_cc_op(s
);
7801 gen_jmp_im(s
->pc
- s
->cs_base
);
7802 gen_helper_rsm(cpu_env
);
7805 case 0x1b8: /* SSE4.2 popcnt */
7806 if ((prefixes
& (PREFIX_REPZ
| PREFIX_LOCK
| PREFIX_REPNZ
)) !=
7809 if (!(s
->cpuid_ext_features
& CPUID_EXT_POPCNT
))
7812 modrm
= cpu_ldub_code(env
, s
->pc
++);
7813 reg
= ((modrm
>> 3) & 7) | rex_r
;
7815 if (s
->prefix
& PREFIX_DATA
) {
7818 ot
= mo_64_32(dflag
);
7821 gen_ldst_modrm(env
, s
, modrm
, ot
, OR_TMP0
, 0);
7822 gen_helper_popcnt(cpu_T
[0], cpu_env
, cpu_T
[0], tcg_const_i32(ot
));
7823 gen_op_mov_reg_v(ot
, reg
, cpu_T
[0]);
7825 set_cc_op(s
, CC_OP_EFLAGS
);
7827 case 0x10e ... 0x10f:
7828 /* 3DNow! instructions, ignore prefixes */
7829 s
->prefix
&= ~(PREFIX_REPZ
| PREFIX_REPNZ
| PREFIX_DATA
);
7830 case 0x110 ... 0x117:
7831 case 0x128 ... 0x12f:
7832 case 0x138 ... 0x13a:
7833 case 0x150 ... 0x179:
7834 case 0x17c ... 0x17f:
7836 case 0x1c4 ... 0x1c6:
7837 case 0x1d0 ... 0x1fe:
7838 gen_sse(env
, s
, b
, pc_start
, rex_r
);
7843 /* lock generation */
7844 if (s
->prefix
& PREFIX_LOCK
)
7845 gen_helper_unlock();
7848 if (s
->prefix
& PREFIX_LOCK
)
7849 gen_helper_unlock();
7850 /* XXX: ensure that no lock was generated */
7851 gen_exception(s
, EXCP06_ILLOP
, pc_start
- s
->cs_base
);
7855 void optimize_flags_init(void)
7857 static const char reg_names
[CPU_NB_REGS
][4] = {
7858 #ifdef TARGET_X86_64
7888 cpu_env
= tcg_global_reg_new_ptr(TCG_AREG0
, "env");
7889 cpu_cc_op
= tcg_global_mem_new_i32(TCG_AREG0
,
7890 offsetof(CPUX86State
, cc_op
), "cc_op");
7891 cpu_cc_dst
= tcg_global_mem_new(TCG_AREG0
, offsetof(CPUX86State
, cc_dst
),
7893 cpu_cc_src
= tcg_global_mem_new(TCG_AREG0
, offsetof(CPUX86State
, cc_src
),
7895 cpu_cc_src2
= tcg_global_mem_new(TCG_AREG0
, offsetof(CPUX86State
, cc_src2
),
7898 for (i
= 0; i
< CPU_NB_REGS
; ++i
) {
7899 cpu_regs
[i
] = tcg_global_mem_new(TCG_AREG0
,
7900 offsetof(CPUX86State
, regs
[i
]),
7905 /* generate intermediate code in gen_opc_buf and gen_opparam_buf for
7906 basic block 'tb'. If search_pc is TRUE, also generate PC
7907 information for each intermediate instruction. */
7908 static inline void gen_intermediate_code_internal(X86CPU
*cpu
,
7909 TranslationBlock
*tb
,
7912 CPUState
*cs
= CPU(cpu
);
7913 CPUX86State
*env
= &cpu
->env
;
7914 DisasContext dc1
, *dc
= &dc1
;
7915 target_ulong pc_ptr
;
7916 uint16_t *gen_opc_end
;
7920 target_ulong pc_start
;
7921 target_ulong cs_base
;
7925 /* generate intermediate code */
7927 cs_base
= tb
->cs_base
;
7930 dc
->pe
= (flags
>> HF_PE_SHIFT
) & 1;
7931 dc
->code32
= (flags
>> HF_CS32_SHIFT
) & 1;
7932 dc
->ss32
= (flags
>> HF_SS32_SHIFT
) & 1;
7933 dc
->addseg
= (flags
>> HF_ADDSEG_SHIFT
) & 1;
7935 dc
->vm86
= (flags
>> VM_SHIFT
) & 1;
7936 dc
->cpl
= (flags
>> HF_CPL_SHIFT
) & 3;
7937 dc
->iopl
= (flags
>> IOPL_SHIFT
) & 3;
7938 dc
->tf
= (flags
>> TF_SHIFT
) & 1;
7939 dc
->singlestep_enabled
= cs
->singlestep_enabled
;
7940 dc
->cc_op
= CC_OP_DYNAMIC
;
7941 dc
->cc_op_dirty
= false;
7942 dc
->cs_base
= cs_base
;
7944 dc
->popl_esp_hack
= 0;
7945 /* select memory access functions */
7947 if (flags
& HF_SOFTMMU_MASK
) {
7948 dc
->mem_index
= cpu_mmu_index(env
);
7950 dc
->cpuid_features
= env
->features
[FEAT_1_EDX
];
7951 dc
->cpuid_ext_features
= env
->features
[FEAT_1_ECX
];
7952 dc
->cpuid_ext2_features
= env
->features
[FEAT_8000_0001_EDX
];
7953 dc
->cpuid_ext3_features
= env
->features
[FEAT_8000_0001_ECX
];
7954 dc
->cpuid_7_0_ebx_features
= env
->features
[FEAT_7_0_EBX
];
7955 #ifdef TARGET_X86_64
7956 dc
->lma
= (flags
>> HF_LMA_SHIFT
) & 1;
7957 dc
->code64
= (flags
>> HF_CS64_SHIFT
) & 1;
7960 dc
->jmp_opt
= !(dc
->tf
|| cs
->singlestep_enabled
||
7961 (flags
& HF_INHIBIT_IRQ_MASK
)
7962 #ifndef CONFIG_SOFTMMU
7963 || (flags
& HF_SOFTMMU_MASK
)
7966 /* Do not optimize repz jumps at all in icount mode, because
7967 rep movsS instructions are execured with different paths
7968 in !repz_opt and repz_opt modes. The first one was used
7969 always except single step mode. And this setting
7970 disables jumps optimization and control paths become
7971 equivalent in run and single step modes.
7972 Now there will be no jump optimization for repz in
7973 record/replay modes and there will always be an
7974 additional step for ecx=0 when icount is enabled.
7976 dc
->repz_opt
= !dc
->jmp_opt
&& !(tb
->cflags
& CF_USE_ICOUNT
);
7978 /* check addseg logic */
7979 if (!dc
->addseg
&& (dc
->vm86
|| !dc
->pe
|| !dc
->code32
))
7980 printf("ERROR addseg\n");
7983 cpu_T
[0] = tcg_temp_new();
7984 cpu_T
[1] = tcg_temp_new();
7985 cpu_A0
= tcg_temp_new();
7987 cpu_tmp0
= tcg_temp_new();
7988 cpu_tmp1_i64
= tcg_temp_new_i64();
7989 cpu_tmp2_i32
= tcg_temp_new_i32();
7990 cpu_tmp3_i32
= tcg_temp_new_i32();
7991 cpu_tmp4
= tcg_temp_new();
7992 cpu_ptr0
= tcg_temp_new_ptr();
7993 cpu_ptr1
= tcg_temp_new_ptr();
7994 cpu_cc_srcT
= tcg_temp_local_new();
7996 gen_opc_end
= tcg_ctx
.gen_opc_buf
+ OPC_MAX_SIZE
;
7998 dc
->is_jmp
= DISAS_NEXT
;
8002 max_insns
= tb
->cflags
& CF_COUNT_MASK
;
8004 max_insns
= CF_COUNT_MASK
;
8008 if (unlikely(!QTAILQ_EMPTY(&cs
->breakpoints
))) {
8009 QTAILQ_FOREACH(bp
, &cs
->breakpoints
, entry
) {
8010 if (bp
->pc
== pc_ptr
&&
8011 !((bp
->flags
& BP_CPU
) && (tb
->flags
& HF_RF_MASK
))) {
8012 gen_debug(dc
, pc_ptr
- dc
->cs_base
);
8013 goto done_generating
;
8018 j
= tcg_ctx
.gen_opc_ptr
- tcg_ctx
.gen_opc_buf
;
8022 tcg_ctx
.gen_opc_instr_start
[lj
++] = 0;
8024 tcg_ctx
.gen_opc_pc
[lj
] = pc_ptr
;
8025 gen_opc_cc_op
[lj
] = dc
->cc_op
;
8026 tcg_ctx
.gen_opc_instr_start
[lj
] = 1;
8027 tcg_ctx
.gen_opc_icount
[lj
] = num_insns
;
8029 if (num_insns
+ 1 == max_insns
&& (tb
->cflags
& CF_LAST_IO
))
8032 pc_ptr
= disas_insn(env
, dc
, pc_ptr
);
8034 /* stop translation if indicated */
8037 /* if single step mode, we generate only one instruction and
8038 generate an exception */
8039 /* if irq were inhibited with HF_INHIBIT_IRQ_MASK, we clear
8040 the flag and abort the translation to give the irqs a
8041 change to be happen */
8042 if (dc
->tf
|| dc
->singlestep_enabled
||
8043 (flags
& HF_INHIBIT_IRQ_MASK
)) {
8044 gen_jmp_im(pc_ptr
- dc
->cs_base
);
8048 /* Do not cross the boundary of the pages in icount mode,
8049 it can cause an exception. Do it only when boundary is
8050 crossed by the first instruction in the block.
8051 If current instruction already crossed the bound - it's ok,
8052 because an exception hasn't stopped this code.
8054 if ((tb
->cflags
& CF_USE_ICOUNT
)
8055 && ((pc_ptr
& TARGET_PAGE_MASK
)
8056 != ((pc_ptr
+ TARGET_MAX_INSN_SIZE
- 1) & TARGET_PAGE_MASK
)
8057 || (pc_ptr
& ~TARGET_PAGE_MASK
) == 0)) {
8058 gen_jmp_im(pc_ptr
- dc
->cs_base
);
8062 /* if too long translation, stop generation too */
8063 if (tcg_ctx
.gen_opc_ptr
>= gen_opc_end
||
8064 (pc_ptr
- pc_start
) >= (TARGET_PAGE_SIZE
- 32) ||
8065 num_insns
>= max_insns
) {
8066 gen_jmp_im(pc_ptr
- dc
->cs_base
);
8071 gen_jmp_im(pc_ptr
- dc
->cs_base
);
8076 if (tb
->cflags
& CF_LAST_IO
)
8079 gen_tb_end(tb
, num_insns
);
8080 *tcg_ctx
.gen_opc_ptr
= INDEX_op_end
;
8081 /* we don't forget to fill the last values */
8083 j
= tcg_ctx
.gen_opc_ptr
- tcg_ctx
.gen_opc_buf
;
8086 tcg_ctx
.gen_opc_instr_start
[lj
++] = 0;
8090 if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM
)) {
8092 qemu_log("----------------\n");
8093 qemu_log("IN: %s\n", lookup_symbol(pc_start
));
8094 #ifdef TARGET_X86_64
8099 disas_flags
= !dc
->code32
;
8100 log_target_disas(env
, pc_start
, pc_ptr
- pc_start
, disas_flags
);
8106 tb
->size
= pc_ptr
- pc_start
;
8107 tb
->icount
= num_insns
;
8111 void gen_intermediate_code(CPUX86State
*env
, TranslationBlock
*tb
)
8113 gen_intermediate_code_internal(x86_env_get_cpu(env
), tb
, false);
8116 void gen_intermediate_code_pc(CPUX86State
*env
, TranslationBlock
*tb
)
8118 gen_intermediate_code_internal(x86_env_get_cpu(env
), tb
, true);
8121 void restore_state_to_opc(CPUX86State
*env
, TranslationBlock
*tb
, int pc_pos
)
8125 if (qemu_loglevel_mask(CPU_LOG_TB_OP
)) {
8127 qemu_log("RESTORE:\n");
8128 for(i
= 0;i
<= pc_pos
; i
++) {
8129 if (tcg_ctx
.gen_opc_instr_start
[i
]) {
8130 qemu_log("0x%04x: " TARGET_FMT_lx
"\n", i
,
8131 tcg_ctx
.gen_opc_pc
[i
]);
8134 qemu_log("pc_pos=0x%x eip=" TARGET_FMT_lx
" cs_base=%x\n",
8135 pc_pos
, tcg_ctx
.gen_opc_pc
[pc_pos
] - tb
->cs_base
,
8136 (uint32_t)tb
->cs_base
);
8139 env
->eip
= tcg_ctx
.gen_opc_pc
[pc_pos
] - tb
->cs_base
;
8140 cc_op
= gen_opc_cc_op
[pc_pos
];
8141 if (cc_op
!= CC_OP_DYNAMIC
)