target/arm: update the cortex-a15 MIDR to latest rev
[qemu/ar7.git] / tcg / tcg.c
blob612a12f58f5531174b55859bae21e6f4d8a02cd4
1 /*
2 * Tiny Code Generator for QEMU
4 * Copyright (c) 2008 Fabrice Bellard
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to deal
8 * in the Software without restriction, including without limitation the rights
9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 * copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22 * THE SOFTWARE.
25 /* define it to use liveness analysis (better code) */
26 #define USE_TCG_OPTIMIZATIONS
28 #include "qemu/osdep.h"
30 /* Define to jump the ELF file used to communicate with GDB. */
31 #undef DEBUG_JIT
33 #include "qemu/error-report.h"
34 #include "qemu/cutils.h"
35 #include "qemu/host-utils.h"
36 #include "qemu/qemu-print.h"
37 #include "qemu/timer.h"
38 #include "qemu/cacheflush.h"
39 #include "qemu/cacheinfo.h"
41 /* Note: the long term plan is to reduce the dependencies on the QEMU
42 CPU definitions. Currently they are used for qemu_ld/st
43 instructions */
44 #define NO_CPU_IO_DEFS
46 #include "exec/exec-all.h"
47 #include "tcg/tcg-op.h"
49 #if UINTPTR_MAX == UINT32_MAX
50 # define ELF_CLASS ELFCLASS32
51 #else
52 # define ELF_CLASS ELFCLASS64
53 #endif
54 #if HOST_BIG_ENDIAN
55 # define ELF_DATA ELFDATA2MSB
56 #else
57 # define ELF_DATA ELFDATA2LSB
58 #endif
60 #include "elf.h"
61 #include "exec/log.h"
62 #include "tcg/tcg-ldst.h"
63 #include "tcg-internal.h"
65 #ifdef CONFIG_TCG_INTERPRETER
66 #include <ffi.h>
67 #endif
69 /* Forward declarations for functions declared in tcg-target.c.inc and
70 used here. */
71 static void tcg_target_init(TCGContext *s);
72 static void tcg_target_qemu_prologue(TCGContext *s);
73 static bool patch_reloc(tcg_insn_unit *code_ptr, int type,
74 intptr_t value, intptr_t addend);
76 /* The CIE and FDE header definitions will be common to all hosts. */
77 typedef struct {
78 uint32_t len __attribute__((aligned((sizeof(void *)))));
79 uint32_t id;
80 uint8_t version;
81 char augmentation[1];
82 uint8_t code_align;
83 uint8_t data_align;
84 uint8_t return_column;
85 } DebugFrameCIE;
87 typedef struct QEMU_PACKED {
88 uint32_t len __attribute__((aligned((sizeof(void *)))));
89 uint32_t cie_offset;
90 uintptr_t func_start;
91 uintptr_t func_len;
92 } DebugFrameFDEHeader;
94 typedef struct QEMU_PACKED {
95 DebugFrameCIE cie;
96 DebugFrameFDEHeader fde;
97 } DebugFrameHeader;
99 static void tcg_register_jit_int(const void *buf, size_t size,
100 const void *debug_frame,
101 size_t debug_frame_size)
102 __attribute__((unused));
104 /* Forward declarations for functions declared and used in tcg-target.c.inc. */
105 static void tcg_out_ld(TCGContext *s, TCGType type, TCGReg ret, TCGReg arg1,
106 intptr_t arg2);
107 static bool tcg_out_mov(TCGContext *s, TCGType type, TCGReg ret, TCGReg arg);
108 static void tcg_out_movi(TCGContext *s, TCGType type,
109 TCGReg ret, tcg_target_long arg);
110 static void tcg_out_op(TCGContext *s, TCGOpcode opc,
111 const TCGArg args[TCG_MAX_OP_ARGS],
112 const int const_args[TCG_MAX_OP_ARGS]);
113 #if TCG_TARGET_MAYBE_vec
114 static bool tcg_out_dup_vec(TCGContext *s, TCGType type, unsigned vece,
115 TCGReg dst, TCGReg src);
116 static bool tcg_out_dupm_vec(TCGContext *s, TCGType type, unsigned vece,
117 TCGReg dst, TCGReg base, intptr_t offset);
118 static void tcg_out_dupi_vec(TCGContext *s, TCGType type, unsigned vece,
119 TCGReg dst, int64_t arg);
120 static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
121 unsigned vecl, unsigned vece,
122 const TCGArg args[TCG_MAX_OP_ARGS],
123 const int const_args[TCG_MAX_OP_ARGS]);
124 #else
125 static inline bool tcg_out_dup_vec(TCGContext *s, TCGType type, unsigned vece,
126 TCGReg dst, TCGReg src)
128 g_assert_not_reached();
130 static inline bool tcg_out_dupm_vec(TCGContext *s, TCGType type, unsigned vece,
131 TCGReg dst, TCGReg base, intptr_t offset)
133 g_assert_not_reached();
135 static inline void tcg_out_dupi_vec(TCGContext *s, TCGType type, unsigned vece,
136 TCGReg dst, int64_t arg)
138 g_assert_not_reached();
140 static inline void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
141 unsigned vecl, unsigned vece,
142 const TCGArg args[TCG_MAX_OP_ARGS],
143 const int const_args[TCG_MAX_OP_ARGS])
145 g_assert_not_reached();
147 #endif
148 static void tcg_out_st(TCGContext *s, TCGType type, TCGReg arg, TCGReg arg1,
149 intptr_t arg2);
150 static bool tcg_out_sti(TCGContext *s, TCGType type, TCGArg val,
151 TCGReg base, intptr_t ofs);
152 #ifdef CONFIG_TCG_INTERPRETER
153 static void tcg_out_call(TCGContext *s, const tcg_insn_unit *target,
154 ffi_cif *cif);
155 #else
156 static void tcg_out_call(TCGContext *s, const tcg_insn_unit *target);
157 #endif
158 static bool tcg_target_const_match(int64_t val, TCGType type, int ct);
159 #ifdef TCG_TARGET_NEED_LDST_LABELS
160 static int tcg_out_ldst_finalize(TCGContext *s);
161 #endif
163 TCGContext tcg_init_ctx;
164 __thread TCGContext *tcg_ctx;
166 TCGContext **tcg_ctxs;
167 unsigned int tcg_cur_ctxs;
168 unsigned int tcg_max_ctxs;
169 TCGv_env cpu_env = 0;
170 const void *tcg_code_gen_epilogue;
171 uintptr_t tcg_splitwx_diff;
173 #ifndef CONFIG_TCG_INTERPRETER
174 tcg_prologue_fn *tcg_qemu_tb_exec;
175 #endif
177 static TCGRegSet tcg_target_available_regs[TCG_TYPE_COUNT];
178 static TCGRegSet tcg_target_call_clobber_regs;
180 #if TCG_TARGET_INSN_UNIT_SIZE == 1
181 static __attribute__((unused)) inline void tcg_out8(TCGContext *s, uint8_t v)
183 *s->code_ptr++ = v;
186 static __attribute__((unused)) inline void tcg_patch8(tcg_insn_unit *p,
187 uint8_t v)
189 *p = v;
191 #endif
193 #if TCG_TARGET_INSN_UNIT_SIZE <= 2
194 static __attribute__((unused)) inline void tcg_out16(TCGContext *s, uint16_t v)
196 if (TCG_TARGET_INSN_UNIT_SIZE == 2) {
197 *s->code_ptr++ = v;
198 } else {
199 tcg_insn_unit *p = s->code_ptr;
200 memcpy(p, &v, sizeof(v));
201 s->code_ptr = p + (2 / TCG_TARGET_INSN_UNIT_SIZE);
205 static __attribute__((unused)) inline void tcg_patch16(tcg_insn_unit *p,
206 uint16_t v)
208 if (TCG_TARGET_INSN_UNIT_SIZE == 2) {
209 *p = v;
210 } else {
211 memcpy(p, &v, sizeof(v));
214 #endif
216 #if TCG_TARGET_INSN_UNIT_SIZE <= 4
217 static __attribute__((unused)) inline void tcg_out32(TCGContext *s, uint32_t v)
219 if (TCG_TARGET_INSN_UNIT_SIZE == 4) {
220 *s->code_ptr++ = v;
221 } else {
222 tcg_insn_unit *p = s->code_ptr;
223 memcpy(p, &v, sizeof(v));
224 s->code_ptr = p + (4 / TCG_TARGET_INSN_UNIT_SIZE);
228 static __attribute__((unused)) inline void tcg_patch32(tcg_insn_unit *p,
229 uint32_t v)
231 if (TCG_TARGET_INSN_UNIT_SIZE == 4) {
232 *p = v;
233 } else {
234 memcpy(p, &v, sizeof(v));
237 #endif
239 #if TCG_TARGET_INSN_UNIT_SIZE <= 8
240 static __attribute__((unused)) inline void tcg_out64(TCGContext *s, uint64_t v)
242 if (TCG_TARGET_INSN_UNIT_SIZE == 8) {
243 *s->code_ptr++ = v;
244 } else {
245 tcg_insn_unit *p = s->code_ptr;
246 memcpy(p, &v, sizeof(v));
247 s->code_ptr = p + (8 / TCG_TARGET_INSN_UNIT_SIZE);
251 static __attribute__((unused)) inline void tcg_patch64(tcg_insn_unit *p,
252 uint64_t v)
254 if (TCG_TARGET_INSN_UNIT_SIZE == 8) {
255 *p = v;
256 } else {
257 memcpy(p, &v, sizeof(v));
260 #endif
262 /* label relocation processing */
264 static void tcg_out_reloc(TCGContext *s, tcg_insn_unit *code_ptr, int type,
265 TCGLabel *l, intptr_t addend)
267 TCGRelocation *r = tcg_malloc(sizeof(TCGRelocation));
269 r->type = type;
270 r->ptr = code_ptr;
271 r->addend = addend;
272 QSIMPLEQ_INSERT_TAIL(&l->relocs, r, next);
275 static void tcg_out_label(TCGContext *s, TCGLabel *l)
277 tcg_debug_assert(!l->has_value);
278 l->has_value = 1;
279 l->u.value_ptr = tcg_splitwx_to_rx(s->code_ptr);
282 TCGLabel *gen_new_label(void)
284 TCGContext *s = tcg_ctx;
285 TCGLabel *l = tcg_malloc(sizeof(TCGLabel));
287 memset(l, 0, sizeof(TCGLabel));
288 l->id = s->nb_labels++;
289 QSIMPLEQ_INIT(&l->relocs);
291 QSIMPLEQ_INSERT_TAIL(&s->labels, l, next);
293 return l;
296 static bool tcg_resolve_relocs(TCGContext *s)
298 TCGLabel *l;
300 QSIMPLEQ_FOREACH(l, &s->labels, next) {
301 TCGRelocation *r;
302 uintptr_t value = l->u.value;
304 QSIMPLEQ_FOREACH(r, &l->relocs, next) {
305 if (!patch_reloc(r->ptr, r->type, value, r->addend)) {
306 return false;
310 return true;
313 static void set_jmp_reset_offset(TCGContext *s, int which)
316 * We will check for overflow at the end of the opcode loop in
317 * tcg_gen_code, where we bound tcg_current_code_size to UINT16_MAX.
319 s->tb_jmp_reset_offset[which] = tcg_current_code_size(s);
322 /* Signal overflow, starting over with fewer guest insns. */
323 static G_NORETURN
324 void tcg_raise_tb_overflow(TCGContext *s)
326 siglongjmp(s->jmp_trans, -2);
329 #define C_PFX1(P, A) P##A
330 #define C_PFX2(P, A, B) P##A##_##B
331 #define C_PFX3(P, A, B, C) P##A##_##B##_##C
332 #define C_PFX4(P, A, B, C, D) P##A##_##B##_##C##_##D
333 #define C_PFX5(P, A, B, C, D, E) P##A##_##B##_##C##_##D##_##E
334 #define C_PFX6(P, A, B, C, D, E, F) P##A##_##B##_##C##_##D##_##E##_##F
336 /* Define an enumeration for the various combinations. */
338 #define C_O0_I1(I1) C_PFX1(c_o0_i1_, I1),
339 #define C_O0_I2(I1, I2) C_PFX2(c_o0_i2_, I1, I2),
340 #define C_O0_I3(I1, I2, I3) C_PFX3(c_o0_i3_, I1, I2, I3),
341 #define C_O0_I4(I1, I2, I3, I4) C_PFX4(c_o0_i4_, I1, I2, I3, I4),
343 #define C_O1_I1(O1, I1) C_PFX2(c_o1_i1_, O1, I1),
344 #define C_O1_I2(O1, I1, I2) C_PFX3(c_o1_i2_, O1, I1, I2),
345 #define C_O1_I3(O1, I1, I2, I3) C_PFX4(c_o1_i3_, O1, I1, I2, I3),
346 #define C_O1_I4(O1, I1, I2, I3, I4) C_PFX5(c_o1_i4_, O1, I1, I2, I3, I4),
348 #define C_N1_I2(O1, I1, I2) C_PFX3(c_n1_i2_, O1, I1, I2),
350 #define C_O2_I1(O1, O2, I1) C_PFX3(c_o2_i1_, O1, O2, I1),
351 #define C_O2_I2(O1, O2, I1, I2) C_PFX4(c_o2_i2_, O1, O2, I1, I2),
352 #define C_O2_I3(O1, O2, I1, I2, I3) C_PFX5(c_o2_i3_, O1, O2, I1, I2, I3),
353 #define C_O2_I4(O1, O2, I1, I2, I3, I4) C_PFX6(c_o2_i4_, O1, O2, I1, I2, I3, I4),
355 typedef enum {
356 #include "tcg-target-con-set.h"
357 } TCGConstraintSetIndex;
359 static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode);
361 #undef C_O0_I1
362 #undef C_O0_I2
363 #undef C_O0_I3
364 #undef C_O0_I4
365 #undef C_O1_I1
366 #undef C_O1_I2
367 #undef C_O1_I3
368 #undef C_O1_I4
369 #undef C_N1_I2
370 #undef C_O2_I1
371 #undef C_O2_I2
372 #undef C_O2_I3
373 #undef C_O2_I4
375 /* Put all of the constraint sets into an array, indexed by the enum. */
377 #define C_O0_I1(I1) { .args_ct_str = { #I1 } },
378 #define C_O0_I2(I1, I2) { .args_ct_str = { #I1, #I2 } },
379 #define C_O0_I3(I1, I2, I3) { .args_ct_str = { #I1, #I2, #I3 } },
380 #define C_O0_I4(I1, I2, I3, I4) { .args_ct_str = { #I1, #I2, #I3, #I4 } },
382 #define C_O1_I1(O1, I1) { .args_ct_str = { #O1, #I1 } },
383 #define C_O1_I2(O1, I1, I2) { .args_ct_str = { #O1, #I1, #I2 } },
384 #define C_O1_I3(O1, I1, I2, I3) { .args_ct_str = { #O1, #I1, #I2, #I3 } },
385 #define C_O1_I4(O1, I1, I2, I3, I4) { .args_ct_str = { #O1, #I1, #I2, #I3, #I4 } },
387 #define C_N1_I2(O1, I1, I2) { .args_ct_str = { "&" #O1, #I1, #I2 } },
389 #define C_O2_I1(O1, O2, I1) { .args_ct_str = { #O1, #O2, #I1 } },
390 #define C_O2_I2(O1, O2, I1, I2) { .args_ct_str = { #O1, #O2, #I1, #I2 } },
391 #define C_O2_I3(O1, O2, I1, I2, I3) { .args_ct_str = { #O1, #O2, #I1, #I2, #I3 } },
392 #define C_O2_I4(O1, O2, I1, I2, I3, I4) { .args_ct_str = { #O1, #O2, #I1, #I2, #I3, #I4 } },
394 static const TCGTargetOpDef constraint_sets[] = {
395 #include "tcg-target-con-set.h"
399 #undef C_O0_I1
400 #undef C_O0_I2
401 #undef C_O0_I3
402 #undef C_O0_I4
403 #undef C_O1_I1
404 #undef C_O1_I2
405 #undef C_O1_I3
406 #undef C_O1_I4
407 #undef C_N1_I2
408 #undef C_O2_I1
409 #undef C_O2_I2
410 #undef C_O2_I3
411 #undef C_O2_I4
413 /* Expand the enumerator to be returned from tcg_target_op_def(). */
415 #define C_O0_I1(I1) C_PFX1(c_o0_i1_, I1)
416 #define C_O0_I2(I1, I2) C_PFX2(c_o0_i2_, I1, I2)
417 #define C_O0_I3(I1, I2, I3) C_PFX3(c_o0_i3_, I1, I2, I3)
418 #define C_O0_I4(I1, I2, I3, I4) C_PFX4(c_o0_i4_, I1, I2, I3, I4)
420 #define C_O1_I1(O1, I1) C_PFX2(c_o1_i1_, O1, I1)
421 #define C_O1_I2(O1, I1, I2) C_PFX3(c_o1_i2_, O1, I1, I2)
422 #define C_O1_I3(O1, I1, I2, I3) C_PFX4(c_o1_i3_, O1, I1, I2, I3)
423 #define C_O1_I4(O1, I1, I2, I3, I4) C_PFX5(c_o1_i4_, O1, I1, I2, I3, I4)
425 #define C_N1_I2(O1, I1, I2) C_PFX3(c_n1_i2_, O1, I1, I2)
427 #define C_O2_I1(O1, O2, I1) C_PFX3(c_o2_i1_, O1, O2, I1)
428 #define C_O2_I2(O1, O2, I1, I2) C_PFX4(c_o2_i2_, O1, O2, I1, I2)
429 #define C_O2_I3(O1, O2, I1, I2, I3) C_PFX5(c_o2_i3_, O1, O2, I1, I2, I3)
430 #define C_O2_I4(O1, O2, I1, I2, I3, I4) C_PFX6(c_o2_i4_, O1, O2, I1, I2, I3, I4)
432 #include "tcg-target.c.inc"
434 static void alloc_tcg_plugin_context(TCGContext *s)
436 #ifdef CONFIG_PLUGIN
437 s->plugin_tb = g_new0(struct qemu_plugin_tb, 1);
438 s->plugin_tb->insns =
439 g_ptr_array_new_with_free_func(qemu_plugin_insn_cleanup_fn);
440 #endif
444 * All TCG threads except the parent (i.e. the one that called tcg_context_init
445 * and registered the target's TCG globals) must register with this function
446 * before initiating translation.
448 * In user-mode we just point tcg_ctx to tcg_init_ctx. See the documentation
449 * of tcg_region_init() for the reasoning behind this.
451 * In softmmu each caller registers its context in tcg_ctxs[]. Note that in
452 * softmmu tcg_ctxs[] does not track tcg_ctx_init, since the initial context
453 * is not used anymore for translation once this function is called.
455 * Not tracking tcg_init_ctx in tcg_ctxs[] in softmmu keeps code that iterates
456 * over the array (e.g. tcg_code_size() the same for both softmmu and user-mode.
458 #ifdef CONFIG_USER_ONLY
459 void tcg_register_thread(void)
461 tcg_ctx = &tcg_init_ctx;
463 #else
464 void tcg_register_thread(void)
466 TCGContext *s = g_malloc(sizeof(*s));
467 unsigned int i, n;
469 *s = tcg_init_ctx;
471 /* Relink mem_base. */
472 for (i = 0, n = tcg_init_ctx.nb_globals; i < n; ++i) {
473 if (tcg_init_ctx.temps[i].mem_base) {
474 ptrdiff_t b = tcg_init_ctx.temps[i].mem_base - tcg_init_ctx.temps;
475 tcg_debug_assert(b >= 0 && b < n);
476 s->temps[i].mem_base = &s->temps[b];
480 /* Claim an entry in tcg_ctxs */
481 n = qatomic_fetch_inc(&tcg_cur_ctxs);
482 g_assert(n < tcg_max_ctxs);
483 qatomic_set(&tcg_ctxs[n], s);
485 if (n > 0) {
486 alloc_tcg_plugin_context(s);
487 tcg_region_initial_alloc(s);
490 tcg_ctx = s;
492 #endif /* !CONFIG_USER_ONLY */
494 /* pool based memory allocation */
495 void *tcg_malloc_internal(TCGContext *s, int size)
497 TCGPool *p;
498 int pool_size;
500 if (size > TCG_POOL_CHUNK_SIZE) {
501 /* big malloc: insert a new pool (XXX: could optimize) */
502 p = g_malloc(sizeof(TCGPool) + size);
503 p->size = size;
504 p->next = s->pool_first_large;
505 s->pool_first_large = p;
506 return p->data;
507 } else {
508 p = s->pool_current;
509 if (!p) {
510 p = s->pool_first;
511 if (!p)
512 goto new_pool;
513 } else {
514 if (!p->next) {
515 new_pool:
516 pool_size = TCG_POOL_CHUNK_SIZE;
517 p = g_malloc(sizeof(TCGPool) + pool_size);
518 p->size = pool_size;
519 p->next = NULL;
520 if (s->pool_current)
521 s->pool_current->next = p;
522 else
523 s->pool_first = p;
524 } else {
525 p = p->next;
529 s->pool_current = p;
530 s->pool_cur = p->data + size;
531 s->pool_end = p->data + p->size;
532 return p->data;
535 void tcg_pool_reset(TCGContext *s)
537 TCGPool *p, *t;
538 for (p = s->pool_first_large; p; p = t) {
539 t = p->next;
540 g_free(p);
542 s->pool_first_large = NULL;
543 s->pool_cur = s->pool_end = NULL;
544 s->pool_current = NULL;
547 #include "exec/helper-proto.h"
549 static const TCGHelperInfo all_helpers[] = {
550 #include "exec/helper-tcg.h"
552 static GHashTable *helper_table;
554 #ifdef CONFIG_TCG_INTERPRETER
555 static GHashTable *ffi_table;
557 static ffi_type * const typecode_to_ffi[8] = {
558 [dh_typecode_void] = &ffi_type_void,
559 [dh_typecode_i32] = &ffi_type_uint32,
560 [dh_typecode_s32] = &ffi_type_sint32,
561 [dh_typecode_i64] = &ffi_type_uint64,
562 [dh_typecode_s64] = &ffi_type_sint64,
563 [dh_typecode_ptr] = &ffi_type_pointer,
565 #endif
567 static int indirect_reg_alloc_order[ARRAY_SIZE(tcg_target_reg_alloc_order)];
568 static void process_op_defs(TCGContext *s);
569 static TCGTemp *tcg_global_reg_new_internal(TCGContext *s, TCGType type,
570 TCGReg reg, const char *name);
572 static void tcg_context_init(unsigned max_cpus)
574 TCGContext *s = &tcg_init_ctx;
575 int op, total_args, n, i;
576 TCGOpDef *def;
577 TCGArgConstraint *args_ct;
578 TCGTemp *ts;
580 memset(s, 0, sizeof(*s));
581 s->nb_globals = 0;
583 /* Count total number of arguments and allocate the corresponding
584 space */
585 total_args = 0;
586 for(op = 0; op < NB_OPS; op++) {
587 def = &tcg_op_defs[op];
588 n = def->nb_iargs + def->nb_oargs;
589 total_args += n;
592 args_ct = g_new0(TCGArgConstraint, total_args);
594 for(op = 0; op < NB_OPS; op++) {
595 def = &tcg_op_defs[op];
596 def->args_ct = args_ct;
597 n = def->nb_iargs + def->nb_oargs;
598 args_ct += n;
601 /* Register helpers. */
602 /* Use g_direct_hash/equal for direct pointer comparisons on func. */
603 helper_table = g_hash_table_new(NULL, NULL);
605 for (i = 0; i < ARRAY_SIZE(all_helpers); ++i) {
606 g_hash_table_insert(helper_table, (gpointer)all_helpers[i].func,
607 (gpointer)&all_helpers[i]);
610 #ifdef CONFIG_TCG_INTERPRETER
611 /* g_direct_hash/equal for direct comparisons on uint32_t. */
612 ffi_table = g_hash_table_new(NULL, NULL);
613 for (i = 0; i < ARRAY_SIZE(all_helpers); ++i) {
614 struct {
615 ffi_cif cif;
616 ffi_type *args[];
617 } *ca;
618 uint32_t typemask = all_helpers[i].typemask;
619 gpointer hash = (gpointer)(uintptr_t)typemask;
620 ffi_status status;
621 int nargs;
623 if (g_hash_table_lookup(ffi_table, hash)) {
624 continue;
627 /* Ignoring the return type, find the last non-zero field. */
628 nargs = 32 - clz32(typemask >> 3);
629 nargs = DIV_ROUND_UP(nargs, 3);
631 ca = g_malloc0(sizeof(*ca) + nargs * sizeof(ffi_type *));
632 ca->cif.rtype = typecode_to_ffi[typemask & 7];
633 ca->cif.nargs = nargs;
635 if (nargs != 0) {
636 ca->cif.arg_types = ca->args;
637 for (i = 0; i < nargs; ++i) {
638 int typecode = extract32(typemask, (i + 1) * 3, 3);
639 ca->args[i] = typecode_to_ffi[typecode];
643 status = ffi_prep_cif(&ca->cif, FFI_DEFAULT_ABI, nargs,
644 ca->cif.rtype, ca->cif.arg_types);
645 assert(status == FFI_OK);
647 g_hash_table_insert(ffi_table, hash, (gpointer)&ca->cif);
649 #endif
651 tcg_target_init(s);
652 process_op_defs(s);
654 /* Reverse the order of the saved registers, assuming they're all at
655 the start of tcg_target_reg_alloc_order. */
656 for (n = 0; n < ARRAY_SIZE(tcg_target_reg_alloc_order); ++n) {
657 int r = tcg_target_reg_alloc_order[n];
658 if (tcg_regset_test_reg(tcg_target_call_clobber_regs, r)) {
659 break;
662 for (i = 0; i < n; ++i) {
663 indirect_reg_alloc_order[i] = tcg_target_reg_alloc_order[n - 1 - i];
665 for (; i < ARRAY_SIZE(tcg_target_reg_alloc_order); ++i) {
666 indirect_reg_alloc_order[i] = tcg_target_reg_alloc_order[i];
669 alloc_tcg_plugin_context(s);
671 tcg_ctx = s;
673 * In user-mode we simply share the init context among threads, since we
674 * use a single region. See the documentation tcg_region_init() for the
675 * reasoning behind this.
676 * In softmmu we will have at most max_cpus TCG threads.
678 #ifdef CONFIG_USER_ONLY
679 tcg_ctxs = &tcg_ctx;
680 tcg_cur_ctxs = 1;
681 tcg_max_ctxs = 1;
682 #else
683 tcg_max_ctxs = max_cpus;
684 tcg_ctxs = g_new0(TCGContext *, max_cpus);
685 #endif
687 tcg_debug_assert(!tcg_regset_test_reg(s->reserved_regs, TCG_AREG0));
688 ts = tcg_global_reg_new_internal(s, TCG_TYPE_PTR, TCG_AREG0, "env");
689 cpu_env = temp_tcgv_ptr(ts);
692 void tcg_init(size_t tb_size, int splitwx, unsigned max_cpus)
694 tcg_context_init(max_cpus);
695 tcg_region_init(tb_size, splitwx, max_cpus);
699 * Allocate TBs right before their corresponding translated code, making
700 * sure that TBs and code are on different cache lines.
702 TranslationBlock *tcg_tb_alloc(TCGContext *s)
704 uintptr_t align = qemu_icache_linesize;
705 TranslationBlock *tb;
706 void *next;
708 retry:
709 tb = (void *)ROUND_UP((uintptr_t)s->code_gen_ptr, align);
710 next = (void *)ROUND_UP((uintptr_t)(tb + 1), align);
712 if (unlikely(next > s->code_gen_highwater)) {
713 if (tcg_region_alloc(s)) {
714 return NULL;
716 goto retry;
718 qatomic_set(&s->code_gen_ptr, next);
719 s->data_gen_ptr = NULL;
720 return tb;
723 void tcg_prologue_init(TCGContext *s)
725 size_t prologue_size;
727 s->code_ptr = s->code_gen_ptr;
728 s->code_buf = s->code_gen_ptr;
729 s->data_gen_ptr = NULL;
731 #ifndef CONFIG_TCG_INTERPRETER
732 tcg_qemu_tb_exec = (tcg_prologue_fn *)tcg_splitwx_to_rx(s->code_ptr);
733 #endif
735 #ifdef TCG_TARGET_NEED_POOL_LABELS
736 s->pool_labels = NULL;
737 #endif
739 qemu_thread_jit_write();
740 /* Generate the prologue. */
741 tcg_target_qemu_prologue(s);
743 #ifdef TCG_TARGET_NEED_POOL_LABELS
744 /* Allow the prologue to put e.g. guest_base into a pool entry. */
746 int result = tcg_out_pool_finalize(s);
747 tcg_debug_assert(result == 0);
749 #endif
751 prologue_size = tcg_current_code_size(s);
753 #ifndef CONFIG_TCG_INTERPRETER
754 flush_idcache_range((uintptr_t)tcg_splitwx_to_rx(s->code_buf),
755 (uintptr_t)s->code_buf, prologue_size);
756 #endif
758 #ifdef DEBUG_DISAS
759 if (qemu_loglevel_mask(CPU_LOG_TB_OUT_ASM)) {
760 FILE *logfile = qemu_log_trylock();
761 if (logfile) {
762 fprintf(logfile, "PROLOGUE: [size=%zu]\n", prologue_size);
763 if (s->data_gen_ptr) {
764 size_t code_size = s->data_gen_ptr - s->code_gen_ptr;
765 size_t data_size = prologue_size - code_size;
766 size_t i;
768 disas(logfile, s->code_gen_ptr, code_size);
770 for (i = 0; i < data_size; i += sizeof(tcg_target_ulong)) {
771 if (sizeof(tcg_target_ulong) == 8) {
772 fprintf(logfile,
773 "0x%08" PRIxPTR ": .quad 0x%016" PRIx64 "\n",
774 (uintptr_t)s->data_gen_ptr + i,
775 *(uint64_t *)(s->data_gen_ptr + i));
776 } else {
777 fprintf(logfile,
778 "0x%08" PRIxPTR ": .long 0x%08x\n",
779 (uintptr_t)s->data_gen_ptr + i,
780 *(uint32_t *)(s->data_gen_ptr + i));
783 } else {
784 disas(logfile, s->code_gen_ptr, prologue_size);
786 fprintf(logfile, "\n");
787 qemu_log_unlock(logfile);
790 #endif
792 #ifndef CONFIG_TCG_INTERPRETER
794 * Assert that goto_ptr is implemented completely, setting an epilogue.
795 * For tci, we use NULL as the signal to return from the interpreter,
796 * so skip this check.
798 tcg_debug_assert(tcg_code_gen_epilogue != NULL);
799 #endif
801 tcg_region_prologue_set(s);
804 void tcg_func_start(TCGContext *s)
806 tcg_pool_reset(s);
807 s->nb_temps = s->nb_globals;
809 /* No temps have been previously allocated for size or locality. */
810 memset(s->free_temps, 0, sizeof(s->free_temps));
812 /* No constant temps have been previously allocated. */
813 for (int i = 0; i < TCG_TYPE_COUNT; ++i) {
814 if (s->const_table[i]) {
815 g_hash_table_remove_all(s->const_table[i]);
819 s->nb_ops = 0;
820 s->nb_labels = 0;
821 s->current_frame_offset = s->frame_start;
823 #ifdef CONFIG_DEBUG_TCG
824 s->goto_tb_issue_mask = 0;
825 #endif
827 QTAILQ_INIT(&s->ops);
828 QTAILQ_INIT(&s->free_ops);
829 QSIMPLEQ_INIT(&s->labels);
832 static TCGTemp *tcg_temp_alloc(TCGContext *s)
834 int n = s->nb_temps++;
836 if (n >= TCG_MAX_TEMPS) {
837 tcg_raise_tb_overflow(s);
839 return memset(&s->temps[n], 0, sizeof(TCGTemp));
842 static TCGTemp *tcg_global_alloc(TCGContext *s)
844 TCGTemp *ts;
846 tcg_debug_assert(s->nb_globals == s->nb_temps);
847 tcg_debug_assert(s->nb_globals < TCG_MAX_TEMPS);
848 s->nb_globals++;
849 ts = tcg_temp_alloc(s);
850 ts->kind = TEMP_GLOBAL;
852 return ts;
855 static TCGTemp *tcg_global_reg_new_internal(TCGContext *s, TCGType type,
856 TCGReg reg, const char *name)
858 TCGTemp *ts;
860 if (TCG_TARGET_REG_BITS == 32 && type != TCG_TYPE_I32) {
861 tcg_abort();
864 ts = tcg_global_alloc(s);
865 ts->base_type = type;
866 ts->type = type;
867 ts->kind = TEMP_FIXED;
868 ts->reg = reg;
869 ts->name = name;
870 tcg_regset_set_reg(s->reserved_regs, reg);
872 return ts;
875 void tcg_set_frame(TCGContext *s, TCGReg reg, intptr_t start, intptr_t size)
877 s->frame_start = start;
878 s->frame_end = start + size;
879 s->frame_temp
880 = tcg_global_reg_new_internal(s, TCG_TYPE_PTR, reg, "_frame");
883 TCGTemp *tcg_global_mem_new_internal(TCGType type, TCGv_ptr base,
884 intptr_t offset, const char *name)
886 TCGContext *s = tcg_ctx;
887 TCGTemp *base_ts = tcgv_ptr_temp(base);
888 TCGTemp *ts = tcg_global_alloc(s);
889 int indirect_reg = 0, bigendian = 0;
890 #if HOST_BIG_ENDIAN
891 bigendian = 1;
892 #endif
894 switch (base_ts->kind) {
895 case TEMP_FIXED:
896 break;
897 case TEMP_GLOBAL:
898 /* We do not support double-indirect registers. */
899 tcg_debug_assert(!base_ts->indirect_reg);
900 base_ts->indirect_base = 1;
901 s->nb_indirects += (TCG_TARGET_REG_BITS == 32 && type == TCG_TYPE_I64
902 ? 2 : 1);
903 indirect_reg = 1;
904 break;
905 default:
906 g_assert_not_reached();
909 if (TCG_TARGET_REG_BITS == 32 && type == TCG_TYPE_I64) {
910 TCGTemp *ts2 = tcg_global_alloc(s);
911 char buf[64];
913 ts->base_type = TCG_TYPE_I64;
914 ts->type = TCG_TYPE_I32;
915 ts->indirect_reg = indirect_reg;
916 ts->mem_allocated = 1;
917 ts->mem_base = base_ts;
918 ts->mem_offset = offset + bigendian * 4;
919 pstrcpy(buf, sizeof(buf), name);
920 pstrcat(buf, sizeof(buf), "_0");
921 ts->name = strdup(buf);
923 tcg_debug_assert(ts2 == ts + 1);
924 ts2->base_type = TCG_TYPE_I64;
925 ts2->type = TCG_TYPE_I32;
926 ts2->indirect_reg = indirect_reg;
927 ts2->mem_allocated = 1;
928 ts2->mem_base = base_ts;
929 ts2->mem_offset = offset + (1 - bigendian) * 4;
930 pstrcpy(buf, sizeof(buf), name);
931 pstrcat(buf, sizeof(buf), "_1");
932 ts2->name = strdup(buf);
933 } else {
934 ts->base_type = type;
935 ts->type = type;
936 ts->indirect_reg = indirect_reg;
937 ts->mem_allocated = 1;
938 ts->mem_base = base_ts;
939 ts->mem_offset = offset;
940 ts->name = name;
942 return ts;
945 TCGTemp *tcg_temp_new_internal(TCGType type, bool temp_local)
947 TCGContext *s = tcg_ctx;
948 TCGTempKind kind = temp_local ? TEMP_LOCAL : TEMP_NORMAL;
949 TCGTemp *ts;
950 int idx, k;
952 k = type + (temp_local ? TCG_TYPE_COUNT : 0);
953 idx = find_first_bit(s->free_temps[k].l, TCG_MAX_TEMPS);
954 if (idx < TCG_MAX_TEMPS) {
955 /* There is already an available temp with the right type. */
956 clear_bit(idx, s->free_temps[k].l);
958 ts = &s->temps[idx];
959 ts->temp_allocated = 1;
960 tcg_debug_assert(ts->base_type == type);
961 tcg_debug_assert(ts->kind == kind);
962 } else {
963 ts = tcg_temp_alloc(s);
964 if (TCG_TARGET_REG_BITS == 32 && type == TCG_TYPE_I64) {
965 TCGTemp *ts2 = tcg_temp_alloc(s);
967 ts->base_type = type;
968 ts->type = TCG_TYPE_I32;
969 ts->temp_allocated = 1;
970 ts->kind = kind;
972 tcg_debug_assert(ts2 == ts + 1);
973 ts2->base_type = TCG_TYPE_I64;
974 ts2->type = TCG_TYPE_I32;
975 ts2->temp_allocated = 1;
976 ts2->kind = kind;
977 } else {
978 ts->base_type = type;
979 ts->type = type;
980 ts->temp_allocated = 1;
981 ts->kind = kind;
985 #if defined(CONFIG_DEBUG_TCG)
986 s->temps_in_use++;
987 #endif
988 return ts;
991 TCGv_vec tcg_temp_new_vec(TCGType type)
993 TCGTemp *t;
995 #ifdef CONFIG_DEBUG_TCG
996 switch (type) {
997 case TCG_TYPE_V64:
998 assert(TCG_TARGET_HAS_v64);
999 break;
1000 case TCG_TYPE_V128:
1001 assert(TCG_TARGET_HAS_v128);
1002 break;
1003 case TCG_TYPE_V256:
1004 assert(TCG_TARGET_HAS_v256);
1005 break;
1006 default:
1007 g_assert_not_reached();
1009 #endif
1011 t = tcg_temp_new_internal(type, 0);
1012 return temp_tcgv_vec(t);
1015 /* Create a new temp of the same type as an existing temp. */
1016 TCGv_vec tcg_temp_new_vec_matching(TCGv_vec match)
1018 TCGTemp *t = tcgv_vec_temp(match);
1020 tcg_debug_assert(t->temp_allocated != 0);
1022 t = tcg_temp_new_internal(t->base_type, 0);
1023 return temp_tcgv_vec(t);
1026 void tcg_temp_free_internal(TCGTemp *ts)
1028 TCGContext *s = tcg_ctx;
1029 int k, idx;
1031 switch (ts->kind) {
1032 case TEMP_CONST:
1034 * In order to simplify users of tcg_constant_*,
1035 * silently ignore free.
1037 return;
1038 case TEMP_NORMAL:
1039 case TEMP_LOCAL:
1040 break;
1041 default:
1042 g_assert_not_reached();
1045 #if defined(CONFIG_DEBUG_TCG)
1046 s->temps_in_use--;
1047 if (s->temps_in_use < 0) {
1048 fprintf(stderr, "More temporaries freed than allocated!\n");
1050 #endif
1052 tcg_debug_assert(ts->temp_allocated != 0);
1053 ts->temp_allocated = 0;
1055 idx = temp_idx(ts);
1056 k = ts->base_type + (ts->kind == TEMP_NORMAL ? 0 : TCG_TYPE_COUNT);
1057 set_bit(idx, s->free_temps[k].l);
1060 TCGTemp *tcg_constant_internal(TCGType type, int64_t val)
1062 TCGContext *s = tcg_ctx;
1063 GHashTable *h = s->const_table[type];
1064 TCGTemp *ts;
1066 if (h == NULL) {
1067 h = g_hash_table_new(g_int64_hash, g_int64_equal);
1068 s->const_table[type] = h;
1071 ts = g_hash_table_lookup(h, &val);
1072 if (ts == NULL) {
1073 ts = tcg_temp_alloc(s);
1075 if (TCG_TARGET_REG_BITS == 32 && type == TCG_TYPE_I64) {
1076 TCGTemp *ts2 = tcg_temp_alloc(s);
1078 ts->base_type = TCG_TYPE_I64;
1079 ts->type = TCG_TYPE_I32;
1080 ts->kind = TEMP_CONST;
1081 ts->temp_allocated = 1;
1083 * Retain the full value of the 64-bit constant in the low
1084 * part, so that the hash table works. Actual uses will
1085 * truncate the value to the low part.
1087 ts->val = val;
1089 tcg_debug_assert(ts2 == ts + 1);
1090 ts2->base_type = TCG_TYPE_I64;
1091 ts2->type = TCG_TYPE_I32;
1092 ts2->kind = TEMP_CONST;
1093 ts2->temp_allocated = 1;
1094 ts2->val = val >> 32;
1095 } else {
1096 ts->base_type = type;
1097 ts->type = type;
1098 ts->kind = TEMP_CONST;
1099 ts->temp_allocated = 1;
1100 ts->val = val;
1102 g_hash_table_insert(h, &ts->val, ts);
1105 return ts;
1108 TCGv_vec tcg_constant_vec(TCGType type, unsigned vece, int64_t val)
1110 val = dup_const(vece, val);
1111 return temp_tcgv_vec(tcg_constant_internal(type, val));
1114 TCGv_vec tcg_constant_vec_matching(TCGv_vec match, unsigned vece, int64_t val)
1116 TCGTemp *t = tcgv_vec_temp(match);
1118 tcg_debug_assert(t->temp_allocated != 0);
1119 return tcg_constant_vec(t->base_type, vece, val);
1122 TCGv_i32 tcg_const_i32(int32_t val)
1124 TCGv_i32 t0;
1125 t0 = tcg_temp_new_i32();
1126 tcg_gen_movi_i32(t0, val);
1127 return t0;
1130 TCGv_i64 tcg_const_i64(int64_t val)
1132 TCGv_i64 t0;
1133 t0 = tcg_temp_new_i64();
1134 tcg_gen_movi_i64(t0, val);
1135 return t0;
1138 TCGv_i32 tcg_const_local_i32(int32_t val)
1140 TCGv_i32 t0;
1141 t0 = tcg_temp_local_new_i32();
1142 tcg_gen_movi_i32(t0, val);
1143 return t0;
1146 TCGv_i64 tcg_const_local_i64(int64_t val)
1148 TCGv_i64 t0;
1149 t0 = tcg_temp_local_new_i64();
1150 tcg_gen_movi_i64(t0, val);
1151 return t0;
1154 #if defined(CONFIG_DEBUG_TCG)
1155 void tcg_clear_temp_count(void)
1157 TCGContext *s = tcg_ctx;
1158 s->temps_in_use = 0;
1161 int tcg_check_temp_count(void)
1163 TCGContext *s = tcg_ctx;
1164 if (s->temps_in_use) {
1165 /* Clear the count so that we don't give another
1166 * warning immediately next time around.
1168 s->temps_in_use = 0;
1169 return 1;
1171 return 0;
1173 #endif
1175 /* Return true if OP may appear in the opcode stream.
1176 Test the runtime variable that controls each opcode. */
1177 bool tcg_op_supported(TCGOpcode op)
1179 const bool have_vec
1180 = TCG_TARGET_HAS_v64 | TCG_TARGET_HAS_v128 | TCG_TARGET_HAS_v256;
1182 switch (op) {
1183 case INDEX_op_discard:
1184 case INDEX_op_set_label:
1185 case INDEX_op_call:
1186 case INDEX_op_br:
1187 case INDEX_op_mb:
1188 case INDEX_op_insn_start:
1189 case INDEX_op_exit_tb:
1190 case INDEX_op_goto_tb:
1191 case INDEX_op_goto_ptr:
1192 case INDEX_op_qemu_ld_i32:
1193 case INDEX_op_qemu_st_i32:
1194 case INDEX_op_qemu_ld_i64:
1195 case INDEX_op_qemu_st_i64:
1196 return true;
1198 case INDEX_op_qemu_st8_i32:
1199 return TCG_TARGET_HAS_qemu_st8_i32;
1201 case INDEX_op_mov_i32:
1202 case INDEX_op_setcond_i32:
1203 case INDEX_op_brcond_i32:
1204 case INDEX_op_ld8u_i32:
1205 case INDEX_op_ld8s_i32:
1206 case INDEX_op_ld16u_i32:
1207 case INDEX_op_ld16s_i32:
1208 case INDEX_op_ld_i32:
1209 case INDEX_op_st8_i32:
1210 case INDEX_op_st16_i32:
1211 case INDEX_op_st_i32:
1212 case INDEX_op_add_i32:
1213 case INDEX_op_sub_i32:
1214 case INDEX_op_mul_i32:
1215 case INDEX_op_and_i32:
1216 case INDEX_op_or_i32:
1217 case INDEX_op_xor_i32:
1218 case INDEX_op_shl_i32:
1219 case INDEX_op_shr_i32:
1220 case INDEX_op_sar_i32:
1221 return true;
1223 case INDEX_op_movcond_i32:
1224 return TCG_TARGET_HAS_movcond_i32;
1225 case INDEX_op_div_i32:
1226 case INDEX_op_divu_i32:
1227 return TCG_TARGET_HAS_div_i32;
1228 case INDEX_op_rem_i32:
1229 case INDEX_op_remu_i32:
1230 return TCG_TARGET_HAS_rem_i32;
1231 case INDEX_op_div2_i32:
1232 case INDEX_op_divu2_i32:
1233 return TCG_TARGET_HAS_div2_i32;
1234 case INDEX_op_rotl_i32:
1235 case INDEX_op_rotr_i32:
1236 return TCG_TARGET_HAS_rot_i32;
1237 case INDEX_op_deposit_i32:
1238 return TCG_TARGET_HAS_deposit_i32;
1239 case INDEX_op_extract_i32:
1240 return TCG_TARGET_HAS_extract_i32;
1241 case INDEX_op_sextract_i32:
1242 return TCG_TARGET_HAS_sextract_i32;
1243 case INDEX_op_extract2_i32:
1244 return TCG_TARGET_HAS_extract2_i32;
1245 case INDEX_op_add2_i32:
1246 return TCG_TARGET_HAS_add2_i32;
1247 case INDEX_op_sub2_i32:
1248 return TCG_TARGET_HAS_sub2_i32;
1249 case INDEX_op_mulu2_i32:
1250 return TCG_TARGET_HAS_mulu2_i32;
1251 case INDEX_op_muls2_i32:
1252 return TCG_TARGET_HAS_muls2_i32;
1253 case INDEX_op_muluh_i32:
1254 return TCG_TARGET_HAS_muluh_i32;
1255 case INDEX_op_mulsh_i32:
1256 return TCG_TARGET_HAS_mulsh_i32;
1257 case INDEX_op_ext8s_i32:
1258 return TCG_TARGET_HAS_ext8s_i32;
1259 case INDEX_op_ext16s_i32:
1260 return TCG_TARGET_HAS_ext16s_i32;
1261 case INDEX_op_ext8u_i32:
1262 return TCG_TARGET_HAS_ext8u_i32;
1263 case INDEX_op_ext16u_i32:
1264 return TCG_TARGET_HAS_ext16u_i32;
1265 case INDEX_op_bswap16_i32:
1266 return TCG_TARGET_HAS_bswap16_i32;
1267 case INDEX_op_bswap32_i32:
1268 return TCG_TARGET_HAS_bswap32_i32;
1269 case INDEX_op_not_i32:
1270 return TCG_TARGET_HAS_not_i32;
1271 case INDEX_op_neg_i32:
1272 return TCG_TARGET_HAS_neg_i32;
1273 case INDEX_op_andc_i32:
1274 return TCG_TARGET_HAS_andc_i32;
1275 case INDEX_op_orc_i32:
1276 return TCG_TARGET_HAS_orc_i32;
1277 case INDEX_op_eqv_i32:
1278 return TCG_TARGET_HAS_eqv_i32;
1279 case INDEX_op_nand_i32:
1280 return TCG_TARGET_HAS_nand_i32;
1281 case INDEX_op_nor_i32:
1282 return TCG_TARGET_HAS_nor_i32;
1283 case INDEX_op_clz_i32:
1284 return TCG_TARGET_HAS_clz_i32;
1285 case INDEX_op_ctz_i32:
1286 return TCG_TARGET_HAS_ctz_i32;
1287 case INDEX_op_ctpop_i32:
1288 return TCG_TARGET_HAS_ctpop_i32;
1290 case INDEX_op_brcond2_i32:
1291 case INDEX_op_setcond2_i32:
1292 return TCG_TARGET_REG_BITS == 32;
1294 case INDEX_op_mov_i64:
1295 case INDEX_op_setcond_i64:
1296 case INDEX_op_brcond_i64:
1297 case INDEX_op_ld8u_i64:
1298 case INDEX_op_ld8s_i64:
1299 case INDEX_op_ld16u_i64:
1300 case INDEX_op_ld16s_i64:
1301 case INDEX_op_ld32u_i64:
1302 case INDEX_op_ld32s_i64:
1303 case INDEX_op_ld_i64:
1304 case INDEX_op_st8_i64:
1305 case INDEX_op_st16_i64:
1306 case INDEX_op_st32_i64:
1307 case INDEX_op_st_i64:
1308 case INDEX_op_add_i64:
1309 case INDEX_op_sub_i64:
1310 case INDEX_op_mul_i64:
1311 case INDEX_op_and_i64:
1312 case INDEX_op_or_i64:
1313 case INDEX_op_xor_i64:
1314 case INDEX_op_shl_i64:
1315 case INDEX_op_shr_i64:
1316 case INDEX_op_sar_i64:
1317 case INDEX_op_ext_i32_i64:
1318 case INDEX_op_extu_i32_i64:
1319 return TCG_TARGET_REG_BITS == 64;
1321 case INDEX_op_movcond_i64:
1322 return TCG_TARGET_HAS_movcond_i64;
1323 case INDEX_op_div_i64:
1324 case INDEX_op_divu_i64:
1325 return TCG_TARGET_HAS_div_i64;
1326 case INDEX_op_rem_i64:
1327 case INDEX_op_remu_i64:
1328 return TCG_TARGET_HAS_rem_i64;
1329 case INDEX_op_div2_i64:
1330 case INDEX_op_divu2_i64:
1331 return TCG_TARGET_HAS_div2_i64;
1332 case INDEX_op_rotl_i64:
1333 case INDEX_op_rotr_i64:
1334 return TCG_TARGET_HAS_rot_i64;
1335 case INDEX_op_deposit_i64:
1336 return TCG_TARGET_HAS_deposit_i64;
1337 case INDEX_op_extract_i64:
1338 return TCG_TARGET_HAS_extract_i64;
1339 case INDEX_op_sextract_i64:
1340 return TCG_TARGET_HAS_sextract_i64;
1341 case INDEX_op_extract2_i64:
1342 return TCG_TARGET_HAS_extract2_i64;
1343 case INDEX_op_extrl_i64_i32:
1344 return TCG_TARGET_HAS_extrl_i64_i32;
1345 case INDEX_op_extrh_i64_i32:
1346 return TCG_TARGET_HAS_extrh_i64_i32;
1347 case INDEX_op_ext8s_i64:
1348 return TCG_TARGET_HAS_ext8s_i64;
1349 case INDEX_op_ext16s_i64:
1350 return TCG_TARGET_HAS_ext16s_i64;
1351 case INDEX_op_ext32s_i64:
1352 return TCG_TARGET_HAS_ext32s_i64;
1353 case INDEX_op_ext8u_i64:
1354 return TCG_TARGET_HAS_ext8u_i64;
1355 case INDEX_op_ext16u_i64:
1356 return TCG_TARGET_HAS_ext16u_i64;
1357 case INDEX_op_ext32u_i64:
1358 return TCG_TARGET_HAS_ext32u_i64;
1359 case INDEX_op_bswap16_i64:
1360 return TCG_TARGET_HAS_bswap16_i64;
1361 case INDEX_op_bswap32_i64:
1362 return TCG_TARGET_HAS_bswap32_i64;
1363 case INDEX_op_bswap64_i64:
1364 return TCG_TARGET_HAS_bswap64_i64;
1365 case INDEX_op_not_i64:
1366 return TCG_TARGET_HAS_not_i64;
1367 case INDEX_op_neg_i64:
1368 return TCG_TARGET_HAS_neg_i64;
1369 case INDEX_op_andc_i64:
1370 return TCG_TARGET_HAS_andc_i64;
1371 case INDEX_op_orc_i64:
1372 return TCG_TARGET_HAS_orc_i64;
1373 case INDEX_op_eqv_i64:
1374 return TCG_TARGET_HAS_eqv_i64;
1375 case INDEX_op_nand_i64:
1376 return TCG_TARGET_HAS_nand_i64;
1377 case INDEX_op_nor_i64:
1378 return TCG_TARGET_HAS_nor_i64;
1379 case INDEX_op_clz_i64:
1380 return TCG_TARGET_HAS_clz_i64;
1381 case INDEX_op_ctz_i64:
1382 return TCG_TARGET_HAS_ctz_i64;
1383 case INDEX_op_ctpop_i64:
1384 return TCG_TARGET_HAS_ctpop_i64;
1385 case INDEX_op_add2_i64:
1386 return TCG_TARGET_HAS_add2_i64;
1387 case INDEX_op_sub2_i64:
1388 return TCG_TARGET_HAS_sub2_i64;
1389 case INDEX_op_mulu2_i64:
1390 return TCG_TARGET_HAS_mulu2_i64;
1391 case INDEX_op_muls2_i64:
1392 return TCG_TARGET_HAS_muls2_i64;
1393 case INDEX_op_muluh_i64:
1394 return TCG_TARGET_HAS_muluh_i64;
1395 case INDEX_op_mulsh_i64:
1396 return TCG_TARGET_HAS_mulsh_i64;
1398 case INDEX_op_mov_vec:
1399 case INDEX_op_dup_vec:
1400 case INDEX_op_dupm_vec:
1401 case INDEX_op_ld_vec:
1402 case INDEX_op_st_vec:
1403 case INDEX_op_add_vec:
1404 case INDEX_op_sub_vec:
1405 case INDEX_op_and_vec:
1406 case INDEX_op_or_vec:
1407 case INDEX_op_xor_vec:
1408 case INDEX_op_cmp_vec:
1409 return have_vec;
1410 case INDEX_op_dup2_vec:
1411 return have_vec && TCG_TARGET_REG_BITS == 32;
1412 case INDEX_op_not_vec:
1413 return have_vec && TCG_TARGET_HAS_not_vec;
1414 case INDEX_op_neg_vec:
1415 return have_vec && TCG_TARGET_HAS_neg_vec;
1416 case INDEX_op_abs_vec:
1417 return have_vec && TCG_TARGET_HAS_abs_vec;
1418 case INDEX_op_andc_vec:
1419 return have_vec && TCG_TARGET_HAS_andc_vec;
1420 case INDEX_op_orc_vec:
1421 return have_vec && TCG_TARGET_HAS_orc_vec;
1422 case INDEX_op_nand_vec:
1423 return have_vec && TCG_TARGET_HAS_nand_vec;
1424 case INDEX_op_nor_vec:
1425 return have_vec && TCG_TARGET_HAS_nor_vec;
1426 case INDEX_op_eqv_vec:
1427 return have_vec && TCG_TARGET_HAS_eqv_vec;
1428 case INDEX_op_mul_vec:
1429 return have_vec && TCG_TARGET_HAS_mul_vec;
1430 case INDEX_op_shli_vec:
1431 case INDEX_op_shri_vec:
1432 case INDEX_op_sari_vec:
1433 return have_vec && TCG_TARGET_HAS_shi_vec;
1434 case INDEX_op_shls_vec:
1435 case INDEX_op_shrs_vec:
1436 case INDEX_op_sars_vec:
1437 return have_vec && TCG_TARGET_HAS_shs_vec;
1438 case INDEX_op_shlv_vec:
1439 case INDEX_op_shrv_vec:
1440 case INDEX_op_sarv_vec:
1441 return have_vec && TCG_TARGET_HAS_shv_vec;
1442 case INDEX_op_rotli_vec:
1443 return have_vec && TCG_TARGET_HAS_roti_vec;
1444 case INDEX_op_rotls_vec:
1445 return have_vec && TCG_TARGET_HAS_rots_vec;
1446 case INDEX_op_rotlv_vec:
1447 case INDEX_op_rotrv_vec:
1448 return have_vec && TCG_TARGET_HAS_rotv_vec;
1449 case INDEX_op_ssadd_vec:
1450 case INDEX_op_usadd_vec:
1451 case INDEX_op_sssub_vec:
1452 case INDEX_op_ussub_vec:
1453 return have_vec && TCG_TARGET_HAS_sat_vec;
1454 case INDEX_op_smin_vec:
1455 case INDEX_op_umin_vec:
1456 case INDEX_op_smax_vec:
1457 case INDEX_op_umax_vec:
1458 return have_vec && TCG_TARGET_HAS_minmax_vec;
1459 case INDEX_op_bitsel_vec:
1460 return have_vec && TCG_TARGET_HAS_bitsel_vec;
1461 case INDEX_op_cmpsel_vec:
1462 return have_vec && TCG_TARGET_HAS_cmpsel_vec;
1464 default:
1465 tcg_debug_assert(op > INDEX_op_last_generic && op < NB_OPS);
1466 return true;
1470 /* Note: we convert the 64 bit args to 32 bit and do some alignment
1471 and endian swap. Maybe it would be better to do the alignment
1472 and endian swap in tcg_reg_alloc_call(). */
1473 void tcg_gen_callN(void *func, TCGTemp *ret, int nargs, TCGTemp **args)
1475 int i, real_args, nb_rets, pi;
1476 unsigned typemask;
1477 const TCGHelperInfo *info;
1478 TCGOp *op;
1480 info = g_hash_table_lookup(helper_table, (gpointer)func);
1481 typemask = info->typemask;
1483 #ifdef CONFIG_PLUGIN
1484 /* detect non-plugin helpers */
1485 if (tcg_ctx->plugin_insn && unlikely(strncmp(info->name, "plugin_", 7))) {
1486 tcg_ctx->plugin_insn->calls_helpers = true;
1488 #endif
1490 #if defined(__sparc__) && !defined(__arch64__) \
1491 && !defined(CONFIG_TCG_INTERPRETER)
1492 /* We have 64-bit values in one register, but need to pass as two
1493 separate parameters. Split them. */
1494 int orig_typemask = typemask;
1495 int orig_nargs = nargs;
1496 TCGv_i64 retl, reth;
1497 TCGTemp *split_args[MAX_OPC_PARAM];
1499 retl = NULL;
1500 reth = NULL;
1501 typemask = 0;
1502 for (i = real_args = 0; i < nargs; ++i) {
1503 int argtype = extract32(orig_typemask, (i + 1) * 3, 3);
1504 bool is_64bit = (argtype & ~1) == dh_typecode_i64;
1506 if (is_64bit) {
1507 TCGv_i64 orig = temp_tcgv_i64(args[i]);
1508 TCGv_i32 h = tcg_temp_new_i32();
1509 TCGv_i32 l = tcg_temp_new_i32();
1510 tcg_gen_extr_i64_i32(l, h, orig);
1511 split_args[real_args++] = tcgv_i32_temp(h);
1512 typemask |= dh_typecode_i32 << (real_args * 3);
1513 split_args[real_args++] = tcgv_i32_temp(l);
1514 typemask |= dh_typecode_i32 << (real_args * 3);
1515 } else {
1516 split_args[real_args++] = args[i];
1517 typemask |= argtype << (real_args * 3);
1520 nargs = real_args;
1521 args = split_args;
1522 #elif defined(TCG_TARGET_EXTEND_ARGS) && TCG_TARGET_REG_BITS == 64
1523 for (i = 0; i < nargs; ++i) {
1524 int argtype = extract32(typemask, (i + 1) * 3, 3);
1525 bool is_32bit = (argtype & ~1) == dh_typecode_i32;
1526 bool is_signed = argtype & 1;
1528 if (is_32bit) {
1529 TCGv_i64 temp = tcg_temp_new_i64();
1530 TCGv_i32 orig = temp_tcgv_i32(args[i]);
1531 if (is_signed) {
1532 tcg_gen_ext_i32_i64(temp, orig);
1533 } else {
1534 tcg_gen_extu_i32_i64(temp, orig);
1536 args[i] = tcgv_i64_temp(temp);
1539 #endif /* TCG_TARGET_EXTEND_ARGS */
1541 op = tcg_emit_op(INDEX_op_call);
1543 pi = 0;
1544 if (ret != NULL) {
1545 #if defined(__sparc__) && !defined(__arch64__) \
1546 && !defined(CONFIG_TCG_INTERPRETER)
1547 if ((typemask & 6) == dh_typecode_i64) {
1548 /* The 32-bit ABI is going to return the 64-bit value in
1549 the %o0/%o1 register pair. Prepare for this by using
1550 two return temporaries, and reassemble below. */
1551 retl = tcg_temp_new_i64();
1552 reth = tcg_temp_new_i64();
1553 op->args[pi++] = tcgv_i64_arg(reth);
1554 op->args[pi++] = tcgv_i64_arg(retl);
1555 nb_rets = 2;
1556 } else {
1557 op->args[pi++] = temp_arg(ret);
1558 nb_rets = 1;
1560 #else
1561 if (TCG_TARGET_REG_BITS < 64 && (typemask & 6) == dh_typecode_i64) {
1562 #if HOST_BIG_ENDIAN
1563 op->args[pi++] = temp_arg(ret + 1);
1564 op->args[pi++] = temp_arg(ret);
1565 #else
1566 op->args[pi++] = temp_arg(ret);
1567 op->args[pi++] = temp_arg(ret + 1);
1568 #endif
1569 nb_rets = 2;
1570 } else {
1571 op->args[pi++] = temp_arg(ret);
1572 nb_rets = 1;
1574 #endif
1575 } else {
1576 nb_rets = 0;
1578 TCGOP_CALLO(op) = nb_rets;
1580 real_args = 0;
1581 for (i = 0; i < nargs; i++) {
1582 int argtype = extract32(typemask, (i + 1) * 3, 3);
1583 bool is_64bit = (argtype & ~1) == dh_typecode_i64;
1584 bool want_align = false;
1586 #if defined(CONFIG_TCG_INTERPRETER)
1588 * Align all arguments, so that they land in predictable places
1589 * for passing off to ffi_call.
1591 want_align = true;
1592 #elif defined(TCG_TARGET_CALL_ALIGN_ARGS)
1593 /* Some targets want aligned 64 bit args */
1594 want_align = is_64bit;
1595 #endif
1597 if (TCG_TARGET_REG_BITS < 64 && want_align && (real_args & 1)) {
1598 op->args[pi++] = TCG_CALL_DUMMY_ARG;
1599 real_args++;
1602 if (TCG_TARGET_REG_BITS < 64 && is_64bit) {
1604 * If stack grows up, then we will be placing successive
1605 * arguments at lower addresses, which means we need to
1606 * reverse the order compared to how we would normally
1607 * treat either big or little-endian. For those arguments
1608 * that will wind up in registers, this still works for
1609 * HPPA (the only current STACK_GROWSUP target) since the
1610 * argument registers are *also* allocated in decreasing
1611 * order. If another such target is added, this logic may
1612 * have to get more complicated to differentiate between
1613 * stack arguments and register arguments.
1615 #if HOST_BIG_ENDIAN != defined(TCG_TARGET_STACK_GROWSUP)
1616 op->args[pi++] = temp_arg(args[i] + 1);
1617 op->args[pi++] = temp_arg(args[i]);
1618 #else
1619 op->args[pi++] = temp_arg(args[i]);
1620 op->args[pi++] = temp_arg(args[i] + 1);
1621 #endif
1622 real_args += 2;
1623 continue;
1626 op->args[pi++] = temp_arg(args[i]);
1627 real_args++;
1629 op->args[pi++] = (uintptr_t)func;
1630 op->args[pi++] = (uintptr_t)info;
1631 TCGOP_CALLI(op) = real_args;
1633 /* Make sure the fields didn't overflow. */
1634 tcg_debug_assert(TCGOP_CALLI(op) == real_args);
1635 tcg_debug_assert(pi <= ARRAY_SIZE(op->args));
1637 #if defined(__sparc__) && !defined(__arch64__) \
1638 && !defined(CONFIG_TCG_INTERPRETER)
1639 /* Free all of the parts we allocated above. */
1640 for (i = real_args = 0; i < orig_nargs; ++i) {
1641 int argtype = extract32(orig_typemask, (i + 1) * 3, 3);
1642 bool is_64bit = (argtype & ~1) == dh_typecode_i64;
1644 if (is_64bit) {
1645 tcg_temp_free_internal(args[real_args++]);
1646 tcg_temp_free_internal(args[real_args++]);
1647 } else {
1648 real_args++;
1651 if ((orig_typemask & 6) == dh_typecode_i64) {
1652 /* The 32-bit ABI returned two 32-bit pieces. Re-assemble them.
1653 Note that describing these as TCGv_i64 eliminates an unnecessary
1654 zero-extension that tcg_gen_concat_i32_i64 would create. */
1655 tcg_gen_concat32_i64(temp_tcgv_i64(ret), retl, reth);
1656 tcg_temp_free_i64(retl);
1657 tcg_temp_free_i64(reth);
1659 #elif defined(TCG_TARGET_EXTEND_ARGS) && TCG_TARGET_REG_BITS == 64
1660 for (i = 0; i < nargs; ++i) {
1661 int argtype = extract32(typemask, (i + 1) * 3, 3);
1662 bool is_32bit = (argtype & ~1) == dh_typecode_i32;
1664 if (is_32bit) {
1665 tcg_temp_free_internal(args[i]);
1668 #endif /* TCG_TARGET_EXTEND_ARGS */
1671 static void tcg_reg_alloc_start(TCGContext *s)
1673 int i, n;
1675 for (i = 0, n = s->nb_temps; i < n; i++) {
1676 TCGTemp *ts = &s->temps[i];
1677 TCGTempVal val = TEMP_VAL_MEM;
1679 switch (ts->kind) {
1680 case TEMP_CONST:
1681 val = TEMP_VAL_CONST;
1682 break;
1683 case TEMP_FIXED:
1684 val = TEMP_VAL_REG;
1685 break;
1686 case TEMP_GLOBAL:
1687 break;
1688 case TEMP_NORMAL:
1689 case TEMP_EBB:
1690 val = TEMP_VAL_DEAD;
1691 /* fall through */
1692 case TEMP_LOCAL:
1693 ts->mem_allocated = 0;
1694 break;
1695 default:
1696 g_assert_not_reached();
1698 ts->val_type = val;
1701 memset(s->reg_to_temp, 0, sizeof(s->reg_to_temp));
1704 static char *tcg_get_arg_str_ptr(TCGContext *s, char *buf, int buf_size,
1705 TCGTemp *ts)
1707 int idx = temp_idx(ts);
1709 switch (ts->kind) {
1710 case TEMP_FIXED:
1711 case TEMP_GLOBAL:
1712 pstrcpy(buf, buf_size, ts->name);
1713 break;
1714 case TEMP_LOCAL:
1715 snprintf(buf, buf_size, "loc%d", idx - s->nb_globals);
1716 break;
1717 case TEMP_EBB:
1718 snprintf(buf, buf_size, "ebb%d", idx - s->nb_globals);
1719 break;
1720 case TEMP_NORMAL:
1721 snprintf(buf, buf_size, "tmp%d", idx - s->nb_globals);
1722 break;
1723 case TEMP_CONST:
1724 switch (ts->type) {
1725 case TCG_TYPE_I32:
1726 snprintf(buf, buf_size, "$0x%x", (int32_t)ts->val);
1727 break;
1728 #if TCG_TARGET_REG_BITS > 32
1729 case TCG_TYPE_I64:
1730 snprintf(buf, buf_size, "$0x%" PRIx64, ts->val);
1731 break;
1732 #endif
1733 case TCG_TYPE_V64:
1734 case TCG_TYPE_V128:
1735 case TCG_TYPE_V256:
1736 snprintf(buf, buf_size, "v%d$0x%" PRIx64,
1737 64 << (ts->type - TCG_TYPE_V64), ts->val);
1738 break;
1739 default:
1740 g_assert_not_reached();
1742 break;
1744 return buf;
1747 static char *tcg_get_arg_str(TCGContext *s, char *buf,
1748 int buf_size, TCGArg arg)
1750 return tcg_get_arg_str_ptr(s, buf, buf_size, arg_temp(arg));
1753 static const char * const cond_name[] =
1755 [TCG_COND_NEVER] = "never",
1756 [TCG_COND_ALWAYS] = "always",
1757 [TCG_COND_EQ] = "eq",
1758 [TCG_COND_NE] = "ne",
1759 [TCG_COND_LT] = "lt",
1760 [TCG_COND_GE] = "ge",
1761 [TCG_COND_LE] = "le",
1762 [TCG_COND_GT] = "gt",
1763 [TCG_COND_LTU] = "ltu",
1764 [TCG_COND_GEU] = "geu",
1765 [TCG_COND_LEU] = "leu",
1766 [TCG_COND_GTU] = "gtu"
1769 static const char * const ldst_name[] =
1771 [MO_UB] = "ub",
1772 [MO_SB] = "sb",
1773 [MO_LEUW] = "leuw",
1774 [MO_LESW] = "lesw",
1775 [MO_LEUL] = "leul",
1776 [MO_LESL] = "lesl",
1777 [MO_LEUQ] = "leq",
1778 [MO_BEUW] = "beuw",
1779 [MO_BESW] = "besw",
1780 [MO_BEUL] = "beul",
1781 [MO_BESL] = "besl",
1782 [MO_BEUQ] = "beq",
1785 static const char * const alignment_name[(MO_AMASK >> MO_ASHIFT) + 1] = {
1786 #ifdef TARGET_ALIGNED_ONLY
1787 [MO_UNALN >> MO_ASHIFT] = "un+",
1788 [MO_ALIGN >> MO_ASHIFT] = "",
1789 #else
1790 [MO_UNALN >> MO_ASHIFT] = "",
1791 [MO_ALIGN >> MO_ASHIFT] = "al+",
1792 #endif
1793 [MO_ALIGN_2 >> MO_ASHIFT] = "al2+",
1794 [MO_ALIGN_4 >> MO_ASHIFT] = "al4+",
1795 [MO_ALIGN_8 >> MO_ASHIFT] = "al8+",
1796 [MO_ALIGN_16 >> MO_ASHIFT] = "al16+",
1797 [MO_ALIGN_32 >> MO_ASHIFT] = "al32+",
1798 [MO_ALIGN_64 >> MO_ASHIFT] = "al64+",
1801 static const char bswap_flag_name[][6] = {
1802 [TCG_BSWAP_IZ] = "iz",
1803 [TCG_BSWAP_OZ] = "oz",
1804 [TCG_BSWAP_OS] = "os",
1805 [TCG_BSWAP_IZ | TCG_BSWAP_OZ] = "iz,oz",
1806 [TCG_BSWAP_IZ | TCG_BSWAP_OS] = "iz,os",
1809 static inline bool tcg_regset_single(TCGRegSet d)
1811 return (d & (d - 1)) == 0;
1814 static inline TCGReg tcg_regset_first(TCGRegSet d)
1816 if (TCG_TARGET_NB_REGS <= 32) {
1817 return ctz32(d);
1818 } else {
1819 return ctz64(d);
1823 /* Return only the number of characters output -- no error return. */
1824 #define ne_fprintf(...) \
1825 ({ int ret_ = fprintf(__VA_ARGS__); ret_ >= 0 ? ret_ : 0; })
1827 static void tcg_dump_ops(TCGContext *s, FILE *f, bool have_prefs)
1829 char buf[128];
1830 TCGOp *op;
1832 QTAILQ_FOREACH(op, &s->ops, link) {
1833 int i, k, nb_oargs, nb_iargs, nb_cargs;
1834 const TCGOpDef *def;
1835 TCGOpcode c;
1836 int col = 0;
1838 c = op->opc;
1839 def = &tcg_op_defs[c];
1841 if (c == INDEX_op_insn_start) {
1842 nb_oargs = 0;
1843 col += ne_fprintf(f, "\n ----");
1845 for (i = 0; i < TARGET_INSN_START_WORDS; ++i) {
1846 target_ulong a;
1847 #if TARGET_LONG_BITS > TCG_TARGET_REG_BITS
1848 a = deposit64(op->args[i * 2], 32, 32, op->args[i * 2 + 1]);
1849 #else
1850 a = op->args[i];
1851 #endif
1852 col += ne_fprintf(f, " " TARGET_FMT_lx, a);
1854 } else if (c == INDEX_op_call) {
1855 const TCGHelperInfo *info = tcg_call_info(op);
1856 void *func = tcg_call_func(op);
1858 /* variable number of arguments */
1859 nb_oargs = TCGOP_CALLO(op);
1860 nb_iargs = TCGOP_CALLI(op);
1861 nb_cargs = def->nb_cargs;
1863 col += ne_fprintf(f, " %s ", def->name);
1866 * Print the function name from TCGHelperInfo, if available.
1867 * Note that plugins have a template function for the info,
1868 * but the actual function pointer comes from the plugin.
1870 if (func == info->func) {
1871 col += ne_fprintf(f, "%s", info->name);
1872 } else {
1873 col += ne_fprintf(f, "plugin(%p)", func);
1876 col += ne_fprintf(f, ",$0x%x,$%d", info->flags, nb_oargs);
1877 for (i = 0; i < nb_oargs; i++) {
1878 col += ne_fprintf(f, ",%s", tcg_get_arg_str(s, buf, sizeof(buf),
1879 op->args[i]));
1881 for (i = 0; i < nb_iargs; i++) {
1882 TCGArg arg = op->args[nb_oargs + i];
1883 const char *t = "<dummy>";
1884 if (arg != TCG_CALL_DUMMY_ARG) {
1885 t = tcg_get_arg_str(s, buf, sizeof(buf), arg);
1887 col += ne_fprintf(f, ",%s", t);
1889 } else {
1890 col += ne_fprintf(f, " %s ", def->name);
1892 nb_oargs = def->nb_oargs;
1893 nb_iargs = def->nb_iargs;
1894 nb_cargs = def->nb_cargs;
1896 if (def->flags & TCG_OPF_VECTOR) {
1897 col += ne_fprintf(f, "v%d,e%d,", 64 << TCGOP_VECL(op),
1898 8 << TCGOP_VECE(op));
1901 k = 0;
1902 for (i = 0; i < nb_oargs; i++) {
1903 const char *sep = k ? "," : "";
1904 col += ne_fprintf(f, "%s%s", sep,
1905 tcg_get_arg_str(s, buf, sizeof(buf),
1906 op->args[k++]));
1908 for (i = 0; i < nb_iargs; i++) {
1909 const char *sep = k ? "," : "";
1910 col += ne_fprintf(f, "%s%s", sep,
1911 tcg_get_arg_str(s, buf, sizeof(buf),
1912 op->args[k++]));
1914 switch (c) {
1915 case INDEX_op_brcond_i32:
1916 case INDEX_op_setcond_i32:
1917 case INDEX_op_movcond_i32:
1918 case INDEX_op_brcond2_i32:
1919 case INDEX_op_setcond2_i32:
1920 case INDEX_op_brcond_i64:
1921 case INDEX_op_setcond_i64:
1922 case INDEX_op_movcond_i64:
1923 case INDEX_op_cmp_vec:
1924 case INDEX_op_cmpsel_vec:
1925 if (op->args[k] < ARRAY_SIZE(cond_name)
1926 && cond_name[op->args[k]]) {
1927 col += ne_fprintf(f, ",%s", cond_name[op->args[k++]]);
1928 } else {
1929 col += ne_fprintf(f, ",$0x%" TCG_PRIlx, op->args[k++]);
1931 i = 1;
1932 break;
1933 case INDEX_op_qemu_ld_i32:
1934 case INDEX_op_qemu_st_i32:
1935 case INDEX_op_qemu_st8_i32:
1936 case INDEX_op_qemu_ld_i64:
1937 case INDEX_op_qemu_st_i64:
1939 MemOpIdx oi = op->args[k++];
1940 MemOp op = get_memop(oi);
1941 unsigned ix = get_mmuidx(oi);
1943 if (op & ~(MO_AMASK | MO_BSWAP | MO_SSIZE)) {
1944 col += ne_fprintf(f, ",$0x%x,%u", op, ix);
1945 } else {
1946 const char *s_al, *s_op;
1947 s_al = alignment_name[(op & MO_AMASK) >> MO_ASHIFT];
1948 s_op = ldst_name[op & (MO_BSWAP | MO_SSIZE)];
1949 col += ne_fprintf(f, ",%s%s,%u", s_al, s_op, ix);
1951 i = 1;
1953 break;
1954 case INDEX_op_bswap16_i32:
1955 case INDEX_op_bswap16_i64:
1956 case INDEX_op_bswap32_i32:
1957 case INDEX_op_bswap32_i64:
1958 case INDEX_op_bswap64_i64:
1960 TCGArg flags = op->args[k];
1961 const char *name = NULL;
1963 if (flags < ARRAY_SIZE(bswap_flag_name)) {
1964 name = bswap_flag_name[flags];
1966 if (name) {
1967 col += ne_fprintf(f, ",%s", name);
1968 } else {
1969 col += ne_fprintf(f, ",$0x%" TCG_PRIlx, flags);
1971 i = k = 1;
1973 break;
1974 default:
1975 i = 0;
1976 break;
1978 switch (c) {
1979 case INDEX_op_set_label:
1980 case INDEX_op_br:
1981 case INDEX_op_brcond_i32:
1982 case INDEX_op_brcond_i64:
1983 case INDEX_op_brcond2_i32:
1984 col += ne_fprintf(f, "%s$L%d", k ? "," : "",
1985 arg_label(op->args[k])->id);
1986 i++, k++;
1987 break;
1988 default:
1989 break;
1991 for (; i < nb_cargs; i++, k++) {
1992 col += ne_fprintf(f, "%s$0x%" TCG_PRIlx, k ? "," : "",
1993 op->args[k]);
1997 if (have_prefs || op->life) {
1998 for (; col < 40; ++col) {
1999 putc(' ', f);
2003 if (op->life) {
2004 unsigned life = op->life;
2006 if (life & (SYNC_ARG * 3)) {
2007 ne_fprintf(f, " sync:");
2008 for (i = 0; i < 2; ++i) {
2009 if (life & (SYNC_ARG << i)) {
2010 ne_fprintf(f, " %d", i);
2014 life /= DEAD_ARG;
2015 if (life) {
2016 ne_fprintf(f, " dead:");
2017 for (i = 0; life; ++i, life >>= 1) {
2018 if (life & 1) {
2019 ne_fprintf(f, " %d", i);
2025 if (have_prefs) {
2026 for (i = 0; i < nb_oargs; ++i) {
2027 TCGRegSet set = op->output_pref[i];
2029 if (i == 0) {
2030 ne_fprintf(f, " pref=");
2031 } else {
2032 ne_fprintf(f, ",");
2034 if (set == 0) {
2035 ne_fprintf(f, "none");
2036 } else if (set == MAKE_64BIT_MASK(0, TCG_TARGET_NB_REGS)) {
2037 ne_fprintf(f, "all");
2038 #ifdef CONFIG_DEBUG_TCG
2039 } else if (tcg_regset_single(set)) {
2040 TCGReg reg = tcg_regset_first(set);
2041 ne_fprintf(f, "%s", tcg_target_reg_names[reg]);
2042 #endif
2043 } else if (TCG_TARGET_NB_REGS <= 32) {
2044 ne_fprintf(f, "0x%x", (uint32_t)set);
2045 } else {
2046 ne_fprintf(f, "0x%" PRIx64, (uint64_t)set);
2051 putc('\n', f);
2055 /* we give more priority to constraints with less registers */
2056 static int get_constraint_priority(const TCGOpDef *def, int k)
2058 const TCGArgConstraint *arg_ct = &def->args_ct[k];
2059 int n;
2061 if (arg_ct->oalias) {
2062 /* an alias is equivalent to a single register */
2063 n = 1;
2064 } else {
2065 n = ctpop64(arg_ct->regs);
2067 return TCG_TARGET_NB_REGS - n + 1;
2070 /* sort from highest priority to lowest */
2071 static void sort_constraints(TCGOpDef *def, int start, int n)
2073 int i, j;
2074 TCGArgConstraint *a = def->args_ct;
2076 for (i = 0; i < n; i++) {
2077 a[start + i].sort_index = start + i;
2079 if (n <= 1) {
2080 return;
2082 for (i = 0; i < n - 1; i++) {
2083 for (j = i + 1; j < n; j++) {
2084 int p1 = get_constraint_priority(def, a[start + i].sort_index);
2085 int p2 = get_constraint_priority(def, a[start + j].sort_index);
2086 if (p1 < p2) {
2087 int tmp = a[start + i].sort_index;
2088 a[start + i].sort_index = a[start + j].sort_index;
2089 a[start + j].sort_index = tmp;
2095 static void process_op_defs(TCGContext *s)
2097 TCGOpcode op;
2099 for (op = 0; op < NB_OPS; op++) {
2100 TCGOpDef *def = &tcg_op_defs[op];
2101 const TCGTargetOpDef *tdefs;
2102 int i, nb_args;
2104 if (def->flags & TCG_OPF_NOT_PRESENT) {
2105 continue;
2108 nb_args = def->nb_iargs + def->nb_oargs;
2109 if (nb_args == 0) {
2110 continue;
2114 * Macro magic should make it impossible, but double-check that
2115 * the array index is in range. Since the signness of an enum
2116 * is implementation defined, force the result to unsigned.
2118 unsigned con_set = tcg_target_op_def(op);
2119 tcg_debug_assert(con_set < ARRAY_SIZE(constraint_sets));
2120 tdefs = &constraint_sets[con_set];
2122 for (i = 0; i < nb_args; i++) {
2123 const char *ct_str = tdefs->args_ct_str[i];
2124 /* Incomplete TCGTargetOpDef entry. */
2125 tcg_debug_assert(ct_str != NULL);
2127 while (*ct_str != '\0') {
2128 switch(*ct_str) {
2129 case '0' ... '9':
2131 int oarg = *ct_str - '0';
2132 tcg_debug_assert(ct_str == tdefs->args_ct_str[i]);
2133 tcg_debug_assert(oarg < def->nb_oargs);
2134 tcg_debug_assert(def->args_ct[oarg].regs != 0);
2135 def->args_ct[i] = def->args_ct[oarg];
2136 /* The output sets oalias. */
2137 def->args_ct[oarg].oalias = true;
2138 def->args_ct[oarg].alias_index = i;
2139 /* The input sets ialias. */
2140 def->args_ct[i].ialias = true;
2141 def->args_ct[i].alias_index = oarg;
2143 ct_str++;
2144 break;
2145 case '&':
2146 def->args_ct[i].newreg = true;
2147 ct_str++;
2148 break;
2149 case 'i':
2150 def->args_ct[i].ct |= TCG_CT_CONST;
2151 ct_str++;
2152 break;
2154 /* Include all of the target-specific constraints. */
2156 #undef CONST
2157 #define CONST(CASE, MASK) \
2158 case CASE: def->args_ct[i].ct |= MASK; ct_str++; break;
2159 #define REGS(CASE, MASK) \
2160 case CASE: def->args_ct[i].regs |= MASK; ct_str++; break;
2162 #include "tcg-target-con-str.h"
2164 #undef REGS
2165 #undef CONST
2166 default:
2167 /* Typo in TCGTargetOpDef constraint. */
2168 g_assert_not_reached();
2173 /* TCGTargetOpDef entry with too much information? */
2174 tcg_debug_assert(i == TCG_MAX_OP_ARGS || tdefs->args_ct_str[i] == NULL);
2176 /* sort the constraints (XXX: this is just an heuristic) */
2177 sort_constraints(def, 0, def->nb_oargs);
2178 sort_constraints(def, def->nb_oargs, def->nb_iargs);
2182 void tcg_op_remove(TCGContext *s, TCGOp *op)
2184 TCGLabel *label;
2186 switch (op->opc) {
2187 case INDEX_op_br:
2188 label = arg_label(op->args[0]);
2189 label->refs--;
2190 break;
2191 case INDEX_op_brcond_i32:
2192 case INDEX_op_brcond_i64:
2193 label = arg_label(op->args[3]);
2194 label->refs--;
2195 break;
2196 case INDEX_op_brcond2_i32:
2197 label = arg_label(op->args[5]);
2198 label->refs--;
2199 break;
2200 default:
2201 break;
2204 QTAILQ_REMOVE(&s->ops, op, link);
2205 QTAILQ_INSERT_TAIL(&s->free_ops, op, link);
2206 s->nb_ops--;
2208 #ifdef CONFIG_PROFILER
2209 qatomic_set(&s->prof.del_op_count, s->prof.del_op_count + 1);
2210 #endif
2213 void tcg_remove_ops_after(TCGOp *op)
2215 TCGContext *s = tcg_ctx;
2217 while (true) {
2218 TCGOp *last = tcg_last_op();
2219 if (last == op) {
2220 return;
2222 tcg_op_remove(s, last);
2226 static TCGOp *tcg_op_alloc(TCGOpcode opc)
2228 TCGContext *s = tcg_ctx;
2229 TCGOp *op;
2231 if (likely(QTAILQ_EMPTY(&s->free_ops))) {
2232 op = tcg_malloc(sizeof(TCGOp));
2233 } else {
2234 op = QTAILQ_FIRST(&s->free_ops);
2235 QTAILQ_REMOVE(&s->free_ops, op, link);
2237 memset(op, 0, offsetof(TCGOp, link));
2238 op->opc = opc;
2239 s->nb_ops++;
2241 return op;
2244 TCGOp *tcg_emit_op(TCGOpcode opc)
2246 TCGOp *op = tcg_op_alloc(opc);
2247 QTAILQ_INSERT_TAIL(&tcg_ctx->ops, op, link);
2248 return op;
2251 TCGOp *tcg_op_insert_before(TCGContext *s, TCGOp *old_op, TCGOpcode opc)
2253 TCGOp *new_op = tcg_op_alloc(opc);
2254 QTAILQ_INSERT_BEFORE(old_op, new_op, link);
2255 return new_op;
2258 TCGOp *tcg_op_insert_after(TCGContext *s, TCGOp *old_op, TCGOpcode opc)
2260 TCGOp *new_op = tcg_op_alloc(opc);
2261 QTAILQ_INSERT_AFTER(&s->ops, old_op, new_op, link);
2262 return new_op;
2265 /* Reachable analysis : remove unreachable code. */
2266 static void reachable_code_pass(TCGContext *s)
2268 TCGOp *op, *op_next;
2269 bool dead = false;
2271 QTAILQ_FOREACH_SAFE(op, &s->ops, link, op_next) {
2272 bool remove = dead;
2273 TCGLabel *label;
2275 switch (op->opc) {
2276 case INDEX_op_set_label:
2277 label = arg_label(op->args[0]);
2278 if (label->refs == 0) {
2280 * While there is an occasional backward branch, virtually
2281 * all branches generated by the translators are forward.
2282 * Which means that generally we will have already removed
2283 * all references to the label that will be, and there is
2284 * little to be gained by iterating.
2286 remove = true;
2287 } else {
2288 /* Once we see a label, insns become live again. */
2289 dead = false;
2290 remove = false;
2293 * Optimization can fold conditional branches to unconditional.
2294 * If we find a label with one reference which is preceded by
2295 * an unconditional branch to it, remove both. This needed to
2296 * wait until the dead code in between them was removed.
2298 if (label->refs == 1) {
2299 TCGOp *op_prev = QTAILQ_PREV(op, link);
2300 if (op_prev->opc == INDEX_op_br &&
2301 label == arg_label(op_prev->args[0])) {
2302 tcg_op_remove(s, op_prev);
2303 remove = true;
2307 break;
2309 case INDEX_op_br:
2310 case INDEX_op_exit_tb:
2311 case INDEX_op_goto_ptr:
2312 /* Unconditional branches; everything following is dead. */
2313 dead = true;
2314 break;
2316 case INDEX_op_call:
2317 /* Notice noreturn helper calls, raising exceptions. */
2318 if (tcg_call_flags(op) & TCG_CALL_NO_RETURN) {
2319 dead = true;
2321 break;
2323 case INDEX_op_insn_start:
2324 /* Never remove -- we need to keep these for unwind. */
2325 remove = false;
2326 break;
2328 default:
2329 break;
2332 if (remove) {
2333 tcg_op_remove(s, op);
2338 #define TS_DEAD 1
2339 #define TS_MEM 2
2341 #define IS_DEAD_ARG(n) (arg_life & (DEAD_ARG << (n)))
2342 #define NEED_SYNC_ARG(n) (arg_life & (SYNC_ARG << (n)))
2344 /* For liveness_pass_1, the register preferences for a given temp. */
2345 static inline TCGRegSet *la_temp_pref(TCGTemp *ts)
2347 return ts->state_ptr;
2350 /* For liveness_pass_1, reset the preferences for a given temp to the
2351 * maximal regset for its type.
2353 static inline void la_reset_pref(TCGTemp *ts)
2355 *la_temp_pref(ts)
2356 = (ts->state == TS_DEAD ? 0 : tcg_target_available_regs[ts->type]);
2359 /* liveness analysis: end of function: all temps are dead, and globals
2360 should be in memory. */
2361 static void la_func_end(TCGContext *s, int ng, int nt)
2363 int i;
2365 for (i = 0; i < ng; ++i) {
2366 s->temps[i].state = TS_DEAD | TS_MEM;
2367 la_reset_pref(&s->temps[i]);
2369 for (i = ng; i < nt; ++i) {
2370 s->temps[i].state = TS_DEAD;
2371 la_reset_pref(&s->temps[i]);
2375 /* liveness analysis: end of basic block: all temps are dead, globals
2376 and local temps should be in memory. */
2377 static void la_bb_end(TCGContext *s, int ng, int nt)
2379 int i;
2381 for (i = 0; i < nt; ++i) {
2382 TCGTemp *ts = &s->temps[i];
2383 int state;
2385 switch (ts->kind) {
2386 case TEMP_FIXED:
2387 case TEMP_GLOBAL:
2388 case TEMP_LOCAL:
2389 state = TS_DEAD | TS_MEM;
2390 break;
2391 case TEMP_NORMAL:
2392 case TEMP_EBB:
2393 case TEMP_CONST:
2394 state = TS_DEAD;
2395 break;
2396 default:
2397 g_assert_not_reached();
2399 ts->state = state;
2400 la_reset_pref(ts);
2404 /* liveness analysis: sync globals back to memory. */
2405 static void la_global_sync(TCGContext *s, int ng)
2407 int i;
2409 for (i = 0; i < ng; ++i) {
2410 int state = s->temps[i].state;
2411 s->temps[i].state = state | TS_MEM;
2412 if (state == TS_DEAD) {
2413 /* If the global was previously dead, reset prefs. */
2414 la_reset_pref(&s->temps[i]);
2420 * liveness analysis: conditional branch: all temps are dead unless
2421 * explicitly live-across-conditional-branch, globals and local temps
2422 * should be synced.
2424 static void la_bb_sync(TCGContext *s, int ng, int nt)
2426 la_global_sync(s, ng);
2428 for (int i = ng; i < nt; ++i) {
2429 TCGTemp *ts = &s->temps[i];
2430 int state;
2432 switch (ts->kind) {
2433 case TEMP_LOCAL:
2434 state = ts->state;
2435 ts->state = state | TS_MEM;
2436 if (state != TS_DEAD) {
2437 continue;
2439 break;
2440 case TEMP_NORMAL:
2441 s->temps[i].state = TS_DEAD;
2442 break;
2443 case TEMP_EBB:
2444 case TEMP_CONST:
2445 continue;
2446 default:
2447 g_assert_not_reached();
2449 la_reset_pref(&s->temps[i]);
2453 /* liveness analysis: sync globals back to memory and kill. */
2454 static void la_global_kill(TCGContext *s, int ng)
2456 int i;
2458 for (i = 0; i < ng; i++) {
2459 s->temps[i].state = TS_DEAD | TS_MEM;
2460 la_reset_pref(&s->temps[i]);
2464 /* liveness analysis: note live globals crossing calls. */
2465 static void la_cross_call(TCGContext *s, int nt)
2467 TCGRegSet mask = ~tcg_target_call_clobber_regs;
2468 int i;
2470 for (i = 0; i < nt; i++) {
2471 TCGTemp *ts = &s->temps[i];
2472 if (!(ts->state & TS_DEAD)) {
2473 TCGRegSet *pset = la_temp_pref(ts);
2474 TCGRegSet set = *pset;
2476 set &= mask;
2477 /* If the combination is not possible, restart. */
2478 if (set == 0) {
2479 set = tcg_target_available_regs[ts->type] & mask;
2481 *pset = set;
2486 /* Liveness analysis : update the opc_arg_life array to tell if a
2487 given input arguments is dead. Instructions updating dead
2488 temporaries are removed. */
2489 static void liveness_pass_1(TCGContext *s)
2491 int nb_globals = s->nb_globals;
2492 int nb_temps = s->nb_temps;
2493 TCGOp *op, *op_prev;
2494 TCGRegSet *prefs;
2495 int i;
2497 prefs = tcg_malloc(sizeof(TCGRegSet) * nb_temps);
2498 for (i = 0; i < nb_temps; ++i) {
2499 s->temps[i].state_ptr = prefs + i;
2502 /* ??? Should be redundant with the exit_tb that ends the TB. */
2503 la_func_end(s, nb_globals, nb_temps);
2505 QTAILQ_FOREACH_REVERSE_SAFE(op, &s->ops, link, op_prev) {
2506 int nb_iargs, nb_oargs;
2507 TCGOpcode opc_new, opc_new2;
2508 bool have_opc_new2;
2509 TCGLifeData arg_life = 0;
2510 TCGTemp *ts;
2511 TCGOpcode opc = op->opc;
2512 const TCGOpDef *def = &tcg_op_defs[opc];
2514 switch (opc) {
2515 case INDEX_op_call:
2517 int call_flags;
2518 int nb_call_regs;
2520 nb_oargs = TCGOP_CALLO(op);
2521 nb_iargs = TCGOP_CALLI(op);
2522 call_flags = tcg_call_flags(op);
2524 /* pure functions can be removed if their result is unused */
2525 if (call_flags & TCG_CALL_NO_SIDE_EFFECTS) {
2526 for (i = 0; i < nb_oargs; i++) {
2527 ts = arg_temp(op->args[i]);
2528 if (ts->state != TS_DEAD) {
2529 goto do_not_remove_call;
2532 goto do_remove;
2534 do_not_remove_call:
2536 /* Output args are dead. */
2537 for (i = 0; i < nb_oargs; i++) {
2538 ts = arg_temp(op->args[i]);
2539 if (ts->state & TS_DEAD) {
2540 arg_life |= DEAD_ARG << i;
2542 if (ts->state & TS_MEM) {
2543 arg_life |= SYNC_ARG << i;
2545 ts->state = TS_DEAD;
2546 la_reset_pref(ts);
2548 /* Not used -- it will be tcg_target_call_oarg_regs[i]. */
2549 op->output_pref[i] = 0;
2552 if (!(call_flags & (TCG_CALL_NO_WRITE_GLOBALS |
2553 TCG_CALL_NO_READ_GLOBALS))) {
2554 la_global_kill(s, nb_globals);
2555 } else if (!(call_flags & TCG_CALL_NO_READ_GLOBALS)) {
2556 la_global_sync(s, nb_globals);
2559 /* Record arguments that die in this helper. */
2560 for (i = nb_oargs; i < nb_iargs + nb_oargs; i++) {
2561 ts = arg_temp(op->args[i]);
2562 if (ts && ts->state & TS_DEAD) {
2563 arg_life |= DEAD_ARG << i;
2567 /* For all live registers, remove call-clobbered prefs. */
2568 la_cross_call(s, nb_temps);
2570 nb_call_regs = ARRAY_SIZE(tcg_target_call_iarg_regs);
2572 /* Input arguments are live for preceding opcodes. */
2573 for (i = 0; i < nb_iargs; i++) {
2574 ts = arg_temp(op->args[i + nb_oargs]);
2575 if (ts && ts->state & TS_DEAD) {
2576 /* For those arguments that die, and will be allocated
2577 * in registers, clear the register set for that arg,
2578 * to be filled in below. For args that will be on
2579 * the stack, reset to any available reg.
2581 *la_temp_pref(ts)
2582 = (i < nb_call_regs ? 0 :
2583 tcg_target_available_regs[ts->type]);
2584 ts->state &= ~TS_DEAD;
2588 /* For each input argument, add its input register to prefs.
2589 If a temp is used once, this produces a single set bit. */
2590 for (i = 0; i < MIN(nb_call_regs, nb_iargs); i++) {
2591 ts = arg_temp(op->args[i + nb_oargs]);
2592 if (ts) {
2593 tcg_regset_set_reg(*la_temp_pref(ts),
2594 tcg_target_call_iarg_regs[i]);
2598 break;
2599 case INDEX_op_insn_start:
2600 break;
2601 case INDEX_op_discard:
2602 /* mark the temporary as dead */
2603 ts = arg_temp(op->args[0]);
2604 ts->state = TS_DEAD;
2605 la_reset_pref(ts);
2606 break;
2608 case INDEX_op_add2_i32:
2609 opc_new = INDEX_op_add_i32;
2610 goto do_addsub2;
2611 case INDEX_op_sub2_i32:
2612 opc_new = INDEX_op_sub_i32;
2613 goto do_addsub2;
2614 case INDEX_op_add2_i64:
2615 opc_new = INDEX_op_add_i64;
2616 goto do_addsub2;
2617 case INDEX_op_sub2_i64:
2618 opc_new = INDEX_op_sub_i64;
2619 do_addsub2:
2620 nb_iargs = 4;
2621 nb_oargs = 2;
2622 /* Test if the high part of the operation is dead, but not
2623 the low part. The result can be optimized to a simple
2624 add or sub. This happens often for x86_64 guest when the
2625 cpu mode is set to 32 bit. */
2626 if (arg_temp(op->args[1])->state == TS_DEAD) {
2627 if (arg_temp(op->args[0])->state == TS_DEAD) {
2628 goto do_remove;
2630 /* Replace the opcode and adjust the args in place,
2631 leaving 3 unused args at the end. */
2632 op->opc = opc = opc_new;
2633 op->args[1] = op->args[2];
2634 op->args[2] = op->args[4];
2635 /* Fall through and mark the single-word operation live. */
2636 nb_iargs = 2;
2637 nb_oargs = 1;
2639 goto do_not_remove;
2641 case INDEX_op_mulu2_i32:
2642 opc_new = INDEX_op_mul_i32;
2643 opc_new2 = INDEX_op_muluh_i32;
2644 have_opc_new2 = TCG_TARGET_HAS_muluh_i32;
2645 goto do_mul2;
2646 case INDEX_op_muls2_i32:
2647 opc_new = INDEX_op_mul_i32;
2648 opc_new2 = INDEX_op_mulsh_i32;
2649 have_opc_new2 = TCG_TARGET_HAS_mulsh_i32;
2650 goto do_mul2;
2651 case INDEX_op_mulu2_i64:
2652 opc_new = INDEX_op_mul_i64;
2653 opc_new2 = INDEX_op_muluh_i64;
2654 have_opc_new2 = TCG_TARGET_HAS_muluh_i64;
2655 goto do_mul2;
2656 case INDEX_op_muls2_i64:
2657 opc_new = INDEX_op_mul_i64;
2658 opc_new2 = INDEX_op_mulsh_i64;
2659 have_opc_new2 = TCG_TARGET_HAS_mulsh_i64;
2660 goto do_mul2;
2661 do_mul2:
2662 nb_iargs = 2;
2663 nb_oargs = 2;
2664 if (arg_temp(op->args[1])->state == TS_DEAD) {
2665 if (arg_temp(op->args[0])->state == TS_DEAD) {
2666 /* Both parts of the operation are dead. */
2667 goto do_remove;
2669 /* The high part of the operation is dead; generate the low. */
2670 op->opc = opc = opc_new;
2671 op->args[1] = op->args[2];
2672 op->args[2] = op->args[3];
2673 } else if (arg_temp(op->args[0])->state == TS_DEAD && have_opc_new2) {
2674 /* The low part of the operation is dead; generate the high. */
2675 op->opc = opc = opc_new2;
2676 op->args[0] = op->args[1];
2677 op->args[1] = op->args[2];
2678 op->args[2] = op->args[3];
2679 } else {
2680 goto do_not_remove;
2682 /* Mark the single-word operation live. */
2683 nb_oargs = 1;
2684 goto do_not_remove;
2686 default:
2687 /* XXX: optimize by hardcoding common cases (e.g. triadic ops) */
2688 nb_iargs = def->nb_iargs;
2689 nb_oargs = def->nb_oargs;
2691 /* Test if the operation can be removed because all
2692 its outputs are dead. We assume that nb_oargs == 0
2693 implies side effects */
2694 if (!(def->flags & TCG_OPF_SIDE_EFFECTS) && nb_oargs != 0) {
2695 for (i = 0; i < nb_oargs; i++) {
2696 if (arg_temp(op->args[i])->state != TS_DEAD) {
2697 goto do_not_remove;
2700 goto do_remove;
2702 goto do_not_remove;
2704 do_remove:
2705 tcg_op_remove(s, op);
2706 break;
2708 do_not_remove:
2709 for (i = 0; i < nb_oargs; i++) {
2710 ts = arg_temp(op->args[i]);
2712 /* Remember the preference of the uses that followed. */
2713 op->output_pref[i] = *la_temp_pref(ts);
2715 /* Output args are dead. */
2716 if (ts->state & TS_DEAD) {
2717 arg_life |= DEAD_ARG << i;
2719 if (ts->state & TS_MEM) {
2720 arg_life |= SYNC_ARG << i;
2722 ts->state = TS_DEAD;
2723 la_reset_pref(ts);
2726 /* If end of basic block, update. */
2727 if (def->flags & TCG_OPF_BB_EXIT) {
2728 la_func_end(s, nb_globals, nb_temps);
2729 } else if (def->flags & TCG_OPF_COND_BRANCH) {
2730 la_bb_sync(s, nb_globals, nb_temps);
2731 } else if (def->flags & TCG_OPF_BB_END) {
2732 la_bb_end(s, nb_globals, nb_temps);
2733 } else if (def->flags & TCG_OPF_SIDE_EFFECTS) {
2734 la_global_sync(s, nb_globals);
2735 if (def->flags & TCG_OPF_CALL_CLOBBER) {
2736 la_cross_call(s, nb_temps);
2740 /* Record arguments that die in this opcode. */
2741 for (i = nb_oargs; i < nb_oargs + nb_iargs; i++) {
2742 ts = arg_temp(op->args[i]);
2743 if (ts->state & TS_DEAD) {
2744 arg_life |= DEAD_ARG << i;
2748 /* Input arguments are live for preceding opcodes. */
2749 for (i = nb_oargs; i < nb_oargs + nb_iargs; i++) {
2750 ts = arg_temp(op->args[i]);
2751 if (ts->state & TS_DEAD) {
2752 /* For operands that were dead, initially allow
2753 all regs for the type. */
2754 *la_temp_pref(ts) = tcg_target_available_regs[ts->type];
2755 ts->state &= ~TS_DEAD;
2759 /* Incorporate constraints for this operand. */
2760 switch (opc) {
2761 case INDEX_op_mov_i32:
2762 case INDEX_op_mov_i64:
2763 /* Note that these are TCG_OPF_NOT_PRESENT and do not
2764 have proper constraints. That said, special case
2765 moves to propagate preferences backward. */
2766 if (IS_DEAD_ARG(1)) {
2767 *la_temp_pref(arg_temp(op->args[0]))
2768 = *la_temp_pref(arg_temp(op->args[1]));
2770 break;
2772 default:
2773 for (i = nb_oargs; i < nb_oargs + nb_iargs; i++) {
2774 const TCGArgConstraint *ct = &def->args_ct[i];
2775 TCGRegSet set, *pset;
2777 ts = arg_temp(op->args[i]);
2778 pset = la_temp_pref(ts);
2779 set = *pset;
2781 set &= ct->regs;
2782 if (ct->ialias) {
2783 set &= op->output_pref[ct->alias_index];
2785 /* If the combination is not possible, restart. */
2786 if (set == 0) {
2787 set = ct->regs;
2789 *pset = set;
2791 break;
2793 break;
2795 op->life = arg_life;
2799 /* Liveness analysis: Convert indirect regs to direct temporaries. */
2800 static bool liveness_pass_2(TCGContext *s)
2802 int nb_globals = s->nb_globals;
2803 int nb_temps, i;
2804 bool changes = false;
2805 TCGOp *op, *op_next;
2807 /* Create a temporary for each indirect global. */
2808 for (i = 0; i < nb_globals; ++i) {
2809 TCGTemp *its = &s->temps[i];
2810 if (its->indirect_reg) {
2811 TCGTemp *dts = tcg_temp_alloc(s);
2812 dts->type = its->type;
2813 dts->base_type = its->base_type;
2814 dts->kind = TEMP_EBB;
2815 its->state_ptr = dts;
2816 } else {
2817 its->state_ptr = NULL;
2819 /* All globals begin dead. */
2820 its->state = TS_DEAD;
2822 for (nb_temps = s->nb_temps; i < nb_temps; ++i) {
2823 TCGTemp *its = &s->temps[i];
2824 its->state_ptr = NULL;
2825 its->state = TS_DEAD;
2828 QTAILQ_FOREACH_SAFE(op, &s->ops, link, op_next) {
2829 TCGOpcode opc = op->opc;
2830 const TCGOpDef *def = &tcg_op_defs[opc];
2831 TCGLifeData arg_life = op->life;
2832 int nb_iargs, nb_oargs, call_flags;
2833 TCGTemp *arg_ts, *dir_ts;
2835 if (opc == INDEX_op_call) {
2836 nb_oargs = TCGOP_CALLO(op);
2837 nb_iargs = TCGOP_CALLI(op);
2838 call_flags = tcg_call_flags(op);
2839 } else {
2840 nb_iargs = def->nb_iargs;
2841 nb_oargs = def->nb_oargs;
2843 /* Set flags similar to how calls require. */
2844 if (def->flags & TCG_OPF_COND_BRANCH) {
2845 /* Like reading globals: sync_globals */
2846 call_flags = TCG_CALL_NO_WRITE_GLOBALS;
2847 } else if (def->flags & TCG_OPF_BB_END) {
2848 /* Like writing globals: save_globals */
2849 call_flags = 0;
2850 } else if (def->flags & TCG_OPF_SIDE_EFFECTS) {
2851 /* Like reading globals: sync_globals */
2852 call_flags = TCG_CALL_NO_WRITE_GLOBALS;
2853 } else {
2854 /* No effect on globals. */
2855 call_flags = (TCG_CALL_NO_READ_GLOBALS |
2856 TCG_CALL_NO_WRITE_GLOBALS);
2860 /* Make sure that input arguments are available. */
2861 for (i = nb_oargs; i < nb_iargs + nb_oargs; i++) {
2862 arg_ts = arg_temp(op->args[i]);
2863 if (arg_ts) {
2864 dir_ts = arg_ts->state_ptr;
2865 if (dir_ts && arg_ts->state == TS_DEAD) {
2866 TCGOpcode lopc = (arg_ts->type == TCG_TYPE_I32
2867 ? INDEX_op_ld_i32
2868 : INDEX_op_ld_i64);
2869 TCGOp *lop = tcg_op_insert_before(s, op, lopc);
2871 lop->args[0] = temp_arg(dir_ts);
2872 lop->args[1] = temp_arg(arg_ts->mem_base);
2873 lop->args[2] = arg_ts->mem_offset;
2875 /* Loaded, but synced with memory. */
2876 arg_ts->state = TS_MEM;
2881 /* Perform input replacement, and mark inputs that became dead.
2882 No action is required except keeping temp_state up to date
2883 so that we reload when needed. */
2884 for (i = nb_oargs; i < nb_iargs + nb_oargs; i++) {
2885 arg_ts = arg_temp(op->args[i]);
2886 if (arg_ts) {
2887 dir_ts = arg_ts->state_ptr;
2888 if (dir_ts) {
2889 op->args[i] = temp_arg(dir_ts);
2890 changes = true;
2891 if (IS_DEAD_ARG(i)) {
2892 arg_ts->state = TS_DEAD;
2898 /* Liveness analysis should ensure that the following are
2899 all correct, for call sites and basic block end points. */
2900 if (call_flags & TCG_CALL_NO_READ_GLOBALS) {
2901 /* Nothing to do */
2902 } else if (call_flags & TCG_CALL_NO_WRITE_GLOBALS) {
2903 for (i = 0; i < nb_globals; ++i) {
2904 /* Liveness should see that globals are synced back,
2905 that is, either TS_DEAD or TS_MEM. */
2906 arg_ts = &s->temps[i];
2907 tcg_debug_assert(arg_ts->state_ptr == 0
2908 || arg_ts->state != 0);
2910 } else {
2911 for (i = 0; i < nb_globals; ++i) {
2912 /* Liveness should see that globals are saved back,
2913 that is, TS_DEAD, waiting to be reloaded. */
2914 arg_ts = &s->temps[i];
2915 tcg_debug_assert(arg_ts->state_ptr == 0
2916 || arg_ts->state == TS_DEAD);
2920 /* Outputs become available. */
2921 if (opc == INDEX_op_mov_i32 || opc == INDEX_op_mov_i64) {
2922 arg_ts = arg_temp(op->args[0]);
2923 dir_ts = arg_ts->state_ptr;
2924 if (dir_ts) {
2925 op->args[0] = temp_arg(dir_ts);
2926 changes = true;
2928 /* The output is now live and modified. */
2929 arg_ts->state = 0;
2931 if (NEED_SYNC_ARG(0)) {
2932 TCGOpcode sopc = (arg_ts->type == TCG_TYPE_I32
2933 ? INDEX_op_st_i32
2934 : INDEX_op_st_i64);
2935 TCGOp *sop = tcg_op_insert_after(s, op, sopc);
2936 TCGTemp *out_ts = dir_ts;
2938 if (IS_DEAD_ARG(0)) {
2939 out_ts = arg_temp(op->args[1]);
2940 arg_ts->state = TS_DEAD;
2941 tcg_op_remove(s, op);
2942 } else {
2943 arg_ts->state = TS_MEM;
2946 sop->args[0] = temp_arg(out_ts);
2947 sop->args[1] = temp_arg(arg_ts->mem_base);
2948 sop->args[2] = arg_ts->mem_offset;
2949 } else {
2950 tcg_debug_assert(!IS_DEAD_ARG(0));
2953 } else {
2954 for (i = 0; i < nb_oargs; i++) {
2955 arg_ts = arg_temp(op->args[i]);
2956 dir_ts = arg_ts->state_ptr;
2957 if (!dir_ts) {
2958 continue;
2960 op->args[i] = temp_arg(dir_ts);
2961 changes = true;
2963 /* The output is now live and modified. */
2964 arg_ts->state = 0;
2966 /* Sync outputs upon their last write. */
2967 if (NEED_SYNC_ARG(i)) {
2968 TCGOpcode sopc = (arg_ts->type == TCG_TYPE_I32
2969 ? INDEX_op_st_i32
2970 : INDEX_op_st_i64);
2971 TCGOp *sop = tcg_op_insert_after(s, op, sopc);
2973 sop->args[0] = temp_arg(dir_ts);
2974 sop->args[1] = temp_arg(arg_ts->mem_base);
2975 sop->args[2] = arg_ts->mem_offset;
2977 arg_ts->state = TS_MEM;
2979 /* Drop outputs that are dead. */
2980 if (IS_DEAD_ARG(i)) {
2981 arg_ts->state = TS_DEAD;
2987 return changes;
2990 #ifdef CONFIG_DEBUG_TCG
2991 static void dump_regs(TCGContext *s)
2993 TCGTemp *ts;
2994 int i;
2995 char buf[64];
2997 for(i = 0; i < s->nb_temps; i++) {
2998 ts = &s->temps[i];
2999 printf(" %10s: ", tcg_get_arg_str_ptr(s, buf, sizeof(buf), ts));
3000 switch(ts->val_type) {
3001 case TEMP_VAL_REG:
3002 printf("%s", tcg_target_reg_names[ts->reg]);
3003 break;
3004 case TEMP_VAL_MEM:
3005 printf("%d(%s)", (int)ts->mem_offset,
3006 tcg_target_reg_names[ts->mem_base->reg]);
3007 break;
3008 case TEMP_VAL_CONST:
3009 printf("$0x%" PRIx64, ts->val);
3010 break;
3011 case TEMP_VAL_DEAD:
3012 printf("D");
3013 break;
3014 default:
3015 printf("???");
3016 break;
3018 printf("\n");
3021 for(i = 0; i < TCG_TARGET_NB_REGS; i++) {
3022 if (s->reg_to_temp[i] != NULL) {
3023 printf("%s: %s\n",
3024 tcg_target_reg_names[i],
3025 tcg_get_arg_str_ptr(s, buf, sizeof(buf), s->reg_to_temp[i]));
3030 static void check_regs(TCGContext *s)
3032 int reg;
3033 int k;
3034 TCGTemp *ts;
3035 char buf[64];
3037 for (reg = 0; reg < TCG_TARGET_NB_REGS; reg++) {
3038 ts = s->reg_to_temp[reg];
3039 if (ts != NULL) {
3040 if (ts->val_type != TEMP_VAL_REG || ts->reg != reg) {
3041 printf("Inconsistency for register %s:\n",
3042 tcg_target_reg_names[reg]);
3043 goto fail;
3047 for (k = 0; k < s->nb_temps; k++) {
3048 ts = &s->temps[k];
3049 if (ts->val_type == TEMP_VAL_REG
3050 && ts->kind != TEMP_FIXED
3051 && s->reg_to_temp[ts->reg] != ts) {
3052 printf("Inconsistency for temp %s:\n",
3053 tcg_get_arg_str_ptr(s, buf, sizeof(buf), ts));
3054 fail:
3055 printf("reg state:\n");
3056 dump_regs(s);
3057 tcg_abort();
3061 #endif
3063 static void temp_allocate_frame(TCGContext *s, TCGTemp *ts)
3065 intptr_t off, size, align;
3067 switch (ts->type) {
3068 case TCG_TYPE_I32:
3069 size = align = 4;
3070 break;
3071 case TCG_TYPE_I64:
3072 case TCG_TYPE_V64:
3073 size = align = 8;
3074 break;
3075 case TCG_TYPE_V128:
3076 size = align = 16;
3077 break;
3078 case TCG_TYPE_V256:
3079 /* Note that we do not require aligned storage for V256. */
3080 size = 32, align = 16;
3081 break;
3082 default:
3083 g_assert_not_reached();
3087 * Assume the stack is sufficiently aligned.
3088 * This affects e.g. ARM NEON, where we have 8 byte stack alignment
3089 * and do not require 16 byte vector alignment. This seems slightly
3090 * easier than fully parameterizing the above switch statement.
3092 align = MIN(TCG_TARGET_STACK_ALIGN, align);
3093 off = ROUND_UP(s->current_frame_offset, align);
3095 /* If we've exhausted the stack frame, restart with a smaller TB. */
3096 if (off + size > s->frame_end) {
3097 tcg_raise_tb_overflow(s);
3099 s->current_frame_offset = off + size;
3101 ts->mem_offset = off;
3102 #if defined(__sparc__)
3103 ts->mem_offset += TCG_TARGET_STACK_BIAS;
3104 #endif
3105 ts->mem_base = s->frame_temp;
3106 ts->mem_allocated = 1;
3109 static void temp_load(TCGContext *, TCGTemp *, TCGRegSet, TCGRegSet, TCGRegSet);
3111 /* Mark a temporary as free or dead. If 'free_or_dead' is negative,
3112 mark it free; otherwise mark it dead. */
3113 static void temp_free_or_dead(TCGContext *s, TCGTemp *ts, int free_or_dead)
3115 TCGTempVal new_type;
3117 switch (ts->kind) {
3118 case TEMP_FIXED:
3119 return;
3120 case TEMP_GLOBAL:
3121 case TEMP_LOCAL:
3122 new_type = TEMP_VAL_MEM;
3123 break;
3124 case TEMP_NORMAL:
3125 case TEMP_EBB:
3126 new_type = free_or_dead < 0 ? TEMP_VAL_MEM : TEMP_VAL_DEAD;
3127 break;
3128 case TEMP_CONST:
3129 new_type = TEMP_VAL_CONST;
3130 break;
3131 default:
3132 g_assert_not_reached();
3134 if (ts->val_type == TEMP_VAL_REG) {
3135 s->reg_to_temp[ts->reg] = NULL;
3137 ts->val_type = new_type;
3140 /* Mark a temporary as dead. */
3141 static inline void temp_dead(TCGContext *s, TCGTemp *ts)
3143 temp_free_or_dead(s, ts, 1);
3146 /* Sync a temporary to memory. 'allocated_regs' is used in case a temporary
3147 registers needs to be allocated to store a constant. If 'free_or_dead'
3148 is non-zero, subsequently release the temporary; if it is positive, the
3149 temp is dead; if it is negative, the temp is free. */
3150 static void temp_sync(TCGContext *s, TCGTemp *ts, TCGRegSet allocated_regs,
3151 TCGRegSet preferred_regs, int free_or_dead)
3153 if (!temp_readonly(ts) && !ts->mem_coherent) {
3154 if (!ts->mem_allocated) {
3155 temp_allocate_frame(s, ts);
3157 switch (ts->val_type) {
3158 case TEMP_VAL_CONST:
3159 /* If we're going to free the temp immediately, then we won't
3160 require it later in a register, so attempt to store the
3161 constant to memory directly. */
3162 if (free_or_dead
3163 && tcg_out_sti(s, ts->type, ts->val,
3164 ts->mem_base->reg, ts->mem_offset)) {
3165 break;
3167 temp_load(s, ts, tcg_target_available_regs[ts->type],
3168 allocated_regs, preferred_regs);
3169 /* fallthrough */
3171 case TEMP_VAL_REG:
3172 tcg_out_st(s, ts->type, ts->reg,
3173 ts->mem_base->reg, ts->mem_offset);
3174 break;
3176 case TEMP_VAL_MEM:
3177 break;
3179 case TEMP_VAL_DEAD:
3180 default:
3181 tcg_abort();
3183 ts->mem_coherent = 1;
3185 if (free_or_dead) {
3186 temp_free_or_dead(s, ts, free_or_dead);
3190 /* free register 'reg' by spilling the corresponding temporary if necessary */
3191 static void tcg_reg_free(TCGContext *s, TCGReg reg, TCGRegSet allocated_regs)
3193 TCGTemp *ts = s->reg_to_temp[reg];
3194 if (ts != NULL) {
3195 temp_sync(s, ts, allocated_regs, 0, -1);
3200 * tcg_reg_alloc:
3201 * @required_regs: Set of registers in which we must allocate.
3202 * @allocated_regs: Set of registers which must be avoided.
3203 * @preferred_regs: Set of registers we should prefer.
3204 * @rev: True if we search the registers in "indirect" order.
3206 * The allocated register must be in @required_regs & ~@allocated_regs,
3207 * but if we can put it in @preferred_regs we may save a move later.
3209 static TCGReg tcg_reg_alloc(TCGContext *s, TCGRegSet required_regs,
3210 TCGRegSet allocated_regs,
3211 TCGRegSet preferred_regs, bool rev)
3213 int i, j, f, n = ARRAY_SIZE(tcg_target_reg_alloc_order);
3214 TCGRegSet reg_ct[2];
3215 const int *order;
3217 reg_ct[1] = required_regs & ~allocated_regs;
3218 tcg_debug_assert(reg_ct[1] != 0);
3219 reg_ct[0] = reg_ct[1] & preferred_regs;
3221 /* Skip the preferred_regs option if it cannot be satisfied,
3222 or if the preference made no difference. */
3223 f = reg_ct[0] == 0 || reg_ct[0] == reg_ct[1];
3225 order = rev ? indirect_reg_alloc_order : tcg_target_reg_alloc_order;
3227 /* Try free registers, preferences first. */
3228 for (j = f; j < 2; j++) {
3229 TCGRegSet set = reg_ct[j];
3231 if (tcg_regset_single(set)) {
3232 /* One register in the set. */
3233 TCGReg reg = tcg_regset_first(set);
3234 if (s->reg_to_temp[reg] == NULL) {
3235 return reg;
3237 } else {
3238 for (i = 0; i < n; i++) {
3239 TCGReg reg = order[i];
3240 if (s->reg_to_temp[reg] == NULL &&
3241 tcg_regset_test_reg(set, reg)) {
3242 return reg;
3248 /* We must spill something. */
3249 for (j = f; j < 2; j++) {
3250 TCGRegSet set = reg_ct[j];
3252 if (tcg_regset_single(set)) {
3253 /* One register in the set. */
3254 TCGReg reg = tcg_regset_first(set);
3255 tcg_reg_free(s, reg, allocated_regs);
3256 return reg;
3257 } else {
3258 for (i = 0; i < n; i++) {
3259 TCGReg reg = order[i];
3260 if (tcg_regset_test_reg(set, reg)) {
3261 tcg_reg_free(s, reg, allocated_regs);
3262 return reg;
3268 tcg_abort();
3271 /* Make sure the temporary is in a register. If needed, allocate the register
3272 from DESIRED while avoiding ALLOCATED. */
3273 static void temp_load(TCGContext *s, TCGTemp *ts, TCGRegSet desired_regs,
3274 TCGRegSet allocated_regs, TCGRegSet preferred_regs)
3276 TCGReg reg;
3278 switch (ts->val_type) {
3279 case TEMP_VAL_REG:
3280 return;
3281 case TEMP_VAL_CONST:
3282 reg = tcg_reg_alloc(s, desired_regs, allocated_regs,
3283 preferred_regs, ts->indirect_base);
3284 if (ts->type <= TCG_TYPE_I64) {
3285 tcg_out_movi(s, ts->type, reg, ts->val);
3286 } else {
3287 uint64_t val = ts->val;
3288 MemOp vece = MO_64;
3291 * Find the minimal vector element that matches the constant.
3292 * The targets will, in general, have to do this search anyway,
3293 * do this generically.
3295 if (val == dup_const(MO_8, val)) {
3296 vece = MO_8;
3297 } else if (val == dup_const(MO_16, val)) {
3298 vece = MO_16;
3299 } else if (val == dup_const(MO_32, val)) {
3300 vece = MO_32;
3303 tcg_out_dupi_vec(s, ts->type, vece, reg, ts->val);
3305 ts->mem_coherent = 0;
3306 break;
3307 case TEMP_VAL_MEM:
3308 reg = tcg_reg_alloc(s, desired_regs, allocated_regs,
3309 preferred_regs, ts->indirect_base);
3310 tcg_out_ld(s, ts->type, reg, ts->mem_base->reg, ts->mem_offset);
3311 ts->mem_coherent = 1;
3312 break;
3313 case TEMP_VAL_DEAD:
3314 default:
3315 tcg_abort();
3317 ts->reg = reg;
3318 ts->val_type = TEMP_VAL_REG;
3319 s->reg_to_temp[reg] = ts;
3322 /* Save a temporary to memory. 'allocated_regs' is used in case a
3323 temporary registers needs to be allocated to store a constant. */
3324 static void temp_save(TCGContext *s, TCGTemp *ts, TCGRegSet allocated_regs)
3326 /* The liveness analysis already ensures that globals are back
3327 in memory. Keep an tcg_debug_assert for safety. */
3328 tcg_debug_assert(ts->val_type == TEMP_VAL_MEM || temp_readonly(ts));
3331 /* save globals to their canonical location and assume they can be
3332 modified be the following code. 'allocated_regs' is used in case a
3333 temporary registers needs to be allocated to store a constant. */
3334 static void save_globals(TCGContext *s, TCGRegSet allocated_regs)
3336 int i, n;
3338 for (i = 0, n = s->nb_globals; i < n; i++) {
3339 temp_save(s, &s->temps[i], allocated_regs);
3343 /* sync globals to their canonical location and assume they can be
3344 read by the following code. 'allocated_regs' is used in case a
3345 temporary registers needs to be allocated to store a constant. */
3346 static void sync_globals(TCGContext *s, TCGRegSet allocated_regs)
3348 int i, n;
3350 for (i = 0, n = s->nb_globals; i < n; i++) {
3351 TCGTemp *ts = &s->temps[i];
3352 tcg_debug_assert(ts->val_type != TEMP_VAL_REG
3353 || ts->kind == TEMP_FIXED
3354 || ts->mem_coherent);
3358 /* at the end of a basic block, we assume all temporaries are dead and
3359 all globals are stored at their canonical location. */
3360 static void tcg_reg_alloc_bb_end(TCGContext *s, TCGRegSet allocated_regs)
3362 int i;
3364 for (i = s->nb_globals; i < s->nb_temps; i++) {
3365 TCGTemp *ts = &s->temps[i];
3367 switch (ts->kind) {
3368 case TEMP_LOCAL:
3369 temp_save(s, ts, allocated_regs);
3370 break;
3371 case TEMP_NORMAL:
3372 case TEMP_EBB:
3373 /* The liveness analysis already ensures that temps are dead.
3374 Keep an tcg_debug_assert for safety. */
3375 tcg_debug_assert(ts->val_type == TEMP_VAL_DEAD);
3376 break;
3377 case TEMP_CONST:
3378 /* Similarly, we should have freed any allocated register. */
3379 tcg_debug_assert(ts->val_type == TEMP_VAL_CONST);
3380 break;
3381 default:
3382 g_assert_not_reached();
3386 save_globals(s, allocated_regs);
3390 * At a conditional branch, we assume all temporaries are dead unless
3391 * explicitly live-across-conditional-branch; all globals and local
3392 * temps are synced to their location.
3394 static void tcg_reg_alloc_cbranch(TCGContext *s, TCGRegSet allocated_regs)
3396 sync_globals(s, allocated_regs);
3398 for (int i = s->nb_globals; i < s->nb_temps; i++) {
3399 TCGTemp *ts = &s->temps[i];
3401 * The liveness analysis already ensures that temps are dead.
3402 * Keep tcg_debug_asserts for safety.
3404 switch (ts->kind) {
3405 case TEMP_LOCAL:
3406 tcg_debug_assert(ts->val_type != TEMP_VAL_REG || ts->mem_coherent);
3407 break;
3408 case TEMP_NORMAL:
3409 tcg_debug_assert(ts->val_type == TEMP_VAL_DEAD);
3410 break;
3411 case TEMP_EBB:
3412 case TEMP_CONST:
3413 break;
3414 default:
3415 g_assert_not_reached();
3421 * Specialized code generation for INDEX_op_mov_* with a constant.
3423 static void tcg_reg_alloc_do_movi(TCGContext *s, TCGTemp *ots,
3424 tcg_target_ulong val, TCGLifeData arg_life,
3425 TCGRegSet preferred_regs)
3427 /* ENV should not be modified. */
3428 tcg_debug_assert(!temp_readonly(ots));
3430 /* The movi is not explicitly generated here. */
3431 if (ots->val_type == TEMP_VAL_REG) {
3432 s->reg_to_temp[ots->reg] = NULL;
3434 ots->val_type = TEMP_VAL_CONST;
3435 ots->val = val;
3436 ots->mem_coherent = 0;
3437 if (NEED_SYNC_ARG(0)) {
3438 temp_sync(s, ots, s->reserved_regs, preferred_regs, IS_DEAD_ARG(0));
3439 } else if (IS_DEAD_ARG(0)) {
3440 temp_dead(s, ots);
3445 * Specialized code generation for INDEX_op_mov_*.
3447 static void tcg_reg_alloc_mov(TCGContext *s, const TCGOp *op)
3449 const TCGLifeData arg_life = op->life;
3450 TCGRegSet allocated_regs, preferred_regs;
3451 TCGTemp *ts, *ots;
3452 TCGType otype, itype;
3454 allocated_regs = s->reserved_regs;
3455 preferred_regs = op->output_pref[0];
3456 ots = arg_temp(op->args[0]);
3457 ts = arg_temp(op->args[1]);
3459 /* ENV should not be modified. */
3460 tcg_debug_assert(!temp_readonly(ots));
3462 /* Note that otype != itype for no-op truncation. */
3463 otype = ots->type;
3464 itype = ts->type;
3466 if (ts->val_type == TEMP_VAL_CONST) {
3467 /* propagate constant or generate sti */
3468 tcg_target_ulong val = ts->val;
3469 if (IS_DEAD_ARG(1)) {
3470 temp_dead(s, ts);
3472 tcg_reg_alloc_do_movi(s, ots, val, arg_life, preferred_regs);
3473 return;
3476 /* If the source value is in memory we're going to be forced
3477 to have it in a register in order to perform the copy. Copy
3478 the SOURCE value into its own register first, that way we
3479 don't have to reload SOURCE the next time it is used. */
3480 if (ts->val_type == TEMP_VAL_MEM) {
3481 temp_load(s, ts, tcg_target_available_regs[itype],
3482 allocated_regs, preferred_regs);
3485 tcg_debug_assert(ts->val_type == TEMP_VAL_REG);
3486 if (IS_DEAD_ARG(0)) {
3487 /* mov to a non-saved dead register makes no sense (even with
3488 liveness analysis disabled). */
3489 tcg_debug_assert(NEED_SYNC_ARG(0));
3490 if (!ots->mem_allocated) {
3491 temp_allocate_frame(s, ots);
3493 tcg_out_st(s, otype, ts->reg, ots->mem_base->reg, ots->mem_offset);
3494 if (IS_DEAD_ARG(1)) {
3495 temp_dead(s, ts);
3497 temp_dead(s, ots);
3498 } else {
3499 if (IS_DEAD_ARG(1) && ts->kind != TEMP_FIXED) {
3500 /* the mov can be suppressed */
3501 if (ots->val_type == TEMP_VAL_REG) {
3502 s->reg_to_temp[ots->reg] = NULL;
3504 ots->reg = ts->reg;
3505 temp_dead(s, ts);
3506 } else {
3507 if (ots->val_type != TEMP_VAL_REG) {
3508 /* When allocating a new register, make sure to not spill the
3509 input one. */
3510 tcg_regset_set_reg(allocated_regs, ts->reg);
3511 ots->reg = tcg_reg_alloc(s, tcg_target_available_regs[otype],
3512 allocated_regs, preferred_regs,
3513 ots->indirect_base);
3515 if (!tcg_out_mov(s, otype, ots->reg, ts->reg)) {
3517 * Cross register class move not supported.
3518 * Store the source register into the destination slot
3519 * and leave the destination temp as TEMP_VAL_MEM.
3521 assert(!temp_readonly(ots));
3522 if (!ts->mem_allocated) {
3523 temp_allocate_frame(s, ots);
3525 tcg_out_st(s, ts->type, ts->reg,
3526 ots->mem_base->reg, ots->mem_offset);
3527 ots->mem_coherent = 1;
3528 temp_free_or_dead(s, ots, -1);
3529 return;
3532 ots->val_type = TEMP_VAL_REG;
3533 ots->mem_coherent = 0;
3534 s->reg_to_temp[ots->reg] = ots;
3535 if (NEED_SYNC_ARG(0)) {
3536 temp_sync(s, ots, allocated_regs, 0, 0);
3542 * Specialized code generation for INDEX_op_dup_vec.
3544 static void tcg_reg_alloc_dup(TCGContext *s, const TCGOp *op)
3546 const TCGLifeData arg_life = op->life;
3547 TCGRegSet dup_out_regs, dup_in_regs;
3548 TCGTemp *its, *ots;
3549 TCGType itype, vtype;
3550 intptr_t endian_fixup;
3551 unsigned vece;
3552 bool ok;
3554 ots = arg_temp(op->args[0]);
3555 its = arg_temp(op->args[1]);
3557 /* ENV should not be modified. */
3558 tcg_debug_assert(!temp_readonly(ots));
3560 itype = its->type;
3561 vece = TCGOP_VECE(op);
3562 vtype = TCGOP_VECL(op) + TCG_TYPE_V64;
3564 if (its->val_type == TEMP_VAL_CONST) {
3565 /* Propagate constant via movi -> dupi. */
3566 tcg_target_ulong val = its->val;
3567 if (IS_DEAD_ARG(1)) {
3568 temp_dead(s, its);
3570 tcg_reg_alloc_do_movi(s, ots, val, arg_life, op->output_pref[0]);
3571 return;
3574 dup_out_regs = tcg_op_defs[INDEX_op_dup_vec].args_ct[0].regs;
3575 dup_in_regs = tcg_op_defs[INDEX_op_dup_vec].args_ct[1].regs;
3577 /* Allocate the output register now. */
3578 if (ots->val_type != TEMP_VAL_REG) {
3579 TCGRegSet allocated_regs = s->reserved_regs;
3581 if (!IS_DEAD_ARG(1) && its->val_type == TEMP_VAL_REG) {
3582 /* Make sure to not spill the input register. */
3583 tcg_regset_set_reg(allocated_regs, its->reg);
3585 ots->reg = tcg_reg_alloc(s, dup_out_regs, allocated_regs,
3586 op->output_pref[0], ots->indirect_base);
3587 ots->val_type = TEMP_VAL_REG;
3588 ots->mem_coherent = 0;
3589 s->reg_to_temp[ots->reg] = ots;
3592 switch (its->val_type) {
3593 case TEMP_VAL_REG:
3595 * The dup constriaints must be broad, covering all possible VECE.
3596 * However, tcg_op_dup_vec() gets to see the VECE and we allow it
3597 * to fail, indicating that extra moves are required for that case.
3599 if (tcg_regset_test_reg(dup_in_regs, its->reg)) {
3600 if (tcg_out_dup_vec(s, vtype, vece, ots->reg, its->reg)) {
3601 goto done;
3603 /* Try again from memory or a vector input register. */
3605 if (!its->mem_coherent) {
3607 * The input register is not synced, and so an extra store
3608 * would be required to use memory. Attempt an integer-vector
3609 * register move first. We do not have a TCGRegSet for this.
3611 if (tcg_out_mov(s, itype, ots->reg, its->reg)) {
3612 break;
3614 /* Sync the temp back to its slot and load from there. */
3615 temp_sync(s, its, s->reserved_regs, 0, 0);
3617 /* fall through */
3619 case TEMP_VAL_MEM:
3620 #if HOST_BIG_ENDIAN
3621 endian_fixup = itype == TCG_TYPE_I32 ? 4 : 8;
3622 endian_fixup -= 1 << vece;
3623 #else
3624 endian_fixup = 0;
3625 #endif
3626 if (tcg_out_dupm_vec(s, vtype, vece, ots->reg, its->mem_base->reg,
3627 its->mem_offset + endian_fixup)) {
3628 goto done;
3630 tcg_out_ld(s, itype, ots->reg, its->mem_base->reg, its->mem_offset);
3631 break;
3633 default:
3634 g_assert_not_reached();
3637 /* We now have a vector input register, so dup must succeed. */
3638 ok = tcg_out_dup_vec(s, vtype, vece, ots->reg, ots->reg);
3639 tcg_debug_assert(ok);
3641 done:
3642 if (IS_DEAD_ARG(1)) {
3643 temp_dead(s, its);
3645 if (NEED_SYNC_ARG(0)) {
3646 temp_sync(s, ots, s->reserved_regs, 0, 0);
3648 if (IS_DEAD_ARG(0)) {
3649 temp_dead(s, ots);
3653 static void tcg_reg_alloc_op(TCGContext *s, const TCGOp *op)
3655 const TCGLifeData arg_life = op->life;
3656 const TCGOpDef * const def = &tcg_op_defs[op->opc];
3657 TCGRegSet i_allocated_regs;
3658 TCGRegSet o_allocated_regs;
3659 int i, k, nb_iargs, nb_oargs;
3660 TCGReg reg;
3661 TCGArg arg;
3662 const TCGArgConstraint *arg_ct;
3663 TCGTemp *ts;
3664 TCGArg new_args[TCG_MAX_OP_ARGS];
3665 int const_args[TCG_MAX_OP_ARGS];
3667 nb_oargs = def->nb_oargs;
3668 nb_iargs = def->nb_iargs;
3670 /* copy constants */
3671 memcpy(new_args + nb_oargs + nb_iargs,
3672 op->args + nb_oargs + nb_iargs,
3673 sizeof(TCGArg) * def->nb_cargs);
3675 i_allocated_regs = s->reserved_regs;
3676 o_allocated_regs = s->reserved_regs;
3678 /* satisfy input constraints */
3679 for (k = 0; k < nb_iargs; k++) {
3680 TCGRegSet i_preferred_regs, o_preferred_regs;
3682 i = def->args_ct[nb_oargs + k].sort_index;
3683 arg = op->args[i];
3684 arg_ct = &def->args_ct[i];
3685 ts = arg_temp(arg);
3687 if (ts->val_type == TEMP_VAL_CONST
3688 && tcg_target_const_match(ts->val, ts->type, arg_ct->ct)) {
3689 /* constant is OK for instruction */
3690 const_args[i] = 1;
3691 new_args[i] = ts->val;
3692 continue;
3695 i_preferred_regs = o_preferred_regs = 0;
3696 if (arg_ct->ialias) {
3697 o_preferred_regs = op->output_pref[arg_ct->alias_index];
3700 * If the input is readonly, then it cannot also be an
3701 * output and aliased to itself. If the input is not
3702 * dead after the instruction, we must allocate a new
3703 * register and move it.
3705 if (temp_readonly(ts) || !IS_DEAD_ARG(i)) {
3706 goto allocate_in_reg;
3710 * Check if the current register has already been allocated
3711 * for another input aliased to an output.
3713 if (ts->val_type == TEMP_VAL_REG) {
3714 reg = ts->reg;
3715 for (int k2 = 0; k2 < k; k2++) {
3716 int i2 = def->args_ct[nb_oargs + k2].sort_index;
3717 if (def->args_ct[i2].ialias && reg == new_args[i2]) {
3718 goto allocate_in_reg;
3722 i_preferred_regs = o_preferred_regs;
3725 temp_load(s, ts, arg_ct->regs, i_allocated_regs, i_preferred_regs);
3726 reg = ts->reg;
3728 if (!tcg_regset_test_reg(arg_ct->regs, reg)) {
3729 allocate_in_reg:
3731 * Allocate a new register matching the constraint
3732 * and move the temporary register into it.
3734 temp_load(s, ts, tcg_target_available_regs[ts->type],
3735 i_allocated_regs, 0);
3736 reg = tcg_reg_alloc(s, arg_ct->regs, i_allocated_regs,
3737 o_preferred_regs, ts->indirect_base);
3738 if (!tcg_out_mov(s, ts->type, reg, ts->reg)) {
3740 * Cross register class move not supported. Sync the
3741 * temp back to its slot and load from there.
3743 temp_sync(s, ts, i_allocated_regs, 0, 0);
3744 tcg_out_ld(s, ts->type, reg,
3745 ts->mem_base->reg, ts->mem_offset);
3748 new_args[i] = reg;
3749 const_args[i] = 0;
3750 tcg_regset_set_reg(i_allocated_regs, reg);
3753 /* mark dead temporaries and free the associated registers */
3754 for (i = nb_oargs; i < nb_oargs + nb_iargs; i++) {
3755 if (IS_DEAD_ARG(i)) {
3756 temp_dead(s, arg_temp(op->args[i]));
3760 if (def->flags & TCG_OPF_COND_BRANCH) {
3761 tcg_reg_alloc_cbranch(s, i_allocated_regs);
3762 } else if (def->flags & TCG_OPF_BB_END) {
3763 tcg_reg_alloc_bb_end(s, i_allocated_regs);
3764 } else {
3765 if (def->flags & TCG_OPF_CALL_CLOBBER) {
3766 /* XXX: permit generic clobber register list ? */
3767 for (i = 0; i < TCG_TARGET_NB_REGS; i++) {
3768 if (tcg_regset_test_reg(tcg_target_call_clobber_regs, i)) {
3769 tcg_reg_free(s, i, i_allocated_regs);
3773 if (def->flags & TCG_OPF_SIDE_EFFECTS) {
3774 /* sync globals if the op has side effects and might trigger
3775 an exception. */
3776 sync_globals(s, i_allocated_regs);
3779 /* satisfy the output constraints */
3780 for(k = 0; k < nb_oargs; k++) {
3781 i = def->args_ct[k].sort_index;
3782 arg = op->args[i];
3783 arg_ct = &def->args_ct[i];
3784 ts = arg_temp(arg);
3786 /* ENV should not be modified. */
3787 tcg_debug_assert(!temp_readonly(ts));
3789 if (arg_ct->oalias && !const_args[arg_ct->alias_index]) {
3790 reg = new_args[arg_ct->alias_index];
3791 } else if (arg_ct->newreg) {
3792 reg = tcg_reg_alloc(s, arg_ct->regs,
3793 i_allocated_regs | o_allocated_regs,
3794 op->output_pref[k], ts->indirect_base);
3795 } else {
3796 reg = tcg_reg_alloc(s, arg_ct->regs, o_allocated_regs,
3797 op->output_pref[k], ts->indirect_base);
3799 tcg_regset_set_reg(o_allocated_regs, reg);
3800 if (ts->val_type == TEMP_VAL_REG) {
3801 s->reg_to_temp[ts->reg] = NULL;
3803 ts->val_type = TEMP_VAL_REG;
3804 ts->reg = reg;
3806 * Temp value is modified, so the value kept in memory is
3807 * potentially not the same.
3809 ts->mem_coherent = 0;
3810 s->reg_to_temp[reg] = ts;
3811 new_args[i] = reg;
3815 /* emit instruction */
3816 if (def->flags & TCG_OPF_VECTOR) {
3817 tcg_out_vec_op(s, op->opc, TCGOP_VECL(op), TCGOP_VECE(op),
3818 new_args, const_args);
3819 } else {
3820 tcg_out_op(s, op->opc, new_args, const_args);
3823 /* move the outputs in the correct register if needed */
3824 for(i = 0; i < nb_oargs; i++) {
3825 ts = arg_temp(op->args[i]);
3827 /* ENV should not be modified. */
3828 tcg_debug_assert(!temp_readonly(ts));
3830 if (NEED_SYNC_ARG(i)) {
3831 temp_sync(s, ts, o_allocated_regs, 0, IS_DEAD_ARG(i));
3832 } else if (IS_DEAD_ARG(i)) {
3833 temp_dead(s, ts);
3838 static bool tcg_reg_alloc_dup2(TCGContext *s, const TCGOp *op)
3840 const TCGLifeData arg_life = op->life;
3841 TCGTemp *ots, *itsl, *itsh;
3842 TCGType vtype = TCGOP_VECL(op) + TCG_TYPE_V64;
3844 /* This opcode is only valid for 32-bit hosts, for 64-bit elements. */
3845 tcg_debug_assert(TCG_TARGET_REG_BITS == 32);
3846 tcg_debug_assert(TCGOP_VECE(op) == MO_64);
3848 ots = arg_temp(op->args[0]);
3849 itsl = arg_temp(op->args[1]);
3850 itsh = arg_temp(op->args[2]);
3852 /* ENV should not be modified. */
3853 tcg_debug_assert(!temp_readonly(ots));
3855 /* Allocate the output register now. */
3856 if (ots->val_type != TEMP_VAL_REG) {
3857 TCGRegSet allocated_regs = s->reserved_regs;
3858 TCGRegSet dup_out_regs =
3859 tcg_op_defs[INDEX_op_dup_vec].args_ct[0].regs;
3861 /* Make sure to not spill the input registers. */
3862 if (!IS_DEAD_ARG(1) && itsl->val_type == TEMP_VAL_REG) {
3863 tcg_regset_set_reg(allocated_regs, itsl->reg);
3865 if (!IS_DEAD_ARG(2) && itsh->val_type == TEMP_VAL_REG) {
3866 tcg_regset_set_reg(allocated_regs, itsh->reg);
3869 ots->reg = tcg_reg_alloc(s, dup_out_regs, allocated_regs,
3870 op->output_pref[0], ots->indirect_base);
3871 ots->val_type = TEMP_VAL_REG;
3872 ots->mem_coherent = 0;
3873 s->reg_to_temp[ots->reg] = ots;
3876 /* Promote dup2 of immediates to dupi_vec. */
3877 if (itsl->val_type == TEMP_VAL_CONST && itsh->val_type == TEMP_VAL_CONST) {
3878 uint64_t val = deposit64(itsl->val, 32, 32, itsh->val);
3879 MemOp vece = MO_64;
3881 if (val == dup_const(MO_8, val)) {
3882 vece = MO_8;
3883 } else if (val == dup_const(MO_16, val)) {
3884 vece = MO_16;
3885 } else if (val == dup_const(MO_32, val)) {
3886 vece = MO_32;
3889 tcg_out_dupi_vec(s, vtype, vece, ots->reg, val);
3890 goto done;
3893 /* If the two inputs form one 64-bit value, try dupm_vec. */
3894 if (itsl + 1 == itsh && itsl->base_type == TCG_TYPE_I64) {
3895 if (!itsl->mem_coherent) {
3896 temp_sync(s, itsl, s->reserved_regs, 0, 0);
3898 if (!itsh->mem_coherent) {
3899 temp_sync(s, itsh, s->reserved_regs, 0, 0);
3901 #if HOST_BIG_ENDIAN
3902 TCGTemp *its = itsh;
3903 #else
3904 TCGTemp *its = itsl;
3905 #endif
3906 if (tcg_out_dupm_vec(s, vtype, MO_64, ots->reg,
3907 its->mem_base->reg, its->mem_offset)) {
3908 goto done;
3912 /* Fall back to generic expansion. */
3913 return false;
3915 done:
3916 if (IS_DEAD_ARG(1)) {
3917 temp_dead(s, itsl);
3919 if (IS_DEAD_ARG(2)) {
3920 temp_dead(s, itsh);
3922 if (NEED_SYNC_ARG(0)) {
3923 temp_sync(s, ots, s->reserved_regs, 0, IS_DEAD_ARG(0));
3924 } else if (IS_DEAD_ARG(0)) {
3925 temp_dead(s, ots);
3927 return true;
3930 #ifdef TCG_TARGET_STACK_GROWSUP
3931 #define STACK_DIR(x) (-(x))
3932 #else
3933 #define STACK_DIR(x) (x)
3934 #endif
3936 static void tcg_reg_alloc_call(TCGContext *s, TCGOp *op)
3938 const int nb_oargs = TCGOP_CALLO(op);
3939 const int nb_iargs = TCGOP_CALLI(op);
3940 const TCGLifeData arg_life = op->life;
3941 const TCGHelperInfo *info;
3942 int flags, nb_regs, i;
3943 TCGReg reg;
3944 TCGArg arg;
3945 TCGTemp *ts;
3946 intptr_t stack_offset;
3947 size_t call_stack_size;
3948 tcg_insn_unit *func_addr;
3949 int allocate_args;
3950 TCGRegSet allocated_regs;
3952 func_addr = tcg_call_func(op);
3953 info = tcg_call_info(op);
3954 flags = info->flags;
3956 nb_regs = ARRAY_SIZE(tcg_target_call_iarg_regs);
3957 if (nb_regs > nb_iargs) {
3958 nb_regs = nb_iargs;
3961 /* assign stack slots first */
3962 call_stack_size = (nb_iargs - nb_regs) * sizeof(tcg_target_long);
3963 call_stack_size = (call_stack_size + TCG_TARGET_STACK_ALIGN - 1) &
3964 ~(TCG_TARGET_STACK_ALIGN - 1);
3965 allocate_args = (call_stack_size > TCG_STATIC_CALL_ARGS_SIZE);
3966 if (allocate_args) {
3967 /* XXX: if more than TCG_STATIC_CALL_ARGS_SIZE is needed,
3968 preallocate call stack */
3969 tcg_abort();
3972 stack_offset = TCG_TARGET_CALL_STACK_OFFSET;
3973 for (i = nb_regs; i < nb_iargs; i++) {
3974 arg = op->args[nb_oargs + i];
3975 #ifdef TCG_TARGET_STACK_GROWSUP
3976 stack_offset -= sizeof(tcg_target_long);
3977 #endif
3978 if (arg != TCG_CALL_DUMMY_ARG) {
3979 ts = arg_temp(arg);
3980 temp_load(s, ts, tcg_target_available_regs[ts->type],
3981 s->reserved_regs, 0);
3982 tcg_out_st(s, ts->type, ts->reg, TCG_REG_CALL_STACK, stack_offset);
3984 #ifndef TCG_TARGET_STACK_GROWSUP
3985 stack_offset += sizeof(tcg_target_long);
3986 #endif
3989 /* assign input registers */
3990 allocated_regs = s->reserved_regs;
3991 for (i = 0; i < nb_regs; i++) {
3992 arg = op->args[nb_oargs + i];
3993 if (arg != TCG_CALL_DUMMY_ARG) {
3994 ts = arg_temp(arg);
3995 reg = tcg_target_call_iarg_regs[i];
3997 if (ts->val_type == TEMP_VAL_REG) {
3998 if (ts->reg != reg) {
3999 tcg_reg_free(s, reg, allocated_regs);
4000 if (!tcg_out_mov(s, ts->type, reg, ts->reg)) {
4002 * Cross register class move not supported. Sync the
4003 * temp back to its slot and load from there.
4005 temp_sync(s, ts, allocated_regs, 0, 0);
4006 tcg_out_ld(s, ts->type, reg,
4007 ts->mem_base->reg, ts->mem_offset);
4010 } else {
4011 TCGRegSet arg_set = 0;
4013 tcg_reg_free(s, reg, allocated_regs);
4014 tcg_regset_set_reg(arg_set, reg);
4015 temp_load(s, ts, arg_set, allocated_regs, 0);
4018 tcg_regset_set_reg(allocated_regs, reg);
4022 /* mark dead temporaries and free the associated registers */
4023 for (i = nb_oargs; i < nb_iargs + nb_oargs; i++) {
4024 if (IS_DEAD_ARG(i)) {
4025 temp_dead(s, arg_temp(op->args[i]));
4029 /* clobber call registers */
4030 for (i = 0; i < TCG_TARGET_NB_REGS; i++) {
4031 if (tcg_regset_test_reg(tcg_target_call_clobber_regs, i)) {
4032 tcg_reg_free(s, i, allocated_regs);
4036 /* Save globals if they might be written by the helper, sync them if
4037 they might be read. */
4038 if (flags & TCG_CALL_NO_READ_GLOBALS) {
4039 /* Nothing to do */
4040 } else if (flags & TCG_CALL_NO_WRITE_GLOBALS) {
4041 sync_globals(s, allocated_regs);
4042 } else {
4043 save_globals(s, allocated_regs);
4046 #ifdef CONFIG_TCG_INTERPRETER
4048 gpointer hash = (gpointer)(uintptr_t)info->typemask;
4049 ffi_cif *cif = g_hash_table_lookup(ffi_table, hash);
4050 assert(cif != NULL);
4051 tcg_out_call(s, func_addr, cif);
4053 #else
4054 tcg_out_call(s, func_addr);
4055 #endif
4057 /* assign output registers and emit moves if needed */
4058 for(i = 0; i < nb_oargs; i++) {
4059 arg = op->args[i];
4060 ts = arg_temp(arg);
4062 /* ENV should not be modified. */
4063 tcg_debug_assert(!temp_readonly(ts));
4065 reg = tcg_target_call_oarg_regs[i];
4066 tcg_debug_assert(s->reg_to_temp[reg] == NULL);
4067 if (ts->val_type == TEMP_VAL_REG) {
4068 s->reg_to_temp[ts->reg] = NULL;
4070 ts->val_type = TEMP_VAL_REG;
4071 ts->reg = reg;
4072 ts->mem_coherent = 0;
4073 s->reg_to_temp[reg] = ts;
4074 if (NEED_SYNC_ARG(i)) {
4075 temp_sync(s, ts, allocated_regs, 0, IS_DEAD_ARG(i));
4076 } else if (IS_DEAD_ARG(i)) {
4077 temp_dead(s, ts);
4082 #ifdef CONFIG_PROFILER
4084 /* avoid copy/paste errors */
4085 #define PROF_ADD(to, from, field) \
4086 do { \
4087 (to)->field += qatomic_read(&((from)->field)); \
4088 } while (0)
4090 #define PROF_MAX(to, from, field) \
4091 do { \
4092 typeof((from)->field) val__ = qatomic_read(&((from)->field)); \
4093 if (val__ > (to)->field) { \
4094 (to)->field = val__; \
4096 } while (0)
4098 /* Pass in a zero'ed @prof */
4099 static inline
4100 void tcg_profile_snapshot(TCGProfile *prof, bool counters, bool table)
4102 unsigned int n_ctxs = qatomic_read(&tcg_cur_ctxs);
4103 unsigned int i;
4105 for (i = 0; i < n_ctxs; i++) {
4106 TCGContext *s = qatomic_read(&tcg_ctxs[i]);
4107 const TCGProfile *orig = &s->prof;
4109 if (counters) {
4110 PROF_ADD(prof, orig, cpu_exec_time);
4111 PROF_ADD(prof, orig, tb_count1);
4112 PROF_ADD(prof, orig, tb_count);
4113 PROF_ADD(prof, orig, op_count);
4114 PROF_MAX(prof, orig, op_count_max);
4115 PROF_ADD(prof, orig, temp_count);
4116 PROF_MAX(prof, orig, temp_count_max);
4117 PROF_ADD(prof, orig, del_op_count);
4118 PROF_ADD(prof, orig, code_in_len);
4119 PROF_ADD(prof, orig, code_out_len);
4120 PROF_ADD(prof, orig, search_out_len);
4121 PROF_ADD(prof, orig, interm_time);
4122 PROF_ADD(prof, orig, code_time);
4123 PROF_ADD(prof, orig, la_time);
4124 PROF_ADD(prof, orig, opt_time);
4125 PROF_ADD(prof, orig, restore_count);
4126 PROF_ADD(prof, orig, restore_time);
4128 if (table) {
4129 int i;
4131 for (i = 0; i < NB_OPS; i++) {
4132 PROF_ADD(prof, orig, table_op_count[i]);
4138 #undef PROF_ADD
4139 #undef PROF_MAX
4141 static void tcg_profile_snapshot_counters(TCGProfile *prof)
4143 tcg_profile_snapshot(prof, true, false);
4146 static void tcg_profile_snapshot_table(TCGProfile *prof)
4148 tcg_profile_snapshot(prof, false, true);
4151 void tcg_dump_op_count(GString *buf)
4153 TCGProfile prof = {};
4154 int i;
4156 tcg_profile_snapshot_table(&prof);
4157 for (i = 0; i < NB_OPS; i++) {
4158 g_string_append_printf(buf, "%s %" PRId64 "\n", tcg_op_defs[i].name,
4159 prof.table_op_count[i]);
4163 int64_t tcg_cpu_exec_time(void)
4165 unsigned int n_ctxs = qatomic_read(&tcg_cur_ctxs);
4166 unsigned int i;
4167 int64_t ret = 0;
4169 for (i = 0; i < n_ctxs; i++) {
4170 const TCGContext *s = qatomic_read(&tcg_ctxs[i]);
4171 const TCGProfile *prof = &s->prof;
4173 ret += qatomic_read(&prof->cpu_exec_time);
4175 return ret;
4177 #else
4178 void tcg_dump_op_count(GString *buf)
4180 g_string_append_printf(buf, "[TCG profiler not compiled]\n");
4183 int64_t tcg_cpu_exec_time(void)
4185 error_report("%s: TCG profiler not compiled", __func__);
4186 exit(EXIT_FAILURE);
4188 #endif
4191 int tcg_gen_code(TCGContext *s, TranslationBlock *tb, target_ulong pc_start)
4193 #ifdef CONFIG_PROFILER
4194 TCGProfile *prof = &s->prof;
4195 #endif
4196 int i, num_insns;
4197 TCGOp *op;
4199 #ifdef CONFIG_PROFILER
4201 int n = 0;
4203 QTAILQ_FOREACH(op, &s->ops, link) {
4204 n++;
4206 qatomic_set(&prof->op_count, prof->op_count + n);
4207 if (n > prof->op_count_max) {
4208 qatomic_set(&prof->op_count_max, n);
4211 n = s->nb_temps;
4212 qatomic_set(&prof->temp_count, prof->temp_count + n);
4213 if (n > prof->temp_count_max) {
4214 qatomic_set(&prof->temp_count_max, n);
4217 #endif
4219 #ifdef DEBUG_DISAS
4220 if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP)
4221 && qemu_log_in_addr_range(pc_start))) {
4222 FILE *logfile = qemu_log_trylock();
4223 if (logfile) {
4224 fprintf(logfile, "OP:\n");
4225 tcg_dump_ops(s, logfile, false);
4226 fprintf(logfile, "\n");
4227 qemu_log_unlock(logfile);
4230 #endif
4232 #ifdef CONFIG_DEBUG_TCG
4233 /* Ensure all labels referenced have been emitted. */
4235 TCGLabel *l;
4236 bool error = false;
4238 QSIMPLEQ_FOREACH(l, &s->labels, next) {
4239 if (unlikely(!l->present) && l->refs) {
4240 qemu_log_mask(CPU_LOG_TB_OP,
4241 "$L%d referenced but not present.\n", l->id);
4242 error = true;
4245 assert(!error);
4247 #endif
4249 #ifdef CONFIG_PROFILER
4250 qatomic_set(&prof->opt_time, prof->opt_time - profile_getclock());
4251 #endif
4253 #ifdef USE_TCG_OPTIMIZATIONS
4254 tcg_optimize(s);
4255 #endif
4257 #ifdef CONFIG_PROFILER
4258 qatomic_set(&prof->opt_time, prof->opt_time + profile_getclock());
4259 qatomic_set(&prof->la_time, prof->la_time - profile_getclock());
4260 #endif
4262 reachable_code_pass(s);
4263 liveness_pass_1(s);
4265 if (s->nb_indirects > 0) {
4266 #ifdef DEBUG_DISAS
4267 if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP_IND)
4268 && qemu_log_in_addr_range(pc_start))) {
4269 FILE *logfile = qemu_log_trylock();
4270 if (logfile) {
4271 fprintf(logfile, "OP before indirect lowering:\n");
4272 tcg_dump_ops(s, logfile, false);
4273 fprintf(logfile, "\n");
4274 qemu_log_unlock(logfile);
4277 #endif
4278 /* Replace indirect temps with direct temps. */
4279 if (liveness_pass_2(s)) {
4280 /* If changes were made, re-run liveness. */
4281 liveness_pass_1(s);
4285 #ifdef CONFIG_PROFILER
4286 qatomic_set(&prof->la_time, prof->la_time + profile_getclock());
4287 #endif
4289 #ifdef DEBUG_DISAS
4290 if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP_OPT)
4291 && qemu_log_in_addr_range(pc_start))) {
4292 FILE *logfile = qemu_log_trylock();
4293 if (logfile) {
4294 fprintf(logfile, "OP after optimization and liveness analysis:\n");
4295 tcg_dump_ops(s, logfile, true);
4296 fprintf(logfile, "\n");
4297 qemu_log_unlock(logfile);
4300 #endif
4302 tcg_reg_alloc_start(s);
4305 * Reset the buffer pointers when restarting after overflow.
4306 * TODO: Move this into translate-all.c with the rest of the
4307 * buffer management. Having only this done here is confusing.
4309 s->code_buf = tcg_splitwx_to_rw(tb->tc.ptr);
4310 s->code_ptr = s->code_buf;
4312 #ifdef TCG_TARGET_NEED_LDST_LABELS
4313 QSIMPLEQ_INIT(&s->ldst_labels);
4314 #endif
4315 #ifdef TCG_TARGET_NEED_POOL_LABELS
4316 s->pool_labels = NULL;
4317 #endif
4319 num_insns = -1;
4320 QTAILQ_FOREACH(op, &s->ops, link) {
4321 TCGOpcode opc = op->opc;
4323 #ifdef CONFIG_PROFILER
4324 qatomic_set(&prof->table_op_count[opc], prof->table_op_count[opc] + 1);
4325 #endif
4327 switch (opc) {
4328 case INDEX_op_mov_i32:
4329 case INDEX_op_mov_i64:
4330 case INDEX_op_mov_vec:
4331 tcg_reg_alloc_mov(s, op);
4332 break;
4333 case INDEX_op_dup_vec:
4334 tcg_reg_alloc_dup(s, op);
4335 break;
4336 case INDEX_op_insn_start:
4337 if (num_insns >= 0) {
4338 size_t off = tcg_current_code_size(s);
4339 s->gen_insn_end_off[num_insns] = off;
4340 /* Assert that we do not overflow our stored offset. */
4341 assert(s->gen_insn_end_off[num_insns] == off);
4343 num_insns++;
4344 for (i = 0; i < TARGET_INSN_START_WORDS; ++i) {
4345 target_ulong a;
4346 #if TARGET_LONG_BITS > TCG_TARGET_REG_BITS
4347 a = deposit64(op->args[i * 2], 32, 32, op->args[i * 2 + 1]);
4348 #else
4349 a = op->args[i];
4350 #endif
4351 s->gen_insn_data[num_insns][i] = a;
4353 break;
4354 case INDEX_op_discard:
4355 temp_dead(s, arg_temp(op->args[0]));
4356 break;
4357 case INDEX_op_set_label:
4358 tcg_reg_alloc_bb_end(s, s->reserved_regs);
4359 tcg_out_label(s, arg_label(op->args[0]));
4360 break;
4361 case INDEX_op_call:
4362 tcg_reg_alloc_call(s, op);
4363 break;
4364 case INDEX_op_dup2_vec:
4365 if (tcg_reg_alloc_dup2(s, op)) {
4366 break;
4368 /* fall through */
4369 default:
4370 /* Sanity check that we've not introduced any unhandled opcodes. */
4371 tcg_debug_assert(tcg_op_supported(opc));
4372 /* Note: in order to speed up the code, it would be much
4373 faster to have specialized register allocator functions for
4374 some common argument patterns */
4375 tcg_reg_alloc_op(s, op);
4376 break;
4378 #ifdef CONFIG_DEBUG_TCG
4379 check_regs(s);
4380 #endif
4381 /* Test for (pending) buffer overflow. The assumption is that any
4382 one operation beginning below the high water mark cannot overrun
4383 the buffer completely. Thus we can test for overflow after
4384 generating code without having to check during generation. */
4385 if (unlikely((void *)s->code_ptr > s->code_gen_highwater)) {
4386 return -1;
4388 /* Test for TB overflow, as seen by gen_insn_end_off. */
4389 if (unlikely(tcg_current_code_size(s) > UINT16_MAX)) {
4390 return -2;
4393 tcg_debug_assert(num_insns >= 0);
4394 s->gen_insn_end_off[num_insns] = tcg_current_code_size(s);
4396 /* Generate TB finalization at the end of block */
4397 #ifdef TCG_TARGET_NEED_LDST_LABELS
4398 i = tcg_out_ldst_finalize(s);
4399 if (i < 0) {
4400 return i;
4402 #endif
4403 #ifdef TCG_TARGET_NEED_POOL_LABELS
4404 i = tcg_out_pool_finalize(s);
4405 if (i < 0) {
4406 return i;
4408 #endif
4409 if (!tcg_resolve_relocs(s)) {
4410 return -2;
4413 #ifndef CONFIG_TCG_INTERPRETER
4414 /* flush instruction cache */
4415 flush_idcache_range((uintptr_t)tcg_splitwx_to_rx(s->code_buf),
4416 (uintptr_t)s->code_buf,
4417 tcg_ptr_byte_diff(s->code_ptr, s->code_buf));
4418 #endif
4420 return tcg_current_code_size(s);
4423 #ifdef CONFIG_PROFILER
4424 void tcg_dump_info(GString *buf)
4426 TCGProfile prof = {};
4427 const TCGProfile *s;
4428 int64_t tb_count;
4429 int64_t tb_div_count;
4430 int64_t tot;
4432 tcg_profile_snapshot_counters(&prof);
4433 s = &prof;
4434 tb_count = s->tb_count;
4435 tb_div_count = tb_count ? tb_count : 1;
4436 tot = s->interm_time + s->code_time;
4438 g_string_append_printf(buf, "JIT cycles %" PRId64
4439 " (%0.3f s at 2.4 GHz)\n",
4440 tot, tot / 2.4e9);
4441 g_string_append_printf(buf, "translated TBs %" PRId64
4442 " (aborted=%" PRId64 " %0.1f%%)\n",
4443 tb_count, s->tb_count1 - tb_count,
4444 (double)(s->tb_count1 - s->tb_count)
4445 / (s->tb_count1 ? s->tb_count1 : 1) * 100.0);
4446 g_string_append_printf(buf, "avg ops/TB %0.1f max=%d\n",
4447 (double)s->op_count / tb_div_count, s->op_count_max);
4448 g_string_append_printf(buf, "deleted ops/TB %0.2f\n",
4449 (double)s->del_op_count / tb_div_count);
4450 g_string_append_printf(buf, "avg temps/TB %0.2f max=%d\n",
4451 (double)s->temp_count / tb_div_count,
4452 s->temp_count_max);
4453 g_string_append_printf(buf, "avg host code/TB %0.1f\n",
4454 (double)s->code_out_len / tb_div_count);
4455 g_string_append_printf(buf, "avg search data/TB %0.1f\n",
4456 (double)s->search_out_len / tb_div_count);
4458 g_string_append_printf(buf, "cycles/op %0.1f\n",
4459 s->op_count ? (double)tot / s->op_count : 0);
4460 g_string_append_printf(buf, "cycles/in byte %0.1f\n",
4461 s->code_in_len ? (double)tot / s->code_in_len : 0);
4462 g_string_append_printf(buf, "cycles/out byte %0.1f\n",
4463 s->code_out_len ? (double)tot / s->code_out_len : 0);
4464 g_string_append_printf(buf, "cycles/search byte %0.1f\n",
4465 s->search_out_len ?
4466 (double)tot / s->search_out_len : 0);
4467 if (tot == 0) {
4468 tot = 1;
4470 g_string_append_printf(buf, " gen_interm time %0.1f%%\n",
4471 (double)s->interm_time / tot * 100.0);
4472 g_string_append_printf(buf, " gen_code time %0.1f%%\n",
4473 (double)s->code_time / tot * 100.0);
4474 g_string_append_printf(buf, "optim./code time %0.1f%%\n",
4475 (double)s->opt_time / (s->code_time ?
4476 s->code_time : 1)
4477 * 100.0);
4478 g_string_append_printf(buf, "liveness/code time %0.1f%%\n",
4479 (double)s->la_time / (s->code_time ?
4480 s->code_time : 1) * 100.0);
4481 g_string_append_printf(buf, "cpu_restore count %" PRId64 "\n",
4482 s->restore_count);
4483 g_string_append_printf(buf, " avg cycles %0.1f\n",
4484 s->restore_count ?
4485 (double)s->restore_time / s->restore_count : 0);
4487 #else
4488 void tcg_dump_info(GString *buf)
4490 g_string_append_printf(buf, "[TCG profiler not compiled]\n");
4492 #endif
4494 #ifdef ELF_HOST_MACHINE
4495 /* In order to use this feature, the backend needs to do three things:
4497 (1) Define ELF_HOST_MACHINE to indicate both what value to
4498 put into the ELF image and to indicate support for the feature.
4500 (2) Define tcg_register_jit. This should create a buffer containing
4501 the contents of a .debug_frame section that describes the post-
4502 prologue unwind info for the tcg machine.
4504 (3) Call tcg_register_jit_int, with the constructed .debug_frame.
4507 /* Begin GDB interface. THE FOLLOWING MUST MATCH GDB DOCS. */
4508 typedef enum {
4509 JIT_NOACTION = 0,
4510 JIT_REGISTER_FN,
4511 JIT_UNREGISTER_FN
4512 } jit_actions_t;
4514 struct jit_code_entry {
4515 struct jit_code_entry *next_entry;
4516 struct jit_code_entry *prev_entry;
4517 const void *symfile_addr;
4518 uint64_t symfile_size;
4521 struct jit_descriptor {
4522 uint32_t version;
4523 uint32_t action_flag;
4524 struct jit_code_entry *relevant_entry;
4525 struct jit_code_entry *first_entry;
4528 void __jit_debug_register_code(void) __attribute__((noinline));
4529 void __jit_debug_register_code(void)
4531 asm("");
4534 /* Must statically initialize the version, because GDB may check
4535 the version before we can set it. */
4536 struct jit_descriptor __jit_debug_descriptor = { 1, 0, 0, 0 };
4538 /* End GDB interface. */
4540 static int find_string(const char *strtab, const char *str)
4542 const char *p = strtab + 1;
4544 while (1) {
4545 if (strcmp(p, str) == 0) {
4546 return p - strtab;
4548 p += strlen(p) + 1;
4552 static void tcg_register_jit_int(const void *buf_ptr, size_t buf_size,
4553 const void *debug_frame,
4554 size_t debug_frame_size)
4556 struct __attribute__((packed)) DebugInfo {
4557 uint32_t len;
4558 uint16_t version;
4559 uint32_t abbrev;
4560 uint8_t ptr_size;
4561 uint8_t cu_die;
4562 uint16_t cu_lang;
4563 uintptr_t cu_low_pc;
4564 uintptr_t cu_high_pc;
4565 uint8_t fn_die;
4566 char fn_name[16];
4567 uintptr_t fn_low_pc;
4568 uintptr_t fn_high_pc;
4569 uint8_t cu_eoc;
4572 struct ElfImage {
4573 ElfW(Ehdr) ehdr;
4574 ElfW(Phdr) phdr;
4575 ElfW(Shdr) shdr[7];
4576 ElfW(Sym) sym[2];
4577 struct DebugInfo di;
4578 uint8_t da[24];
4579 char str[80];
4582 struct ElfImage *img;
4584 static const struct ElfImage img_template = {
4585 .ehdr = {
4586 .e_ident[EI_MAG0] = ELFMAG0,
4587 .e_ident[EI_MAG1] = ELFMAG1,
4588 .e_ident[EI_MAG2] = ELFMAG2,
4589 .e_ident[EI_MAG3] = ELFMAG3,
4590 .e_ident[EI_CLASS] = ELF_CLASS,
4591 .e_ident[EI_DATA] = ELF_DATA,
4592 .e_ident[EI_VERSION] = EV_CURRENT,
4593 .e_type = ET_EXEC,
4594 .e_machine = ELF_HOST_MACHINE,
4595 .e_version = EV_CURRENT,
4596 .e_phoff = offsetof(struct ElfImage, phdr),
4597 .e_shoff = offsetof(struct ElfImage, shdr),
4598 .e_ehsize = sizeof(ElfW(Shdr)),
4599 .e_phentsize = sizeof(ElfW(Phdr)),
4600 .e_phnum = 1,
4601 .e_shentsize = sizeof(ElfW(Shdr)),
4602 .e_shnum = ARRAY_SIZE(img->shdr),
4603 .e_shstrndx = ARRAY_SIZE(img->shdr) - 1,
4604 #ifdef ELF_HOST_FLAGS
4605 .e_flags = ELF_HOST_FLAGS,
4606 #endif
4607 #ifdef ELF_OSABI
4608 .e_ident[EI_OSABI] = ELF_OSABI,
4609 #endif
4611 .phdr = {
4612 .p_type = PT_LOAD,
4613 .p_flags = PF_X,
4615 .shdr = {
4616 [0] = { .sh_type = SHT_NULL },
4617 /* Trick: The contents of code_gen_buffer are not present in
4618 this fake ELF file; that got allocated elsewhere. Therefore
4619 we mark .text as SHT_NOBITS (similar to .bss) so that readers
4620 will not look for contents. We can record any address. */
4621 [1] = { /* .text */
4622 .sh_type = SHT_NOBITS,
4623 .sh_flags = SHF_EXECINSTR | SHF_ALLOC,
4625 [2] = { /* .debug_info */
4626 .sh_type = SHT_PROGBITS,
4627 .sh_offset = offsetof(struct ElfImage, di),
4628 .sh_size = sizeof(struct DebugInfo),
4630 [3] = { /* .debug_abbrev */
4631 .sh_type = SHT_PROGBITS,
4632 .sh_offset = offsetof(struct ElfImage, da),
4633 .sh_size = sizeof(img->da),
4635 [4] = { /* .debug_frame */
4636 .sh_type = SHT_PROGBITS,
4637 .sh_offset = sizeof(struct ElfImage),
4639 [5] = { /* .symtab */
4640 .sh_type = SHT_SYMTAB,
4641 .sh_offset = offsetof(struct ElfImage, sym),
4642 .sh_size = sizeof(img->sym),
4643 .sh_info = 1,
4644 .sh_link = ARRAY_SIZE(img->shdr) - 1,
4645 .sh_entsize = sizeof(ElfW(Sym)),
4647 [6] = { /* .strtab */
4648 .sh_type = SHT_STRTAB,
4649 .sh_offset = offsetof(struct ElfImage, str),
4650 .sh_size = sizeof(img->str),
4653 .sym = {
4654 [1] = { /* code_gen_buffer */
4655 .st_info = ELF_ST_INFO(STB_GLOBAL, STT_FUNC),
4656 .st_shndx = 1,
4659 .di = {
4660 .len = sizeof(struct DebugInfo) - 4,
4661 .version = 2,
4662 .ptr_size = sizeof(void *),
4663 .cu_die = 1,
4664 .cu_lang = 0x8001, /* DW_LANG_Mips_Assembler */
4665 .fn_die = 2,
4666 .fn_name = "code_gen_buffer"
4668 .da = {
4669 1, /* abbrev number (the cu) */
4670 0x11, 1, /* DW_TAG_compile_unit, has children */
4671 0x13, 0x5, /* DW_AT_language, DW_FORM_data2 */
4672 0x11, 0x1, /* DW_AT_low_pc, DW_FORM_addr */
4673 0x12, 0x1, /* DW_AT_high_pc, DW_FORM_addr */
4674 0, 0, /* end of abbrev */
4675 2, /* abbrev number (the fn) */
4676 0x2e, 0, /* DW_TAG_subprogram, no children */
4677 0x3, 0x8, /* DW_AT_name, DW_FORM_string */
4678 0x11, 0x1, /* DW_AT_low_pc, DW_FORM_addr */
4679 0x12, 0x1, /* DW_AT_high_pc, DW_FORM_addr */
4680 0, 0, /* end of abbrev */
4681 0 /* no more abbrev */
4683 .str = "\0" ".text\0" ".debug_info\0" ".debug_abbrev\0"
4684 ".debug_frame\0" ".symtab\0" ".strtab\0" "code_gen_buffer",
4687 /* We only need a single jit entry; statically allocate it. */
4688 static struct jit_code_entry one_entry;
4690 uintptr_t buf = (uintptr_t)buf_ptr;
4691 size_t img_size = sizeof(struct ElfImage) + debug_frame_size;
4692 DebugFrameHeader *dfh;
4694 img = g_malloc(img_size);
4695 *img = img_template;
4697 img->phdr.p_vaddr = buf;
4698 img->phdr.p_paddr = buf;
4699 img->phdr.p_memsz = buf_size;
4701 img->shdr[1].sh_name = find_string(img->str, ".text");
4702 img->shdr[1].sh_addr = buf;
4703 img->shdr[1].sh_size = buf_size;
4705 img->shdr[2].sh_name = find_string(img->str, ".debug_info");
4706 img->shdr[3].sh_name = find_string(img->str, ".debug_abbrev");
4708 img->shdr[4].sh_name = find_string(img->str, ".debug_frame");
4709 img->shdr[4].sh_size = debug_frame_size;
4711 img->shdr[5].sh_name = find_string(img->str, ".symtab");
4712 img->shdr[6].sh_name = find_string(img->str, ".strtab");
4714 img->sym[1].st_name = find_string(img->str, "code_gen_buffer");
4715 img->sym[1].st_value = buf;
4716 img->sym[1].st_size = buf_size;
4718 img->di.cu_low_pc = buf;
4719 img->di.cu_high_pc = buf + buf_size;
4720 img->di.fn_low_pc = buf;
4721 img->di.fn_high_pc = buf + buf_size;
4723 dfh = (DebugFrameHeader *)(img + 1);
4724 memcpy(dfh, debug_frame, debug_frame_size);
4725 dfh->fde.func_start = buf;
4726 dfh->fde.func_len = buf_size;
4728 #ifdef DEBUG_JIT
4729 /* Enable this block to be able to debug the ELF image file creation.
4730 One can use readelf, objdump, or other inspection utilities. */
4732 FILE *f = fopen("/tmp/qemu.jit", "w+b");
4733 if (f) {
4734 if (fwrite(img, img_size, 1, f) != img_size) {
4735 /* Avoid stupid unused return value warning for fwrite. */
4737 fclose(f);
4740 #endif
4742 one_entry.symfile_addr = img;
4743 one_entry.symfile_size = img_size;
4745 __jit_debug_descriptor.action_flag = JIT_REGISTER_FN;
4746 __jit_debug_descriptor.relevant_entry = &one_entry;
4747 __jit_debug_descriptor.first_entry = &one_entry;
4748 __jit_debug_register_code();
4750 #else
4751 /* No support for the feature. Provide the entry point expected by exec.c,
4752 and implement the internal function we declared earlier. */
4754 static void tcg_register_jit_int(const void *buf, size_t size,
4755 const void *debug_frame,
4756 size_t debug_frame_size)
4760 void tcg_register_jit(const void *buf, size_t buf_size)
4763 #endif /* ELF_HOST_MACHINE */
4765 #if !TCG_TARGET_MAYBE_vec
4766 void tcg_expand_vec_op(TCGOpcode o, TCGType t, unsigned e, TCGArg a0, ...)
4768 g_assert_not_reached();
4770 #endif