1 /* architecture-dependent code generation for x86_64 */
5 /* x86-64 registers, without r8-r15 */
35 #define MIN(a, b) ((a) < (b) ? (a) : (b))
36 #define ALIGN(x, a) (((x) + (a) - 1) & ~((a) - 1))
38 int tmpregs
[] = {0, 7, 6, 2, 1, 8, 9, 10, 11, 3, 12, 13, 14, 15};
39 int argregs
[] = {7, 6, 2, 1, 8, 9};
41 #define OP2(o2, o1) (0x010000 | ((o2) << 8) | (o1))
42 #define O2(op) (((op) >> 8) & 0xff)
43 #define O1(op) ((op) & 0xff)
44 #define MODRM(m, r1, r2) ((m) << 6 | (r1) << 3 | (r2))
45 #define REX(r1, r2) (0x48 | (((r1) & 8) >> 1) | (((r2) & 8) >> 3))
47 static struct mem cs
; /* generated code */
49 /* code generation functions */
50 static void os(void *s
, int n
)
55 static char *ointbuf(long n
, int l
)
59 for (i
= 0; i
< l
; i
++) {
66 static void oi(long n
, int l
)
68 mem_put(&cs
, ointbuf(n
, l
), l
);
71 static void oi_at(long pos
, long n
, int l
)
73 mem_cpy(&cs
, pos
, ointbuf(n
, l
), l
);
76 static long opos(void)
81 static void op_x(int op
, int r1
, int r2
, int bt
)
99 oi(sz
== 1 ? O1(op
) & ~0x1 : O1(op
), 1);
104 /* op_*(): r=reg, m=mem, i=imm, s=sym */
105 static void op_rm(int op
, int src
, int base
, int off
, int bt
)
107 int dis
= off
== (char) off
? 1 : 4;
108 int mod
= dis
== 4 ? 2 : 1;
109 if (!off
&& (base
& 7) != R_RBP
)
111 op_x(op
, src
, base
, bt
);
112 oi(MODRM(mod
, src
& 0x07, base
& 0x07), 1);
113 if ((base
& 7) == R_RSP
)
119 static void op_rr(int op
, int src
, int dst
, int bt
)
121 op_x(op
, src
, dst
, bt
);
122 oi(MODRM(3, src
& 0x07, dst
& 0x07), 1);
125 #define movrx_bt(bt) (((bt) == 4) ? 4 : LONGSZ)
127 static int movrx_op(int bt
, int mov
)
131 return bt
& T_MSIGN
? I_MOVSXD
: mov
;
133 return OP2(0x0f, bt
& T_MSIGN
? 0xbf : 0xb7);
135 return OP2(0x0f, bt
& T_MSIGN
? 0xbe : 0xb6);
139 static void mov_r2r(int rd
, int r1
, unsigned bt
)
141 if (rd
!= r1
|| T_SZ(bt
) != LONGSZ
)
142 op_rr(movrx_op(bt
, I_MOVR
), rd
, r1
, movrx_bt(bt
));
145 static void i_push(int reg
)
147 op_x(I_PUSH
| (reg
& 0x7), 0, reg
, LONGSZ
);
150 void i_mov(int rd
, int rn
)
152 op_rr(movrx_op(LONGSZ
, I_MOVR
), rd
, rn
, movrx_bt(LONGSZ
));
155 static void i_add(int op
, int rd
, int r1
, int r2
)
157 /* opcode for O_ADD, O_SUB, O_AND, O_OR, O_XOR */
158 static int rx
[] = {0003, 0053, 0043, 0013, 0063};
159 op_rr(rx
[op
& 0x0f], rd
, r2
, LONGSZ
);
162 static void i_add_imm(int op
, int rd
, int rn
, long n
)
164 /* opcode for O_ADD, O_SUB, O_AND, O_OR, O_XOR */
165 static int rx
[] = {0xc0, 0xe8, 0xe0, 0xc8, 0xf0};
166 unsigned char s
[4] = {REX(0, rd
), 0x83, rx
[op
& 0x0f] | (rd
& 7), n
& 0xff};
170 static void i_num(int rd
, long n
)
173 op_rr(I_XOR
, rd
, rd
, 4);
176 if (n
< 0 && -n
<= 0xffffffff) {
177 op_rr(I_MOVI
, 0, rd
, LONGSZ
);
181 if (n
> 0 && n
<= 0xffffffff)
183 op_x(I_MOVIR
+ (rd
& 7), 0, rd
, len
);
188 static void i_mul(int rd
, int r1
, int r2
)
192 op_rr(I_MUL
, 4, r2
, LONGSZ
);
195 static void i_div(int op
, int rd
, int r1
, int r2
)
200 op_x(I_CQO
, R_RAX
, R_RDX
, LONGSZ
);
204 op_rr(I_MUL
, bt
& T_MSIGN
? 7 : 6, r2
, LONGSZ
);
207 static void i_tst(int rn
, int rm
)
209 op_rr(I_TST
, rn
, rm
, LONGSZ
);
212 static void i_cmp(int rn
, int rm
)
214 op_rr(I_CMP
, rn
, rm
, LONGSZ
);
217 static void i_cmp_imm(int rn
, long n
)
219 unsigned char s
[4] = {REX(0, rn
), 0x83, 0xf8 | rn
, n
& 0xff};
223 static void i_shl(int op
, int rd
, int r1
, int rs
)
227 if ((op
& 0x0f) == 1)
228 sm
= bt
& T_MSIGN
? 7 : 5;
229 op_rr(I_SHX
, sm
, rd
, LONGSZ
);
232 static void i_shl_imm(int op
, int rd
, int rn
, long n
)
235 int sm
= (op
& 0x1) ? (bt
& T_MSIGN
? 0xf8 : 0xe8) : 0xe0;
236 char s
[4] = {REX(0, rn
), 0xc1, sm
| (rn
& 7), n
& 0xff};
240 static void i_neg(int rd
)
242 op_rr(I_NOT
, 3, rd
, LONGSZ
);
245 static void i_not(int rd
)
247 op_rr(I_NOT
, 2, rd
, LONGSZ
);
250 static int i_cond(long op
)
252 /* lt, ge, eq, ne, le, gt */
253 static int ucond
[] = {0x92, 0x93, 0x94, 0x95, 0x96, 0x97};
254 static int scond
[] = {0x9c, 0x9d, 0x94, 0x95, 0x9e, 0x9f};
256 return bt
& T_MSIGN
? scond
[op
& 0x0f] : ucond
[op
& 0x0f];
259 static void i_set(long op
, int rd
)
261 char set
[] = "\x0f\x00\xc0";
263 os(set
, 3); /* setl al */
264 os("\x48\x0f\xb6\xc0", 4); /* movzx rax, al */
267 static void i_lnot(int rd
)
269 char cmp
[] = "\x00\x83\xf8\x00";
272 os(cmp
, 4); /* cmp rax, 0 */
276 static void jx(int x
, int nbytes
)
280 op
[0] = 0x70 | (x
& 0x0f);
281 os(op
, 1); /* jx $addr */
284 os(op
, 2); /* jx $addr */
288 /* generate cmp or tst before a conditional jump */
289 static void i_jcmp(long op
, long rn
, long rm
)
301 /* generate a jump instruction and return the of its displacement */
302 static long i_jmp(long op
, int nb
)
305 jx(O_C(op
) == O_JZ
? 0x84 : 0x85, nb
);
307 jx(i_cond(op
) & ~0x10, nb
);
309 os(nb
== 1 ? "\xeb" : "\xe9", 1);
314 /* the length of a jump instruction opcode */
315 static int i_jlen(long op
, int nb
)
317 if (op
& (O_JZ
| O_JCC
))
323 static void i_zx(int rd
, int r1
, int bits
)
326 i_shl_imm(O_SHL
, rd
, rd
, LONGSZ
* 8 - bits
);
327 i_shl_imm(O_SHR
, rd
, rd
, LONGSZ
* 8 - bits
);
329 mov_r2r(rd
, r1
, bits
>> 3);
334 static void i_sx(int rd
, int r1
, int bits
)
336 mov_r2r(rd
, r1
, T_MSIGN
| (bits
>> 3));
339 static void i_cast(int rd
, int rn
, int bt
)
346 i_sx(rd
, rn
, T_SZ(bt
) * 8);
348 i_zx(rd
, rn
, T_SZ(bt
) * 8);
352 static void i_add_anyimm(int rd
, int rn
, long n
)
354 op_rm(I_LEA
, rd
, rn
, n
, LONGSZ
);
357 static long *rel_sym
; /* relocation symbols */
358 static long *rel_flg
; /* relocation flags */
359 static long *rel_off
; /* relocation offsets */
360 static long rel_n
, rel_sz
; /* relocation count */
362 static long lab_sz
; /* label count */
363 static long *lab_loc
; /* label offsets in cs */
364 static long jmp_n
, jmp_sz
; /* jump count */
365 static long *jmp_off
; /* jump offsets */
366 static long *jmp_dst
; /* jump destinations */
367 static long *jmp_op
; /* jump opcode */
368 static long jmp_ret
; /* the position of the last return jmp */
370 static void lab_add(long id
)
372 while (id
>= lab_sz
) {
374 lab_sz
= MAX(128, lab_sz
* 2);
375 lab_loc
= mextend(lab_loc
, lab_n
, lab_sz
, sizeof(*lab_loc
));
377 lab_loc
[id
] = opos();
380 static void jmp_add(long op
, long off
, long dst
)
382 if (jmp_n
== jmp_sz
) {
383 jmp_sz
= MAX(128, jmp_sz
* 2);
384 jmp_off
= mextend(jmp_off
, jmp_n
, jmp_sz
, sizeof(*jmp_off
));
385 jmp_dst
= mextend(jmp_dst
, jmp_n
, jmp_sz
, sizeof(*jmp_dst
));
386 jmp_op
= mextend(jmp_op
, jmp_n
, jmp_sz
, sizeof(*jmp_op
));
388 jmp_off
[jmp_n
] = off
;
389 jmp_dst
[jmp_n
] = dst
;
394 void i_label(long id
)
399 static void i_rel(long sym
, long flg
, long off
)
401 if (rel_n
== rel_sz
) {
402 rel_sz
= MAX(128, rel_sz
* 2);
403 rel_sym
= mextend(rel_sym
, rel_n
, rel_sz
, sizeof(*rel_sym
));
404 rel_flg
= mextend(rel_flg
, rel_n
, rel_sz
, sizeof(*rel_flg
));
405 rel_off
= mextend(rel_off
, rel_n
, rel_sz
, sizeof(*rel_off
));
407 rel_sym
[rel_n
] = sym
;
408 rel_flg
[rel_n
] = flg
;
409 rel_off
[rel_n
] = off
;
413 static void i_sym(int rd
, int sym
, int off
)
415 int sz
= X64_ABS_RL
& OUT_RL32
? 4 : LONGSZ
;
416 if (X64_ABS_RL
& OUT_RLSX
)
417 op_rr(I_MOVI
, 0, rd
, sz
);
419 op_x(I_MOVIR
+ (rd
& 7), 0, rd
, sz
);
420 i_rel(sym
, OUT_CS
| X64_ABS_RL
, opos());
424 static void i_saveargs(long sargs
)
427 os("\x58", 1); /* pop rax */
428 for (i
= N_ARGS
- 1; i
>= 0; i
--)
429 if ((1 << argregs
[i
]) & sargs
)
431 os("\x50", 1); /* push rax */
434 static void i_saveregs(long sregs
, long sregs_pos
, int st
)
438 for (i
= 0; i
< N_TMPS
; i
++)
439 if ((1 << tmpregs
[i
]) & sregs
)
440 op_rm(st
? I_MOV
: I_MOVR
, tmpregs
[i
], REG_FP
,
441 sregs_pos
+ nsregs
++ * ULNG
, ULNG
);
444 void i_wrap(int argc
, long sargs
, long spsub
, int initfp
, long sregs
, long sregs_pos
)
448 long diff
; /* prologue length */
449 int nsargs
= 0; /* number of saved arguments */
450 int mod16
; /* 16-byte alignment */
452 /* removing the last jmp to the epilogue */
453 if (jmp_ret
+ i_jlen(O_JMP
, 4) + 4 == opos()) {
454 mem_cut(&cs
, jmp_ret
);
457 lab_add(0); /* the return label */
458 body_n
= mem_len(&cs
);
460 /* generating function prologue */
464 os("\x55", 1); /* push rbp */
465 os("\x48\x89\xe5", 3); /* mov rbp, rsp */
467 for (i
= 0; i
< N_ARGS
; i
++)
468 if ((1 << argregs
[i
]) & sargs
)
470 mod16
= (spsub
+ nsargs
* LONGSZ
) % 16; /* forcing 16-byte alignment */
472 os("\x48\x81\xec", 3);
473 spsub
= spsub
+ (16 - mod16
);
476 i_saveregs(sregs
, sregs_pos
, 1); /* saving registers */
478 mem_put(&cs
, body
, body_n
);
480 /* generating function epilogue */
481 i_saveregs(sregs
, sregs_pos
, 0); /* restoring saved registers */
483 os("\xc9", 1); /* leave */
485 os("\xc2", 1); /* ret n */
486 oi(nsargs
* LONGSZ
, 2);
488 os("\xc3", 1); /* ret */
490 /* adjusting code offsets */
491 for (i
= 0; i
< rel_n
; i
++)
493 for (i
= 0; i
< jmp_n
; i
++)
495 for (i
= 0; i
< lab_sz
; i
++)
499 /* introduce shorter jumps, if possible */
500 static void i_shortjumps(int *nb
)
502 long off
= 0; /* current code offset */
503 long dif
= 0; /* the difference after changing jump instructions */
504 int rel
= 0; /* current relocation */
505 int lab
= 1; /* current label */
506 long c_len
= mem_len(&cs
);
507 char *c
= mem_get(&cs
);
509 for (i
= 0; i
< jmp_n
; i
++)
510 nb
[i
] = abs(lab_loc
[jmp_dst
[i
]] - jmp_off
[i
]) < 0x70 ? 1 : 4;
511 for (i
= 0; i
< jmp_n
; i
++) {
512 long cur
= jmp_off
[i
] - i_jlen(jmp_op
[i
], 4);
513 while (rel
< rel_n
&& rel_off
[rel
] <= cur
)
514 rel_off
[rel
++] += dif
;
515 while (lab
< lab_sz
&& lab_loc
[lab
] <= cur
)
516 lab_loc
[lab
++] += dif
;
517 mem_put(&cs
, c
+ off
, cur
- off
);
518 jmp_off
[i
] = i_jmp(jmp_op
[i
], nb
[i
]);
519 off
= cur
+ i_jlen(jmp_op
[i
], 4) + 4;
520 dif
= mem_len(&cs
) - off
;
523 rel_off
[rel
++] += dif
;
525 lab_loc
[lab
++] += dif
;
527 mem_put(&cs
, c
+ off
, c_len
- off
);
531 void i_code(char **c
, long *c_len
, long **rsym
, long **rflg
, long **roff
, long *rcnt
)
533 int *nb
; /* number of bytes necessary for jump displacements */
535 /* more compact jmp instructions */
536 nb
= malloc(jmp_n
* sizeof(nb
[0]));
537 for (i
= 0; i
< jmp_n
; i
++)
540 for (i
= 0; i
< jmp_n
; i
++) /* filling jmp destinations */
541 oi_at(jmp_off
[i
], lab_loc
[jmp_dst
[i
]] -
542 jmp_off
[i
] - nb
[i
], nb
[i
]);
544 *c_len
= mem_len(&cs
);
566 long i_reg(long op
, long *rd
, long *r1
, long *r2
, long *r3
, long *tmp
)
576 *r1
= oc
& (O_NUM
| O_SYM
) ? 0 : R_TMPS
;
581 *r2
= oc
& O_NUM
? (oc
== O_ADD
? 32 : 8) : R_TMPS
;
597 *rd
= oc
== O_MOD
? (1 << R_RDX
) : (1 << R_RAX
);
599 *r2
= R_TMPS
& ~*rd
& ~*r1
;
601 *r2
&= ~(1 << R_RDX
);
602 *tmp
= (1 << R_RDX
) | (1 << R_RAX
);
608 *r2
= oc
& O_NUM
? 8 : R_TMPS
;
622 *tmp
= (1 << R_RDI
) | (1 << R_RCX
);
629 *tmp
= (1 << R_RDI
) | (1 << R_RSI
) | (1 << R_RCX
);
633 *r1
= (1 << REG_RET
);
637 *rd
= (1 << REG_RET
);
638 *r1
= oc
& O_SYM
? 0 : R_TMPS
;
639 *tmp
= R_TMPS
& ~R_PERM
;
645 *r2
= oc
& O_NUM
? 0 : R_TMPS
;
651 *r3
= oc
& O_NUM
? 0 : R_TMPS
;
660 *r2
= oc
& O_NUM
? 8 : R_TMPS
;
668 int i_imm(long lim
, long n
)
670 long max
= (1 << (lim
- 1)) - 1;
671 return n
<= max
&& n
+ 1 >= -max
;
674 long i_ins(long op
, long rd
, long r1
, long r2
, long r3
)
680 if (rd
== r1
&& r2
<= 127 && r2
>= -128)
681 i_add_imm(op
, r1
, r1
, r2
);
683 i_add_anyimm(rd
, r1
, r2
);
685 i_add(op
, r1
, r1
, r2
);
690 i_shl_imm(op
, r1
, r1
, r2
);
692 i_shl(op
, r1
, r1
, r2
);
696 i_mul(R_RAX
, r1
, r2
);
698 i_div(op
, R_RAX
, r1
, r2
);
700 i_div(op
, R_RDX
, r1
, r2
);
711 if (oc
& O_UOP
) { /* uop */
721 op_rr(I_CALL
, 2, r1
, LONGSZ
);
724 if (oc
== (O_CALL
| O_SYM
)) {
725 os("\xe8", 1); /* call $x */
726 i_rel(r1
, OUT_CS
| OUT_RLREL
, opos());
730 if (oc
== (O_MOV
| O_SYM
)) {
734 if (oc
== (O_MOV
| O_NUM
)) {
739 os("\xfc\xf3\xaa", 3); /* cld; rep stosb */
743 os("\xfc\xf3\xa4", 3); /* cld; rep movs */
748 jmp_add(O_JMP
, i_jmp(op
, 4), 0);
751 if (oc
== (O_LD
| O_NUM
)) {
752 op_rm(movrx_op(bt
, I_MOVR
), rd
, r1
, r2
, movrx_bt(bt
));
755 if (oc
== (O_ST
| O_NUM
)) {
756 op_rm(I_MOV
, r1
, r2
, r3
, bt
);
765 jmp_add(op
, i_jmp(op
, 4), r3
+ 1);