riscv: Fix tcctest.c
[tinycc.git] / riscv64-gen.c
blob09c011e30c9cd3c577c8f9de12d18fee46f7d415
1 #ifdef TARGET_DEFS_ONLY
3 // Number of registers available to allocator:
4 #define NB_REGS 19 // x10-x17 aka a0-a7, f10-f17 aka fa0-fa7, xxx, ra, sp
6 #define TREG_R(x) (x) // x = 0..7
7 #define TREG_F(x) (x + 8) // x = 0..7
9 // Register classes sorted from more general to more precise:
10 #define RC_INT (1 << 0)
11 #define RC_FLOAT (1 << 1)
12 #define RC_R(x) (1 << (2 + (x))) // x = 0..7
13 #define RC_F(x) (1 << (10 + (x))) // x = 0..7
15 #define RC_IRET (RC_R(0)) // int return register class
16 #define RC_FRET (RC_F(0)) // float return register class
18 #define REG_IRET (TREG_R(0)) // int return register number
19 #define REG_FRET (TREG_F(0)) // float return register number
21 #define PTR_SIZE 8
23 #define LDOUBLE_SIZE 16
24 #define LDOUBLE_ALIGN 16
26 #define MAX_ALIGN 16
28 #define CHAR_IS_UNSIGNED
30 #else
31 #include "tcc.h"
32 #include <assert.h>
34 #define XLEN 8
36 #define TREG_RA 17
37 #define TREG_SP 18
39 ST_DATA const int reg_classes[NB_REGS] = {
40 RC_INT | RC_R(0),
41 RC_INT | RC_R(1),
42 RC_INT | RC_R(2),
43 RC_INT | RC_R(3),
44 RC_INT | RC_R(4),
45 RC_INT | RC_R(5),
46 RC_INT | RC_R(6),
47 RC_INT | RC_R(7),
48 RC_FLOAT | RC_F(0),
49 RC_FLOAT | RC_F(1),
50 RC_FLOAT | RC_F(2),
51 RC_FLOAT | RC_F(3),
52 RC_FLOAT | RC_F(4),
53 RC_FLOAT | RC_F(5),
54 RC_FLOAT | RC_F(6),
55 RC_FLOAT | RC_F(7),
57 1 << TREG_RA,
58 1 << TREG_SP
61 static int ireg(int r)
63 if (r == TREG_RA)
64 return 1; // ra
65 if (r == TREG_SP)
66 return 2; // sp
67 assert(r >= 0 && r < 8);
68 return r + 10; // tccrX --> aX == x(10+X)
71 static int is_ireg(int r)
73 return r < 8 || r == TREG_RA || r == TREG_SP;
76 static int freg(int r)
78 assert(r >= 8 && r < 16);
79 return r - 8 + 10; // tccfX --> faX == f(10+X)
82 static int is_freg(int r)
84 return r >= 8 && r < 16;
87 ST_FUNC void o(unsigned int c)
89 int ind1 = ind + 4;
90 if (nocode_wanted)
91 return;
92 if (ind1 > cur_text_section->data_allocated)
93 section_realloc(cur_text_section, ind1);
94 write32le(cur_text_section->data + ind, c);
95 ind = ind1;
98 static void EIu(uint32_t opcode, uint32_t func3,
99 uint32_t rd, uint32_t rs1, uint32_t imm)
101 o(opcode | (func3 << 12) | (rd << 7) | (rs1 << 15) | (imm << 20));
104 static void EI(uint32_t opcode, uint32_t func3,
105 uint32_t rd, uint32_t rs1, uint32_t imm)
107 assert(! ((imm + (1 << 11)) >> 12));
108 EIu(opcode, func3, rd, rs1, imm);
111 static void ES(uint32_t opcode, uint32_t func3,
112 uint32_t rs1, uint32_t rs2, uint32_t imm)
114 assert(! ((imm + (1 << 11)) >> 12));
115 o(opcode | (func3 << 12) | ((imm & 0x1f) << 7) | (rs1 << 15)
116 | (rs2 << 20) | ((imm >> 5) << 25));
119 // Patch all branches in list pointed to by t to branch to a:
120 ST_FUNC void gsym_addr(int t_, int a_)
122 uint32_t t = t_;
123 uint32_t a = a_;
124 while (t) {
125 unsigned char *ptr = cur_text_section->data + t;
126 uint32_t next = read32le(ptr);
127 uint32_t r = a - t, imm;
128 if ((r + (1 << 21)) & ~((1U << 22) - 2))
129 tcc_error("out-of-range branch chain");
130 imm = (((r >> 12) & 0xff) << 12)
131 | (((r >> 11) & 1) << 20)
132 | (((r >> 1) & 0x3ff) << 21)
133 | (((r >> 20) & 1) << 31);
134 write32le(ptr, r == 4 ? 0x33 : 0x6f | imm); // nop || j imm
135 t = next;
139 ST_FUNC void load(int r, SValue *sv)
141 int fr = sv->r;
142 int v = fr & VT_VALMASK;
143 int rr = is_ireg(r) ? ireg(r) : freg(r);
144 int fc = sv->c.i;
145 int bt = sv->type.t & VT_BTYPE;
146 int align, size = type_size(&sv->type, &align);
147 if (fr & VT_LVAL) {
148 int func3, opcode = 0x03, doload = 0;
149 if (is_freg(r)) {
150 assert(bt == VT_DOUBLE || bt == VT_FLOAT);
151 opcode = 0x07;
152 func3 = bt == VT_DOUBLE ? 3 : 2;
153 } else {
154 assert(is_ireg(r));
155 if (bt == VT_FUNC)
156 size = PTR_SIZE;
157 func3 = size == 1 ? 0 : size == 2 ? 1 : size == 4 ? 2 : 3;
158 if (size < 4 && !is_float(sv->type.t) && (sv->type.t & VT_UNSIGNED))
159 func3 |= 4;
161 if (v == VT_LOCAL) {
162 int br = 8; // s0
163 if (fc != sv->c.i)
164 tcc_error("unimp: load1(giant local ofs) (0x%llx)", (long long)sv->c.i);
165 if (((unsigned)fc + (1 << 11)) >> 12) {
166 br = is_ireg(r) ? rr : 5;
167 o(0x37 | (br << 7) | ((0x800 + fc) & 0xfffff000)); //lui BR, upper(fc)
168 o(0x33 | (br << 7) | (br << 15) | (8 << 20)); // add BR, BR, s0
169 fc = fc << 20 >> 20;
171 EI(opcode, func3, rr, br, fc); // l[bhwd][u]/fl[wd] RR, fc(BR)
172 } else if (v < VT_CONST) {
173 /*if (((unsigned)fc + (1 << 11)) >> 12)
174 tcc_error("unimp: load(large addend) (0x%x)", fc);*/
175 fc = 0; // XXX store ofs in LVAL(reg)
176 EI(opcode, func3, rr, ireg(v), fc); // l[bhwd][u] RR, 0(V)
177 } else if (v == VT_CONST && (fr & VT_SYM)) {
178 static Sym label;
179 int tempr;
180 if (sv->sym->type.t & VT_STATIC) { // XXX do this per linker relax
181 greloca(cur_text_section, sv->sym, ind,
182 R_RISCV_PCREL_HI20, sv->c.i);
183 fc = 0;
184 sv->c.i = 0;
185 } else {
186 if (((unsigned)fc + (1 << 11)) >> 12)
187 tcc_error("unimp: large addend for global address (0x%llx)", sv->c.i);
188 greloca(cur_text_section, sv->sym, ind,
189 R_RISCV_GOT_HI20, 0);
190 doload = 1;
192 if (!label.v) {
193 label.v = tok_alloc(".L0 ", 4)->tok;
194 label.type.t = VT_VOID | VT_STATIC;
196 label.c = 0; /* force new local ELF symbol */
197 put_extern_sym(&label, cur_text_section, ind, 0);
198 tempr = is_ireg(r) ? rr : 5;
199 o(0x17 | (tempr << 7)); // auipc TR, 0 %pcrel_hi(sym)+addend
200 greloca(cur_text_section, &label, ind,
201 R_RISCV_PCREL_LO12_I, 0);
202 if (doload) {
203 EI(0x03, 3, tempr, tempr, 0); // ld TR, 0(TR)
204 if (fc)
205 EI(0x13, 0, tempr, tempr, fc << 20 >> 20); // addi TR, TR, FC
206 fc = 0;
208 EI(opcode, func3, rr, tempr, fc); // l[bhwd][u] RR, fc(TR)
209 } else if (v == VT_LLOCAL) {
210 int br = 8, tempr = is_ireg(r) ? rr : 5;
211 if (fc != sv->c.i)
212 tcc_error("unimp: load2(giant local ofs) (0x%llx)", (long long)sv->c.i);
213 if (((unsigned)fc + (1 << 11)) >> 12) {
214 br = tempr;
215 o(0x37 | (br << 7) | ((0x800 + fc) & 0xfffff000)); //lui BR, upper(fc)
216 o(0x33 | (br << 7) | (br << 15) | (8 << 20)); // add BR, BR, s0
217 fc = fc << 20 >> 20;
219 EI(0x03, 3, tempr, br, fc); // ld TEMPR, fc(BR)
220 EI(opcode, func3, rr, tempr, 0); // l[bhwd][u] RR, 0(TEMPR)
221 } else {
222 tcc_error("unimp: load(non-local lval)");
224 } else if (v == VT_CONST) {
225 int rb = 0, do32bit = 8, doload = 0, zext = 0;
226 assert((!is_float(sv->type.t) && is_ireg(r)) || bt == VT_LDOUBLE);
227 if (fr & VT_SYM) {
228 static Sym label;
229 if (sv->sym->type.t & VT_STATIC) { // XXX do this per linker relax
230 greloca(cur_text_section, sv->sym, ind,
231 R_RISCV_PCREL_HI20, sv->c.i);
232 fc = 0;
233 sv->c.i = 0;
234 } else {
235 if (((unsigned)fc + (1 << 11)) >> 12)
236 tcc_error("unimp: large addend for global address (0x%llx)", sv->c.i);
237 greloca(cur_text_section, sv->sym, ind,
238 R_RISCV_GOT_HI20, 0);
239 doload = 1;
241 if (!label.v) {
242 label.v = tok_alloc(".L0 ", 4)->tok;
243 label.type.t = VT_VOID | VT_STATIC;
245 label.c = 0; /* force new local ELF symbol */
246 put_extern_sym(&label, cur_text_section, ind, 0);
247 o(0x17 | (rr << 7)); // auipc RR, 0 %call(func)
248 greloca(cur_text_section, &label, ind,
249 R_RISCV_PCREL_LO12_I, 0);
250 rb = rr;
251 do32bit = 0;
253 if (is_float(sv->type.t) && bt != VT_LDOUBLE)
254 tcc_error("unimp: load(float)");
255 if (fc != sv->c.i) {
256 int64_t si = sv->c.i;
257 uint32_t pi;
258 si >>= 32;
259 if (si != 0) {
260 pi = si;
261 if (fc < 0)
262 pi++;
263 o(0x37 | (rr << 7) | (((pi + 0x800) & 0xfffff000))); // lui RR, up(up(fc))
264 EI(0x13, 0, rr, rr, (int)pi << 20 >> 20); // addi RR, RR, lo(up(fc))
265 EI(0x13, 1, rr, rr, 12); // slli RR, RR, 12
266 EI(0x13, 0, rr, rr, (fc + (1 << 19)) >> 20); // addi RR, RR, up(lo(fc))
267 EI(0x13, 1, rr, rr, 12); // slli RR, RR, 12
268 fc = fc << 12 >> 12;
269 EI(0x13, 0, rr, rr, fc >> 8); // addi RR, RR, lo1(lo(fc))
270 EI(0x13, 1, rr, rr, 8); // slli RR, RR, 8
271 fc &= 0xff;
272 rb = rr;
273 do32bit = 0;
274 } else if (bt == VT_LLONG) {
275 /* A 32bit unsigned constant for a 64bit type.
276 lui always sign extends, so we need to do an explicit zext.*/
277 zext = 1;
280 if (((unsigned)fc + (1 << 11)) >> 12)
281 o(0x37 | (rr << 7) | ((0x800 + fc) & 0xfffff000)), rb = rr; //lui RR, upper(fc)
282 if (doload) {
283 EI(0x03, 3, rr, rr, 0); // ld RR, 0(RR)
284 if (fc)
285 EI(0x13 | do32bit, 0, rr, rr, fc << 20 >> 20); // addi[w] R, x0|R, FC
286 } else
287 EI(0x13 | do32bit, 0, rr, rb, fc << 20 >> 20); // addi[w] R, x0|R, FC
288 if (zext) {
289 EI(0x13, 1, rr, rr, 32); // slli RR, RR, 32
290 EI(0x13, 5, rr, rr, 32); // srli RR, RR, 32
292 } else if (v == VT_LOCAL) {
293 int br = 8; // s0
294 assert(is_ireg(r));
295 if (fc != sv->c.i)
296 tcc_error("unimp: load(addr giant local ofs) (0xll%x)", (long long)sv->c.i);
297 if (((unsigned)fc + (1 << 11)) >> 12) {
298 o(0x37 | (rr << 7) | ((0x800 + fc) & 0xfffff000)); //lui RR, upper(fc)
299 o(0x33 | (rr << 7) | (rr << 15) | (8 << 20)); // add RR, RR, s0
300 fc = fc << 20 >> 20;
301 br = rr;
303 EI(0x13, 0, rr, br, fc); // addi R, s0, FC
304 } else if (v < VT_CONST) {
305 /* reg-reg */
306 //assert(!fc); XXX support offseted regs
307 if (is_freg(r) && is_freg(v))
308 o(0x53 | (rr << 7) | (freg(v) << 15) | (freg(v) << 20) | ((bt == VT_DOUBLE ? 0x11 : 0x10) << 25)); //fsgnj.[sd] RR, V, V == fmv.[sd] RR, V
309 else if (is_ireg(r) && is_ireg(v))
310 EI(0x13, 0, rr, ireg(v), 0); // addi RR, V, 0 == mv RR, V
311 else {
312 int func7 = is_ireg(r) ? 0x70 : 0x78;
313 if (size == 8)
314 func7 |= 1;
315 assert(size == 4 || size == 8);
316 o(0x53 | (rr << 7) | ((is_freg(v) ? freg(v) : ireg(v)) << 15)
317 | (func7 << 25)); // fmv.{w.x, x.w, d.x, x.d} RR, VR
319 } else if (v == VT_CMP) { // we rely on cmp_r to be the correct result
320 EI(0x13, 0, rr, vtop->cmp_r, 0); // mv RR, CMP_R
321 } else if ((v & ~1) == VT_JMP) {
322 int t = v & 1;
323 assert(is_ireg(r));
324 EI(0x13, 0, rr, 0, t); // addi RR, x0, t
325 gjmp_addr(ind + 8);
326 gsym(fc);
327 EI(0x13, 0, rr, 0, t ^ 1); // addi RR, x0, !t
328 } else
329 tcc_error("unimp: load(non-const)");
332 ST_FUNC void store(int r, SValue *sv)
334 int fr = sv->r & VT_VALMASK;
335 int rr = is_ireg(r) ? ireg(r) : freg(r);
336 int fc = sv->c.i;
337 int ft = sv->type.t;
338 int bt = ft & VT_BTYPE;
339 int align, size = type_size(&sv->type, &align);
340 assert(!is_float(bt) || is_freg(r) || bt == VT_LDOUBLE);
341 /* long doubles are in two integer registers, but the load/store
342 primitives only deal with one, so do as if it's one reg. */
343 if (bt == VT_LDOUBLE)
344 size = align = 8;
345 if (bt == VT_STRUCT)
346 tcc_error("unimp: store(struct)");
347 if (size > 8)
348 tcc_error("unimp: large sized store");
349 assert(sv->r & VT_LVAL);
350 if (fr == VT_LOCAL) {
351 int br = 8; // s0
352 if (fc != sv->c.i)
353 tcc_error("unimp: store(giant local off) (0x%llx)", (long long)sv->c.i);
354 if (((unsigned)fc + (1 << 11)) >> 12) {
355 br = 5; // t0
356 o(0x37 | (br << 7) | ((0x800 + fc) & 0xfffff000)); //lui BR, upper(fc)
357 o(0x33 | (br << 7) | (br << 15) | (8 << 20)); // add BR, BR, s0
358 fc = fc << 20 >> 20;
360 if (is_freg(r))
361 ES(0x27, size == 4 ? 2 : 3, br, rr, fc); // fs[wd] RR, fc(base)
362 else
363 ES(0x23, size == 1 ? 0 : size == 2 ? 1 : size == 4 ? 2 : 3,
364 br, rr, fc); // s[bhwd] RR, fc(base)
365 } else if (fr < VT_CONST) {
366 int ptrreg = ireg(fr);
367 /*if (((unsigned)fc + (1 << 11)) >> 12)
368 tcc_error("unimp: store(large addend) (0x%x)", fc);*/
369 fc = 0; // XXX support offsets regs
370 if (is_freg(r))
371 ES(0x27, size == 4 ? 2 : 3, ptrreg, rr, fc); // fs[wd] RR, fc(PTRREG)
372 else
373 ES(0x23, size == 1 ? 0 : size == 2 ? 1 : size == 4 ? 2 : 3,
374 ptrreg, rr, fc); // s[bhwd] RR, fc(PTRREG)
375 } else if ((sv->r & ~VT_LVAL_TYPE) == (VT_CONST | VT_SYM | VT_LVAL)) {
376 static Sym label;
377 int tempr, doload = 0;
378 tempr = 5; // t0
379 if (sv->sym->type.t & VT_STATIC) { // XXX do this per linker relax
380 greloca(cur_text_section, sv->sym, ind,
381 R_RISCV_PCREL_HI20, sv->c.i);
382 fc = 0;
383 sv->c.i = 0;
384 } else {
385 if (((unsigned)fc + (1 << 11)) >> 12)
386 tcc_error("unimp: large addend for global address (0x%llx)", sv->c.i);
387 greloca(cur_text_section, sv->sym, ind,
388 R_RISCV_GOT_HI20, 0);
389 doload = 1;
391 if (!label.v) {
392 label.v = tok_alloc(".L0 ", 4)->tok;
393 label.type.t = VT_VOID | VT_STATIC;
395 label.c = 0; /* force new local ELF symbol */
396 put_extern_sym(&label, cur_text_section, ind, 0);
397 o(0x17 | (tempr << 7)); // auipc TEMPR, 0 %pcrel_hi(sym)+addend
398 greloca(cur_text_section, &label, ind,
399 doload ? R_RISCV_PCREL_LO12_I : R_RISCV_PCREL_LO12_S, 0);
400 if (doload) {
401 EI(0x03, 3, tempr, tempr, 0); // ld TR, 0(TR)
402 if (fc)
403 EI(0x13, 0, tempr, tempr, fc << 20 >> 20); // addi TR, TR, FC
404 fc = 0;
406 if (is_freg(r))
407 ES(0x27, size == 4 ? 2 : 3, tempr, rr, fc); // fs[wd] RR, fc(TEMPR)
408 else
409 ES(0x23, size == 1 ? 0 : size == 2 ? 1 : size == 4 ? 2 : 3,
410 tempr, rr, fc); // s[bhwd] RR, fc(TEMPR)
411 } else
412 tcc_error("implement me: %s(!local)", __FUNCTION__);
415 static void gcall_or_jmp(int docall)
417 int tr = docall ? 1 : 5; // ra or t0
418 if ((vtop->r & (VT_VALMASK | VT_LVAL)) == VT_CONST &&
419 ((vtop->r & VT_SYM) && vtop->c.i == (int)vtop->c.i)) {
420 /* constant symbolic case -> simple relocation */
421 greloca(cur_text_section, vtop->sym, ind,
422 R_RISCV_CALL_PLT, (int)vtop->c.i);
423 o(0x17 | (tr << 7)); // auipc TR, 0 %call(func)
424 EI(0x67, 0, tr, tr, 0);// jalr TR, r(TR)
425 } else if (vtop->r < VT_CONST) {
426 int r = ireg(vtop->r);
427 EI(0x67, 0, tr, r, 0); // jalr TR, 0(R)
428 } else {
429 int r = TREG_RA;
430 load(r, vtop);
431 r = ireg(r);
432 EI(0x67, 0, tr, r, 0); // jalr TR, 0(R)
436 ST_FUNC void gfunc_call(int nb_args)
438 int i, align, size, aireg, afreg;
439 int info[nb_args ? nb_args : 1];
440 int stack_adj = 0, tempspace = 0, ofs, splitofs = 0;
441 int force_stack = 0;
442 SValue *sv;
443 Sym *sa;
444 aireg = afreg = 0;
445 sa = vtop[-nb_args].type.ref->next;
446 for (i = 0; i < nb_args; i++) {
447 int *pareg, nregs, infreg = 0, byref = 0, tempofs;
448 sv = &vtop[1 + i - nb_args];
449 sv->type.t &= ~VT_ARRAY; // XXX this should be done in tccgen.c
450 size = type_size(&sv->type, &align);
451 if (size > 16) {
452 if (align < XLEN)
453 align = XLEN;
454 tempspace = (tempspace + align - 1) & -align;
455 tempofs = tempspace;
456 tempspace += size;
457 size = align = 8;
458 byref = 1;
460 if (size > 8)
461 nregs = 2;
462 else
463 nregs = 1;
464 if ((sv->type.t & VT_BTYPE) == VT_LDOUBLE) {
465 infreg = 0;
466 } else
467 infreg = sa && is_float(sv->type.t);
468 if (!infreg && !sa && align == 2*XLEN && size <= 2*XLEN)
469 aireg = (aireg + 1) & ~1;
470 pareg = infreg ? &afreg : &aireg;
471 if ((*pareg < 8) && !force_stack) {
472 info[i] = *pareg + (infreg ? 8 : 0);
473 (*pareg)++;
474 if (nregs == 1)
476 else if (*pareg < 8)
477 (*pareg)++;
478 else {
479 info[i] |= 16;
480 stack_adj += 8;
482 } else {
483 info[i] = 32;
484 if (align < XLEN)
485 align = XLEN;
486 stack_adj += (size + align - 1) & -align;
487 if (!sa)
488 force_stack = 1;
490 if (byref)
491 info[i] |= 64 | (tempofs << 7);
492 if (sa)
493 sa = sa->next;
495 stack_adj = (stack_adj + 15) & -16;
496 tempspace = (tempspace + 15) & -16;
497 if (stack_adj + tempspace) {
498 EI(0x13, 0, 2, 2, -(stack_adj + tempspace)); // addi sp, sp, -adj
499 for (i = ofs = 0; i < nb_args; i++) {
500 if (info[i] >= 32) {
501 vrotb(nb_args - i);
502 size = type_size(&vtop->type, &align);
503 if (info[i] & 64) {
504 vset(&char_pointer_type, TREG_SP, 0);
505 vpushi(stack_adj + (info[i] >> 7));
506 gen_op('+');
507 vpushv(vtop); // this replaces the old argument
508 vrott(3);
509 indir();
510 vtop->type = vtop[-1].type;
511 vswap();
512 vstore();
513 vpop();
514 size = align = 8;
516 if (info[i] & 32) {
517 if (align < XLEN)
518 align = XLEN;
519 /* Once we support offseted regs we can do this:
520 vset(&vtop->type, TREG_SP | VT_LVAL, ofs);
521 to construct the lvalue for the outgoing stack slot,
522 until then we have to jump through hoops. */
523 vset(&char_pointer_type, TREG_SP, 0);
524 ofs = (ofs + align - 1) & -align;
525 vpushi(ofs);
526 gen_op('+');
527 indir();
528 vtop->type = vtop[-1].type;
529 vswap();
530 vstore();
531 vtop->r = vtop->r2 = VT_CONST; // this arg is done
532 ofs += size;
534 vrott(nb_args - i);
535 } else if (info[i] & 16) {
536 assert(!splitofs);
537 splitofs = ofs;
538 ofs += 8;
542 for (i = 0; i < nb_args; i++) {
543 int r = info[nb_args - 1 - i];
544 if (!(r & 32)) {
545 CType origtype;
546 r &= 15;
547 vrotb(i+1);
548 origtype = vtop->type;
549 size = type_size(&vtop->type, &align);
550 if (size > 8 && (vtop->type.t & VT_BTYPE) == VT_STRUCT)
551 vtop->type.t = VT_LDOUBLE; // force loading a pair of regs
552 gv(r < 8 ? RC_R(r) : RC_F(r - 8));
553 vtop->type = origtype;
554 if (size > 8) {
555 assert((vtop->type.t & VT_BTYPE) == VT_LDOUBLE
556 || (vtop->type.t & VT_BTYPE) == VT_STRUCT);
557 assert(vtop->r2 < VT_CONST);
558 if (info[nb_args - 1 - i] & 16) {
559 ES(0x23, 3, 2, ireg(vtop->r2), splitofs); // sd t0, ofs(sp)
560 } else if (vtop->r2 != 1 + vtop->r) {
561 assert(vtop->r < 7);
562 /* XXX we'd like to have 'gv' move directly into
563 the right class instead of us fixing it up. */
564 EI(0x13, 0, ireg(vtop->r) + 1, ireg(vtop->r2), 0); // mv Ra+1, RR2
565 vtop->r2 = 1 + vtop->r;
568 vrott(i+1);
571 vrotb(nb_args + 1);
572 save_regs(nb_args + 1);
573 gcall_or_jmp(1);
574 vtop -= nb_args + 1;
575 if (stack_adj + tempspace)
576 EI(0x13, 0, 2, 2, stack_adj + tempspace); // addi sp, sp, adj
579 static int func_sub_sp_offset, num_va_regs;
581 ST_FUNC void gfunc_prolog(CType *func_type)
583 int i, addr, align, size;
584 int param_addr = 0;
585 int aireg, afreg;
586 Sym *sym;
587 CType *type;
589 sym = func_type->ref;
590 func_vt = sym->type;
591 loc = -16; // for ra and s0
592 func_sub_sp_offset = ind;
593 ind += 5 * 4;
595 aireg = afreg = 0;
596 addr = 0; // XXX not correct
597 /* if the function returns by reference, then add an
598 implicit pointer parameter */
599 size = type_size(&func_vt, &align);
600 if (size > 2 * XLEN) {
601 loc -= 8;
602 func_vc = loc;
603 ES(0x23, 3, 8, 10 + aireg, loc); // sd a0, loc(s0)
604 aireg++;
606 /* define parameters */
607 while ((sym = sym->next) != NULL) {
608 int byref = 0;
609 type = &sym->type;
610 size = type_size(type, &align);
611 if (size > 2 * XLEN) {
612 type = &char_pointer_type;
613 size = align = byref = 8;
615 if (size > 2 * XLEN) {
616 from_stack:
617 if (align < XLEN)
618 align = XLEN;
619 addr = (addr + align - 1) & -align;
620 param_addr = addr;
621 addr += size;
622 } else {
623 int regcount = 1, *pareg = &aireg;
624 if (is_float(type->t) && (type->t & VT_BTYPE) != VT_LDOUBLE)
625 pareg = &afreg;
626 if (regcount + *pareg > 8)
627 goto from_stack;
628 if (size > XLEN)
629 regcount++;
630 loc -= regcount * 8; // XXX could reserve only 'size' bytes
631 param_addr = loc;
632 for (i = 0; i < regcount; i++) {
633 if (*pareg >= 8) {
634 assert(i == 1 && regcount == 2 && !(addr & 7));
635 EI(0x03, 3, 5, 8, addr); // ld t0, addr(s0)
636 addr += 8;
637 ES(0x23, 3, 8, 5, loc + i*8); // sd t0, loc(s0)
638 continue;
640 if (pareg == &afreg) {
641 assert(type->t == VT_FLOAT || type->t == VT_DOUBLE);
642 ES(0x27, size == 4 ? 2 : 3, 8, 10 + *pareg, loc + i*8); // fs[wd] FAi, loc(s0)
643 } else {
644 ES(0x23, 3, 8, 10 + *pareg, loc + i*8); // sd aX, loc(s0) // XXX
646 (*pareg)++;
649 sym_push(sym->v & ~SYM_FIELD, &sym->type,
650 (byref ? VT_LLOCAL : VT_LOCAL) | lvalue_type(sym->type.t),
651 param_addr);
653 num_va_regs = 0;
654 if (func_type->ref->f.func_type == FUNC_ELLIPSIS) {
655 for (; aireg < 8; aireg++) {
656 num_va_regs++;
657 ES(0x23, 3, 8, 10 + aireg, -8 + num_va_regs * 8); // sd aX, loc(s0)
662 ST_FUNC int gfunc_sret(CType *vt, int variadic, CType *ret,
663 int *ret_align, int *regsize)
665 /* generic code can only deal with structs of pow(2) sizes
666 (it always deals with whole registers), so go through our own
667 code. */
668 int align, size = type_size(vt, &align);
669 *ret_align = 1;
670 *regsize = 8;
671 if (size > 16)
672 return 0;
673 if (size > 8)
674 ret->t = VT_LLONG;
675 else if (size > 4)
676 ret->t = VT_LLONG;
677 else if (size > 2)
678 ret->t = VT_INT;
679 else if (size > 1)
680 ret->t = VT_SHORT;
681 else
682 ret->t = VT_BYTE;
683 return (size + 7) / 8;
686 ST_FUNC void gfunc_return(CType *func_type)
688 int align, size = type_size(func_type, &align), nregs;
689 CType type = *func_type;
690 if (size > 2 * XLEN) {
691 mk_pointer(&type);
692 vset(&type, VT_LOCAL | VT_LVAL, func_vc);
693 indir();
694 vswap();
695 vstore();
696 vpop();
697 return;
699 nregs = (size + 7) / 8;
700 if (nregs == 2)
701 vtop->type.t = VT_LDOUBLE;
703 if (is_float(func_type->t) && (vtop->type.t & VT_BTYPE) != VT_LDOUBLE)
704 gv(RC_FRET);
705 else
706 gv(RC_IRET);
707 vtop--;
710 ST_FUNC void gfunc_epilog(void)
712 int v, saved_ind, d, large_ofs_ind;
714 loc = (loc - num_va_regs * 8);
715 d = v = (-loc + 15) & -16;
717 if (v >= (1 << 11)) {
718 d = 16;
719 o(0x37 | (5 << 7) | ((0x800 + (v-16)) & 0xfffff000)); //lui t0, upper(v)
720 EI(0x13, 0, 5, 5, (v-16) << 20 >> 20); // addi t0, t0, lo(v)
721 o(0x33 | (2 << 7) | (2 << 15) | (5 << 20)); //add sp, sp, t0
723 EI(0x03, 3, 1, 2, d - 8 - num_va_regs * 8); // ld ra, v-8(sp)
724 EI(0x03, 3, 8, 2, d - 16 - num_va_regs * 8); // ld s0, v-16(sp)
725 EI(0x13, 0, 2, 2, d); // addi sp, sp, v
726 EI(0x67, 0, 0, 1, 0); // jalr x0, 0(x1), aka ret
727 if (v >= (1 << 11)) {
728 large_ofs_ind = ind;
729 EI(0x13, 0, 8, 2, d - num_va_regs * 8); // addi s0, sp, d
730 o(0x37 | (5 << 7) | ((0x800 + (v-16)) & 0xfffff000)); //lui t0, upper(v)
731 EI(0x13, 0, 5, 5, (v-16) << 20 >> 20); // addi t0, t0, lo(v)
732 o(0x33 | (2 << 7) | (2 << 15) | (5 << 20) | (0x20 << 25)); //sub sp, sp, t0
733 gjmp_addr(func_sub_sp_offset + 5*4);
735 saved_ind = ind;
737 ind = func_sub_sp_offset;
738 EI(0x13, 0, 2, 2, -d); // addi sp, sp, -d
739 ES(0x23, 3, 2, 1, d - 8 - num_va_regs * 8); // sd ra, d-8(sp)
740 ES(0x23, 3, 2, 8, d - 16 - num_va_regs * 8); // sd s0, d-16(sp)
741 if (v < (1 << 11))
742 EI(0x13, 0, 8, 2, d - num_va_regs * 8); // addi s0, sp, d
743 else
744 gjmp_addr(large_ofs_ind);
745 if ((ind - func_sub_sp_offset) != 5*4)
746 EI(0x13, 0, 0, 0, 0); // addi x0, x0, 0 == nop
747 ind = saved_ind;
750 ST_FUNC void gen_va_start(void)
752 tcc_error("implement me: %s", __FUNCTION__);
755 ST_FUNC void gen_va_arg(CType *t)
757 tcc_error("implement me: %s", __FUNCTION__);
760 ST_FUNC void gen_fill_nops(int bytes)
762 if ((bytes & 3))
763 tcc_error("alignment of code section not multiple of 4");
764 while (bytes > 0) {
765 EI(0x13, 0, 0, 0, 0); // addi x0, x0, 0 == nop
766 bytes -= 4;
770 // Generate forward branch to label:
771 ST_FUNC int gjmp(int t)
773 if (nocode_wanted)
774 return t;
775 o(t);
776 return ind - 4;
779 // Generate branch to known address:
780 ST_FUNC void gjmp_addr(int a)
782 uint32_t r = a - ind, imm;
783 if ((r + (1 << 21)) & ~((1U << 22) - 2)) {
784 o(0x17 | (5 << 7) | (((r + 0x800) & 0xfffff000))); // lui RR, up(r)
785 r = (int)r << 20 >> 20;
786 EI(0x67, 0, 0, 5, r); // jalr x0, r(t0)
787 } else {
788 imm = (((r >> 12) & 0xff) << 12)
789 | (((r >> 11) & 1) << 20)
790 | (((r >> 1) & 0x3ff) << 21)
791 | (((r >> 20) & 1) << 31);
792 o(0x6f | imm); // jal x0, imm == j imm
796 ST_FUNC int gjmp_cond(int op, int t)
798 int inv = op & 1;
799 assert(op == TOK_EQ || op == TOK_NE);
800 assert(vtop->cmp_r >= 10 && vtop->cmp_r < 18);
801 o(0x63 | (!inv << 12) | (vtop->cmp_r << 15) | (8 << 7)); // bne/beq x0,r,+4
802 return gjmp(t);
805 ST_FUNC int gjmp_append(int n, int t)
807 void *p;
808 /* insert jump list n into t */
809 if (n) {
810 uint32_t n1 = n, n2;
811 while ((n2 = read32le(p = cur_text_section->data + n1)))
812 n1 = n2;
813 write32le(p, t);
814 t = n;
816 return t;
819 static void gen_opil(int op, int ll)
821 int a, b, d;
822 int inv = 0;
823 int func3 = 0, func7 = 0;
824 /* XXX We could special-case some constant args. */
825 gv2(RC_INT, RC_INT);
826 a = ireg(vtop[-1].r);
827 b = ireg(vtop[0].r);
828 vtop -= 2;
829 d = get_reg(RC_INT);
830 vtop++;
831 vtop[0].r = d;
832 d = ireg(d);
833 ll = ll ? 0 : 8;
834 switch (op) {
835 default:
836 tcc_error("implement me: %s(%s)", __FUNCTION__, get_tok_str(op, NULL));
838 case '+':
839 o(0x33 | (d << 7) | (a << 15) | (b << 20)); // add d, a, b
840 break;
841 case '-':
842 o(0x33 | (d << 7) | (a << 15) | (b << 20) | (0x20 << 25)); //sub d, a, b
843 break;
844 case TOK_SAR:
845 o(0x33 | ll | (d << 7) | (a << 15) | (b << 20) | (5 << 12) | (1 << 30)); //sra d, a, b
846 break;
847 case TOK_SHR:
848 o(0x33 | ll | (d << 7) | (a << 15) | (b << 20) | (5 << 12)); //srl d, a, b
849 break;
850 case TOK_SHL:
851 o(0x33 | (d << 7) | (a << 15) | (b << 20) | (1 << 12)); //sll d, a, b
852 break;
853 case '*':
854 o(0x33 | (d << 7) | (a << 15) | (b << 20) | (0x01 << 25)); //mul d, a, b
855 break;
856 case '/':
857 o(0x33 | (d << 7) | (a << 15) | (b << 20) | (0x01 << 25) | (4 << 12)); //div d, a, b
858 break;
859 case '&':
860 o(0x33 | (d << 7) | (a << 15) | (b << 20) | (7 << 12)); // and d, a, b
861 break;
862 case '^':
863 o(0x33 | (d << 7) | (a << 15) | (b << 20) | (4 << 12)); // xor d, a, b
864 break;
865 case '|':
866 o(0x33 | (d << 7) | (a << 15) | (b << 20) | (6 << 12)); // or d, a, b
867 break;
868 case '%':
869 o(0x33 | (d << 7) | (a << 15) | (b << 20) | (0x01 << 25) | (6 << 12)); //rem d, a, b
870 break;
871 case TOK_UMOD:
872 o(0x33 | (d << 7) | (a << 15) | (b << 20) | (0x01 << 25) | (7 << 12)); //remu d, a, b
873 break;
874 case TOK_PDIV:
875 case TOK_UDIV:
876 o(0x33 | (d << 7) | (a << 15) | (b << 20) | (0x01 << 25) | (5 << 12)); //divu d, a, b
877 break;
879 case TOK_ULT:
880 case TOK_UGE:
881 case TOK_ULE:
882 case TOK_UGT:
883 case TOK_LT:
884 case TOK_GE:
885 case TOK_LE:
886 case TOK_GT:
887 if (op & 1) { // remove [U]GE,GT
888 inv = 1;
889 op--;
891 if ((op & 7) == 6) { // [U]LE
892 int t = a; a = b; b = t;
893 inv ^= 1;
895 o(0x33 | (d << 7) | (a << 15) | (b << 20) | (((op > TOK_UGT) ? 2 : 3) << 12)); // slt[u] d, a, b
896 if (inv)
897 EI(0x13, 4, d, d, 1); // xori d, d, 1
898 vset_VT_CMP(TOK_NE);
899 vtop->cmp_r = d;
900 break;
901 case TOK_NE:
902 case TOK_EQ:
903 o(0x33 | (d << 7) | (a << 15) | (b << 20) | (0x20 << 25)); // sub d, a, b
904 if (op == TOK_NE)
905 o(0x33 | (3 << 12) | (d << 7) | (0 << 15) | (d << 20)); // sltu d, x0, d == snez d,d
906 else
907 EI(0x13, 3, d, d, 1); // sltiu d, d, 1 == seqz d,d
908 vset_VT_CMP(TOK_NE);
909 vtop->cmp_r = d;
910 break;
914 ST_FUNC void gen_opi(int op)
916 gen_opil(op, 0);
919 ST_FUNC void gen_opl(int op)
921 gen_opil(op, 1);
924 ST_FUNC void gen_opf(int op)
926 int rs1, rs2, rd, dbl, invert;
927 if (vtop[0].type.t == VT_LDOUBLE) {
928 CType type = vtop[0].type;
929 int func = 0;
930 int cond = -1;
931 switch (op) {
932 case '*': func = TOK___multf3; break;
933 case '+': func = TOK___addtf3; break;
934 case '-': func = TOK___subtf3; break;
935 case '/': func = TOK___divtf3; break;
936 case TOK_EQ: func = TOK___eqtf2; cond = 1; break;
937 case TOK_NE: func = TOK___netf2; cond = 0; break;
938 case TOK_LT: func = TOK___lttf2; cond = 10; break;
939 case TOK_GE: func = TOK___getf2; cond = 11; break;
940 case TOK_LE: func = TOK___letf2; cond = 12; break;
941 case TOK_GT: func = TOK___gttf2; cond = 13; break;
942 default: assert(0); break;
944 vpush_global_sym(&func_old_type, func);
945 vrott(3);
946 gfunc_call(2);
947 vpushi(0);
948 vtop->r = REG_IRET;
949 vtop->r2 = cond < 0 ? TREG_R(1) : VT_CONST;
950 if (cond < 0)
951 vtop->type = type;
952 else {
953 vpushi(0);
954 gen_opil(op, 1);
956 return;
959 gv2(RC_FLOAT, RC_FLOAT);
960 assert(vtop->type.t == VT_DOUBLE || vtop->type.t == VT_FLOAT);
961 dbl = vtop->type.t == VT_DOUBLE;
962 rs1 = freg(vtop[-1].r);
963 rs2 = freg(vtop->r);
964 vtop--;
965 invert = 0;
966 switch(op) {
967 default:
968 assert(0);
969 case '+':
970 op = 0; // fadd
971 arithop:
972 rd = get_reg(RC_FLOAT);
973 vtop->r = rd;
974 rd = freg(rd);
975 o(0x53 | (rd << 7) | (rs1 << 15) | (rs2 << 20) | (7 << 12) | (dbl << 25) | (op << 27)); // fop.[sd] RD, RS1, RS2 (dyn rm)
976 break;
977 case '-':
978 op = 1; // fsub
979 goto arithop;
980 case '*':
981 op = 2; // fmul
982 goto arithop;
983 case '/':
984 op = 3; // fdiv
985 goto arithop;
986 case TOK_EQ:
987 op = 2; // EQ
988 cmpop:
989 rd = get_reg(RC_INT);
990 vtop->r = rd;
991 rd = ireg(rd);
992 o(0x53 | (rd << 7) | (rs1 << 15) | (rs2 << 20) | (op << 12) | (dbl << 25) | (0x14 << 27)); // fcmp.[sd] RD, RS1, RS2 (op == eq/lt/le)
993 if (invert)
994 EI(0x13, 4, rd, rd, 1); // xori RD, 1
995 break;
996 case TOK_NE:
997 invert = 1;
998 op = 2; // EQ
999 goto cmpop;
1000 case TOK_LT:
1001 op = 1; // LT
1002 goto cmpop;
1003 case TOK_LE:
1004 op = 0; // LE
1005 goto cmpop;
1006 case TOK_GT:
1007 op = 1; // LT
1008 rd = rs1, rs1 = rs2, rs2 = rd;
1009 goto cmpop;
1010 case TOK_GE:
1011 op = 0; // LE
1012 rd = rs1, rs1 = rs2, rs2 = rd;
1013 goto cmpop;
1017 ST_FUNC void gen_cvt_sxtw(void)
1019 /* XXX on risc-v the registers are usually sign-extended already.
1020 Let's try to not do anything here. */
1023 ST_FUNC void gen_cvt_itof(int t)
1025 int rr = ireg(gv(RC_INT)), dr;
1026 int u = vtop->type.t & VT_UNSIGNED;
1027 int l = (vtop->type.t & VT_BTYPE) == VT_LLONG;
1028 if (t == VT_LDOUBLE) {
1029 int func = l ?
1030 (u ? TOK___floatunditf : TOK___floatditf) :
1031 (u ? TOK___floatunsitf : TOK___floatsitf);
1032 vpush_global_sym(&func_old_type, func);
1033 vrott(2);
1034 gfunc_call(1);
1035 vpushi(0);
1036 vtop->type.t = t;
1037 vtop->r = REG_IRET;
1038 vtop->r2 = TREG_R(1);
1039 } else {
1040 vtop--;
1041 dr = get_reg(RC_FLOAT);
1042 vtop++;
1043 vtop->r = dr;
1044 dr = freg(dr);
1045 EIu(0x53, 7, dr, rr, ((0x68 | (t == VT_DOUBLE ? 1 : 0)) << 5) | (u ? 1 : 0) | (l ? 2 : 0)); // fcvt.[sd].[wl][u]
1049 ST_FUNC void gen_cvt_ftoi(int t)
1051 int ft = vtop->type.t & VT_BTYPE;
1052 int l = (t & VT_BTYPE) == VT_LLONG;
1053 int u = t & VT_UNSIGNED;
1054 if (ft == VT_LDOUBLE) {
1055 int func = l ?
1056 (u ? TOK___fixunstfdi : TOK___fixtfdi) :
1057 (u ? TOK___fixunstfsi : TOK___fixtfsi);
1058 vpush_global_sym(&func_old_type, func);
1059 vrott(2);
1060 gfunc_call(1);
1061 vpushi(0);
1062 vtop->type.t = t;
1063 vtop->r = REG_IRET;
1064 } else {
1065 int rr = freg(gv(RC_FLOAT)), dr;
1066 vtop--;
1067 dr = get_reg(RC_INT);
1068 vtop++;
1069 vtop->r = dr;
1070 dr = ireg(dr);
1071 EIu(0x53, 1, dr, rr, ((0x60 | (ft == VT_DOUBLE ? 1 : 0)) << 5) | (u ? 1 : 0) | (l ? 2 : 0)); // fcvt.[wl][u].[sd] rtz
1075 ST_FUNC void gen_cvt_ftof(int dt)
1077 int st = vtop->type.t & VT_BTYPE, rs, rd;
1078 dt &= VT_BTYPE;
1079 if (st == dt)
1080 return;
1081 if (dt == VT_LDOUBLE || st == VT_LDOUBLE) {
1082 int func = (dt == VT_LDOUBLE) ?
1083 (st == VT_FLOAT ? TOK___extendsftf2 : TOK___extenddftf2) :
1084 (dt == VT_FLOAT ? TOK___trunctfsf2 : TOK___trunctfdf2);
1085 /* We can't use gfunc_call, as func_old_type works like vararg
1086 functions, and on riscv unnamed float args are passed like
1087 integers. But we really need them in the float argument registers
1088 for extendsftf2/extenddftf2. So, do it explicitely. */
1089 save_regs(1);
1090 if (dt == VT_LDOUBLE)
1091 gv(RC_F(0));
1092 else {
1093 gv(RC_R(0));
1094 assert(vtop->r2 < 7);
1095 if (vtop->r2 != 1 + vtop->r) {
1096 EI(0x13, 0, ireg(vtop->r) + 1, ireg(vtop->r2), 0); // mv Ra+1, RR2
1097 vtop->r2 = 1 + vtop->r;
1100 vpush_global_sym(&func_old_type, func);
1101 gcall_or_jmp(1);
1102 vtop -= 2;
1103 vpushi(0);
1104 vtop->type.t = dt;
1105 if (dt == VT_LDOUBLE)
1106 vtop->r = REG_IRET, vtop->r2 = REG_IRET+1;
1107 else
1108 vtop->r = REG_FRET;
1109 } else {
1110 assert (dt == VT_FLOAT || dt == VT_DOUBLE);
1111 assert (st == VT_FLOAT || st == VT_DOUBLE);
1112 rs = gv(RC_FLOAT);
1113 rd = get_reg(RC_FLOAT);
1114 if (dt == VT_DOUBLE)
1115 EI(0x53, 7, freg(rd), freg(rs), 0x21 << 5); // fcvt.d.s RD, RS (dyn rm)
1116 else
1117 EI(0x53, 7, freg(rd), freg(rs), (0x20 << 5) | 1); // fcvt.s.d RD, RS
1118 vtop->r = rd;
1122 ST_FUNC void ggoto(void)
1124 gcall_or_jmp(0);
1125 vtop--;
1128 ST_FUNC void gen_vla_sp_save(int addr)
1130 ES(0x23, 3, 8, 2, addr); // sd sp, fc(s0)
1133 ST_FUNC void gen_vla_sp_restore(int addr)
1135 EI(0x03, 3, 2, 8, addr); // ld sp, fc(s0)
1138 ST_FUNC void gen_vla_alloc(CType *type, int align)
1140 int rr = ireg(gv(RC_INT));
1141 EI(0x13, 0, rr, rr, 15); // addi RR, RR, 15
1142 EI(0x13, 7, rr, rr, -16); // andi, RR, RR, -16
1143 o(0x33 | (2 << 7) | (2 << 15) | (rr << 20) | (0x20 << 25)); //sub sp, sp, rr
1144 vpop();
1146 #endif