Add a one-deep CC-cache for condition code setting and usage.
[smatch.git] / example.c
blobf77b4aab91b455439cffac191a90e6da556ae357
1 /*
2 * Example of how to write a compiler with sparse
3 */
4 #include <stdio.h>
5 #include <stdlib.h>
6 #include <stdarg.h>
7 #include <assert.h>
9 #include "symbol.h"
10 #include "expression.h"
11 #include "linearize.h"
12 #include "flow.h"
13 #include "storage.h"
15 static const char* opcodes[] = {
16 [OP_BADOP] = "bad_op",
18 /* Fn entrypoint */
19 [OP_ENTRY] = "<entry-point>",
21 /* Terminator */
22 [OP_RET] = "ret",
23 [OP_BR] = "br",
24 [OP_SWITCH] = "switch",
25 [OP_INVOKE] = "invoke",
26 [OP_COMPUTEDGOTO] = "jmp *",
27 [OP_UNWIND] = "unwind",
29 /* Binary */
30 [OP_ADD] = "add",
31 [OP_SUB] = "sub",
32 [OP_MUL] = "mul",
33 [OP_DIV] = "div",
34 [OP_MOD] = "mod",
35 [OP_SHL] = "shl",
36 [OP_SHR] = "shr",
38 /* Logical */
39 [OP_AND] = "and",
40 [OP_OR] = "or",
41 [OP_XOR] = "xor",
42 [OP_AND_BOOL] = "and-bool",
43 [OP_OR_BOOL] = "or-bool",
45 /* Binary comparison */
46 [OP_SET_EQ] = "seteq",
47 [OP_SET_NE] = "setne",
48 [OP_SET_LE] = "setle",
49 [OP_SET_GE] = "setge",
50 [OP_SET_LT] = "setlt",
51 [OP_SET_GT] = "setgt",
52 [OP_SET_B] = "setb",
53 [OP_SET_A] = "seta",
54 [OP_SET_BE] = "setbe",
55 [OP_SET_AE] = "setae",
57 /* Uni */
58 [OP_NOT] = "not",
59 [OP_NEG] = "neg",
61 /* Special three-input */
62 [OP_SEL] = "select",
64 /* Memory */
65 [OP_MALLOC] = "malloc",
66 [OP_FREE] = "free",
67 [OP_ALLOCA] = "alloca",
68 [OP_LOAD] = "load",
69 [OP_STORE] = "store",
70 [OP_SETVAL] = "set",
71 [OP_GET_ELEMENT_PTR] = "getelem",
73 /* Other */
74 [OP_PHI] = "phi",
75 [OP_PHISOURCE] = "phisrc",
76 [OP_CAST] = "cast",
77 [OP_PTRCAST] = "ptrcast",
78 [OP_CALL] = "call",
79 [OP_VANEXT] = "va_next",
80 [OP_VAARG] = "va_arg",
81 [OP_SLICE] = "slice",
82 [OP_SNOP] = "snop",
83 [OP_LNOP] = "lnop",
84 [OP_NOP] = "nop",
85 [OP_DEATHNOTE] = "dead",
86 [OP_ASM] = "asm",
88 /* Sparse tagging (line numbers, context, whatever) */
89 [OP_CONTEXT] = "context",
92 struct hardreg {
93 const char *name;
94 struct pseudo_list *contains;
95 unsigned busy:16,
96 dead:8,
97 used:1;
100 #define TAG_DEAD 1
101 #define TAG_DIRTY 2
103 /* Our "switch" generation is very very stupid. */
104 #define SWITCH_REG (1)
106 static void output_bb(struct basic_block *bb, unsigned long generation);
109 * We only know about the caller-clobbered registers
110 * right now.
112 static struct hardreg hardregs[] = {
113 { .name = "%eax" },
114 { .name = "%edx" },
115 { .name = "%ecx" },
116 { .name = "%ebx" },
117 { .name = "%esi" },
118 { .name = "%edi" },
120 #define REGNO (sizeof(hardregs)/sizeof(struct hardreg))
122 struct bb_state {
123 struct position pos;
124 unsigned long stack_offset;
125 struct storage_hash_list *inputs;
126 struct storage_hash_list *outputs;
127 struct storage_hash_list *internal;
129 /* CC cache.. */
130 int cc_opcode, cc_dead;
131 pseudo_t cc_target;
134 static struct storage_hash *find_storage_hash(pseudo_t pseudo, struct storage_hash_list *list)
136 struct storage_hash *entry;
137 FOR_EACH_PTR(list, entry) {
138 if (entry->pseudo == pseudo)
139 return entry;
140 } END_FOR_EACH_PTR(entry);
141 return NULL;
144 static struct storage_hash *find_or_create_hash(pseudo_t pseudo, struct storage_hash_list **listp)
146 struct storage_hash *entry;
148 entry = find_storage_hash(pseudo, *listp);
149 if (!entry) {
150 entry = alloc_storage_hash(alloc_storage());
151 entry->pseudo = pseudo;
152 add_ptr_list(listp, entry);
154 return entry;
157 /* Eventually we should just build it up in memory */
158 static void output_line(struct bb_state *state, const char *fmt, ...)
160 va_list args;
162 va_start(args, fmt);
163 vprintf(fmt, args);
164 va_end(args);
167 static void output_label(struct bb_state *state, const char *fmt, ...)
169 static char buffer[512];
170 va_list args;
172 va_start(args, fmt);
173 vsnprintf(buffer, sizeof(buffer), fmt, args);
174 va_end(args);
176 output_line(state, "%s:\n", buffer);
179 static void output_insn(struct bb_state *state, const char *fmt, ...)
181 static char buffer[512];
182 va_list args;
184 va_start(args, fmt);
185 vsnprintf(buffer, sizeof(buffer), fmt, args);
186 va_end(args);
188 output_line(state, "\t%s\n", buffer);
191 static void output_comment(struct bb_state *state, const char *fmt, ...)
193 static char buffer[512];
194 va_list args;
196 if (!verbose)
197 return;
198 va_start(args, fmt);
199 vsnprintf(buffer, sizeof(buffer), fmt, args);
200 va_end(args);
202 output_line(state, "\t# %s\n", buffer);
205 static const char *show_memop(struct storage *storage)
207 static char buffer[1000];
209 if (!storage)
210 return "undef";
211 switch (storage->type) {
212 case REG_FRAME:
213 sprintf(buffer, "%d(FP)", storage->offset);
214 break;
215 case REG_STACK:
216 sprintf(buffer, "%d(SP)", storage->offset);
217 break;
218 case REG_REG:
219 return hardregs[storage->regno].name;
220 default:
221 return show_storage(storage);
223 return buffer;
226 static void alloc_stack(struct bb_state *state, struct storage *storage)
228 storage->type = REG_STACK;
229 storage->offset = state->stack_offset;
230 state->stack_offset += 4;
234 * Can we re-generate the pseudo, so that we don't need to
235 * flush it to memory? We can regenerate:
236 * - immediates and symbol addresses
237 * - pseudos we got as input in non-registers
238 * - pseudos we've already saved off earlier..
240 static int can_regenerate(struct bb_state *state, pseudo_t pseudo)
242 struct storage_hash *in;
244 switch (pseudo->type) {
245 case PSEUDO_VAL:
246 case PSEUDO_SYM:
247 return 1;
249 default:
250 in = find_storage_hash(pseudo, state->inputs);
251 if (in && in->storage->type != REG_REG)
252 return 1;
253 in = find_storage_hash(pseudo, state->internal);
254 if (in)
255 return 1;
257 return 0;
260 static void flush_one_pseudo(struct bb_state *state, struct hardreg *hardreg, pseudo_t pseudo)
262 struct storage_hash *out;
263 struct storage *storage;
265 if (can_regenerate(state, pseudo))
266 return;
268 output_comment(state, "flushing %s from %s", show_pseudo(pseudo), hardreg->name);
269 out = find_storage_hash(pseudo, state->internal);
270 if (!out) {
271 out = find_storage_hash(pseudo, state->outputs);
272 if (!out)
273 out = find_or_create_hash(pseudo, &state->internal);
275 storage = out->storage;
276 switch (storage->type) {
277 default:
279 * Aieee - the next user wants it in a register, but we
280 * need to flush it to memory in between. Which means that
281 * we need to allocate an internal one, dammit..
283 out = find_or_create_hash(pseudo, &state->internal);
284 storage = out->storage;
285 /* Fall through */
286 case REG_UDEF:
287 alloc_stack(state, storage);
288 /* Fall through */
289 case REG_STACK:
290 output_insn(state, "movl %s,%s", hardreg->name, show_memop(storage));
291 break;
295 /* Flush a hardreg out to the storage it has.. */
296 static void flush_reg(struct bb_state *state, struct hardreg *hardreg)
298 pseudo_t pseudo;
300 if (!hardreg->busy)
301 return;
302 hardreg->busy = 0;
303 hardreg->dead = 0;
304 hardreg->used = 1;
305 FOR_EACH_PTR(hardreg->contains, pseudo) {
306 if (CURRENT_TAG(pseudo) & TAG_DEAD)
307 continue;
308 if (!(CURRENT_TAG(pseudo) & TAG_DIRTY))
309 continue;
310 flush_one_pseudo(state, hardreg, pseudo);
311 } END_FOR_EACH_PTR(pseudo);
312 free_ptr_list(&hardreg->contains);
315 static struct storage_hash *find_pseudo_storage(struct bb_state *state, pseudo_t pseudo, struct hardreg *reg)
317 struct storage_hash *src;
319 src = find_storage_hash(pseudo, state->internal);
320 if (!src) {
321 src = find_storage_hash(pseudo, state->inputs);
322 if (!src) {
323 src = find_storage_hash(pseudo, state->outputs);
324 /* Undefined? Screw it! */
325 if (!src)
326 return NULL;
329 * If we found output storage, it had better be local stack
330 * that we flushed to earlier..
332 if (src->storage->type != REG_STACK)
333 return NULL;
338 * Incoming pseudo with out any pre-set storage allocation?
339 * We can make up our own, and obviously prefer to get it
340 * in the register we already selected (if it hasn't been
341 * used yet).
343 if (src->storage->type == REG_UDEF) {
344 if (reg && !reg->used) {
345 src->storage->type = REG_REG;
346 src->storage->regno = reg - hardregs;
347 } else
348 alloc_stack(state, src->storage);
350 return src;
353 static void mark_reg_dead(struct bb_state *state, pseudo_t pseudo, struct hardreg *reg)
355 pseudo_t p;
357 FOR_EACH_PTR(reg->contains, p) {
358 if (p != pseudo)
359 continue;
360 if (CURRENT_TAG(p) & TAG_DEAD)
361 continue;
362 output_comment(state, "marking pseudo %s in reg %s dead", show_pseudo(pseudo), reg->name);
363 TAG_CURRENT(p, TAG_DEAD);
364 reg->dead++;
365 } END_FOR_EACH_PTR(p);
368 static void add_pseudo_reg(struct bb_state *state, pseudo_t pseudo, struct hardreg *reg)
370 output_comment(state, "added pseudo %s to reg %s", show_pseudo(pseudo), reg->name);
371 add_ptr_list_tag(&reg->contains, pseudo, TAG_DIRTY);
372 reg->busy++;
375 static int last_reg;
377 static struct hardreg *preferred_reg(struct bb_state *state, pseudo_t target)
379 struct storage_hash *dst;
381 dst = find_storage_hash(target, state->outputs);
382 if (dst) {
383 struct storage *storage = dst->storage;
384 if (storage->type == REG_REG)
385 return hardregs + storage->regno;
387 return NULL;
390 static struct hardreg *empty_reg(struct bb_state *state)
392 int i;
393 struct hardreg *reg = hardregs;
395 for (i = 0; i < REGNO; i++, reg++) {
396 if (!reg->busy)
397 return reg;
399 return NULL;
402 static struct hardreg *target_reg(struct bb_state *state, pseudo_t pseudo, pseudo_t target)
404 int i;
405 struct hardreg *reg;
407 /* First, see if we have a preferred target register.. */
408 reg = preferred_reg(state, target);
409 if (reg && !reg->busy)
410 goto found;
412 reg = empty_reg(state);
413 if (reg)
414 goto found;
416 i = last_reg+1;
417 if (i >= REGNO)
418 i = 0;
419 last_reg = i;
420 reg = hardregs + i;
421 flush_reg(state, reg);
423 found:
424 add_pseudo_reg(state, pseudo, reg);
425 return reg;
428 static struct hardreg *find_in_reg(struct bb_state *state, pseudo_t pseudo)
430 int i;
431 struct hardreg *reg;
433 for (i = 0; i < REGNO; i++) {
434 pseudo_t p;
436 reg = hardregs + i;
437 FOR_EACH_PTR(reg->contains, p) {
438 if (p == pseudo) {
439 last_reg = i;
440 output_comment(state, "found pseudo %s in reg %s", show_pseudo(pseudo), reg->name);
441 return reg;
443 } END_FOR_EACH_PTR(p);
445 return NULL;
448 static void flush_cc_cache_to_reg(struct bb_state *state, pseudo_t pseudo, struct hardreg *reg)
450 int opcode = state->cc_opcode;
452 state->cc_opcode = 0;
453 state->cc_target = NULL;
454 output_insn(state, "%s %s", opcodes[opcode], reg->name);
457 static void flush_cc_cache(struct bb_state *state)
459 pseudo_t pseudo = state->cc_target;
461 if (pseudo) {
462 struct hardreg *dst;
464 state->cc_target = NULL;
466 if (!state->cc_dead) {
467 dst = target_reg(state, pseudo, pseudo);
468 flush_cc_cache_to_reg(state, pseudo, dst);
473 static void add_cc_cache(struct bb_state *state, int opcode, pseudo_t pseudo)
475 assert(!state->cc_target);
476 state->cc_target = pseudo;
477 state->cc_opcode = opcode;
478 state->cc_dead = 0;
479 output_comment(state, "caching %s", opcodes[opcode]);
482 /* Fill a hardreg with the pseudo it has */
483 static struct hardreg *fill_reg(struct bb_state *state, struct hardreg *hardreg, pseudo_t pseudo)
485 struct storage_hash *src;
486 struct instruction *def;
488 if (state->cc_target == pseudo) {
489 flush_cc_cache_to_reg(state, pseudo, hardreg);
490 return hardreg;
493 switch (pseudo->type) {
494 case PSEUDO_VAL:
495 output_insn(state, "movl $%lld,%s", pseudo->value, hardreg->name);
496 break;
497 case PSEUDO_SYM:
498 output_insn(state, "movl $<%s>,%s", show_pseudo(pseudo), hardreg->name);
499 break;
500 case PSEUDO_ARG:
501 case PSEUDO_REG:
502 def = pseudo->def;
503 if (def->opcode == OP_SETVAL) {
504 output_insn(state, "movl $<%s>,%s", show_pseudo(def->symbol), hardreg->name);
505 break;
507 src = find_pseudo_storage(state, pseudo, hardreg);
508 if (!src)
509 break;
510 if (src->flags & TAG_DEAD)
511 mark_reg_dead(state, pseudo, hardreg);
512 output_insn(state, "mov.%d %s,%s", 32, show_memop(src->storage), hardreg->name);
513 break;
514 default:
515 output_insn(state, "reload %s from %s", hardreg->name, show_pseudo(pseudo));
516 break;
518 return hardreg;
521 static struct hardreg *getreg(struct bb_state *state, pseudo_t pseudo, pseudo_t target)
523 struct hardreg *reg;
525 reg = find_in_reg(state, pseudo);
526 if (reg)
527 return reg;
528 reg = target_reg(state, pseudo, target);
529 return fill_reg(state, reg, pseudo);
532 static struct hardreg *copy_reg(struct bb_state *state, struct hardreg *src, pseudo_t target)
534 int i;
535 struct hardreg *reg;
537 if (!src->busy)
538 return src;
540 reg = preferred_reg(state, target);
541 if (reg && !reg->busy) {
542 output_comment(state, "copying %s to preferred target %s", show_pseudo(target), reg->name);
543 output_insn(state, "movl %s,%s", src->name, reg->name);
544 return reg;
547 for (i = 0; i < REGNO; i++) {
548 struct hardreg *reg = hardregs + i;
549 if (!reg->busy) {
550 output_comment(state, "copying %s to %s", show_pseudo(target), reg->name);
551 output_insn(state, "movl %s,%s", src->name, reg->name);
552 return reg;
556 flush_reg(state, src);
557 return src;
560 static const char *generic(struct bb_state *state, pseudo_t pseudo)
562 struct hardreg *reg;
563 struct storage_hash *src;
565 switch (pseudo->type) {
566 case PSEUDO_SYM:
567 case PSEUDO_VAL:
568 return show_pseudo(pseudo);
569 default:
570 reg = find_in_reg(state, pseudo);
571 if (reg)
572 return reg->name;
573 src = find_pseudo_storage(state, pseudo, NULL);
574 if (!src)
575 return "undef";
576 return show_memop(src->storage);
580 static const char *address(struct bb_state *state, struct instruction *memop)
582 struct symbol *sym;
583 struct hardreg *base;
584 static char buffer[100];
585 pseudo_t addr = memop->src;
587 switch(addr->type) {
588 case PSEUDO_SYM:
589 sym = addr->sym;
590 if (sym->ctype.modifiers & MOD_NONLOCAL) {
591 sprintf(buffer, "%s+%d", show_ident(sym->ident), memop->offset);
592 return buffer;
594 sprintf(buffer, "%d+%s(SP)", memop->offset, show_ident(sym->ident));
595 return buffer;
596 default:
597 base = getreg(state, addr, NULL);
598 sprintf(buffer, "%d(%s)", memop->offset, base->name);
599 return buffer;
603 static const char *reg_or_imm(struct bb_state *state, pseudo_t pseudo)
605 switch(pseudo->type) {
606 case PSEUDO_VAL:
607 return show_pseudo(pseudo);
608 default:
609 return getreg(state, pseudo, NULL)->name;
613 static void kill_dead_reg(struct hardreg *reg)
615 if (reg->dead) {
616 pseudo_t p;
618 FOR_EACH_PTR(reg->contains, p) {
619 if (CURRENT_TAG(p) & TAG_DEAD) {
620 DELETE_CURRENT_PTR(p);
621 reg->busy--;
622 reg->dead--;
624 } END_FOR_EACH_PTR(p);
625 PACK_PTR_LIST(&reg->contains);
626 assert(!reg->dead);
630 static struct hardreg *target_copy_reg(struct bb_state *state, struct hardreg *src, pseudo_t target)
632 kill_dead_reg(src);
633 return copy_reg(state, src, target);
636 static void do_binop(struct bb_state *state, struct instruction *insn, pseudo_t val1, pseudo_t val2)
638 const char *op = opcodes[insn->opcode];
639 struct hardreg *src = getreg(state, val1, insn->target);
640 const char *src2 = generic(state, val2);
641 struct hardreg *dst;
643 dst = target_copy_reg(state, src, insn->target);
644 output_insn(state, "%s.%d %s,%s", op, insn->size, src2, dst->name);
645 add_pseudo_reg(state, insn->target, dst);
648 static void generate_binop(struct bb_state *state, struct instruction *insn)
650 flush_cc_cache(state);
651 do_binop(state, insn, insn->src1, insn->src2);
654 static int is_dead_reg(struct bb_state *state, pseudo_t pseudo, struct hardreg *reg)
656 pseudo_t p;
657 FOR_EACH_PTR(reg->contains, p) {
658 if (p == pseudo)
659 return CURRENT_TAG(p) & TAG_DEAD;
660 } END_FOR_EACH_PTR(p);
661 return 0;
665 * Commutative binops are much more flexible, since we can switch the
666 * sources around to satisfy the target register, or to avoid having
667 * to load one of them into a register..
669 static void generate_commutative_binop(struct bb_state *state, struct instruction *insn)
671 pseudo_t src1, src2;
672 struct hardreg *reg1, *reg2;
674 flush_cc_cache(state);
675 src1 = insn->src1;
676 src2 = insn->src2;
677 reg2 = find_in_reg(state, src2);
678 if (!reg2)
679 goto dont_switch;
680 reg1 = find_in_reg(state, src1);
681 if (!reg1)
682 goto do_switch;
683 if (!is_dead_reg(state, src2, reg2))
684 goto dont_switch;
685 if (!is_dead_reg(state, src1, reg1))
686 goto do_switch;
688 /* Both are dead. Is one preferrable? */
689 if (reg2 != preferred_reg(state, insn->target))
690 goto dont_switch;
692 do_switch:
693 src1 = src2;
694 src2 = insn->src1;
695 dont_switch:
696 do_binop(state, insn, src1, src2);
700 * This marks a pseudo dead. It still stays on the hardreg list (the hardreg
701 * still has its value), but it's scheduled to be killed after the next
702 * "sequence point" when we call "kill_read_pseudos()"
704 static void mark_pseudo_dead(struct bb_state *state, pseudo_t pseudo)
706 int i;
707 struct storage_hash *src;
709 if (state->cc_target == pseudo)
710 state->cc_dead = 1;
711 src = find_pseudo_storage(state, pseudo, NULL);
712 if (src)
713 src->flags |= TAG_DEAD;
714 for (i = 0; i < REGNO; i++)
715 mark_reg_dead(state, pseudo, hardregs + i);
718 static void kill_dead_pseudos(struct bb_state *state)
720 int i;
722 for (i = 0; i < REGNO; i++) {
723 kill_dead_reg(hardregs + i);
728 * A PHI source can define a pseudo that we already
729 * have in another register. We need to invalidate the
730 * old register so that we don't end up with the same
731 * pseudo in "two places".
733 static void remove_pseudo_reg(struct bb_state *state, pseudo_t pseudo)
735 int i;
737 output_comment(state, "pseudo %s died", show_pseudo(pseudo));
738 for (i = 0; i < REGNO; i++) {
739 struct hardreg *reg = hardregs + i;
740 pseudo_t p;
741 FOR_EACH_PTR(reg->contains, p) {
742 if (p != pseudo)
743 continue;
744 if (CURRENT_TAG(p) & TAG_DEAD)
745 reg->dead--;
746 reg->busy--;
747 DELETE_CURRENT_PTR(p);
748 output_comment(state, "removed pseudo %s from reg %s", show_pseudo(pseudo), reg->name);
749 } END_FOR_EACH_PTR(p);
750 PACK_PTR_LIST(&reg->contains);
754 static void generate_store(struct instruction *insn, struct bb_state *state)
756 output_insn(state, "mov.%d %s,%s", insn->size, reg_or_imm(state, insn->target), address(state, insn));
759 static void generate_load(struct instruction *insn, struct bb_state *state)
761 const char *input = address(state, insn);
762 struct hardreg *dst;
764 kill_dead_pseudos(state);
765 dst = target_reg(state, insn->target, NULL);
766 output_insn(state, "mov.%d %s,%s", insn->size, input, dst->name);
769 static void generate_phisource(struct instruction *insn, struct bb_state *state)
771 struct instruction *user;
772 struct hardreg *reg;
774 /* Mark all the target pseudos dead first */
775 FOR_EACH_PTR(insn->phi_users, user) {
776 mark_pseudo_dead(state, user->target);
777 } END_FOR_EACH_PTR(user);
779 reg = NULL;
780 FOR_EACH_PTR(insn->phi_users, user) {
781 if (!reg)
782 reg = getreg(state, insn->phi_src, user->target);
783 remove_pseudo_reg(state, user->target);
784 add_pseudo_reg(state, user->target, reg);
785 } END_FOR_EACH_PTR(user);
788 static void generate_cast(struct bb_state *state, struct instruction *insn)
790 struct hardreg *src = getreg(state, insn->src, insn->target);
791 struct hardreg *dst;
792 unsigned int old = insn->orig_type ? insn->orig_type->bit_size : 0;
793 unsigned int new = insn->size;
796 * Cast to smaller type? Ignore the high bits, we
797 * just keep both pseudos in the same register.
799 if (old >= new) {
800 add_pseudo_reg(state, insn->target, src);
801 return;
804 dst = target_copy_reg(state, src, insn->target);
806 if (insn->orig_type && (insn->orig_type->ctype.modifiers & MOD_SIGNED)) {
807 output_insn(state, "sext.%d.%d %s", old, new, dst->name);
808 } else {
809 unsigned long long mask;
810 mask = ~(~0ULL << old);
811 mask &= ~(~0ULL << new);
812 output_insn(state, "andl.%d $%#llx,%s", insn->size, mask, dst->name);
814 add_pseudo_reg(state, insn->target, dst);
817 static void generate_output_storage(struct bb_state *state);
819 static void generate_branch(struct bb_state *state, struct instruction *br)
821 const char *branch = "jXXX";
822 struct basic_block *target;
824 if (br->cond) {
825 if (state->cc_target == br->cond) {
826 static const char *branches[] = {
827 [OP_SET_EQ] = "je",
828 [OP_SET_NE] = "jne",
829 [OP_SET_LE] = "jle",
830 [OP_SET_GE] = "jge",
831 [OP_SET_LT] = "jlt",
832 [OP_SET_GT] = "jgt",
833 [OP_SET_B] = "jb",
834 [OP_SET_A] = "ja",
835 [OP_SET_BE] = "jbe",
836 [OP_SET_AE] = "jae"
838 branch = branches[state->cc_opcode];
839 } else {
840 struct hardreg *reg = getreg(state, br->cond, NULL);
841 output_insn(state, "testl %s,%s", reg->name, reg->name);
842 branch = "jne";
845 generate_output_storage(state);
846 target = br->bb_true;
847 if (br->cond) {
848 output_insn(state, "%s .L%p", branch, target);
849 target = br->bb_false;
851 output_insn(state, "jmp .L%p", target);
854 /* We've made sure that there is a dummy reg live for the output */
855 static void generate_switch(struct bb_state *state, struct instruction *insn)
857 struct hardreg *reg = hardregs + SWITCH_REG;
859 generate_output_storage(state);
860 output_insn(state, "switch on %s", reg->name);
861 output_insn(state, "unimplemented: %s", show_instruction(insn));
864 static void generate_ret(struct bb_state *state, struct instruction *ret)
866 if (ret->src && ret->src != VOID) {
867 struct hardreg *wants = hardregs+0;
868 struct hardreg *reg = getreg(state, ret->src, NULL);
869 if (reg != wants)
870 output_insn(state, "movl %s,%s", reg->name, wants->name);
872 output_insn(state, "ret");
876 * Fake "call" linearization just as a taster..
878 static void generate_call(struct bb_state *state, struct instruction *insn)
880 int offset = 0;
881 pseudo_t arg;
883 FOR_EACH_PTR(insn->arguments, arg) {
884 output_insn(state, "pushl %s", generic(state, arg));
885 offset += 4;
886 } END_FOR_EACH_PTR(arg);
887 flush_reg(state, hardregs+0);
888 flush_reg(state, hardregs+1);
889 flush_reg(state, hardregs+2);
890 output_insn(state, "call %s", show_pseudo(insn->func));
891 if (offset)
892 output_insn(state, "addl $%d,%%esp", offset);
893 add_pseudo_reg(state, insn->target, hardregs+0);
896 static void generate_select(struct bb_state *state, struct instruction *insn)
898 struct hardreg *src1, *src2, *cond, *dst;
900 cond = getreg(state, insn->src1, NULL);
901 output_insn(state, "testl %s,%s", cond->name, cond->name);
903 src1 = getreg(state, insn->src2, NULL);
904 dst = copy_reg(state, src1, insn->target);
905 add_pseudo_reg(state, insn->target, dst);
906 src2 = getreg(state, insn->src3, insn->target);
907 output_insn(state, "sele %s,%s", src2->name, dst->name);
910 static void generate_asm_inputs(struct bb_state *state, struct asm_constraint_list *list)
912 struct asm_constraint *entry;
914 FOR_EACH_PTR(list, entry) {
915 const char *constraint = entry->constraint;
916 pseudo_t pseudo = entry->pseudo;
918 output_insn(state, "# asm input \"%s\": %s", constraint, show_pseudo(pseudo));
919 } END_FOR_EACH_PTR(entry);
922 static void generate_asm_outputs(struct bb_state *state, struct asm_constraint_list *list)
924 struct asm_constraint *entry;
926 FOR_EACH_PTR(list, entry) {
927 const char *constraint = entry->constraint;
928 pseudo_t pseudo = entry->pseudo;
930 while (*constraint == '=' || *constraint == '+')
931 constraint++;
933 output_insn(state, "# asm output \"%s\": %s", constraint, show_pseudo(pseudo));
934 } END_FOR_EACH_PTR(entry);
937 static void generate_asm(struct bb_state *state, struct instruction *insn)
939 generate_asm_inputs(state, insn->asm_rules->inputs);
940 output_insn(state, "ASM: %s", insn->string);
941 generate_asm_outputs(state, insn->asm_rules->outputs);
944 static void generate_compare(struct bb_state *state, struct instruction *insn)
946 struct hardreg *src;
947 const char *src2;
948 int opcode;
950 flush_cc_cache(state);
951 opcode = insn->opcode;
954 * We should try to switch these around if necessary,
955 * and update the opcode to match..
957 src = getreg(state, insn->src1, insn->target);
958 src2 = generic(state, insn->src2);
960 output_insn(state, "cmp.%d %s,%s", insn->size, src2, src->name);
962 add_cc_cache(state, opcode, insn->target);
965 static void generate_one_insn(struct instruction *insn, struct bb_state *state)
967 if (verbose)
968 output_comment(state, "%s", show_instruction(insn));
970 switch (insn->opcode) {
971 case OP_ENTRY: {
972 struct symbol *sym = insn->bb->ep->name;
973 const char *name = show_ident(sym->ident);
974 if (sym->ctype.modifiers & MOD_STATIC)
975 printf("\n\n%s:\n", name);
976 else
977 printf("\n\n.globl %s\n%s:\n", name, name);
978 break;
982 * OP_PHI doesn't actually generate any code. It has been
983 * done by the storage allocator and the OP_PHISOURCE.
985 case OP_PHI:
986 break;
988 case OP_PHISOURCE:
989 generate_phisource(insn, state);
990 break;
993 * OP_SETVAL likewise doesn't actually generate any
994 * code. On use, the "def" of the pseudo will be
995 * looked up.
997 case OP_SETVAL:
998 break;
1000 case OP_STORE:
1001 generate_store(insn, state);
1002 break;
1004 case OP_LOAD:
1005 generate_load(insn, state);
1006 break;
1008 case OP_DEATHNOTE:
1009 mark_pseudo_dead(state, insn->target);
1010 return;
1012 case OP_ADD: case OP_MUL:
1013 case OP_AND: case OP_OR: case OP_XOR:
1014 case OP_AND_BOOL: case OP_OR_BOOL:
1015 generate_commutative_binop(state, insn);
1016 break;
1018 case OP_SUB: case OP_DIV: case OP_MOD:
1019 case OP_SHL: case OP_SHR:
1020 generate_binop(state, insn);
1021 break;
1023 case OP_BINCMP ... OP_BINCMP_END:
1024 generate_compare(state, insn);
1025 break;
1027 case OP_CAST: case OP_PTRCAST:
1028 generate_cast(state, insn);
1029 break;
1031 case OP_SEL:
1032 generate_select(state, insn);
1033 break;
1035 case OP_BR:
1036 generate_branch(state, insn);
1037 break;
1039 case OP_SWITCH:
1040 generate_switch(state, insn);
1041 break;
1043 case OP_CALL:
1044 generate_call(state, insn);
1045 break;
1047 case OP_RET:
1048 generate_ret(state, insn);
1049 break;
1051 case OP_ASM:
1052 generate_asm(state, insn);
1053 break;
1055 default:
1056 output_insn(state, "unimplemented: %s", show_instruction(insn));
1057 break;
1059 kill_dead_pseudos(state);
1062 #define VERY_BUSY 1000
1063 #define REG_FIXED 2000
1065 static void write_reg_to_storage(struct bb_state *state, struct hardreg *reg, pseudo_t pseudo, struct storage *storage)
1067 int i;
1068 struct hardreg *out;
1070 switch (storage->type) {
1071 case REG_REG:
1072 out = hardregs + storage->regno;
1073 if (reg == out)
1074 return;
1075 output_insn(state, "movl %s,%s", reg->name, out->name);
1076 return;
1077 case REG_UDEF:
1078 if (reg->busy < VERY_BUSY) {
1079 storage->type = REG_REG;
1080 storage->regno = reg - hardregs;
1081 reg->busy = REG_FIXED;
1082 return;
1085 /* Try to find a non-busy register.. */
1086 for (i = 0; i < REGNO; i++) {
1087 out = hardregs + i;
1088 if (out->busy)
1089 continue;
1090 output_insn(state, "movl %s,%s", reg->name, out->name);
1091 storage->type = REG_REG;
1092 storage->regno = i;
1093 reg->busy = REG_FIXED;
1094 return;
1097 /* Fall back on stack allocation ... */
1098 alloc_stack(state, storage);
1099 /* Fallthroigh */
1100 default:
1101 output_insn(state, "movl %s,%s", reg->name, show_memop(storage));
1102 return;
1106 static void write_val_to_storage(struct bb_state *state, pseudo_t src, struct storage *storage)
1108 struct hardreg *out;
1110 switch (storage->type) {
1111 case REG_UDEF:
1112 alloc_stack(state, storage);
1113 default:
1114 output_insn(state, "movl %s,%s", show_pseudo(src), show_memop(storage));
1115 break;
1116 case REG_REG:
1117 out = hardregs + storage->regno;
1118 output_insn(state, "movl %s,%s", show_pseudo(src), out->name);
1122 static void fill_output(struct bb_state *state, pseudo_t pseudo, struct storage *out)
1124 int i;
1125 struct storage_hash *in;
1126 struct instruction *def;
1128 /* Is that pseudo a constant value? */
1129 switch (pseudo->type) {
1130 case PSEUDO_VAL:
1131 write_val_to_storage(state, pseudo, out);
1132 return;
1133 case PSEUDO_REG:
1134 def = pseudo->def;
1135 if (def->opcode == OP_SETVAL) {
1136 write_val_to_storage(state, def->symbol, out);
1137 return;
1139 default:
1140 break;
1143 /* See if we have that pseudo in a register.. */
1144 for (i = 0; i < REGNO; i++) {
1145 struct hardreg *reg = hardregs + i;
1146 pseudo_t p;
1148 FOR_EACH_PTR(reg->contains, p) {
1149 if (p == pseudo) {
1150 write_reg_to_storage(state, reg, pseudo, out);
1151 return;
1153 } END_FOR_EACH_PTR(p);
1156 /* Do we have it in another storage? */
1157 in = find_storage_hash(pseudo, state->internal);
1158 if (!in) {
1159 in = find_storage_hash(pseudo, state->inputs);
1160 /* Undefined? */
1161 if (!in)
1162 return;
1164 switch (out->type) {
1165 case REG_UDEF:
1166 *out = *in->storage;
1167 break;
1168 case REG_REG:
1169 output_insn(state, "movl %s,%s", show_memop(in->storage), hardregs[out->regno].name);
1170 break;
1171 default:
1172 if (out == in->storage)
1173 break;
1174 if (out->type == in->storage->type == out->regno == in->storage->regno)
1175 break;
1176 output_insn(state, "movl %s,%s", show_memop(in->storage), show_memop(out));
1177 break;
1179 return;
1182 static int final_pseudo_flush(struct bb_state *state, pseudo_t pseudo, struct hardreg *reg)
1184 struct storage_hash *hash;
1185 struct storage *out;
1186 struct hardreg *dst;
1189 * Since this pseudo is live at exit, we'd better have output
1190 * storage for it..
1192 hash = find_storage_hash(pseudo, state->outputs);
1193 if (!hash)
1194 return 1;
1195 out = hash->storage;
1197 /* If the output is in a register, try to get it there.. */
1198 if (out->type == REG_REG) {
1199 dst = hardregs + out->regno;
1201 * Two good cases: nobody is using the right register,
1202 * or we've already set it aside for output..
1204 if (!dst->busy || dst->busy > VERY_BUSY)
1205 goto copy_to_dst;
1207 /* Aiee. Try to keep it in a register.. */
1208 dst = empty_reg(state);
1209 if (dst)
1210 goto copy_to_dst;
1212 return 0;
1215 /* If the output is undefined, let's see if we can put it in a register.. */
1216 if (out->type == REG_UDEF) {
1217 dst = empty_reg(state);
1218 if (dst) {
1219 out->type = REG_REG;
1220 out->regno = dst - hardregs;
1221 goto copy_to_dst;
1223 /* Uhhuh. Not so good. No empty registers right now */
1224 return 0;
1227 /* If we know we need to flush it, just do so already .. */
1228 output_insn(state, "movl %s,%s", reg->name, show_memop(out));
1229 return 1;
1231 copy_to_dst:
1232 if (reg == dst)
1233 return 1;
1234 output_insn(state, "movl %s,%s", reg->name, dst->name);
1235 add_pseudo_reg(state, pseudo, dst);
1236 return 1;
1240 * This tries to make sure that we put all the pseudos that are
1241 * live on exit into the proper storage
1243 static void generate_output_storage(struct bb_state *state)
1245 struct storage_hash *entry;
1247 /* Go through the fixed outputs, making sure we have those regs free */
1248 FOR_EACH_PTR(state->outputs, entry) {
1249 struct storage *out = entry->storage;
1250 if (out->type == REG_REG) {
1251 struct hardreg *reg = hardregs + out->regno;
1252 pseudo_t p;
1253 int flushme = 0;
1255 reg->busy = REG_FIXED;
1256 FOR_EACH_PTR(reg->contains, p) {
1257 if (p == entry->pseudo) {
1258 flushme = -100;
1259 continue;
1261 if (CURRENT_TAG(p) & TAG_DEAD)
1262 continue;
1264 /* Try to write back the pseudo to where it should go ... */
1265 if (final_pseudo_flush(state, p, reg)) {
1266 DELETE_CURRENT_PTR(p);
1267 reg->busy--;
1268 continue;
1270 flushme++;
1271 } END_FOR_EACH_PTR(p);
1272 PACK_PTR_LIST(&reg->contains);
1273 if (flushme > 0)
1274 flush_reg(state, reg);
1276 } END_FOR_EACH_PTR(entry);
1278 FOR_EACH_PTR(state->outputs, entry) {
1279 fill_output(state, entry->pseudo, entry->storage);
1280 } END_FOR_EACH_PTR(entry);
1283 static void generate(struct basic_block *bb, struct bb_state *state)
1285 int i;
1286 struct storage_hash *entry;
1287 struct instruction *insn;
1289 for (i = 0; i < REGNO; i++) {
1290 free_ptr_list(&hardregs[i].contains);
1291 hardregs[i].busy = 0;
1292 hardregs[i].dead = 0;
1293 hardregs[i].used = 0;
1296 FOR_EACH_PTR(state->inputs, entry) {
1297 struct storage *storage = entry->storage;
1298 const char *name = show_storage(storage);
1299 output_comment(state, "incoming %s in %s", show_pseudo(entry->pseudo), name);
1300 if (storage->type == REG_REG) {
1301 int regno = storage->regno;
1302 add_pseudo_reg(state, entry->pseudo, hardregs + regno);
1303 name = hardregs[regno].name;
1305 } END_FOR_EACH_PTR(entry);
1307 output_label(state, ".L%p", bb);
1308 FOR_EACH_PTR(bb->insns, insn) {
1309 if (!insn->bb)
1310 continue;
1311 generate_one_insn(insn, state);
1312 } END_FOR_EACH_PTR(insn);
1314 if (verbose) {
1315 output_comment(state, "--- in ---");
1316 FOR_EACH_PTR(state->inputs, entry) {
1317 output_comment(state, "%s <- %s", show_pseudo(entry->pseudo), show_storage(entry->storage));
1318 } END_FOR_EACH_PTR(entry);
1319 output_comment(state, "--- spill ---");
1320 FOR_EACH_PTR(state->internal, entry) {
1321 output_comment(state, "%s <-> %s", show_pseudo(entry->pseudo), show_storage(entry->storage));
1322 } END_FOR_EACH_PTR(entry);
1323 output_comment(state, "--- out ---");
1324 FOR_EACH_PTR(state->outputs, entry) {
1325 output_comment(state, "%s -> %s", show_pseudo(entry->pseudo), show_storage(entry->storage));
1326 } END_FOR_EACH_PTR(entry);
1328 printf("\n");
1331 static void generate_list(struct basic_block_list *list, unsigned long generation)
1333 struct basic_block *bb;
1334 FOR_EACH_PTR(list, bb) {
1335 if (bb->generation == generation)
1336 continue;
1337 output_bb(bb, generation);
1338 } END_FOR_EACH_PTR(bb);
1341 static void output_bb(struct basic_block *bb, unsigned long generation)
1343 struct bb_state state;
1345 bb->generation = generation;
1347 /* Make sure all parents have been generated first */
1348 generate_list(bb->parents, generation);
1350 state.pos = bb->pos;
1351 state.inputs = gather_storage(bb, STOR_IN);
1352 state.outputs = gather_storage(bb, STOR_OUT);
1353 state.internal = NULL;
1354 state.stack_offset = 0;
1355 state.cc_opcode = 0;
1356 state.cc_target = NULL;
1358 generate(bb, &state);
1360 free_ptr_list(&state.inputs);
1361 free_ptr_list(&state.outputs);
1363 /* Generate all children... */
1364 generate_list(bb->children, generation);
1367 static void set_up_arch_entry(struct entrypoint *ep, struct instruction *entry)
1369 int i;
1370 pseudo_t arg;
1373 * We should set up argument sources here..
1375 * Things like "first three arguments in registers" etc
1376 * are all for this place.
1378 i = 0;
1379 FOR_EACH_PTR(entry->arg_list, arg) {
1380 struct storage *in = lookup_storage(entry->bb, arg, STOR_IN);
1381 if (!in) {
1382 in = alloc_storage();
1383 add_storage(in, entry->bb, arg, STOR_IN);
1385 if (i < 3) {
1386 in->type = REG_REG;
1387 in->regno = i;
1388 } else {
1389 in->type = REG_FRAME;
1390 in->offset = (i-3)*4;
1392 i++;
1393 } END_FOR_EACH_PTR(arg);
1397 * Set up storage information for "return"
1399 * Not strictly necessary, since the code generator will
1400 * certainly move the return value to the right register,
1401 * but it can help register allocation if the allocator
1402 * sees that the target register is going to return in %eax.
1404 static void set_up_arch_exit(struct basic_block *bb, struct instruction *ret)
1406 pseudo_t pseudo = ret->src;
1408 if (pseudo && pseudo != VOID) {
1409 struct storage *out = lookup_storage(bb, pseudo, STOR_OUT);
1410 if (!out) {
1411 out = alloc_storage();
1412 add_storage(out, bb, pseudo, STOR_OUT);
1414 out->type = REG_REG;
1415 out->regno = 0;
1420 * Set up dummy/silly output storage information for a switch
1421 * instruction. We need to make sure that a register is available
1422 * when we generate code for switch, so force that by creating
1423 * a dummy output rule.
1425 static void set_up_arch_switch(struct basic_block *bb, struct instruction *insn)
1427 pseudo_t pseudo = insn->cond;
1428 struct storage *out = lookup_storage(bb, pseudo, STOR_OUT);
1429 if (!out) {
1430 out = alloc_storage();
1431 add_storage(out, bb, pseudo, STOR_OUT);
1433 out->type = REG_REG;
1434 out->regno = SWITCH_REG;
1437 static void arch_set_up_storage(struct entrypoint *ep)
1439 struct basic_block *bb;
1441 /* Argument storage etc.. */
1442 set_up_arch_entry(ep, ep->entry);
1444 FOR_EACH_PTR(ep->bbs, bb) {
1445 struct instruction *insn = last_instruction(bb->insns);
1446 if (!insn)
1447 continue;
1448 switch (insn->opcode) {
1449 case OP_RET:
1450 set_up_arch_exit(bb, insn);
1451 break;
1452 case OP_SWITCH:
1453 set_up_arch_switch(bb, insn);
1454 break;
1455 default:
1456 /* nothing */;
1458 } END_FOR_EACH_PTR(bb);
1461 static void output(struct entrypoint *ep)
1463 unsigned long generation = ++bb_generation;
1465 last_reg = -1;
1467 /* Set up initial inter-bb storage links */
1468 set_up_storage(ep);
1470 /* Architecture-specific storage rules.. */
1471 arch_set_up_storage(ep);
1473 /* Show the results ... */
1474 output_bb(ep->entry->bb, generation);
1476 /* Clear the storage hashes for the next function.. */
1477 free_storage();
1480 static int compile(struct symbol_list *list)
1482 struct symbol *sym;
1483 FOR_EACH_PTR(list, sym) {
1484 struct entrypoint *ep;
1485 expand_symbol(sym);
1486 ep = linearize_symbol(sym);
1487 if (ep)
1488 output(ep);
1489 } END_FOR_EACH_PTR(sym);
1491 return 0;
1494 int main(int argc, char **argv)
1496 return compile(sparse(argc, argv));