2 * Example of how to write a compiler with sparse
11 #include "expression.h"
12 #include "linearize.h"
16 static const char* opcodes
[] = {
17 [OP_BADOP
] = "bad_op",
20 [OP_ENTRY
] = "<entry-point>",
25 [OP_SWITCH
] = "switch",
26 [OP_INVOKE
] = "invoke",
27 [OP_COMPUTEDGOTO
] = "jmp *",
28 [OP_UNWIND
] = "unwind",
43 [OP_AND_BOOL
] = "and-bool",
44 [OP_OR_BOOL
] = "or-bool",
46 /* Binary comparison */
47 [OP_SET_EQ
] = "seteq",
48 [OP_SET_NE
] = "setne",
49 [OP_SET_LE
] = "setle",
50 [OP_SET_GE
] = "setge",
51 [OP_SET_LT
] = "setlt",
52 [OP_SET_GT
] = "setgt",
55 [OP_SET_BE
] = "setbe",
56 [OP_SET_AE
] = "setae",
62 /* Special three-input */
66 [OP_MALLOC
] = "malloc",
68 [OP_ALLOCA
] = "alloca",
72 [OP_GET_ELEMENT_PTR
] = "getelem",
76 [OP_PHISOURCE
] = "phisrc",
78 [OP_PTRCAST
] = "ptrcast",
80 [OP_VANEXT
] = "va_next",
81 [OP_VAARG
] = "va_arg",
86 [OP_DEATHNOTE
] = "dead",
89 /* Sparse tagging (line numbers, context, whatever) */
90 [OP_CONTEXT
] = "context",
93 static int last_reg
, stack_offset
;
97 struct pseudo_list
*contains
;
106 /* Our "switch" generation is very very stupid. */
107 #define SWITCH_REG (1)
109 static void output_bb(struct basic_block
*bb
, unsigned long generation
);
112 * We only know about the caller-clobbered registers
115 static struct hardreg hardregs
[] = {
132 struct storage_hash_list
*inputs
;
133 struct storage_hash_list
*outputs
;
134 struct storage_hash_list
*internal
;
137 int cc_opcode
, cc_dead
;
155 struct /* OP_MEM and OP_ADDR */ {
159 struct hardreg
*base
;
160 struct hardreg
*index
;
165 static const char *show_op(struct bb_state
*state
, struct operand
*op
)
167 static char buf
[256][4];
172 nr
= (bufnr
+ 1) & 3;
180 return op
->reg
->name
;
182 sprintf(p
, "$%lld", op
->value
);
187 p
+= sprintf(p
, "%d", op
->offset
);
189 p
+= sprintf(p
, "%s%s",
190 op
->offset
? "+" : "",
191 show_ident(op
->sym
->ident
));
192 if (op
->base
|| op
->index
) {
193 p
+= sprintf(p
, "(%s%s%s",
194 op
->base
? op
->base
->name
: "",
195 (op
->base
&& op
->index
) ? "," : "",
196 op
->index
? op
->index
->name
: "");
198 p
+= sprintf(p
, ",%d", op
->scale
);
207 static struct storage_hash
*find_storage_hash(pseudo_t pseudo
, struct storage_hash_list
*list
)
209 struct storage_hash
*entry
;
210 FOR_EACH_PTR(list
, entry
) {
211 if (entry
->pseudo
== pseudo
)
213 } END_FOR_EACH_PTR(entry
);
217 static struct storage_hash
*find_or_create_hash(pseudo_t pseudo
, struct storage_hash_list
**listp
)
219 struct storage_hash
*entry
;
221 entry
= find_storage_hash(pseudo
, *listp
);
223 entry
= alloc_storage_hash(alloc_storage());
224 entry
->pseudo
= pseudo
;
225 add_ptr_list(listp
, entry
);
230 /* Eventually we should just build it up in memory */
231 static void output_line(struct bb_state
*state
, const char *fmt
, ...)
240 static void output_label(struct bb_state
*state
, const char *fmt
, ...)
242 static char buffer
[512];
246 vsnprintf(buffer
, sizeof(buffer
), fmt
, args
);
249 output_line(state
, "%s:\n", buffer
);
252 static void output_insn(struct bb_state
*state
, const char *fmt
, ...)
254 static char buffer
[512];
258 vsnprintf(buffer
, sizeof(buffer
), fmt
, args
);
261 output_line(state
, "\t%s\n", buffer
);
264 static void output_comment(struct bb_state
*state
, const char *fmt
, ...)
266 static char buffer
[512];
272 vsnprintf(buffer
, sizeof(buffer
), fmt
, args
);
275 output_line(state
, "\t# %s\n", buffer
);
278 static const char *show_memop(struct storage
*storage
)
280 static char buffer
[1000];
284 switch (storage
->type
) {
286 sprintf(buffer
, "%d(FP)", storage
->offset
);
289 sprintf(buffer
, "%d(SP)", storage
->offset
);
292 return hardregs
[storage
->regno
].name
;
294 return show_storage(storage
);
299 static void alloc_stack(struct bb_state
*state
, struct storage
*storage
)
301 storage
->type
= REG_STACK
;
302 storage
->offset
= stack_offset
;
307 * Can we re-generate the pseudo, so that we don't need to
308 * flush it to memory? We can regenerate:
309 * - immediates and symbol addresses
310 * - pseudos we got as input in non-registers
311 * - pseudos we've already saved off earlier..
313 static int can_regenerate(struct bb_state
*state
, pseudo_t pseudo
)
315 struct storage_hash
*in
;
317 switch (pseudo
->type
) {
323 in
= find_storage_hash(pseudo
, state
->inputs
);
324 if (in
&& in
->storage
->type
!= REG_REG
)
326 in
= find_storage_hash(pseudo
, state
->internal
);
333 static void flush_one_pseudo(struct bb_state
*state
, struct hardreg
*hardreg
, pseudo_t pseudo
)
335 struct storage_hash
*out
;
336 struct storage
*storage
;
338 if (can_regenerate(state
, pseudo
))
341 output_comment(state
, "flushing %s from %s", show_pseudo(pseudo
), hardreg
->name
);
342 out
= find_storage_hash(pseudo
, state
->internal
);
344 out
= find_storage_hash(pseudo
, state
->outputs
);
346 out
= find_or_create_hash(pseudo
, &state
->internal
);
348 storage
= out
->storage
;
349 switch (storage
->type
) {
352 * Aieee - the next user wants it in a register, but we
353 * need to flush it to memory in between. Which means that
354 * we need to allocate an internal one, dammit..
356 out
= find_or_create_hash(pseudo
, &state
->internal
);
357 storage
= out
->storage
;
360 alloc_stack(state
, storage
);
363 output_insn(state
, "movl %s,%s", hardreg
->name
, show_memop(storage
));
368 /* Flush a hardreg out to the storage it has.. */
369 static void flush_reg(struct bb_state
*state
, struct hardreg
*hardreg
)
378 FOR_EACH_PTR(hardreg
->contains
, pseudo
) {
379 if (CURRENT_TAG(pseudo
) & TAG_DEAD
)
381 if (!(CURRENT_TAG(pseudo
) & TAG_DIRTY
))
383 flush_one_pseudo(state
, hardreg
, pseudo
);
384 } END_FOR_EACH_PTR(pseudo
);
385 free_ptr_list(&hardreg
->contains
);
388 static struct storage_hash
*find_pseudo_storage(struct bb_state
*state
, pseudo_t pseudo
, struct hardreg
*reg
)
390 struct storage_hash
*src
;
392 src
= find_storage_hash(pseudo
, state
->internal
);
394 src
= find_storage_hash(pseudo
, state
->inputs
);
396 src
= find_storage_hash(pseudo
, state
->outputs
);
397 /* Undefined? Screw it! */
402 * If we found output storage, it had better be local stack
403 * that we flushed to earlier..
405 if (src
->storage
->type
!= REG_STACK
)
411 * Incoming pseudo with out any pre-set storage allocation?
412 * We can make up our own, and obviously prefer to get it
413 * in the register we already selected (if it hasn't been
416 if (src
->storage
->type
== REG_UDEF
) {
417 if (reg
&& !reg
->used
) {
418 src
->storage
->type
= REG_REG
;
419 src
->storage
->regno
= reg
- hardregs
;
422 alloc_stack(state
, src
->storage
);
427 static void mark_reg_dead(struct bb_state
*state
, pseudo_t pseudo
, struct hardreg
*reg
)
431 FOR_EACH_PTR(reg
->contains
, p
) {
434 if (CURRENT_TAG(p
) & TAG_DEAD
)
436 output_comment(state
, "marking pseudo %s in reg %s dead", show_pseudo(pseudo
), reg
->name
);
437 TAG_CURRENT(p
, TAG_DEAD
);
439 } END_FOR_EACH_PTR(p
);
442 static void add_pseudo_reg(struct bb_state
*state
, pseudo_t pseudo
, struct hardreg
*reg
)
444 output_comment(state
, "added pseudo %s to reg %s", show_pseudo(pseudo
), reg
->name
);
445 add_ptr_list_tag(®
->contains
, pseudo
, TAG_DIRTY
);
449 static struct hardreg
*preferred_reg(struct bb_state
*state
, pseudo_t target
)
451 struct storage_hash
*dst
;
453 dst
= find_storage_hash(target
, state
->outputs
);
455 struct storage
*storage
= dst
->storage
;
456 if (storage
->type
== REG_REG
)
457 return hardregs
+ storage
->regno
;
462 static struct hardreg
*empty_reg(struct bb_state
*state
)
465 struct hardreg
*reg
= hardregs
;
467 for (i
= 0; i
< REGNO
; i
++, reg
++) {
474 static struct hardreg
*target_reg(struct bb_state
*state
, pseudo_t pseudo
, pseudo_t target
)
479 /* First, see if we have a preferred target register.. */
480 reg
= preferred_reg(state
, target
);
481 if (reg
&& !reg
->busy
)
484 reg
= empty_reg(state
);
493 flush_reg(state
, reg
);
496 add_pseudo_reg(state
, pseudo
, reg
);
500 static struct hardreg
*find_in_reg(struct bb_state
*state
, pseudo_t pseudo
)
505 for (i
= 0; i
< REGNO
; i
++) {
509 FOR_EACH_PTR(reg
->contains
, p
) {
512 output_comment(state
, "found pseudo %s in reg %s", show_pseudo(pseudo
), reg
->name
);
515 } END_FOR_EACH_PTR(p
);
520 static void flush_cc_cache_to_reg(struct bb_state
*state
, pseudo_t pseudo
, struct hardreg
*reg
)
522 int opcode
= state
->cc_opcode
;
524 state
->cc_opcode
= 0;
525 state
->cc_target
= NULL
;
526 output_insn(state
, "%s %s", opcodes
[opcode
], reg
->name
);
529 static void flush_cc_cache(struct bb_state
*state
)
531 pseudo_t pseudo
= state
->cc_target
;
536 state
->cc_target
= NULL
;
538 if (!state
->cc_dead
) {
539 dst
= target_reg(state
, pseudo
, pseudo
);
540 flush_cc_cache_to_reg(state
, pseudo
, dst
);
545 static void add_cc_cache(struct bb_state
*state
, int opcode
, pseudo_t pseudo
)
547 assert(!state
->cc_target
);
548 state
->cc_target
= pseudo
;
549 state
->cc_opcode
= opcode
;
551 output_comment(state
, "caching %s", opcodes
[opcode
]);
554 /* Fill a hardreg with the pseudo it has */
555 static struct hardreg
*fill_reg(struct bb_state
*state
, struct hardreg
*hardreg
, pseudo_t pseudo
)
557 struct storage_hash
*src
;
558 struct instruction
*def
;
560 if (state
->cc_target
== pseudo
) {
561 flush_cc_cache_to_reg(state
, pseudo
, hardreg
);
565 switch (pseudo
->type
) {
567 output_insn(state
, "movl $%lld,%s", pseudo
->value
, hardreg
->name
);
570 output_insn(state
, "movl $<%s>,%s", show_pseudo(pseudo
), hardreg
->name
);
575 if (def
->opcode
== OP_SETVAL
) {
576 output_insn(state
, "movl $<%s>,%s", show_pseudo(def
->target
), hardreg
->name
);
579 src
= find_pseudo_storage(state
, pseudo
, hardreg
);
582 if (src
->flags
& TAG_DEAD
)
583 mark_reg_dead(state
, pseudo
, hardreg
);
584 output_insn(state
, "mov.%d %s,%s", 32, show_memop(src
->storage
), hardreg
->name
);
587 output_insn(state
, "reload %s from %s", hardreg
->name
, show_pseudo(pseudo
));
593 static struct hardreg
*getreg(struct bb_state
*state
, pseudo_t pseudo
, pseudo_t target
)
597 reg
= find_in_reg(state
, pseudo
);
600 reg
= target_reg(state
, pseudo
, target
);
601 return fill_reg(state
, reg
, pseudo
);
604 static void move_reg(struct bb_state
*state
, struct hardreg
*src
, struct hardreg
*dst
)
606 output_insn(state
, "movl %s,%s", src
->name
, dst
->name
);
609 static struct hardreg
*copy_reg(struct bb_state
*state
, struct hardreg
*src
, pseudo_t target
)
617 reg
= preferred_reg(state
, target
);
618 if (reg
&& !reg
->busy
) {
619 output_comment(state
, "copying %s to preferred target %s", show_pseudo(target
), reg
->name
);
620 move_reg(state
, src
, reg
);
624 for (i
= 0; i
< REGNO
; i
++) {
625 struct hardreg
*reg
= hardregs
+ i
;
627 output_comment(state
, "copying %s to %s", show_pseudo(target
), reg
->name
);
628 output_insn(state
, "movl %s,%s", src
->name
, reg
->name
);
633 flush_reg(state
, src
);
637 static void put_operand(struct bb_state
*state
, struct operand
*op
)
641 static struct operand
*alloc_op(void)
643 struct operand
*op
= malloc(sizeof(*op
));
644 memset(op
, 0, sizeof(*op
));
648 static struct operand
*get_register_operand(struct bb_state
*state
, pseudo_t pseudo
, pseudo_t target
)
650 struct operand
*op
= alloc_op();
652 op
->reg
= getreg(state
, pseudo
, target
);
656 static struct operand
*get_generic_operand(struct bb_state
*state
, pseudo_t pseudo
)
660 struct storage_hash
*hash
;
661 struct operand
*op
= malloc(sizeof(*op
));
663 memset(op
, 0, sizeof(*op
));
664 switch (pseudo
->type
) {
667 op
->value
= pseudo
->value
;
672 op
->sym
= pseudo
->sym
;
676 reg
= find_in_reg(state
, pseudo
);
682 hash
= find_pseudo_storage(state
, pseudo
, NULL
);
689 op
->reg
= hardregs
+ src
->regno
;
693 op
->offset
= src
->offset
;
694 op
->base
= hardregs
+ REG_EBP
;
698 op
->offset
= src
->offset
;
699 op
->base
= hardregs
+ REG_ESP
;
708 static const char *generic(struct bb_state
*state
, pseudo_t pseudo
)
710 return show_op(state
, get_generic_operand(state
, pseudo
));
713 static const char *address(struct bb_state
*state
, struct instruction
*memop
)
716 struct hardreg
*base
;
717 static char buffer
[100];
718 pseudo_t addr
= memop
->src
;
723 if (sym
->ctype
.modifiers
& MOD_NONLOCAL
) {
724 sprintf(buffer
, "%s+%d", show_ident(sym
->ident
), memop
->offset
);
727 sprintf(buffer
, "%d+%s(SP)", memop
->offset
, show_ident(sym
->ident
));
730 base
= getreg(state
, addr
, NULL
);
731 sprintf(buffer
, "%d(%s)", memop
->offset
, base
->name
);
736 static const char *reg_or_imm(struct bb_state
*state
, pseudo_t pseudo
)
738 switch(pseudo
->type
) {
740 return show_pseudo(pseudo
);
742 return getreg(state
, pseudo
, NULL
)->name
;
746 static void kill_dead_reg(struct hardreg
*reg
)
751 FOR_EACH_PTR(reg
->contains
, p
) {
752 if (CURRENT_TAG(p
) & TAG_DEAD
) {
753 DELETE_CURRENT_PTR(p
);
757 } END_FOR_EACH_PTR(p
);
758 PACK_PTR_LIST(®
->contains
);
763 static struct hardreg
*target_copy_reg(struct bb_state
*state
, struct hardreg
*src
, pseudo_t target
)
766 return copy_reg(state
, src
, target
);
769 static void do_binop(struct bb_state
*state
, struct instruction
*insn
, pseudo_t val1
, pseudo_t val2
)
771 const char *op
= opcodes
[insn
->opcode
];
772 struct operand
*src
= get_register_operand(state
, val1
, insn
->target
);
773 struct operand
*src2
= get_generic_operand(state
, val2
);
776 dst
= target_copy_reg(state
, src
->reg
, insn
->target
);
777 output_insn(state
, "%s.%d %s,%s", op
, insn
->size
, show_op(state
, src2
), dst
->name
);
778 put_operand(state
, src
);
779 put_operand(state
, src2
);
780 add_pseudo_reg(state
, insn
->target
, dst
);
783 static void generate_binop(struct bb_state
*state
, struct instruction
*insn
)
785 flush_cc_cache(state
);
786 do_binop(state
, insn
, insn
->src1
, insn
->src2
);
789 static int is_dead_reg(struct bb_state
*state
, pseudo_t pseudo
, struct hardreg
*reg
)
792 FOR_EACH_PTR(reg
->contains
, p
) {
794 return CURRENT_TAG(p
) & TAG_DEAD
;
795 } END_FOR_EACH_PTR(p
);
800 * Commutative binops are much more flexible, since we can switch the
801 * sources around to satisfy the target register, or to avoid having
802 * to load one of them into a register..
804 static void generate_commutative_binop(struct bb_state
*state
, struct instruction
*insn
)
807 struct hardreg
*reg1
, *reg2
;
809 flush_cc_cache(state
);
812 reg2
= find_in_reg(state
, src2
);
815 reg1
= find_in_reg(state
, src1
);
818 if (!is_dead_reg(state
, src2
, reg2
))
820 if (!is_dead_reg(state
, src1
, reg1
))
823 /* Both are dead. Is one preferrable? */
824 if (reg2
!= preferred_reg(state
, insn
->target
))
831 do_binop(state
, insn
, src1
, src2
);
835 * This marks a pseudo dead. It still stays on the hardreg list (the hardreg
836 * still has its value), but it's scheduled to be killed after the next
837 * "sequence point" when we call "kill_read_pseudos()"
839 static void mark_pseudo_dead(struct bb_state
*state
, pseudo_t pseudo
)
842 struct storage_hash
*src
;
844 if (state
->cc_target
== pseudo
)
846 src
= find_pseudo_storage(state
, pseudo
, NULL
);
848 src
->flags
|= TAG_DEAD
;
849 for (i
= 0; i
< REGNO
; i
++)
850 mark_reg_dead(state
, pseudo
, hardregs
+ i
);
853 static void kill_dead_pseudos(struct bb_state
*state
)
857 for (i
= 0; i
< REGNO
; i
++) {
858 kill_dead_reg(hardregs
+ i
);
863 * A PHI source can define a pseudo that we already
864 * have in another register. We need to invalidate the
865 * old register so that we don't end up with the same
866 * pseudo in "two places".
868 static void remove_pseudo_reg(struct bb_state
*state
, pseudo_t pseudo
)
872 output_comment(state
, "pseudo %s died", show_pseudo(pseudo
));
873 for (i
= 0; i
< REGNO
; i
++) {
874 struct hardreg
*reg
= hardregs
+ i
;
876 FOR_EACH_PTR(reg
->contains
, p
) {
879 if (CURRENT_TAG(p
) & TAG_DEAD
)
882 DELETE_CURRENT_PTR(p
);
883 output_comment(state
, "removed pseudo %s from reg %s", show_pseudo(pseudo
), reg
->name
);
884 } END_FOR_EACH_PTR(p
);
885 PACK_PTR_LIST(®
->contains
);
889 static void generate_store(struct instruction
*insn
, struct bb_state
*state
)
891 output_insn(state
, "mov.%d %s,%s", insn
->size
, reg_or_imm(state
, insn
->target
), address(state
, insn
));
894 static void generate_load(struct instruction
*insn
, struct bb_state
*state
)
896 const char *input
= address(state
, insn
);
899 kill_dead_pseudos(state
);
900 dst
= target_reg(state
, insn
->target
, NULL
);
901 output_insn(state
, "mov.%d %s,%s", insn
->size
, input
, dst
->name
);
904 static void generate_phisource(struct instruction
*insn
, struct bb_state
*state
)
906 struct instruction
*user
;
909 /* Mark all the target pseudos dead first */
910 FOR_EACH_PTR(insn
->phi_users
, user
) {
911 mark_pseudo_dead(state
, user
->target
);
912 } END_FOR_EACH_PTR(user
);
915 FOR_EACH_PTR(insn
->phi_users
, user
) {
917 reg
= getreg(state
, insn
->phi_src
, user
->target
);
918 remove_pseudo_reg(state
, user
->target
);
919 add_pseudo_reg(state
, user
->target
, reg
);
920 } END_FOR_EACH_PTR(user
);
923 static void generate_cast(struct bb_state
*state
, struct instruction
*insn
)
925 struct hardreg
*src
= getreg(state
, insn
->src
, insn
->target
);
927 unsigned int old
= insn
->orig_type
? insn
->orig_type
->bit_size
: 0;
928 unsigned int new = insn
->size
;
931 * Cast to smaller type? Ignore the high bits, we
932 * just keep both pseudos in the same register.
935 add_pseudo_reg(state
, insn
->target
, src
);
939 dst
= target_copy_reg(state
, src
, insn
->target
);
941 if (insn
->orig_type
&& (insn
->orig_type
->ctype
.modifiers
& MOD_SIGNED
)) {
942 output_insn(state
, "sext.%d.%d %s", old
, new, dst
->name
);
944 unsigned long long mask
;
945 mask
= ~(~0ULL << old
);
946 mask
&= ~(~0ULL << new);
947 output_insn(state
, "andl.%d $%#llx,%s", insn
->size
, mask
, dst
->name
);
949 add_pseudo_reg(state
, insn
->target
, dst
);
952 static void generate_output_storage(struct bb_state
*state
);
954 static const char *conditional
[] = {
968 static void generate_branch(struct bb_state
*state
, struct instruction
*br
)
970 const char *cond
= "XXX";
971 struct basic_block
*target
;
974 if (state
->cc_target
== br
->cond
) {
975 cond
= conditional
[state
->cc_opcode
];
977 struct hardreg
*reg
= getreg(state
, br
->cond
, NULL
);
978 output_insn(state
, "testl %s,%s", reg
->name
, reg
->name
);
982 generate_output_storage(state
);
983 target
= br
->bb_true
;
985 output_insn(state
, "j%s .L%p", cond
, target
);
986 target
= br
->bb_false
;
988 output_insn(state
, "jmp .L%p", target
);
991 /* We've made sure that there is a dummy reg live for the output */
992 static void generate_switch(struct bb_state
*state
, struct instruction
*insn
)
994 struct hardreg
*reg
= hardregs
+ SWITCH_REG
;
996 generate_output_storage(state
);
997 output_insn(state
, "switch on %s", reg
->name
);
998 output_insn(state
, "unimplemented: %s", show_instruction(insn
));
1001 static void generate_ret(struct bb_state
*state
, struct instruction
*ret
)
1003 if (ret
->src
&& ret
->src
!= VOID
) {
1004 struct hardreg
*wants
= hardregs
+0;
1005 struct hardreg
*reg
= getreg(state
, ret
->src
, NULL
);
1007 output_insn(state
, "movl %s,%s", reg
->name
, wants
->name
);
1009 output_insn(state
, "ret");
1013 * Fake "call" linearization just as a taster..
1015 static void generate_call(struct bb_state
*state
, struct instruction
*insn
)
1020 FOR_EACH_PTR(insn
->arguments
, arg
) {
1021 output_insn(state
, "pushl %s", generic(state
, arg
));
1023 } END_FOR_EACH_PTR(arg
);
1024 flush_reg(state
, hardregs
+0);
1025 flush_reg(state
, hardregs
+1);
1026 flush_reg(state
, hardregs
+2);
1027 output_insn(state
, "call %s", show_pseudo(insn
->func
));
1029 output_insn(state
, "addl $%d,%%esp", offset
);
1030 add_pseudo_reg(state
, insn
->target
, hardregs
+0);
1033 static void generate_select(struct bb_state
*state
, struct instruction
*insn
)
1036 struct hardreg
*src1
, *src2
, *dst
;
1038 src1
= getreg(state
, insn
->src2
, NULL
);
1039 dst
= copy_reg(state
, src1
, insn
->target
);
1040 add_pseudo_reg(state
, insn
->target
, dst
);
1041 src2
= getreg(state
, insn
->src3
, insn
->target
);
1043 if (state
->cc_target
== insn
->src1
) {
1044 cond
= conditional
[state
->cc_opcode
];
1046 struct hardreg
*reg
= getreg(state
, insn
->src1
, NULL
);
1047 output_insn(state
, "testl %s,%s", reg
->name
, reg
->name
);
1051 output_insn(state
, "sel%s %s,%s", cond
, src2
->name
, dst
->name
);
1055 const struct ident
*name
;
1058 struct hardreg
*reg
;
1061 static void replace_asm_arg(char **dst_p
, struct asm_arg
*arg
)
1064 int len
= strlen(arg
->value
);
1066 memcpy(dst
, arg
->value
, len
);
1070 static void replace_asm_percent(const char **src_p
, char **dst_p
, struct asm_arg
*args
, int nr
)
1072 const char *src
= *src_p
;
1081 replace_asm_arg(dst_p
, args
+index
);
1088 static void replace_asm_named(const char **src_p
, char **dst_p
, struct asm_arg
*args
, int nr
)
1090 const char *src
= *src_p
;
1091 const char *end
= src
;
1101 for (i
= 0; i
< nr
; i
++) {
1102 const struct ident
*ident
= args
[i
].name
;
1107 if (memcmp(src
, ident
->name
, len
))
1109 replace_asm_arg(dst_p
, args
+i
);
1116 static const char *replace_asm_args(const char *str
, struct asm_arg
*args
, int nr
)
1118 static char buffer
[1000];
1134 replace_asm_percent(&str
, &p
, args
, nr
);
1137 replace_asm_named(&str
, &p
, args
, nr
);
1146 #define MAX_ASM_ARG (50)
1147 static struct asm_arg asm_arguments
[MAX_ASM_ARG
];
1149 static struct asm_arg
*generate_asm_inputs(struct bb_state
*state
, struct asm_constraint_list
*list
, struct asm_arg
*arg
)
1151 struct asm_constraint
*entry
;
1153 FOR_EACH_PTR(list
, entry
) {
1154 const char *constraint
= entry
->constraint
;
1155 pseudo_t pseudo
= entry
->pseudo
;
1156 struct hardreg
*reg
, *orig
;
1161 switch (*constraint
) {
1163 string
= getreg(state
, pseudo
, NULL
)->name
;
1166 index
= *constraint
- '0';
1167 reg
= asm_arguments
[index
].reg
;
1168 orig
= find_in_reg(state
, pseudo
);
1170 move_reg(state
, orig
, reg
);
1172 fill_reg(state
, reg
, pseudo
);
1176 string
= generic(state
, pseudo
);
1180 output_insn(state
, "# asm input \"%s\": %s : %s", constraint
, show_pseudo(pseudo
), string
);
1182 arg
->name
= entry
->ident
;
1183 arg
->value
= string
;
1187 } END_FOR_EACH_PTR(entry
);
1191 static struct asm_arg
*generate_asm_outputs(struct bb_state
*state
, struct asm_constraint_list
*list
, struct asm_arg
*arg
)
1193 struct asm_constraint
*entry
;
1195 FOR_EACH_PTR(list
, entry
) {
1196 const char *constraint
= entry
->constraint
;
1197 pseudo_t pseudo
= entry
->pseudo
;
1198 struct hardreg
*reg
;
1201 while (*constraint
== '=' || *constraint
== '+')
1205 switch (*constraint
) {
1208 reg
= target_reg(state
, pseudo
, NULL
);
1209 arg
->pseudo
= pseudo
;
1215 output_insn(state
, "# asm output \"%s\": %s : %s", constraint
, show_pseudo(pseudo
), string
);
1217 arg
->name
= entry
->ident
;
1218 arg
->value
= string
;
1220 } END_FOR_EACH_PTR(entry
);
1224 static void generate_asm(struct bb_state
*state
, struct instruction
*insn
)
1226 const char *str
= insn
->string
;
1228 if (insn
->asm_rules
->outputs
|| insn
->asm_rules
->inputs
) {
1229 struct asm_arg
*arg
;
1231 arg
= generate_asm_outputs(state
, insn
->asm_rules
->outputs
, asm_arguments
);
1232 arg
= generate_asm_inputs(state
, insn
->asm_rules
->inputs
, arg
);
1233 str
= replace_asm_args(str
, asm_arguments
, arg
- asm_arguments
);
1235 output_insn(state
, "%s", str
);
1238 static void generate_compare(struct bb_state
*state
, struct instruction
*insn
)
1240 struct hardreg
*src
;
1244 flush_cc_cache(state
);
1245 opcode
= insn
->opcode
;
1248 * We should try to switch these around if necessary,
1249 * and update the opcode to match..
1251 src
= getreg(state
, insn
->src1
, insn
->target
);
1252 src2
= generic(state
, insn
->src2
);
1254 output_insn(state
, "cmp.%d %s,%s", insn
->size
, src2
, src
->name
);
1256 add_cc_cache(state
, opcode
, insn
->target
);
1259 static void generate_one_insn(struct instruction
*insn
, struct bb_state
*state
)
1262 output_comment(state
, "%s", show_instruction(insn
));
1264 switch (insn
->opcode
) {
1266 struct symbol
*sym
= insn
->bb
->ep
->name
;
1267 const char *name
= show_ident(sym
->ident
);
1268 if (sym
->ctype
.modifiers
& MOD_STATIC
)
1269 printf("\n\n%s:\n", name
);
1271 printf("\n\n.globl %s\n%s:\n", name
, name
);
1276 * OP_PHI doesn't actually generate any code. It has been
1277 * done by the storage allocator and the OP_PHISOURCE.
1283 generate_phisource(insn
, state
);
1287 * OP_SETVAL likewise doesn't actually generate any
1288 * code. On use, the "def" of the pseudo will be
1295 generate_store(insn
, state
);
1299 generate_load(insn
, state
);
1303 mark_pseudo_dead(state
, insn
->target
);
1306 case OP_ADD
: case OP_MUL
:
1307 case OP_AND
: case OP_OR
: case OP_XOR
:
1308 case OP_AND_BOOL
: case OP_OR_BOOL
:
1309 generate_commutative_binop(state
, insn
);
1312 case OP_SUB
: case OP_DIV
: case OP_MOD
:
1313 case OP_SHL
: case OP_SHR
:
1314 generate_binop(state
, insn
);
1317 case OP_BINCMP
... OP_BINCMP_END
:
1318 generate_compare(state
, insn
);
1321 case OP_CAST
: case OP_PTRCAST
:
1322 generate_cast(state
, insn
);
1326 generate_select(state
, insn
);
1330 generate_branch(state
, insn
);
1334 generate_switch(state
, insn
);
1338 generate_call(state
, insn
);
1342 generate_ret(state
, insn
);
1346 generate_asm(state
, insn
);
1350 output_insn(state
, "unimplemented: %s", show_instruction(insn
));
1353 kill_dead_pseudos(state
);
1356 #define VERY_BUSY 1000
1357 #define REG_FIXED 2000
1359 static void write_reg_to_storage(struct bb_state
*state
, struct hardreg
*reg
, pseudo_t pseudo
, struct storage
*storage
)
1362 struct hardreg
*out
;
1364 switch (storage
->type
) {
1366 out
= hardregs
+ storage
->regno
;
1369 output_insn(state
, "movl %s,%s", reg
->name
, out
->name
);
1372 if (reg
->busy
< VERY_BUSY
) {
1373 storage
->type
= REG_REG
;
1374 storage
->regno
= reg
- hardregs
;
1375 reg
->busy
= REG_FIXED
;
1379 /* Try to find a non-busy register.. */
1380 for (i
= 0; i
< REGNO
; i
++) {
1384 output_insn(state
, "movl %s,%s", reg
->name
, out
->name
);
1385 storage
->type
= REG_REG
;
1387 out
->busy
= REG_FIXED
;
1391 /* Fall back on stack allocation ... */
1392 alloc_stack(state
, storage
);
1395 output_insn(state
, "movl %s,%s", reg
->name
, show_memop(storage
));
1400 static void write_val_to_storage(struct bb_state
*state
, pseudo_t src
, struct storage
*storage
)
1402 struct hardreg
*out
;
1404 switch (storage
->type
) {
1406 alloc_stack(state
, storage
);
1408 output_insn(state
, "movl %s,%s", show_pseudo(src
), show_memop(storage
));
1411 out
= hardregs
+ storage
->regno
;
1412 output_insn(state
, "movl %s,%s", show_pseudo(src
), out
->name
);
1416 static void fill_output(struct bb_state
*state
, pseudo_t pseudo
, struct storage
*out
)
1419 struct storage_hash
*in
;
1420 struct instruction
*def
;
1422 /* Is that pseudo a constant value? */
1423 switch (pseudo
->type
) {
1425 write_val_to_storage(state
, pseudo
, out
);
1429 if (def
->opcode
== OP_SETVAL
) {
1430 write_val_to_storage(state
, pseudo
, out
);
1437 /* See if we have that pseudo in a register.. */
1438 for (i
= 0; i
< REGNO
; i
++) {
1439 struct hardreg
*reg
= hardregs
+ i
;
1442 FOR_EACH_PTR(reg
->contains
, p
) {
1444 write_reg_to_storage(state
, reg
, pseudo
, out
);
1447 } END_FOR_EACH_PTR(p
);
1450 /* Do we have it in another storage? */
1451 in
= find_storage_hash(pseudo
, state
->internal
);
1453 in
= find_storage_hash(pseudo
, state
->inputs
);
1458 switch (out
->type
) {
1460 *out
= *in
->storage
;
1463 output_insn(state
, "movl %s,%s", show_memop(in
->storage
), hardregs
[out
->regno
].name
);
1466 if (out
== in
->storage
)
1468 if (out
->type
== in
->storage
->type
== out
->regno
== in
->storage
->regno
)
1470 output_insn(state
, "movl %s,%s", show_memop(in
->storage
), show_memop(out
));
1476 static int final_pseudo_flush(struct bb_state
*state
, pseudo_t pseudo
, struct hardreg
*reg
)
1478 struct storage_hash
*hash
;
1479 struct storage
*out
;
1480 struct hardreg
*dst
;
1483 * Since this pseudo is live at exit, we'd better have output
1486 hash
= find_storage_hash(pseudo
, state
->outputs
);
1489 out
= hash
->storage
;
1491 /* If the output is in a register, try to get it there.. */
1492 if (out
->type
== REG_REG
) {
1493 dst
= hardregs
+ out
->regno
;
1495 * Two good cases: nobody is using the right register,
1496 * or we've already set it aside for output..
1498 if (!dst
->busy
|| dst
->busy
> VERY_BUSY
)
1501 /* Aiee. Try to keep it in a register.. */
1502 dst
= empty_reg(state
);
1509 /* If the output is undefined, let's see if we can put it in a register.. */
1510 if (out
->type
== REG_UDEF
) {
1511 dst
= empty_reg(state
);
1513 out
->type
= REG_REG
;
1514 out
->regno
= dst
- hardregs
;
1517 /* Uhhuh. Not so good. No empty registers right now */
1521 /* If we know we need to flush it, just do so already .. */
1522 output_insn(state
, "movl %s,%s", reg
->name
, show_memop(out
));
1528 output_insn(state
, "movl %s,%s", reg
->name
, dst
->name
);
1529 add_pseudo_reg(state
, pseudo
, dst
);
1534 * This tries to make sure that we put all the pseudos that are
1535 * live on exit into the proper storage
1537 static void generate_output_storage(struct bb_state
*state
)
1539 struct storage_hash
*entry
;
1541 /* Go through the fixed outputs, making sure we have those regs free */
1542 FOR_EACH_PTR(state
->outputs
, entry
) {
1543 struct storage
*out
= entry
->storage
;
1544 if (out
->type
== REG_REG
) {
1545 struct hardreg
*reg
= hardregs
+ out
->regno
;
1549 reg
->busy
= REG_FIXED
;
1550 FOR_EACH_PTR(reg
->contains
, p
) {
1551 if (p
== entry
->pseudo
) {
1555 if (CURRENT_TAG(p
) & TAG_DEAD
)
1558 /* Try to write back the pseudo to where it should go ... */
1559 if (final_pseudo_flush(state
, p
, reg
)) {
1560 DELETE_CURRENT_PTR(p
);
1565 } END_FOR_EACH_PTR(p
);
1566 PACK_PTR_LIST(®
->contains
);
1568 flush_reg(state
, reg
);
1570 } END_FOR_EACH_PTR(entry
);
1572 FOR_EACH_PTR(state
->outputs
, entry
) {
1573 fill_output(state
, entry
->pseudo
, entry
->storage
);
1574 } END_FOR_EACH_PTR(entry
);
1577 static void generate(struct basic_block
*bb
, struct bb_state
*state
)
1580 struct storage_hash
*entry
;
1581 struct instruction
*insn
;
1583 for (i
= 0; i
< REGNO
; i
++) {
1584 free_ptr_list(&hardregs
[i
].contains
);
1585 hardregs
[i
].busy
= 0;
1586 hardregs
[i
].dead
= 0;
1587 hardregs
[i
].used
= 0;
1590 FOR_EACH_PTR(state
->inputs
, entry
) {
1591 struct storage
*storage
= entry
->storage
;
1592 const char *name
= show_storage(storage
);
1593 output_comment(state
, "incoming %s in %s", show_pseudo(entry
->pseudo
), name
);
1594 if (storage
->type
== REG_REG
) {
1595 int regno
= storage
->regno
;
1596 add_pseudo_reg(state
, entry
->pseudo
, hardregs
+ regno
);
1597 name
= hardregs
[regno
].name
;
1599 } END_FOR_EACH_PTR(entry
);
1601 output_label(state
, ".L%p", bb
);
1602 FOR_EACH_PTR(bb
->insns
, insn
) {
1605 generate_one_insn(insn
, state
);
1606 } END_FOR_EACH_PTR(insn
);
1609 output_comment(state
, "--- in ---");
1610 FOR_EACH_PTR(state
->inputs
, entry
) {
1611 output_comment(state
, "%s <- %s", show_pseudo(entry
->pseudo
), show_storage(entry
->storage
));
1612 } END_FOR_EACH_PTR(entry
);
1613 output_comment(state
, "--- spill ---");
1614 FOR_EACH_PTR(state
->internal
, entry
) {
1615 output_comment(state
, "%s <-> %s", show_pseudo(entry
->pseudo
), show_storage(entry
->storage
));
1616 } END_FOR_EACH_PTR(entry
);
1617 output_comment(state
, "--- out ---");
1618 FOR_EACH_PTR(state
->outputs
, entry
) {
1619 output_comment(state
, "%s -> %s", show_pseudo(entry
->pseudo
), show_storage(entry
->storage
));
1620 } END_FOR_EACH_PTR(entry
);
1625 static void generate_list(struct basic_block_list
*list
, unsigned long generation
)
1627 struct basic_block
*bb
;
1628 FOR_EACH_PTR(list
, bb
) {
1629 if (bb
->generation
== generation
)
1631 output_bb(bb
, generation
);
1632 } END_FOR_EACH_PTR(bb
);
1636 * Mark all the output registers of all the parents
1637 * as being "used" - this does not mean that we cannot
1638 * re-use them, but it means that we cannot ask the
1639 * parents to pass in another pseudo in one of those
1640 * registers that it already uses for another child.
1642 static void mark_used_registers(struct basic_block
*bb
, struct bb_state
*state
)
1644 struct basic_block
*parent
;
1646 FOR_EACH_PTR(bb
->parents
, parent
) {
1647 struct storage_hash_list
*outputs
= gather_storage(parent
, STOR_OUT
);
1648 struct storage_hash
*entry
;
1650 FOR_EACH_PTR(outputs
, entry
) {
1651 struct storage
*s
= entry
->storage
;
1652 if (s
->type
== REG_REG
) {
1653 struct hardreg
*reg
= hardregs
+ s
->regno
;
1656 } END_FOR_EACH_PTR(entry
);
1657 } END_FOR_EACH_PTR(parent
);
1660 static void output_bb(struct basic_block
*bb
, unsigned long generation
)
1662 struct bb_state state
;
1664 bb
->generation
= generation
;
1666 /* Make sure all parents have been generated first */
1667 generate_list(bb
->parents
, generation
);
1669 state
.pos
= bb
->pos
;
1670 state
.inputs
= gather_storage(bb
, STOR_IN
);
1671 state
.outputs
= gather_storage(bb
, STOR_OUT
);
1672 state
.internal
= NULL
;
1673 state
.cc_opcode
= 0;
1674 state
.cc_target
= NULL
;
1676 /* Mark incoming registers used */
1677 mark_used_registers(bb
, &state
);
1679 generate(bb
, &state
);
1681 free_ptr_list(&state
.inputs
);
1682 free_ptr_list(&state
.outputs
);
1684 /* Generate all children... */
1685 generate_list(bb
->children
, generation
);
1688 static void set_up_arch_entry(struct entrypoint
*ep
, struct instruction
*entry
)
1694 * We should set up argument sources here..
1696 * Things like "first three arguments in registers" etc
1697 * are all for this place.
1700 FOR_EACH_PTR(entry
->arg_list
, arg
) {
1701 struct storage
*in
= lookup_storage(entry
->bb
, arg
, STOR_IN
);
1703 in
= alloc_storage();
1704 add_storage(in
, entry
->bb
, arg
, STOR_IN
);
1710 in
->type
= REG_FRAME
;
1711 in
->offset
= (i
-3)*4;
1714 } END_FOR_EACH_PTR(arg
);
1718 * Set up storage information for "return"
1720 * Not strictly necessary, since the code generator will
1721 * certainly move the return value to the right register,
1722 * but it can help register allocation if the allocator
1723 * sees that the target register is going to return in %eax.
1725 static void set_up_arch_exit(struct basic_block
*bb
, struct instruction
*ret
)
1727 pseudo_t pseudo
= ret
->src
;
1729 if (pseudo
&& pseudo
!= VOID
) {
1730 struct storage
*out
= lookup_storage(bb
, pseudo
, STOR_OUT
);
1732 out
= alloc_storage();
1733 add_storage(out
, bb
, pseudo
, STOR_OUT
);
1735 out
->type
= REG_REG
;
1741 * Set up dummy/silly output storage information for a switch
1742 * instruction. We need to make sure that a register is available
1743 * when we generate code for switch, so force that by creating
1744 * a dummy output rule.
1746 static void set_up_arch_switch(struct basic_block
*bb
, struct instruction
*insn
)
1748 pseudo_t pseudo
= insn
->cond
;
1749 struct storage
*out
= lookup_storage(bb
, pseudo
, STOR_OUT
);
1751 out
= alloc_storage();
1752 add_storage(out
, bb
, pseudo
, STOR_OUT
);
1754 out
->type
= REG_REG
;
1755 out
->regno
= SWITCH_REG
;
1758 static void arch_set_up_storage(struct entrypoint
*ep
)
1760 struct basic_block
*bb
;
1762 /* Argument storage etc.. */
1763 set_up_arch_entry(ep
, ep
->entry
);
1765 FOR_EACH_PTR(ep
->bbs
, bb
) {
1766 struct instruction
*insn
= last_instruction(bb
->insns
);
1769 switch (insn
->opcode
) {
1771 set_up_arch_exit(bb
, insn
);
1774 set_up_arch_switch(bb
, insn
);
1779 } END_FOR_EACH_PTR(bb
);
1782 static void output(struct entrypoint
*ep
)
1784 unsigned long generation
= ++bb_generation
;
1789 /* Set up initial inter-bb storage links */
1792 /* Architecture-specific storage rules.. */
1793 arch_set_up_storage(ep
);
1795 /* Show the results ... */
1796 output_bb(ep
->entry
->bb
, generation
);
1798 /* Clear the storage hashes for the next function.. */
1802 static int compile(struct symbol_list
*list
)
1805 FOR_EACH_PTR(list
, sym
) {
1806 struct entrypoint
*ep
;
1808 ep
= linearize_symbol(sym
);
1811 } END_FOR_EACH_PTR(sym
);
1816 int main(int argc
, char **argv
)
1818 return compile(sparse(argc
, argv
));