* rw.po: Remove.
[official-gcc.git] / gcc / cfgrtl.c
blob4b801a57df8836719762c2d0997d09bad4ef6b46
1 /* Control flow graph manipulation code for GNU compiler.
2 Copyright (C) 1987, 1988, 1992, 1993, 1994, 1995, 1996, 1997, 1998,
3 1999, 2000, 2001, 2002, 2003, 2004, 2005, 2007 Free Software Foundation, Inc.
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify it under
8 the terms of the GNU General Public License as published by the Free
9 Software Foundation; either version 3, or (at your option) any later
10 version.
12 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
15 for more details.
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
21 /* This file contains low level functions to manipulate the CFG and analyze it
22 that are aware of the RTL intermediate language.
24 Available functionality:
25 - Basic CFG/RTL manipulation API documented in cfghooks.h
26 - CFG-aware instruction chain manipulation
27 delete_insn, delete_insn_chain
28 - Edge splitting and committing to edges
29 insert_insn_on_edge, commit_edge_insertions
30 - CFG updating after insn simplification
31 purge_dead_edges, purge_all_dead_edges
33 Functions not supposed for generic use:
34 - Infrastructure to determine quickly basic block for insn
35 compute_bb_for_insn, update_bb_for_insn, set_block_for_insn,
36 - Edge redirection with updating and optimizing of insn chain
37 block_label, tidy_fallthru_edge, force_nonfallthru */
39 #include "config.h"
40 #include "system.h"
41 #include "coretypes.h"
42 #include "tm.h"
43 #include "tree.h"
44 #include "rtl.h"
45 #include "hard-reg-set.h"
46 #include "basic-block.h"
47 #include "regs.h"
48 #include "flags.h"
49 #include "output.h"
50 #include "function.h"
51 #include "except.h"
52 #include "toplev.h"
53 #include "tm_p.h"
54 #include "obstack.h"
55 #include "insn-config.h"
56 #include "cfglayout.h"
57 #include "expr.h"
58 #include "target.h"
59 #include "cfgloop.h"
60 #include "ggc.h"
61 #include "tree-pass.h"
63 static int can_delete_note_p (rtx);
64 static int can_delete_label_p (rtx);
65 static void commit_one_edge_insertion (edge, int);
66 static basic_block rtl_split_edge (edge);
67 static bool rtl_move_block_after (basic_block, basic_block);
68 static int rtl_verify_flow_info (void);
69 static basic_block cfg_layout_split_block (basic_block, void *);
70 static edge cfg_layout_redirect_edge_and_branch (edge, basic_block);
71 static basic_block cfg_layout_redirect_edge_and_branch_force (edge, basic_block);
72 static void cfg_layout_delete_block (basic_block);
73 static void rtl_delete_block (basic_block);
74 static basic_block rtl_redirect_edge_and_branch_force (edge, basic_block);
75 static edge rtl_redirect_edge_and_branch (edge, basic_block);
76 static basic_block rtl_split_block (basic_block, void *);
77 static void rtl_dump_bb (basic_block, FILE *, int);
78 static int rtl_verify_flow_info_1 (void);
79 static void rtl_make_forwarder_block (edge);
81 /* Return true if NOTE is not one of the ones that must be kept paired,
82 so that we may simply delete it. */
84 static int
85 can_delete_note_p (rtx note)
87 return (NOTE_LINE_NUMBER (note) == NOTE_INSN_DELETED
88 || NOTE_LINE_NUMBER (note) == NOTE_INSN_BASIC_BLOCK);
91 /* True if a given label can be deleted. */
93 static int
94 can_delete_label_p (rtx label)
96 return (!LABEL_PRESERVE_P (label)
97 /* User declared labels must be preserved. */
98 && LABEL_NAME (label) == 0
99 && !in_expr_list_p (forced_labels, label));
102 /* Delete INSN by patching it out. Return the next insn. */
105 delete_insn (rtx insn)
107 rtx next = NEXT_INSN (insn);
108 rtx note;
109 bool really_delete = true;
111 if (LABEL_P (insn))
113 /* Some labels can't be directly removed from the INSN chain, as they
114 might be references via variables, constant pool etc.
115 Convert them to the special NOTE_INSN_DELETED_LABEL note. */
116 if (! can_delete_label_p (insn))
118 const char *name = LABEL_NAME (insn);
120 really_delete = false;
121 PUT_CODE (insn, NOTE);
122 NOTE_LINE_NUMBER (insn) = NOTE_INSN_DELETED_LABEL;
123 NOTE_DELETED_LABEL_NAME (insn) = name;
126 remove_node_from_expr_list (insn, &nonlocal_goto_handler_labels);
129 if (really_delete)
131 /* If this insn has already been deleted, something is very wrong. */
132 gcc_assert (!INSN_DELETED_P (insn));
133 remove_insn (insn);
134 INSN_DELETED_P (insn) = 1;
137 /* If deleting a jump, decrement the use count of the label. Deleting
138 the label itself should happen in the normal course of block merging. */
139 if (JUMP_P (insn)
140 && JUMP_LABEL (insn)
141 && LABEL_P (JUMP_LABEL (insn)))
142 LABEL_NUSES (JUMP_LABEL (insn))--;
144 /* Also if deleting an insn that references a label. */
145 else
147 while ((note = find_reg_note (insn, REG_LABEL, NULL_RTX)) != NULL_RTX
148 && LABEL_P (XEXP (note, 0)))
150 LABEL_NUSES (XEXP (note, 0))--;
151 remove_note (insn, note);
155 if (JUMP_P (insn)
156 && (GET_CODE (PATTERN (insn)) == ADDR_VEC
157 || GET_CODE (PATTERN (insn)) == ADDR_DIFF_VEC))
159 rtx pat = PATTERN (insn);
160 int diff_vec_p = GET_CODE (PATTERN (insn)) == ADDR_DIFF_VEC;
161 int len = XVECLEN (pat, diff_vec_p);
162 int i;
164 for (i = 0; i < len; i++)
166 rtx label = XEXP (XVECEXP (pat, diff_vec_p, i), 0);
168 /* When deleting code in bulk (e.g. removing many unreachable
169 blocks) we can delete a label that's a target of the vector
170 before deleting the vector itself. */
171 if (!NOTE_P (label))
172 LABEL_NUSES (label)--;
176 return next;
179 /* Like delete_insn but also purge dead edges from BB. */
181 delete_insn_and_edges (rtx insn)
183 rtx x;
184 bool purge = false;
186 if (INSN_P (insn)
187 && BLOCK_FOR_INSN (insn)
188 && BB_END (BLOCK_FOR_INSN (insn)) == insn)
189 purge = true;
190 x = delete_insn (insn);
191 if (purge)
192 purge_dead_edges (BLOCK_FOR_INSN (insn));
193 return x;
196 /* Unlink a chain of insns between START and FINISH, leaving notes
197 that must be paired. */
199 void
200 delete_insn_chain (rtx start, rtx finish)
202 rtx next;
204 /* Unchain the insns one by one. It would be quicker to delete all of these
205 with a single unchaining, rather than one at a time, but we need to keep
206 the NOTE's. */
207 while (1)
209 next = NEXT_INSN (start);
210 if (NOTE_P (start) && !can_delete_note_p (start))
212 else
213 next = delete_insn (start);
215 if (start == finish)
216 break;
217 start = next;
221 /* Like delete_insn but also purge dead edges from BB. */
222 void
223 delete_insn_chain_and_edges (rtx first, rtx last)
225 bool purge = false;
227 if (INSN_P (last)
228 && BLOCK_FOR_INSN (last)
229 && BB_END (BLOCK_FOR_INSN (last)) == last)
230 purge = true;
231 delete_insn_chain (first, last);
232 if (purge)
233 purge_dead_edges (BLOCK_FOR_INSN (last));
236 /* Create a new basic block consisting of the instructions between HEAD and END
237 inclusive. This function is designed to allow fast BB construction - reuses
238 the note and basic block struct in BB_NOTE, if any and do not grow
239 BASIC_BLOCK chain and should be used directly only by CFG construction code.
240 END can be NULL in to create new empty basic block before HEAD. Both END
241 and HEAD can be NULL to create basic block at the end of INSN chain.
242 AFTER is the basic block we should be put after. */
244 basic_block
245 create_basic_block_structure (rtx head, rtx end, rtx bb_note, basic_block after)
247 basic_block bb;
249 if (bb_note
250 && (bb = NOTE_BASIC_BLOCK (bb_note)) != NULL
251 && bb->aux == NULL)
253 /* If we found an existing note, thread it back onto the chain. */
255 rtx after;
257 if (LABEL_P (head))
258 after = head;
259 else
261 after = PREV_INSN (head);
262 head = bb_note;
265 if (after != bb_note && NEXT_INSN (after) != bb_note)
266 reorder_insns_nobb (bb_note, bb_note, after);
268 else
270 /* Otherwise we must create a note and a basic block structure. */
272 bb = alloc_block ();
274 init_rtl_bb_info (bb);
275 if (!head && !end)
276 head = end = bb_note
277 = emit_note_after (NOTE_INSN_BASIC_BLOCK, get_last_insn ());
278 else if (LABEL_P (head) && end)
280 bb_note = emit_note_after (NOTE_INSN_BASIC_BLOCK, head);
281 if (head == end)
282 end = bb_note;
284 else
286 bb_note = emit_note_before (NOTE_INSN_BASIC_BLOCK, head);
287 head = bb_note;
288 if (!end)
289 end = head;
292 NOTE_BASIC_BLOCK (bb_note) = bb;
295 /* Always include the bb note in the block. */
296 if (NEXT_INSN (end) == bb_note)
297 end = bb_note;
299 BB_HEAD (bb) = head;
300 BB_END (bb) = end;
301 bb->index = last_basic_block++;
302 bb->flags = BB_NEW | BB_RTL;
303 link_block (bb, after);
304 SET_BASIC_BLOCK (bb->index, bb);
305 update_bb_for_insn (bb);
306 BB_SET_PARTITION (bb, BB_UNPARTITIONED);
308 /* Tag the block so that we know it has been used when considering
309 other basic block notes. */
310 bb->aux = bb;
312 return bb;
315 /* Create new basic block consisting of instructions in between HEAD and END
316 and place it to the BB chain after block AFTER. END can be NULL in to
317 create new empty basic block before HEAD. Both END and HEAD can be NULL to
318 create basic block at the end of INSN chain. */
320 static basic_block
321 rtl_create_basic_block (void *headp, void *endp, basic_block after)
323 rtx head = headp, end = endp;
324 basic_block bb;
326 /* Grow the basic block array if needed. */
327 if ((size_t) last_basic_block >= VEC_length (basic_block, basic_block_info))
329 size_t old_size = VEC_length (basic_block, basic_block_info);
330 size_t new_size = last_basic_block + (last_basic_block + 3) / 4;
331 basic_block *p;
332 VEC_safe_grow (basic_block, gc, basic_block_info, new_size);
333 p = VEC_address (basic_block, basic_block_info);
334 memset (&p[old_size], 0, sizeof (basic_block) * (new_size - old_size));
337 n_basic_blocks++;
339 bb = create_basic_block_structure (head, end, NULL, after);
340 bb->aux = NULL;
341 return bb;
344 static basic_block
345 cfg_layout_create_basic_block (void *head, void *end, basic_block after)
347 basic_block newbb = rtl_create_basic_block (head, end, after);
349 return newbb;
352 /* Delete the insns in a (non-live) block. We physically delete every
353 non-deleted-note insn, and update the flow graph appropriately.
355 Return nonzero if we deleted an exception handler. */
357 /* ??? Preserving all such notes strikes me as wrong. It would be nice
358 to post-process the stream to remove empty blocks, loops, ranges, etc. */
360 static void
361 rtl_delete_block (basic_block b)
363 rtx insn, end;
365 /* If the head of this block is a CODE_LABEL, then it might be the
366 label for an exception handler which can't be reached. We need
367 to remove the label from the exception_handler_label list. */
368 insn = BB_HEAD (b);
369 if (LABEL_P (insn))
370 maybe_remove_eh_handler (insn);
372 end = get_last_bb_insn (b);
374 /* Selectively delete the entire chain. */
375 BB_HEAD (b) = NULL;
376 delete_insn_chain (insn, end);
377 if (b->il.rtl->global_live_at_start)
379 FREE_REG_SET (b->il.rtl->global_live_at_start);
380 FREE_REG_SET (b->il.rtl->global_live_at_end);
381 b->il.rtl->global_live_at_start = NULL;
382 b->il.rtl->global_live_at_end = NULL;
386 /* Records the basic block struct in BLOCK_FOR_INSN for every insn. */
388 void
389 compute_bb_for_insn (void)
391 basic_block bb;
393 FOR_EACH_BB (bb)
395 rtx end = BB_END (bb);
396 rtx insn;
398 for (insn = BB_HEAD (bb); ; insn = NEXT_INSN (insn))
400 BLOCK_FOR_INSN (insn) = bb;
401 if (insn == end)
402 break;
407 /* Release the basic_block_for_insn array. */
409 unsigned int
410 free_bb_for_insn (void)
412 rtx insn;
413 for (insn = get_insns (); insn; insn = NEXT_INSN (insn))
414 if (!BARRIER_P (insn))
415 BLOCK_FOR_INSN (insn) = NULL;
416 return 0;
419 struct tree_opt_pass pass_free_cfg =
421 NULL, /* name */
422 NULL, /* gate */
423 free_bb_for_insn, /* execute */
424 NULL, /* sub */
425 NULL, /* next */
426 0, /* static_pass_number */
427 0, /* tv_id */
428 0, /* properties_required */
429 0, /* properties_provided */
430 PROP_cfg, /* properties_destroyed */
431 0, /* todo_flags_start */
432 0, /* todo_flags_finish */
433 0 /* letter */
436 /* Return RTX to emit after when we want to emit code on the entry of function. */
438 entry_of_function (void)
440 return (n_basic_blocks > NUM_FIXED_BLOCKS ?
441 BB_HEAD (ENTRY_BLOCK_PTR->next_bb) : get_insns ());
444 /* Emit INSN at the entry point of the function, ensuring that it is only
445 executed once per function. */
446 void
447 emit_insn_at_entry (rtx insn)
449 edge_iterator ei = ei_start (ENTRY_BLOCK_PTR->succs);
450 edge e = ei_safe_edge (ei);
451 gcc_assert (e->flags & EDGE_FALLTHRU);
453 insert_insn_on_edge (insn, e);
454 commit_edge_insertions ();
457 /* Update insns block within BB. */
459 void
460 update_bb_for_insn (basic_block bb)
462 rtx insn;
464 for (insn = BB_HEAD (bb); ; insn = NEXT_INSN (insn))
466 if (!BARRIER_P (insn))
467 set_block_for_insn (insn, bb);
468 if (insn == BB_END (bb))
469 break;
473 /* Creates a new basic block just after basic block B by splitting
474 everything after specified instruction I. */
476 static basic_block
477 rtl_split_block (basic_block bb, void *insnp)
479 basic_block new_bb;
480 rtx insn = insnp;
481 edge e;
482 edge_iterator ei;
484 if (!insn)
486 insn = first_insn_after_basic_block_note (bb);
488 if (insn)
489 insn = PREV_INSN (insn);
490 else
491 insn = get_last_insn ();
494 /* We probably should check type of the insn so that we do not create
495 inconsistent cfg. It is checked in verify_flow_info anyway, so do not
496 bother. */
497 if (insn == BB_END (bb))
498 emit_note_after (NOTE_INSN_DELETED, insn);
500 /* Create the new basic block. */
501 new_bb = create_basic_block (NEXT_INSN (insn), BB_END (bb), bb);
502 BB_COPY_PARTITION (new_bb, bb);
503 BB_END (bb) = insn;
505 /* Redirect the outgoing edges. */
506 new_bb->succs = bb->succs;
507 bb->succs = NULL;
508 FOR_EACH_EDGE (e, ei, new_bb->succs)
509 e->src = new_bb;
511 if (bb->il.rtl->global_live_at_start)
513 new_bb->il.rtl->global_live_at_start = ALLOC_REG_SET (&reg_obstack);
514 new_bb->il.rtl->global_live_at_end = ALLOC_REG_SET (&reg_obstack);
515 COPY_REG_SET (new_bb->il.rtl->global_live_at_end, bb->il.rtl->global_live_at_end);
517 /* We now have to calculate which registers are live at the end
518 of the split basic block and at the start of the new basic
519 block. Start with those registers that are known to be live
520 at the end of the original basic block and get
521 propagate_block to determine which registers are live. */
522 COPY_REG_SET (new_bb->il.rtl->global_live_at_start, bb->il.rtl->global_live_at_end);
523 propagate_block (new_bb, new_bb->il.rtl->global_live_at_start, NULL, NULL, 0);
524 COPY_REG_SET (bb->il.rtl->global_live_at_end,
525 new_bb->il.rtl->global_live_at_start);
526 #ifdef HAVE_conditional_execution
527 /* In the presence of conditional execution we are not able to update
528 liveness precisely. */
529 if (reload_completed)
531 bb->flags |= BB_DIRTY;
532 new_bb->flags |= BB_DIRTY;
534 #endif
537 return new_bb;
540 /* Blocks A and B are to be merged into a single block A. The insns
541 are already contiguous. */
543 static void
544 rtl_merge_blocks (basic_block a, basic_block b)
546 rtx b_head = BB_HEAD (b), b_end = BB_END (b), a_end = BB_END (a);
547 rtx del_first = NULL_RTX, del_last = NULL_RTX;
548 int b_empty = 0;
550 /* If there was a CODE_LABEL beginning B, delete it. */
551 if (LABEL_P (b_head))
553 /* This might have been an EH label that no longer has incoming
554 EH edges. Update data structures to match. */
555 maybe_remove_eh_handler (b_head);
557 /* Detect basic blocks with nothing but a label. This can happen
558 in particular at the end of a function. */
559 if (b_head == b_end)
560 b_empty = 1;
562 del_first = del_last = b_head;
563 b_head = NEXT_INSN (b_head);
566 /* Delete the basic block note and handle blocks containing just that
567 note. */
568 if (NOTE_INSN_BASIC_BLOCK_P (b_head))
570 if (b_head == b_end)
571 b_empty = 1;
572 if (! del_last)
573 del_first = b_head;
575 del_last = b_head;
576 b_head = NEXT_INSN (b_head);
579 /* If there was a jump out of A, delete it. */
580 if (JUMP_P (a_end))
582 rtx prev;
584 for (prev = PREV_INSN (a_end); ; prev = PREV_INSN (prev))
585 if (!NOTE_P (prev)
586 || NOTE_LINE_NUMBER (prev) == NOTE_INSN_BASIC_BLOCK
587 || prev == BB_HEAD (a))
588 break;
590 del_first = a_end;
592 #ifdef HAVE_cc0
593 /* If this was a conditional jump, we need to also delete
594 the insn that set cc0. */
595 if (only_sets_cc0_p (prev))
597 rtx tmp = prev;
599 prev = prev_nonnote_insn (prev);
600 if (!prev)
601 prev = BB_HEAD (a);
602 del_first = tmp;
604 #endif
606 a_end = PREV_INSN (del_first);
608 else if (BARRIER_P (NEXT_INSN (a_end)))
609 del_first = NEXT_INSN (a_end);
611 /* Delete everything marked above as well as crap that might be
612 hanging out between the two blocks. */
613 BB_HEAD (b) = NULL;
614 delete_insn_chain (del_first, del_last);
616 /* Reassociate the insns of B with A. */
617 if (!b_empty)
619 rtx x;
621 for (x = a_end; x != b_end; x = NEXT_INSN (x))
622 set_block_for_insn (x, a);
624 set_block_for_insn (b_end, a);
626 a_end = b_end;
629 BB_END (a) = a_end;
630 a->il.rtl->global_live_at_end = b->il.rtl->global_live_at_end;
633 /* Return true when block A and B can be merged. */
634 static bool
635 rtl_can_merge_blocks (basic_block a,basic_block b)
637 /* If we are partitioning hot/cold basic blocks, we don't want to
638 mess up unconditional or indirect jumps that cross between hot
639 and cold sections.
641 Basic block partitioning may result in some jumps that appear to
642 be optimizable (or blocks that appear to be mergeable), but which really
643 must be left untouched (they are required to make it safely across
644 partition boundaries). See the comments at the top of
645 bb-reorder.c:partition_hot_cold_basic_blocks for complete details. */
647 if (BB_PARTITION (a) != BB_PARTITION (b))
648 return false;
650 /* There must be exactly one edge in between the blocks. */
651 return (single_succ_p (a)
652 && single_succ (a) == b
653 && single_pred_p (b)
654 && a != b
655 /* Must be simple edge. */
656 && !(single_succ_edge (a)->flags & EDGE_COMPLEX)
657 && a->next_bb == b
658 && a != ENTRY_BLOCK_PTR && b != EXIT_BLOCK_PTR
659 /* If the jump insn has side effects,
660 we can't kill the edge. */
661 && (!JUMP_P (BB_END (a))
662 || (reload_completed
663 ? simplejump_p (BB_END (a)) : onlyjump_p (BB_END (a)))));
666 /* Return the label in the head of basic block BLOCK. Create one if it doesn't
667 exist. */
670 block_label (basic_block block)
672 if (block == EXIT_BLOCK_PTR)
673 return NULL_RTX;
675 if (!LABEL_P (BB_HEAD (block)))
677 BB_HEAD (block) = emit_label_before (gen_label_rtx (), BB_HEAD (block));
680 return BB_HEAD (block);
683 /* Attempt to perform edge redirection by replacing possibly complex jump
684 instruction by unconditional jump or removing jump completely. This can
685 apply only if all edges now point to the same block. The parameters and
686 return values are equivalent to redirect_edge_and_branch. */
688 edge
689 try_redirect_by_replacing_jump (edge e, basic_block target, bool in_cfglayout)
691 basic_block src = e->src;
692 rtx insn = BB_END (src), kill_from;
693 rtx set;
694 int fallthru = 0;
696 /* If we are partitioning hot/cold basic blocks, we don't want to
697 mess up unconditional or indirect jumps that cross between hot
698 and cold sections.
700 Basic block partitioning may result in some jumps that appear to
701 be optimizable (or blocks that appear to be mergeable), but which really
702 must be left untouched (they are required to make it safely across
703 partition boundaries). See the comments at the top of
704 bb-reorder.c:partition_hot_cold_basic_blocks for complete details. */
706 if (find_reg_note (insn, REG_CROSSING_JUMP, NULL_RTX)
707 || BB_PARTITION (src) != BB_PARTITION (target))
708 return NULL;
710 /* We can replace or remove a complex jump only when we have exactly
711 two edges. Also, if we have exactly one outgoing edge, we can
712 redirect that. */
713 if (EDGE_COUNT (src->succs) >= 3
714 /* Verify that all targets will be TARGET. Specifically, the
715 edge that is not E must also go to TARGET. */
716 || (EDGE_COUNT (src->succs) == 2
717 && EDGE_SUCC (src, EDGE_SUCC (src, 0) == e)->dest != target))
718 return NULL;
720 if (!onlyjump_p (insn))
721 return NULL;
722 if ((!optimize || reload_completed) && tablejump_p (insn, NULL, NULL))
723 return NULL;
725 /* Avoid removing branch with side effects. */
726 set = single_set (insn);
727 if (!set || side_effects_p (set))
728 return NULL;
730 /* In case we zap a conditional jump, we'll need to kill
731 the cc0 setter too. */
732 kill_from = insn;
733 #ifdef HAVE_cc0
734 if (reg_mentioned_p (cc0_rtx, PATTERN (insn)))
735 kill_from = PREV_INSN (insn);
736 #endif
738 /* See if we can create the fallthru edge. */
739 if (in_cfglayout || can_fallthru (src, target))
741 if (dump_file)
742 fprintf (dump_file, "Removing jump %i.\n", INSN_UID (insn));
743 fallthru = 1;
745 /* Selectively unlink whole insn chain. */
746 if (in_cfglayout)
748 rtx insn = src->il.rtl->footer;
750 delete_insn_chain (kill_from, BB_END (src));
752 /* Remove barriers but keep jumptables. */
753 while (insn)
755 if (BARRIER_P (insn))
757 if (PREV_INSN (insn))
758 NEXT_INSN (PREV_INSN (insn)) = NEXT_INSN (insn);
759 else
760 src->il.rtl->footer = NEXT_INSN (insn);
761 if (NEXT_INSN (insn))
762 PREV_INSN (NEXT_INSN (insn)) = PREV_INSN (insn);
764 if (LABEL_P (insn))
765 break;
766 insn = NEXT_INSN (insn);
769 else
770 delete_insn_chain (kill_from, PREV_INSN (BB_HEAD (target)));
773 /* If this already is simplejump, redirect it. */
774 else if (simplejump_p (insn))
776 if (e->dest == target)
777 return NULL;
778 if (dump_file)
779 fprintf (dump_file, "Redirecting jump %i from %i to %i.\n",
780 INSN_UID (insn), e->dest->index, target->index);
781 if (!redirect_jump (insn, block_label (target), 0))
783 gcc_assert (target == EXIT_BLOCK_PTR);
784 return NULL;
788 /* Cannot do anything for target exit block. */
789 else if (target == EXIT_BLOCK_PTR)
790 return NULL;
792 /* Or replace possibly complicated jump insn by simple jump insn. */
793 else
795 rtx target_label = block_label (target);
796 rtx barrier, label, table;
798 emit_jump_insn_after_noloc (gen_jump (target_label), insn);
799 JUMP_LABEL (BB_END (src)) = target_label;
800 LABEL_NUSES (target_label)++;
801 if (dump_file)
802 fprintf (dump_file, "Replacing insn %i by jump %i\n",
803 INSN_UID (insn), INSN_UID (BB_END (src)));
806 delete_insn_chain (kill_from, insn);
808 /* Recognize a tablejump that we are converting to a
809 simple jump and remove its associated CODE_LABEL
810 and ADDR_VEC or ADDR_DIFF_VEC. */
811 if (tablejump_p (insn, &label, &table))
812 delete_insn_chain (label, table);
814 barrier = next_nonnote_insn (BB_END (src));
815 if (!barrier || !BARRIER_P (barrier))
816 emit_barrier_after (BB_END (src));
817 else
819 if (barrier != NEXT_INSN (BB_END (src)))
821 /* Move the jump before barrier so that the notes
822 which originally were or were created before jump table are
823 inside the basic block. */
824 rtx new_insn = BB_END (src);
825 rtx tmp;
827 for (tmp = NEXT_INSN (BB_END (src)); tmp != barrier;
828 tmp = NEXT_INSN (tmp))
829 set_block_for_insn (tmp, src);
831 NEXT_INSN (PREV_INSN (new_insn)) = NEXT_INSN (new_insn);
832 PREV_INSN (NEXT_INSN (new_insn)) = PREV_INSN (new_insn);
834 NEXT_INSN (new_insn) = barrier;
835 NEXT_INSN (PREV_INSN (barrier)) = new_insn;
837 PREV_INSN (new_insn) = PREV_INSN (barrier);
838 PREV_INSN (barrier) = new_insn;
843 /* Keep only one edge out and set proper flags. */
844 if (!single_succ_p (src))
845 remove_edge (e);
846 gcc_assert (single_succ_p (src));
848 e = single_succ_edge (src);
849 if (fallthru)
850 e->flags = EDGE_FALLTHRU;
851 else
852 e->flags = 0;
854 e->probability = REG_BR_PROB_BASE;
855 e->count = src->count;
857 /* We don't want a block to end on a line-number note since that has
858 the potential of changing the code between -g and not -g. */
859 while (NOTE_P (BB_END (e->src))
860 && NOTE_LINE_NUMBER (BB_END (e->src)) >= 0)
861 delete_insn (BB_END (e->src));
863 if (e->dest != target)
864 redirect_edge_succ (e, target);
866 return e;
869 /* Redirect edge representing branch of (un)conditional jump or tablejump,
870 NULL on failure */
871 static edge
872 redirect_branch_edge (edge e, basic_block target)
874 rtx tmp;
875 rtx old_label = BB_HEAD (e->dest);
876 basic_block src = e->src;
877 rtx insn = BB_END (src);
879 /* We can only redirect non-fallthru edges of jump insn. */
880 if (e->flags & EDGE_FALLTHRU)
881 return NULL;
882 else if (!JUMP_P (insn))
883 return NULL;
885 /* Recognize a tablejump and adjust all matching cases. */
886 if (tablejump_p (insn, NULL, &tmp))
888 rtvec vec;
889 int j;
890 rtx new_label = block_label (target);
892 if (target == EXIT_BLOCK_PTR)
893 return NULL;
894 if (GET_CODE (PATTERN (tmp)) == ADDR_VEC)
895 vec = XVEC (PATTERN (tmp), 0);
896 else
897 vec = XVEC (PATTERN (tmp), 1);
899 for (j = GET_NUM_ELEM (vec) - 1; j >= 0; --j)
900 if (XEXP (RTVEC_ELT (vec, j), 0) == old_label)
902 RTVEC_ELT (vec, j) = gen_rtx_LABEL_REF (Pmode, new_label);
903 --LABEL_NUSES (old_label);
904 ++LABEL_NUSES (new_label);
907 /* Handle casesi dispatch insns. */
908 if ((tmp = single_set (insn)) != NULL
909 && SET_DEST (tmp) == pc_rtx
910 && GET_CODE (SET_SRC (tmp)) == IF_THEN_ELSE
911 && GET_CODE (XEXP (SET_SRC (tmp), 2)) == LABEL_REF
912 && XEXP (XEXP (SET_SRC (tmp), 2), 0) == old_label)
914 XEXP (SET_SRC (tmp), 2) = gen_rtx_LABEL_REF (Pmode,
915 new_label);
916 --LABEL_NUSES (old_label);
917 ++LABEL_NUSES (new_label);
920 else
922 /* ?? We may play the games with moving the named labels from
923 one basic block to the other in case only one computed_jump is
924 available. */
925 if (computed_jump_p (insn)
926 /* A return instruction can't be redirected. */
927 || returnjump_p (insn))
928 return NULL;
930 /* If the insn doesn't go where we think, we're confused. */
931 gcc_assert (JUMP_LABEL (insn) == old_label);
933 /* If the substitution doesn't succeed, die. This can happen
934 if the back end emitted unrecognizable instructions or if
935 target is exit block on some arches. */
936 if (!redirect_jump (insn, block_label (target), 0))
938 gcc_assert (target == EXIT_BLOCK_PTR);
939 return NULL;
943 if (dump_file)
944 fprintf (dump_file, "Edge %i->%i redirected to %i\n",
945 e->src->index, e->dest->index, target->index);
947 if (e->dest != target)
948 e = redirect_edge_succ_nodup (e, target);
949 return e;
952 /* Attempt to change code to redirect edge E to TARGET. Don't do that on
953 expense of adding new instructions or reordering basic blocks.
955 Function can be also called with edge destination equivalent to the TARGET.
956 Then it should try the simplifications and do nothing if none is possible.
958 Return edge representing the branch if transformation succeeded. Return NULL
959 on failure.
960 We still return NULL in case E already destinated TARGET and we didn't
961 managed to simplify instruction stream. */
963 static edge
964 rtl_redirect_edge_and_branch (edge e, basic_block target)
966 edge ret;
967 basic_block src = e->src;
969 if (e->flags & (EDGE_ABNORMAL_CALL | EDGE_EH))
970 return NULL;
972 if (e->dest == target)
973 return e;
975 if ((ret = try_redirect_by_replacing_jump (e, target, false)) != NULL)
977 src->flags |= BB_DIRTY;
978 return ret;
981 ret = redirect_branch_edge (e, target);
982 if (!ret)
983 return NULL;
985 src->flags |= BB_DIRTY;
986 return ret;
989 /* Like force_nonfallthru below, but additionally performs redirection
990 Used by redirect_edge_and_branch_force. */
992 static basic_block
993 force_nonfallthru_and_redirect (edge e, basic_block target)
995 basic_block jump_block, new_bb = NULL, src = e->src;
996 rtx note;
997 edge new_edge;
998 int abnormal_edge_flags = 0;
1000 /* In the case the last instruction is conditional jump to the next
1001 instruction, first redirect the jump itself and then continue
1002 by creating a basic block afterwards to redirect fallthru edge. */
1003 if (e->src != ENTRY_BLOCK_PTR && e->dest != EXIT_BLOCK_PTR
1004 && any_condjump_p (BB_END (e->src))
1005 && JUMP_LABEL (BB_END (e->src)) == BB_HEAD (e->dest))
1007 rtx note;
1008 edge b = unchecked_make_edge (e->src, target, 0);
1009 bool redirected;
1011 redirected = redirect_jump (BB_END (e->src), block_label (target), 0);
1012 gcc_assert (redirected);
1014 note = find_reg_note (BB_END (e->src), REG_BR_PROB, NULL_RTX);
1015 if (note)
1017 int prob = INTVAL (XEXP (note, 0));
1019 b->probability = prob;
1020 b->count = e->count * prob / REG_BR_PROB_BASE;
1021 e->probability -= e->probability;
1022 e->count -= b->count;
1023 if (e->probability < 0)
1024 e->probability = 0;
1025 if (e->count < 0)
1026 e->count = 0;
1030 if (e->flags & EDGE_ABNORMAL)
1032 /* Irritating special case - fallthru edge to the same block as abnormal
1033 edge.
1034 We can't redirect abnormal edge, but we still can split the fallthru
1035 one and create separate abnormal edge to original destination.
1036 This allows bb-reorder to make such edge non-fallthru. */
1037 gcc_assert (e->dest == target);
1038 abnormal_edge_flags = e->flags & ~(EDGE_FALLTHRU | EDGE_CAN_FALLTHRU);
1039 e->flags &= EDGE_FALLTHRU | EDGE_CAN_FALLTHRU;
1041 else
1043 gcc_assert (e->flags & EDGE_FALLTHRU);
1044 if (e->src == ENTRY_BLOCK_PTR)
1046 /* We can't redirect the entry block. Create an empty block
1047 at the start of the function which we use to add the new
1048 jump. */
1049 edge tmp;
1050 edge_iterator ei;
1051 bool found = false;
1053 basic_block bb = create_basic_block (BB_HEAD (e->dest), NULL, ENTRY_BLOCK_PTR);
1055 /* Change the existing edge's source to be the new block, and add
1056 a new edge from the entry block to the new block. */
1057 e->src = bb;
1058 for (ei = ei_start (ENTRY_BLOCK_PTR->succs); (tmp = ei_safe_edge (ei)); )
1060 if (tmp == e)
1062 VEC_unordered_remove (edge, ENTRY_BLOCK_PTR->succs, ei.index);
1063 found = true;
1064 break;
1066 else
1067 ei_next (&ei);
1070 gcc_assert (found);
1072 VEC_safe_push (edge, gc, bb->succs, e);
1073 make_single_succ_edge (ENTRY_BLOCK_PTR, bb, EDGE_FALLTHRU);
1077 if (EDGE_COUNT (e->src->succs) >= 2 || abnormal_edge_flags)
1079 /* Create the new structures. */
1081 /* If the old block ended with a tablejump, skip its table
1082 by searching forward from there. Otherwise start searching
1083 forward from the last instruction of the old block. */
1084 if (!tablejump_p (BB_END (e->src), NULL, &note))
1085 note = BB_END (e->src);
1086 note = NEXT_INSN (note);
1088 jump_block = create_basic_block (note, NULL, e->src);
1089 jump_block->count = e->count;
1090 jump_block->frequency = EDGE_FREQUENCY (e);
1091 jump_block->loop_depth = target->loop_depth;
1093 if (target->il.rtl->global_live_at_start)
1095 jump_block->il.rtl->global_live_at_start = ALLOC_REG_SET (&reg_obstack);
1096 jump_block->il.rtl->global_live_at_end = ALLOC_REG_SET (&reg_obstack);
1097 COPY_REG_SET (jump_block->il.rtl->global_live_at_start,
1098 target->il.rtl->global_live_at_start);
1099 COPY_REG_SET (jump_block->il.rtl->global_live_at_end,
1100 target->il.rtl->global_live_at_start);
1103 /* Make sure new block ends up in correct hot/cold section. */
1105 BB_COPY_PARTITION (jump_block, e->src);
1106 if (flag_reorder_blocks_and_partition
1107 && targetm.have_named_sections
1108 && JUMP_P (BB_END (jump_block))
1109 && !any_condjump_p (BB_END (jump_block))
1110 && (EDGE_SUCC (jump_block, 0)->flags & EDGE_CROSSING))
1111 REG_NOTES (BB_END (jump_block)) = gen_rtx_EXPR_LIST (REG_CROSSING_JUMP,
1112 NULL_RTX,
1113 REG_NOTES
1114 (BB_END
1115 (jump_block)));
1117 /* Wire edge in. */
1118 new_edge = make_edge (e->src, jump_block, EDGE_FALLTHRU);
1119 new_edge->probability = e->probability;
1120 new_edge->count = e->count;
1122 /* Redirect old edge. */
1123 redirect_edge_pred (e, jump_block);
1124 e->probability = REG_BR_PROB_BASE;
1126 new_bb = jump_block;
1128 else
1129 jump_block = e->src;
1131 e->flags &= ~EDGE_FALLTHRU;
1132 if (target == EXIT_BLOCK_PTR)
1134 #ifdef HAVE_return
1135 emit_jump_insn_after_noloc (gen_return (), BB_END (jump_block));
1136 #else
1137 gcc_unreachable ();
1138 #endif
1140 else
1142 rtx label = block_label (target);
1143 emit_jump_insn_after_noloc (gen_jump (label), BB_END (jump_block));
1144 JUMP_LABEL (BB_END (jump_block)) = label;
1145 LABEL_NUSES (label)++;
1148 emit_barrier_after (BB_END (jump_block));
1149 redirect_edge_succ_nodup (e, target);
1151 if (abnormal_edge_flags)
1152 make_edge (src, target, abnormal_edge_flags);
1154 return new_bb;
1157 /* Edge E is assumed to be fallthru edge. Emit needed jump instruction
1158 (and possibly create new basic block) to make edge non-fallthru.
1159 Return newly created BB or NULL if none. */
1161 basic_block
1162 force_nonfallthru (edge e)
1164 return force_nonfallthru_and_redirect (e, e->dest);
1167 /* Redirect edge even at the expense of creating new jump insn or
1168 basic block. Return new basic block if created, NULL otherwise.
1169 Conversion must be possible. */
1171 static basic_block
1172 rtl_redirect_edge_and_branch_force (edge e, basic_block target)
1174 if (redirect_edge_and_branch (e, target)
1175 || e->dest == target)
1176 return NULL;
1178 /* In case the edge redirection failed, try to force it to be non-fallthru
1179 and redirect newly created simplejump. */
1180 e->src->flags |= BB_DIRTY;
1181 return force_nonfallthru_and_redirect (e, target);
1184 /* The given edge should potentially be a fallthru edge. If that is in
1185 fact true, delete the jump and barriers that are in the way. */
1187 static void
1188 rtl_tidy_fallthru_edge (edge e)
1190 rtx q;
1191 basic_block b = e->src, c = b->next_bb;
1193 /* ??? In a late-running flow pass, other folks may have deleted basic
1194 blocks by nopping out blocks, leaving multiple BARRIERs between here
1195 and the target label. They ought to be chastised and fixed.
1197 We can also wind up with a sequence of undeletable labels between
1198 one block and the next.
1200 So search through a sequence of barriers, labels, and notes for
1201 the head of block C and assert that we really do fall through. */
1203 for (q = NEXT_INSN (BB_END (b)); q != BB_HEAD (c); q = NEXT_INSN (q))
1204 if (INSN_P (q))
1205 return;
1207 /* Remove what will soon cease being the jump insn from the source block.
1208 If block B consisted only of this single jump, turn it into a deleted
1209 note. */
1210 q = BB_END (b);
1211 if (JUMP_P (q)
1212 && onlyjump_p (q)
1213 && (any_uncondjump_p (q)
1214 || single_succ_p (b)))
1216 #ifdef HAVE_cc0
1217 /* If this was a conditional jump, we need to also delete
1218 the insn that set cc0. */
1219 if (any_condjump_p (q) && only_sets_cc0_p (PREV_INSN (q)))
1220 q = PREV_INSN (q);
1221 #endif
1223 q = PREV_INSN (q);
1225 /* We don't want a block to end on a line-number note since that has
1226 the potential of changing the code between -g and not -g. */
1227 while (NOTE_P (q) && NOTE_LINE_NUMBER (q) >= 0)
1228 q = PREV_INSN (q);
1231 /* Selectively unlink the sequence. */
1232 if (q != PREV_INSN (BB_HEAD (c)))
1233 delete_insn_chain (NEXT_INSN (q), PREV_INSN (BB_HEAD (c)));
1235 e->flags |= EDGE_FALLTHRU;
1238 /* Should move basic block BB after basic block AFTER. NIY. */
1240 static bool
1241 rtl_move_block_after (basic_block bb ATTRIBUTE_UNUSED,
1242 basic_block after ATTRIBUTE_UNUSED)
1244 return false;
1247 /* Split a (typically critical) edge. Return the new block.
1248 The edge must not be abnormal.
1250 ??? The code generally expects to be called on critical edges.
1251 The case of a block ending in an unconditional jump to a
1252 block with multiple predecessors is not handled optimally. */
1254 static basic_block
1255 rtl_split_edge (edge edge_in)
1257 basic_block bb;
1258 rtx before;
1260 /* Abnormal edges cannot be split. */
1261 gcc_assert (!(edge_in->flags & EDGE_ABNORMAL));
1263 /* We are going to place the new block in front of edge destination.
1264 Avoid existence of fallthru predecessors. */
1265 if ((edge_in->flags & EDGE_FALLTHRU) == 0)
1267 edge e;
1268 edge_iterator ei;
1270 FOR_EACH_EDGE (e, ei, edge_in->dest->preds)
1271 if (e->flags & EDGE_FALLTHRU)
1272 break;
1274 if (e)
1275 force_nonfallthru (e);
1278 /* Create the basic block note. */
1279 if (edge_in->dest != EXIT_BLOCK_PTR)
1280 before = BB_HEAD (edge_in->dest);
1281 else
1282 before = NULL_RTX;
1284 /* If this is a fall through edge to the exit block, the blocks might be
1285 not adjacent, and the right place is the after the source. */
1286 if (edge_in->flags & EDGE_FALLTHRU && edge_in->dest == EXIT_BLOCK_PTR)
1288 before = NEXT_INSN (BB_END (edge_in->src));
1289 bb = create_basic_block (before, NULL, edge_in->src);
1290 BB_COPY_PARTITION (bb, edge_in->src);
1292 else
1294 bb = create_basic_block (before, NULL, edge_in->dest->prev_bb);
1295 /* ??? Why not edge_in->dest->prev_bb here? */
1296 BB_COPY_PARTITION (bb, edge_in->dest);
1299 /* ??? This info is likely going to be out of date very soon. */
1300 if (edge_in->dest->il.rtl->global_live_at_start)
1302 bb->il.rtl->global_live_at_start = ALLOC_REG_SET (&reg_obstack);
1303 bb->il.rtl->global_live_at_end = ALLOC_REG_SET (&reg_obstack);
1304 COPY_REG_SET (bb->il.rtl->global_live_at_start,
1305 edge_in->dest->il.rtl->global_live_at_start);
1306 COPY_REG_SET (bb->il.rtl->global_live_at_end,
1307 edge_in->dest->il.rtl->global_live_at_start);
1310 make_single_succ_edge (bb, edge_in->dest, EDGE_FALLTHRU);
1312 /* For non-fallthru edges, we must adjust the predecessor's
1313 jump instruction to target our new block. */
1314 if ((edge_in->flags & EDGE_FALLTHRU) == 0)
1316 edge redirected = redirect_edge_and_branch (edge_in, bb);
1317 gcc_assert (redirected);
1319 else
1320 redirect_edge_succ (edge_in, bb);
1322 return bb;
1325 /* Queue instructions for insertion on an edge between two basic blocks.
1326 The new instructions and basic blocks (if any) will not appear in the
1327 CFG until commit_edge_insertions is called. */
1329 void
1330 insert_insn_on_edge (rtx pattern, edge e)
1332 /* We cannot insert instructions on an abnormal critical edge.
1333 It will be easier to find the culprit if we die now. */
1334 gcc_assert (!((e->flags & EDGE_ABNORMAL) && EDGE_CRITICAL_P (e)));
1336 if (e->insns.r == NULL_RTX)
1337 start_sequence ();
1338 else
1339 push_to_sequence (e->insns.r);
1341 emit_insn (pattern);
1343 e->insns.r = get_insns ();
1344 end_sequence ();
1347 /* Update the CFG for the instructions queued on edge E. */
1349 static void
1350 commit_one_edge_insertion (edge e, int watch_calls)
1352 rtx before = NULL_RTX, after = NULL_RTX, insns, tmp, last;
1353 basic_block bb = NULL;
1355 /* Pull the insns off the edge now since the edge might go away. */
1356 insns = e->insns.r;
1357 e->insns.r = NULL_RTX;
1359 /* Special case -- avoid inserting code between call and storing
1360 its return value. */
1361 if (watch_calls && (e->flags & EDGE_FALLTHRU)
1362 && single_pred_p (e->dest)
1363 && e->src != ENTRY_BLOCK_PTR
1364 && CALL_P (BB_END (e->src)))
1366 rtx next = next_nonnote_insn (BB_END (e->src));
1368 after = BB_HEAD (e->dest);
1369 /* The first insn after the call may be a stack pop, skip it. */
1370 while (next
1371 && keep_with_call_p (next))
1373 after = next;
1374 next = next_nonnote_insn (next);
1376 bb = e->dest;
1378 if (!before && !after)
1380 /* Figure out where to put these things. If the destination has
1381 one predecessor, insert there. Except for the exit block. */
1382 if (single_pred_p (e->dest) && e->dest != EXIT_BLOCK_PTR)
1384 bb = e->dest;
1386 /* Get the location correct wrt a code label, and "nice" wrt
1387 a basic block note, and before everything else. */
1388 tmp = BB_HEAD (bb);
1389 if (LABEL_P (tmp))
1390 tmp = NEXT_INSN (tmp);
1391 if (NOTE_INSN_BASIC_BLOCK_P (tmp))
1392 tmp = NEXT_INSN (tmp);
1393 if (tmp == BB_HEAD (bb))
1394 before = tmp;
1395 else if (tmp)
1396 after = PREV_INSN (tmp);
1397 else
1398 after = get_last_insn ();
1401 /* If the source has one successor and the edge is not abnormal,
1402 insert there. Except for the entry block. */
1403 else if ((e->flags & EDGE_ABNORMAL) == 0
1404 && single_succ_p (e->src)
1405 && e->src != ENTRY_BLOCK_PTR)
1407 bb = e->src;
1409 /* It is possible to have a non-simple jump here. Consider a target
1410 where some forms of unconditional jumps clobber a register. This
1411 happens on the fr30 for example.
1413 We know this block has a single successor, so we can just emit
1414 the queued insns before the jump. */
1415 if (JUMP_P (BB_END (bb)))
1416 before = BB_END (bb);
1417 else
1419 /* We'd better be fallthru, or we've lost track of
1420 what's what. */
1421 gcc_assert (e->flags & EDGE_FALLTHRU);
1423 after = BB_END (bb);
1426 /* Otherwise we must split the edge. */
1427 else
1429 bb = split_edge (e);
1430 after = BB_END (bb);
1432 if (flag_reorder_blocks_and_partition
1433 && targetm.have_named_sections
1434 && e->src != ENTRY_BLOCK_PTR
1435 && BB_PARTITION (e->src) == BB_COLD_PARTITION
1436 && !(e->flags & EDGE_CROSSING))
1438 rtx bb_note, cur_insn;
1440 bb_note = NULL_RTX;
1441 for (cur_insn = BB_HEAD (bb); cur_insn != NEXT_INSN (BB_END (bb));
1442 cur_insn = NEXT_INSN (cur_insn))
1443 if (NOTE_P (cur_insn)
1444 && NOTE_LINE_NUMBER (cur_insn) == NOTE_INSN_BASIC_BLOCK)
1446 bb_note = cur_insn;
1447 break;
1450 if (JUMP_P (BB_END (bb))
1451 && !any_condjump_p (BB_END (bb))
1452 && (single_succ_edge (bb)->flags & EDGE_CROSSING))
1453 REG_NOTES (BB_END (bb)) = gen_rtx_EXPR_LIST
1454 (REG_CROSSING_JUMP, NULL_RTX, REG_NOTES (BB_END (bb)));
1459 /* Now that we've found the spot, do the insertion. */
1461 if (before)
1463 emit_insn_before_noloc (insns, before);
1464 last = prev_nonnote_insn (before);
1466 else
1467 last = emit_insn_after_noloc (insns, after);
1469 if (returnjump_p (last))
1471 /* ??? Remove all outgoing edges from BB and add one for EXIT.
1472 This is not currently a problem because this only happens
1473 for the (single) epilogue, which already has a fallthru edge
1474 to EXIT. */
1476 e = single_succ_edge (bb);
1477 gcc_assert (e->dest == EXIT_BLOCK_PTR
1478 && single_succ_p (bb) && (e->flags & EDGE_FALLTHRU));
1480 e->flags &= ~EDGE_FALLTHRU;
1481 emit_barrier_after (last);
1483 if (before)
1484 delete_insn (before);
1486 else
1487 gcc_assert (!JUMP_P (last));
1489 /* Mark the basic block for find_many_sub_basic_blocks. */
1490 bb->aux = &bb->aux;
1493 /* Update the CFG for all queued instructions. */
1495 void
1496 commit_edge_insertions (void)
1498 basic_block bb;
1499 sbitmap blocks;
1500 bool changed = false;
1502 #ifdef ENABLE_CHECKING
1503 verify_flow_info ();
1504 #endif
1506 FOR_BB_BETWEEN (bb, ENTRY_BLOCK_PTR, EXIT_BLOCK_PTR, next_bb)
1508 edge e;
1509 edge_iterator ei;
1511 FOR_EACH_EDGE (e, ei, bb->succs)
1512 if (e->insns.r)
1514 changed = true;
1515 commit_one_edge_insertion (e, false);
1519 if (!changed)
1520 return;
1522 blocks = sbitmap_alloc (last_basic_block);
1523 sbitmap_zero (blocks);
1524 FOR_EACH_BB (bb)
1525 if (bb->aux)
1527 SET_BIT (blocks, bb->index);
1528 /* Check for forgotten bb->aux values before commit_edge_insertions
1529 call. */
1530 gcc_assert (bb->aux == &bb->aux);
1531 bb->aux = NULL;
1533 find_many_sub_basic_blocks (blocks);
1534 sbitmap_free (blocks);
1537 /* Update the CFG for all queued instructions, taking special care of inserting
1538 code on edges between call and storing its return value. */
1540 void
1541 commit_edge_insertions_watch_calls (void)
1543 basic_block bb;
1544 sbitmap blocks;
1545 bool changed = false;
1547 #ifdef ENABLE_CHECKING
1548 verify_flow_info ();
1549 #endif
1551 FOR_BB_BETWEEN (bb, ENTRY_BLOCK_PTR, EXIT_BLOCK_PTR, next_bb)
1553 edge e;
1554 edge_iterator ei;
1556 FOR_EACH_EDGE (e, ei, bb->succs)
1557 if (e->insns.r)
1559 changed = true;
1560 commit_one_edge_insertion (e, true);
1564 if (!changed)
1565 return;
1567 blocks = sbitmap_alloc (last_basic_block);
1568 sbitmap_zero (blocks);
1569 FOR_EACH_BB (bb)
1570 if (bb->aux)
1572 SET_BIT (blocks, bb->index);
1573 /* Check for forgotten bb->aux values before commit_edge_insertions
1574 call. */
1575 gcc_assert (bb->aux == &bb->aux);
1576 bb->aux = NULL;
1578 find_many_sub_basic_blocks (blocks);
1579 sbitmap_free (blocks);
1582 /* Print out RTL-specific basic block information (live information
1583 at start and end). */
1585 static void
1586 rtl_dump_bb (basic_block bb, FILE *outf, int indent)
1588 rtx insn;
1589 rtx last;
1590 char *s_indent;
1592 s_indent = alloca ((size_t) indent + 1);
1593 memset (s_indent, ' ', (size_t) indent);
1594 s_indent[indent] = '\0';
1596 fprintf (outf, ";;%s Registers live at start: ", s_indent);
1597 dump_regset (bb->il.rtl->global_live_at_start, outf);
1598 putc ('\n', outf);
1600 for (insn = BB_HEAD (bb), last = NEXT_INSN (BB_END (bb)); insn != last;
1601 insn = NEXT_INSN (insn))
1602 print_rtl_single (outf, insn);
1604 fprintf (outf, ";;%s Registers live at end: ", s_indent);
1605 dump_regset (bb->il.rtl->global_live_at_end, outf);
1606 putc ('\n', outf);
1609 /* Like print_rtl, but also print out live information for the start of each
1610 basic block. */
1612 void
1613 print_rtl_with_bb (FILE *outf, rtx rtx_first)
1615 rtx tmp_rtx;
1617 if (rtx_first == 0)
1618 fprintf (outf, "(nil)\n");
1619 else
1621 enum bb_state { NOT_IN_BB, IN_ONE_BB, IN_MULTIPLE_BB };
1622 int max_uid = get_max_uid ();
1623 basic_block *start = XCNEWVEC (basic_block, max_uid);
1624 basic_block *end = XCNEWVEC (basic_block, max_uid);
1625 enum bb_state *in_bb_p = XCNEWVEC (enum bb_state, max_uid);
1627 basic_block bb;
1629 FOR_EACH_BB_REVERSE (bb)
1631 rtx x;
1633 start[INSN_UID (BB_HEAD (bb))] = bb;
1634 end[INSN_UID (BB_END (bb))] = bb;
1635 for (x = BB_HEAD (bb); x != NULL_RTX; x = NEXT_INSN (x))
1637 enum bb_state state = IN_MULTIPLE_BB;
1639 if (in_bb_p[INSN_UID (x)] == NOT_IN_BB)
1640 state = IN_ONE_BB;
1641 in_bb_p[INSN_UID (x)] = state;
1643 if (x == BB_END (bb))
1644 break;
1648 for (tmp_rtx = rtx_first; NULL != tmp_rtx; tmp_rtx = NEXT_INSN (tmp_rtx))
1650 int did_output;
1652 if ((bb = start[INSN_UID (tmp_rtx)]) != NULL)
1654 fprintf (outf, ";; Start of basic block %d, registers live:",
1655 bb->index);
1656 dump_regset (bb->il.rtl->global_live_at_start, outf);
1657 putc ('\n', outf);
1660 if (in_bb_p[INSN_UID (tmp_rtx)] == NOT_IN_BB
1661 && !NOTE_P (tmp_rtx)
1662 && !BARRIER_P (tmp_rtx))
1663 fprintf (outf, ";; Insn is not within a basic block\n");
1664 else if (in_bb_p[INSN_UID (tmp_rtx)] == IN_MULTIPLE_BB)
1665 fprintf (outf, ";; Insn is in multiple basic blocks\n");
1667 did_output = print_rtl_single (outf, tmp_rtx);
1669 if ((bb = end[INSN_UID (tmp_rtx)]) != NULL)
1671 fprintf (outf, ";; End of basic block %d, registers live:\n",
1672 bb->index);
1673 dump_regset (bb->il.rtl->global_live_at_end, outf);
1674 putc ('\n', outf);
1677 if (did_output)
1678 putc ('\n', outf);
1681 free (start);
1682 free (end);
1683 free (in_bb_p);
1686 if (current_function_epilogue_delay_list != 0)
1688 fprintf (outf, "\n;; Insns in epilogue delay list:\n\n");
1689 for (tmp_rtx = current_function_epilogue_delay_list; tmp_rtx != 0;
1690 tmp_rtx = XEXP (tmp_rtx, 1))
1691 print_rtl_single (outf, XEXP (tmp_rtx, 0));
1695 void
1696 update_br_prob_note (basic_block bb)
1698 rtx note;
1699 if (!JUMP_P (BB_END (bb)))
1700 return;
1701 note = find_reg_note (BB_END (bb), REG_BR_PROB, NULL_RTX);
1702 if (!note || INTVAL (XEXP (note, 0)) == BRANCH_EDGE (bb)->probability)
1703 return;
1704 XEXP (note, 0) = GEN_INT (BRANCH_EDGE (bb)->probability);
1707 /* Get the last insn associated with block BB (that includes barriers and
1708 tablejumps after BB). */
1710 get_last_bb_insn (basic_block bb)
1712 rtx tmp;
1713 rtx end = BB_END (bb);
1715 /* Include any jump table following the basic block. */
1716 if (tablejump_p (end, NULL, &tmp))
1717 end = tmp;
1719 /* Include any barriers that may follow the basic block. */
1720 tmp = next_nonnote_insn (end);
1721 while (tmp && BARRIER_P (tmp))
1723 end = tmp;
1724 tmp = next_nonnote_insn (end);
1727 return end;
1730 /* Verify the CFG and RTL consistency common for both underlying RTL and
1731 cfglayout RTL.
1733 Currently it does following checks:
1735 - test head/end pointers
1736 - overlapping of basic blocks
1737 - headers of basic blocks (the NOTE_INSN_BASIC_BLOCK note)
1738 - tails of basic blocks (ensure that boundary is necessary)
1739 - scans body of the basic block for JUMP_INSN, CODE_LABEL
1740 and NOTE_INSN_BASIC_BLOCK
1741 - verify that no fall_thru edge crosses hot/cold partition boundaries
1743 In future it can be extended check a lot of other stuff as well
1744 (reachability of basic blocks, life information, etc. etc.). */
1746 static int
1747 rtl_verify_flow_info_1 (void)
1749 const int max_uid = get_max_uid ();
1750 rtx last_head = get_last_insn ();
1751 basic_block *bb_info;
1752 rtx x;
1753 int err = 0;
1754 basic_block bb;
1756 bb_info = XCNEWVEC (basic_block, max_uid);
1758 FOR_EACH_BB_REVERSE (bb)
1760 rtx head = BB_HEAD (bb);
1761 rtx end = BB_END (bb);
1763 /* Verify the end of the basic block is in the INSN chain. */
1764 for (x = last_head; x != NULL_RTX; x = PREV_INSN (x))
1765 if (x == end)
1766 break;
1768 if (!(bb->flags & BB_RTL))
1770 error ("BB_RTL flag not set for block %d", bb->index);
1771 err = 1;
1774 if (!x)
1776 error ("end insn %d for block %d not found in the insn stream",
1777 INSN_UID (end), bb->index);
1778 err = 1;
1781 /* Work backwards from the end to the head of the basic block
1782 to verify the head is in the RTL chain. */
1783 for (; x != NULL_RTX; x = PREV_INSN (x))
1785 /* While walking over the insn chain, verify insns appear
1786 in only one basic block and initialize the BB_INFO array
1787 used by other passes. */
1788 if (bb_info[INSN_UID (x)] != NULL)
1790 error ("insn %d is in multiple basic blocks (%d and %d)",
1791 INSN_UID (x), bb->index, bb_info[INSN_UID (x)]->index);
1792 err = 1;
1795 bb_info[INSN_UID (x)] = bb;
1797 if (x == head)
1798 break;
1800 if (!x)
1802 error ("head insn %d for block %d not found in the insn stream",
1803 INSN_UID (head), bb->index);
1804 err = 1;
1807 last_head = x;
1810 /* Now check the basic blocks (boundaries etc.) */
1811 FOR_EACH_BB_REVERSE (bb)
1813 int n_fallthru = 0, n_eh = 0, n_call = 0, n_abnormal = 0, n_branch = 0;
1814 edge e, fallthru = NULL;
1815 rtx note;
1816 edge_iterator ei;
1818 if (JUMP_P (BB_END (bb))
1819 && (note = find_reg_note (BB_END (bb), REG_BR_PROB, NULL_RTX))
1820 && EDGE_COUNT (bb->succs) >= 2
1821 && any_condjump_p (BB_END (bb)))
1823 if (INTVAL (XEXP (note, 0)) != BRANCH_EDGE (bb)->probability
1824 && profile_status != PROFILE_ABSENT)
1826 error ("verify_flow_info: REG_BR_PROB does not match cfg %wi %i",
1827 INTVAL (XEXP (note, 0)), BRANCH_EDGE (bb)->probability);
1828 err = 1;
1831 FOR_EACH_EDGE (e, ei, bb->succs)
1833 if (e->flags & EDGE_FALLTHRU)
1835 n_fallthru++, fallthru = e;
1836 if ((e->flags & EDGE_CROSSING)
1837 || (BB_PARTITION (e->src) != BB_PARTITION (e->dest)
1838 && e->src != ENTRY_BLOCK_PTR
1839 && e->dest != EXIT_BLOCK_PTR))
1841 error ("fallthru edge crosses section boundary (bb %i)",
1842 e->src->index);
1843 err = 1;
1847 if ((e->flags & ~(EDGE_DFS_BACK
1848 | EDGE_CAN_FALLTHRU
1849 | EDGE_IRREDUCIBLE_LOOP
1850 | EDGE_LOOP_EXIT
1851 | EDGE_CROSSING)) == 0)
1852 n_branch++;
1854 if (e->flags & EDGE_ABNORMAL_CALL)
1855 n_call++;
1857 if (e->flags & EDGE_EH)
1858 n_eh++;
1859 else if (e->flags & EDGE_ABNORMAL)
1860 n_abnormal++;
1863 if (n_eh && GET_CODE (PATTERN (BB_END (bb))) != RESX
1864 && !find_reg_note (BB_END (bb), REG_EH_REGION, NULL_RTX))
1866 error ("missing REG_EH_REGION note in the end of bb %i", bb->index);
1867 err = 1;
1869 if (n_branch
1870 && (!JUMP_P (BB_END (bb))
1871 || (n_branch > 1 && (any_uncondjump_p (BB_END (bb))
1872 || any_condjump_p (BB_END (bb))))))
1874 error ("too many outgoing branch edges from bb %i", bb->index);
1875 err = 1;
1877 if (n_fallthru && any_uncondjump_p (BB_END (bb)))
1879 error ("fallthru edge after unconditional jump %i", bb->index);
1880 err = 1;
1882 if (n_branch != 1 && any_uncondjump_p (BB_END (bb)))
1884 error ("wrong amount of branch edges after unconditional jump %i", bb->index);
1885 err = 1;
1887 if (n_branch != 1 && any_condjump_p (BB_END (bb))
1888 && JUMP_LABEL (BB_END (bb)) != BB_HEAD (fallthru->dest))
1890 error ("wrong amount of branch edges after conditional jump %i",
1891 bb->index);
1892 err = 1;
1894 if (n_call && !CALL_P (BB_END (bb)))
1896 error ("call edges for non-call insn in bb %i", bb->index);
1897 err = 1;
1899 if (n_abnormal
1900 && (!CALL_P (BB_END (bb)) && n_call != n_abnormal)
1901 && (!JUMP_P (BB_END (bb))
1902 || any_condjump_p (BB_END (bb))
1903 || any_uncondjump_p (BB_END (bb))))
1905 error ("abnormal edges for no purpose in bb %i", bb->index);
1906 err = 1;
1909 for (x = BB_HEAD (bb); x != NEXT_INSN (BB_END (bb)); x = NEXT_INSN (x))
1910 /* We may have a barrier inside a basic block before dead code
1911 elimination. There is no BLOCK_FOR_INSN field in a barrier. */
1912 if (!BARRIER_P (x) && BLOCK_FOR_INSN (x) != bb)
1914 debug_rtx (x);
1915 if (! BLOCK_FOR_INSN (x))
1916 error
1917 ("insn %d inside basic block %d but block_for_insn is NULL",
1918 INSN_UID (x), bb->index);
1919 else
1920 error
1921 ("insn %d inside basic block %d but block_for_insn is %i",
1922 INSN_UID (x), bb->index, BLOCK_FOR_INSN (x)->index);
1924 err = 1;
1927 /* OK pointers are correct. Now check the header of basic
1928 block. It ought to contain optional CODE_LABEL followed
1929 by NOTE_BASIC_BLOCK. */
1930 x = BB_HEAD (bb);
1931 if (LABEL_P (x))
1933 if (BB_END (bb) == x)
1935 error ("NOTE_INSN_BASIC_BLOCK is missing for block %d",
1936 bb->index);
1937 err = 1;
1940 x = NEXT_INSN (x);
1943 if (!NOTE_INSN_BASIC_BLOCK_P (x) || NOTE_BASIC_BLOCK (x) != bb)
1945 error ("NOTE_INSN_BASIC_BLOCK is missing for block %d",
1946 bb->index);
1947 err = 1;
1950 if (BB_END (bb) == x)
1951 /* Do checks for empty blocks here. */
1953 else
1954 for (x = NEXT_INSN (x); x; x = NEXT_INSN (x))
1956 if (NOTE_INSN_BASIC_BLOCK_P (x))
1958 error ("NOTE_INSN_BASIC_BLOCK %d in middle of basic block %d",
1959 INSN_UID (x), bb->index);
1960 err = 1;
1963 if (x == BB_END (bb))
1964 break;
1966 if (control_flow_insn_p (x))
1968 error ("in basic block %d:", bb->index);
1969 fatal_insn ("flow control insn inside a basic block", x);
1974 /* Clean up. */
1975 free (bb_info);
1976 return err;
1979 /* Verify the CFG and RTL consistency common for both underlying RTL and
1980 cfglayout RTL.
1982 Currently it does following checks:
1983 - all checks of rtl_verify_flow_info_1
1984 - check that all insns are in the basic blocks
1985 (except the switch handling code, barriers and notes)
1986 - check that all returns are followed by barriers
1987 - check that all fallthru edge points to the adjacent blocks. */
1988 static int
1989 rtl_verify_flow_info (void)
1991 basic_block bb;
1992 int err = rtl_verify_flow_info_1 ();
1993 rtx x;
1994 int num_bb_notes;
1995 const rtx rtx_first = get_insns ();
1996 basic_block last_bb_seen = ENTRY_BLOCK_PTR, curr_bb = NULL;
1998 FOR_EACH_BB_REVERSE (bb)
2000 edge e;
2001 edge_iterator ei;
2003 if (bb->predictions)
2005 error ("bb prediction set for block %i, but it is not used in RTL land", bb->index);
2006 err = 1;
2009 FOR_EACH_EDGE (e, ei, bb->succs)
2010 if (e->flags & EDGE_FALLTHRU)
2011 break;
2012 if (!e)
2014 rtx insn;
2016 /* Ensure existence of barrier in BB with no fallthru edges. */
2017 for (insn = BB_END (bb); !insn || !BARRIER_P (insn);
2018 insn = NEXT_INSN (insn))
2019 if (!insn
2020 || (NOTE_P (insn)
2021 && NOTE_LINE_NUMBER (insn) == NOTE_INSN_BASIC_BLOCK))
2023 error ("missing barrier after block %i", bb->index);
2024 err = 1;
2025 break;
2028 else if (e->src != ENTRY_BLOCK_PTR
2029 && e->dest != EXIT_BLOCK_PTR)
2031 rtx insn;
2033 if (e->src->next_bb != e->dest)
2035 error
2036 ("verify_flow_info: Incorrect blocks for fallthru %i->%i",
2037 e->src->index, e->dest->index);
2038 err = 1;
2040 else
2041 for (insn = NEXT_INSN (BB_END (e->src)); insn != BB_HEAD (e->dest);
2042 insn = NEXT_INSN (insn))
2043 if (BARRIER_P (insn) || INSN_P (insn))
2045 error ("verify_flow_info: Incorrect fallthru %i->%i",
2046 e->src->index, e->dest->index);
2047 fatal_insn ("wrong insn in the fallthru edge", insn);
2048 err = 1;
2053 num_bb_notes = 0;
2054 last_bb_seen = ENTRY_BLOCK_PTR;
2056 for (x = rtx_first; x; x = NEXT_INSN (x))
2058 if (NOTE_INSN_BASIC_BLOCK_P (x))
2060 bb = NOTE_BASIC_BLOCK (x);
2062 num_bb_notes++;
2063 if (bb != last_bb_seen->next_bb)
2064 internal_error ("basic blocks not laid down consecutively");
2066 curr_bb = last_bb_seen = bb;
2069 if (!curr_bb)
2071 switch (GET_CODE (x))
2073 case BARRIER:
2074 case NOTE:
2075 break;
2077 case CODE_LABEL:
2078 /* An addr_vec is placed outside any basic block. */
2079 if (NEXT_INSN (x)
2080 && JUMP_P (NEXT_INSN (x))
2081 && (GET_CODE (PATTERN (NEXT_INSN (x))) == ADDR_DIFF_VEC
2082 || GET_CODE (PATTERN (NEXT_INSN (x))) == ADDR_VEC))
2083 x = NEXT_INSN (x);
2085 /* But in any case, non-deletable labels can appear anywhere. */
2086 break;
2088 default:
2089 fatal_insn ("insn outside basic block", x);
2093 if (JUMP_P (x)
2094 && returnjump_p (x) && ! condjump_p (x)
2095 && ! (NEXT_INSN (x) && BARRIER_P (NEXT_INSN (x))))
2096 fatal_insn ("return not followed by barrier", x);
2097 if (curr_bb && x == BB_END (curr_bb))
2098 curr_bb = NULL;
2101 if (num_bb_notes != n_basic_blocks - NUM_FIXED_BLOCKS)
2102 internal_error
2103 ("number of bb notes in insn chain (%d) != n_basic_blocks (%d)",
2104 num_bb_notes, n_basic_blocks);
2106 return err;
2109 /* Assume that the preceding pass has possibly eliminated jump instructions
2110 or converted the unconditional jumps. Eliminate the edges from CFG.
2111 Return true if any edges are eliminated. */
2113 bool
2114 purge_dead_edges (basic_block bb)
2116 edge e;
2117 rtx insn = BB_END (bb), note;
2118 bool purged = false;
2119 bool found;
2120 edge_iterator ei;
2122 /* If this instruction cannot trap, remove REG_EH_REGION notes. */
2123 if (NONJUMP_INSN_P (insn)
2124 && (note = find_reg_note (insn, REG_EH_REGION, NULL)))
2126 rtx eqnote;
2128 if (! may_trap_p (PATTERN (insn))
2129 || ((eqnote = find_reg_equal_equiv_note (insn))
2130 && ! may_trap_p (XEXP (eqnote, 0))))
2131 remove_note (insn, note);
2134 /* Cleanup abnormal edges caused by exceptions or non-local gotos. */
2135 for (ei = ei_start (bb->succs); (e = ei_safe_edge (ei)); )
2137 /* There are three types of edges we need to handle correctly here: EH
2138 edges, abnormal call EH edges, and abnormal call non-EH edges. The
2139 latter can appear when nonlocal gotos are used. */
2140 if (e->flags & EDGE_EH)
2142 if (can_throw_internal (BB_END (bb))
2143 /* If this is a call edge, verify that this is a call insn. */
2144 && (! (e->flags & EDGE_ABNORMAL_CALL)
2145 || CALL_P (BB_END (bb))))
2147 ei_next (&ei);
2148 continue;
2151 else if (e->flags & EDGE_ABNORMAL_CALL)
2153 if (CALL_P (BB_END (bb))
2154 && (! (note = find_reg_note (insn, REG_EH_REGION, NULL))
2155 || INTVAL (XEXP (note, 0)) >= 0))
2157 ei_next (&ei);
2158 continue;
2161 else
2163 ei_next (&ei);
2164 continue;
2167 remove_edge (e);
2168 bb->flags |= BB_DIRTY;
2169 purged = true;
2172 if (JUMP_P (insn))
2174 rtx note;
2175 edge b,f;
2176 edge_iterator ei;
2178 /* We do care only about conditional jumps and simplejumps. */
2179 if (!any_condjump_p (insn)
2180 && !returnjump_p (insn)
2181 && !simplejump_p (insn))
2182 return purged;
2184 /* Branch probability/prediction notes are defined only for
2185 condjumps. We've possibly turned condjump into simplejump. */
2186 if (simplejump_p (insn))
2188 note = find_reg_note (insn, REG_BR_PROB, NULL);
2189 if (note)
2190 remove_note (insn, note);
2191 while ((note = find_reg_note (insn, REG_BR_PRED, NULL)))
2192 remove_note (insn, note);
2195 for (ei = ei_start (bb->succs); (e = ei_safe_edge (ei)); )
2197 /* Avoid abnormal flags to leak from computed jumps turned
2198 into simplejumps. */
2200 e->flags &= ~EDGE_ABNORMAL;
2202 /* See if this edge is one we should keep. */
2203 if ((e->flags & EDGE_FALLTHRU) && any_condjump_p (insn))
2204 /* A conditional jump can fall through into the next
2205 block, so we should keep the edge. */
2207 ei_next (&ei);
2208 continue;
2210 else if (e->dest != EXIT_BLOCK_PTR
2211 && BB_HEAD (e->dest) == JUMP_LABEL (insn))
2212 /* If the destination block is the target of the jump,
2213 keep the edge. */
2215 ei_next (&ei);
2216 continue;
2218 else if (e->dest == EXIT_BLOCK_PTR && returnjump_p (insn))
2219 /* If the destination block is the exit block, and this
2220 instruction is a return, then keep the edge. */
2222 ei_next (&ei);
2223 continue;
2225 else if ((e->flags & EDGE_EH) && can_throw_internal (insn))
2226 /* Keep the edges that correspond to exceptions thrown by
2227 this instruction and rematerialize the EDGE_ABNORMAL
2228 flag we just cleared above. */
2230 e->flags |= EDGE_ABNORMAL;
2231 ei_next (&ei);
2232 continue;
2235 /* We do not need this edge. */
2236 bb->flags |= BB_DIRTY;
2237 purged = true;
2238 remove_edge (e);
2241 if (EDGE_COUNT (bb->succs) == 0 || !purged)
2242 return purged;
2244 if (dump_file)
2245 fprintf (dump_file, "Purged edges from bb %i\n", bb->index);
2247 if (!optimize)
2248 return purged;
2250 /* Redistribute probabilities. */
2251 if (single_succ_p (bb))
2253 single_succ_edge (bb)->probability = REG_BR_PROB_BASE;
2254 single_succ_edge (bb)->count = bb->count;
2256 else
2258 note = find_reg_note (insn, REG_BR_PROB, NULL);
2259 if (!note)
2260 return purged;
2262 b = BRANCH_EDGE (bb);
2263 f = FALLTHRU_EDGE (bb);
2264 b->probability = INTVAL (XEXP (note, 0));
2265 f->probability = REG_BR_PROB_BASE - b->probability;
2266 b->count = bb->count * b->probability / REG_BR_PROB_BASE;
2267 f->count = bb->count * f->probability / REG_BR_PROB_BASE;
2270 return purged;
2272 else if (CALL_P (insn) && SIBLING_CALL_P (insn))
2274 /* First, there should not be any EH or ABCALL edges resulting
2275 from non-local gotos and the like. If there were, we shouldn't
2276 have created the sibcall in the first place. Second, there
2277 should of course never have been a fallthru edge. */
2278 gcc_assert (single_succ_p (bb));
2279 gcc_assert (single_succ_edge (bb)->flags
2280 == (EDGE_SIBCALL | EDGE_ABNORMAL));
2282 return 0;
2285 /* If we don't see a jump insn, we don't know exactly why the block would
2286 have been broken at this point. Look for a simple, non-fallthru edge,
2287 as these are only created by conditional branches. If we find such an
2288 edge we know that there used to be a jump here and can then safely
2289 remove all non-fallthru edges. */
2290 found = false;
2291 FOR_EACH_EDGE (e, ei, bb->succs)
2292 if (! (e->flags & (EDGE_COMPLEX | EDGE_FALLTHRU)))
2294 found = true;
2295 break;
2298 if (!found)
2299 return purged;
2301 /* Remove all but the fake and fallthru edges. The fake edge may be
2302 the only successor for this block in the case of noreturn
2303 calls. */
2304 for (ei = ei_start (bb->succs); (e = ei_safe_edge (ei)); )
2306 if (!(e->flags & (EDGE_FALLTHRU | EDGE_FAKE)))
2308 bb->flags |= BB_DIRTY;
2309 remove_edge (e);
2310 purged = true;
2312 else
2313 ei_next (&ei);
2316 gcc_assert (single_succ_p (bb));
2318 single_succ_edge (bb)->probability = REG_BR_PROB_BASE;
2319 single_succ_edge (bb)->count = bb->count;
2321 if (dump_file)
2322 fprintf (dump_file, "Purged non-fallthru edges from bb %i\n",
2323 bb->index);
2324 return purged;
2327 /* Search all basic blocks for potentially dead edges and purge them. Return
2328 true if some edge has been eliminated. */
2330 bool
2331 purge_all_dead_edges (void)
2333 int purged = false;
2334 basic_block bb;
2336 FOR_EACH_BB (bb)
2338 bool purged_here = purge_dead_edges (bb);
2340 purged |= purged_here;
2343 return purged;
2346 /* Same as split_block but update cfg_layout structures. */
2348 static basic_block
2349 cfg_layout_split_block (basic_block bb, void *insnp)
2351 rtx insn = insnp;
2352 basic_block new_bb = rtl_split_block (bb, insn);
2354 new_bb->il.rtl->footer = bb->il.rtl->footer;
2355 bb->il.rtl->footer = NULL;
2357 return new_bb;
2361 /* Redirect Edge to DEST. */
2362 static edge
2363 cfg_layout_redirect_edge_and_branch (edge e, basic_block dest)
2365 basic_block src = e->src;
2366 edge ret;
2368 if (e->flags & (EDGE_ABNORMAL_CALL | EDGE_EH))
2369 return NULL;
2371 if (e->dest == dest)
2372 return e;
2374 if (e->src != ENTRY_BLOCK_PTR
2375 && (ret = try_redirect_by_replacing_jump (e, dest, true)))
2377 src->flags |= BB_DIRTY;
2378 return ret;
2381 if (e->src == ENTRY_BLOCK_PTR
2382 && (e->flags & EDGE_FALLTHRU) && !(e->flags & EDGE_COMPLEX))
2384 if (dump_file)
2385 fprintf (dump_file, "Redirecting entry edge from bb %i to %i\n",
2386 e->src->index, dest->index);
2388 e->src->flags |= BB_DIRTY;
2389 redirect_edge_succ (e, dest);
2390 return e;
2393 /* Redirect_edge_and_branch may decide to turn branch into fallthru edge
2394 in the case the basic block appears to be in sequence. Avoid this
2395 transformation. */
2397 if (e->flags & EDGE_FALLTHRU)
2399 /* Redirect any branch edges unified with the fallthru one. */
2400 if (JUMP_P (BB_END (src))
2401 && label_is_jump_target_p (BB_HEAD (e->dest),
2402 BB_END (src)))
2404 edge redirected;
2406 if (dump_file)
2407 fprintf (dump_file, "Fallthru edge unified with branch "
2408 "%i->%i redirected to %i\n",
2409 e->src->index, e->dest->index, dest->index);
2410 e->flags &= ~EDGE_FALLTHRU;
2411 redirected = redirect_branch_edge (e, dest);
2412 gcc_assert (redirected);
2413 e->flags |= EDGE_FALLTHRU;
2414 e->src->flags |= BB_DIRTY;
2415 return e;
2417 /* In case we are redirecting fallthru edge to the branch edge
2418 of conditional jump, remove it. */
2419 if (EDGE_COUNT (src->succs) == 2)
2421 /* Find the edge that is different from E. */
2422 edge s = EDGE_SUCC (src, EDGE_SUCC (src, 0) == e);
2424 if (s->dest == dest
2425 && any_condjump_p (BB_END (src))
2426 && onlyjump_p (BB_END (src)))
2427 delete_insn (BB_END (src));
2429 ret = redirect_edge_succ_nodup (e, dest);
2430 if (dump_file)
2431 fprintf (dump_file, "Fallthru edge %i->%i redirected to %i\n",
2432 e->src->index, e->dest->index, dest->index);
2434 else
2435 ret = redirect_branch_edge (e, dest);
2437 /* We don't want simplejumps in the insn stream during cfglayout. */
2438 gcc_assert (!simplejump_p (BB_END (src)));
2440 src->flags |= BB_DIRTY;
2441 return ret;
2444 /* Simple wrapper as we always can redirect fallthru edges. */
2445 static basic_block
2446 cfg_layout_redirect_edge_and_branch_force (edge e, basic_block dest)
2448 edge redirected = cfg_layout_redirect_edge_and_branch (e, dest);
2450 gcc_assert (redirected);
2451 return NULL;
2454 /* Same as delete_basic_block but update cfg_layout structures. */
2456 static void
2457 cfg_layout_delete_block (basic_block bb)
2459 rtx insn, next, prev = PREV_INSN (BB_HEAD (bb)), *to, remaints;
2461 if (bb->il.rtl->header)
2463 next = BB_HEAD (bb);
2464 if (prev)
2465 NEXT_INSN (prev) = bb->il.rtl->header;
2466 else
2467 set_first_insn (bb->il.rtl->header);
2468 PREV_INSN (bb->il.rtl->header) = prev;
2469 insn = bb->il.rtl->header;
2470 while (NEXT_INSN (insn))
2471 insn = NEXT_INSN (insn);
2472 NEXT_INSN (insn) = next;
2473 PREV_INSN (next) = insn;
2475 next = NEXT_INSN (BB_END (bb));
2476 if (bb->il.rtl->footer)
2478 insn = bb->il.rtl->footer;
2479 while (insn)
2481 if (BARRIER_P (insn))
2483 if (PREV_INSN (insn))
2484 NEXT_INSN (PREV_INSN (insn)) = NEXT_INSN (insn);
2485 else
2486 bb->il.rtl->footer = NEXT_INSN (insn);
2487 if (NEXT_INSN (insn))
2488 PREV_INSN (NEXT_INSN (insn)) = PREV_INSN (insn);
2490 if (LABEL_P (insn))
2491 break;
2492 insn = NEXT_INSN (insn);
2494 if (bb->il.rtl->footer)
2496 insn = BB_END (bb);
2497 NEXT_INSN (insn) = bb->il.rtl->footer;
2498 PREV_INSN (bb->il.rtl->footer) = insn;
2499 while (NEXT_INSN (insn))
2500 insn = NEXT_INSN (insn);
2501 NEXT_INSN (insn) = next;
2502 if (next)
2503 PREV_INSN (next) = insn;
2504 else
2505 set_last_insn (insn);
2508 if (bb->next_bb != EXIT_BLOCK_PTR)
2509 to = &bb->next_bb->il.rtl->header;
2510 else
2511 to = &cfg_layout_function_footer;
2513 rtl_delete_block (bb);
2515 if (prev)
2516 prev = NEXT_INSN (prev);
2517 else
2518 prev = get_insns ();
2519 if (next)
2520 next = PREV_INSN (next);
2521 else
2522 next = get_last_insn ();
2524 if (next && NEXT_INSN (next) != prev)
2526 remaints = unlink_insn_chain (prev, next);
2527 insn = remaints;
2528 while (NEXT_INSN (insn))
2529 insn = NEXT_INSN (insn);
2530 NEXT_INSN (insn) = *to;
2531 if (*to)
2532 PREV_INSN (*to) = insn;
2533 *to = remaints;
2537 /* Return true when blocks A and B can be safely merged. */
2538 static bool
2539 cfg_layout_can_merge_blocks_p (basic_block a, basic_block b)
2541 /* If we are partitioning hot/cold basic blocks, we don't want to
2542 mess up unconditional or indirect jumps that cross between hot
2543 and cold sections.
2545 Basic block partitioning may result in some jumps that appear to
2546 be optimizable (or blocks that appear to be mergeable), but which really
2547 must be left untouched (they are required to make it safely across
2548 partition boundaries). See the comments at the top of
2549 bb-reorder.c:partition_hot_cold_basic_blocks for complete details. */
2551 if (BB_PARTITION (a) != BB_PARTITION (b))
2552 return false;
2554 /* There must be exactly one edge in between the blocks. */
2555 return (single_succ_p (a)
2556 && single_succ (a) == b
2557 && single_pred_p (b) == 1
2558 && a != b
2559 /* Must be simple edge. */
2560 && !(single_succ_edge (a)->flags & EDGE_COMPLEX)
2561 && a != ENTRY_BLOCK_PTR && b != EXIT_BLOCK_PTR
2562 /* If the jump insn has side effects,
2563 we can't kill the edge. */
2564 && (!JUMP_P (BB_END (a))
2565 || (reload_completed
2566 ? simplejump_p (BB_END (a)) : onlyjump_p (BB_END (a)))));
2569 /* Merge block A and B. The blocks must be mergeable. */
2571 static void
2572 cfg_layout_merge_blocks (basic_block a, basic_block b)
2574 #ifdef ENABLE_CHECKING
2575 gcc_assert (cfg_layout_can_merge_blocks_p (a, b));
2576 #endif
2578 /* If there was a CODE_LABEL beginning B, delete it. */
2579 if (LABEL_P (BB_HEAD (b)))
2581 /* This might have been an EH label that no longer has incoming
2582 EH edges. Update data structures to match. */
2583 maybe_remove_eh_handler (BB_HEAD (b));
2585 delete_insn (BB_HEAD (b));
2588 /* We should have fallthru edge in a, or we can do dummy redirection to get
2589 it cleaned up. */
2590 if (JUMP_P (BB_END (a)))
2591 try_redirect_by_replacing_jump (EDGE_SUCC (a, 0), b, true);
2592 gcc_assert (!JUMP_P (BB_END (a)));
2594 /* Possible line number notes should appear in between. */
2595 if (b->il.rtl->header)
2597 rtx first = BB_END (a), last;
2599 last = emit_insn_after_noloc (b->il.rtl->header, BB_END (a));
2600 delete_insn_chain (NEXT_INSN (first), last);
2601 b->il.rtl->header = NULL;
2604 /* In the case basic blocks are not adjacent, move them around. */
2605 if (NEXT_INSN (BB_END (a)) != BB_HEAD (b))
2607 rtx first = unlink_insn_chain (BB_HEAD (b), BB_END (b));
2609 emit_insn_after_noloc (first, BB_END (a));
2610 /* Skip possible DELETED_LABEL insn. */
2611 if (!NOTE_INSN_BASIC_BLOCK_P (first))
2612 first = NEXT_INSN (first);
2613 gcc_assert (NOTE_INSN_BASIC_BLOCK_P (first));
2614 BB_HEAD (b) = NULL;
2615 delete_insn (first);
2617 /* Otherwise just re-associate the instructions. */
2618 else
2620 rtx insn;
2622 for (insn = BB_HEAD (b);
2623 insn != NEXT_INSN (BB_END (b));
2624 insn = NEXT_INSN (insn))
2625 set_block_for_insn (insn, a);
2626 insn = BB_HEAD (b);
2627 /* Skip possible DELETED_LABEL insn. */
2628 if (!NOTE_INSN_BASIC_BLOCK_P (insn))
2629 insn = NEXT_INSN (insn);
2630 gcc_assert (NOTE_INSN_BASIC_BLOCK_P (insn));
2631 BB_HEAD (b) = NULL;
2632 BB_END (a) = BB_END (b);
2633 delete_insn (insn);
2636 /* Possible tablejumps and barriers should appear after the block. */
2637 if (b->il.rtl->footer)
2639 if (!a->il.rtl->footer)
2640 a->il.rtl->footer = b->il.rtl->footer;
2641 else
2643 rtx last = a->il.rtl->footer;
2645 while (NEXT_INSN (last))
2646 last = NEXT_INSN (last);
2647 NEXT_INSN (last) = b->il.rtl->footer;
2648 PREV_INSN (b->il.rtl->footer) = last;
2650 b->il.rtl->footer = NULL;
2652 a->il.rtl->global_live_at_end = b->il.rtl->global_live_at_end;
2654 if (dump_file)
2655 fprintf (dump_file, "Merged blocks %d and %d.\n",
2656 a->index, b->index);
2659 /* Split edge E. */
2661 static basic_block
2662 cfg_layout_split_edge (edge e)
2664 basic_block new_bb =
2665 create_basic_block (e->src != ENTRY_BLOCK_PTR
2666 ? NEXT_INSN (BB_END (e->src)) : get_insns (),
2667 NULL_RTX, e->src);
2669 /* ??? This info is likely going to be out of date very soon, but we must
2670 create it to avoid getting an ICE later. */
2671 if (e->dest->il.rtl->global_live_at_start)
2673 new_bb->il.rtl->global_live_at_start = ALLOC_REG_SET (&reg_obstack);
2674 new_bb->il.rtl->global_live_at_end = ALLOC_REG_SET (&reg_obstack);
2675 COPY_REG_SET (new_bb->il.rtl->global_live_at_start,
2676 e->dest->il.rtl->global_live_at_start);
2677 COPY_REG_SET (new_bb->il.rtl->global_live_at_end,
2678 e->dest->il.rtl->global_live_at_start);
2681 make_edge (new_bb, e->dest, EDGE_FALLTHRU);
2682 redirect_edge_and_branch_force (e, new_bb);
2684 return new_bb;
2687 /* Do postprocessing after making a forwarder block joined by edge FALLTHRU. */
2689 static void
2690 rtl_make_forwarder_block (edge fallthru ATTRIBUTE_UNUSED)
2694 /* Return 1 if BB ends with a call, possibly followed by some
2695 instructions that must stay with the call, 0 otherwise. */
2697 static bool
2698 rtl_block_ends_with_call_p (basic_block bb)
2700 rtx insn = BB_END (bb);
2702 while (!CALL_P (insn)
2703 && insn != BB_HEAD (bb)
2704 && keep_with_call_p (insn))
2705 insn = PREV_INSN (insn);
2706 return (CALL_P (insn));
2709 /* Return 1 if BB ends with a conditional branch, 0 otherwise. */
2711 static bool
2712 rtl_block_ends_with_condjump_p (basic_block bb)
2714 return any_condjump_p (BB_END (bb));
2717 /* Return true if we need to add fake edge to exit.
2718 Helper function for rtl_flow_call_edges_add. */
2720 static bool
2721 need_fake_edge_p (rtx insn)
2723 if (!INSN_P (insn))
2724 return false;
2726 if ((CALL_P (insn)
2727 && !SIBLING_CALL_P (insn)
2728 && !find_reg_note (insn, REG_NORETURN, NULL)
2729 && !CONST_OR_PURE_CALL_P (insn)))
2730 return true;
2732 return ((GET_CODE (PATTERN (insn)) == ASM_OPERANDS
2733 && MEM_VOLATILE_P (PATTERN (insn)))
2734 || (GET_CODE (PATTERN (insn)) == PARALLEL
2735 && asm_noperands (insn) != -1
2736 && MEM_VOLATILE_P (XVECEXP (PATTERN (insn), 0, 0)))
2737 || GET_CODE (PATTERN (insn)) == ASM_INPUT);
2740 /* Add fake edges to the function exit for any non constant and non noreturn
2741 calls, volatile inline assembly in the bitmap of blocks specified by
2742 BLOCKS or to the whole CFG if BLOCKS is zero. Return the number of blocks
2743 that were split.
2745 The goal is to expose cases in which entering a basic block does not imply
2746 that all subsequent instructions must be executed. */
2748 static int
2749 rtl_flow_call_edges_add (sbitmap blocks)
2751 int i;
2752 int blocks_split = 0;
2753 int last_bb = last_basic_block;
2754 bool check_last_block = false;
2756 if (n_basic_blocks == NUM_FIXED_BLOCKS)
2757 return 0;
2759 if (! blocks)
2760 check_last_block = true;
2761 else
2762 check_last_block = TEST_BIT (blocks, EXIT_BLOCK_PTR->prev_bb->index);
2764 /* In the last basic block, before epilogue generation, there will be
2765 a fallthru edge to EXIT. Special care is required if the last insn
2766 of the last basic block is a call because make_edge folds duplicate
2767 edges, which would result in the fallthru edge also being marked
2768 fake, which would result in the fallthru edge being removed by
2769 remove_fake_edges, which would result in an invalid CFG.
2771 Moreover, we can't elide the outgoing fake edge, since the block
2772 profiler needs to take this into account in order to solve the minimal
2773 spanning tree in the case that the call doesn't return.
2775 Handle this by adding a dummy instruction in a new last basic block. */
2776 if (check_last_block)
2778 basic_block bb = EXIT_BLOCK_PTR->prev_bb;
2779 rtx insn = BB_END (bb);
2781 /* Back up past insns that must be kept in the same block as a call. */
2782 while (insn != BB_HEAD (bb)
2783 && keep_with_call_p (insn))
2784 insn = PREV_INSN (insn);
2786 if (need_fake_edge_p (insn))
2788 edge e;
2790 e = find_edge (bb, EXIT_BLOCK_PTR);
2791 if (e)
2793 insert_insn_on_edge (gen_rtx_USE (VOIDmode, const0_rtx), e);
2794 commit_edge_insertions ();
2799 /* Now add fake edges to the function exit for any non constant
2800 calls since there is no way that we can determine if they will
2801 return or not... */
2803 for (i = NUM_FIXED_BLOCKS; i < last_bb; i++)
2805 basic_block bb = BASIC_BLOCK (i);
2806 rtx insn;
2807 rtx prev_insn;
2809 if (!bb)
2810 continue;
2812 if (blocks && !TEST_BIT (blocks, i))
2813 continue;
2815 for (insn = BB_END (bb); ; insn = prev_insn)
2817 prev_insn = PREV_INSN (insn);
2818 if (need_fake_edge_p (insn))
2820 edge e;
2821 rtx split_at_insn = insn;
2823 /* Don't split the block between a call and an insn that should
2824 remain in the same block as the call. */
2825 if (CALL_P (insn))
2826 while (split_at_insn != BB_END (bb)
2827 && keep_with_call_p (NEXT_INSN (split_at_insn)))
2828 split_at_insn = NEXT_INSN (split_at_insn);
2830 /* The handling above of the final block before the epilogue
2831 should be enough to verify that there is no edge to the exit
2832 block in CFG already. Calling make_edge in such case would
2833 cause us to mark that edge as fake and remove it later. */
2835 #ifdef ENABLE_CHECKING
2836 if (split_at_insn == BB_END (bb))
2838 e = find_edge (bb, EXIT_BLOCK_PTR);
2839 gcc_assert (e == NULL);
2841 #endif
2843 /* Note that the following may create a new basic block
2844 and renumber the existing basic blocks. */
2845 if (split_at_insn != BB_END (bb))
2847 e = split_block (bb, split_at_insn);
2848 if (e)
2849 blocks_split++;
2852 make_edge (bb, EXIT_BLOCK_PTR, EDGE_FAKE);
2855 if (insn == BB_HEAD (bb))
2856 break;
2860 if (blocks_split)
2861 verify_flow_info ();
2863 return blocks_split;
2866 /* Add COMP_RTX as a condition at end of COND_BB. FIRST_HEAD is
2867 the conditional branch target, SECOND_HEAD should be the fall-thru
2868 there is no need to handle this here the loop versioning code handles
2869 this. the reason for SECON_HEAD is that it is needed for condition
2870 in trees, and this should be of the same type since it is a hook. */
2871 static void
2872 rtl_lv_add_condition_to_bb (basic_block first_head ,
2873 basic_block second_head ATTRIBUTE_UNUSED,
2874 basic_block cond_bb, void *comp_rtx)
2876 rtx label, seq, jump;
2877 rtx op0 = XEXP ((rtx)comp_rtx, 0);
2878 rtx op1 = XEXP ((rtx)comp_rtx, 1);
2879 enum rtx_code comp = GET_CODE ((rtx)comp_rtx);
2880 enum machine_mode mode;
2883 label = block_label (first_head);
2884 mode = GET_MODE (op0);
2885 if (mode == VOIDmode)
2886 mode = GET_MODE (op1);
2888 start_sequence ();
2889 op0 = force_operand (op0, NULL_RTX);
2890 op1 = force_operand (op1, NULL_RTX);
2891 do_compare_rtx_and_jump (op0, op1, comp, 0,
2892 mode, NULL_RTX, NULL_RTX, label);
2893 jump = get_last_insn ();
2894 JUMP_LABEL (jump) = label;
2895 LABEL_NUSES (label)++;
2896 seq = get_insns ();
2897 end_sequence ();
2899 /* Add the new cond , in the new head. */
2900 emit_insn_after(seq, BB_END(cond_bb));
2904 /* Given a block B with unconditional branch at its end, get the
2905 store the return the branch edge and the fall-thru edge in
2906 BRANCH_EDGE and FALLTHRU_EDGE respectively. */
2907 static void
2908 rtl_extract_cond_bb_edges (basic_block b, edge *branch_edge,
2909 edge *fallthru_edge)
2911 edge e = EDGE_SUCC (b, 0);
2913 if (e->flags & EDGE_FALLTHRU)
2915 *fallthru_edge = e;
2916 *branch_edge = EDGE_SUCC (b, 1);
2918 else
2920 *branch_edge = e;
2921 *fallthru_edge = EDGE_SUCC (b, 1);
2925 void
2926 init_rtl_bb_info (basic_block bb)
2928 gcc_assert (!bb->il.rtl);
2929 bb->il.rtl = ggc_alloc_cleared (sizeof (struct rtl_bb_info));
2933 /* Add EXPR to the end of basic block BB. */
2936 insert_insn_end_bb_new (rtx pat, basic_block bb)
2938 rtx insn = BB_END (bb);
2939 rtx new_insn;
2940 rtx pat_end = pat;
2942 while (NEXT_INSN (pat_end) != NULL_RTX)
2943 pat_end = NEXT_INSN (pat_end);
2945 /* If the last insn is a jump, insert EXPR in front [taking care to
2946 handle cc0, etc. properly]. Similarly we need to care trapping
2947 instructions in presence of non-call exceptions. */
2949 if (JUMP_P (insn)
2950 || (NONJUMP_INSN_P (insn)
2951 && (!single_succ_p (bb)
2952 || single_succ_edge (bb)->flags & EDGE_ABNORMAL)))
2954 #ifdef HAVE_cc0
2955 rtx note;
2956 #endif
2957 /* If this is a jump table, then we can't insert stuff here. Since
2958 we know the previous real insn must be the tablejump, we insert
2959 the new instruction just before the tablejump. */
2960 if (GET_CODE (PATTERN (insn)) == ADDR_VEC
2961 || GET_CODE (PATTERN (insn)) == ADDR_DIFF_VEC)
2962 insn = prev_real_insn (insn);
2964 #ifdef HAVE_cc0
2965 /* FIXME: 'twould be nice to call prev_cc0_setter here but it aborts
2966 if cc0 isn't set. */
2967 note = find_reg_note (insn, REG_CC_SETTER, NULL_RTX);
2968 if (note)
2969 insn = XEXP (note, 0);
2970 else
2972 rtx maybe_cc0_setter = prev_nonnote_insn (insn);
2973 if (maybe_cc0_setter
2974 && INSN_P (maybe_cc0_setter)
2975 && sets_cc0_p (PATTERN (maybe_cc0_setter)))
2976 insn = maybe_cc0_setter;
2978 #endif
2979 /* FIXME: What if something in cc0/jump uses value set in new
2980 insn? */
2981 new_insn = emit_insn_before_noloc (pat, insn);
2984 /* Likewise if the last insn is a call, as will happen in the presence
2985 of exception handling. */
2986 else if (CALL_P (insn)
2987 && (!single_succ_p (bb)
2988 || single_succ_edge (bb)->flags & EDGE_ABNORMAL))
2990 /* Keeping in mind SMALL_REGISTER_CLASSES and parameters in registers,
2991 we search backward and place the instructions before the first
2992 parameter is loaded. Do this for everyone for consistency and a
2993 presumption that we'll get better code elsewhere as well. */
2995 /* Since different machines initialize their parameter registers
2996 in different orders, assume nothing. Collect the set of all
2997 parameter registers. */
2998 insn = find_first_parameter_load (insn, BB_HEAD (bb));
3000 /* If we found all the parameter loads, then we want to insert
3001 before the first parameter load.
3003 If we did not find all the parameter loads, then we might have
3004 stopped on the head of the block, which could be a CODE_LABEL.
3005 If we inserted before the CODE_LABEL, then we would be putting
3006 the insn in the wrong basic block. In that case, put the insn
3007 after the CODE_LABEL. Also, respect NOTE_INSN_BASIC_BLOCK. */
3008 while (LABEL_P (insn)
3009 || NOTE_INSN_BASIC_BLOCK_P (insn))
3010 insn = NEXT_INSN (insn);
3012 new_insn = emit_insn_before_noloc (pat, insn);
3014 else
3015 new_insn = emit_insn_after_noloc (pat, insn);
3017 return new_insn;
3020 /* Implementation of CFG manipulation for linearized RTL. */
3021 struct cfg_hooks rtl_cfg_hooks = {
3022 "rtl",
3023 rtl_verify_flow_info,
3024 rtl_dump_bb,
3025 rtl_create_basic_block,
3026 rtl_redirect_edge_and_branch,
3027 rtl_redirect_edge_and_branch_force,
3028 rtl_delete_block,
3029 rtl_split_block,
3030 rtl_move_block_after,
3031 rtl_can_merge_blocks, /* can_merge_blocks_p */
3032 rtl_merge_blocks,
3033 rtl_predict_edge,
3034 rtl_predicted_by_p,
3035 NULL, /* can_duplicate_block_p */
3036 NULL, /* duplicate_block */
3037 rtl_split_edge,
3038 rtl_make_forwarder_block,
3039 rtl_tidy_fallthru_edge,
3040 rtl_block_ends_with_call_p,
3041 rtl_block_ends_with_condjump_p,
3042 rtl_flow_call_edges_add,
3043 NULL, /* execute_on_growing_pred */
3044 NULL, /* execute_on_shrinking_pred */
3045 NULL, /* duplicate loop for trees */
3046 NULL, /* lv_add_condition_to_bb */
3047 NULL, /* lv_adjust_loop_header_phi*/
3048 NULL, /* extract_cond_bb_edges */
3049 NULL /* flush_pending_stmts */
3052 /* Implementation of CFG manipulation for cfg layout RTL, where
3053 basic block connected via fallthru edges does not have to be adjacent.
3054 This representation will hopefully become the default one in future
3055 version of the compiler. */
3057 /* We do not want to declare these functions in a header file, since they
3058 should only be used through the cfghooks interface, and we do not want to
3059 move them here since it would require also moving quite a lot of related
3060 code. */
3061 extern bool cfg_layout_can_duplicate_bb_p (basic_block);
3062 extern basic_block cfg_layout_duplicate_bb (basic_block);
3064 struct cfg_hooks cfg_layout_rtl_cfg_hooks = {
3065 "cfglayout mode",
3066 rtl_verify_flow_info_1,
3067 rtl_dump_bb,
3068 cfg_layout_create_basic_block,
3069 cfg_layout_redirect_edge_and_branch,
3070 cfg_layout_redirect_edge_and_branch_force,
3071 cfg_layout_delete_block,
3072 cfg_layout_split_block,
3073 rtl_move_block_after,
3074 cfg_layout_can_merge_blocks_p,
3075 cfg_layout_merge_blocks,
3076 rtl_predict_edge,
3077 rtl_predicted_by_p,
3078 cfg_layout_can_duplicate_bb_p,
3079 cfg_layout_duplicate_bb,
3080 cfg_layout_split_edge,
3081 rtl_make_forwarder_block,
3082 NULL,
3083 rtl_block_ends_with_call_p,
3084 rtl_block_ends_with_condjump_p,
3085 rtl_flow_call_edges_add,
3086 NULL, /* execute_on_growing_pred */
3087 NULL, /* execute_on_shrinking_pred */
3088 duplicate_loop_to_header_edge, /* duplicate loop for trees */
3089 rtl_lv_add_condition_to_bb, /* lv_add_condition_to_bb */
3090 NULL, /* lv_adjust_loop_header_phi*/
3091 rtl_extract_cond_bb_edges, /* extract_cond_bb_edges */
3092 NULL /* flush_pending_stmts */