1 /* CFG cleanup for trees.
2 Copyright (C) 2001, 2002, 2003, 2004, 2005 Free Software Foundation, Inc.
4 This file is part of GCC.
6 GCC is free software; you can redistribute it and/or modify
7 it under the terms of the GNU General Public License as published by
8 the Free Software Foundation; either version 2, or (at your option)
11 GCC is distributed in the hope that it will be useful,
12 but WITHOUT ANY WARRANTY; without even the implied warranty of
13 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 GNU General Public License for more details.
16 You should have received a copy of the GNU General Public License
17 along with GCC; see the file COPYING. If not, write to
18 the Free Software Foundation, 51 Franklin Street, Fifth Floor,
19 Boston, MA 02110-1301, USA. */
23 #include "coretypes.h"
28 #include "hard-reg-set.h"
29 #include "basic-block.h"
36 #include "langhooks.h"
37 #include "diagnostic.h"
38 #include "tree-flow.h"
40 #include "tree-dump.h"
41 #include "tree-pass.h"
45 #include "cfglayout.h"
47 #include "tree-ssa-propagate.h"
48 #include "tree-scalar-evolution.h"
50 /* Remove any fallthru edge from EV. Return true if an edge was removed. */
53 remove_fallthru_edge (VEC(edge
,gc
) *ev
)
58 FOR_EACH_EDGE (e
, ei
, ev
)
59 if ((e
->flags
& EDGE_FALLTHRU
) != 0)
67 /* Disconnect an unreachable block in the control expression starting
71 cleanup_control_expr_graph (basic_block bb
, block_stmt_iterator bsi
)
75 tree expr
= bsi_stmt (bsi
), val
;
77 if (!single_succ_p (bb
))
82 switch (TREE_CODE (expr
))
85 val
= fold (COND_EXPR_COND (expr
));
89 val
= fold (SWITCH_COND (expr
));
90 if (TREE_CODE (val
) != INTEGER_CST
)
98 taken_edge
= find_taken_edge (bb
, val
);
102 /* Remove all the edges except the one that is always executed. */
103 for (ei
= ei_start (bb
->succs
); (e
= ei_safe_edge (ei
)); )
107 taken_edge
->probability
+= e
->probability
;
108 taken_edge
->count
+= e
->count
;
115 if (taken_edge
->probability
> REG_BR_PROB_BASE
)
116 taken_edge
->probability
= REG_BR_PROB_BASE
;
119 taken_edge
= single_succ_edge (bb
);
121 bsi_remove (&bsi
, true);
122 taken_edge
->flags
= EDGE_FALLTHRU
;
124 /* We removed some paths from the cfg. */
125 free_dominance_info (CDI_DOMINATORS
);
130 /* A list of all the noreturn calls passed to modify_stmt.
131 cleanup_control_flow uses it to detect cases where a mid-block
132 indirect call has been turned into a noreturn call. When this
133 happens, all the instructions after the call are no longer
134 reachable and must be deleted as dead. */
136 VEC(tree
,gc
) *modified_noreturn_calls
;
138 /* Try to remove superfluous control structures. */
141 cleanup_control_flow (void)
144 block_stmt_iterator bsi
;
148 /* Detect cases where a mid-block call is now known not to return. */
149 while (VEC_length (tree
, modified_noreturn_calls
))
151 stmt
= VEC_pop (tree
, modified_noreturn_calls
);
152 bb
= bb_for_stmt (stmt
);
153 if (bb
!= NULL
&& last_stmt (bb
) != stmt
&& noreturn_call_p (stmt
))
154 split_block (bb
, stmt
);
161 /* If the last statement of the block could throw and now cannot,
162 we need to prune cfg. */
163 tree_purge_dead_eh_edges (bb
);
168 stmt
= bsi_stmt (bsi
);
170 if (TREE_CODE (stmt
) == COND_EXPR
171 || TREE_CODE (stmt
) == SWITCH_EXPR
)
172 retval
|= cleanup_control_expr_graph (bb
, bsi
);
173 /* If we had a computed goto which has a compile-time determinable
174 destination, then we can eliminate the goto. */
175 else if (TREE_CODE (stmt
) == GOTO_EXPR
176 && TREE_CODE (GOTO_DESTINATION (stmt
)) == ADDR_EXPR
177 && (TREE_CODE (TREE_OPERAND (GOTO_DESTINATION (stmt
), 0))
183 basic_block target_block
;
184 bool removed_edge
= false;
186 /* First look at all the outgoing edges. Delete any outgoing
187 edges which do not go to the right block. For the one
188 edge which goes to the right block, fix up its flags. */
189 label
= TREE_OPERAND (GOTO_DESTINATION (stmt
), 0);
190 target_block
= label_to_block (label
);
191 for (ei
= ei_start (bb
->succs
); (e
= ei_safe_edge (ei
)); )
193 if (e
->dest
!= target_block
)
200 /* Turn off the EDGE_ABNORMAL flag. */
201 e
->flags
&= ~EDGE_ABNORMAL
;
203 /* And set EDGE_FALLTHRU. */
204 e
->flags
|= EDGE_FALLTHRU
;
209 /* If we removed one or more edges, then we will need to fix the
210 dominators. It may be possible to incrementally update them. */
212 free_dominance_info (CDI_DOMINATORS
);
214 /* Remove the GOTO_EXPR as it is not needed. The CFG has all the
215 relevant information we need. */
216 bsi_remove (&bsi
, true);
220 /* Check for indirect calls that have been turned into
222 else if (noreturn_call_p (stmt
) && remove_fallthru_edge (bb
->succs
))
224 free_dominance_info (CDI_DOMINATORS
);
231 /* Return true if basic block BB does nothing except pass control
232 flow to another block and that we can safely insert a label at
233 the start of the successor block.
235 As a precondition, we require that BB be not equal to
239 tree_forwarder_block_p (basic_block bb
, bool phi_wanted
)
241 block_stmt_iterator bsi
;
246 /* BB must have a single outgoing edge. */
247 if (single_succ_p (bb
) != 1
248 /* If PHI_WANTED is false, BB must not have any PHI nodes.
249 Otherwise, BB must have PHI nodes. */
250 || (phi_nodes (bb
) != NULL_TREE
) != phi_wanted
251 /* BB may not be a predecessor of EXIT_BLOCK_PTR. */
252 || single_succ (bb
) == EXIT_BLOCK_PTR
253 /* Nor should this be an infinite loop. */
254 || single_succ (bb
) == bb
255 /* BB may not have an abnormal outgoing edge. */
256 || (single_succ_edge (bb
)->flags
& EDGE_ABNORMAL
))
260 gcc_assert (bb
!= ENTRY_BLOCK_PTR
);
263 /* Now walk through the statements backward. We can ignore labels,
264 anything else means this is not a forwarder block. */
265 for (bsi
= bsi_last (bb
); !bsi_end_p (bsi
); bsi_prev (&bsi
))
267 tree stmt
= bsi_stmt (bsi
);
269 switch (TREE_CODE (stmt
))
272 if (DECL_NONLOCAL (LABEL_EXPR_LABEL (stmt
)))
281 if (find_edge (ENTRY_BLOCK_PTR
, bb
))
287 /* Protect loop latches, headers and preheaders. */
288 if (bb
->loop_father
->header
== bb
)
290 dest
= EDGE_SUCC (bb
, 0)->dest
;
292 if (dest
->loop_father
->header
== dest
)
296 /* If we have an EH edge leaving this block, make sure that the
297 destination of this block has only one predecessor. This ensures
298 that we don't get into the situation where we try to remove two
299 forwarders that go to the same basic block but are handlers for
300 different EH regions. */
301 succ
= single_succ_edge (bb
);
303 FOR_EACH_EDGE (e
, ei
, bb
->preds
)
305 if (e
->flags
& EDGE_EH
)
307 if (!single_pred_p (dest
))
315 /* Return true if BB has at least one abnormal incoming edge. */
318 has_abnormal_incoming_edge_p (basic_block bb
)
323 FOR_EACH_EDGE (e
, ei
, bb
->preds
)
324 if (e
->flags
& EDGE_ABNORMAL
)
330 /* If all the PHI nodes in DEST have alternatives for E1 and E2 and
331 those alternatives are equal in each of the PHI nodes, then return
332 true, else return false. */
335 phi_alternatives_equal (basic_block dest
, edge e1
, edge e2
)
337 int n1
= e1
->dest_idx
;
338 int n2
= e2
->dest_idx
;
341 for (phi
= phi_nodes (dest
); phi
; phi
= PHI_CHAIN (phi
))
343 tree val1
= PHI_ARG_DEF (phi
, n1
);
344 tree val2
= PHI_ARG_DEF (phi
, n2
);
346 gcc_assert (val1
!= NULL_TREE
);
347 gcc_assert (val2
!= NULL_TREE
);
349 if (!operand_equal_for_phi_arg_p (val1
, val2
))
356 /* Removes forwarder block BB. Returns false if this failed. If a new
357 forwarder block is created due to redirection of edges, it is
358 stored to worklist. */
361 remove_forwarder_block (basic_block bb
, basic_block
**worklist
)
363 edge succ
= single_succ_edge (bb
), e
, s
;
364 basic_block dest
= succ
->dest
;
368 block_stmt_iterator bsi
, bsi_to
;
369 bool seen_abnormal_edge
= false;
371 /* We check for infinite loops already in tree_forwarder_block_p.
372 However it may happen that the infinite loop is created
373 afterwards due to removal of forwarders. */
377 /* If the destination block consists of a nonlocal label, do not merge
379 label
= first_stmt (dest
);
381 && TREE_CODE (label
) == LABEL_EXPR
382 && DECL_NONLOCAL (LABEL_EXPR_LABEL (label
)))
385 /* If there is an abnormal edge to basic block BB, but not into
386 dest, problems might occur during removal of the phi node at out
387 of ssa due to overlapping live ranges of registers.
389 If there is an abnormal edge in DEST, the problems would occur
390 anyway since cleanup_dead_labels would then merge the labels for
391 two different eh regions, and rest of exception handling code
394 So if there is an abnormal edge to BB, proceed only if there is
395 no abnormal edge to DEST and there are no phi nodes in DEST. */
396 if (has_abnormal_incoming_edge_p (bb
))
398 seen_abnormal_edge
= true;
400 if (has_abnormal_incoming_edge_p (dest
)
401 || phi_nodes (dest
) != NULL_TREE
)
405 /* If there are phi nodes in DEST, and some of the blocks that are
406 predecessors of BB are also predecessors of DEST, check that the
407 phi node arguments match. */
408 if (phi_nodes (dest
))
410 FOR_EACH_EDGE (e
, ei
, bb
->preds
)
412 s
= find_edge (e
->src
, dest
);
416 if (!phi_alternatives_equal (dest
, succ
, s
))
421 /* Redirect the edges. */
422 for (ei
= ei_start (bb
->preds
); (e
= ei_safe_edge (ei
)); )
424 if (e
->flags
& EDGE_ABNORMAL
)
426 /* If there is an abnormal edge, redirect it anyway, and
427 move the labels to the new block to make it legal. */
428 s
= redirect_edge_succ_nodup (e
, dest
);
431 s
= redirect_edge_and_branch (e
, dest
);
435 /* Create arguments for the phi nodes, since the edge was not
437 for (phi
= phi_nodes (dest
); phi
; phi
= PHI_CHAIN (phi
))
438 add_phi_arg (phi
, PHI_ARG_DEF (phi
, succ
->dest_idx
), s
);
442 /* The source basic block might become a forwarder. We know
443 that it was not a forwarder before, since it used to have
444 at least two outgoing edges, so we may just add it to
446 if (tree_forwarder_block_p (s
->src
, false))
447 *(*worklist
)++ = s
->src
;
451 if (seen_abnormal_edge
)
453 /* Move the labels to the new block, so that the redirection of
454 the abnormal edges works. */
456 bsi_to
= bsi_start (dest
);
457 for (bsi
= bsi_start (bb
); !bsi_end_p (bsi
); )
459 label
= bsi_stmt (bsi
);
460 gcc_assert (TREE_CODE (label
) == LABEL_EXPR
);
461 bsi_remove (&bsi
, false);
462 bsi_insert_before (&bsi_to
, label
, BSI_CONTINUE_LINKING
);
466 /* Update the dominators. */
467 if (dom_info_available_p (CDI_DOMINATORS
))
469 basic_block dom
, dombb
, domdest
;
471 dombb
= get_immediate_dominator (CDI_DOMINATORS
, bb
);
472 domdest
= get_immediate_dominator (CDI_DOMINATORS
, dest
);
475 /* Shortcut to avoid calling (relatively expensive)
476 nearest_common_dominator unless necessary. */
480 dom
= nearest_common_dominator (CDI_DOMINATORS
, domdest
, dombb
);
482 set_immediate_dominator (CDI_DOMINATORS
, dest
, dom
);
485 /* And kill the forwarder block. */
486 delete_basic_block (bb
);
491 /* Removes forwarder blocks. */
494 cleanup_forwarder_blocks (void)
497 bool changed
= false;
498 basic_block
*worklist
= XNEWVEC (basic_block
, n_basic_blocks
);
499 basic_block
*current
= worklist
;
503 if (tree_forwarder_block_p (bb
, false))
507 while (current
!= worklist
)
510 changed
|= remove_forwarder_block (bb
, ¤t
);
517 /* Do one round of CFG cleanup. */
520 cleanup_tree_cfg_1 (void)
524 retval
= cleanup_control_flow ();
525 retval
|= delete_unreachable_blocks ();
527 /* Forwarder blocks can carry line number information which is
528 useful when debugging, so we only clean them up when
533 /* cleanup_forwarder_blocks can redirect edges out of
534 SWITCH_EXPRs, which can get expensive. So we want to enable
535 recording of edge to CASE_LABEL_EXPR mappings around the call
536 to cleanup_forwarder_blocks. */
537 start_recording_case_labels ();
538 retval
|= cleanup_forwarder_blocks ();
539 end_recording_case_labels ();
542 /* Merging the blocks may create new opportunities for folding
543 conditional branches (due to the elimination of single-valued PHI
545 retval
|= merge_seq_blocks ();
551 /* Remove unreachable blocks and other miscellaneous clean up work.
552 Return true if the flowgraph was modified, false otherwise. */
555 cleanup_tree_cfg (void)
557 bool retval
, changed
;
559 timevar_push (TV_TREE_CLEANUP_CFG
);
561 /* Iterate until there are no more cleanups left to do. If any
562 iteration changed the flowgraph, set CHANGED to true. */
566 retval
= cleanup_tree_cfg_1 ();
573 #ifdef ENABLE_CHECKING
577 timevar_pop (TV_TREE_CLEANUP_CFG
);
582 /* Cleanup cfg and repair loop structures. */
585 cleanup_tree_cfg_loop (void)
587 bool changed
= cleanup_tree_cfg ();
591 bitmap changed_bbs
= BITMAP_ALLOC (NULL
);
592 fix_loop_structure (current_loops
, changed_bbs
);
593 calculate_dominance_info (CDI_DOMINATORS
);
595 /* This usually does nothing. But sometimes parts of cfg that originally
596 were inside a loop get out of it due to edge removal (since they
597 become unreachable by back edges from latch). */
598 rewrite_into_loop_closed_ssa (changed_bbs
, TODO_update_ssa
);
600 BITMAP_FREE (changed_bbs
);
602 #ifdef ENABLE_CHECKING
603 verify_loop_structure (current_loops
);
609 /* Merge the PHI nodes at BB into those at BB's sole successor. */
612 remove_forwarder_block_with_phi (basic_block bb
)
614 edge succ
= single_succ_edge (bb
);
615 basic_block dest
= succ
->dest
;
617 basic_block dombb
, domdest
, dom
;
619 /* We check for infinite loops already in tree_forwarder_block_p.
620 However it may happen that the infinite loop is created
621 afterwards due to removal of forwarders. */
625 /* If the destination block consists of a nonlocal label, do not
627 label
= first_stmt (dest
);
629 && TREE_CODE (label
) == LABEL_EXPR
630 && DECL_NONLOCAL (LABEL_EXPR_LABEL (label
)))
633 /* Redirect each incoming edge to BB to DEST. */
634 while (EDGE_COUNT (bb
->preds
) > 0)
636 edge e
= EDGE_PRED (bb
, 0), s
;
639 s
= find_edge (e
->src
, dest
);
642 /* We already have an edge S from E->src to DEST. If S and
643 E->dest's sole successor edge have the same PHI arguments
644 at DEST, redirect S to DEST. */
645 if (phi_alternatives_equal (dest
, s
, succ
))
647 e
= redirect_edge_and_branch (e
, dest
);
648 PENDING_STMT (e
) = NULL_TREE
;
652 /* PHI arguments are different. Create a forwarder block by
653 splitting E so that we can merge PHI arguments on E to
655 e
= single_succ_edge (split_edge (e
));
658 s
= redirect_edge_and_branch (e
, dest
);
660 /* redirect_edge_and_branch must not create a new edge. */
663 /* Add to the PHI nodes at DEST each PHI argument removed at the
665 for (phi
= phi_nodes (dest
); phi
; phi
= PHI_CHAIN (phi
))
667 tree def
= PHI_ARG_DEF (phi
, succ
->dest_idx
);
669 if (TREE_CODE (def
) == SSA_NAME
)
673 /* If DEF is one of the results of PHI nodes removed during
674 redirection, replace it with the PHI argument that used
676 for (var
= PENDING_STMT (e
); var
; var
= TREE_CHAIN (var
))
678 tree old_arg
= TREE_PURPOSE (var
);
679 tree new_arg
= TREE_VALUE (var
);
689 add_phi_arg (phi
, def
, s
);
692 PENDING_STMT (e
) = NULL
;
695 /* Update the dominators. */
696 dombb
= get_immediate_dominator (CDI_DOMINATORS
, bb
);
697 domdest
= get_immediate_dominator (CDI_DOMINATORS
, dest
);
700 /* Shortcut to avoid calling (relatively expensive)
701 nearest_common_dominator unless necessary. */
705 dom
= nearest_common_dominator (CDI_DOMINATORS
, domdest
, dombb
);
707 set_immediate_dominator (CDI_DOMINATORS
, dest
, dom
);
709 /* Remove BB since all of BB's incoming edges have been redirected
711 delete_basic_block (bb
);
714 /* This pass merges PHI nodes if one feeds into another. For example,
715 suppose we have the following:
722 # tem_6 = PHI <tem_17(8), tem_23(7)>;
725 # tem_3 = PHI <tem_6(9), tem_2(5)>;
728 Then we merge the first PHI node into the second one like so:
735 # tem_3 = PHI <tem_23(7), tem_2(5), tem_17(8)>;
740 merge_phi_nodes (void)
742 basic_block
*worklist
= XNEWVEC (basic_block
, n_basic_blocks
);
743 basic_block
*current
= worklist
;
746 calculate_dominance_info (CDI_DOMINATORS
);
748 /* Find all PHI nodes that we may be able to merge. */
753 /* Look for a forwarder block with PHI nodes. */
754 if (!tree_forwarder_block_p (bb
, true))
757 dest
= single_succ (bb
);
759 /* We have to feed into another basic block with PHI
761 if (!phi_nodes (dest
)
762 /* We don't want to deal with a basic block with
764 || has_abnormal_incoming_edge_p (bb
))
767 if (!dominated_by_p (CDI_DOMINATORS
, dest
, bb
))
769 /* If BB does not dominate DEST, then the PHI nodes at
770 DEST must be the only users of the results of the PHI
777 unsigned int dest_idx
= single_succ_edge (bb
)->dest_idx
;
779 /* BB dominates DEST. There may be many users of the PHI
780 nodes in BB. However, there is still a trivial case we
781 can handle. If the result of every PHI in BB is used
782 only by a PHI in DEST, then we can trivially merge the
783 PHI nodes from BB into DEST. */
784 for (phi
= phi_nodes (bb
); phi
; phi
= PHI_CHAIN (phi
))
786 tree result
= PHI_RESULT (phi
);
787 use_operand_p imm_use
;
790 /* If the PHI's result is never used, then we can just
792 if (has_zero_uses (result
))
795 /* Get the single use of the result of this PHI node. */
796 if (!single_imm_use (result
, &imm_use
, &use_stmt
)
797 || TREE_CODE (use_stmt
) != PHI_NODE
798 || bb_for_stmt (use_stmt
) != dest
799 || PHI_ARG_DEF (use_stmt
, dest_idx
) != result
)
803 /* If the loop above iterated through all the PHI nodes
804 in BB, then we can merge the PHIs from BB into DEST. */
810 /* Now let's drain WORKLIST. */
811 while (current
!= worklist
)
814 remove_forwarder_block_with_phi (bb
);
822 gate_merge_phi (void)
827 struct tree_opt_pass pass_merge_phi
= {
828 "mergephi", /* name */
829 gate_merge_phi
, /* gate */
830 merge_phi_nodes
, /* execute */
833 0, /* static_pass_number */
834 TV_TREE_MERGE_PHI
, /* tv_id */
835 PROP_cfg
| PROP_ssa
, /* properties_required */
836 0, /* properties_provided */
837 0, /* properties_destroyed */
838 0, /* todo_flags_start */
839 TODO_dump_func
| TODO_ggc_collect
/* todo_flags_finish */