Add execution + assembler tests of AArch64 TRN Intrinsics.
[official-gcc.git] / gcc / tree-ssa-dom.c
blob5b5adca90b2f53c6c52a10c75a8b9beb54aad023
1 /* SSA Dominator optimizations for trees
2 Copyright (C) 2001-2014 Free Software Foundation, Inc.
3 Contributed by Diego Novillo <dnovillo@redhat.com>
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License as published by
9 the Free Software Foundation; either version 3, or (at your option)
10 any later version.
12 GCC is distributed in the hope that it will be useful,
13 but WITHOUT ANY WARRANTY; without even the implied warranty of
14 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 GNU General Public License for more details.
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
21 #include "config.h"
22 #include "system.h"
23 #include "coretypes.h"
24 #include "hash-table.h"
25 #include "tm.h"
26 #include "tree.h"
27 #include "stor-layout.h"
28 #include "flags.h"
29 #include "tm_p.h"
30 #include "basic-block.h"
31 #include "cfgloop.h"
32 #include "function.h"
33 #include "gimple-pretty-print.h"
34 #include "tree-ssa-alias.h"
35 #include "internal-fn.h"
36 #include "gimple-fold.h"
37 #include "tree-eh.h"
38 #include "gimple-expr.h"
39 #include "is-a.h"
40 #include "gimple.h"
41 #include "gimple-iterator.h"
42 #include "gimple-ssa.h"
43 #include "tree-cfg.h"
44 #include "tree-phinodes.h"
45 #include "ssa-iterators.h"
46 #include "stringpool.h"
47 #include "tree-ssanames.h"
48 #include "tree-into-ssa.h"
49 #include "domwalk.h"
50 #include "tree-pass.h"
51 #include "tree-ssa-propagate.h"
52 #include "tree-ssa-threadupdate.h"
53 #include "langhooks.h"
54 #include "params.h"
55 #include "tree-ssa-threadedge.h"
56 #include "tree-ssa-dom.h"
58 /* This file implements optimizations on the dominator tree. */
60 /* Representation of a "naked" right-hand-side expression, to be used
61 in recording available expressions in the expression hash table. */
63 enum expr_kind
65 EXPR_SINGLE,
66 EXPR_UNARY,
67 EXPR_BINARY,
68 EXPR_TERNARY,
69 EXPR_CALL,
70 EXPR_PHI
73 struct hashable_expr
75 tree type;
76 enum expr_kind kind;
77 union {
78 struct { tree rhs; } single;
79 struct { enum tree_code op; tree opnd; } unary;
80 struct { enum tree_code op; tree opnd0, opnd1; } binary;
81 struct { enum tree_code op; tree opnd0, opnd1, opnd2; } ternary;
82 struct { gimple fn_from; bool pure; size_t nargs; tree *args; } call;
83 struct { size_t nargs; tree *args; } phi;
84 } ops;
87 /* Structure for recording known values of a conditional expression
88 at the exits from its block. */
90 typedef struct cond_equivalence_s
92 struct hashable_expr cond;
93 tree value;
94 } cond_equivalence;
97 /* Structure for recording edge equivalences as well as any pending
98 edge redirections during the dominator optimizer.
100 Computing and storing the edge equivalences instead of creating
101 them on-demand can save significant amounts of time, particularly
102 for pathological cases involving switch statements.
104 These structures live for a single iteration of the dominator
105 optimizer in the edge's AUX field. At the end of an iteration we
106 free each of these structures and update the AUX field to point
107 to any requested redirection target (the code for updating the
108 CFG and SSA graph for edge redirection expects redirection edge
109 targets to be in the AUX field for each edge. */
111 struct edge_info
113 /* If this edge creates a simple equivalence, the LHS and RHS of
114 the equivalence will be stored here. */
115 tree lhs;
116 tree rhs;
118 /* Traversing an edge may also indicate one or more particular conditions
119 are true or false. */
120 vec<cond_equivalence> cond_equivalences;
123 /* Stack of available expressions in AVAIL_EXPRs. Each block pushes any
124 expressions it enters into the hash table along with a marker entry
125 (null). When we finish processing the block, we pop off entries and
126 remove the expressions from the global hash table until we hit the
127 marker. */
128 typedef struct expr_hash_elt * expr_hash_elt_t;
130 static vec<expr_hash_elt_t> avail_exprs_stack;
132 /* Structure for entries in the expression hash table. */
134 struct expr_hash_elt
136 /* The value (lhs) of this expression. */
137 tree lhs;
139 /* The expression (rhs) we want to record. */
140 struct hashable_expr expr;
142 /* The stmt pointer if this element corresponds to a statement. */
143 gimple stmt;
145 /* The hash value for RHS. */
146 hashval_t hash;
148 /* A unique stamp, typically the address of the hash
149 element itself, used in removing entries from the table. */
150 struct expr_hash_elt *stamp;
153 /* Hashtable helpers. */
155 static bool hashable_expr_equal_p (const struct hashable_expr *,
156 const struct hashable_expr *);
157 static void free_expr_hash_elt (void *);
159 struct expr_elt_hasher
161 typedef expr_hash_elt value_type;
162 typedef expr_hash_elt compare_type;
163 static inline hashval_t hash (const value_type *);
164 static inline bool equal (const value_type *, const compare_type *);
165 static inline void remove (value_type *);
168 inline hashval_t
169 expr_elt_hasher::hash (const value_type *p)
171 return p->hash;
174 inline bool
175 expr_elt_hasher::equal (const value_type *p1, const compare_type *p2)
177 gimple stmt1 = p1->stmt;
178 const struct hashable_expr *expr1 = &p1->expr;
179 const struct expr_hash_elt *stamp1 = p1->stamp;
180 gimple stmt2 = p2->stmt;
181 const struct hashable_expr *expr2 = &p2->expr;
182 const struct expr_hash_elt *stamp2 = p2->stamp;
184 /* This case should apply only when removing entries from the table. */
185 if (stamp1 == stamp2)
186 return true;
188 /* FIXME tuples:
189 We add stmts to a hash table and them modify them. To detect the case
190 that we modify a stmt and then search for it, we assume that the hash
191 is always modified by that change.
192 We have to fully check why this doesn't happen on trunk or rewrite
193 this in a more reliable (and easier to understand) way. */
194 if (((const struct expr_hash_elt *)p1)->hash
195 != ((const struct expr_hash_elt *)p2)->hash)
196 return false;
198 /* In case of a collision, both RHS have to be identical and have the
199 same VUSE operands. */
200 if (hashable_expr_equal_p (expr1, expr2)
201 && types_compatible_p (expr1->type, expr2->type))
203 /* Note that STMT1 and/or STMT2 may be NULL. */
204 return ((stmt1 ? gimple_vuse (stmt1) : NULL_TREE)
205 == (stmt2 ? gimple_vuse (stmt2) : NULL_TREE));
208 return false;
211 /* Delete an expr_hash_elt and reclaim its storage. */
213 inline void
214 expr_elt_hasher::remove (value_type *element)
216 free_expr_hash_elt (element);
219 /* Hash table with expressions made available during the renaming process.
220 When an assignment of the form X_i = EXPR is found, the statement is
221 stored in this table. If the same expression EXPR is later found on the
222 RHS of another statement, it is replaced with X_i (thus performing
223 global redundancy elimination). Similarly as we pass through conditionals
224 we record the conditional itself as having either a true or false value
225 in this table. */
226 static hash_table <expr_elt_hasher> avail_exprs;
228 /* Stack of dest,src pairs that need to be restored during finalization.
230 A NULL entry is used to mark the end of pairs which need to be
231 restored during finalization of this block. */
232 static vec<tree> const_and_copies_stack;
234 /* Track whether or not we have changed the control flow graph. */
235 static bool cfg_altered;
237 /* Bitmap of blocks that have had EH statements cleaned. We should
238 remove their dead edges eventually. */
239 static bitmap need_eh_cleanup;
241 /* Statistics for dominator optimizations. */
242 struct opt_stats_d
244 long num_stmts;
245 long num_exprs_considered;
246 long num_re;
247 long num_const_prop;
248 long num_copy_prop;
251 static struct opt_stats_d opt_stats;
253 /* Local functions. */
254 static void optimize_stmt (basic_block, gimple_stmt_iterator);
255 static tree lookup_avail_expr (gimple, bool);
256 static hashval_t avail_expr_hash (const void *);
257 static void htab_statistics (FILE *, hash_table <expr_elt_hasher>);
258 static void record_cond (cond_equivalence *);
259 static void record_const_or_copy (tree, tree);
260 static void record_equality (tree, tree);
261 static void record_equivalences_from_phis (basic_block);
262 static void record_equivalences_from_incoming_edge (basic_block);
263 static void eliminate_redundant_computations (gimple_stmt_iterator *);
264 static void record_equivalences_from_stmt (gimple, int);
265 static void remove_local_expressions_from_table (void);
266 static void restore_vars_to_original_value (void);
267 static edge single_incoming_edge_ignoring_loop_edges (basic_block);
270 /* Given a statement STMT, initialize the hash table element pointed to
271 by ELEMENT. */
273 static void
274 initialize_hash_element (gimple stmt, tree lhs,
275 struct expr_hash_elt *element)
277 enum gimple_code code = gimple_code (stmt);
278 struct hashable_expr *expr = &element->expr;
280 if (code == GIMPLE_ASSIGN)
282 enum tree_code subcode = gimple_assign_rhs_code (stmt);
284 switch (get_gimple_rhs_class (subcode))
286 case GIMPLE_SINGLE_RHS:
287 expr->kind = EXPR_SINGLE;
288 expr->type = TREE_TYPE (gimple_assign_rhs1 (stmt));
289 expr->ops.single.rhs = gimple_assign_rhs1 (stmt);
290 break;
291 case GIMPLE_UNARY_RHS:
292 expr->kind = EXPR_UNARY;
293 expr->type = TREE_TYPE (gimple_assign_lhs (stmt));
294 expr->ops.unary.op = subcode;
295 expr->ops.unary.opnd = gimple_assign_rhs1 (stmt);
296 break;
297 case GIMPLE_BINARY_RHS:
298 expr->kind = EXPR_BINARY;
299 expr->type = TREE_TYPE (gimple_assign_lhs (stmt));
300 expr->ops.binary.op = subcode;
301 expr->ops.binary.opnd0 = gimple_assign_rhs1 (stmt);
302 expr->ops.binary.opnd1 = gimple_assign_rhs2 (stmt);
303 break;
304 case GIMPLE_TERNARY_RHS:
305 expr->kind = EXPR_TERNARY;
306 expr->type = TREE_TYPE (gimple_assign_lhs (stmt));
307 expr->ops.ternary.op = subcode;
308 expr->ops.ternary.opnd0 = gimple_assign_rhs1 (stmt);
309 expr->ops.ternary.opnd1 = gimple_assign_rhs2 (stmt);
310 expr->ops.ternary.opnd2 = gimple_assign_rhs3 (stmt);
311 break;
312 default:
313 gcc_unreachable ();
316 else if (code == GIMPLE_COND)
318 expr->type = boolean_type_node;
319 expr->kind = EXPR_BINARY;
320 expr->ops.binary.op = gimple_cond_code (stmt);
321 expr->ops.binary.opnd0 = gimple_cond_lhs (stmt);
322 expr->ops.binary.opnd1 = gimple_cond_rhs (stmt);
324 else if (code == GIMPLE_CALL)
326 size_t nargs = gimple_call_num_args (stmt);
327 size_t i;
329 gcc_assert (gimple_call_lhs (stmt));
331 expr->type = TREE_TYPE (gimple_call_lhs (stmt));
332 expr->kind = EXPR_CALL;
333 expr->ops.call.fn_from = stmt;
335 if (gimple_call_flags (stmt) & (ECF_CONST | ECF_PURE))
336 expr->ops.call.pure = true;
337 else
338 expr->ops.call.pure = false;
340 expr->ops.call.nargs = nargs;
341 expr->ops.call.args = XCNEWVEC (tree, nargs);
342 for (i = 0; i < nargs; i++)
343 expr->ops.call.args[i] = gimple_call_arg (stmt, i);
345 else if (code == GIMPLE_SWITCH)
347 expr->type = TREE_TYPE (gimple_switch_index (stmt));
348 expr->kind = EXPR_SINGLE;
349 expr->ops.single.rhs = gimple_switch_index (stmt);
351 else if (code == GIMPLE_GOTO)
353 expr->type = TREE_TYPE (gimple_goto_dest (stmt));
354 expr->kind = EXPR_SINGLE;
355 expr->ops.single.rhs = gimple_goto_dest (stmt);
357 else if (code == GIMPLE_PHI)
359 size_t nargs = gimple_phi_num_args (stmt);
360 size_t i;
362 expr->type = TREE_TYPE (gimple_phi_result (stmt));
363 expr->kind = EXPR_PHI;
364 expr->ops.phi.nargs = nargs;
365 expr->ops.phi.args = XCNEWVEC (tree, nargs);
367 for (i = 0; i < nargs; i++)
368 expr->ops.phi.args[i] = gimple_phi_arg_def (stmt, i);
370 else
371 gcc_unreachable ();
373 element->lhs = lhs;
374 element->stmt = stmt;
375 element->hash = avail_expr_hash (element);
376 element->stamp = element;
379 /* Given a conditional expression COND as a tree, initialize
380 a hashable_expr expression EXPR. The conditional must be a
381 comparison or logical negation. A constant or a variable is
382 not permitted. */
384 static void
385 initialize_expr_from_cond (tree cond, struct hashable_expr *expr)
387 expr->type = boolean_type_node;
389 if (COMPARISON_CLASS_P (cond))
391 expr->kind = EXPR_BINARY;
392 expr->ops.binary.op = TREE_CODE (cond);
393 expr->ops.binary.opnd0 = TREE_OPERAND (cond, 0);
394 expr->ops.binary.opnd1 = TREE_OPERAND (cond, 1);
396 else if (TREE_CODE (cond) == TRUTH_NOT_EXPR)
398 expr->kind = EXPR_UNARY;
399 expr->ops.unary.op = TRUTH_NOT_EXPR;
400 expr->ops.unary.opnd = TREE_OPERAND (cond, 0);
402 else
403 gcc_unreachable ();
406 /* Given a hashable_expr expression EXPR and an LHS,
407 initialize the hash table element pointed to by ELEMENT. */
409 static void
410 initialize_hash_element_from_expr (struct hashable_expr *expr,
411 tree lhs,
412 struct expr_hash_elt *element)
414 element->expr = *expr;
415 element->lhs = lhs;
416 element->stmt = NULL;
417 element->hash = avail_expr_hash (element);
418 element->stamp = element;
421 /* Compare two hashable_expr structures for equivalence.
422 They are considered equivalent when the the expressions
423 they denote must necessarily be equal. The logic is intended
424 to follow that of operand_equal_p in fold-const.c */
426 static bool
427 hashable_expr_equal_p (const struct hashable_expr *expr0,
428 const struct hashable_expr *expr1)
430 tree type0 = expr0->type;
431 tree type1 = expr1->type;
433 /* If either type is NULL, there is nothing to check. */
434 if ((type0 == NULL_TREE) ^ (type1 == NULL_TREE))
435 return false;
437 /* If both types don't have the same signedness, precision, and mode,
438 then we can't consider them equal. */
439 if (type0 != type1
440 && (TREE_CODE (type0) == ERROR_MARK
441 || TREE_CODE (type1) == ERROR_MARK
442 || TYPE_UNSIGNED (type0) != TYPE_UNSIGNED (type1)
443 || TYPE_PRECISION (type0) != TYPE_PRECISION (type1)
444 || TYPE_MODE (type0) != TYPE_MODE (type1)))
445 return false;
447 if (expr0->kind != expr1->kind)
448 return false;
450 switch (expr0->kind)
452 case EXPR_SINGLE:
453 return operand_equal_p (expr0->ops.single.rhs,
454 expr1->ops.single.rhs, 0);
456 case EXPR_UNARY:
457 if (expr0->ops.unary.op != expr1->ops.unary.op)
458 return false;
460 if ((CONVERT_EXPR_CODE_P (expr0->ops.unary.op)
461 || expr0->ops.unary.op == NON_LVALUE_EXPR)
462 && TYPE_UNSIGNED (expr0->type) != TYPE_UNSIGNED (expr1->type))
463 return false;
465 return operand_equal_p (expr0->ops.unary.opnd,
466 expr1->ops.unary.opnd, 0);
468 case EXPR_BINARY:
469 if (expr0->ops.binary.op != expr1->ops.binary.op)
470 return false;
472 if (operand_equal_p (expr0->ops.binary.opnd0,
473 expr1->ops.binary.opnd0, 0)
474 && operand_equal_p (expr0->ops.binary.opnd1,
475 expr1->ops.binary.opnd1, 0))
476 return true;
478 /* For commutative ops, allow the other order. */
479 return (commutative_tree_code (expr0->ops.binary.op)
480 && operand_equal_p (expr0->ops.binary.opnd0,
481 expr1->ops.binary.opnd1, 0)
482 && operand_equal_p (expr0->ops.binary.opnd1,
483 expr1->ops.binary.opnd0, 0));
485 case EXPR_TERNARY:
486 if (expr0->ops.ternary.op != expr1->ops.ternary.op
487 || !operand_equal_p (expr0->ops.ternary.opnd2,
488 expr1->ops.ternary.opnd2, 0))
489 return false;
491 if (operand_equal_p (expr0->ops.ternary.opnd0,
492 expr1->ops.ternary.opnd0, 0)
493 && operand_equal_p (expr0->ops.ternary.opnd1,
494 expr1->ops.ternary.opnd1, 0))
495 return true;
497 /* For commutative ops, allow the other order. */
498 return (commutative_ternary_tree_code (expr0->ops.ternary.op)
499 && operand_equal_p (expr0->ops.ternary.opnd0,
500 expr1->ops.ternary.opnd1, 0)
501 && operand_equal_p (expr0->ops.ternary.opnd1,
502 expr1->ops.ternary.opnd0, 0));
504 case EXPR_CALL:
506 size_t i;
508 /* If the calls are to different functions, then they
509 clearly cannot be equal. */
510 if (!gimple_call_same_target_p (expr0->ops.call.fn_from,
511 expr1->ops.call.fn_from))
512 return false;
514 if (! expr0->ops.call.pure)
515 return false;
517 if (expr0->ops.call.nargs != expr1->ops.call.nargs)
518 return false;
520 for (i = 0; i < expr0->ops.call.nargs; i++)
521 if (! operand_equal_p (expr0->ops.call.args[i],
522 expr1->ops.call.args[i], 0))
523 return false;
525 return true;
528 case EXPR_PHI:
530 size_t i;
532 if (expr0->ops.phi.nargs != expr1->ops.phi.nargs)
533 return false;
535 for (i = 0; i < expr0->ops.phi.nargs; i++)
536 if (! operand_equal_p (expr0->ops.phi.args[i],
537 expr1->ops.phi.args[i], 0))
538 return false;
540 return true;
543 default:
544 gcc_unreachable ();
548 /* Generate a hash value for a pair of expressions. This can be used
549 iteratively by passing a previous result as the VAL argument.
551 The same hash value is always returned for a given pair of expressions,
552 regardless of the order in which they are presented. This is useful in
553 hashing the operands of commutative functions. */
555 static hashval_t
556 iterative_hash_exprs_commutative (const_tree t1,
557 const_tree t2, hashval_t val)
559 hashval_t one = iterative_hash_expr (t1, 0);
560 hashval_t two = iterative_hash_expr (t2, 0);
561 hashval_t t;
563 if (one > two)
564 t = one, one = two, two = t;
565 val = iterative_hash_hashval_t (one, val);
566 val = iterative_hash_hashval_t (two, val);
568 return val;
571 /* Compute a hash value for a hashable_expr value EXPR and a
572 previously accumulated hash value VAL. If two hashable_expr
573 values compare equal with hashable_expr_equal_p, they must
574 hash to the same value, given an identical value of VAL.
575 The logic is intended to follow iterative_hash_expr in tree.c. */
577 static hashval_t
578 iterative_hash_hashable_expr (const struct hashable_expr *expr, hashval_t val)
580 switch (expr->kind)
582 case EXPR_SINGLE:
583 val = iterative_hash_expr (expr->ops.single.rhs, val);
584 break;
586 case EXPR_UNARY:
587 val = iterative_hash_object (expr->ops.unary.op, val);
589 /* Make sure to include signedness in the hash computation.
590 Don't hash the type, that can lead to having nodes which
591 compare equal according to operand_equal_p, but which
592 have different hash codes. */
593 if (CONVERT_EXPR_CODE_P (expr->ops.unary.op)
594 || expr->ops.unary.op == NON_LVALUE_EXPR)
595 val += TYPE_UNSIGNED (expr->type);
597 val = iterative_hash_expr (expr->ops.unary.opnd, val);
598 break;
600 case EXPR_BINARY:
601 val = iterative_hash_object (expr->ops.binary.op, val);
602 if (commutative_tree_code (expr->ops.binary.op))
603 val = iterative_hash_exprs_commutative (expr->ops.binary.opnd0,
604 expr->ops.binary.opnd1, val);
605 else
607 val = iterative_hash_expr (expr->ops.binary.opnd0, val);
608 val = iterative_hash_expr (expr->ops.binary.opnd1, val);
610 break;
612 case EXPR_TERNARY:
613 val = iterative_hash_object (expr->ops.ternary.op, val);
614 if (commutative_ternary_tree_code (expr->ops.ternary.op))
615 val = iterative_hash_exprs_commutative (expr->ops.ternary.opnd0,
616 expr->ops.ternary.opnd1, val);
617 else
619 val = iterative_hash_expr (expr->ops.ternary.opnd0, val);
620 val = iterative_hash_expr (expr->ops.ternary.opnd1, val);
622 val = iterative_hash_expr (expr->ops.ternary.opnd2, val);
623 break;
625 case EXPR_CALL:
627 size_t i;
628 enum tree_code code = CALL_EXPR;
629 gimple fn_from;
631 val = iterative_hash_object (code, val);
632 fn_from = expr->ops.call.fn_from;
633 if (gimple_call_internal_p (fn_from))
634 val = iterative_hash_hashval_t
635 ((hashval_t) gimple_call_internal_fn (fn_from), val);
636 else
637 val = iterative_hash_expr (gimple_call_fn (fn_from), val);
638 for (i = 0; i < expr->ops.call.nargs; i++)
639 val = iterative_hash_expr (expr->ops.call.args[i], val);
641 break;
643 case EXPR_PHI:
645 size_t i;
647 for (i = 0; i < expr->ops.phi.nargs; i++)
648 val = iterative_hash_expr (expr->ops.phi.args[i], val);
650 break;
652 default:
653 gcc_unreachable ();
656 return val;
659 /* Print a diagnostic dump of an expression hash table entry. */
661 static void
662 print_expr_hash_elt (FILE * stream, const struct expr_hash_elt *element)
664 if (element->stmt)
665 fprintf (stream, "STMT ");
666 else
667 fprintf (stream, "COND ");
669 if (element->lhs)
671 print_generic_expr (stream, element->lhs, 0);
672 fprintf (stream, " = ");
675 switch (element->expr.kind)
677 case EXPR_SINGLE:
678 print_generic_expr (stream, element->expr.ops.single.rhs, 0);
679 break;
681 case EXPR_UNARY:
682 fprintf (stream, "%s ", get_tree_code_name (element->expr.ops.unary.op));
683 print_generic_expr (stream, element->expr.ops.unary.opnd, 0);
684 break;
686 case EXPR_BINARY:
687 print_generic_expr (stream, element->expr.ops.binary.opnd0, 0);
688 fprintf (stream, " %s ", get_tree_code_name (element->expr.ops.binary.op));
689 print_generic_expr (stream, element->expr.ops.binary.opnd1, 0);
690 break;
692 case EXPR_TERNARY:
693 fprintf (stream, " %s <", get_tree_code_name (element->expr.ops.ternary.op));
694 print_generic_expr (stream, element->expr.ops.ternary.opnd0, 0);
695 fputs (", ", stream);
696 print_generic_expr (stream, element->expr.ops.ternary.opnd1, 0);
697 fputs (", ", stream);
698 print_generic_expr (stream, element->expr.ops.ternary.opnd2, 0);
699 fputs (">", stream);
700 break;
702 case EXPR_CALL:
704 size_t i;
705 size_t nargs = element->expr.ops.call.nargs;
706 gimple fn_from;
708 fn_from = element->expr.ops.call.fn_from;
709 if (gimple_call_internal_p (fn_from))
710 fputs (internal_fn_name (gimple_call_internal_fn (fn_from)),
711 stream);
712 else
713 print_generic_expr (stream, gimple_call_fn (fn_from), 0);
714 fprintf (stream, " (");
715 for (i = 0; i < nargs; i++)
717 print_generic_expr (stream, element->expr.ops.call.args[i], 0);
718 if (i + 1 < nargs)
719 fprintf (stream, ", ");
721 fprintf (stream, ")");
723 break;
725 case EXPR_PHI:
727 size_t i;
728 size_t nargs = element->expr.ops.phi.nargs;
730 fprintf (stream, "PHI <");
731 for (i = 0; i < nargs; i++)
733 print_generic_expr (stream, element->expr.ops.phi.args[i], 0);
734 if (i + 1 < nargs)
735 fprintf (stream, ", ");
737 fprintf (stream, ">");
739 break;
741 fprintf (stream, "\n");
743 if (element->stmt)
745 fprintf (stream, " ");
746 print_gimple_stmt (stream, element->stmt, 0, 0);
750 /* Delete variable sized pieces of the expr_hash_elt ELEMENT. */
752 static void
753 free_expr_hash_elt_contents (struct expr_hash_elt *element)
755 if (element->expr.kind == EXPR_CALL)
756 free (element->expr.ops.call.args);
757 else if (element->expr.kind == EXPR_PHI)
758 free (element->expr.ops.phi.args);
761 /* Delete an expr_hash_elt and reclaim its storage. */
763 static void
764 free_expr_hash_elt (void *elt)
766 struct expr_hash_elt *element = ((struct expr_hash_elt *)elt);
767 free_expr_hash_elt_contents (element);
768 free (element);
771 /* Allocate an EDGE_INFO for edge E and attach it to E.
772 Return the new EDGE_INFO structure. */
774 static struct edge_info *
775 allocate_edge_info (edge e)
777 struct edge_info *edge_info;
779 edge_info = XCNEW (struct edge_info);
781 e->aux = edge_info;
782 return edge_info;
785 /* Free all EDGE_INFO structures associated with edges in the CFG.
786 If a particular edge can be threaded, copy the redirection
787 target from the EDGE_INFO structure into the edge's AUX field
788 as required by code to update the CFG and SSA graph for
789 jump threading. */
791 static void
792 free_all_edge_infos (void)
794 basic_block bb;
795 edge_iterator ei;
796 edge e;
798 FOR_EACH_BB_FN (bb, cfun)
800 FOR_EACH_EDGE (e, ei, bb->preds)
802 struct edge_info *edge_info = (struct edge_info *) e->aux;
804 if (edge_info)
806 edge_info->cond_equivalences.release ();
807 free (edge_info);
808 e->aux = NULL;
814 class dom_opt_dom_walker : public dom_walker
816 public:
817 dom_opt_dom_walker (cdi_direction direction)
818 : dom_walker (direction), m_dummy_cond (NULL) {}
820 virtual void before_dom_children (basic_block);
821 virtual void after_dom_children (basic_block);
823 private:
824 void thread_across_edge (edge);
826 gimple m_dummy_cond;
829 /* Jump threading, redundancy elimination and const/copy propagation.
831 This pass may expose new symbols that need to be renamed into SSA. For
832 every new symbol exposed, its corresponding bit will be set in
833 VARS_TO_RENAME. */
835 namespace {
837 const pass_data pass_data_dominator =
839 GIMPLE_PASS, /* type */
840 "dom", /* name */
841 OPTGROUP_NONE, /* optinfo_flags */
842 true, /* has_execute */
843 TV_TREE_SSA_DOMINATOR_OPTS, /* tv_id */
844 ( PROP_cfg | PROP_ssa ), /* properties_required */
845 0, /* properties_provided */
846 0, /* properties_destroyed */
847 0, /* todo_flags_start */
848 ( TODO_cleanup_cfg | TODO_update_ssa ), /* todo_flags_finish */
851 class pass_dominator : public gimple_opt_pass
853 public:
854 pass_dominator (gcc::context *ctxt)
855 : gimple_opt_pass (pass_data_dominator, ctxt)
858 /* opt_pass methods: */
859 opt_pass * clone () { return new pass_dominator (m_ctxt); }
860 virtual bool gate (function *) { return flag_tree_dom != 0; }
861 virtual unsigned int execute (function *);
863 }; // class pass_dominator
865 unsigned int
866 pass_dominator::execute (function *fun)
868 memset (&opt_stats, 0, sizeof (opt_stats));
870 /* Create our hash tables. */
871 avail_exprs.create (1024);
872 avail_exprs_stack.create (20);
873 const_and_copies_stack.create (20);
874 need_eh_cleanup = BITMAP_ALLOC (NULL);
876 calculate_dominance_info (CDI_DOMINATORS);
877 cfg_altered = false;
879 /* We need to know loop structures in order to avoid destroying them
880 in jump threading. Note that we still can e.g. thread through loop
881 headers to an exit edge, or through loop header to the loop body, assuming
882 that we update the loop info.
884 TODO: We don't need to set LOOPS_HAVE_PREHEADERS generally, but due
885 to several overly conservative bail-outs in jump threading, case
886 gcc.dg/tree-ssa/pr21417.c can't be threaded if loop preheader is
887 missing. We should improve jump threading in future then
888 LOOPS_HAVE_PREHEADERS won't be needed here. */
889 loop_optimizer_init (LOOPS_HAVE_PREHEADERS | LOOPS_HAVE_SIMPLE_LATCHES);
891 /* Initialize the value-handle array. */
892 threadedge_initialize_values ();
894 /* We need accurate information regarding back edges in the CFG
895 for jump threading; this may include back edges that are not part of
896 a single loop. */
897 mark_dfs_back_edges ();
899 /* Recursively walk the dominator tree optimizing statements. */
900 dom_opt_dom_walker (CDI_DOMINATORS).walk (fun->cfg->x_entry_block_ptr);
903 gimple_stmt_iterator gsi;
904 basic_block bb;
905 FOR_EACH_BB_FN (bb, fun)
907 for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
908 update_stmt_if_modified (gsi_stmt (gsi));
912 /* If we exposed any new variables, go ahead and put them into
913 SSA form now, before we handle jump threading. This simplifies
914 interactions between rewriting of _DECL nodes into SSA form
915 and rewriting SSA_NAME nodes into SSA form after block
916 duplication and CFG manipulation. */
917 update_ssa (TODO_update_ssa);
919 free_all_edge_infos ();
921 /* Thread jumps, creating duplicate blocks as needed. */
922 cfg_altered |= thread_through_all_blocks (first_pass_instance);
924 if (cfg_altered)
925 free_dominance_info (CDI_DOMINATORS);
927 /* Removal of statements may make some EH edges dead. Purge
928 such edges from the CFG as needed. */
929 if (!bitmap_empty_p (need_eh_cleanup))
931 unsigned i;
932 bitmap_iterator bi;
934 /* Jump threading may have created forwarder blocks from blocks
935 needing EH cleanup; the new successor of these blocks, which
936 has inherited from the original block, needs the cleanup.
937 Don't clear bits in the bitmap, as that can break the bitmap
938 iterator. */
939 EXECUTE_IF_SET_IN_BITMAP (need_eh_cleanup, 0, i, bi)
941 basic_block bb = BASIC_BLOCK_FOR_FN (fun, i);
942 if (bb == NULL)
943 continue;
944 while (single_succ_p (bb)
945 && (single_succ_edge (bb)->flags & EDGE_EH) == 0)
946 bb = single_succ (bb);
947 if (bb == EXIT_BLOCK_PTR_FOR_FN (fun))
948 continue;
949 if ((unsigned) bb->index != i)
950 bitmap_set_bit (need_eh_cleanup, bb->index);
953 gimple_purge_all_dead_eh_edges (need_eh_cleanup);
954 bitmap_clear (need_eh_cleanup);
957 statistics_counter_event (fun, "Redundant expressions eliminated",
958 opt_stats.num_re);
959 statistics_counter_event (fun, "Constants propagated",
960 opt_stats.num_const_prop);
961 statistics_counter_event (fun, "Copies propagated",
962 opt_stats.num_copy_prop);
964 /* Debugging dumps. */
965 if (dump_file && (dump_flags & TDF_STATS))
966 dump_dominator_optimization_stats (dump_file);
968 loop_optimizer_finalize ();
970 /* Delete our main hashtable. */
971 avail_exprs.dispose ();
973 /* Free asserted bitmaps and stacks. */
974 BITMAP_FREE (need_eh_cleanup);
976 avail_exprs_stack.release ();
977 const_and_copies_stack.release ();
979 /* Free the value-handle array. */
980 threadedge_finalize_values ();
982 return 0;
985 } // anon namespace
987 gimple_opt_pass *
988 make_pass_dominator (gcc::context *ctxt)
990 return new pass_dominator (ctxt);
994 /* Given a conditional statement CONDSTMT, convert the
995 condition to a canonical form. */
997 static void
998 canonicalize_comparison (gimple condstmt)
1000 tree op0;
1001 tree op1;
1002 enum tree_code code;
1004 gcc_assert (gimple_code (condstmt) == GIMPLE_COND);
1006 op0 = gimple_cond_lhs (condstmt);
1007 op1 = gimple_cond_rhs (condstmt);
1009 code = gimple_cond_code (condstmt);
1011 /* If it would be profitable to swap the operands, then do so to
1012 canonicalize the statement, enabling better optimization.
1014 By placing canonicalization of such expressions here we
1015 transparently keep statements in canonical form, even
1016 when the statement is modified. */
1017 if (tree_swap_operands_p (op0, op1, false))
1019 /* For relationals we need to swap the operands
1020 and change the code. */
1021 if (code == LT_EXPR
1022 || code == GT_EXPR
1023 || code == LE_EXPR
1024 || code == GE_EXPR)
1026 code = swap_tree_comparison (code);
1028 gimple_cond_set_code (condstmt, code);
1029 gimple_cond_set_lhs (condstmt, op1);
1030 gimple_cond_set_rhs (condstmt, op0);
1032 update_stmt (condstmt);
1037 /* Initialize local stacks for this optimizer and record equivalences
1038 upon entry to BB. Equivalences can come from the edge traversed to
1039 reach BB or they may come from PHI nodes at the start of BB. */
1041 /* Remove all the expressions in LOCALS from TABLE, stopping when there are
1042 LIMIT entries left in LOCALs. */
1044 static void
1045 remove_local_expressions_from_table (void)
1047 /* Remove all the expressions made available in this block. */
1048 while (avail_exprs_stack.length () > 0)
1050 expr_hash_elt_t victim = avail_exprs_stack.pop ();
1051 expr_hash_elt **slot;
1053 if (victim == NULL)
1054 break;
1056 /* This must precede the actual removal from the hash table,
1057 as ELEMENT and the table entry may share a call argument
1058 vector which will be freed during removal. */
1059 if (dump_file && (dump_flags & TDF_DETAILS))
1061 fprintf (dump_file, "<<<< ");
1062 print_expr_hash_elt (dump_file, victim);
1065 slot = avail_exprs.find_slot_with_hash (victim, victim->hash, NO_INSERT);
1066 gcc_assert (slot && *slot == victim);
1067 avail_exprs.clear_slot (slot);
1071 /* Use the source/dest pairs in CONST_AND_COPIES_STACK to restore
1072 CONST_AND_COPIES to its original state, stopping when we hit a
1073 NULL marker. */
1075 static void
1076 restore_vars_to_original_value (void)
1078 while (const_and_copies_stack.length () > 0)
1080 tree prev_value, dest;
1082 dest = const_and_copies_stack.pop ();
1084 if (dest == NULL)
1085 break;
1087 if (dump_file && (dump_flags & TDF_DETAILS))
1089 fprintf (dump_file, "<<<< COPY ");
1090 print_generic_expr (dump_file, dest, 0);
1091 fprintf (dump_file, " = ");
1092 print_generic_expr (dump_file, SSA_NAME_VALUE (dest), 0);
1093 fprintf (dump_file, "\n");
1096 prev_value = const_and_copies_stack.pop ();
1097 set_ssa_name_value (dest, prev_value);
1101 /* A trivial wrapper so that we can present the generic jump
1102 threading code with a simple API for simplifying statements. */
1103 static tree
1104 simplify_stmt_for_jump_threading (gimple stmt,
1105 gimple within_stmt ATTRIBUTE_UNUSED)
1107 return lookup_avail_expr (stmt, false);
1110 /* Record into the equivalence tables any equivalences implied by
1111 traversing edge E (which are cached in E->aux).
1113 Callers are responsible for managing the unwinding markers. */
1114 static void
1115 record_temporary_equivalences (edge e)
1117 int i;
1118 struct edge_info *edge_info = (struct edge_info *) e->aux;
1120 /* If we have info associated with this edge, record it into
1121 our equivalence tables. */
1122 if (edge_info)
1124 cond_equivalence *eq;
1125 tree lhs = edge_info->lhs;
1126 tree rhs = edge_info->rhs;
1128 /* If we have a simple NAME = VALUE equivalence, record it. */
1129 if (lhs && TREE_CODE (lhs) == SSA_NAME)
1130 record_const_or_copy (lhs, rhs);
1132 /* If we have 0 = COND or 1 = COND equivalences, record them
1133 into our expression hash tables. */
1134 for (i = 0; edge_info->cond_equivalences.iterate (i, &eq); ++i)
1135 record_cond (eq);
1139 /* Wrapper for common code to attempt to thread an edge. For example,
1140 it handles lazily building the dummy condition and the bookkeeping
1141 when jump threading is successful. */
1143 void
1144 dom_opt_dom_walker::thread_across_edge (edge e)
1146 if (! m_dummy_cond)
1147 m_dummy_cond =
1148 gimple_build_cond (NE_EXPR,
1149 integer_zero_node, integer_zero_node,
1150 NULL, NULL);
1152 /* Push a marker on both stacks so we can unwind the tables back to their
1153 current state. */
1154 avail_exprs_stack.safe_push (NULL);
1155 const_and_copies_stack.safe_push (NULL_TREE);
1157 /* Traversing E may result in equivalences we can utilize. */
1158 record_temporary_equivalences (e);
1160 /* With all the edge equivalences in the tables, go ahead and attempt
1161 to thread through E->dest. */
1162 ::thread_across_edge (m_dummy_cond, e, false,
1163 &const_and_copies_stack,
1164 simplify_stmt_for_jump_threading);
1166 /* And restore the various tables to their state before
1167 we threaded this edge.
1169 XXX The code in tree-ssa-threadedge.c will restore the state of
1170 the const_and_copies table. We we just have to restore the expression
1171 table. */
1172 remove_local_expressions_from_table ();
1175 /* PHI nodes can create equivalences too.
1177 Ignoring any alternatives which are the same as the result, if
1178 all the alternatives are equal, then the PHI node creates an
1179 equivalence. */
1181 static void
1182 record_equivalences_from_phis (basic_block bb)
1184 gimple_stmt_iterator gsi;
1186 for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi))
1188 gimple phi = gsi_stmt (gsi);
1190 tree lhs = gimple_phi_result (phi);
1191 tree rhs = NULL;
1192 size_t i;
1194 for (i = 0; i < gimple_phi_num_args (phi); i++)
1196 tree t = gimple_phi_arg_def (phi, i);
1198 /* Ignore alternatives which are the same as our LHS. Since
1199 LHS is a PHI_RESULT, it is known to be a SSA_NAME, so we
1200 can simply compare pointers. */
1201 if (lhs == t)
1202 continue;
1204 /* If we have not processed an alternative yet, then set
1205 RHS to this alternative. */
1206 if (rhs == NULL)
1207 rhs = t;
1208 /* If we have processed an alternative (stored in RHS), then
1209 see if it is equal to this one. If it isn't, then stop
1210 the search. */
1211 else if (! operand_equal_for_phi_arg_p (rhs, t))
1212 break;
1215 /* If we had no interesting alternatives, then all the RHS alternatives
1216 must have been the same as LHS. */
1217 if (!rhs)
1218 rhs = lhs;
1220 /* If we managed to iterate through each PHI alternative without
1221 breaking out of the loop, then we have a PHI which may create
1222 a useful equivalence. We do not need to record unwind data for
1223 this, since this is a true assignment and not an equivalence
1224 inferred from a comparison. All uses of this ssa name are dominated
1225 by this assignment, so unwinding just costs time and space. */
1226 if (i == gimple_phi_num_args (phi) && may_propagate_copy (lhs, rhs))
1227 set_ssa_name_value (lhs, rhs);
1231 /* Ignoring loop backedges, if BB has precisely one incoming edge then
1232 return that edge. Otherwise return NULL. */
1233 static edge
1234 single_incoming_edge_ignoring_loop_edges (basic_block bb)
1236 edge retval = NULL;
1237 edge e;
1238 edge_iterator ei;
1240 FOR_EACH_EDGE (e, ei, bb->preds)
1242 /* A loop back edge can be identified by the destination of
1243 the edge dominating the source of the edge. */
1244 if (dominated_by_p (CDI_DOMINATORS, e->src, e->dest))
1245 continue;
1247 /* If we have already seen a non-loop edge, then we must have
1248 multiple incoming non-loop edges and thus we return NULL. */
1249 if (retval)
1250 return NULL;
1252 /* This is the first non-loop incoming edge we have found. Record
1253 it. */
1254 retval = e;
1257 return retval;
1260 /* Record any equivalences created by the incoming edge to BB. If BB
1261 has more than one incoming edge, then no equivalence is created. */
1263 static void
1264 record_equivalences_from_incoming_edge (basic_block bb)
1266 edge e;
1267 basic_block parent;
1268 struct edge_info *edge_info;
1270 /* If our parent block ended with a control statement, then we may be
1271 able to record some equivalences based on which outgoing edge from
1272 the parent was followed. */
1273 parent = get_immediate_dominator (CDI_DOMINATORS, bb);
1275 e = single_incoming_edge_ignoring_loop_edges (bb);
1277 /* If we had a single incoming edge from our parent block, then enter
1278 any data associated with the edge into our tables. */
1279 if (e && e->src == parent)
1281 unsigned int i;
1283 edge_info = (struct edge_info *) e->aux;
1285 if (edge_info)
1287 tree lhs = edge_info->lhs;
1288 tree rhs = edge_info->rhs;
1289 cond_equivalence *eq;
1291 if (lhs)
1292 record_equality (lhs, rhs);
1294 /* If LHS is an SSA_NAME and RHS is a constant integer and LHS was
1295 set via a widening type conversion, then we may be able to record
1296 additional equivalences. */
1297 if (lhs
1298 && TREE_CODE (lhs) == SSA_NAME
1299 && is_gimple_constant (rhs)
1300 && TREE_CODE (rhs) == INTEGER_CST)
1302 gimple defstmt = SSA_NAME_DEF_STMT (lhs);
1304 if (defstmt
1305 && is_gimple_assign (defstmt)
1306 && CONVERT_EXPR_CODE_P (gimple_assign_rhs_code (defstmt)))
1308 tree old_rhs = gimple_assign_rhs1 (defstmt);
1310 /* If the conversion widens the original value and
1311 the constant is in the range of the type of OLD_RHS,
1312 then convert the constant and record the equivalence.
1314 Note that int_fits_type_p does not check the precision
1315 if the upper and lower bounds are OK. */
1316 if (INTEGRAL_TYPE_P (TREE_TYPE (old_rhs))
1317 && (TYPE_PRECISION (TREE_TYPE (lhs))
1318 > TYPE_PRECISION (TREE_TYPE (old_rhs)))
1319 && int_fits_type_p (rhs, TREE_TYPE (old_rhs)))
1321 tree newval = fold_convert (TREE_TYPE (old_rhs), rhs);
1322 record_equality (old_rhs, newval);
1327 for (i = 0; edge_info->cond_equivalences.iterate (i, &eq); ++i)
1328 record_cond (eq);
1333 /* Dump SSA statistics on FILE. */
1335 void
1336 dump_dominator_optimization_stats (FILE *file)
1338 fprintf (file, "Total number of statements: %6ld\n\n",
1339 opt_stats.num_stmts);
1340 fprintf (file, "Exprs considered for dominator optimizations: %6ld\n",
1341 opt_stats.num_exprs_considered);
1343 fprintf (file, "\nHash table statistics:\n");
1345 fprintf (file, " avail_exprs: ");
1346 htab_statistics (file, avail_exprs);
1350 /* Dump SSA statistics on stderr. */
1352 DEBUG_FUNCTION void
1353 debug_dominator_optimization_stats (void)
1355 dump_dominator_optimization_stats (stderr);
1359 /* Dump statistics for the hash table HTAB. */
1361 static void
1362 htab_statistics (FILE *file, hash_table <expr_elt_hasher> htab)
1364 fprintf (file, "size %ld, %ld elements, %f collision/search ratio\n",
1365 (long) htab.size (),
1366 (long) htab.elements (),
1367 htab.collisions ());
1371 /* Enter condition equivalence into the expression hash table.
1372 This indicates that a conditional expression has a known
1373 boolean value. */
1375 static void
1376 record_cond (cond_equivalence *p)
1378 struct expr_hash_elt *element = XCNEW (struct expr_hash_elt);
1379 expr_hash_elt **slot;
1381 initialize_hash_element_from_expr (&p->cond, p->value, element);
1383 slot = avail_exprs.find_slot_with_hash (element, element->hash, INSERT);
1384 if (*slot == NULL)
1386 *slot = element;
1388 if (dump_file && (dump_flags & TDF_DETAILS))
1390 fprintf (dump_file, "1>>> ");
1391 print_expr_hash_elt (dump_file, element);
1394 avail_exprs_stack.safe_push (element);
1396 else
1397 free_expr_hash_elt (element);
1400 /* Build a cond_equivalence record indicating that the comparison
1401 CODE holds between operands OP0 and OP1 and push it to **P. */
1403 static void
1404 build_and_record_new_cond (enum tree_code code,
1405 tree op0, tree op1,
1406 vec<cond_equivalence> *p)
1408 cond_equivalence c;
1409 struct hashable_expr *cond = &c.cond;
1411 gcc_assert (TREE_CODE_CLASS (code) == tcc_comparison);
1413 cond->type = boolean_type_node;
1414 cond->kind = EXPR_BINARY;
1415 cond->ops.binary.op = code;
1416 cond->ops.binary.opnd0 = op0;
1417 cond->ops.binary.opnd1 = op1;
1419 c.value = boolean_true_node;
1420 p->safe_push (c);
1423 /* Record that COND is true and INVERTED is false into the edge information
1424 structure. Also record that any conditions dominated by COND are true
1425 as well.
1427 For example, if a < b is true, then a <= b must also be true. */
1429 static void
1430 record_conditions (struct edge_info *edge_info, tree cond, tree inverted)
1432 tree op0, op1;
1433 cond_equivalence c;
1435 if (!COMPARISON_CLASS_P (cond))
1436 return;
1438 op0 = TREE_OPERAND (cond, 0);
1439 op1 = TREE_OPERAND (cond, 1);
1441 switch (TREE_CODE (cond))
1443 case LT_EXPR:
1444 case GT_EXPR:
1445 if (FLOAT_TYPE_P (TREE_TYPE (op0)))
1447 build_and_record_new_cond (ORDERED_EXPR, op0, op1,
1448 &edge_info->cond_equivalences);
1449 build_and_record_new_cond (LTGT_EXPR, op0, op1,
1450 &edge_info->cond_equivalences);
1453 build_and_record_new_cond ((TREE_CODE (cond) == LT_EXPR
1454 ? LE_EXPR : GE_EXPR),
1455 op0, op1, &edge_info->cond_equivalences);
1456 build_and_record_new_cond (NE_EXPR, op0, op1,
1457 &edge_info->cond_equivalences);
1458 break;
1460 case GE_EXPR:
1461 case LE_EXPR:
1462 if (FLOAT_TYPE_P (TREE_TYPE (op0)))
1464 build_and_record_new_cond (ORDERED_EXPR, op0, op1,
1465 &edge_info->cond_equivalences);
1467 break;
1469 case EQ_EXPR:
1470 if (FLOAT_TYPE_P (TREE_TYPE (op0)))
1472 build_and_record_new_cond (ORDERED_EXPR, op0, op1,
1473 &edge_info->cond_equivalences);
1475 build_and_record_new_cond (LE_EXPR, op0, op1,
1476 &edge_info->cond_equivalences);
1477 build_and_record_new_cond (GE_EXPR, op0, op1,
1478 &edge_info->cond_equivalences);
1479 break;
1481 case UNORDERED_EXPR:
1482 build_and_record_new_cond (NE_EXPR, op0, op1,
1483 &edge_info->cond_equivalences);
1484 build_and_record_new_cond (UNLE_EXPR, op0, op1,
1485 &edge_info->cond_equivalences);
1486 build_and_record_new_cond (UNGE_EXPR, op0, op1,
1487 &edge_info->cond_equivalences);
1488 build_and_record_new_cond (UNEQ_EXPR, op0, op1,
1489 &edge_info->cond_equivalences);
1490 build_and_record_new_cond (UNLT_EXPR, op0, op1,
1491 &edge_info->cond_equivalences);
1492 build_and_record_new_cond (UNGT_EXPR, op0, op1,
1493 &edge_info->cond_equivalences);
1494 break;
1496 case UNLT_EXPR:
1497 case UNGT_EXPR:
1498 build_and_record_new_cond ((TREE_CODE (cond) == UNLT_EXPR
1499 ? UNLE_EXPR : UNGE_EXPR),
1500 op0, op1, &edge_info->cond_equivalences);
1501 build_and_record_new_cond (NE_EXPR, op0, op1,
1502 &edge_info->cond_equivalences);
1503 break;
1505 case UNEQ_EXPR:
1506 build_and_record_new_cond (UNLE_EXPR, op0, op1,
1507 &edge_info->cond_equivalences);
1508 build_and_record_new_cond (UNGE_EXPR, op0, op1,
1509 &edge_info->cond_equivalences);
1510 break;
1512 case LTGT_EXPR:
1513 build_and_record_new_cond (NE_EXPR, op0, op1,
1514 &edge_info->cond_equivalences);
1515 build_and_record_new_cond (ORDERED_EXPR, op0, op1,
1516 &edge_info->cond_equivalences);
1517 break;
1519 default:
1520 break;
1523 /* Now store the original true and false conditions into the first
1524 two slots. */
1525 initialize_expr_from_cond (cond, &c.cond);
1526 c.value = boolean_true_node;
1527 edge_info->cond_equivalences.safe_push (c);
1529 /* It is possible for INVERTED to be the negation of a comparison,
1530 and not a valid RHS or GIMPLE_COND condition. This happens because
1531 invert_truthvalue may return such an expression when asked to invert
1532 a floating-point comparison. These comparisons are not assumed to
1533 obey the trichotomy law. */
1534 initialize_expr_from_cond (inverted, &c.cond);
1535 c.value = boolean_false_node;
1536 edge_info->cond_equivalences.safe_push (c);
1539 /* A helper function for record_const_or_copy and record_equality.
1540 Do the work of recording the value and undo info. */
1542 static void
1543 record_const_or_copy_1 (tree x, tree y, tree prev_x)
1545 set_ssa_name_value (x, y);
1547 if (dump_file && (dump_flags & TDF_DETAILS))
1549 fprintf (dump_file, "0>>> COPY ");
1550 print_generic_expr (dump_file, x, 0);
1551 fprintf (dump_file, " = ");
1552 print_generic_expr (dump_file, y, 0);
1553 fprintf (dump_file, "\n");
1556 const_and_copies_stack.reserve (2);
1557 const_and_copies_stack.quick_push (prev_x);
1558 const_and_copies_stack.quick_push (x);
1561 /* Return the loop depth of the basic block of the defining statement of X.
1562 This number should not be treated as absolutely correct because the loop
1563 information may not be completely up-to-date when dom runs. However, it
1564 will be relatively correct, and as more passes are taught to keep loop info
1565 up to date, the result will become more and more accurate. */
1568 loop_depth_of_name (tree x)
1570 gimple defstmt;
1571 basic_block defbb;
1573 /* If it's not an SSA_NAME, we have no clue where the definition is. */
1574 if (TREE_CODE (x) != SSA_NAME)
1575 return 0;
1577 /* Otherwise return the loop depth of the defining statement's bb.
1578 Note that there may not actually be a bb for this statement, if the
1579 ssa_name is live on entry. */
1580 defstmt = SSA_NAME_DEF_STMT (x);
1581 defbb = gimple_bb (defstmt);
1582 if (!defbb)
1583 return 0;
1585 return bb_loop_depth (defbb);
1588 /* Record that X is equal to Y in const_and_copies. Record undo
1589 information in the block-local vector. */
1591 static void
1592 record_const_or_copy (tree x, tree y)
1594 tree prev_x = SSA_NAME_VALUE (x);
1596 gcc_assert (TREE_CODE (x) == SSA_NAME);
1598 if (TREE_CODE (y) == SSA_NAME)
1600 tree tmp = SSA_NAME_VALUE (y);
1601 if (tmp)
1602 y = tmp;
1605 record_const_or_copy_1 (x, y, prev_x);
1608 /* Similarly, but assume that X and Y are the two operands of an EQ_EXPR.
1609 This constrains the cases in which we may treat this as assignment. */
1611 static void
1612 record_equality (tree x, tree y)
1614 tree prev_x = NULL, prev_y = NULL;
1616 if (TREE_CODE (x) == SSA_NAME)
1617 prev_x = SSA_NAME_VALUE (x);
1618 if (TREE_CODE (y) == SSA_NAME)
1619 prev_y = SSA_NAME_VALUE (y);
1621 /* If one of the previous values is invariant, or invariant in more loops
1622 (by depth), then use that.
1623 Otherwise it doesn't matter which value we choose, just so
1624 long as we canonicalize on one value. */
1625 if (is_gimple_min_invariant (y))
1627 else if (is_gimple_min_invariant (x)
1628 || (loop_depth_of_name (x) <= loop_depth_of_name (y)))
1629 prev_x = x, x = y, y = prev_x, prev_x = prev_y;
1630 else if (prev_x && is_gimple_min_invariant (prev_x))
1631 x = y, y = prev_x, prev_x = prev_y;
1632 else if (prev_y)
1633 y = prev_y;
1635 /* After the swapping, we must have one SSA_NAME. */
1636 if (TREE_CODE (x) != SSA_NAME)
1637 return;
1639 /* For IEEE, -0.0 == 0.0, so we don't necessarily know the sign of a
1640 variable compared against zero. If we're honoring signed zeros,
1641 then we cannot record this value unless we know that the value is
1642 nonzero. */
1643 if (HONOR_SIGNED_ZEROS (TYPE_MODE (TREE_TYPE (x)))
1644 && (TREE_CODE (y) != REAL_CST
1645 || REAL_VALUES_EQUAL (dconst0, TREE_REAL_CST (y))))
1646 return;
1648 record_const_or_copy_1 (x, y, prev_x);
1651 /* Returns true when STMT is a simple iv increment. It detects the
1652 following situation:
1654 i_1 = phi (..., i_2)
1655 i_2 = i_1 +/- ... */
1657 bool
1658 simple_iv_increment_p (gimple stmt)
1660 enum tree_code code;
1661 tree lhs, preinc;
1662 gimple phi;
1663 size_t i;
1665 if (gimple_code (stmt) != GIMPLE_ASSIGN)
1666 return false;
1668 lhs = gimple_assign_lhs (stmt);
1669 if (TREE_CODE (lhs) != SSA_NAME)
1670 return false;
1672 code = gimple_assign_rhs_code (stmt);
1673 if (code != PLUS_EXPR
1674 && code != MINUS_EXPR
1675 && code != POINTER_PLUS_EXPR)
1676 return false;
1678 preinc = gimple_assign_rhs1 (stmt);
1679 if (TREE_CODE (preinc) != SSA_NAME)
1680 return false;
1682 phi = SSA_NAME_DEF_STMT (preinc);
1683 if (gimple_code (phi) != GIMPLE_PHI)
1684 return false;
1686 for (i = 0; i < gimple_phi_num_args (phi); i++)
1687 if (gimple_phi_arg_def (phi, i) == lhs)
1688 return true;
1690 return false;
1693 /* CONST_AND_COPIES is a table which maps an SSA_NAME to the current
1694 known value for that SSA_NAME (or NULL if no value is known).
1696 Propagate values from CONST_AND_COPIES into the PHI nodes of the
1697 successors of BB. */
1699 static void
1700 cprop_into_successor_phis (basic_block bb)
1702 edge e;
1703 edge_iterator ei;
1705 FOR_EACH_EDGE (e, ei, bb->succs)
1707 int indx;
1708 gimple_stmt_iterator gsi;
1710 /* If this is an abnormal edge, then we do not want to copy propagate
1711 into the PHI alternative associated with this edge. */
1712 if (e->flags & EDGE_ABNORMAL)
1713 continue;
1715 gsi = gsi_start_phis (e->dest);
1716 if (gsi_end_p (gsi))
1717 continue;
1719 /* We may have an equivalence associated with this edge. While
1720 we can not propagate it into non-dominated blocks, we can
1721 propagate them into PHIs in non-dominated blocks. */
1723 /* Push the unwind marker so we can reset the const and copies
1724 table back to its original state after processing this edge. */
1725 const_and_copies_stack.safe_push (NULL_TREE);
1727 /* Extract and record any simple NAME = VALUE equivalences.
1729 Don't bother with [01] = COND equivalences, they're not useful
1730 here. */
1731 struct edge_info *edge_info = (struct edge_info *) e->aux;
1732 if (edge_info)
1734 tree lhs = edge_info->lhs;
1735 tree rhs = edge_info->rhs;
1737 if (lhs && TREE_CODE (lhs) == SSA_NAME)
1738 record_const_or_copy (lhs, rhs);
1741 indx = e->dest_idx;
1742 for ( ; !gsi_end_p (gsi); gsi_next (&gsi))
1744 tree new_val;
1745 use_operand_p orig_p;
1746 tree orig_val;
1747 gimple phi = gsi_stmt (gsi);
1749 /* The alternative may be associated with a constant, so verify
1750 it is an SSA_NAME before doing anything with it. */
1751 orig_p = gimple_phi_arg_imm_use_ptr (phi, indx);
1752 orig_val = get_use_from_ptr (orig_p);
1753 if (TREE_CODE (orig_val) != SSA_NAME)
1754 continue;
1756 /* If we have *ORIG_P in our constant/copy table, then replace
1757 ORIG_P with its value in our constant/copy table. */
1758 new_val = SSA_NAME_VALUE (orig_val);
1759 if (new_val
1760 && new_val != orig_val
1761 && (TREE_CODE (new_val) == SSA_NAME
1762 || is_gimple_min_invariant (new_val))
1763 && may_propagate_copy (orig_val, new_val))
1764 propagate_value (orig_p, new_val);
1767 restore_vars_to_original_value ();
1771 /* We have finished optimizing BB, record any information implied by
1772 taking a specific outgoing edge from BB. */
1774 static void
1775 record_edge_info (basic_block bb)
1777 gimple_stmt_iterator gsi = gsi_last_bb (bb);
1778 struct edge_info *edge_info;
1780 if (! gsi_end_p (gsi))
1782 gimple stmt = gsi_stmt (gsi);
1783 location_t loc = gimple_location (stmt);
1785 if (gimple_code (stmt) == GIMPLE_SWITCH)
1787 tree index = gimple_switch_index (stmt);
1789 if (TREE_CODE (index) == SSA_NAME)
1791 int i;
1792 int n_labels = gimple_switch_num_labels (stmt);
1793 tree *info = XCNEWVEC (tree, last_basic_block_for_fn (cfun));
1794 edge e;
1795 edge_iterator ei;
1797 for (i = 0; i < n_labels; i++)
1799 tree label = gimple_switch_label (stmt, i);
1800 basic_block target_bb = label_to_block (CASE_LABEL (label));
1801 if (CASE_HIGH (label)
1802 || !CASE_LOW (label)
1803 || info[target_bb->index])
1804 info[target_bb->index] = error_mark_node;
1805 else
1806 info[target_bb->index] = label;
1809 FOR_EACH_EDGE (e, ei, bb->succs)
1811 basic_block target_bb = e->dest;
1812 tree label = info[target_bb->index];
1814 if (label != NULL && label != error_mark_node)
1816 tree x = fold_convert_loc (loc, TREE_TYPE (index),
1817 CASE_LOW (label));
1818 edge_info = allocate_edge_info (e);
1819 edge_info->lhs = index;
1820 edge_info->rhs = x;
1823 free (info);
1827 /* A COND_EXPR may create equivalences too. */
1828 if (gimple_code (stmt) == GIMPLE_COND)
1830 edge true_edge;
1831 edge false_edge;
1833 tree op0 = gimple_cond_lhs (stmt);
1834 tree op1 = gimple_cond_rhs (stmt);
1835 enum tree_code code = gimple_cond_code (stmt);
1837 extract_true_false_edges_from_block (bb, &true_edge, &false_edge);
1839 /* Special case comparing booleans against a constant as we
1840 know the value of OP0 on both arms of the branch. i.e., we
1841 can record an equivalence for OP0 rather than COND. */
1842 if ((code == EQ_EXPR || code == NE_EXPR)
1843 && TREE_CODE (op0) == SSA_NAME
1844 && TREE_CODE (TREE_TYPE (op0)) == BOOLEAN_TYPE
1845 && is_gimple_min_invariant (op1))
1847 if (code == EQ_EXPR)
1849 edge_info = allocate_edge_info (true_edge);
1850 edge_info->lhs = op0;
1851 edge_info->rhs = (integer_zerop (op1)
1852 ? boolean_false_node
1853 : boolean_true_node);
1855 edge_info = allocate_edge_info (false_edge);
1856 edge_info->lhs = op0;
1857 edge_info->rhs = (integer_zerop (op1)
1858 ? boolean_true_node
1859 : boolean_false_node);
1861 else
1863 edge_info = allocate_edge_info (true_edge);
1864 edge_info->lhs = op0;
1865 edge_info->rhs = (integer_zerop (op1)
1866 ? boolean_true_node
1867 : boolean_false_node);
1869 edge_info = allocate_edge_info (false_edge);
1870 edge_info->lhs = op0;
1871 edge_info->rhs = (integer_zerop (op1)
1872 ? boolean_false_node
1873 : boolean_true_node);
1876 else if (is_gimple_min_invariant (op0)
1877 && (TREE_CODE (op1) == SSA_NAME
1878 || is_gimple_min_invariant (op1)))
1880 tree cond = build2 (code, boolean_type_node, op0, op1);
1881 tree inverted = invert_truthvalue_loc (loc, cond);
1882 bool can_infer_simple_equiv
1883 = !(HONOR_SIGNED_ZEROS (TYPE_MODE (TREE_TYPE (op0)))
1884 && real_zerop (op0));
1885 struct edge_info *edge_info;
1887 edge_info = allocate_edge_info (true_edge);
1888 record_conditions (edge_info, cond, inverted);
1890 if (can_infer_simple_equiv && code == EQ_EXPR)
1892 edge_info->lhs = op1;
1893 edge_info->rhs = op0;
1896 edge_info = allocate_edge_info (false_edge);
1897 record_conditions (edge_info, inverted, cond);
1899 if (can_infer_simple_equiv && TREE_CODE (inverted) == EQ_EXPR)
1901 edge_info->lhs = op1;
1902 edge_info->rhs = op0;
1906 else if (TREE_CODE (op0) == SSA_NAME
1907 && (TREE_CODE (op1) == SSA_NAME
1908 || is_gimple_min_invariant (op1)))
1910 tree cond = build2 (code, boolean_type_node, op0, op1);
1911 tree inverted = invert_truthvalue_loc (loc, cond);
1912 bool can_infer_simple_equiv
1913 = !(HONOR_SIGNED_ZEROS (TYPE_MODE (TREE_TYPE (op1)))
1914 && (TREE_CODE (op1) == SSA_NAME || real_zerop (op1)));
1915 struct edge_info *edge_info;
1917 edge_info = allocate_edge_info (true_edge);
1918 record_conditions (edge_info, cond, inverted);
1920 if (can_infer_simple_equiv && code == EQ_EXPR)
1922 edge_info->lhs = op0;
1923 edge_info->rhs = op1;
1926 edge_info = allocate_edge_info (false_edge);
1927 record_conditions (edge_info, inverted, cond);
1929 if (can_infer_simple_equiv && TREE_CODE (inverted) == EQ_EXPR)
1931 edge_info->lhs = op0;
1932 edge_info->rhs = op1;
1937 /* ??? TRUTH_NOT_EXPR can create an equivalence too. */
1941 void
1942 dom_opt_dom_walker::before_dom_children (basic_block bb)
1944 gimple_stmt_iterator gsi;
1946 if (dump_file && (dump_flags & TDF_DETAILS))
1947 fprintf (dump_file, "\n\nOptimizing block #%d\n\n", bb->index);
1949 /* Push a marker on the stacks of local information so that we know how
1950 far to unwind when we finalize this block. */
1951 avail_exprs_stack.safe_push (NULL);
1952 const_and_copies_stack.safe_push (NULL_TREE);
1954 record_equivalences_from_incoming_edge (bb);
1956 /* PHI nodes can create equivalences too. */
1957 record_equivalences_from_phis (bb);
1959 /* Create equivalences from redundant PHIs. PHIs are only truly
1960 redundant when they exist in the same block, so push another
1961 marker and unwind right afterwards. */
1962 avail_exprs_stack.safe_push (NULL);
1963 for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi))
1964 eliminate_redundant_computations (&gsi);
1965 remove_local_expressions_from_table ();
1967 for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
1968 optimize_stmt (bb, gsi);
1970 /* Now prepare to process dominated blocks. */
1971 record_edge_info (bb);
1972 cprop_into_successor_phis (bb);
1975 /* We have finished processing the dominator children of BB, perform
1976 any finalization actions in preparation for leaving this node in
1977 the dominator tree. */
1979 void
1980 dom_opt_dom_walker::after_dom_children (basic_block bb)
1982 gimple last;
1984 /* If we have an outgoing edge to a block with multiple incoming and
1985 outgoing edges, then we may be able to thread the edge, i.e., we
1986 may be able to statically determine which of the outgoing edges
1987 will be traversed when the incoming edge from BB is traversed. */
1988 if (single_succ_p (bb)
1989 && (single_succ_edge (bb)->flags & EDGE_ABNORMAL) == 0
1990 && potentially_threadable_block (single_succ (bb)))
1992 thread_across_edge (single_succ_edge (bb));
1994 else if ((last = last_stmt (bb))
1995 && gimple_code (last) == GIMPLE_COND
1996 && EDGE_COUNT (bb->succs) == 2
1997 && (EDGE_SUCC (bb, 0)->flags & EDGE_ABNORMAL) == 0
1998 && (EDGE_SUCC (bb, 1)->flags & EDGE_ABNORMAL) == 0)
2000 edge true_edge, false_edge;
2002 extract_true_false_edges_from_block (bb, &true_edge, &false_edge);
2004 /* Only try to thread the edge if it reaches a target block with
2005 more than one predecessor and more than one successor. */
2006 if (potentially_threadable_block (true_edge->dest))
2007 thread_across_edge (true_edge);
2009 /* Similarly for the ELSE arm. */
2010 if (potentially_threadable_block (false_edge->dest))
2011 thread_across_edge (false_edge);
2015 /* These remove expressions local to BB from the tables. */
2016 remove_local_expressions_from_table ();
2017 restore_vars_to_original_value ();
2020 /* Search for redundant computations in STMT. If any are found, then
2021 replace them with the variable holding the result of the computation.
2023 If safe, record this expression into the available expression hash
2024 table. */
2026 static void
2027 eliminate_redundant_computations (gimple_stmt_iterator* gsi)
2029 tree expr_type;
2030 tree cached_lhs;
2031 tree def;
2032 bool insert = true;
2033 bool assigns_var_p = false;
2035 gimple stmt = gsi_stmt (*gsi);
2037 if (gimple_code (stmt) == GIMPLE_PHI)
2038 def = gimple_phi_result (stmt);
2039 else
2040 def = gimple_get_lhs (stmt);
2042 /* Certain expressions on the RHS can be optimized away, but can not
2043 themselves be entered into the hash tables. */
2044 if (! def
2045 || TREE_CODE (def) != SSA_NAME
2046 || SSA_NAME_OCCURS_IN_ABNORMAL_PHI (def)
2047 || gimple_vdef (stmt)
2048 /* Do not record equivalences for increments of ivs. This would create
2049 overlapping live ranges for a very questionable gain. */
2050 || simple_iv_increment_p (stmt))
2051 insert = false;
2053 /* Check if the expression has been computed before. */
2054 cached_lhs = lookup_avail_expr (stmt, insert);
2056 opt_stats.num_exprs_considered++;
2058 /* Get the type of the expression we are trying to optimize. */
2059 if (is_gimple_assign (stmt))
2061 expr_type = TREE_TYPE (gimple_assign_lhs (stmt));
2062 assigns_var_p = true;
2064 else if (gimple_code (stmt) == GIMPLE_COND)
2065 expr_type = boolean_type_node;
2066 else if (is_gimple_call (stmt))
2068 gcc_assert (gimple_call_lhs (stmt));
2069 expr_type = TREE_TYPE (gimple_call_lhs (stmt));
2070 assigns_var_p = true;
2072 else if (gimple_code (stmt) == GIMPLE_SWITCH)
2073 expr_type = TREE_TYPE (gimple_switch_index (stmt));
2074 else if (gimple_code (stmt) == GIMPLE_PHI)
2075 /* We can't propagate into a phi, so the logic below doesn't apply.
2076 Instead record an equivalence between the cached LHS and the
2077 PHI result of this statement, provided they are in the same block.
2078 This should be sufficient to kill the redundant phi. */
2080 if (def && cached_lhs)
2081 record_const_or_copy (def, cached_lhs);
2082 return;
2084 else
2085 gcc_unreachable ();
2087 if (!cached_lhs)
2088 return;
2090 /* It is safe to ignore types here since we have already done
2091 type checking in the hashing and equality routines. In fact
2092 type checking here merely gets in the way of constant
2093 propagation. Also, make sure that it is safe to propagate
2094 CACHED_LHS into the expression in STMT. */
2095 if ((TREE_CODE (cached_lhs) != SSA_NAME
2096 && (assigns_var_p
2097 || useless_type_conversion_p (expr_type, TREE_TYPE (cached_lhs))))
2098 || may_propagate_copy_into_stmt (stmt, cached_lhs))
2100 gcc_checking_assert (TREE_CODE (cached_lhs) == SSA_NAME
2101 || is_gimple_min_invariant (cached_lhs));
2103 if (dump_file && (dump_flags & TDF_DETAILS))
2105 fprintf (dump_file, " Replaced redundant expr '");
2106 print_gimple_expr (dump_file, stmt, 0, dump_flags);
2107 fprintf (dump_file, "' with '");
2108 print_generic_expr (dump_file, cached_lhs, dump_flags);
2109 fprintf (dump_file, "'\n");
2112 opt_stats.num_re++;
2114 if (assigns_var_p
2115 && !useless_type_conversion_p (expr_type, TREE_TYPE (cached_lhs)))
2116 cached_lhs = fold_convert (expr_type, cached_lhs);
2118 propagate_tree_value_into_stmt (gsi, cached_lhs);
2120 /* Since it is always necessary to mark the result as modified,
2121 perhaps we should move this into propagate_tree_value_into_stmt
2122 itself. */
2123 gimple_set_modified (gsi_stmt (*gsi), true);
2127 /* STMT, a GIMPLE_ASSIGN, may create certain equivalences, in either
2128 the available expressions table or the const_and_copies table.
2129 Detect and record those equivalences. */
2130 /* We handle only very simple copy equivalences here. The heavy
2131 lifing is done by eliminate_redundant_computations. */
2133 static void
2134 record_equivalences_from_stmt (gimple stmt, int may_optimize_p)
2136 tree lhs;
2137 enum tree_code lhs_code;
2139 gcc_assert (is_gimple_assign (stmt));
2141 lhs = gimple_assign_lhs (stmt);
2142 lhs_code = TREE_CODE (lhs);
2144 if (lhs_code == SSA_NAME
2145 && gimple_assign_single_p (stmt))
2147 tree rhs = gimple_assign_rhs1 (stmt);
2149 /* If the RHS of the assignment is a constant or another variable that
2150 may be propagated, register it in the CONST_AND_COPIES table. We
2151 do not need to record unwind data for this, since this is a true
2152 assignment and not an equivalence inferred from a comparison. All
2153 uses of this ssa name are dominated by this assignment, so unwinding
2154 just costs time and space. */
2155 if (may_optimize_p
2156 && (TREE_CODE (rhs) == SSA_NAME
2157 || is_gimple_min_invariant (rhs)))
2159 if (dump_file && (dump_flags & TDF_DETAILS))
2161 fprintf (dump_file, "==== ASGN ");
2162 print_generic_expr (dump_file, lhs, 0);
2163 fprintf (dump_file, " = ");
2164 print_generic_expr (dump_file, rhs, 0);
2165 fprintf (dump_file, "\n");
2168 set_ssa_name_value (lhs, rhs);
2172 /* A memory store, even an aliased store, creates a useful
2173 equivalence. By exchanging the LHS and RHS, creating suitable
2174 vops and recording the result in the available expression table,
2175 we may be able to expose more redundant loads. */
2176 if (!gimple_has_volatile_ops (stmt)
2177 && gimple_references_memory_p (stmt)
2178 && gimple_assign_single_p (stmt)
2179 && (TREE_CODE (gimple_assign_rhs1 (stmt)) == SSA_NAME
2180 || is_gimple_min_invariant (gimple_assign_rhs1 (stmt)))
2181 && !is_gimple_reg (lhs))
2183 tree rhs = gimple_assign_rhs1 (stmt);
2184 gimple new_stmt;
2186 /* Build a new statement with the RHS and LHS exchanged. */
2187 if (TREE_CODE (rhs) == SSA_NAME)
2189 /* NOTE tuples. The call to gimple_build_assign below replaced
2190 a call to build_gimple_modify_stmt, which did not set the
2191 SSA_NAME_DEF_STMT on the LHS of the assignment. Doing so
2192 may cause an SSA validation failure, as the LHS may be a
2193 default-initialized name and should have no definition. I'm
2194 a bit dubious of this, as the artificial statement that we
2195 generate here may in fact be ill-formed, but it is simply
2196 used as an internal device in this pass, and never becomes
2197 part of the CFG. */
2198 gimple defstmt = SSA_NAME_DEF_STMT (rhs);
2199 new_stmt = gimple_build_assign (rhs, lhs);
2200 SSA_NAME_DEF_STMT (rhs) = defstmt;
2202 else
2203 new_stmt = gimple_build_assign (rhs, lhs);
2205 gimple_set_vuse (new_stmt, gimple_vdef (stmt));
2207 /* Finally enter the statement into the available expression
2208 table. */
2209 lookup_avail_expr (new_stmt, true);
2213 /* Replace *OP_P in STMT with any known equivalent value for *OP_P from
2214 CONST_AND_COPIES. */
2216 static void
2217 cprop_operand (gimple stmt, use_operand_p op_p)
2219 tree val;
2220 tree op = USE_FROM_PTR (op_p);
2222 /* If the operand has a known constant value or it is known to be a
2223 copy of some other variable, use the value or copy stored in
2224 CONST_AND_COPIES. */
2225 val = SSA_NAME_VALUE (op);
2226 if (val && val != op)
2228 /* Do not replace hard register operands in asm statements. */
2229 if (gimple_code (stmt) == GIMPLE_ASM
2230 && !may_propagate_copy_into_asm (op))
2231 return;
2233 /* Certain operands are not allowed to be copy propagated due
2234 to their interaction with exception handling and some GCC
2235 extensions. */
2236 if (!may_propagate_copy (op, val))
2237 return;
2239 /* Do not propagate addresses that point to volatiles into memory
2240 stmts without volatile operands. */
2241 if (POINTER_TYPE_P (TREE_TYPE (val))
2242 && TYPE_VOLATILE (TREE_TYPE (TREE_TYPE (val)))
2243 && gimple_has_mem_ops (stmt)
2244 && !gimple_has_volatile_ops (stmt))
2245 return;
2247 /* Do not propagate copies if the propagated value is at a deeper loop
2248 depth than the propagatee. Otherwise, this may move loop variant
2249 variables outside of their loops and prevent coalescing
2250 opportunities. If the value was loop invariant, it will be hoisted
2251 by LICM and exposed for copy propagation. */
2252 if (loop_depth_of_name (val) > loop_depth_of_name (op))
2253 return;
2255 /* Do not propagate copies into simple IV increment statements.
2256 See PR23821 for how this can disturb IV analysis. */
2257 if (TREE_CODE (val) != INTEGER_CST
2258 && simple_iv_increment_p (stmt))
2259 return;
2261 /* Dump details. */
2262 if (dump_file && (dump_flags & TDF_DETAILS))
2264 fprintf (dump_file, " Replaced '");
2265 print_generic_expr (dump_file, op, dump_flags);
2266 fprintf (dump_file, "' with %s '",
2267 (TREE_CODE (val) != SSA_NAME ? "constant" : "variable"));
2268 print_generic_expr (dump_file, val, dump_flags);
2269 fprintf (dump_file, "'\n");
2272 if (TREE_CODE (val) != SSA_NAME)
2273 opt_stats.num_const_prop++;
2274 else
2275 opt_stats.num_copy_prop++;
2277 propagate_value (op_p, val);
2279 /* And note that we modified this statement. This is now
2280 safe, even if we changed virtual operands since we will
2281 rescan the statement and rewrite its operands again. */
2282 gimple_set_modified (stmt, true);
2286 /* CONST_AND_COPIES is a table which maps an SSA_NAME to the current
2287 known value for that SSA_NAME (or NULL if no value is known).
2289 Propagate values from CONST_AND_COPIES into the uses, vuses and
2290 vdef_ops of STMT. */
2292 static void
2293 cprop_into_stmt (gimple stmt)
2295 use_operand_p op_p;
2296 ssa_op_iter iter;
2298 FOR_EACH_SSA_USE_OPERAND (op_p, stmt, iter, SSA_OP_USE)
2299 cprop_operand (stmt, op_p);
2302 /* Optimize the statement pointed to by iterator SI.
2304 We try to perform some simplistic global redundancy elimination and
2305 constant propagation:
2307 1- To detect global redundancy, we keep track of expressions that have
2308 been computed in this block and its dominators. If we find that the
2309 same expression is computed more than once, we eliminate repeated
2310 computations by using the target of the first one.
2312 2- Constant values and copy assignments. This is used to do very
2313 simplistic constant and copy propagation. When a constant or copy
2314 assignment is found, we map the value on the RHS of the assignment to
2315 the variable in the LHS in the CONST_AND_COPIES table. */
2317 static void
2318 optimize_stmt (basic_block bb, gimple_stmt_iterator si)
2320 gimple stmt, old_stmt;
2321 bool may_optimize_p;
2322 bool modified_p = false;
2324 old_stmt = stmt = gsi_stmt (si);
2326 if (dump_file && (dump_flags & TDF_DETAILS))
2328 fprintf (dump_file, "Optimizing statement ");
2329 print_gimple_stmt (dump_file, stmt, 0, TDF_SLIM);
2332 if (gimple_code (stmt) == GIMPLE_COND)
2333 canonicalize_comparison (stmt);
2335 update_stmt_if_modified (stmt);
2336 opt_stats.num_stmts++;
2338 /* Const/copy propagate into USES, VUSES and the RHS of VDEFs. */
2339 cprop_into_stmt (stmt);
2341 /* If the statement has been modified with constant replacements,
2342 fold its RHS before checking for redundant computations. */
2343 if (gimple_modified_p (stmt))
2345 tree rhs = NULL;
2347 /* Try to fold the statement making sure that STMT is kept
2348 up to date. */
2349 if (fold_stmt (&si))
2351 stmt = gsi_stmt (si);
2352 gimple_set_modified (stmt, true);
2354 if (dump_file && (dump_flags & TDF_DETAILS))
2356 fprintf (dump_file, " Folded to: ");
2357 print_gimple_stmt (dump_file, stmt, 0, TDF_SLIM);
2361 /* We only need to consider cases that can yield a gimple operand. */
2362 if (gimple_assign_single_p (stmt))
2363 rhs = gimple_assign_rhs1 (stmt);
2364 else if (gimple_code (stmt) == GIMPLE_GOTO)
2365 rhs = gimple_goto_dest (stmt);
2366 else if (gimple_code (stmt) == GIMPLE_SWITCH)
2367 /* This should never be an ADDR_EXPR. */
2368 rhs = gimple_switch_index (stmt);
2370 if (rhs && TREE_CODE (rhs) == ADDR_EXPR)
2371 recompute_tree_invariant_for_addr_expr (rhs);
2373 /* Indicate that maybe_clean_or_replace_eh_stmt needs to be called,
2374 even if fold_stmt updated the stmt already and thus cleared
2375 gimple_modified_p flag on it. */
2376 modified_p = true;
2379 /* Check for redundant computations. Do this optimization only
2380 for assignments that have no volatile ops and conditionals. */
2381 may_optimize_p = (!gimple_has_side_effects (stmt)
2382 && (is_gimple_assign (stmt)
2383 || (is_gimple_call (stmt)
2384 && gimple_call_lhs (stmt) != NULL_TREE)
2385 || gimple_code (stmt) == GIMPLE_COND
2386 || gimple_code (stmt) == GIMPLE_SWITCH));
2388 if (may_optimize_p)
2390 if (gimple_code (stmt) == GIMPLE_CALL)
2392 /* Resolve __builtin_constant_p. If it hasn't been
2393 folded to integer_one_node by now, it's fairly
2394 certain that the value simply isn't constant. */
2395 tree callee = gimple_call_fndecl (stmt);
2396 if (callee
2397 && DECL_BUILT_IN_CLASS (callee) == BUILT_IN_NORMAL
2398 && DECL_FUNCTION_CODE (callee) == BUILT_IN_CONSTANT_P)
2400 propagate_tree_value_into_stmt (&si, integer_zero_node);
2401 stmt = gsi_stmt (si);
2405 update_stmt_if_modified (stmt);
2406 eliminate_redundant_computations (&si);
2407 stmt = gsi_stmt (si);
2409 /* Perform simple redundant store elimination. */
2410 if (gimple_assign_single_p (stmt)
2411 && TREE_CODE (gimple_assign_lhs (stmt)) != SSA_NAME)
2413 tree lhs = gimple_assign_lhs (stmt);
2414 tree rhs = gimple_assign_rhs1 (stmt);
2415 tree cached_lhs;
2416 gimple new_stmt;
2417 if (TREE_CODE (rhs) == SSA_NAME)
2419 tree tem = SSA_NAME_VALUE (rhs);
2420 if (tem)
2421 rhs = tem;
2423 /* Build a new statement with the RHS and LHS exchanged. */
2424 if (TREE_CODE (rhs) == SSA_NAME)
2426 gimple defstmt = SSA_NAME_DEF_STMT (rhs);
2427 new_stmt = gimple_build_assign (rhs, lhs);
2428 SSA_NAME_DEF_STMT (rhs) = defstmt;
2430 else
2431 new_stmt = gimple_build_assign (rhs, lhs);
2432 gimple_set_vuse (new_stmt, gimple_vuse (stmt));
2433 cached_lhs = lookup_avail_expr (new_stmt, false);
2434 if (cached_lhs
2435 && rhs == cached_lhs)
2437 basic_block bb = gimple_bb (stmt);
2438 unlink_stmt_vdef (stmt);
2439 if (gsi_remove (&si, true))
2441 bitmap_set_bit (need_eh_cleanup, bb->index);
2442 if (dump_file && (dump_flags & TDF_DETAILS))
2443 fprintf (dump_file, " Flagged to clear EH edges.\n");
2445 release_defs (stmt);
2446 return;
2451 /* Record any additional equivalences created by this statement. */
2452 if (is_gimple_assign (stmt))
2453 record_equivalences_from_stmt (stmt, may_optimize_p);
2455 /* If STMT is a COND_EXPR and it was modified, then we may know
2456 where it goes. If that is the case, then mark the CFG as altered.
2458 This will cause us to later call remove_unreachable_blocks and
2459 cleanup_tree_cfg when it is safe to do so. It is not safe to
2460 clean things up here since removal of edges and such can trigger
2461 the removal of PHI nodes, which in turn can release SSA_NAMEs to
2462 the manager.
2464 That's all fine and good, except that once SSA_NAMEs are released
2465 to the manager, we must not call create_ssa_name until all references
2466 to released SSA_NAMEs have been eliminated.
2468 All references to the deleted SSA_NAMEs can not be eliminated until
2469 we remove unreachable blocks.
2471 We can not remove unreachable blocks until after we have completed
2472 any queued jump threading.
2474 We can not complete any queued jump threads until we have taken
2475 appropriate variables out of SSA form. Taking variables out of
2476 SSA form can call create_ssa_name and thus we lose.
2478 Ultimately I suspect we're going to need to change the interface
2479 into the SSA_NAME manager. */
2480 if (gimple_modified_p (stmt) || modified_p)
2482 tree val = NULL;
2484 update_stmt_if_modified (stmt);
2486 if (gimple_code (stmt) == GIMPLE_COND)
2487 val = fold_binary_loc (gimple_location (stmt),
2488 gimple_cond_code (stmt), boolean_type_node,
2489 gimple_cond_lhs (stmt), gimple_cond_rhs (stmt));
2490 else if (gimple_code (stmt) == GIMPLE_SWITCH)
2491 val = gimple_switch_index (stmt);
2493 if (val && TREE_CODE (val) == INTEGER_CST && find_taken_edge (bb, val))
2494 cfg_altered = true;
2496 /* If we simplified a statement in such a way as to be shown that it
2497 cannot trap, update the eh information and the cfg to match. */
2498 if (maybe_clean_or_replace_eh_stmt (old_stmt, stmt))
2500 bitmap_set_bit (need_eh_cleanup, bb->index);
2501 if (dump_file && (dump_flags & TDF_DETAILS))
2502 fprintf (dump_file, " Flagged to clear EH edges.\n");
2507 /* Search for an existing instance of STMT in the AVAIL_EXPRS table.
2508 If found, return its LHS. Otherwise insert STMT in the table and
2509 return NULL_TREE.
2511 Also, when an expression is first inserted in the table, it is also
2512 is also added to AVAIL_EXPRS_STACK, so that it can be removed when
2513 we finish processing this block and its children. */
2515 static tree
2516 lookup_avail_expr (gimple stmt, bool insert)
2518 expr_hash_elt **slot;
2519 tree lhs;
2520 tree temp;
2521 struct expr_hash_elt element;
2523 /* Get LHS of phi, assignment, or call; else NULL_TREE. */
2524 if (gimple_code (stmt) == GIMPLE_PHI)
2525 lhs = gimple_phi_result (stmt);
2526 else
2527 lhs = gimple_get_lhs (stmt);
2529 initialize_hash_element (stmt, lhs, &element);
2531 if (dump_file && (dump_flags & TDF_DETAILS))
2533 fprintf (dump_file, "LKUP ");
2534 print_expr_hash_elt (dump_file, &element);
2537 /* Don't bother remembering constant assignments and copy operations.
2538 Constants and copy operations are handled by the constant/copy propagator
2539 in optimize_stmt. */
2540 if (element.expr.kind == EXPR_SINGLE
2541 && (TREE_CODE (element.expr.ops.single.rhs) == SSA_NAME
2542 || is_gimple_min_invariant (element.expr.ops.single.rhs)))
2543 return NULL_TREE;
2545 /* Finally try to find the expression in the main expression hash table. */
2546 slot = avail_exprs.find_slot_with_hash (&element, element.hash,
2547 (insert ? INSERT : NO_INSERT));
2548 if (slot == NULL)
2550 free_expr_hash_elt_contents (&element);
2551 return NULL_TREE;
2553 else if (*slot == NULL)
2555 struct expr_hash_elt *element2 = XNEW (struct expr_hash_elt);
2556 *element2 = element;
2557 element2->stamp = element2;
2558 *slot = element2;
2560 if (dump_file && (dump_flags & TDF_DETAILS))
2562 fprintf (dump_file, "2>>> ");
2563 print_expr_hash_elt (dump_file, element2);
2566 avail_exprs_stack.safe_push (element2);
2567 return NULL_TREE;
2569 else
2570 free_expr_hash_elt_contents (&element);
2572 /* Extract the LHS of the assignment so that it can be used as the current
2573 definition of another variable. */
2574 lhs = ((struct expr_hash_elt *)*slot)->lhs;
2576 /* See if the LHS appears in the CONST_AND_COPIES table. If it does, then
2577 use the value from the const_and_copies table. */
2578 if (TREE_CODE (lhs) == SSA_NAME)
2580 temp = SSA_NAME_VALUE (lhs);
2581 if (temp)
2582 lhs = temp;
2585 if (dump_file && (dump_flags & TDF_DETAILS))
2587 fprintf (dump_file, "FIND: ");
2588 print_generic_expr (dump_file, lhs, 0);
2589 fprintf (dump_file, "\n");
2592 return lhs;
2595 /* Hashing and equality functions for AVAIL_EXPRS. We compute a value number
2596 for expressions using the code of the expression and the SSA numbers of
2597 its operands. */
2599 static hashval_t
2600 avail_expr_hash (const void *p)
2602 gimple stmt = ((const struct expr_hash_elt *)p)->stmt;
2603 const struct hashable_expr *expr = &((const struct expr_hash_elt *)p)->expr;
2604 tree vuse;
2605 hashval_t val = 0;
2607 val = iterative_hash_hashable_expr (expr, val);
2609 /* If the hash table entry is not associated with a statement, then we
2610 can just hash the expression and not worry about virtual operands
2611 and such. */
2612 if (!stmt)
2613 return val;
2615 /* Add the SSA version numbers of the vuse operand. This is important
2616 because compound variables like arrays are not renamed in the
2617 operands. Rather, the rename is done on the virtual variable
2618 representing all the elements of the array. */
2619 if ((vuse = gimple_vuse (stmt)))
2620 val = iterative_hash_expr (vuse, val);
2622 return val;
2625 /* PHI-ONLY copy and constant propagation. This pass is meant to clean
2626 up degenerate PHIs created by or exposed by jump threading. */
2628 /* Given a statement STMT, which is either a PHI node or an assignment,
2629 remove it from the IL. */
2631 static void
2632 remove_stmt_or_phi (gimple stmt)
2634 gimple_stmt_iterator gsi = gsi_for_stmt (stmt);
2636 if (gimple_code (stmt) == GIMPLE_PHI)
2637 remove_phi_node (&gsi, true);
2638 else
2640 gsi_remove (&gsi, true);
2641 release_defs (stmt);
2645 /* Given a statement STMT, which is either a PHI node or an assignment,
2646 return the "rhs" of the node, in the case of a non-degenerate
2647 phi, NULL is returned. */
2649 static tree
2650 get_rhs_or_phi_arg (gimple stmt)
2652 if (gimple_code (stmt) == GIMPLE_PHI)
2653 return degenerate_phi_result (stmt);
2654 else if (gimple_assign_single_p (stmt))
2655 return gimple_assign_rhs1 (stmt);
2656 else
2657 gcc_unreachable ();
2661 /* Given a statement STMT, which is either a PHI node or an assignment,
2662 return the "lhs" of the node. */
2664 static tree
2665 get_lhs_or_phi_result (gimple stmt)
2667 if (gimple_code (stmt) == GIMPLE_PHI)
2668 return gimple_phi_result (stmt);
2669 else if (is_gimple_assign (stmt))
2670 return gimple_assign_lhs (stmt);
2671 else
2672 gcc_unreachable ();
2675 /* Propagate RHS into all uses of LHS (when possible).
2677 RHS and LHS are derived from STMT, which is passed in solely so
2678 that we can remove it if propagation is successful.
2680 When propagating into a PHI node or into a statement which turns
2681 into a trivial copy or constant initialization, set the
2682 appropriate bit in INTERESTING_NAMEs so that we will visit those
2683 nodes as well in an effort to pick up secondary optimization
2684 opportunities. */
2686 static void
2687 propagate_rhs_into_lhs (gimple stmt, tree lhs, tree rhs, bitmap interesting_names)
2689 /* First verify that propagation is valid and isn't going to move a
2690 loop variant variable outside its loop. */
2691 if (! SSA_NAME_OCCURS_IN_ABNORMAL_PHI (lhs)
2692 && (TREE_CODE (rhs) != SSA_NAME
2693 || ! SSA_NAME_OCCURS_IN_ABNORMAL_PHI (rhs))
2694 && may_propagate_copy (lhs, rhs)
2695 && loop_depth_of_name (lhs) >= loop_depth_of_name (rhs))
2697 use_operand_p use_p;
2698 imm_use_iterator iter;
2699 gimple use_stmt;
2700 bool all = true;
2702 /* Dump details. */
2703 if (dump_file && (dump_flags & TDF_DETAILS))
2705 fprintf (dump_file, " Replacing '");
2706 print_generic_expr (dump_file, lhs, dump_flags);
2707 fprintf (dump_file, "' with %s '",
2708 (TREE_CODE (rhs) != SSA_NAME ? "constant" : "variable"));
2709 print_generic_expr (dump_file, rhs, dump_flags);
2710 fprintf (dump_file, "'\n");
2713 /* Walk over every use of LHS and try to replace the use with RHS.
2714 At this point the only reason why such a propagation would not
2715 be successful would be if the use occurs in an ASM_EXPR. */
2716 FOR_EACH_IMM_USE_STMT (use_stmt, iter, lhs)
2718 /* Leave debug stmts alone. If we succeed in propagating
2719 all non-debug uses, we'll drop the DEF, and propagation
2720 into debug stmts will occur then. */
2721 if (gimple_debug_bind_p (use_stmt))
2722 continue;
2724 /* It's not always safe to propagate into an ASM_EXPR. */
2725 if (gimple_code (use_stmt) == GIMPLE_ASM
2726 && ! may_propagate_copy_into_asm (lhs))
2728 all = false;
2729 continue;
2732 /* It's not ok to propagate into the definition stmt of RHS.
2733 <bb 9>:
2734 # prephitmp.12_36 = PHI <g_67.1_6(9)>
2735 g_67.1_6 = prephitmp.12_36;
2736 goto <bb 9>;
2737 While this is strictly all dead code we do not want to
2738 deal with this here. */
2739 if (TREE_CODE (rhs) == SSA_NAME
2740 && SSA_NAME_DEF_STMT (rhs) == use_stmt)
2742 all = false;
2743 continue;
2746 /* Dump details. */
2747 if (dump_file && (dump_flags & TDF_DETAILS))
2749 fprintf (dump_file, " Original statement:");
2750 print_gimple_stmt (dump_file, use_stmt, 0, dump_flags);
2753 /* Propagate the RHS into this use of the LHS. */
2754 FOR_EACH_IMM_USE_ON_STMT (use_p, iter)
2755 propagate_value (use_p, rhs);
2757 /* Special cases to avoid useless calls into the folding
2758 routines, operand scanning, etc.
2760 Propagation into a PHI may cause the PHI to become
2761 a degenerate, so mark the PHI as interesting. No other
2762 actions are necessary. */
2763 if (gimple_code (use_stmt) == GIMPLE_PHI)
2765 tree result;
2767 /* Dump details. */
2768 if (dump_file && (dump_flags & TDF_DETAILS))
2770 fprintf (dump_file, " Updated statement:");
2771 print_gimple_stmt (dump_file, use_stmt, 0, dump_flags);
2774 result = get_lhs_or_phi_result (use_stmt);
2775 bitmap_set_bit (interesting_names, SSA_NAME_VERSION (result));
2776 continue;
2779 /* From this point onward we are propagating into a
2780 real statement. Folding may (or may not) be possible,
2781 we may expose new operands, expose dead EH edges,
2782 etc. */
2783 /* NOTE tuples. In the tuples world, fold_stmt_inplace
2784 cannot fold a call that simplifies to a constant,
2785 because the GIMPLE_CALL must be replaced by a
2786 GIMPLE_ASSIGN, and there is no way to effect such a
2787 transformation in-place. We might want to consider
2788 using the more general fold_stmt here. */
2790 gimple_stmt_iterator gsi = gsi_for_stmt (use_stmt);
2791 fold_stmt_inplace (&gsi);
2794 /* Sometimes propagation can expose new operands to the
2795 renamer. */
2796 update_stmt (use_stmt);
2798 /* Dump details. */
2799 if (dump_file && (dump_flags & TDF_DETAILS))
2801 fprintf (dump_file, " Updated statement:");
2802 print_gimple_stmt (dump_file, use_stmt, 0, dump_flags);
2805 /* If we replaced a variable index with a constant, then
2806 we would need to update the invariant flag for ADDR_EXPRs. */
2807 if (gimple_assign_single_p (use_stmt)
2808 && TREE_CODE (gimple_assign_rhs1 (use_stmt)) == ADDR_EXPR)
2809 recompute_tree_invariant_for_addr_expr
2810 (gimple_assign_rhs1 (use_stmt));
2812 /* If we cleaned up EH information from the statement,
2813 mark its containing block as needing EH cleanups. */
2814 if (maybe_clean_or_replace_eh_stmt (use_stmt, use_stmt))
2816 bitmap_set_bit (need_eh_cleanup, gimple_bb (use_stmt)->index);
2817 if (dump_file && (dump_flags & TDF_DETAILS))
2818 fprintf (dump_file, " Flagged to clear EH edges.\n");
2821 /* Propagation may expose new trivial copy/constant propagation
2822 opportunities. */
2823 if (gimple_assign_single_p (use_stmt)
2824 && TREE_CODE (gimple_assign_lhs (use_stmt)) == SSA_NAME
2825 && (TREE_CODE (gimple_assign_rhs1 (use_stmt)) == SSA_NAME
2826 || is_gimple_min_invariant (gimple_assign_rhs1 (use_stmt))))
2828 tree result = get_lhs_or_phi_result (use_stmt);
2829 bitmap_set_bit (interesting_names, SSA_NAME_VERSION (result));
2832 /* Propagation into these nodes may make certain edges in
2833 the CFG unexecutable. We want to identify them as PHI nodes
2834 at the destination of those unexecutable edges may become
2835 degenerates. */
2836 else if (gimple_code (use_stmt) == GIMPLE_COND
2837 || gimple_code (use_stmt) == GIMPLE_SWITCH
2838 || gimple_code (use_stmt) == GIMPLE_GOTO)
2840 tree val;
2842 if (gimple_code (use_stmt) == GIMPLE_COND)
2843 val = fold_binary_loc (gimple_location (use_stmt),
2844 gimple_cond_code (use_stmt),
2845 boolean_type_node,
2846 gimple_cond_lhs (use_stmt),
2847 gimple_cond_rhs (use_stmt));
2848 else if (gimple_code (use_stmt) == GIMPLE_SWITCH)
2849 val = gimple_switch_index (use_stmt);
2850 else
2851 val = gimple_goto_dest (use_stmt);
2853 if (val && is_gimple_min_invariant (val))
2855 basic_block bb = gimple_bb (use_stmt);
2856 edge te = find_taken_edge (bb, val);
2857 edge_iterator ei;
2858 edge e;
2859 gimple_stmt_iterator gsi, psi;
2861 /* Remove all outgoing edges except TE. */
2862 for (ei = ei_start (bb->succs); (e = ei_safe_edge (ei));)
2864 if (e != te)
2866 /* Mark all the PHI nodes at the destination of
2867 the unexecutable edge as interesting. */
2868 for (psi = gsi_start_phis (e->dest);
2869 !gsi_end_p (psi);
2870 gsi_next (&psi))
2872 gimple phi = gsi_stmt (psi);
2874 tree result = gimple_phi_result (phi);
2875 int version = SSA_NAME_VERSION (result);
2877 bitmap_set_bit (interesting_names, version);
2880 te->probability += e->probability;
2882 te->count += e->count;
2883 remove_edge (e);
2884 cfg_altered = true;
2886 else
2887 ei_next (&ei);
2890 gsi = gsi_last_bb (gimple_bb (use_stmt));
2891 gsi_remove (&gsi, true);
2893 /* And fixup the flags on the single remaining edge. */
2894 te->flags &= ~(EDGE_TRUE_VALUE | EDGE_FALSE_VALUE);
2895 te->flags &= ~EDGE_ABNORMAL;
2896 te->flags |= EDGE_FALLTHRU;
2897 if (te->probability > REG_BR_PROB_BASE)
2898 te->probability = REG_BR_PROB_BASE;
2903 /* Ensure there is nothing else to do. */
2904 gcc_assert (!all || has_zero_uses (lhs));
2906 /* If we were able to propagate away all uses of LHS, then
2907 we can remove STMT. */
2908 if (all)
2909 remove_stmt_or_phi (stmt);
2913 /* STMT is either a PHI node (potentially a degenerate PHI node) or
2914 a statement that is a trivial copy or constant initialization.
2916 Attempt to eliminate T by propagating its RHS into all uses of
2917 its LHS. This may in turn set new bits in INTERESTING_NAMES
2918 for nodes we want to revisit later.
2920 All exit paths should clear INTERESTING_NAMES for the result
2921 of STMT. */
2923 static void
2924 eliminate_const_or_copy (gimple stmt, bitmap interesting_names)
2926 tree lhs = get_lhs_or_phi_result (stmt);
2927 tree rhs;
2928 int version = SSA_NAME_VERSION (lhs);
2930 /* If the LHS of this statement or PHI has no uses, then we can
2931 just eliminate it. This can occur if, for example, the PHI
2932 was created by block duplication due to threading and its only
2933 use was in the conditional at the end of the block which was
2934 deleted. */
2935 if (has_zero_uses (lhs))
2937 bitmap_clear_bit (interesting_names, version);
2938 remove_stmt_or_phi (stmt);
2939 return;
2942 /* Get the RHS of the assignment or PHI node if the PHI is a
2943 degenerate. */
2944 rhs = get_rhs_or_phi_arg (stmt);
2945 if (!rhs)
2947 bitmap_clear_bit (interesting_names, version);
2948 return;
2951 if (!virtual_operand_p (lhs))
2952 propagate_rhs_into_lhs (stmt, lhs, rhs, interesting_names);
2953 else
2955 gimple use_stmt;
2956 imm_use_iterator iter;
2957 use_operand_p use_p;
2958 /* For virtual operands we have to propagate into all uses as
2959 otherwise we will create overlapping life-ranges. */
2960 FOR_EACH_IMM_USE_STMT (use_stmt, iter, lhs)
2961 FOR_EACH_IMM_USE_ON_STMT (use_p, iter)
2962 SET_USE (use_p, rhs);
2963 if (SSA_NAME_OCCURS_IN_ABNORMAL_PHI (lhs))
2964 SSA_NAME_OCCURS_IN_ABNORMAL_PHI (rhs) = 1;
2965 remove_stmt_or_phi (stmt);
2968 /* Note that STMT may well have been deleted by now, so do
2969 not access it, instead use the saved version # to clear
2970 T's entry in the worklist. */
2971 bitmap_clear_bit (interesting_names, version);
2974 /* The first phase in degenerate PHI elimination.
2976 Eliminate the degenerate PHIs in BB, then recurse on the
2977 dominator children of BB. */
2979 static void
2980 eliminate_degenerate_phis_1 (basic_block bb, bitmap interesting_names)
2982 gimple_stmt_iterator gsi;
2983 basic_block son;
2985 for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi))
2987 gimple phi = gsi_stmt (gsi);
2989 eliminate_const_or_copy (phi, interesting_names);
2992 /* Recurse into the dominator children of BB. */
2993 for (son = first_dom_son (CDI_DOMINATORS, bb);
2994 son;
2995 son = next_dom_son (CDI_DOMINATORS, son))
2996 eliminate_degenerate_phis_1 (son, interesting_names);
3000 /* A very simple pass to eliminate degenerate PHI nodes from the
3001 IL. This is meant to be fast enough to be able to be run several
3002 times in the optimization pipeline.
3004 Certain optimizations, particularly those which duplicate blocks
3005 or remove edges from the CFG can create or expose PHIs which are
3006 trivial copies or constant initializations.
3008 While we could pick up these optimizations in DOM or with the
3009 combination of copy-prop and CCP, those solutions are far too
3010 heavy-weight for our needs.
3012 This implementation has two phases so that we can efficiently
3013 eliminate the first order degenerate PHIs and second order
3014 degenerate PHIs.
3016 The first phase performs a dominator walk to identify and eliminate
3017 the vast majority of the degenerate PHIs. When a degenerate PHI
3018 is identified and eliminated any affected statements or PHIs
3019 are put on a worklist.
3021 The second phase eliminates degenerate PHIs and trivial copies
3022 or constant initializations using the worklist. This is how we
3023 pick up the secondary optimization opportunities with minimal
3024 cost. */
3026 namespace {
3028 const pass_data pass_data_phi_only_cprop =
3030 GIMPLE_PASS, /* type */
3031 "phicprop", /* name */
3032 OPTGROUP_NONE, /* optinfo_flags */
3033 true, /* has_execute */
3034 TV_TREE_PHI_CPROP, /* tv_id */
3035 ( PROP_cfg | PROP_ssa ), /* properties_required */
3036 0, /* properties_provided */
3037 0, /* properties_destroyed */
3038 0, /* todo_flags_start */
3039 ( TODO_cleanup_cfg | TODO_update_ssa ), /* todo_flags_finish */
3042 class pass_phi_only_cprop : public gimple_opt_pass
3044 public:
3045 pass_phi_only_cprop (gcc::context *ctxt)
3046 : gimple_opt_pass (pass_data_phi_only_cprop, ctxt)
3049 /* opt_pass methods: */
3050 opt_pass * clone () { return new pass_phi_only_cprop (m_ctxt); }
3051 virtual bool gate (function *) { return flag_tree_dom != 0; }
3052 virtual unsigned int execute (function *);
3054 }; // class pass_phi_only_cprop
3056 unsigned int
3057 pass_phi_only_cprop::execute (function *fun)
3059 bitmap interesting_names;
3060 bitmap interesting_names1;
3062 /* Bitmap of blocks which need EH information updated. We can not
3063 update it on-the-fly as doing so invalidates the dominator tree. */
3064 need_eh_cleanup = BITMAP_ALLOC (NULL);
3066 /* INTERESTING_NAMES is effectively our worklist, indexed by
3067 SSA_NAME_VERSION.
3069 A set bit indicates that the statement or PHI node which
3070 defines the SSA_NAME should be (re)examined to determine if
3071 it has become a degenerate PHI or trivial const/copy propagation
3072 opportunity.
3074 Experiments have show we generally get better compilation
3075 time behavior with bitmaps rather than sbitmaps. */
3076 interesting_names = BITMAP_ALLOC (NULL);
3077 interesting_names1 = BITMAP_ALLOC (NULL);
3079 calculate_dominance_info (CDI_DOMINATORS);
3080 cfg_altered = false;
3082 /* First phase. Eliminate degenerate PHIs via a dominator
3083 walk of the CFG.
3085 Experiments have indicated that we generally get better
3086 compile-time behavior by visiting blocks in the first
3087 phase in dominator order. Presumably this is because walking
3088 in dominator order leaves fewer PHIs for later examination
3089 by the worklist phase. */
3090 eliminate_degenerate_phis_1 (ENTRY_BLOCK_PTR_FOR_FN (fun),
3091 interesting_names);
3093 /* Second phase. Eliminate second order degenerate PHIs as well
3094 as trivial copies or constant initializations identified by
3095 the first phase or this phase. Basically we keep iterating
3096 until our set of INTERESTING_NAMEs is empty. */
3097 while (!bitmap_empty_p (interesting_names))
3099 unsigned int i;
3100 bitmap_iterator bi;
3102 /* EXECUTE_IF_SET_IN_BITMAP does not like its bitmap
3103 changed during the loop. Copy it to another bitmap and
3104 use that. */
3105 bitmap_copy (interesting_names1, interesting_names);
3107 EXECUTE_IF_SET_IN_BITMAP (interesting_names1, 0, i, bi)
3109 tree name = ssa_name (i);
3111 /* Ignore SSA_NAMEs that have been released because
3112 their defining statement was deleted (unreachable). */
3113 if (name)
3114 eliminate_const_or_copy (SSA_NAME_DEF_STMT (ssa_name (i)),
3115 interesting_names);
3119 if (cfg_altered)
3121 free_dominance_info (CDI_DOMINATORS);
3122 /* If we changed the CFG schedule loops for fixup by cfgcleanup. */
3123 if (current_loops)
3124 loops_state_set (LOOPS_NEED_FIXUP);
3127 /* Propagation of const and copies may make some EH edges dead. Purge
3128 such edges from the CFG as needed. */
3129 if (!bitmap_empty_p (need_eh_cleanup))
3131 gimple_purge_all_dead_eh_edges (need_eh_cleanup);
3132 BITMAP_FREE (need_eh_cleanup);
3135 BITMAP_FREE (interesting_names);
3136 BITMAP_FREE (interesting_names1);
3137 return 0;
3140 } // anon namespace
3142 gimple_opt_pass *
3143 make_pass_phi_only_cprop (gcc::context *ctxt)
3145 return new pass_phi_only_cprop (ctxt);