simplify-rtx.c (simplify_rtx): Use simplify_subreg rather than simplify_gen_subreg.
[official-gcc.git] / gcc / tree-ssa-pre.c
blob80986eb086041880f11ab88744ba8db45f7fb598
1 /* SSA-PRE for trees.
2 Copyright (C) 2001, 2002, 2003, 2004, 2005 Free Software Foundation, Inc.
3 Contributed by Daniel Berlin <dan@dberlin.org> and Steven Bosscher
4 <stevenb@suse.de>
6 This file is part of GCC.
8 GCC is free software; you can redistribute it and/or modify
9 it under the terms of the GNU General Public License as published by
10 the Free Software Foundation; either version 2, or (at your option)
11 any later version.
13 GCC is distributed in the hope that it will be useful,
14 but WITHOUT ANY WARRANTY; without even the implied warranty of
15 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 GNU General Public License for more details.
18 You should have received a copy of the GNU General Public License
19 along with GCC; see the file COPYING. If not, write to
20 the Free Software Foundation, 51 Franklin Street, Fifth Floor,
21 Boston, MA 02110-1301, USA. */
23 #include "config.h"
24 #include "system.h"
25 #include "coretypes.h"
26 #include "tm.h"
27 #include "ggc.h"
28 #include "tree.h"
29 #include "basic-block.h"
30 #include "diagnostic.h"
31 #include "tree-inline.h"
32 #include "tree-flow.h"
33 #include "tree-gimple.h"
34 #include "tree-dump.h"
35 #include "timevar.h"
36 #include "fibheap.h"
37 #include "hashtab.h"
38 #include "tree-iterator.h"
39 #include "real.h"
40 #include "alloc-pool.h"
41 #include "tree-pass.h"
42 #include "flags.h"
43 #include "bitmap.h"
44 #include "langhooks.h"
45 #include "cfgloop.h"
47 /* TODO:
49 1. Avail sets can be shared by making an avail_find_leader that
50 walks up the dominator tree and looks in those avail sets.
51 This might affect code optimality, it's unclear right now.
52 2. Strength reduction can be performed by anticipating expressions
53 we can repair later on.
54 3. We can do back-substitution or smarter value numbering to catch
55 commutative expressions split up over multiple statements.
56 4. ANTIC_SAFE_LOADS could be a lot smarter than it is now.
57 Right now, it is simply calculating loads that occur before
58 any store in a block, instead of loads that occur before
59 stores that affect them. This is relatively more expensive, and
60 it's not clear how much more it will buy us.
63 /* For ease of terminology, "expression node" in the below refers to
64 every expression node but MODIFY_EXPR, because MODIFY_EXPR's represent
65 the actual statement containing the expressions we care about, and
66 we cache the value number by putting it in the expression. */
68 /* Basic algorithm
70 First we walk the statements to generate the AVAIL sets, the
71 EXP_GEN sets, and the tmp_gen sets. EXP_GEN sets represent the
72 generation of values/expressions by a given block. We use them
73 when computing the ANTIC sets. The AVAIL sets consist of
74 SSA_NAME's that represent values, so we know what values are
75 available in what blocks. AVAIL is a forward dataflow problem. In
76 SSA, values are never killed, so we don't need a kill set, or a
77 fixpoint iteration, in order to calculate the AVAIL sets. In
78 traditional parlance, AVAIL sets tell us the downsafety of the
79 expressions/values.
81 Next, we generate the ANTIC sets. These sets represent the
82 anticipatable expressions. ANTIC is a backwards dataflow
83 problem. An expression is anticipatable in a given block if it could
84 be generated in that block. This means that if we had to perform
85 an insertion in that block, of the value of that expression, we
86 could. Calculating the ANTIC sets requires phi translation of
87 expressions, because the flow goes backwards through phis. We must
88 iterate to a fixpoint of the ANTIC sets, because we have a kill
89 set. Even in SSA form, values are not live over the entire
90 function, only from their definition point onwards. So we have to
91 remove values from the ANTIC set once we go past the definition
92 point of the leaders that make them up.
93 compute_antic/compute_antic_aux performs this computation.
95 Third, we perform insertions to make partially redundant
96 expressions fully redundant.
98 An expression is partially redundant (excluding partial
99 anticipation) if:
101 1. It is AVAIL in some, but not all, of the predecessors of a
102 given block.
103 2. It is ANTIC in all the predecessors.
105 In order to make it fully redundant, we insert the expression into
106 the predecessors where it is not available, but is ANTIC.
108 For the partial anticipation case, we only perform insertion if it
109 is partially anticipated in some block, and fully available in all
110 of the predecessors.
112 insert/insert_aux/do_regular_insertion/do_partial_partial_insertion
113 performs these steps.
115 Fourth, we eliminate fully redundant expressions.
116 This is a simple statement walk that replaces redundant
117 calculations with the now available values. */
119 /* Representations of value numbers:
121 Value numbers are represented using the "value handle" approach.
122 This means that each SSA_NAME (and for other reasons to be
123 disclosed in a moment, expression nodes) has a value handle that
124 can be retrieved through get_value_handle. This value handle *is*
125 the value number of the SSA_NAME. You can pointer compare the
126 value handles for equivalence purposes.
128 For debugging reasons, the value handle is internally more than
129 just a number, it is a VALUE_HANDLE named "VH.x", where x is a
130 unique number for each value number in use. This allows
131 expressions with SSA_NAMES replaced by value handles to still be
132 pretty printed in a sane way. They simply print as "VH.3 *
133 VH.5", etc.
135 Expression nodes have value handles associated with them as a
136 cache. Otherwise, we'd have to look them up again in the hash
137 table This makes significant difference (factor of two or more) on
138 some test cases. They can be thrown away after the pass is
139 finished. */
141 /* Representation of expressions on value numbers:
143 In some portions of this code, you will notice we allocate "fake"
144 analogues to the expression we are value numbering, and replace the
145 operands with the values of the expression. Since we work on
146 values, and not just names, we canonicalize expressions to value
147 expressions for use in the ANTIC sets, the EXP_GEN set, etc.
149 This is theoretically unnecessary, it just saves a bunch of
150 repeated get_value_handle and find_leader calls in the remainder of
151 the code, trading off temporary memory usage for speed. The tree
152 nodes aren't actually creating more garbage, since they are
153 allocated in a special pools which are thrown away at the end of
154 this pass.
156 All of this also means that if you print the EXP_GEN or ANTIC sets,
157 you will see "VH.5 + VH.7" in the set, instead of "a_55 +
158 b_66" or something. The only thing that actually cares about
159 seeing the value leaders is phi translation, and it needs to be
160 able to find the leader for a value in an arbitrary block, so this
161 "value expression" form is perfect for it (otherwise you'd do
162 get_value_handle->find_leader->translate->get_value_handle->find_leader).*/
165 /* Representation of sets:
167 There are currently two types of sets used, hopefully to be unified soon.
168 The AVAIL sets do not need to be sorted in any particular order,
169 and thus, are simply represented as two bitmaps, one that keeps
170 track of values present in the set, and one that keeps track of
171 expressions present in the set.
173 The other sets are represented as doubly linked lists kept in topological
174 order, with an optional supporting bitmap of values present in the
175 set. The sets represent values, and the elements can be values or
176 expressions. The elements can appear in different sets, but each
177 element can only appear once in each set.
179 Since each node in the set represents a value, we also want to be
180 able to map expression, set pairs to something that tells us
181 whether the value is present is a set. We use a per-set bitmap for
182 that. The value handles also point to a linked list of the
183 expressions they represent via a tree annotation. This is mainly
184 useful only for debugging, since we don't do identity lookups. */
187 /* Next global expression id number. */
188 static unsigned int next_expression_id;
190 /* Mapping from expression to id number we can use in bitmap sets. */
191 static VEC(tree, heap) *expressions;
193 /* Allocate an expression id for EXPR. */
195 static inline unsigned int
196 alloc_expression_id (tree expr)
198 tree_ann_common_t ann;
200 ann = get_tree_common_ann (expr);
202 /* Make sure we won't overflow. */
203 gcc_assert (next_expression_id + 1 > next_expression_id);
205 ann->aux = XNEW (unsigned int);
206 * ((unsigned int *)ann->aux) = next_expression_id++;
207 VEC_safe_push (tree, heap, expressions, expr);
208 return next_expression_id - 1;
211 /* Return the expression id for tree EXPR. */
213 static inline unsigned int
214 get_expression_id (tree expr)
216 tree_ann_common_t ann = tree_common_ann (expr);
217 gcc_assert (ann);
218 gcc_assert (ann->aux);
220 return *((unsigned int *)ann->aux);
223 /* Return the existing expression id for EXPR, or create one if one
224 does not exist yet. */
226 static inline unsigned int
227 get_or_alloc_expression_id (tree expr)
229 tree_ann_common_t ann = tree_common_ann (expr);
231 if (ann == NULL || !ann->aux)
232 return alloc_expression_id (expr);
234 return get_expression_id (expr);
237 /* Return the expression that has expression id ID */
239 static inline tree
240 expression_for_id (unsigned int id)
242 return VEC_index (tree, expressions, id);
245 /* Free the expression id field in all of our expressions,
246 and then destroy the expressions array. */
248 static void
249 clear_expression_ids (void)
251 int i;
252 tree expr;
254 for (i = 0; VEC_iterate (tree, expressions, i, expr); i++)
256 free (tree_common_ann (expr)->aux);
257 tree_common_ann (expr)->aux = NULL;
259 VEC_free (tree, heap, expressions);
262 static bool in_fre = false;
264 /* An unordered bitmap set. One bitmap tracks values, the other,
265 expressions. */
266 typedef struct bitmap_set
268 bitmap expressions;
269 bitmap values;
270 } *bitmap_set_t;
272 #define FOR_EACH_EXPR_ID_IN_SET(set, id, bi) \
273 EXECUTE_IF_SET_IN_BITMAP(set->expressions, 0, id, bi)
275 /* Sets that we need to keep track of. */
276 typedef struct bb_bitmap_sets
278 /* The EXP_GEN set, which represents expressions/values generated in
279 a basic block. */
280 bitmap_set_t exp_gen;
282 /* The PHI_GEN set, which represents PHI results generated in a
283 basic block. */
284 bitmap_set_t phi_gen;
286 /* The TMP_GEN set, which represents results/temporaries generated
287 in a basic block. IE the LHS of an expression. */
288 bitmap_set_t tmp_gen;
290 /* The AVAIL_OUT set, which represents which values are available in
291 a given basic block. */
292 bitmap_set_t avail_out;
294 /* The ANTIC_IN set, which represents which values are anticipatable
295 in a given basic block. */
296 bitmap_set_t antic_in;
298 /* The PA_IN set, which represents which values are
299 partially anticipatable in a given basic block. */
300 bitmap_set_t pa_in;
302 /* The NEW_SETS set, which is used during insertion to augment the
303 AVAIL_OUT set of blocks with the new insertions performed during
304 the current iteration. */
305 bitmap_set_t new_sets;
307 /* The RVUSE sets, which are used during ANTIC computation to ensure
308 that we don't mark loads ANTIC once they have died. */
309 bitmap rvuse_in;
310 bitmap rvuse_out;
311 bitmap rvuse_gen;
312 bitmap rvuse_kill;
314 /* For actually occurring loads, as long as they occur before all the
315 other stores in the block, we know they are antic at the top of
316 the block, regardless of RVUSE_KILL. */
317 bitmap_set_t antic_safe_loads;
319 /* True if we have visited this block during ANTIC calculation. */
320 unsigned int visited:1;
322 /* True we have deferred processing this block during ANTIC
323 calculation until its successor is processed. */
324 unsigned int deferred : 1;
325 } *bb_value_sets_t;
327 #define EXP_GEN(BB) ((bb_value_sets_t) ((BB)->aux))->exp_gen
328 #define PHI_GEN(BB) ((bb_value_sets_t) ((BB)->aux))->phi_gen
329 #define TMP_GEN(BB) ((bb_value_sets_t) ((BB)->aux))->tmp_gen
330 #define AVAIL_OUT(BB) ((bb_value_sets_t) ((BB)->aux))->avail_out
331 #define ANTIC_IN(BB) ((bb_value_sets_t) ((BB)->aux))->antic_in
332 #define PA_IN(BB) ((bb_value_sets_t) ((BB)->aux))->pa_in
333 #define RVUSE_IN(BB) ((bb_value_sets_t) ((BB)->aux))->rvuse_in
334 #define RVUSE_GEN(BB) ((bb_value_sets_t) ((BB)->aux))->rvuse_gen
335 #define RVUSE_KILL(BB) ((bb_value_sets_t) ((BB)->aux))->rvuse_kill
336 #define RVUSE_OUT(BB) ((bb_value_sets_t) ((BB)->aux))->rvuse_out
337 #define NEW_SETS(BB) ((bb_value_sets_t) ((BB)->aux))->new_sets
338 #define ANTIC_SAFE_LOADS(BB) ((bb_value_sets_t) ((BB)->aux))->antic_safe_loads
339 #define BB_VISITED(BB) ((bb_value_sets_t) ((BB)->aux))->visited
340 #define BB_DEFERRED(BB) ((bb_value_sets_t) ((BB)->aux))->deferred
342 /* Maximal set of values, used to initialize the ANTIC problem, which
343 is an intersection problem. */
344 static bitmap_set_t maximal_set;
346 /* Basic block list in postorder. */
347 static int *postorder;
349 /* This structure is used to keep track of statistics on what
350 optimization PRE was able to perform. */
351 static struct
353 /* The number of RHS computations eliminated by PRE. */
354 int eliminations;
356 /* The number of new expressions/temporaries generated by PRE. */
357 int insertions;
359 /* The number of inserts found due to partial anticipation */
360 int pa_insert;
362 /* The number of new PHI nodes added by PRE. */
363 int phis;
365 /* The number of values found constant. */
366 int constified;
368 } pre_stats;
370 static bool do_partial_partial;
371 static tree bitmap_find_leader (bitmap_set_t, tree);
372 static void bitmap_value_insert_into_set (bitmap_set_t, tree);
373 static void bitmap_value_replace_in_set (bitmap_set_t, tree);
374 static void bitmap_set_copy (bitmap_set_t, bitmap_set_t);
375 static bool bitmap_set_contains_value (bitmap_set_t, tree);
376 static void bitmap_insert_into_set (bitmap_set_t, tree);
377 static bitmap_set_t bitmap_set_new (void);
378 static bool is_undefined_value (tree);
379 static tree create_expression_by_pieces (basic_block, tree, tree);
380 static tree find_or_generate_expression (basic_block, tree, tree);
382 /* We can add and remove elements and entries to and from sets
383 and hash tables, so we use alloc pools for them. */
385 static alloc_pool bitmap_set_pool;
386 static alloc_pool binary_node_pool;
387 static alloc_pool unary_node_pool;
388 static alloc_pool reference_node_pool;
389 static alloc_pool comparison_node_pool;
390 static alloc_pool expression_node_pool;
391 static alloc_pool list_node_pool;
392 static alloc_pool modify_expr_node_pool;
393 static bitmap_obstack grand_bitmap_obstack;
395 /* To avoid adding 300 temporary variables when we only need one, we
396 only create one temporary variable, on demand, and build ssa names
397 off that. We do have to change the variable if the types don't
398 match the current variable's type. */
399 static tree pretemp;
400 static tree storetemp;
401 static tree mergephitemp;
402 static tree prephitemp;
404 /* Set of blocks with statements that have had its EH information
405 cleaned up. */
406 static bitmap need_eh_cleanup;
408 /* The phi_translate_table caches phi translations for a given
409 expression and predecessor. */
411 static htab_t phi_translate_table;
413 /* A three tuple {e, pred, v} used to cache phi translations in the
414 phi_translate_table. */
416 typedef struct expr_pred_trans_d
418 /* The expression. */
419 tree e;
421 /* The predecessor block along which we translated the expression. */
422 basic_block pred;
424 /* vuses associated with the expression. */
425 VEC (tree, gc) *vuses;
427 /* The value that resulted from the translation. */
428 tree v;
430 /* The hashcode for the expression, pred pair. This is cached for
431 speed reasons. */
432 hashval_t hashcode;
433 } *expr_pred_trans_t;
435 /* Return the hash value for a phi translation table entry. */
437 static hashval_t
438 expr_pred_trans_hash (const void *p)
440 const expr_pred_trans_t ve = (expr_pred_trans_t) p;
441 return ve->hashcode;
444 /* Return true if two phi translation table entries are the same.
445 P1 and P2 should point to the expr_pred_trans_t's to be compared.*/
447 static int
448 expr_pred_trans_eq (const void *p1, const void *p2)
450 const expr_pred_trans_t ve1 = (expr_pred_trans_t) p1;
451 const expr_pred_trans_t ve2 = (expr_pred_trans_t) p2;
452 basic_block b1 = ve1->pred;
453 basic_block b2 = ve2->pred;
454 int i;
455 tree vuse1;
457 /* If they are not translations for the same basic block, they can't
458 be equal. */
459 if (b1 != b2)
460 return false;
463 /* If they are for the same basic block, determine if the
464 expressions are equal. */
465 if (!expressions_equal_p (ve1->e, ve2->e))
466 return false;
468 /* Make sure the vuses are equivalent. */
469 if (ve1->vuses == ve2->vuses)
470 return true;
472 if (VEC_length (tree, ve1->vuses) != VEC_length (tree, ve2->vuses))
473 return false;
475 for (i = 0; VEC_iterate (tree, ve1->vuses, i, vuse1); i++)
477 if (VEC_index (tree, ve2->vuses, i) != vuse1)
478 return false;
481 return true;
484 /* Search in the phi translation table for the translation of
485 expression E in basic block PRED with vuses VUSES.
486 Return the translated value, if found, NULL otherwise. */
488 static inline tree
489 phi_trans_lookup (tree e, basic_block pred, VEC (tree, gc) *vuses)
491 void **slot;
492 struct expr_pred_trans_d ept;
494 ept.e = e;
495 ept.pred = pred;
496 ept.vuses = vuses;
497 ept.hashcode = vn_compute (e, (unsigned long) pred);
498 slot = htab_find_slot_with_hash (phi_translate_table, &ept, ept.hashcode,
499 NO_INSERT);
500 if (!slot)
501 return NULL;
502 else
503 return ((expr_pred_trans_t) *slot)->v;
507 /* Add the tuple mapping from {expression E, basic block PRED, vuses VUSES} to
508 value V, to the phi translation table. */
510 static inline void
511 phi_trans_add (tree e, tree v, basic_block pred, VEC (tree, gc) *vuses)
513 void **slot;
514 expr_pred_trans_t new_pair = XNEW (struct expr_pred_trans_d);
515 new_pair->e = e;
516 new_pair->pred = pred;
517 new_pair->vuses = vuses;
518 new_pair->v = v;
519 new_pair->hashcode = vn_compute (e, (unsigned long) pred);
520 slot = htab_find_slot_with_hash (phi_translate_table, new_pair,
521 new_pair->hashcode, INSERT);
522 if (*slot)
523 free (*slot);
524 *slot = (void *) new_pair;
528 /* Return true if V is a value expression that represents itself.
529 In our world, this is *only* non-value handles. */
531 static inline bool
532 constant_expr_p (tree v)
534 return TREE_CODE (v) != VALUE_HANDLE && is_gimple_min_invariant (v);
535 /* return TREE_CODE (v) != VALUE_HANDLE; */
538 /* Add expression E to the expression set of value V. */
540 void
541 add_to_value (tree v, tree e)
543 /* Constants have no expression sets. */
544 if (constant_expr_p (v))
545 return;
547 if (VALUE_HANDLE_EXPR_SET (v) == NULL)
548 VALUE_HANDLE_EXPR_SET (v) = bitmap_set_new ();
550 bitmap_insert_into_set (VALUE_HANDLE_EXPR_SET (v), e);
553 /* Create a new bitmap set and return it. */
555 static bitmap_set_t
556 bitmap_set_new (void)
558 bitmap_set_t ret = (bitmap_set_t) pool_alloc (bitmap_set_pool);
559 ret->expressions = BITMAP_ALLOC (&grand_bitmap_obstack);
560 ret->values = BITMAP_ALLOC (&grand_bitmap_obstack);
561 return ret;
564 /* Remove an expression EXPR from a bitmapped set. */
566 static void
567 bitmap_remove_from_set (bitmap_set_t set, tree expr)
569 tree val = get_value_handle (expr);
571 gcc_assert (val);
572 if (!constant_expr_p (val))
574 bitmap_clear_bit (set->values, VALUE_HANDLE_ID (val));
575 bitmap_clear_bit (set->expressions, get_expression_id (expr));
579 /* Insert an expression EXPR into a bitmapped set. */
581 static void
582 bitmap_insert_into_set (bitmap_set_t set, tree expr)
584 tree val = get_value_handle (expr);
586 gcc_assert (val);
587 if (!constant_expr_p (val))
589 bitmap_set_bit (set->values, VALUE_HANDLE_ID (val));
590 bitmap_set_bit (set->expressions, get_or_alloc_expression_id (expr));
594 /* Copy a bitmapped set ORIG, into bitmapped set DEST. */
596 static void
597 bitmap_set_copy (bitmap_set_t dest, bitmap_set_t orig)
599 bitmap_copy (dest->expressions, orig->expressions);
600 bitmap_copy (dest->values, orig->values);
604 /* Free memory used up by SET. */
605 static void
606 bitmap_set_free (bitmap_set_t set)
608 BITMAP_FREE (set->expressions);
609 BITMAP_FREE (set->values);
613 /* A comparison function for use in qsort to top sort a bitmap set. Simply
614 subtracts value handle ids, since they are created in topo-order. */
616 static int
617 vh_compare (const void *pa, const void *pb)
619 const tree vha = get_value_handle (*((const tree *)pa));
620 const tree vhb = get_value_handle (*((const tree *)pb));
622 /* This can happen when we constify things. */
623 if (constant_expr_p (vha))
625 if (constant_expr_p (vhb))
626 return -1;
627 return -1;
629 else if (constant_expr_p (vhb))
630 return 1;
631 return VALUE_HANDLE_ID (vha) - VALUE_HANDLE_ID (vhb);
634 /* Generate an topological-ordered array of bitmap set SET. */
636 static VEC(tree, heap) *
637 sorted_array_from_bitmap_set (bitmap_set_t set)
639 unsigned int i;
640 bitmap_iterator bi;
641 VEC(tree, heap) *result = NULL;
643 FOR_EACH_EXPR_ID_IN_SET (set, i, bi)
644 VEC_safe_push (tree, heap, result, expression_for_id (i));
646 qsort (VEC_address (tree, result), VEC_length (tree, result),
647 sizeof (tree), vh_compare);
649 return result;
652 /* Perform bitmapped set operation DEST &= ORIG. */
654 static void
655 bitmap_set_and (bitmap_set_t dest, bitmap_set_t orig)
657 bitmap_iterator bi;
658 unsigned int i;
660 if (dest != orig)
662 bitmap temp = BITMAP_ALLOC (&grand_bitmap_obstack);
664 bitmap_and_into (dest->values, orig->values);
666 bitmap_copy (temp, dest->expressions);
667 EXECUTE_IF_SET_IN_BITMAP (temp, 0, i, bi)
669 tree expr = expression_for_id (i);
670 tree val = get_value_handle (expr);
671 if (!bitmap_bit_p (dest->values, VALUE_HANDLE_ID (val)))
672 bitmap_clear_bit (dest->expressions, i);
674 BITMAP_FREE (temp);
678 /* Subtract all values and expressions contained in ORIG from DEST. */
680 static bitmap_set_t
681 bitmap_set_subtract (bitmap_set_t dest, bitmap_set_t orig)
683 bitmap_set_t result = bitmap_set_new ();
684 bitmap_iterator bi;
685 unsigned int i;
687 bitmap_and_compl (result->expressions, dest->expressions,
688 orig->expressions);
690 FOR_EACH_EXPR_ID_IN_SET (result, i, bi)
692 tree expr = expression_for_id (i);
693 tree val = get_value_handle (expr);
694 bitmap_set_bit (result->values, VALUE_HANDLE_ID (val));
697 return result;
700 /* Subtract all the values in bitmap set B from bitmap set A. */
702 static void
703 bitmap_set_subtract_values (bitmap_set_t a, bitmap_set_t b)
705 unsigned int i;
706 bitmap_iterator bi;
707 bitmap temp = BITMAP_ALLOC (&grand_bitmap_obstack);
709 bitmap_copy (temp, a->expressions);
710 EXECUTE_IF_SET_IN_BITMAP (temp, 0, i, bi)
712 tree expr = expression_for_id (i);
713 if (bitmap_set_contains_value (b, get_value_handle (expr)))
714 bitmap_remove_from_set (a, expr);
716 BITMAP_FREE (temp);
720 /* Return true if bitmapped set SET contains the value VAL. */
722 static bool
723 bitmap_set_contains_value (bitmap_set_t set, tree val)
725 if (constant_expr_p (val))
726 return true;
728 if (!set || bitmap_empty_p (set->expressions))
729 return false;
731 return bitmap_bit_p (set->values, VALUE_HANDLE_ID (val));
734 static inline bool
735 bitmap_set_contains_expr (bitmap_set_t set, tree expr)
737 return bitmap_bit_p (set->expressions, get_expression_id (expr));
740 /* Replace an instance of value LOOKFOR with expression EXPR in SET. */
742 static void
743 bitmap_set_replace_value (bitmap_set_t set, tree lookfor, tree expr)
745 bitmap_set_t exprset;
746 unsigned int i;
747 bitmap_iterator bi;
749 if (constant_expr_p (lookfor))
750 return;
752 if (!bitmap_set_contains_value (set, lookfor))
753 return;
755 /* The number of expressions having a given value is usually
756 significantly less than the total number of expressions in SET.
757 Thus, rather than check, for each expression in SET, whether it
758 has the value LOOKFOR, we walk the reverse mapping that tells us
759 what expressions have a given value, and see if any of those
760 expressions are in our set. For large testcases, this is about
761 5-10x faster than walking the bitmap. If this is somehow a
762 significant lose for some cases, we can choose which set to walk
763 based on the set size. */
764 exprset = VALUE_HANDLE_EXPR_SET (lookfor);
765 FOR_EACH_EXPR_ID_IN_SET (exprset, i, bi)
767 if (bitmap_bit_p (set->expressions, i))
769 bitmap_clear_bit (set->expressions, i);
770 bitmap_set_bit (set->expressions, get_expression_id (expr));
771 return;
776 /* Return true if two bitmap sets are equal. */
778 static bool
779 bitmap_set_equal (bitmap_set_t a, bitmap_set_t b)
781 return bitmap_equal_p (a->values, b->values);
784 /* Replace an instance of EXPR's VALUE with EXPR in SET if it exists,
785 and add it otherwise. */
787 static void
788 bitmap_value_replace_in_set (bitmap_set_t set, tree expr)
790 tree val = get_value_handle (expr);
792 if (bitmap_set_contains_value (set, val))
793 bitmap_set_replace_value (set, val, expr);
794 else
795 bitmap_insert_into_set (set, expr);
798 /* Insert EXPR into SET if EXPR's value is not already present in
799 SET. */
801 static void
802 bitmap_value_insert_into_set (bitmap_set_t set, tree expr)
804 tree val = get_value_handle (expr);
806 if (constant_expr_p (val))
807 return;
809 if (!bitmap_set_contains_value (set, val))
810 bitmap_insert_into_set (set, expr);
813 /* Print out SET to OUTFILE. */
815 static void
816 print_bitmap_set (FILE *outfile, bitmap_set_t set,
817 const char *setname, int blockindex)
819 fprintf (outfile, "%s[%d] := { ", setname, blockindex);
820 if (set)
822 bool first = true;
823 unsigned i;
824 bitmap_iterator bi;
826 FOR_EACH_EXPR_ID_IN_SET (set, i, bi)
828 tree expr = expression_for_id (i);
830 if (!first)
831 fprintf (outfile, ", ");
832 first = false;
833 print_generic_expr (outfile, expr, 0);
835 fprintf (outfile, " (");
836 print_generic_expr (outfile, get_value_handle (expr), 0);
837 fprintf (outfile, ") ");
840 fprintf (outfile, " }\n");
843 void debug_bitmap_set (bitmap_set_t);
845 void
846 debug_bitmap_set (bitmap_set_t set)
848 print_bitmap_set (stderr, set, "debug", 0);
851 /* Print out the expressions that have VAL to OUTFILE. */
853 void
854 print_value_expressions (FILE *outfile, tree val)
856 if (VALUE_HANDLE_EXPR_SET (val))
858 char s[10];
859 sprintf (s, "VH.%04d", VALUE_HANDLE_ID (val));
860 print_bitmap_set (outfile, VALUE_HANDLE_EXPR_SET (val), s, 0);
865 void
866 debug_value_expressions (tree val)
868 print_value_expressions (stderr, val);
871 /* Return the folded version of T if T, when folded, is a gimple
872 min_invariant. Otherwise, return T. */
874 static tree
875 fully_constant_expression (tree t)
877 tree folded;
878 folded = fold (t);
879 if (folded && is_gimple_min_invariant (folded))
880 return folded;
881 return t;
884 /* Return a copy of a chain of nodes, chained through the TREE_CHAIN field.
885 For example, this can copy a list made of TREE_LIST nodes.
886 Allocates the nodes in list_node_pool*/
888 static tree
889 pool_copy_list (tree list)
891 tree head;
892 tree prev, next;
894 if (list == 0)
895 return 0;
896 head = (tree) pool_alloc (list_node_pool);
898 memcpy (head, list, tree_size (list));
899 prev = head;
901 next = TREE_CHAIN (list);
902 while (next)
904 TREE_CHAIN (prev) = (tree) pool_alloc (list_node_pool);
905 memcpy (TREE_CHAIN (prev), next, tree_size (next));
906 prev = TREE_CHAIN (prev);
907 next = TREE_CHAIN (next);
909 return head;
912 /* Translate the vuses in the VUSES vector backwards through phi nodes
913 in PHIBLOCK, so that they have the value they would have in
914 BLOCK. */
916 static VEC(tree, gc) *
917 translate_vuses_through_block (VEC (tree, gc) *vuses,
918 basic_block phiblock,
919 basic_block block)
921 tree oldvuse;
922 VEC(tree, gc) *result = NULL;
923 int i;
925 for (i = 0; VEC_iterate (tree, vuses, i, oldvuse); i++)
927 tree phi = SSA_NAME_DEF_STMT (oldvuse);
928 if (TREE_CODE (phi) == PHI_NODE
929 && bb_for_stmt (phi) == phiblock)
931 edge e = find_edge (block, bb_for_stmt (phi));
932 if (e)
934 tree def = PHI_ARG_DEF (phi, e->dest_idx);
935 if (def != oldvuse)
937 if (!result)
938 result = VEC_copy (tree, gc, vuses);
939 VEC_replace (tree, result, i, def);
945 /* We avoid creating a new copy of the vuses unless something
946 actually changed, so result can be NULL. */
947 if (result)
949 sort_vuses (result);
950 return result;
952 return vuses;
956 /* Like find_leader, but checks for the value existing in SET1 *or*
957 SET2. This is used to avoid making a set consisting of the union
958 of PA_IN and ANTIC_IN during insert. */
960 static inline tree
961 find_leader_in_sets (tree expr, bitmap_set_t set1, bitmap_set_t set2)
963 tree result;
965 result = bitmap_find_leader (set1, expr);
966 if (!result && set2)
967 result = bitmap_find_leader (set2, expr);
968 return result;
971 /* Translate EXPR using phis in PHIBLOCK, so that it has the values of
972 the phis in PRED. Return NULL if we can't find a leader for each
973 part of the translated expression. */
975 static tree
976 phi_translate (tree expr, bitmap_set_t set1, bitmap_set_t set2,
977 basic_block pred, basic_block phiblock)
979 tree phitrans = NULL;
980 tree oldexpr = expr;
982 if (expr == NULL)
983 return NULL;
985 if (is_gimple_min_invariant (expr))
986 return expr;
988 /* Phi translations of a given expression don't change. */
989 if (EXPR_P (expr))
991 tree vh;
993 vh = get_value_handle (expr);
994 if (vh && TREE_CODE (vh) == VALUE_HANDLE)
995 phitrans = phi_trans_lookup (expr, pred, VALUE_HANDLE_VUSES (vh));
996 else
997 phitrans = phi_trans_lookup (expr, pred, NULL);
999 else
1000 phitrans = phi_trans_lookup (expr, pred, NULL);
1002 if (phitrans)
1003 return phitrans;
1005 switch (TREE_CODE_CLASS (TREE_CODE (expr)))
1007 case tcc_expression:
1009 if (TREE_CODE (expr) != CALL_EXPR)
1010 return NULL;
1011 else
1013 tree oldop0 = TREE_OPERAND (expr, 0);
1014 tree oldval0 = oldop0;
1015 tree oldarglist = TREE_OPERAND (expr, 1);
1016 tree oldop2 = TREE_OPERAND (expr, 2);
1017 tree oldval2 = oldop2;
1018 tree newop0;
1019 tree newarglist;
1020 tree newop2 = NULL;
1021 tree oldwalker;
1022 tree newwalker;
1023 tree newexpr;
1024 tree vh = get_value_handle (expr);
1025 bool listchanged = false;
1026 bool invariantarg = false;
1027 VEC (tree, gc) *vuses = VALUE_HANDLE_VUSES (vh);
1028 VEC (tree, gc) *tvuses;
1030 /* Call expressions are kind of weird because they have an
1031 argument list. We don't want to value number the list
1032 as one value number, because that doesn't make much
1033 sense, and just breaks the support functions we call,
1034 which expect TREE_OPERAND (call_expr, 2) to be a
1035 TREE_LIST. */
1036 oldval0 = find_leader_in_sets (oldop0, set1, set2);
1037 newop0 = phi_translate (oldval0, set1, set2, pred, phiblock);
1038 if (newop0 == NULL)
1039 return NULL;
1040 if (oldop2)
1042 oldop2 = find_leader_in_sets (oldop2, set1, set2);
1043 newop2 = phi_translate (oldop2, set1, set2, pred, phiblock);
1044 if (newop2 == NULL)
1045 return NULL;
1048 /* phi translate the argument list piece by piece.
1050 We could actually build the list piece by piece here,
1051 but it's likely to not be worth the memory we will save,
1052 unless you have millions of call arguments. */
1054 newarglist = pool_copy_list (oldarglist);
1055 for (oldwalker = oldarglist, newwalker = newarglist;
1056 oldwalker && newwalker;
1057 oldwalker = TREE_CHAIN (oldwalker),
1058 newwalker = TREE_CHAIN (newwalker))
1061 tree oldval = TREE_VALUE (oldwalker);
1062 tree newval;
1063 if (oldval)
1065 /* This may seem like a weird place for this
1066 check, but it's actually the easiest place to
1067 do it. We can't do it lower on in the
1068 recursion because it's valid for pieces of a
1069 component ref to be of AGGREGATE_TYPE, as long
1070 as the outermost one is not.
1071 To avoid *that* case, we have a check for
1072 AGGREGATE_TYPE_P in insert_aux. However, that
1073 check will *not* catch this case because here
1074 it occurs in the argument list. */
1075 if (AGGREGATE_TYPE_P (TREE_TYPE (oldval)))
1076 return NULL;
1077 oldval = find_leader_in_sets (oldval, set1, set2);
1078 newval = phi_translate (oldval, set1, set2, pred,
1079 phiblock);
1080 if (newval == NULL)
1081 return NULL;
1082 if (newval != oldval)
1084 listchanged = true;
1085 invariantarg |= is_gimple_min_invariant (newval);
1086 TREE_VALUE (newwalker) = get_value_handle (newval);
1091 /* In case of new invariant args we might try to fold the call
1092 again. */
1093 if (invariantarg)
1095 tree tmp = fold_ternary (CALL_EXPR, TREE_TYPE (expr),
1096 newop0, newarglist, newop2);
1097 if (tmp)
1099 STRIP_TYPE_NOPS (tmp);
1100 if (is_gimple_min_invariant (tmp))
1101 return tmp;
1105 if (listchanged)
1106 vn_lookup_or_add (newarglist, NULL);
1108 tvuses = translate_vuses_through_block (vuses, phiblock, pred);
1110 if (listchanged || (newop0 != oldop0) || (oldop2 != newop2)
1111 || vuses != tvuses)
1113 newexpr = (tree) pool_alloc (expression_node_pool);
1114 memcpy (newexpr, expr, tree_size (expr));
1115 TREE_OPERAND (newexpr, 0) = newop0 == oldop0 ? oldval0 : get_value_handle (newop0);
1116 TREE_OPERAND (newexpr, 1) = listchanged ? newarglist : oldarglist;
1117 TREE_OPERAND (newexpr, 2) = newop2 == oldop2 ? oldval2 : get_value_handle (newop2);
1118 newexpr->common.ann = NULL;
1119 vn_lookup_or_add_with_vuses (newexpr, tvuses);
1120 expr = newexpr;
1121 phi_trans_add (oldexpr, newexpr, pred, tvuses);
1125 return expr;
1127 case tcc_declaration:
1129 VEC (tree, gc) * oldvuses = NULL;
1130 VEC (tree, gc) * newvuses = NULL;
1132 oldvuses = VALUE_HANDLE_VUSES (get_value_handle (expr));
1133 if (oldvuses)
1134 newvuses = translate_vuses_through_block (oldvuses, phiblock,
1135 pred);
1137 if (oldvuses != newvuses)
1138 vn_lookup_or_add_with_vuses (expr, newvuses);
1140 phi_trans_add (oldexpr, expr, pred, newvuses);
1142 return expr;
1144 case tcc_reference:
1146 tree oldop0 = TREE_OPERAND (expr, 0);
1147 tree oldop1 = NULL;
1148 tree newop0;
1149 tree newop1 = NULL;
1150 tree oldop2 = NULL;
1151 tree newop2 = NULL;
1152 tree oldop3 = NULL;
1153 tree newop3 = NULL;
1154 tree newexpr;
1155 VEC (tree, gc) * oldvuses = NULL;
1156 VEC (tree, gc) * newvuses = NULL;
1158 if (TREE_CODE (expr) != INDIRECT_REF
1159 && TREE_CODE (expr) != COMPONENT_REF
1160 && TREE_CODE (expr) != ARRAY_REF)
1161 return NULL;
1163 oldop0 = find_leader_in_sets (oldop0, set1, set2);
1164 newop0 = phi_translate (oldop0, set1, set2, pred, phiblock);
1165 if (newop0 == NULL)
1166 return NULL;
1168 if (TREE_CODE (expr) == ARRAY_REF)
1170 oldop1 = TREE_OPERAND (expr, 1);
1171 oldop1 = find_leader_in_sets (oldop1, set1, set2);
1172 newop1 = phi_translate (oldop1, set1, set2, pred, phiblock);
1174 if (newop1 == NULL)
1175 return NULL;
1177 oldop2 = TREE_OPERAND (expr, 2);
1178 if (oldop2)
1180 oldop2 = find_leader_in_sets (oldop2, set1, set2);
1181 newop2 = phi_translate (oldop2, set1, set2, pred, phiblock);
1183 if (newop2 == NULL)
1184 return NULL;
1186 oldop3 = TREE_OPERAND (expr, 3);
1187 if (oldop3)
1189 oldop3 = find_leader_in_sets (oldop3, set1, set2);
1190 newop3 = phi_translate (oldop3, set1, set2, pred, phiblock);
1192 if (newop3 == NULL)
1193 return NULL;
1197 oldvuses = VALUE_HANDLE_VUSES (get_value_handle (expr));
1198 if (oldvuses)
1199 newvuses = translate_vuses_through_block (oldvuses, phiblock,
1200 pred);
1202 if (newop0 != oldop0 || newvuses != oldvuses
1203 || newop1 != oldop1
1204 || newop2 != oldop2
1205 || newop3 != oldop3)
1207 tree t;
1209 newexpr = (tree) pool_alloc (reference_node_pool);
1210 memcpy (newexpr, expr, tree_size (expr));
1211 TREE_OPERAND (newexpr, 0) = get_value_handle (newop0);
1212 if (TREE_CODE (expr) == ARRAY_REF)
1214 TREE_OPERAND (newexpr, 1) = get_value_handle (newop1);
1215 if (newop2)
1216 TREE_OPERAND (newexpr, 2) = get_value_handle (newop2);
1217 if (newop3)
1218 TREE_OPERAND (newexpr, 3) = get_value_handle (newop3);
1221 t = fully_constant_expression (newexpr);
1223 if (t != newexpr)
1225 pool_free (reference_node_pool, newexpr);
1226 newexpr = t;
1228 else
1230 newexpr->common.ann = NULL;
1231 vn_lookup_or_add_with_vuses (newexpr, newvuses);
1233 expr = newexpr;
1234 phi_trans_add (oldexpr, newexpr, pred, newvuses);
1237 return expr;
1238 break;
1240 case tcc_binary:
1241 case tcc_comparison:
1243 tree oldop1 = TREE_OPERAND (expr, 0);
1244 tree oldval1 = oldop1;
1245 tree oldop2 = TREE_OPERAND (expr, 1);
1246 tree oldval2 = oldop2;
1247 tree newop1;
1248 tree newop2;
1249 tree newexpr;
1251 oldop1 = find_leader_in_sets (oldop1, set1, set2);
1252 newop1 = phi_translate (oldop1, set1, set2, pred, phiblock);
1253 if (newop1 == NULL)
1254 return NULL;
1256 oldop2 = find_leader_in_sets (oldop2, set1, set2);
1257 newop2 = phi_translate (oldop2, set1, set2, pred, phiblock);
1258 if (newop2 == NULL)
1259 return NULL;
1260 if (newop1 != oldop1 || newop2 != oldop2)
1262 tree t;
1263 newexpr = (tree) pool_alloc (binary_node_pool);
1264 memcpy (newexpr, expr, tree_size (expr));
1265 TREE_OPERAND (newexpr, 0) = newop1 == oldop1 ? oldval1 : get_value_handle (newop1);
1266 TREE_OPERAND (newexpr, 1) = newop2 == oldop2 ? oldval2 : get_value_handle (newop2);
1267 t = fully_constant_expression (newexpr);
1268 if (t != newexpr)
1270 pool_free (binary_node_pool, newexpr);
1271 newexpr = t;
1273 else
1275 newexpr->common.ann = NULL;
1276 vn_lookup_or_add (newexpr, NULL);
1278 expr = newexpr;
1279 phi_trans_add (oldexpr, newexpr, pred, NULL);
1282 return expr;
1284 case tcc_unary:
1286 tree oldop1 = TREE_OPERAND (expr, 0);
1287 tree newop1;
1288 tree newexpr;
1290 oldop1 = find_leader_in_sets (oldop1, set1, set2);
1291 newop1 = phi_translate (oldop1, set1, set2, pred, phiblock);
1292 if (newop1 == NULL)
1293 return NULL;
1294 if (newop1 != oldop1)
1296 tree t;
1297 newexpr = (tree) pool_alloc (unary_node_pool);
1298 memcpy (newexpr, expr, tree_size (expr));
1299 TREE_OPERAND (newexpr, 0) = get_value_handle (newop1);
1300 t = fully_constant_expression (newexpr);
1301 if (t != newexpr)
1303 pool_free (unary_node_pool, newexpr);
1304 newexpr = t;
1306 else
1308 newexpr->common.ann = NULL;
1309 vn_lookup_or_add (newexpr, NULL);
1311 expr = newexpr;
1312 phi_trans_add (oldexpr, newexpr, pred, NULL);
1315 return expr;
1317 case tcc_exceptional:
1319 tree phi = NULL;
1320 edge e;
1321 tree def_stmt;
1322 gcc_assert (TREE_CODE (expr) == SSA_NAME);
1324 def_stmt = SSA_NAME_DEF_STMT (expr);
1325 if (TREE_CODE (def_stmt) == PHI_NODE
1326 && bb_for_stmt (def_stmt) == phiblock)
1327 phi = def_stmt;
1328 else
1329 return expr;
1331 e = find_edge (pred, bb_for_stmt (phi));
1332 if (e)
1334 if (is_undefined_value (PHI_ARG_DEF (phi, e->dest_idx)))
1335 return NULL;
1336 vn_lookup_or_add (PHI_ARG_DEF (phi, e->dest_idx), NULL);
1337 return PHI_ARG_DEF (phi, e->dest_idx);
1340 return expr;
1342 default:
1343 gcc_unreachable ();
1347 /* For each expression in SET, translate the value handles through phi nodes
1348 in PHIBLOCK using edge PHIBLOCK->PRED, and store the resulting
1349 expressions in DEST. */
1351 static void
1352 phi_translate_set (bitmap_set_t dest, bitmap_set_t set, basic_block pred,
1353 basic_block phiblock)
1355 VEC (tree, heap) *exprs;
1356 tree expr;
1357 int i;
1359 if (!phi_nodes (phiblock))
1361 bitmap_set_copy (dest, set);
1362 return;
1365 exprs = sorted_array_from_bitmap_set (set);
1366 for (i = 0; VEC_iterate (tree, exprs, i, expr); i++)
1368 tree translated;
1369 translated = phi_translate (expr, set, NULL, pred, phiblock);
1371 /* Don't add constants or empty translations to the cache, since
1372 we won't look them up that way, or use the result, anyway. */
1373 if (translated && !is_gimple_min_invariant (translated))
1375 tree vh = get_value_handle (translated);
1376 VEC (tree, gc) *vuses;
1378 /* The value handle itself may also be an invariant, in
1379 which case, it has no vuses. */
1380 vuses = !is_gimple_min_invariant (vh)
1381 ? VALUE_HANDLE_VUSES (vh) : NULL;
1382 phi_trans_add (expr, translated, pred, vuses);
1385 if (translated != NULL)
1386 bitmap_value_insert_into_set (dest, translated);
1388 VEC_free (tree, heap, exprs);
1391 /* Find the leader for a value (i.e., the name representing that
1392 value) in a given set, and return it. Return NULL if no leader is
1393 found. */
1395 static tree
1396 bitmap_find_leader (bitmap_set_t set, tree val)
1398 if (val == NULL)
1399 return NULL;
1401 if (constant_expr_p (val))
1402 return val;
1404 if (bitmap_set_contains_value (set, val))
1406 /* Rather than walk the entire bitmap of expressions, and see
1407 whether any of them has the value we are looking for, we look
1408 at the reverse mapping, which tells us the set of expressions
1409 that have a given value (IE value->expressions with that
1410 value) and see if any of those expressions are in our set.
1411 The number of expressions per value is usually significantly
1412 less than the number of expressions in the set. In fact, for
1413 large testcases, doing it this way is roughly 5-10x faster
1414 than walking the bitmap.
1415 If this is somehow a significant lose for some cases, we can
1416 choose which set to walk based on which set is smaller. */
1417 unsigned int i;
1418 bitmap_iterator bi;
1419 bitmap_set_t exprset = VALUE_HANDLE_EXPR_SET (val);
1421 EXECUTE_IF_AND_IN_BITMAP (exprset->expressions,
1422 set->expressions, 0, i, bi)
1423 return expression_for_id (i);
1425 return NULL;
1428 /* Given the vuse representative map, MAP, and an SSA version number,
1429 ID, return the bitmap of names ID represents, or NULL, if none
1430 exists. */
1432 static bitmap
1433 get_representative (bitmap *map, int id)
1435 if (map[id] != NULL)
1436 return map[id];
1437 return NULL;
1440 /* A vuse is anticipable at the top of block x, from the bottom of the
1441 block, if it reaches the top of the block, and is not killed in the
1442 block. In effect, we are trying to see if the vuse is transparent
1443 backwards in the block. */
1445 static bool
1446 vuses_dies_in_block_x (VEC (tree, gc) *vuses, basic_block block)
1448 int i;
1449 tree vuse;
1451 for (i = 0; VEC_iterate (tree, vuses, i, vuse); i++)
1453 /* Any places where this is too conservative, are places
1454 where we created a new version and shouldn't have. */
1456 if (!bitmap_bit_p (RVUSE_IN (block), SSA_NAME_VERSION (vuse))
1457 || bitmap_bit_p (RVUSE_KILL (block), SSA_NAME_VERSION (vuse)))
1458 return true;
1460 return false;
1463 /* Determine if the expression EXPR is valid in SET1 U SET2.
1464 ONLY SET2 CAN BE NULL.
1465 This means that we have a leader for each part of the expression
1466 (if it consists of values), or the expression is an SSA_NAME.
1468 NB: We never should run into a case where we have SSA_NAME +
1469 SSA_NAME or SSA_NAME + value. The sets valid_in_sets is called on,
1470 the ANTIC sets, will only ever have SSA_NAME's or value expressions
1471 (IE VALUE1 + VALUE2, *VALUE1, VALUE1 < VALUE2) */
1473 #define union_contains_value(SET1, SET2, VAL) \
1474 (bitmap_set_contains_value ((SET1), (VAL)) \
1475 || ((SET2) && bitmap_set_contains_value ((SET2), (VAL))))
1477 static bool
1478 valid_in_sets (bitmap_set_t set1, bitmap_set_t set2, tree expr,
1479 basic_block block)
1481 tree vh = get_value_handle (expr);
1482 switch (TREE_CODE_CLASS (TREE_CODE (expr)))
1484 case tcc_binary:
1485 case tcc_comparison:
1487 tree op1 = TREE_OPERAND (expr, 0);
1488 tree op2 = TREE_OPERAND (expr, 1);
1490 return union_contains_value (set1, set2, op1)
1491 && union_contains_value (set1, set2, op2);
1494 case tcc_unary:
1496 tree op1 = TREE_OPERAND (expr, 0);
1497 return union_contains_value (set1, set2, op1);
1500 case tcc_expression:
1502 if (TREE_CODE (expr) == CALL_EXPR)
1504 tree op0 = TREE_OPERAND (expr, 0);
1505 tree arglist = TREE_OPERAND (expr, 1);
1506 tree op2 = TREE_OPERAND (expr, 2);
1508 /* Check the non-list operands first. */
1509 if (!union_contains_value (set1, set2, op0)
1510 || (op2 && !union_contains_value (set1, set2, op2)))
1511 return false;
1513 /* Now check the operands. */
1514 for (; arglist; arglist = TREE_CHAIN (arglist))
1516 tree arg = TREE_VALUE (arglist);
1517 if (!union_contains_value (set1, set2, arg))
1518 return false;
1520 return !vuses_dies_in_block_x (VALUE_HANDLE_VUSES (vh), block);
1522 return false;
1525 case tcc_reference:
1527 if (TREE_CODE (expr) == INDIRECT_REF
1528 || TREE_CODE (expr) == COMPONENT_REF
1529 || TREE_CODE (expr) == ARRAY_REF)
1531 tree op0 = TREE_OPERAND (expr, 0);
1532 gcc_assert (is_gimple_min_invariant (op0)
1533 || TREE_CODE (op0) == VALUE_HANDLE);
1534 if (!union_contains_value (set1, set2, op0))
1535 return false;
1536 if (TREE_CODE (expr) == ARRAY_REF)
1538 tree op1 = TREE_OPERAND (expr, 1);
1539 tree op2 = TREE_OPERAND (expr, 2);
1540 tree op3 = TREE_OPERAND (expr, 3);
1541 gcc_assert (is_gimple_min_invariant (op1)
1542 || TREE_CODE (op1) == VALUE_HANDLE);
1543 if (!union_contains_value (set1, set2, op1))
1544 return false;
1545 gcc_assert (!op2 || is_gimple_min_invariant (op2)
1546 || TREE_CODE (op2) == VALUE_HANDLE);
1547 if (op2
1548 && !union_contains_value (set1, set2, op2))
1549 return false;
1550 gcc_assert (!op3 || is_gimple_min_invariant (op3)
1551 || TREE_CODE (op3) == VALUE_HANDLE);
1552 if (op3
1553 && !union_contains_value (set1, set2, op3))
1554 return false;
1556 return bitmap_set_contains_value (ANTIC_SAFE_LOADS (block),
1558 || !vuses_dies_in_block_x (VALUE_HANDLE_VUSES (vh),
1559 block);
1562 return false;
1564 case tcc_exceptional:
1566 gcc_assert (TREE_CODE (expr) == SSA_NAME);
1567 return bitmap_set_contains_expr (AVAIL_OUT (block), expr);
1570 case tcc_declaration:
1571 return !vuses_dies_in_block_x (VALUE_HANDLE_VUSES (vh), block);
1573 default:
1574 /* No other cases should be encountered. */
1575 gcc_unreachable ();
1579 /* Clean the set of expressions that are no longer valid in SET1 or
1580 SET2. This means expressions that are made up of values we have no
1581 leaders for in SET1 or SET2. This version is used for partial
1582 anticipation, which means it is not valid in either ANTIC_IN or
1583 PA_IN. */
1585 static void
1586 dependent_clean (bitmap_set_t set1, bitmap_set_t set2, basic_block block)
1588 VEC (tree, heap) *exprs = sorted_array_from_bitmap_set (set1);
1589 tree expr;
1590 int i;
1592 for (i = 0; VEC_iterate (tree, exprs, i, expr); i++)
1594 if (!valid_in_sets (set1, set2, expr, block))
1595 bitmap_remove_from_set (set1, expr);
1597 VEC_free (tree, heap, exprs);
1600 /* Clean the set of expressions that are no longer valid in SET. This
1601 means expressions that are made up of values we have no leaders for
1602 in SET. */
1604 static void
1605 clean (bitmap_set_t set, basic_block block)
1607 VEC (tree, heap) *exprs = sorted_array_from_bitmap_set (set);
1608 tree expr;
1609 int i;
1611 for (i = 0; VEC_iterate (tree, exprs, i, expr); i++)
1613 if (!valid_in_sets (set, NULL, expr, block))
1614 bitmap_remove_from_set (set, expr);
1616 VEC_free (tree, heap, exprs);
1619 static sbitmap has_abnormal_preds;
1622 /* List of blocks that may have changed during ANTIC computation and
1623 thus need to be iterated over. */
1625 static sbitmap changed_blocks;
1626 /* Compute the ANTIC set for BLOCK.
1628 If succs(BLOCK) > 1 then
1629 ANTIC_OUT[BLOCK] = intersection of ANTIC_IN[b] for all succ(BLOCK)
1630 else if succs(BLOCK) == 1 then
1631 ANTIC_OUT[BLOCK] = phi_translate (ANTIC_IN[succ(BLOCK)])
1633 ANTIC_IN[BLOCK] = clean(ANTIC_OUT[BLOCK] U EXP_GEN[BLOCK] - TMP_GEN[BLOCK])
1636 static bool
1637 compute_antic_aux (basic_block block, bool block_has_abnormal_pred_edge)
1639 bool changed = false;
1640 bitmap_set_t S, old, ANTIC_OUT;
1641 bitmap_iterator bi;
1642 unsigned int bii;
1643 edge e;
1644 edge_iterator ei;
1646 old = ANTIC_OUT = S = NULL;
1647 BB_VISITED (block) = 1;
1649 /* If any edges from predecessors are abnormal, antic_in is empty,
1650 so do nothing. */
1651 if (block_has_abnormal_pred_edge)
1652 goto maybe_dump_sets;
1654 old = ANTIC_IN (block);
1655 ANTIC_OUT = bitmap_set_new ();
1657 /* If the block has no successors, ANTIC_OUT is empty. */
1658 if (EDGE_COUNT (block->succs) == 0)
1660 /* If we have one successor, we could have some phi nodes to
1661 translate through. */
1662 else if (single_succ_p (block))
1664 basic_block succ_bb = single_succ (block);
1666 /* We trade iterations of the dataflow equations for having to
1667 phi translate the maximal set, which is incredibly slow
1668 (since the maximal set often has 300+ members, even when you
1669 have a small number of blocks).
1670 Basically, we defer the computation of ANTIC for this block
1671 until we have processed it's successor, which will inveitably
1672 have a *much* smaller set of values to phi translate once
1673 clean has been run on it.
1674 The cost of doing this is that we technically perform more
1675 iterations, however, they are lower cost iterations.
1677 Timings for PRE on tramp3d-v4:
1678 without maximal set fix: 11 seconds
1679 with maximal set fix/without deferring: 26 seconds
1680 with maximal set fix/with deferring: 11 seconds
1683 if (!BB_VISITED (succ_bb))
1685 changed = true;
1686 SET_BIT (changed_blocks, block->index);
1687 BB_VISITED (block) = 0;
1688 BB_DEFERRED (block) = 1;
1689 goto maybe_dump_sets;
1691 else
1692 phi_translate_set (ANTIC_OUT, ANTIC_IN (succ_bb),
1693 block, succ_bb);
1697 /* If we have multiple successors, we take the intersection of all of
1698 them. */
1699 else
1701 VEC(basic_block, heap) * worklist;
1702 size_t i;
1703 basic_block bprime, first;
1705 worklist = VEC_alloc (basic_block, heap, EDGE_COUNT (block->succs));
1706 FOR_EACH_EDGE (e, ei, block->succs)
1707 VEC_quick_push (basic_block, worklist, e->dest);
1708 first = VEC_index (basic_block, worklist, 0);
1710 if (!BB_VISITED (first))
1711 bitmap_set_copy (ANTIC_OUT, maximal_set);
1712 else
1713 bitmap_set_copy (ANTIC_OUT, ANTIC_IN (first));
1715 for (i = 1; VEC_iterate (basic_block, worklist, i, bprime); i++)
1717 if (!BB_VISITED (bprime))
1718 bitmap_set_and (ANTIC_OUT, maximal_set);
1719 else
1720 bitmap_set_and (ANTIC_OUT, ANTIC_IN (bprime));
1722 VEC_free (basic_block, heap, worklist);
1725 /* Generate ANTIC_OUT - TMP_GEN. */
1726 S = bitmap_set_subtract (ANTIC_OUT, TMP_GEN (block));
1728 /* Start ANTIC_IN with EXP_GEN - TMP_GEN. */
1729 ANTIC_IN (block) = bitmap_set_subtract (EXP_GEN (block),
1730 TMP_GEN (block));
1732 /* Then union in the ANTIC_OUT - TMP_GEN values,
1733 to get ANTIC_OUT U EXP_GEN - TMP_GEN */
1734 FOR_EACH_EXPR_ID_IN_SET (S, bii, bi)
1735 bitmap_value_insert_into_set (ANTIC_IN (block),
1736 expression_for_id (bii));
1738 clean (ANTIC_IN (block), block);
1740 /* !old->expressions can happen when we deferred a block. */
1741 if (!old->expressions || !bitmap_set_equal (old, ANTIC_IN (block)))
1743 changed = true;
1744 SET_BIT (changed_blocks, block->index);
1745 FOR_EACH_EDGE (e, ei, block->preds)
1746 SET_BIT (changed_blocks, e->src->index);
1748 else
1749 RESET_BIT (changed_blocks, block->index);
1751 maybe_dump_sets:
1752 if (dump_file && (dump_flags & TDF_DETAILS))
1754 if (!BB_DEFERRED (block) || BB_VISITED (block))
1756 if (ANTIC_OUT)
1757 print_bitmap_set (dump_file, ANTIC_OUT, "ANTIC_OUT", block->index);
1759 if (ANTIC_SAFE_LOADS (block))
1760 print_bitmap_set (dump_file, ANTIC_SAFE_LOADS (block),
1761 "ANTIC_SAFE_LOADS", block->index);
1762 print_bitmap_set (dump_file, ANTIC_IN (block), "ANTIC_IN",
1763 block->index);
1765 if (S)
1766 print_bitmap_set (dump_file, S, "S", block->index);
1768 else
1770 fprintf (dump_file,
1771 "Block %d was deferred for a future iteration.\n",
1772 block->index);
1775 if (old)
1776 bitmap_set_free (old);
1777 if (S)
1778 bitmap_set_free (S);
1779 if (ANTIC_OUT)
1780 bitmap_set_free (ANTIC_OUT);
1781 return changed;
1784 /* Compute PARTIAL_ANTIC for BLOCK.
1786 If succs(BLOCK) > 1 then
1787 PA_OUT[BLOCK] = value wise union of PA_IN[b] + all ANTIC_IN not
1788 in ANTIC_OUT for all succ(BLOCK)
1789 else if succs(BLOCK) == 1 then
1790 PA_OUT[BLOCK] = phi_translate (PA_IN[succ(BLOCK)])
1792 PA_IN[BLOCK] = dependent_clean(PA_OUT[BLOCK] - TMP_GEN[BLOCK]
1793 - ANTIC_IN[BLOCK])
1796 static bool
1797 compute_partial_antic_aux (basic_block block,
1798 bool block_has_abnormal_pred_edge)
1800 bool changed = false;
1801 bitmap_set_t old_PA_IN;
1802 bitmap_set_t PA_OUT;
1803 edge e;
1804 edge_iterator ei;
1806 old_PA_IN = PA_OUT = NULL;
1808 /* If any edges from predecessors are abnormal, antic_in is empty,
1809 so do nothing. */
1810 if (block_has_abnormal_pred_edge)
1811 goto maybe_dump_sets;
1813 old_PA_IN = PA_IN (block);
1814 PA_OUT = bitmap_set_new ();
1816 /* If the block has no successors, ANTIC_OUT is empty. */
1817 if (EDGE_COUNT (block->succs) == 0)
1819 /* If we have one successor, we could have some phi nodes to
1820 translate through. Note that we can't phi translate across DFS
1821 back edges in partial antic, because it uses a union operation
1822 on the successors. For recurrences like IV's, we will end up generating a
1823 new value in the set on each go around (i + 3 (VH.1) VH.1 + 1
1824 (VH.2), VH.2 + 1 (VH.3), etc), forever. */
1825 else if (single_succ_p (block))
1827 basic_block succ = single_succ (block);
1828 if (!(single_succ_edge (block)->flags & EDGE_DFS_BACK))
1829 phi_translate_set (PA_OUT, PA_IN (succ), block, succ);
1831 /* If we have multiple successors, we take the union of all of
1832 them. */
1833 else
1835 VEC(basic_block, heap) * worklist;
1836 size_t i;
1837 basic_block bprime;
1839 worklist = VEC_alloc (basic_block, heap, EDGE_COUNT (block->succs));
1840 FOR_EACH_EDGE (e, ei, block->succs)
1842 if (e->flags & EDGE_DFS_BACK)
1843 continue;
1844 VEC_quick_push (basic_block, worklist, e->dest);
1846 if (VEC_length (basic_block, worklist) > 0)
1848 for (i = 0; VEC_iterate (basic_block, worklist, i, bprime); i++)
1850 unsigned int i;
1851 bitmap_iterator bi;
1853 FOR_EACH_EXPR_ID_IN_SET (ANTIC_IN (bprime), i, bi)
1854 bitmap_value_insert_into_set (PA_OUT,
1855 expression_for_id (i));
1857 FOR_EACH_EXPR_ID_IN_SET (PA_IN (bprime), i, bi)
1858 bitmap_value_insert_into_set (PA_OUT,
1859 expression_for_id (i));
1862 VEC_free (basic_block, heap, worklist);
1865 /* PA_IN starts with PA_OUT - TMP_GEN.
1866 Then we subtract things from ANTIC_IN. */
1867 PA_IN (block) = bitmap_set_subtract (PA_OUT, TMP_GEN (block));
1869 /* For partial antic, we want to put back in the phi results, since
1870 we will properly avoid making them partially antic over backedges. */
1871 bitmap_ior_into (PA_IN (block)->values, PHI_GEN (block)->values);
1872 bitmap_ior_into (PA_IN (block)->expressions, PHI_GEN (block)->expressions);
1874 /* PA_IN[block] = PA_IN[block] - ANTIC_IN[block] */
1875 bitmap_set_subtract_values (PA_IN (block), ANTIC_IN (block));
1877 dependent_clean (PA_IN (block), ANTIC_IN (block), block);
1879 if (!bitmap_set_equal (old_PA_IN, PA_IN (block)))
1881 changed = true;
1882 SET_BIT (changed_blocks, block->index);
1883 FOR_EACH_EDGE (e, ei, block->preds)
1884 SET_BIT (changed_blocks, e->src->index);
1886 else
1887 RESET_BIT (changed_blocks, block->index);
1889 maybe_dump_sets:
1890 if (dump_file && (dump_flags & TDF_DETAILS))
1892 if (PA_OUT)
1893 print_bitmap_set (dump_file, PA_OUT, "PA_OUT", block->index);
1895 print_bitmap_set (dump_file, PA_IN (block), "PA_IN", block->index);
1897 if (old_PA_IN)
1898 bitmap_set_free (old_PA_IN);
1899 if (PA_OUT)
1900 bitmap_set_free (PA_OUT);
1901 return changed;
1904 /* Compute ANTIC and partial ANTIC sets. */
1906 static void
1907 compute_antic (void)
1909 bool changed = true;
1910 int num_iterations = 0;
1911 basic_block block;
1912 int i;
1914 /* If any predecessor edges are abnormal, we punt, so antic_in is empty.
1915 We pre-build the map of blocks with incoming abnormal edges here. */
1916 has_abnormal_preds = sbitmap_alloc (last_basic_block);
1917 sbitmap_zero (has_abnormal_preds);
1919 FOR_EACH_BB (block)
1921 edge_iterator ei;
1922 edge e;
1924 FOR_EACH_EDGE (e, ei, block->preds)
1926 e->flags &= ~EDGE_DFS_BACK;
1927 if (e->flags & EDGE_ABNORMAL)
1929 SET_BIT (has_abnormal_preds, block->index);
1930 break;
1934 BB_VISITED (block) = 0;
1935 BB_DEFERRED (block) = 0;
1936 /* While we are here, give empty ANTIC_IN sets to each block. */
1937 ANTIC_IN (block) = bitmap_set_new ();
1938 PA_IN (block) = bitmap_set_new ();
1941 /* At the exit block we anticipate nothing. */
1942 ANTIC_IN (EXIT_BLOCK_PTR) = bitmap_set_new ();
1943 BB_VISITED (EXIT_BLOCK_PTR) = 1;
1944 PA_IN (EXIT_BLOCK_PTR) = bitmap_set_new ();
1946 changed_blocks = sbitmap_alloc (last_basic_block + 1);
1947 sbitmap_ones (changed_blocks);
1948 while (changed)
1950 if (dump_file && (dump_flags & TDF_DETAILS))
1951 fprintf (dump_file, "Starting iteration %d\n", num_iterations);
1952 num_iterations++;
1953 changed = false;
1954 for (i = 0; i < last_basic_block - NUM_FIXED_BLOCKS; i++)
1956 if (TEST_BIT (changed_blocks, postorder[i]))
1958 basic_block block = BASIC_BLOCK (postorder[i]);
1959 changed |= compute_antic_aux (block,
1960 TEST_BIT (has_abnormal_preds,
1961 block->index));
1964 /* Theoretically possible, but *highly* unlikely. */
1965 gcc_assert (num_iterations < 50);
1968 if (dump_file && (dump_flags & TDF_STATS))
1969 fprintf (dump_file, "compute_antic required %d iterations\n",
1970 num_iterations);
1972 if (do_partial_partial)
1974 sbitmap_ones (changed_blocks);
1975 mark_dfs_back_edges ();
1976 num_iterations = 0;
1977 changed = true;
1978 while (changed)
1980 if (dump_file && (dump_flags & TDF_DETAILS))
1981 fprintf (dump_file, "Starting iteration %d\n", num_iterations);
1982 num_iterations++;
1983 changed = false;
1984 for (i = 0; i < last_basic_block - NUM_FIXED_BLOCKS; i++)
1986 if (TEST_BIT (changed_blocks, postorder[i]))
1988 basic_block block = BASIC_BLOCK (postorder[i]);
1989 changed
1990 |= compute_partial_antic_aux (block,
1991 TEST_BIT (has_abnormal_preds,
1992 block->index));
1995 /* Theoretically possible, but *highly* unlikely. */
1996 gcc_assert (num_iterations < 50);
1998 if (dump_file && (dump_flags & TDF_STATS))
1999 fprintf (dump_file, "compute_partial_antic required %d iterations\n",
2000 num_iterations);
2002 sbitmap_free (has_abnormal_preds);
2003 sbitmap_free (changed_blocks);
2006 /* Print the names represented by the bitmap NAMES, to the file OUT. */
2008 static void
2009 dump_bitmap_of_names (FILE *out, bitmap names)
2011 bitmap_iterator bi;
2012 unsigned int i;
2014 fprintf (out, " { ");
2015 EXECUTE_IF_SET_IN_BITMAP (names, 0, i, bi)
2017 print_generic_expr (out, ssa_name (i), 0);
2018 fprintf (out, " ");
2020 fprintf (out, "}\n");
2023 /* Compute a set of representative vuse versions for each phi. This
2024 is so we can compute conservative kill sets in terms of all vuses
2025 that are killed, instead of continually walking chains.
2027 We also have to be able kill all names associated with a phi when
2028 the phi dies in order to ensure we don't generate overlapping
2029 live ranges, which are not allowed in virtual SSA. */
2031 static bitmap *vuse_names;
2032 static void
2033 compute_vuse_representatives (void)
2035 tree phi;
2036 basic_block bb;
2037 VEC (tree, heap) *phis = NULL;
2038 bool changed = true;
2039 size_t i;
2041 FOR_EACH_BB (bb)
2043 for (phi = phi_nodes (bb);
2044 phi;
2045 phi = PHI_CHAIN (phi))
2046 if (!is_gimple_reg (PHI_RESULT (phi)))
2047 VEC_safe_push (tree, heap, phis, phi);
2050 while (changed)
2052 changed = false;
2054 for (i = 0; VEC_iterate (tree, phis, i, phi); i++)
2056 size_t ver = SSA_NAME_VERSION (PHI_RESULT (phi));
2057 use_operand_p usep;
2058 ssa_op_iter iter;
2060 if (vuse_names[ver] == NULL)
2062 vuse_names[ver] = BITMAP_ALLOC (&grand_bitmap_obstack);
2063 bitmap_set_bit (vuse_names[ver], ver);
2065 FOR_EACH_PHI_ARG (usep, phi, iter, SSA_OP_ALL_USES)
2067 tree use = USE_FROM_PTR (usep);
2068 bitmap usebitmap = get_representative (vuse_names,
2069 SSA_NAME_VERSION (use));
2070 if (usebitmap != NULL)
2072 changed |= bitmap_ior_into (vuse_names[ver],
2073 usebitmap);
2075 else
2077 changed |= !bitmap_bit_p (vuse_names[ver],
2078 SSA_NAME_VERSION (use));
2079 if (changed)
2080 bitmap_set_bit (vuse_names[ver],
2081 SSA_NAME_VERSION (use));
2087 if (dump_file && (dump_flags & TDF_DETAILS))
2088 for (i = 0; VEC_iterate (tree, phis, i, phi); i++)
2090 bitmap reps = get_representative (vuse_names,
2091 SSA_NAME_VERSION (PHI_RESULT (phi)));
2092 if (reps)
2094 print_generic_expr (dump_file, PHI_RESULT (phi), 0);
2095 fprintf (dump_file, " represents ");
2096 dump_bitmap_of_names (dump_file, reps);
2099 VEC_free (tree, heap, phis);
2102 /* Compute reaching vuses and antic safe loads. RVUSE computation is
2103 is a small bit of iterative dataflow to determine what virtual uses
2104 reach what blocks. Because we can't generate overlapping virtual
2105 uses, and virtual uses *do* actually die, this ends up being faster
2106 in most cases than continually walking the virtual use/def chains
2107 to determine whether we are inside a block where a given virtual is
2108 still available to be used.
2110 ANTIC_SAFE_LOADS are those loads that actually occur before any kill to
2111 their vuses in the block,and thus, are safe at the top of the
2112 block.
2114 An example:
2116 <block begin>
2117 b = *a
2118 *a = 9
2119 <block end>
2121 b = *a is an antic safe load because it still safe to consider it
2122 ANTIC at the top of the block.
2124 We currently compute a conservative approximation to
2125 ANTIC_SAFE_LOADS. We compute those loads that occur before *any*
2126 stores in the block. This is not because it is difficult to
2127 compute the precise answer, but because it is expensive. More
2128 testing is necessary to determine whether it is worth computing the
2129 precise answer. */
2131 static void
2132 compute_rvuse_and_antic_safe (void)
2135 unsigned int i;
2136 tree phi;
2137 basic_block bb;
2138 bool changed = true;
2139 unsigned int *first_store_uid;
2141 first_store_uid = xcalloc (n_basic_blocks, sizeof (unsigned int));
2143 compute_vuse_representatives ();
2145 FOR_ALL_BB (bb)
2147 RVUSE_IN (bb) = BITMAP_ALLOC (&grand_bitmap_obstack);
2148 RVUSE_GEN (bb) = BITMAP_ALLOC (&grand_bitmap_obstack);
2149 RVUSE_KILL (bb) = BITMAP_ALLOC (&grand_bitmap_obstack);
2150 RVUSE_OUT (bb) = BITMAP_ALLOC (&grand_bitmap_obstack);
2151 ANTIC_SAFE_LOADS (bb) = NULL;
2154 /* Mark live on entry */
2155 for (i = 0; i < num_ssa_names; i++)
2157 tree name = ssa_name (i);
2158 if (name && !is_gimple_reg (name)
2159 && IS_EMPTY_STMT (SSA_NAME_DEF_STMT (name)))
2160 bitmap_set_bit (RVUSE_OUT (ENTRY_BLOCK_PTR),
2161 SSA_NAME_VERSION (name));
2164 /* Compute local sets for reaching vuses.
2165 GEN(block) = generated in block and not locally killed.
2166 KILL(block) = set of vuses killed in block.
2169 FOR_EACH_BB (bb)
2171 block_stmt_iterator bsi;
2172 ssa_op_iter iter;
2173 def_operand_p defp;
2174 use_operand_p usep;
2176 for (bsi = bsi_start (bb); !bsi_end_p (bsi); bsi_next (&bsi))
2178 tree stmt = bsi_stmt (bsi);
2180 if (first_store_uid[bb->index] == 0
2181 && !ZERO_SSA_OPERANDS (stmt, SSA_OP_VMAYUSE | SSA_OP_VMAYDEF
2182 | SSA_OP_VMUSTDEF | SSA_OP_VMUSTKILL))
2184 first_store_uid[bb->index] = stmt_ann (stmt)->uid;
2188 FOR_EACH_SSA_USE_OPERAND (usep, stmt, iter, SSA_OP_VIRTUAL_KILLS
2189 | SSA_OP_VMAYUSE)
2191 tree use = USE_FROM_PTR (usep);
2192 bitmap repbit = get_representative (vuse_names,
2193 SSA_NAME_VERSION (use));
2194 if (repbit != NULL)
2196 bitmap_and_compl_into (RVUSE_GEN (bb), repbit);
2197 bitmap_ior_into (RVUSE_KILL (bb), repbit);
2199 else
2201 bitmap_set_bit (RVUSE_KILL (bb), SSA_NAME_VERSION (use));
2202 bitmap_clear_bit (RVUSE_GEN (bb), SSA_NAME_VERSION (use));
2205 FOR_EACH_SSA_DEF_OPERAND (defp, stmt, iter, SSA_OP_VIRTUAL_DEFS)
2207 tree def = DEF_FROM_PTR (defp);
2208 bitmap_set_bit (RVUSE_GEN (bb), SSA_NAME_VERSION (def));
2213 FOR_EACH_BB (bb)
2215 for (phi = phi_nodes (bb); phi; phi = PHI_CHAIN (phi))
2217 if (!is_gimple_reg (PHI_RESULT (phi)))
2219 edge e;
2220 edge_iterator ei;
2222 tree def = PHI_RESULT (phi);
2223 /* In reality, the PHI result is generated at the end of
2224 each predecessor block. This will make the value
2225 LVUSE_IN for the bb containing the PHI, which is
2226 correct. */
2227 FOR_EACH_EDGE (e, ei, bb->preds)
2228 bitmap_set_bit (RVUSE_GEN (e->src), SSA_NAME_VERSION (def));
2233 /* Solve reaching vuses.
2235 RVUSE_IN[BB] = Union of RVUSE_OUT of predecessors.
2236 RVUSE_OUT[BB] = RVUSE_GEN[BB] U (RVUSE_IN[BB] - RVUSE_KILL[BB])
2239 changed = true;
2240 while (changed)
2242 int j;
2243 changed = false;
2244 for (j = n_basic_blocks - NUM_FIXED_BLOCKS - 1; j >= 0; j--)
2246 edge e;
2247 edge_iterator ei;
2248 bb = BASIC_BLOCK (postorder[j]);
2250 FOR_EACH_EDGE (e, ei, bb->preds)
2251 bitmap_ior_into (RVUSE_IN (bb), RVUSE_OUT (e->src));
2253 changed |= bitmap_ior_and_compl (RVUSE_OUT (bb),
2254 RVUSE_GEN (bb),
2255 RVUSE_IN (bb),
2256 RVUSE_KILL (bb));
2260 if (dump_file && (dump_flags & TDF_DETAILS))
2262 FOR_ALL_BB (bb)
2264 fprintf (dump_file, "RVUSE_IN (%d) =", bb->index);
2265 dump_bitmap_of_names (dump_file, RVUSE_IN (bb));
2267 fprintf (dump_file, "RVUSE_KILL (%d) =", bb->index);
2268 dump_bitmap_of_names (dump_file, RVUSE_KILL (bb));
2270 fprintf (dump_file, "RVUSE_GEN (%d) =", bb->index);
2271 dump_bitmap_of_names (dump_file, RVUSE_GEN (bb));
2273 fprintf (dump_file, "RVUSE_OUT (%d) =", bb->index);
2274 dump_bitmap_of_names (dump_file, RVUSE_OUT (bb));
2278 FOR_EACH_BB (bb)
2280 bitmap_iterator bi;
2282 if (bitmap_empty_p (RVUSE_KILL (bb)))
2283 continue;
2285 FOR_EACH_EXPR_ID_IN_SET (EXP_GEN (bb), i, bi)
2287 tree expr = expression_for_id (i);
2288 if (REFERENCE_CLASS_P (expr))
2290 tree vh = get_value_handle (expr);
2291 tree maybe = bitmap_find_leader (AVAIL_OUT (bb), vh);
2293 if (maybe)
2295 tree def = SSA_NAME_DEF_STMT (maybe);
2297 if (bb_for_stmt (def) != bb)
2298 continue;
2300 if (TREE_CODE (def) == PHI_NODE
2301 || stmt_ann (def)->uid < first_store_uid[bb->index])
2303 if (ANTIC_SAFE_LOADS (bb) == NULL)
2304 ANTIC_SAFE_LOADS (bb) = bitmap_set_new ();
2305 bitmap_value_insert_into_set (ANTIC_SAFE_LOADS (bb),
2306 expr);
2312 free (first_store_uid);
2315 /* Return true if we can value number the call in STMT. This is true
2316 if we have a pure or constant call. */
2318 static bool
2319 can_value_number_call (tree stmt)
2321 tree call = get_call_expr_in (stmt);
2323 if (call_expr_flags (call) & (ECF_PURE | ECF_CONST))
2324 return true;
2325 return false;
2328 /* Return true if OP is a tree which we can perform value numbering
2329 on. */
2331 static bool
2332 can_value_number_operation (tree op)
2334 return UNARY_CLASS_P (op)
2335 || BINARY_CLASS_P (op)
2336 || COMPARISON_CLASS_P (op)
2337 || REFERENCE_CLASS_P (op)
2338 || (TREE_CODE (op) == CALL_EXPR
2339 && can_value_number_call (op));
2343 /* Return true if OP is a tree which we can perform PRE on
2344 on. This may not match the operations we can value number, but in
2345 a perfect world would. */
2347 static bool
2348 can_PRE_operation (tree op)
2350 return UNARY_CLASS_P (op)
2351 || BINARY_CLASS_P (op)
2352 || COMPARISON_CLASS_P (op)
2353 || TREE_CODE (op) == INDIRECT_REF
2354 || TREE_CODE (op) == COMPONENT_REF
2355 || TREE_CODE (op) == CALL_EXPR
2356 || TREE_CODE (op) == ARRAY_REF;
2360 /* Inserted expressions are placed onto this worklist, which is used
2361 for performing quick dead code elimination of insertions we made
2362 that didn't turn out to be necessary. */
2363 static VEC(tree,heap) *inserted_exprs;
2365 /* Pool allocated fake store expressions are placed onto this
2366 worklist, which, after performing dead code elimination, is walked
2367 to see which expressions need to be put into GC'able memory */
2368 static VEC(tree, heap) *need_creation;
2370 /* For COMPONENT_REF's and ARRAY_REF's, we can't have any intermediates for the
2371 COMPONENT_REF or INDIRECT_REF or ARRAY_REF portion, because we'd end up with
2372 trying to rename aggregates into ssa form directly, which is a no
2375 Thus, this routine doesn't create temporaries, it just builds a
2376 single access expression for the array, calling
2377 find_or_generate_expression to build the innermost pieces.
2379 This function is a subroutine of create_expression_by_pieces, and
2380 should not be called on it's own unless you really know what you
2381 are doing.
2383 static tree
2384 create_component_ref_by_pieces (basic_block block, tree expr, tree stmts)
2386 tree genop = expr;
2387 tree folded;
2389 if (TREE_CODE (genop) == VALUE_HANDLE)
2391 tree found = bitmap_find_leader (AVAIL_OUT (block), expr);
2392 if (found)
2393 return found;
2396 if (TREE_CODE (genop) == VALUE_HANDLE)
2398 bitmap_set_t exprset = VALUE_HANDLE_EXPR_SET (expr);
2399 unsigned int firstbit = bitmap_first_set_bit (exprset->expressions);
2400 genop = expression_for_id (firstbit);
2403 switch TREE_CODE (genop)
2405 case ARRAY_REF:
2407 tree op0;
2408 tree op1, op2, op3;
2409 op0 = create_component_ref_by_pieces (block,
2410 TREE_OPERAND (genop, 0),
2411 stmts);
2412 op1 = TREE_OPERAND (genop, 1);
2413 if (TREE_CODE (op1) == VALUE_HANDLE)
2414 op1 = find_or_generate_expression (block, op1, stmts);
2415 op2 = TREE_OPERAND (genop, 2);
2416 if (op2 && TREE_CODE (op2) == VALUE_HANDLE)
2417 op2 = find_or_generate_expression (block, op2, stmts);
2418 op3 = TREE_OPERAND (genop, 3);
2419 if (op3 && TREE_CODE (op3) == VALUE_HANDLE)
2420 op3 = find_or_generate_expression (block, op3, stmts);
2421 folded = build4 (ARRAY_REF, TREE_TYPE (genop), op0, op1,
2422 op2, op3);
2423 return folded;
2425 case COMPONENT_REF:
2427 bitmap_set_t exprset;
2428 unsigned int firstbit;
2429 tree op0;
2430 tree op1;
2431 op0 = create_component_ref_by_pieces (block,
2432 TREE_OPERAND (genop, 0),
2433 stmts);
2434 exprset = VALUE_HANDLE_EXPR_SET (TREE_OPERAND (genop, 1));
2435 firstbit = bitmap_first_set_bit (exprset->expressions);
2436 op1 = expression_for_id (firstbit);
2437 folded = fold_build3 (COMPONENT_REF, TREE_TYPE (genop), op0, op1,
2438 NULL_TREE);
2439 return folded;
2441 break;
2442 case INDIRECT_REF:
2444 tree op1 = TREE_OPERAND (genop, 0);
2445 tree genop1 = find_or_generate_expression (block, op1, stmts);
2447 folded = fold_build1 (TREE_CODE (genop), TREE_TYPE (genop),
2448 genop1);
2449 return folded;
2451 break;
2452 case VAR_DECL:
2453 case PARM_DECL:
2454 case RESULT_DECL:
2455 case SSA_NAME:
2456 case STRING_CST:
2457 return genop;
2458 default:
2459 gcc_unreachable ();
2462 return NULL_TREE;
2465 /* Find a leader for an expression, or generate one using
2466 create_expression_by_pieces if it's ANTIC but
2467 complex.
2468 BLOCK is the basic_block we are looking for leaders in.
2469 EXPR is the expression to find a leader or generate for.
2470 STMTS is the statement list to put the inserted expressions on.
2471 Returns the SSA_NAME of the LHS of the generated expression or the
2472 leader. */
2474 static tree
2475 find_or_generate_expression (basic_block block, tree expr, tree stmts)
2477 tree genop = bitmap_find_leader (AVAIL_OUT (block), expr);
2479 /* If it's still NULL, it must be a complex expression, so generate
2480 it recursively. */
2481 if (genop == NULL)
2483 bitmap_set_t exprset = VALUE_HANDLE_EXPR_SET (expr);
2484 unsigned int firstbit = bitmap_first_set_bit (exprset->expressions);
2486 genop = expression_for_id (firstbit);
2487 gcc_assert (can_PRE_operation (genop));
2488 genop = create_expression_by_pieces (block, genop, stmts);
2490 return genop;
2493 #define NECESSARY(stmt) stmt->common.asm_written_flag
2494 /* Create an expression in pieces, so that we can handle very complex
2495 expressions that may be ANTIC, but not necessary GIMPLE.
2496 BLOCK is the basic block the expression will be inserted into,
2497 EXPR is the expression to insert (in value form)
2498 STMTS is a statement list to append the necessary insertions into.
2500 This function will die if we hit some value that shouldn't be
2501 ANTIC but is (IE there is no leader for it, or its components).
2502 This function may also generate expressions that are themselves
2503 partially or fully redundant. Those that are will be either made
2504 fully redundant during the next iteration of insert (for partially
2505 redundant ones), or eliminated by eliminate (for fully redundant
2506 ones). */
2508 static tree
2509 create_expression_by_pieces (basic_block block, tree expr, tree stmts)
2511 tree temp, name;
2512 tree folded, forced_stmts, newexpr;
2513 tree v;
2514 tree_stmt_iterator tsi;
2516 switch (TREE_CODE_CLASS (TREE_CODE (expr)))
2518 case tcc_expression:
2520 tree op0, op2;
2521 tree arglist;
2522 tree genop0, genop2;
2523 tree genarglist;
2524 tree walker, genwalker;
2526 gcc_assert (TREE_CODE (expr) == CALL_EXPR);
2527 genop2 = NULL;
2529 op0 = TREE_OPERAND (expr, 0);
2530 arglist = TREE_OPERAND (expr, 1);
2531 op2 = TREE_OPERAND (expr, 2);
2533 genop0 = find_or_generate_expression (block, op0, stmts);
2534 genarglist = copy_list (arglist);
2535 for (walker = arglist, genwalker = genarglist;
2536 genwalker && walker;
2537 genwalker = TREE_CHAIN (genwalker), walker = TREE_CHAIN (walker))
2539 TREE_VALUE (genwalker)
2540 = find_or_generate_expression (block, TREE_VALUE (walker),
2541 stmts);
2544 if (op2)
2545 genop2 = find_or_generate_expression (block, op2, stmts);
2546 folded = fold_build3 (TREE_CODE (expr), TREE_TYPE (expr),
2547 genop0, genarglist, genop2);
2548 break;
2552 break;
2553 case tcc_reference:
2555 if (TREE_CODE (expr) == COMPONENT_REF
2556 || TREE_CODE (expr) == ARRAY_REF)
2558 folded = create_component_ref_by_pieces (block, expr, stmts);
2560 else
2562 tree op1 = TREE_OPERAND (expr, 0);
2563 tree genop1 = find_or_generate_expression (block, op1, stmts);
2565 folded = fold_build1 (TREE_CODE (expr), TREE_TYPE (expr),
2566 genop1);
2568 break;
2571 case tcc_binary:
2572 case tcc_comparison:
2574 tree op1 = TREE_OPERAND (expr, 0);
2575 tree op2 = TREE_OPERAND (expr, 1);
2576 tree genop1 = find_or_generate_expression (block, op1, stmts);
2577 tree genop2 = find_or_generate_expression (block, op2, stmts);
2578 folded = fold_build2 (TREE_CODE (expr), TREE_TYPE (expr),
2579 genop1, genop2);
2580 break;
2583 case tcc_unary:
2585 tree op1 = TREE_OPERAND (expr, 0);
2586 tree genop1 = find_or_generate_expression (block, op1, stmts);
2587 folded = fold_build1 (TREE_CODE (expr), TREE_TYPE (expr),
2588 genop1);
2589 break;
2592 default:
2593 gcc_unreachable ();
2596 /* Force the generated expression to be a sequence of GIMPLE
2597 statements.
2598 We have to call unshare_expr because force_gimple_operand may
2599 modify the tree we pass to it. */
2600 newexpr = force_gimple_operand (unshare_expr (folded), &forced_stmts,
2601 false, NULL);
2603 /* If we have any intermediate expressions to the value sets, add them
2604 to the value sets and chain them on in the instruction stream. */
2605 if (forced_stmts)
2607 tsi = tsi_start (forced_stmts);
2608 for (; !tsi_end_p (tsi); tsi_next (&tsi))
2610 tree stmt = tsi_stmt (tsi);
2611 tree forcedname = TREE_OPERAND (stmt, 0);
2612 tree forcedexpr = TREE_OPERAND (stmt, 1);
2613 tree val = vn_lookup_or_add (forcedexpr, NULL);
2615 VEC_safe_push (tree, heap, inserted_exprs, stmt);
2616 vn_add (forcedname, val);
2617 bitmap_value_replace_in_set (NEW_SETS (block), forcedname);
2618 bitmap_value_replace_in_set (AVAIL_OUT (block), forcedname);
2619 mark_new_vars_to_rename (stmt);
2621 tsi = tsi_last (stmts);
2622 tsi_link_after (&tsi, forced_stmts, TSI_CONTINUE_LINKING);
2625 /* Build and insert the assignment of the end result to the temporary
2626 that we will return. */
2627 if (!pretemp || TREE_TYPE (expr) != TREE_TYPE (pretemp))
2629 pretemp = create_tmp_var (TREE_TYPE (expr), "pretmp");
2630 get_var_ann (pretemp);
2633 temp = pretemp;
2634 add_referenced_var (temp);
2636 if (TREE_CODE (TREE_TYPE (expr)) == COMPLEX_TYPE)
2637 DECL_COMPLEX_GIMPLE_REG_P (temp) = 1;
2639 newexpr = build2 (MODIFY_EXPR, TREE_TYPE (expr), temp, newexpr);
2640 name = make_ssa_name (temp, newexpr);
2641 TREE_OPERAND (newexpr, 0) = name;
2642 NECESSARY (newexpr) = 0;
2644 tsi = tsi_last (stmts);
2645 tsi_link_after (&tsi, newexpr, TSI_CONTINUE_LINKING);
2646 VEC_safe_push (tree, heap, inserted_exprs, newexpr);
2647 mark_new_vars_to_rename (newexpr);
2649 /* Add a value handle to the temporary.
2650 The value may already exist in either NEW_SETS, or AVAIL_OUT, because
2651 we are creating the expression by pieces, and this particular piece of
2652 the expression may have been represented. There is no harm in replacing
2653 here. */
2654 v = get_value_handle (expr);
2655 vn_add (name, v);
2656 get_or_alloc_expression_id (name);
2657 bitmap_value_replace_in_set (NEW_SETS (block), name);
2658 bitmap_value_replace_in_set (AVAIL_OUT (block), name);
2660 pre_stats.insertions++;
2661 if (dump_file && (dump_flags & TDF_DETAILS))
2663 fprintf (dump_file, "Inserted ");
2664 print_generic_expr (dump_file, newexpr, 0);
2665 fprintf (dump_file, " in predecessor %d\n", block->index);
2668 return name;
2671 /* Insert the to-be-made-available values of expression EXPRNUM for each
2672 predecessor, stored in AVAIL, into the predecessors of BLOCK, and
2673 merge the result with a phi node, given the same value handle as
2674 NODE. Return true if we have inserted new stuff. */
2676 static bool
2677 insert_into_preds_of_block (basic_block block, unsigned int exprnum,
2678 tree *avail)
2680 tree expr = expression_for_id (exprnum);
2681 tree val = get_value_handle (expr);
2682 edge pred;
2683 bool insertions = false;
2684 bool nophi = false;
2685 basic_block bprime;
2686 tree eprime;
2687 edge_iterator ei;
2688 tree type = TREE_TYPE (avail[EDGE_PRED (block, 0)->src->index]);
2689 tree temp;
2691 if (dump_file && (dump_flags & TDF_DETAILS))
2693 fprintf (dump_file, "Found partial redundancy for expression ");
2694 print_generic_expr (dump_file, expr, 0);
2695 fprintf (dump_file, " (");
2696 print_generic_expr (dump_file, val, 0);
2697 fprintf (dump_file, ")");
2698 fprintf (dump_file, "\n");
2701 /* Make sure we aren't creating an induction variable. */
2702 if (block->loop_depth > 0 && EDGE_COUNT (block->preds) == 2
2703 && TREE_CODE_CLASS (TREE_CODE (expr)) != tcc_reference )
2705 bool firstinsideloop = false;
2706 bool secondinsideloop = false;
2707 firstinsideloop = flow_bb_inside_loop_p (block->loop_father,
2708 EDGE_PRED (block, 0)->src);
2709 secondinsideloop = flow_bb_inside_loop_p (block->loop_father,
2710 EDGE_PRED (block, 1)->src);
2711 /* Induction variables only have one edge inside the loop. */
2712 if (firstinsideloop ^ secondinsideloop)
2714 if (dump_file && (dump_flags & TDF_DETAILS))
2715 fprintf (dump_file, "Skipping insertion of phi for partial redundancy: Looks like an induction variable\n");
2716 nophi = true;
2721 /* Make the necessary insertions. */
2722 FOR_EACH_EDGE (pred, ei, block->preds)
2724 tree stmts = alloc_stmt_list ();
2725 tree builtexpr;
2726 bprime = pred->src;
2727 eprime = avail[bprime->index];
2729 if (can_PRE_operation (eprime))
2731 #ifdef ENABLE_CHECKING
2732 tree vh;
2734 /* eprime may be an invariant. */
2735 vh = TREE_CODE (eprime) == VALUE_HANDLE
2736 ? eprime
2737 : get_value_handle (eprime);
2739 /* ensure that the virtual uses we need reach our block. */
2740 if (TREE_CODE (vh) == VALUE_HANDLE)
2742 int i;
2743 tree vuse;
2744 for (i = 0;
2745 VEC_iterate (tree, VALUE_HANDLE_VUSES (vh), i, vuse);
2746 i++)
2748 size_t id = SSA_NAME_VERSION (vuse);
2749 gcc_assert (bitmap_bit_p (RVUSE_OUT (bprime), id)
2750 || IS_EMPTY_STMT (SSA_NAME_DEF_STMT (vuse)));
2753 #endif
2754 builtexpr = create_expression_by_pieces (bprime,
2755 eprime,
2756 stmts);
2757 gcc_assert (!(pred->flags & EDGE_ABNORMAL));
2758 bsi_insert_on_edge (pred, stmts);
2759 avail[bprime->index] = builtexpr;
2760 insertions = true;
2763 /* If we didn't want a phi node, and we made insertions, we still have
2764 inserted new stuff, and thus return true. If we didn't want a phi node,
2765 and didn't make insertions, we haven't added anything new, so return
2766 false. */
2767 if (nophi && insertions)
2768 return true;
2769 else if (nophi && !insertions)
2770 return false;
2772 /* Now build a phi for the new variable. */
2773 if (!prephitemp || TREE_TYPE (prephitemp) != type)
2775 prephitemp = create_tmp_var (type, "prephitmp");
2776 get_var_ann (prephitemp);
2779 temp = prephitemp;
2780 add_referenced_var (temp);
2782 if (TREE_CODE (type) == COMPLEX_TYPE)
2783 DECL_COMPLEX_GIMPLE_REG_P (temp) = 1;
2784 temp = create_phi_node (temp, block);
2786 NECESSARY (temp) = 0;
2787 VEC_safe_push (tree, heap, inserted_exprs, temp);
2788 FOR_EACH_EDGE (pred, ei, block->preds)
2789 add_phi_arg (temp, avail[pred->src->index], pred);
2791 vn_add (PHI_RESULT (temp), val);
2793 /* The value should *not* exist in PHI_GEN, or else we wouldn't be doing
2794 this insertion, since we test for the existence of this value in PHI_GEN
2795 before proceeding with the partial redundancy checks in insert_aux.
2797 The value may exist in AVAIL_OUT, in particular, it could be represented
2798 by the expression we are trying to eliminate, in which case we want the
2799 replacement to occur. If it's not existing in AVAIL_OUT, we want it
2800 inserted there.
2802 Similarly, to the PHI_GEN case, the value should not exist in NEW_SETS of
2803 this block, because if it did, it would have existed in our dominator's
2804 AVAIL_OUT, and would have been skipped due to the full redundancy check.
2807 bitmap_insert_into_set (PHI_GEN (block),
2808 PHI_RESULT (temp));
2809 bitmap_value_replace_in_set (AVAIL_OUT (block),
2810 PHI_RESULT (temp));
2811 bitmap_insert_into_set (NEW_SETS (block),
2812 PHI_RESULT (temp));
2814 if (dump_file && (dump_flags & TDF_DETAILS))
2816 fprintf (dump_file, "Created phi ");
2817 print_generic_expr (dump_file, temp, 0);
2818 fprintf (dump_file, " in block %d\n", block->index);
2820 pre_stats.phis++;
2821 return true;
2826 /* Perform insertion of partially redundant values.
2827 For BLOCK, do the following:
2828 1. Propagate the NEW_SETS of the dominator into the current block.
2829 If the block has multiple predecessors,
2830 2a. Iterate over the ANTIC expressions for the block to see if
2831 any of them are partially redundant.
2832 2b. If so, insert them into the necessary predecessors to make
2833 the expression fully redundant.
2834 2c. Insert a new PHI merging the values of the predecessors.
2835 2d. Insert the new PHI, and the new expressions, into the
2836 NEW_SETS set.
2837 3. Recursively call ourselves on the dominator children of BLOCK.
2839 Steps 1, 2a, and 3 are done by insert_aux. 2b, 2c and 2d are done by
2840 do_regular_insertion and do_partial_insertion.
2844 static bool
2845 do_regular_insertion (basic_block block, basic_block dom)
2847 bool new_stuff = false;
2848 VEC (tree, heap) *exprs = sorted_array_from_bitmap_set (ANTIC_IN (block));
2849 tree expr;
2850 int i;
2852 for (i = 0; VEC_iterate (tree, exprs, i, expr); i++)
2854 if (can_PRE_operation (expr) && !AGGREGATE_TYPE_P (TREE_TYPE (expr)))
2856 tree *avail;
2857 tree val;
2858 bool by_some = false;
2859 bool cant_insert = false;
2860 bool all_same = true;
2861 tree first_s = NULL;
2862 edge pred;
2863 basic_block bprime;
2864 tree eprime = NULL_TREE;
2865 edge_iterator ei;
2867 val = get_value_handle (expr);
2868 if (bitmap_set_contains_value (PHI_GEN (block), val))
2869 continue;
2870 if (bitmap_set_contains_value (AVAIL_OUT (dom), val))
2872 if (dump_file && (dump_flags & TDF_DETAILS))
2873 fprintf (dump_file, "Found fully redundant value\n");
2874 continue;
2877 avail = XCNEWVEC (tree, last_basic_block);
2878 FOR_EACH_EDGE (pred, ei, block->preds)
2880 tree vprime;
2881 tree edoubleprime;
2883 /* This can happen in the very weird case
2884 that our fake infinite loop edges have caused a
2885 critical edge to appear. */
2886 if (EDGE_CRITICAL_P (pred))
2888 cant_insert = true;
2889 break;
2891 bprime = pred->src;
2892 eprime = phi_translate (expr, ANTIC_IN (block), NULL,
2893 bprime, block);
2895 /* eprime will generally only be NULL if the
2896 value of the expression, translated
2897 through the PHI for this predecessor, is
2898 undefined. If that is the case, we can't
2899 make the expression fully redundant,
2900 because its value is undefined along a
2901 predecessor path. We can thus break out
2902 early because it doesn't matter what the
2903 rest of the results are. */
2904 if (eprime == NULL)
2906 cant_insert = true;
2907 break;
2910 eprime = fully_constant_expression (eprime);
2911 vprime = get_value_handle (eprime);
2912 gcc_assert (vprime);
2913 edoubleprime = bitmap_find_leader (AVAIL_OUT (bprime),
2914 vprime);
2915 if (edoubleprime == NULL)
2917 avail[bprime->index] = eprime;
2918 all_same = false;
2920 else
2922 avail[bprime->index] = edoubleprime;
2923 by_some = true;
2924 if (first_s == NULL)
2925 first_s = edoubleprime;
2926 else if (!operand_equal_p (first_s, edoubleprime,
2928 all_same = false;
2931 /* If we can insert it, it's not the same value
2932 already existing along every predecessor, and
2933 it's defined by some predecessor, it is
2934 partially redundant. */
2935 if (!cant_insert && !all_same && by_some)
2937 if (insert_into_preds_of_block (block, get_expression_id (expr),
2938 avail))
2939 new_stuff = true;
2941 /* If all edges produce the same value and that value is
2942 an invariant, then the PHI has the same value on all
2943 edges. Note this. */
2944 else if (!cant_insert && all_same && eprime
2945 && is_gimple_min_invariant (eprime)
2946 && !is_gimple_min_invariant (val))
2948 unsigned int j;
2949 bitmap_iterator bi;
2951 bitmap_set_t exprset = VALUE_HANDLE_EXPR_SET (val);
2952 FOR_EACH_EXPR_ID_IN_SET (exprset, j, bi)
2954 tree expr = expression_for_id (j);
2955 if (TREE_CODE (expr) == SSA_NAME)
2957 vn_add (expr, eprime);
2958 pre_stats.constified++;
2962 free (avail);
2966 VEC_free (tree, heap, exprs);
2967 return new_stuff;
2971 /* Perform insertion for partially anticipatable expressions. There
2972 is only one case we will perform insertion for these. This case is
2973 if the expression is partially anticipatable, and fully available.
2974 In this case, we know that putting it earlier will enable us to
2975 remove the later computation. */
2978 static bool
2979 do_partial_partial_insertion (basic_block block, basic_block dom)
2981 bool new_stuff = false;
2982 VEC (tree, heap) *exprs = sorted_array_from_bitmap_set (PA_IN (block));
2983 tree expr;
2984 int i;
2986 for (i = 0; VEC_iterate (tree, exprs, i, expr); i++)
2988 if (can_PRE_operation (expr) && !AGGREGATE_TYPE_P (TREE_TYPE (expr)))
2990 tree *avail;
2991 tree val;
2992 bool by_all = true;
2993 bool cant_insert = false;
2994 edge pred;
2995 basic_block bprime;
2996 tree eprime = NULL_TREE;
2997 edge_iterator ei;
2999 val = get_value_handle (expr);
3000 if (bitmap_set_contains_value (PHI_GEN (block), val))
3001 continue;
3002 if (bitmap_set_contains_value (AVAIL_OUT (dom), val))
3003 continue;
3005 avail = XCNEWVEC (tree, last_basic_block);
3006 FOR_EACH_EDGE (pred, ei, block->preds)
3008 tree vprime;
3009 tree edoubleprime;
3011 /* This can happen in the very weird case
3012 that our fake infinite loop edges have caused a
3013 critical edge to appear. */
3014 if (EDGE_CRITICAL_P (pred))
3016 cant_insert = true;
3017 break;
3019 bprime = pred->src;
3020 eprime = phi_translate (expr, ANTIC_IN (block),
3021 PA_IN (block),
3022 bprime, block);
3024 /* eprime will generally only be NULL if the
3025 value of the expression, translated
3026 through the PHI for this predecessor, is
3027 undefined. If that is the case, we can't
3028 make the expression fully redundant,
3029 because its value is undefined along a
3030 predecessor path. We can thus break out
3031 early because it doesn't matter what the
3032 rest of the results are. */
3033 if (eprime == NULL)
3035 cant_insert = true;
3036 break;
3039 eprime = fully_constant_expression (eprime);
3040 vprime = get_value_handle (eprime);
3041 gcc_assert (vprime);
3042 edoubleprime = bitmap_find_leader (AVAIL_OUT (bprime),
3043 vprime);
3044 if (edoubleprime == NULL)
3046 by_all = false;
3047 break;
3049 else
3050 avail[bprime->index] = edoubleprime;
3054 /* If we can insert it, it's not the same value
3055 already existing along every predecessor, and
3056 it's defined by some predecessor, it is
3057 partially redundant. */
3058 if (!cant_insert && by_all)
3060 pre_stats.pa_insert++;
3061 if (insert_into_preds_of_block (block, get_expression_id (expr),
3062 avail))
3063 new_stuff = true;
3065 free (avail);
3069 VEC_free (tree, heap, exprs);
3070 return new_stuff;
3073 static bool
3074 insert_aux (basic_block block)
3076 basic_block son;
3077 bool new_stuff = false;
3079 if (block)
3081 basic_block dom;
3082 dom = get_immediate_dominator (CDI_DOMINATORS, block);
3083 if (dom)
3085 unsigned i;
3086 bitmap_iterator bi;
3087 bitmap_set_t newset = NEW_SETS (dom);
3088 if (newset)
3090 /* Note that we need to value_replace both NEW_SETS, and
3091 AVAIL_OUT. For both the case of NEW_SETS, the value may be
3092 represented by some non-simple expression here that we want
3093 to replace it with. */
3094 FOR_EACH_EXPR_ID_IN_SET (newset, i, bi)
3096 tree expr = expression_for_id (i);
3097 bitmap_value_replace_in_set (NEW_SETS (block), expr);
3098 bitmap_value_replace_in_set (AVAIL_OUT (block), expr);
3101 if (!single_pred_p (block))
3103 new_stuff |= do_regular_insertion (block, dom);
3104 if (do_partial_partial)
3105 new_stuff |= do_partial_partial_insertion (block, dom);
3109 for (son = first_dom_son (CDI_DOMINATORS, block);
3110 son;
3111 son = next_dom_son (CDI_DOMINATORS, son))
3113 new_stuff |= insert_aux (son);
3116 return new_stuff;
3119 /* Perform insertion of partially redundant values. */
3121 static void
3122 insert (void)
3124 bool new_stuff = true;
3125 basic_block bb;
3126 int num_iterations = 0;
3128 FOR_ALL_BB (bb)
3129 NEW_SETS (bb) = bitmap_set_new ();
3131 while (new_stuff)
3133 num_iterations++;
3134 new_stuff = false;
3135 new_stuff = insert_aux (ENTRY_BLOCK_PTR);
3137 if (num_iterations > 2 && dump_file && (dump_flags & TDF_STATS))
3138 fprintf (dump_file, "insert required %d iterations\n", num_iterations);
3142 /* Return true if VAR is an SSA variable with no defining statement in
3143 this procedure, *AND* isn't a live-on-entry parameter. */
3145 static bool
3146 is_undefined_value (tree expr)
3148 return (TREE_CODE (expr) == SSA_NAME
3149 && IS_EMPTY_STMT (SSA_NAME_DEF_STMT (expr))
3150 /* PARM_DECLs and hard registers are always defined. */
3151 && TREE_CODE (SSA_NAME_VAR (expr)) != PARM_DECL);
3155 /* Given an SSA variable VAR and an expression EXPR, compute the value
3156 number for EXPR and create a value handle (VAL) for it. If VAR and
3157 EXPR are not the same, associate VAL with VAR. Finally, add VAR to
3158 S1 and its value handle to S2, and to the maximal set if
3159 ADD_TO_MAXIMAL is true.
3161 VUSES represent the virtual use operands associated with EXPR (if
3162 any). */
3164 static inline void
3165 add_to_sets (tree var, tree expr, tree stmt, bitmap_set_t s1,
3166 bitmap_set_t s2)
3168 tree val = vn_lookup_or_add (expr, stmt);
3170 /* VAR and EXPR may be the same when processing statements for which
3171 we are not computing value numbers (e.g., non-assignments, or
3172 statements that make aliased stores). In those cases, we are
3173 only interested in making VAR available as its own value. */
3174 if (var != expr)
3175 vn_add (var, val);
3177 if (s1)
3178 bitmap_insert_into_set (s1, var);
3180 /* PHI nodes can't go in the maximal sets because they are not in
3181 TMP_GEN, so it is possible to get into non-monotonic situations
3182 during ANTIC calculation, because it will *add* bits. */
3183 if (!in_fre && TREE_CODE (SSA_NAME_DEF_STMT (var)) != PHI_NODE)
3184 bitmap_value_insert_into_set (maximal_set, var);
3185 bitmap_value_insert_into_set (s2, var);
3188 /* Find existing value expression that is the same as T,
3189 and return it if it exists. */
3191 static inline tree
3192 find_existing_value_expr (tree t, tree stmt)
3194 bitmap_iterator bi;
3195 unsigned int bii;
3196 tree vh;
3197 bitmap_set_t exprset;
3199 if (REFERENCE_CLASS_P (t))
3200 vh = vn_lookup (t, stmt);
3201 else
3202 vh = vn_lookup (t, NULL);
3204 if (!vh)
3205 return NULL;
3206 exprset = VALUE_HANDLE_EXPR_SET (vh);
3207 FOR_EACH_EXPR_ID_IN_SET (exprset, bii, bi)
3209 tree efi = expression_for_id (bii);
3210 if (expressions_equal_p (t, efi))
3211 return efi;
3213 return NULL;
3216 /* Given a unary or binary expression EXPR, create and return a new
3217 expression with the same structure as EXPR but with its operands
3218 replaced with the value handles of each of the operands of EXPR.
3220 VUSES represent the virtual use operands associated with EXPR (if
3221 any). Insert EXPR's operands into the EXP_GEN set for BLOCK. */
3223 static inline tree
3224 create_value_expr_from (tree expr, basic_block block, tree stmt)
3226 int i;
3227 enum tree_code code = TREE_CODE (expr);
3228 tree vexpr;
3229 alloc_pool pool;
3230 tree efi;
3232 gcc_assert (TREE_CODE_CLASS (code) == tcc_unary
3233 || TREE_CODE_CLASS (code) == tcc_binary
3234 || TREE_CODE_CLASS (code) == tcc_comparison
3235 || TREE_CODE_CLASS (code) == tcc_reference
3236 || TREE_CODE_CLASS (code) == tcc_expression
3237 || TREE_CODE_CLASS (code) == tcc_exceptional
3238 || TREE_CODE_CLASS (code) == tcc_declaration);
3240 if (TREE_CODE_CLASS (code) == tcc_unary)
3241 pool = unary_node_pool;
3242 else if (TREE_CODE_CLASS (code) == tcc_reference)
3243 pool = reference_node_pool;
3244 else if (TREE_CODE_CLASS (code) == tcc_binary)
3245 pool = binary_node_pool;
3246 else if (TREE_CODE_CLASS (code) == tcc_comparison)
3247 pool = comparison_node_pool;
3248 else if (TREE_CODE_CLASS (code) == tcc_exceptional)
3250 gcc_assert (code == TREE_LIST);
3251 pool = list_node_pool;
3253 else
3255 gcc_assert (code == CALL_EXPR);
3256 pool = expression_node_pool;
3259 vexpr = (tree) pool_alloc (pool);
3260 memcpy (vexpr, expr, tree_size (expr));
3262 /* This case is only for TREE_LIST's that appear as part of
3263 CALL_EXPR's. Anything else is a bug, but we can't easily verify
3264 this, hence this comment. TREE_LIST is not handled by the
3265 general case below because they don't have a fixed length, or
3266 operands, so you can't access purpose/value/chain through
3267 TREE_OPERAND macros. */
3269 if (code == TREE_LIST)
3271 tree op = NULL_TREE;
3272 tree temp = NULL_TREE;
3273 if (TREE_CHAIN (vexpr))
3274 temp = create_value_expr_from (TREE_CHAIN (vexpr), block, stmt);
3275 TREE_CHAIN (vexpr) = temp ? temp : TREE_CHAIN (vexpr);
3278 /* Recursively value-numberize reference ops. */
3279 if (REFERENCE_CLASS_P (TREE_VALUE (vexpr)))
3281 tree tempop;
3282 op = TREE_VALUE (vexpr);
3283 tempop = create_value_expr_from (op, block, stmt);
3284 op = tempop ? tempop : op;
3286 TREE_VALUE (vexpr) = vn_lookup_or_add (op, stmt);
3288 else
3290 op = TREE_VALUE (vexpr);
3291 TREE_VALUE (vexpr) = vn_lookup_or_add (TREE_VALUE (vexpr), NULL);
3293 /* This is the equivalent of inserting op into EXP_GEN like we
3294 do below */
3295 if (!is_undefined_value (op))
3296 bitmap_value_insert_into_set (EXP_GEN (block), op);
3298 efi = find_existing_value_expr (vexpr, stmt);
3299 if (efi)
3300 return efi;
3301 get_or_alloc_expression_id (vexpr);
3302 return vexpr;
3305 for (i = 0; i < TREE_CODE_LENGTH (code); i++)
3307 tree val, op;
3309 op = TREE_OPERAND (expr, i);
3310 if (op == NULL_TREE)
3311 continue;
3313 /* Recursively value-numberize reference ops and tree lists. */
3314 if (REFERENCE_CLASS_P (op))
3316 tree tempop = create_value_expr_from (op, block, stmt);
3317 op = tempop ? tempop : op;
3318 val = vn_lookup_or_add (op, stmt);
3320 else if (TREE_CODE (op) == TREE_LIST)
3322 tree tempop;
3324 gcc_assert (TREE_CODE (expr) == CALL_EXPR);
3325 tempop = create_value_expr_from (op, block, stmt);
3327 op = tempop ? tempop : op;
3328 vn_lookup_or_add (op, NULL);
3329 /* Unlike everywhere else, we do *not* want to replace the
3330 TREE_LIST itself with a value number, because support
3331 functions we call will blow up. */
3332 val = op;
3334 else
3335 /* Create a value handle for OP and add it to VEXPR. */
3336 val = vn_lookup_or_add (op, NULL);
3338 if (!is_undefined_value (op) && TREE_CODE (op) != TREE_LIST)
3339 bitmap_value_insert_into_set (EXP_GEN (block), op);
3341 if (TREE_CODE (val) == VALUE_HANDLE)
3342 TREE_TYPE (val) = TREE_TYPE (TREE_OPERAND (vexpr, i));
3344 TREE_OPERAND (vexpr, i) = val;
3346 efi = find_existing_value_expr (vexpr, stmt);
3347 if (efi)
3348 return efi;
3349 get_or_alloc_expression_id (vexpr);
3350 return vexpr;
3353 /* Given a statement STMT and its right hand side which is a load, try
3354 to look for the expression stored in the location for the load, and
3355 return true if a useful equivalence was recorded for LHS. */
3357 static bool
3358 try_look_through_load (tree lhs, tree mem_ref, tree stmt, basic_block block)
3360 tree store_stmt = NULL;
3361 tree rhs;
3362 ssa_op_iter i;
3363 tree vuse;
3365 FOR_EACH_SSA_TREE_OPERAND (vuse, stmt, i, SSA_OP_VIRTUAL_USES)
3367 tree def_stmt;
3369 gcc_assert (TREE_CODE (vuse) == SSA_NAME);
3370 def_stmt = SSA_NAME_DEF_STMT (vuse);
3372 /* If there is no useful statement for this VUSE, we'll not find a
3373 useful expression to return either. Likewise, if there is a
3374 statement but it is not a simple assignment or it has virtual
3375 uses, we can stop right here. Note that this means we do
3376 not look through PHI nodes, which is intentional. */
3377 if (!def_stmt
3378 || TREE_CODE (def_stmt) != MODIFY_EXPR
3379 || !ZERO_SSA_OPERANDS (def_stmt, SSA_OP_VIRTUAL_USES))
3380 return false;
3382 /* If this is not the same statement as one we have looked at for
3383 another VUSE of STMT already, we have two statements producing
3384 something that reaches our STMT. */
3385 if (store_stmt && store_stmt != def_stmt)
3386 return false;
3387 else
3389 /* Is this a store to the exact same location as the one we are
3390 loading from in STMT? */
3391 if (!operand_equal_p (TREE_OPERAND (def_stmt, 0), mem_ref, 0))
3392 return false;
3394 /* Otherwise remember this statement and see if all other VUSEs
3395 come from the same statement. */
3396 store_stmt = def_stmt;
3400 /* Alright then, we have visited all VUSEs of STMT and we've determined
3401 that all of them come from the same statement STORE_STMT. See if there
3402 is a useful expression we can deduce from STORE_STMT. */
3403 rhs = TREE_OPERAND (store_stmt, 1);
3404 if ((TREE_CODE (rhs) == SSA_NAME
3405 && !SSA_NAME_OCCURS_IN_ABNORMAL_PHI (rhs))
3406 || is_gimple_min_invariant (rhs)
3407 || TREE_CODE (rhs) == ADDR_EXPR
3408 || TREE_INVARIANT (rhs))
3411 /* Yay! Compute a value number for the RHS of the statement and
3412 add its value to the AVAIL_OUT set for the block. Add the LHS
3413 to TMP_GEN. */
3414 add_to_sets (lhs, rhs, store_stmt, TMP_GEN (block), AVAIL_OUT (block));
3415 if (TREE_CODE (rhs) == SSA_NAME
3416 && !is_undefined_value (rhs))
3417 bitmap_value_insert_into_set (EXP_GEN (block), rhs);
3418 return true;
3421 return false;
3424 /* Return a copy of NODE that is stored in the temporary alloc_pool's.
3425 This is made recursively true, so that the operands are stored in
3426 the pool as well. */
3428 static tree
3429 poolify_tree (tree node)
3431 switch (TREE_CODE (node))
3433 case INDIRECT_REF:
3435 tree temp = (tree) pool_alloc (reference_node_pool);
3436 memcpy (temp, node, tree_size (node));
3437 TREE_OPERAND (temp, 0) = poolify_tree (TREE_OPERAND (temp, 0));
3438 return temp;
3440 break;
3441 case MODIFY_EXPR:
3443 tree temp = (tree) pool_alloc (modify_expr_node_pool);
3444 memcpy (temp, node, tree_size (node));
3445 TREE_OPERAND (temp, 0) = poolify_tree (TREE_OPERAND (temp, 0));
3446 TREE_OPERAND (temp, 1) = poolify_tree (TREE_OPERAND (temp, 1));
3447 return temp;
3449 break;
3450 case SSA_NAME:
3451 case INTEGER_CST:
3452 case STRING_CST:
3453 case REAL_CST:
3454 case PARM_DECL:
3455 case VAR_DECL:
3456 case RESULT_DECL:
3457 return node;
3458 default:
3459 gcc_unreachable ();
3463 static tree modify_expr_template;
3465 /* Allocate a MODIFY_EXPR with TYPE, and operands OP1, OP2 in the
3466 alloc pools and return it. */
3467 static tree
3468 poolify_modify_expr (tree type, tree op1, tree op2)
3470 if (modify_expr_template == NULL)
3471 modify_expr_template = build2 (MODIFY_EXPR, type, op1, op2);
3473 TREE_OPERAND (modify_expr_template, 0) = op1;
3474 TREE_OPERAND (modify_expr_template, 1) = op2;
3475 TREE_TYPE (modify_expr_template) = type;
3477 return poolify_tree (modify_expr_template);
3481 /* For each real store operation of the form
3482 *a = <value> that we see, create a corresponding fake store of the
3483 form storetmp_<version> = *a.
3485 This enables AVAIL computation to mark the results of stores as
3486 available. Without this, you'd need to do some computation to
3487 mark the result of stores as ANTIC and AVAIL at all the right
3488 points.
3489 To save memory, we keep the store
3490 statements pool allocated until we decide whether they are
3491 necessary or not. */
3493 static void
3494 insert_fake_stores (void)
3496 basic_block block;
3498 FOR_ALL_BB (block)
3500 block_stmt_iterator bsi;
3501 for (bsi = bsi_start (block); !bsi_end_p (bsi); bsi_next (&bsi))
3503 tree stmt = bsi_stmt (bsi);
3505 /* We can't generate SSA names for stores that are complex
3506 or aggregate. We also want to ignore things whose
3507 virtual uses occur in abnormal phis. */
3509 if (TREE_CODE (stmt) == MODIFY_EXPR
3510 && TREE_CODE (TREE_OPERAND (stmt, 0)) == INDIRECT_REF
3511 && !AGGREGATE_TYPE_P (TREE_TYPE (TREE_OPERAND (stmt, 0)))
3512 && TREE_CODE (TREE_TYPE (TREE_OPERAND (stmt, 0))) != COMPLEX_TYPE)
3514 ssa_op_iter iter;
3515 def_operand_p defp;
3516 tree lhs = TREE_OPERAND (stmt, 0);
3517 tree rhs = TREE_OPERAND (stmt, 1);
3518 tree new;
3519 bool notokay = false;
3521 FOR_EACH_SSA_DEF_OPERAND (defp, stmt, iter, SSA_OP_VIRTUAL_DEFS)
3523 tree defvar = DEF_FROM_PTR (defp);
3524 if (SSA_NAME_OCCURS_IN_ABNORMAL_PHI (defvar))
3526 notokay = true;
3527 break;
3531 if (notokay)
3532 continue;
3534 if (!storetemp || TREE_TYPE (rhs) != TREE_TYPE (storetemp))
3536 storetemp = create_tmp_var (TREE_TYPE (rhs), "storetmp");
3537 get_var_ann (storetemp);
3540 new = poolify_modify_expr (TREE_TYPE (stmt), storetemp, lhs);
3542 lhs = make_ssa_name (storetemp, new);
3543 TREE_OPERAND (new, 0) = lhs;
3544 create_ssa_artficial_load_stmt (new, stmt);
3546 NECESSARY (new) = 0;
3547 VEC_safe_push (tree, heap, inserted_exprs, new);
3548 VEC_safe_push (tree, heap, need_creation, new);
3549 bsi_insert_after (&bsi, new, BSI_NEW_STMT);
3555 /* Turn the pool allocated fake stores that we created back into real
3556 GC allocated ones if they turned out to be necessary to PRE some
3557 expressions. */
3559 static void
3560 realify_fake_stores (void)
3562 unsigned int i;
3563 tree stmt;
3565 for (i = 0; VEC_iterate (tree, need_creation, i, stmt); i++)
3567 if (NECESSARY (stmt))
3569 block_stmt_iterator bsi;
3570 tree newstmt;
3572 /* Mark the temp variable as referenced */
3573 add_referenced_var (SSA_NAME_VAR (TREE_OPERAND (stmt, 0)));
3575 /* Put the new statement in GC memory, fix up the
3576 SSA_NAME_DEF_STMT on it, and then put it in place of
3577 the old statement before the store in the IR stream
3578 as a plain ssa name copy. */
3579 bsi = bsi_for_stmt (stmt);
3580 bsi_prev (&bsi);
3581 newstmt = build2 (MODIFY_EXPR, void_type_node,
3582 TREE_OPERAND (stmt, 0),
3583 TREE_OPERAND (bsi_stmt (bsi), 1));
3584 SSA_NAME_DEF_STMT (TREE_OPERAND (newstmt, 0)) = newstmt;
3585 bsi_insert_before (&bsi, newstmt, BSI_SAME_STMT);
3586 bsi = bsi_for_stmt (stmt);
3587 bsi_remove (&bsi, true);
3589 else
3590 release_defs (stmt);
3594 /* Tree-combine a value number expression *EXPR_P that does a type
3595 conversion with the value number expression of its operand.
3596 Returns true, if *EXPR_P simplifies to a value number or
3597 gimple min-invariant expression different from EXPR_P and
3598 sets *EXPR_P to the simplified expression value number.
3599 Otherwise returns false and does not change *EXPR_P. */
3601 static bool
3602 try_combine_conversion (tree *expr_p)
3604 tree expr = *expr_p;
3605 tree t;
3606 bitmap_set_t exprset;
3607 unsigned int firstbit;
3609 if (!((TREE_CODE (expr) == NOP_EXPR
3610 || TREE_CODE (expr) == CONVERT_EXPR)
3611 && TREE_CODE (TREE_OPERAND (expr, 0)) == VALUE_HANDLE
3612 && !VALUE_HANDLE_VUSES (TREE_OPERAND (expr, 0))))
3613 return false;
3615 exprset = VALUE_HANDLE_EXPR_SET (TREE_OPERAND (expr, 0));
3616 firstbit = bitmap_first_set_bit (exprset->expressions);
3617 t = fold_unary (TREE_CODE (expr), TREE_TYPE (expr),
3618 expression_for_id (firstbit));
3619 if (!t)
3620 return false;
3622 /* Strip useless type conversions, which is safe in the optimizers but
3623 not generally in fold. */
3624 STRIP_USELESS_TYPE_CONVERSION (t);
3626 /* Disallow value expressions we have no value number for already, as
3627 we would miss a leader for it here. */
3628 if (!(TREE_CODE (t) == VALUE_HANDLE
3629 || is_gimple_min_invariant (t)))
3630 t = vn_lookup (t, NULL);
3632 if (t && t != expr)
3634 *expr_p = t;
3635 return true;
3637 return false;
3640 /* Compute the AVAIL set for all basic blocks.
3642 This function performs value numbering of the statements in each basic
3643 block. The AVAIL sets are built from information we glean while doing
3644 this value numbering, since the AVAIL sets contain only one entry per
3645 value.
3647 AVAIL_IN[BLOCK] = AVAIL_OUT[dom(BLOCK)].
3648 AVAIL_OUT[BLOCK] = AVAIL_IN[BLOCK] U PHI_GEN[BLOCK] U TMP_GEN[BLOCK]. */
3650 static void
3651 compute_avail (void)
3653 basic_block block, son;
3654 basic_block *worklist;
3655 size_t sp = 0;
3656 tree param;
3657 /* For arguments with default definitions, we pretend they are
3658 defined in the entry block. */
3659 for (param = DECL_ARGUMENTS (current_function_decl);
3660 param;
3661 param = TREE_CHAIN (param))
3663 if (gimple_default_def (cfun, param) != NULL)
3665 tree def = gimple_default_def (cfun, param);
3667 vn_lookup_or_add (def, NULL);
3668 bitmap_insert_into_set (TMP_GEN (ENTRY_BLOCK_PTR), def);
3669 if (!in_fre)
3670 bitmap_value_insert_into_set (maximal_set, def);
3671 bitmap_value_insert_into_set (AVAIL_OUT (ENTRY_BLOCK_PTR), def);
3675 /* Likewise for the static chain decl. */
3676 if (cfun->static_chain_decl)
3678 param = cfun->static_chain_decl;
3679 if (gimple_default_def (cfun, param) != NULL)
3681 tree def = gimple_default_def (cfun, param);
3683 vn_lookup_or_add (def, NULL);
3684 bitmap_insert_into_set (TMP_GEN (ENTRY_BLOCK_PTR), def);
3685 if (!in_fre)
3686 bitmap_value_insert_into_set (maximal_set, def);
3687 bitmap_value_insert_into_set (AVAIL_OUT (ENTRY_BLOCK_PTR), def);
3691 /* Allocate the worklist. */
3692 worklist = XNEWVEC (basic_block, n_basic_blocks);
3694 /* Seed the algorithm by putting the dominator children of the entry
3695 block on the worklist. */
3696 for (son = first_dom_son (CDI_DOMINATORS, ENTRY_BLOCK_PTR);
3697 son;
3698 son = next_dom_son (CDI_DOMINATORS, son))
3699 worklist[sp++] = son;
3701 /* Loop until the worklist is empty. */
3702 while (sp)
3704 block_stmt_iterator bsi;
3705 tree stmt, phi;
3706 basic_block dom;
3707 unsigned int stmt_uid = 1;
3709 /* Pick a block from the worklist. */
3710 block = worklist[--sp];
3712 /* Initially, the set of available values in BLOCK is that of
3713 its immediate dominator. */
3714 dom = get_immediate_dominator (CDI_DOMINATORS, block);
3715 if (dom)
3716 bitmap_set_copy (AVAIL_OUT (block), AVAIL_OUT (dom));
3718 /* Generate values for PHI nodes. */
3719 for (phi = phi_nodes (block); phi; phi = PHI_CHAIN (phi))
3721 /* We have no need for virtual phis, as they don't represent
3722 actual computations. */
3723 if (is_gimple_reg (PHI_RESULT (phi)))
3725 add_to_sets (PHI_RESULT (phi), PHI_RESULT (phi), NULL,
3726 PHI_GEN (block), AVAIL_OUT (block));
3730 /* Now compute value numbers and populate value sets with all
3731 the expressions computed in BLOCK. */
3732 for (bsi = bsi_start (block); !bsi_end_p (bsi); bsi_next (&bsi))
3734 stmt_ann_t ann;
3735 ssa_op_iter iter;
3736 tree op;
3738 stmt = bsi_stmt (bsi);
3739 ann = stmt_ann (stmt);
3741 ann->uid = stmt_uid++;
3743 /* For regular value numbering, we are only interested in
3744 assignments of the form X_i = EXPR, where EXPR represents
3745 an "interesting" computation, it has no volatile operands
3746 and X_i doesn't flow through an abnormal edge. */
3747 if (TREE_CODE (stmt) == RETURN_EXPR
3748 && !ann->has_volatile_ops)
3750 tree realstmt = stmt;
3751 tree lhs;
3752 tree rhs;
3754 stmt = TREE_OPERAND (stmt, 0);
3755 if (stmt && TREE_CODE (stmt) == MODIFY_EXPR)
3757 lhs = TREE_OPERAND (stmt, 0);
3758 rhs = TREE_OPERAND (stmt, 1);
3759 if (TREE_CODE (rhs) == SSA_NAME
3760 && !is_undefined_value (rhs))
3761 bitmap_value_insert_into_set (EXP_GEN (block), rhs);
3763 FOR_EACH_SSA_TREE_OPERAND (op, realstmt, iter, SSA_OP_DEF)
3764 add_to_sets (op, op, NULL, TMP_GEN (block),
3765 AVAIL_OUT (block));
3767 continue;
3770 else if (TREE_CODE (stmt) == MODIFY_EXPR
3771 && !ann->has_volatile_ops
3772 && TREE_CODE (TREE_OPERAND (stmt, 0)) == SSA_NAME
3773 && !SSA_NAME_OCCURS_IN_ABNORMAL_PHI (TREE_OPERAND (stmt, 0)))
3775 tree lhs = TREE_OPERAND (stmt, 0);
3776 tree rhs = TREE_OPERAND (stmt, 1);
3778 /* Try to look through loads. */
3779 if (TREE_CODE (lhs) == SSA_NAME
3780 && !ZERO_SSA_OPERANDS (stmt, SSA_OP_VIRTUAL_USES)
3781 && try_look_through_load (lhs, rhs, stmt, block))
3782 continue;
3784 STRIP_USELESS_TYPE_CONVERSION (rhs);
3785 if (can_value_number_operation (rhs))
3787 /* For value numberable operation, create a
3788 duplicate expression with the operands replaced
3789 with the value handles of the original RHS. */
3790 tree newt = create_value_expr_from (rhs, block, stmt);
3791 if (newt)
3793 /* If we can combine a conversion expression
3794 with the expression for its operand just
3795 record the value number for it. */
3796 if (try_combine_conversion (&newt))
3797 vn_add (lhs, newt);
3798 else
3800 tree val = vn_lookup_or_add (newt, stmt);
3801 vn_add (lhs, val);
3802 if (!in_fre)
3803 bitmap_value_insert_into_set (maximal_set, newt);
3804 bitmap_value_insert_into_set (EXP_GEN (block), newt);
3806 bitmap_insert_into_set (TMP_GEN (block), lhs);
3807 bitmap_value_insert_into_set (AVAIL_OUT (block), lhs);
3808 continue;
3811 else if ((TREE_CODE (rhs) == SSA_NAME
3812 && !SSA_NAME_OCCURS_IN_ABNORMAL_PHI (rhs))
3813 || is_gimple_min_invariant (rhs)
3814 || TREE_CODE (rhs) == ADDR_EXPR
3815 || TREE_INVARIANT (rhs)
3816 || DECL_P (rhs))
3818 /* Compute a value number for the RHS of the statement
3819 and add its value to the AVAIL_OUT set for the block.
3820 Add the LHS to TMP_GEN. */
3821 add_to_sets (lhs, rhs, stmt, TMP_GEN (block),
3822 AVAIL_OUT (block));
3824 if (TREE_CODE (rhs) == SSA_NAME
3825 && !is_undefined_value (rhs))
3826 bitmap_value_insert_into_set (EXP_GEN (block), rhs);
3827 continue;
3831 /* For any other statement that we don't recognize, simply
3832 make the names generated by the statement available in
3833 AVAIL_OUT and TMP_GEN. */
3834 FOR_EACH_SSA_TREE_OPERAND (op, stmt, iter, SSA_OP_DEF)
3835 add_to_sets (op, op, NULL, TMP_GEN (block), AVAIL_OUT (block));
3837 FOR_EACH_SSA_TREE_OPERAND (op, stmt, iter, SSA_OP_USE)
3838 add_to_sets (op, op, NULL, NULL , AVAIL_OUT (block));
3841 /* Put the dominator children of BLOCK on the worklist of blocks
3842 to compute available sets for. */
3843 for (son = first_dom_son (CDI_DOMINATORS, block);
3844 son;
3845 son = next_dom_son (CDI_DOMINATORS, son))
3846 worklist[sp++] = son;
3849 free (worklist);
3853 /* Eliminate fully redundant computations. */
3855 static void
3856 eliminate (void)
3858 basic_block b;
3860 FOR_EACH_BB (b)
3862 block_stmt_iterator i;
3864 for (i = bsi_start (b); !bsi_end_p (i); bsi_next (&i))
3866 tree stmt = bsi_stmt (i);
3868 /* Lookup the RHS of the expression, see if we have an
3869 available computation for it. If so, replace the RHS with
3870 the available computation. */
3871 if (TREE_CODE (stmt) == MODIFY_EXPR
3872 && TREE_CODE (TREE_OPERAND (stmt, 0)) == SSA_NAME
3873 && TREE_CODE (TREE_OPERAND (stmt ,1)) != SSA_NAME
3874 && !is_gimple_min_invariant (TREE_OPERAND (stmt, 1))
3875 && !stmt_ann (stmt)->has_volatile_ops)
3877 tree lhs = TREE_OPERAND (stmt, 0);
3878 tree *rhs_p = &TREE_OPERAND (stmt, 1);
3879 tree sprime;
3881 sprime = bitmap_find_leader (AVAIL_OUT (b),
3882 vn_lookup (lhs, NULL));
3883 if (sprime
3884 && sprime != lhs
3885 && (TREE_CODE (*rhs_p) != SSA_NAME
3886 || may_propagate_copy (*rhs_p, sprime)))
3888 gcc_assert (sprime != *rhs_p);
3890 if (dump_file && (dump_flags & TDF_DETAILS))
3892 fprintf (dump_file, "Replaced ");
3893 print_generic_expr (dump_file, *rhs_p, 0);
3894 fprintf (dump_file, " with ");
3895 print_generic_expr (dump_file, sprime, 0);
3896 fprintf (dump_file, " in ");
3897 print_generic_stmt (dump_file, stmt, 0);
3900 if (TREE_CODE (sprime) == SSA_NAME)
3901 NECESSARY (SSA_NAME_DEF_STMT (sprime)) = 1;
3902 /* We need to make sure the new and old types actually match,
3903 which may require adding a simple cast, which fold_convert
3904 will do for us. */
3905 if (TREE_CODE (*rhs_p) != SSA_NAME
3906 && !tree_ssa_useless_type_conversion_1 (TREE_TYPE (*rhs_p),
3907 TREE_TYPE (sprime)))
3908 sprime = fold_convert (TREE_TYPE (*rhs_p), sprime);
3910 pre_stats.eliminations++;
3911 propagate_tree_value (rhs_p, sprime);
3912 update_stmt (stmt);
3914 /* If we removed EH side effects from the statement, clean
3915 its EH information. */
3916 if (maybe_clean_or_replace_eh_stmt (stmt, stmt))
3918 bitmap_set_bit (need_eh_cleanup,
3919 bb_for_stmt (stmt)->index);
3920 if (dump_file && (dump_flags & TDF_DETAILS))
3921 fprintf (dump_file, " Removed EH side effects.\n");
3929 /* Borrow a bit of tree-ssa-dce.c for the moment.
3930 XXX: In 4.1, we should be able to just run a DCE pass after PRE, though
3931 this may be a bit faster, and we may want critical edges kept split. */
3933 /* If OP's defining statement has not already been determined to be necessary,
3934 mark that statement necessary. Return the stmt, if it is newly
3935 necessary. */
3937 static inline tree
3938 mark_operand_necessary (tree op)
3940 tree stmt;
3942 gcc_assert (op);
3944 if (TREE_CODE (op) != SSA_NAME)
3945 return NULL;
3947 stmt = SSA_NAME_DEF_STMT (op);
3948 gcc_assert (stmt);
3950 if (NECESSARY (stmt)
3951 || IS_EMPTY_STMT (stmt))
3952 return NULL;
3954 NECESSARY (stmt) = 1;
3955 return stmt;
3958 /* Because we don't follow exactly the standard PRE algorithm, and decide not
3959 to insert PHI nodes sometimes, and because value numbering of casts isn't
3960 perfect, we sometimes end up inserting dead code. This simple DCE-like
3961 pass removes any insertions we made that weren't actually used. */
3963 static void
3964 remove_dead_inserted_code (void)
3966 VEC(tree,heap) *worklist = NULL;
3967 int i;
3968 tree t;
3970 worklist = VEC_alloc (tree, heap, VEC_length (tree, inserted_exprs));
3971 for (i = 0; VEC_iterate (tree, inserted_exprs, i, t); i++)
3973 if (NECESSARY (t))
3974 VEC_quick_push (tree, worklist, t);
3976 while (VEC_length (tree, worklist) > 0)
3978 t = VEC_pop (tree, worklist);
3980 /* PHI nodes are somewhat special in that each PHI alternative has
3981 data and control dependencies. All the statements feeding the
3982 PHI node's arguments are always necessary. */
3983 if (TREE_CODE (t) == PHI_NODE)
3985 int k;
3987 VEC_reserve (tree, heap, worklist, PHI_NUM_ARGS (t));
3988 for (k = 0; k < PHI_NUM_ARGS (t); k++)
3990 tree arg = PHI_ARG_DEF (t, k);
3991 if (TREE_CODE (arg) == SSA_NAME)
3993 arg = mark_operand_necessary (arg);
3994 if (arg)
3995 VEC_quick_push (tree, worklist, arg);
3999 else
4001 /* Propagate through the operands. Examine all the USE, VUSE and
4002 V_MAY_DEF operands in this statement. Mark all the statements
4003 which feed this statement's uses as necessary. */
4004 ssa_op_iter iter;
4005 tree use;
4007 /* The operands of V_MAY_DEF expressions are also needed as they
4008 represent potential definitions that may reach this
4009 statement (V_MAY_DEF operands allow us to follow def-def
4010 links). */
4012 FOR_EACH_SSA_TREE_OPERAND (use, t, iter, SSA_OP_ALL_USES)
4014 tree n = mark_operand_necessary (use);
4015 if (n)
4016 VEC_safe_push (tree, heap, worklist, n);
4021 for (i = 0; VEC_iterate (tree, inserted_exprs, i, t); i++)
4023 if (!NECESSARY (t))
4025 block_stmt_iterator bsi;
4027 if (dump_file && (dump_flags & TDF_DETAILS))
4029 fprintf (dump_file, "Removing unnecessary insertion:");
4030 print_generic_stmt (dump_file, t, 0);
4033 if (TREE_CODE (t) == PHI_NODE)
4035 remove_phi_node (t, NULL);
4037 else
4039 bsi = bsi_for_stmt (t);
4040 bsi_remove (&bsi, true);
4041 release_defs (t);
4045 VEC_free (tree, heap, worklist);
4048 /* Initialize data structures used by PRE. */
4050 static void
4051 init_pre (bool do_fre)
4053 basic_block bb;
4055 next_expression_id = 0;
4056 expressions = NULL;
4057 in_fre = do_fre;
4059 inserted_exprs = NULL;
4060 need_creation = NULL;
4061 pretemp = NULL_TREE;
4062 storetemp = NULL_TREE;
4063 mergephitemp = NULL_TREE;
4064 prephitemp = NULL_TREE;
4066 vn_init ();
4067 if (!do_fre)
4068 loop_optimizer_init (LOOPS_NORMAL);
4070 connect_infinite_loops_to_exit ();
4071 memset (&pre_stats, 0, sizeof (pre_stats));
4074 postorder = XNEWVEC (int, n_basic_blocks - NUM_FIXED_BLOCKS);
4075 post_order_compute (postorder, false);
4077 FOR_ALL_BB (bb)
4078 bb->aux = xcalloc (1, sizeof (struct bb_bitmap_sets));
4080 calculate_dominance_info (CDI_POST_DOMINATORS);
4081 calculate_dominance_info (CDI_DOMINATORS);
4083 bitmap_obstack_initialize (&grand_bitmap_obstack);
4084 phi_translate_table = htab_create (5110, expr_pred_trans_hash,
4085 expr_pred_trans_eq, free);
4086 bitmap_set_pool = create_alloc_pool ("Bitmap sets",
4087 sizeof (struct bitmap_set), 30);
4088 binary_node_pool = create_alloc_pool ("Binary tree nodes",
4089 tree_code_size (PLUS_EXPR), 30);
4090 unary_node_pool = create_alloc_pool ("Unary tree nodes",
4091 tree_code_size (NEGATE_EXPR), 30);
4092 reference_node_pool = create_alloc_pool ("Reference tree nodes",
4093 tree_code_size (ARRAY_REF), 30);
4094 expression_node_pool = create_alloc_pool ("Expression tree nodes",
4095 tree_code_size (CALL_EXPR), 30);
4096 list_node_pool = create_alloc_pool ("List tree nodes",
4097 tree_code_size (TREE_LIST), 30);
4098 comparison_node_pool = create_alloc_pool ("Comparison tree nodes",
4099 tree_code_size (EQ_EXPR), 30);
4100 modify_expr_node_pool = create_alloc_pool ("MODIFY_EXPR nodes",
4101 tree_code_size (MODIFY_EXPR),
4102 30);
4103 modify_expr_template = NULL;
4105 FOR_ALL_BB (bb)
4107 EXP_GEN (bb) = bitmap_set_new ();
4108 PHI_GEN (bb) = bitmap_set_new ();
4109 TMP_GEN (bb) = bitmap_set_new ();
4110 AVAIL_OUT (bb) = bitmap_set_new ();
4112 maximal_set = in_fre ? NULL : bitmap_set_new ();
4114 need_eh_cleanup = BITMAP_ALLOC (NULL);
4118 /* Deallocate data structures used by PRE. */
4120 static void
4121 fini_pre (bool do_fre)
4123 basic_block bb;
4124 unsigned int i;
4126 free (postorder);
4127 VEC_free (tree, heap, inserted_exprs);
4128 VEC_free (tree, heap, need_creation);
4129 bitmap_obstack_release (&grand_bitmap_obstack);
4130 free_alloc_pool (bitmap_set_pool);
4131 free_alloc_pool (binary_node_pool);
4132 free_alloc_pool (reference_node_pool);
4133 free_alloc_pool (unary_node_pool);
4134 free_alloc_pool (list_node_pool);
4135 free_alloc_pool (expression_node_pool);
4136 free_alloc_pool (comparison_node_pool);
4137 free_alloc_pool (modify_expr_node_pool);
4138 htab_delete (phi_translate_table);
4139 remove_fake_exit_edges ();
4141 FOR_ALL_BB (bb)
4143 free (bb->aux);
4144 bb->aux = NULL;
4147 free_dominance_info (CDI_POST_DOMINATORS);
4148 vn_delete ();
4150 if (!bitmap_empty_p (need_eh_cleanup))
4152 tree_purge_all_dead_eh_edges (need_eh_cleanup);
4153 cleanup_tree_cfg ();
4156 BITMAP_FREE (need_eh_cleanup);
4158 /* Wipe out pointers to VALUE_HANDLEs. In the not terribly distant
4159 future we will want them to be persistent though. */
4160 for (i = 0; i < num_ssa_names; i++)
4162 tree name = ssa_name (i);
4164 if (!name)
4165 continue;
4167 if (SSA_NAME_VALUE (name)
4168 && TREE_CODE (SSA_NAME_VALUE (name)) == VALUE_HANDLE)
4169 SSA_NAME_VALUE (name) = NULL;
4171 if (!do_fre && current_loops)
4172 loop_optimizer_finalize ();
4175 /* Main entry point to the SSA-PRE pass. DO_FRE is true if the caller
4176 only wants to do full redundancy elimination. */
4178 static void
4179 execute_pre (bool do_fre)
4182 do_partial_partial = optimize > 2;
4183 init_pre (do_fre);
4185 if (!do_fre)
4186 insert_fake_stores ();
4188 /* Collect and value number expressions computed in each basic block. */
4189 compute_avail ();
4191 if (dump_file && (dump_flags & TDF_DETAILS))
4193 basic_block bb;
4195 FOR_ALL_BB (bb)
4197 print_bitmap_set (dump_file, EXP_GEN (bb), "exp_gen", bb->index);
4198 print_bitmap_set (dump_file, TMP_GEN (bb), "tmp_gen",
4199 bb->index);
4200 print_bitmap_set (dump_file, AVAIL_OUT (bb), "avail_out",
4201 bb->index);
4205 /* Insert can get quite slow on an incredibly large number of basic
4206 blocks due to some quadratic behavior. Until this behavior is
4207 fixed, don't run it when he have an incredibly large number of
4208 bb's. If we aren't going to run insert, there is no point in
4209 computing ANTIC, either, even though it's plenty fast. */
4210 if (!do_fre && n_basic_blocks < 4000)
4212 vuse_names = XCNEWVEC (bitmap, num_ssa_names);
4213 compute_rvuse_and_antic_safe ();
4214 compute_antic ();
4215 insert ();
4216 free (vuse_names);
4219 /* Remove all the redundant expressions. */
4220 eliminate ();
4222 if (dump_file && (dump_flags & TDF_STATS))
4224 fprintf (dump_file, "Insertions: %d\n", pre_stats.insertions);
4225 fprintf (dump_file, "PA inserted: %d\n", pre_stats.pa_insert);
4226 fprintf (dump_file, "New PHIs: %d\n", pre_stats.phis);
4227 fprintf (dump_file, "Eliminated: %d\n", pre_stats.eliminations);
4228 fprintf (dump_file, "Constified: %d\n", pre_stats.constified);
4230 bsi_commit_edge_inserts ();
4232 clear_expression_ids ();
4233 if (!do_fre)
4235 remove_dead_inserted_code ();
4236 realify_fake_stores ();
4239 fini_pre (do_fre);
4242 /* Gate and execute functions for PRE. */
4244 static unsigned int
4245 do_pre (void)
4247 execute_pre (false);
4248 return 0;
4251 static bool
4252 gate_pre (void)
4254 return flag_tree_pre != 0;
4257 struct tree_opt_pass pass_pre =
4259 "pre", /* name */
4260 gate_pre, /* gate */
4261 do_pre, /* execute */
4262 NULL, /* sub */
4263 NULL, /* next */
4264 0, /* static_pass_number */
4265 TV_TREE_PRE, /* tv_id */
4266 PROP_no_crit_edges | PROP_cfg
4267 | PROP_ssa | PROP_alias, /* properties_required */
4268 0, /* properties_provided */
4269 0, /* properties_destroyed */
4270 0, /* todo_flags_start */
4271 TODO_update_ssa_only_virtuals | TODO_dump_func | TODO_ggc_collect
4272 | TODO_verify_ssa, /* todo_flags_finish */
4273 0 /* letter */
4277 /* Gate and execute functions for FRE. */
4279 static unsigned int
4280 execute_fre (void)
4282 execute_pre (true);
4283 return 0;
4286 static bool
4287 gate_fre (void)
4289 return flag_tree_fre != 0;
4292 struct tree_opt_pass pass_fre =
4294 "fre", /* name */
4295 gate_fre, /* gate */
4296 execute_fre, /* execute */
4297 NULL, /* sub */
4298 NULL, /* next */
4299 0, /* static_pass_number */
4300 TV_TREE_FRE, /* tv_id */
4301 PROP_cfg | PROP_ssa | PROP_alias, /* properties_required */
4302 0, /* properties_provided */
4303 0, /* properties_destroyed */
4304 0, /* todo_flags_start */
4305 TODO_dump_func | TODO_ggc_collect | TODO_verify_ssa, /* todo_flags_finish */
4306 0 /* letter */