Merge branch 'master' r216746-r217593 into gimple-classes-v2-option-3
[official-gcc.git] / gcc / tree-ssa-loop-ivcanon.c
blob23cf4c3e7ed0f2cee378480a275b912beae47891
1 /* Induction variable canonicalization and loop peeling.
2 Copyright (C) 2004-2014 Free Software Foundation, Inc.
4 This file is part of GCC.
6 GCC is free software; you can redistribute it and/or modify it
7 under the terms of the GNU General Public License as published by the
8 Free Software Foundation; either version 3, or (at your option) any
9 later version.
11 GCC is distributed in the hope that it will be useful, but WITHOUT
12 ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
13 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
14 for more details.
16 You should have received a copy of the GNU General Public License
17 along with GCC; see the file COPYING3. If not see
18 <http://www.gnu.org/licenses/>. */
20 /* This pass detects the loops that iterate a constant number of times,
21 adds a canonical induction variable (step -1, tested against 0)
22 and replaces the exit test. This enables the less powerful rtl
23 level analysis to use this information.
25 This might spoil the code in some cases (by increasing register pressure).
26 Note that in the case the new variable is not needed, ivopts will get rid
27 of it, so it might only be a problem when there are no other linear induction
28 variables. In that case the created optimization possibilities are likely
29 to pay up.
31 We also perform
32 - complete unrolling (or peeling) when the loops is rolling few enough
33 times
34 - simple peeling (i.e. copying few initial iterations prior the loop)
35 when number of iteration estimate is known (typically by the profile
36 info). */
38 #include "config.h"
39 #include "system.h"
40 #include "coretypes.h"
41 #include "tm.h"
42 #include "tree.h"
43 #include "tm_p.h"
44 #include "profile.h"
45 #include "predict.h"
46 #include "vec.h"
47 #include "hashtab.h"
48 #include "hash-set.h"
49 #include "machmode.h"
50 #include "hard-reg-set.h"
51 #include "input.h"
52 #include "function.h"
53 #include "dominance.h"
54 #include "cfg.h"
55 #include "basic-block.h"
56 #include "gimple-pretty-print.h"
57 #include "tree-ssa-alias.h"
58 #include "internal-fn.h"
59 #include "gimple-fold.h"
60 #include "tree-eh.h"
61 #include "gimple-expr.h"
62 #include "is-a.h"
63 #include "gimple.h"
64 #include "gimple-iterator.h"
65 #include "gimple-ssa.h"
66 #include "hash-map.h"
67 #include "plugin-api.h"
68 #include "ipa-ref.h"
69 #include "cgraph.h"
70 #include "tree-cfg.h"
71 #include "tree-phinodes.h"
72 #include "ssa-iterators.h"
73 #include "stringpool.h"
74 #include "tree-ssanames.h"
75 #include "tree-ssa-loop-manip.h"
76 #include "tree-ssa-loop-niter.h"
77 #include "tree-ssa-loop.h"
78 #include "tree-into-ssa.h"
79 #include "cfgloop.h"
80 #include "tree-pass.h"
81 #include "tree-chrec.h"
82 #include "tree-scalar-evolution.h"
83 #include "params.h"
84 #include "flags.h"
85 #include "tree-inline.h"
86 #include "target.h"
87 #include "tree-cfgcleanup.h"
88 #include "builtins.h"
90 /* Specifies types of loops that may be unrolled. */
92 enum unroll_level
94 UL_SINGLE_ITER, /* Only loops that exit immediately in the first
95 iteration. */
96 UL_NO_GROWTH, /* Only loops whose unrolling will not cause increase
97 of code size. */
98 UL_ALL /* All suitable loops. */
101 /* Adds a canonical induction variable to LOOP iterating NITER times. EXIT
102 is the exit edge whose condition is replaced. */
104 static void
105 create_canonical_iv (struct loop *loop, edge exit, tree niter)
107 edge in;
108 tree type, var;
109 gcond *cond;
110 gimple_stmt_iterator incr_at;
111 enum tree_code cmp;
113 if (dump_file && (dump_flags & TDF_DETAILS))
115 fprintf (dump_file, "Added canonical iv to loop %d, ", loop->num);
116 print_generic_expr (dump_file, niter, TDF_SLIM);
117 fprintf (dump_file, " iterations.\n");
120 cond = as_a <gcond *> (last_stmt (exit->src));
121 in = EDGE_SUCC (exit->src, 0);
122 if (in == exit)
123 in = EDGE_SUCC (exit->src, 1);
125 /* Note that we do not need to worry about overflows, since
126 type of niter is always unsigned and all comparisons are
127 just for equality/nonequality -- i.e. everything works
128 with a modulo arithmetics. */
130 type = TREE_TYPE (niter);
131 niter = fold_build2 (PLUS_EXPR, type,
132 niter,
133 build_int_cst (type, 1));
134 incr_at = gsi_last_bb (in->src);
135 create_iv (niter,
136 build_int_cst (type, -1),
137 NULL_TREE, loop,
138 &incr_at, false, NULL, &var);
140 cmp = (exit->flags & EDGE_TRUE_VALUE) ? EQ_EXPR : NE_EXPR;
141 gimple_cond_set_code (cond, cmp);
142 gimple_cond_set_lhs (cond, var);
143 gimple_cond_set_rhs (cond, build_int_cst (type, 0));
144 update_stmt (cond);
147 /* Describe size of loop as detected by tree_estimate_loop_size. */
148 struct loop_size
150 /* Number of instructions in the loop. */
151 int overall;
153 /* Number of instructions that will be likely optimized out in
154 peeled iterations of loop (i.e. computation based on induction
155 variable where induction variable starts at known constant.) */
156 int eliminated_by_peeling;
158 /* Same statistics for last iteration of loop: it is smaller because
159 instructions after exit are not executed. */
160 int last_iteration;
161 int last_iteration_eliminated_by_peeling;
163 /* If some IV computation will become constant. */
164 bool constant_iv;
166 /* Number of call stmts that are not a builtin and are pure or const
167 present on the hot path. */
168 int num_pure_calls_on_hot_path;
169 /* Number of call stmts that are not a builtin and are not pure nor const
170 present on the hot path. */
171 int num_non_pure_calls_on_hot_path;
172 /* Number of statements other than calls in the loop. */
173 int non_call_stmts_on_hot_path;
174 /* Number of branches seen on the hot path. */
175 int num_branches_on_hot_path;
178 /* Return true if OP in STMT will be constant after peeling LOOP. */
180 static bool
181 constant_after_peeling (tree op, gimple stmt, struct loop *loop)
183 affine_iv iv;
185 if (is_gimple_min_invariant (op))
186 return true;
188 /* We can still fold accesses to constant arrays when index is known. */
189 if (TREE_CODE (op) != SSA_NAME)
191 tree base = op;
193 /* First make fast look if we see constant array inside. */
194 while (handled_component_p (base))
195 base = TREE_OPERAND (base, 0);
196 if ((DECL_P (base)
197 && ctor_for_folding (base) != error_mark_node)
198 || CONSTANT_CLASS_P (base))
200 /* If so, see if we understand all the indices. */
201 base = op;
202 while (handled_component_p (base))
204 if (TREE_CODE (base) == ARRAY_REF
205 && !constant_after_peeling (TREE_OPERAND (base, 1), stmt, loop))
206 return false;
207 base = TREE_OPERAND (base, 0);
209 return true;
211 return false;
214 /* Induction variables are constants. */
215 if (!simple_iv (loop, loop_containing_stmt (stmt), op, &iv, false))
216 return false;
217 if (!is_gimple_min_invariant (iv.base))
218 return false;
219 if (!is_gimple_min_invariant (iv.step))
220 return false;
221 return true;
224 /* Computes an estimated number of insns in LOOP.
225 EXIT (if non-NULL) is an exite edge that will be eliminated in all but last
226 iteration of the loop.
227 EDGE_TO_CANCEL (if non-NULL) is an non-exit edge eliminated in the last iteration
228 of loop.
229 Return results in SIZE, estimate benefits for complete unrolling exiting by EXIT.
230 Stop estimating after UPPER_BOUND is met. Return true in this case. */
232 static bool
233 tree_estimate_loop_size (struct loop *loop, edge exit, edge edge_to_cancel, struct loop_size *size,
234 int upper_bound)
236 basic_block *body = get_loop_body (loop);
237 gimple_stmt_iterator gsi;
238 unsigned int i;
239 bool after_exit;
240 vec<basic_block> path = get_loop_hot_path (loop);
242 size->overall = 0;
243 size->eliminated_by_peeling = 0;
244 size->last_iteration = 0;
245 size->last_iteration_eliminated_by_peeling = 0;
246 size->num_pure_calls_on_hot_path = 0;
247 size->num_non_pure_calls_on_hot_path = 0;
248 size->non_call_stmts_on_hot_path = 0;
249 size->num_branches_on_hot_path = 0;
250 size->constant_iv = 0;
252 if (dump_file && (dump_flags & TDF_DETAILS))
253 fprintf (dump_file, "Estimating sizes for loop %i\n", loop->num);
254 for (i = 0; i < loop->num_nodes; i++)
256 if (edge_to_cancel && body[i] != edge_to_cancel->src
257 && dominated_by_p (CDI_DOMINATORS, body[i], edge_to_cancel->src))
258 after_exit = true;
259 else
260 after_exit = false;
261 if (dump_file && (dump_flags & TDF_DETAILS))
262 fprintf (dump_file, " BB: %i, after_exit: %i\n", body[i]->index, after_exit);
264 for (gsi = gsi_start_bb (body[i]); !gsi_end_p (gsi); gsi_next (&gsi))
266 gimple stmt = gsi_stmt (gsi);
267 int num = estimate_num_insns (stmt, &eni_size_weights);
268 bool likely_eliminated = false;
269 bool likely_eliminated_last = false;
270 bool likely_eliminated_peeled = false;
272 if (dump_file && (dump_flags & TDF_DETAILS))
274 fprintf (dump_file, " size: %3i ", num);
275 print_gimple_stmt (dump_file, gsi_stmt (gsi), 0, 0);
278 /* Look for reasons why we might optimize this stmt away. */
280 if (gimple_has_side_effects (stmt))
282 /* Exit conditional. */
283 else if (exit && body[i] == exit->src
284 && stmt == last_stmt (exit->src))
286 if (dump_file && (dump_flags & TDF_DETAILS))
287 fprintf (dump_file, " Exit condition will be eliminated "
288 "in peeled copies.\n");
289 likely_eliminated_peeled = true;
291 else if (edge_to_cancel && body[i] == edge_to_cancel->src
292 && stmt == last_stmt (edge_to_cancel->src))
294 if (dump_file && (dump_flags & TDF_DETAILS))
295 fprintf (dump_file, " Exit condition will be eliminated "
296 "in last copy.\n");
297 likely_eliminated_last = true;
299 /* Sets of IV variables */
300 else if (gimple_code (stmt) == GIMPLE_ASSIGN
301 && constant_after_peeling (gimple_assign_lhs (stmt), stmt, loop))
303 if (dump_file && (dump_flags & TDF_DETAILS))
304 fprintf (dump_file, " Induction variable computation will"
305 " be folded away.\n");
306 likely_eliminated = true;
308 /* Assignments of IV variables. */
309 else if (gimple_code (stmt) == GIMPLE_ASSIGN
310 && TREE_CODE (gimple_assign_lhs (stmt)) == SSA_NAME
311 && constant_after_peeling (gimple_assign_rhs1 (stmt), stmt, loop)
312 && (gimple_assign_rhs_class (stmt) != GIMPLE_BINARY_RHS
313 || constant_after_peeling (gimple_assign_rhs2 (stmt),
314 stmt, loop)))
316 size->constant_iv = true;
317 if (dump_file && (dump_flags & TDF_DETAILS))
318 fprintf (dump_file, " Constant expression will be folded away.\n");
319 likely_eliminated = true;
321 /* Conditionals. */
322 else if ((gimple_code (stmt) == GIMPLE_COND
323 && constant_after_peeling (gimple_cond_lhs (stmt), stmt, loop)
324 && constant_after_peeling (gimple_cond_rhs (stmt), stmt, loop))
325 || (gimple_code (stmt) == GIMPLE_SWITCH
326 && constant_after_peeling (gimple_switch_index (
327 as_a <gswitch *> (stmt)),
328 stmt, loop)))
330 if (dump_file && (dump_flags & TDF_DETAILS))
331 fprintf (dump_file, " Constant conditional.\n");
332 likely_eliminated = true;
335 size->overall += num;
336 if (likely_eliminated || likely_eliminated_peeled)
337 size->eliminated_by_peeling += num;
338 if (!after_exit)
340 size->last_iteration += num;
341 if (likely_eliminated || likely_eliminated_last)
342 size->last_iteration_eliminated_by_peeling += num;
344 if ((size->overall * 3 / 2 - size->eliminated_by_peeling
345 - size->last_iteration_eliminated_by_peeling) > upper_bound)
347 free (body);
348 path.release ();
349 return true;
353 while (path.length ())
355 basic_block bb = path.pop ();
356 for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
358 gimple stmt = gsi_stmt (gsi);
359 if (gimple_code (stmt) == GIMPLE_CALL)
361 int flags = gimple_call_flags (stmt);
362 tree decl = gimple_call_fndecl (stmt);
364 if (decl && DECL_IS_BUILTIN (decl)
365 && is_inexpensive_builtin (decl))
367 else if (flags & (ECF_PURE | ECF_CONST))
368 size->num_pure_calls_on_hot_path++;
369 else
370 size->num_non_pure_calls_on_hot_path++;
371 size->num_branches_on_hot_path ++;
373 else if (gimple_code (stmt) != GIMPLE_CALL
374 && gimple_code (stmt) != GIMPLE_DEBUG)
375 size->non_call_stmts_on_hot_path++;
376 if (((gimple_code (stmt) == GIMPLE_COND
377 && (!constant_after_peeling (gimple_cond_lhs (stmt), stmt, loop)
378 || constant_after_peeling (gimple_cond_rhs (stmt), stmt, loop)))
379 || (gimple_code (stmt) == GIMPLE_SWITCH
380 && !constant_after_peeling (gimple_switch_index (
381 as_a <gswitch *> (stmt)),
382 stmt, loop)))
383 && (!exit || bb != exit->src))
384 size->num_branches_on_hot_path++;
387 path.release ();
388 if (dump_file && (dump_flags & TDF_DETAILS))
389 fprintf (dump_file, "size: %i-%i, last_iteration: %i-%i\n", size->overall,
390 size->eliminated_by_peeling, size->last_iteration,
391 size->last_iteration_eliminated_by_peeling);
393 free (body);
394 return false;
397 /* Estimate number of insns of completely unrolled loop.
398 It is (NUNROLL + 1) * size of loop body with taking into account
399 the fact that in last copy everything after exit conditional
400 is dead and that some instructions will be eliminated after
401 peeling.
403 Loop body is likely going to simplify further, this is difficult
404 to guess, we just decrease the result by 1/3. */
406 static unsigned HOST_WIDE_INT
407 estimated_unrolled_size (struct loop_size *size,
408 unsigned HOST_WIDE_INT nunroll)
410 HOST_WIDE_INT unr_insns = ((nunroll)
411 * (HOST_WIDE_INT) (size->overall
412 - size->eliminated_by_peeling));
413 if (!nunroll)
414 unr_insns = 0;
415 unr_insns += size->last_iteration - size->last_iteration_eliminated_by_peeling;
417 unr_insns = unr_insns * 2 / 3;
418 if (unr_insns <= 0)
419 unr_insns = 1;
421 return unr_insns;
424 /* Loop LOOP is known to not loop. See if there is an edge in the loop
425 body that can be remove to make the loop to always exit and at
426 the same time it does not make any code potentially executed
427 during the last iteration dead.
429 After complete unrolling we still may get rid of the conditional
430 on the exit in the last copy even if we have no idea what it does.
431 This is quite common case for loops of form
433 int a[5];
434 for (i=0;i<b;i++)
435 a[i]=0;
437 Here we prove the loop to iterate 5 times but we do not know
438 it from induction variable.
440 For now we handle only simple case where there is exit condition
441 just before the latch block and the latch block contains no statements
442 with side effect that may otherwise terminate the execution of loop
443 (such as by EH or by terminating the program or longjmp).
445 In the general case we may want to cancel the paths leading to statements
446 loop-niter identified as having undefined effect in the last iteration.
447 The other cases are hopefully rare and will be cleaned up later. */
449 static edge
450 loop_edge_to_cancel (struct loop *loop)
452 vec<edge> exits;
453 unsigned i;
454 edge edge_to_cancel;
455 gimple_stmt_iterator gsi;
457 /* We want only one predecestor of the loop. */
458 if (EDGE_COUNT (loop->latch->preds) > 1)
459 return NULL;
461 exits = get_loop_exit_edges (loop);
463 FOR_EACH_VEC_ELT (exits, i, edge_to_cancel)
465 /* Find the other edge than the loop exit
466 leaving the conditoinal. */
467 if (EDGE_COUNT (edge_to_cancel->src->succs) != 2)
468 continue;
469 if (EDGE_SUCC (edge_to_cancel->src, 0) == edge_to_cancel)
470 edge_to_cancel = EDGE_SUCC (edge_to_cancel->src, 1);
471 else
472 edge_to_cancel = EDGE_SUCC (edge_to_cancel->src, 0);
474 /* We only can handle conditionals. */
475 if (!(edge_to_cancel->flags & (EDGE_TRUE_VALUE | EDGE_FALSE_VALUE)))
476 continue;
478 /* We should never have conditionals in the loop latch. */
479 gcc_assert (edge_to_cancel->dest != loop->header);
481 /* Check that it leads to loop latch. */
482 if (edge_to_cancel->dest != loop->latch)
483 continue;
485 exits.release ();
487 /* Verify that the code in loop latch does nothing that may end program
488 execution without really reaching the exit. This may include
489 non-pure/const function calls, EH statements, volatile ASMs etc. */
490 for (gsi = gsi_start_bb (loop->latch); !gsi_end_p (gsi); gsi_next (&gsi))
491 if (gimple_has_side_effects (gsi_stmt (gsi)))
492 return NULL;
493 return edge_to_cancel;
495 exits.release ();
496 return NULL;
499 /* Remove all tests for exits that are known to be taken after LOOP was
500 peeled NPEELED times. Put gcc_unreachable before every statement
501 known to not be executed. */
503 static bool
504 remove_exits_and_undefined_stmts (struct loop *loop, unsigned int npeeled)
506 struct nb_iter_bound *elt;
507 bool changed = false;
509 for (elt = loop->bounds; elt; elt = elt->next)
511 /* If statement is known to be undefined after peeling, turn it
512 into unreachable (or trap when debugging experience is supposed
513 to be good). */
514 if (!elt->is_exit
515 && wi::ltu_p (elt->bound, npeeled))
517 gimple_stmt_iterator gsi = gsi_for_stmt (elt->stmt);
518 gcall *stmt = gimple_build_call
519 (builtin_decl_implicit (BUILT_IN_UNREACHABLE), 0);
521 gimple_set_location (stmt, gimple_location (elt->stmt));
522 gsi_insert_before (&gsi, stmt, GSI_NEW_STMT);
523 changed = true;
524 if (dump_file && (dump_flags & TDF_DETAILS))
526 fprintf (dump_file, "Forced statement unreachable: ");
527 print_gimple_stmt (dump_file, elt->stmt, 0, 0);
530 /* If we know the exit will be taken after peeling, update. */
531 else if (elt->is_exit
532 && wi::leu_p (elt->bound, npeeled))
534 basic_block bb = gimple_bb (elt->stmt);
535 edge exit_edge = EDGE_SUCC (bb, 0);
537 if (dump_file && (dump_flags & TDF_DETAILS))
539 fprintf (dump_file, "Forced exit to be taken: ");
540 print_gimple_stmt (dump_file, elt->stmt, 0, 0);
542 if (!loop_exit_edge_p (loop, exit_edge))
543 exit_edge = EDGE_SUCC (bb, 1);
544 gcc_checking_assert (loop_exit_edge_p (loop, exit_edge));
545 gcond *cond_stmt = as_a <gcond *> (elt->stmt);
546 if (exit_edge->flags & EDGE_TRUE_VALUE)
547 gimple_cond_make_true (cond_stmt);
548 else
549 gimple_cond_make_false (cond_stmt);
550 update_stmt (cond_stmt);
551 changed = true;
554 return changed;
557 /* Remove all exits that are known to be never taken because of the loop bound
558 discovered. */
560 static bool
561 remove_redundant_iv_tests (struct loop *loop)
563 struct nb_iter_bound *elt;
564 bool changed = false;
566 if (!loop->any_upper_bound)
567 return false;
568 for (elt = loop->bounds; elt; elt = elt->next)
570 /* Exit is pointless if it won't be taken before loop reaches
571 upper bound. */
572 if (elt->is_exit && loop->any_upper_bound
573 && wi::ltu_p (loop->nb_iterations_upper_bound, elt->bound))
575 basic_block bb = gimple_bb (elt->stmt);
576 edge exit_edge = EDGE_SUCC (bb, 0);
577 struct tree_niter_desc niter;
579 if (!loop_exit_edge_p (loop, exit_edge))
580 exit_edge = EDGE_SUCC (bb, 1);
582 /* Only when we know the actual number of iterations, not
583 just a bound, we can remove the exit. */
584 if (!number_of_iterations_exit (loop, exit_edge,
585 &niter, false, false)
586 || !integer_onep (niter.assumptions)
587 || !integer_zerop (niter.may_be_zero)
588 || !niter.niter
589 || TREE_CODE (niter.niter) != INTEGER_CST
590 || !wi::ltu_p (loop->nb_iterations_upper_bound,
591 wi::to_widest (niter.niter)))
592 continue;
594 if (dump_file && (dump_flags & TDF_DETAILS))
596 fprintf (dump_file, "Removed pointless exit: ");
597 print_gimple_stmt (dump_file, elt->stmt, 0, 0);
599 gcond *cond_stmt = as_a <gcond *> (elt->stmt);
600 if (exit_edge->flags & EDGE_TRUE_VALUE)
601 gimple_cond_make_false (cond_stmt);
602 else
603 gimple_cond_make_true (cond_stmt);
604 update_stmt (cond_stmt);
605 changed = true;
608 return changed;
611 /* Stores loops that will be unlooped after we process whole loop tree. */
612 static vec<loop_p> loops_to_unloop;
613 static vec<int> loops_to_unloop_nunroll;
615 /* Cancel all fully unrolled loops by putting __builtin_unreachable
616 on the latch edge.
617 We do it after all unrolling since unlooping moves basic blocks
618 across loop boundaries trashing loop closed SSA form as well
619 as SCEV info needed to be intact during unrolling.
621 IRRED_INVALIDATED is used to bookkeep if information about
622 irreducible regions may become invalid as a result
623 of the transformation.
624 LOOP_CLOSED_SSA_INVALIDATED is used to bookkepp the case
625 when we need to go into loop closed SSA form. */
627 static void
628 unloop_loops (bitmap loop_closed_ssa_invalidated,
629 bool *irred_invalidated)
631 while (loops_to_unloop.length ())
633 struct loop *loop = loops_to_unloop.pop ();
634 int n_unroll = loops_to_unloop_nunroll.pop ();
635 basic_block latch = loop->latch;
636 edge latch_edge = loop_latch_edge (loop);
637 int flags = latch_edge->flags;
638 location_t locus = latch_edge->goto_locus;
639 gcall *stmt;
640 gimple_stmt_iterator gsi;
642 remove_exits_and_undefined_stmts (loop, n_unroll);
644 /* Unloop destroys the latch edge. */
645 unloop (loop, irred_invalidated, loop_closed_ssa_invalidated);
647 /* Create new basic block for the latch edge destination and wire
648 it in. */
649 stmt = gimple_build_call (builtin_decl_implicit (BUILT_IN_UNREACHABLE), 0);
650 latch_edge = make_edge (latch, create_basic_block (NULL, NULL, latch), flags);
651 latch_edge->probability = 0;
652 latch_edge->count = 0;
653 latch_edge->flags |= flags;
654 latch_edge->goto_locus = locus;
656 latch_edge->dest->loop_father = current_loops->tree_root;
657 latch_edge->dest->count = 0;
658 latch_edge->dest->frequency = 0;
659 set_immediate_dominator (CDI_DOMINATORS, latch_edge->dest, latch_edge->src);
661 gsi = gsi_start_bb (latch_edge->dest);
662 gsi_insert_after (&gsi, stmt, GSI_NEW_STMT);
664 loops_to_unloop.release ();
665 loops_to_unloop_nunroll.release ();
668 /* Tries to unroll LOOP completely, i.e. NITER times.
669 UL determines which loops we are allowed to unroll.
670 EXIT is the exit of the loop that should be eliminated.
671 MAXITER specfy bound on number of iterations, -1 if it is
672 not known or too large for HOST_WIDE_INT. The location
673 LOCUS corresponding to the loop is used when emitting
674 a summary of the unroll to the dump file. */
676 static bool
677 try_unroll_loop_completely (struct loop *loop,
678 edge exit, tree niter,
679 enum unroll_level ul,
680 HOST_WIDE_INT maxiter,
681 location_t locus)
683 unsigned HOST_WIDE_INT n_unroll = 0, ninsns, max_unroll, unr_insns;
684 struct loop_size size;
685 bool n_unroll_found = false;
686 edge edge_to_cancel = NULL;
687 int report_flags = MSG_OPTIMIZED_LOCATIONS | TDF_RTL | TDF_DETAILS;
689 /* See if we proved number of iterations to be low constant.
691 EXIT is an edge that will be removed in all but last iteration of
692 the loop.
694 EDGE_TO_CACNEL is an edge that will be removed from the last iteration
695 of the unrolled sequence and is expected to make the final loop not
696 rolling.
698 If the number of execution of loop is determined by standard induction
699 variable test, then EXIT and EDGE_TO_CANCEL are the two edges leaving
700 from the iv test. */
701 if (tree_fits_uhwi_p (niter))
703 n_unroll = tree_to_uhwi (niter);
704 n_unroll_found = true;
705 edge_to_cancel = EDGE_SUCC (exit->src, 0);
706 if (edge_to_cancel == exit)
707 edge_to_cancel = EDGE_SUCC (exit->src, 1);
709 /* We do not know the number of iterations and thus we can not eliminate
710 the EXIT edge. */
711 else
712 exit = NULL;
714 /* See if we can improve our estimate by using recorded loop bounds. */
715 if (maxiter >= 0
716 && (!n_unroll_found || (unsigned HOST_WIDE_INT)maxiter < n_unroll))
718 n_unroll = maxiter;
719 n_unroll_found = true;
720 /* Loop terminates before the IV variable test, so we can not
721 remove it in the last iteration. */
722 edge_to_cancel = NULL;
725 if (!n_unroll_found)
726 return false;
728 max_unroll = PARAM_VALUE (PARAM_MAX_COMPLETELY_PEEL_TIMES);
729 if (n_unroll > max_unroll)
730 return false;
732 if (!edge_to_cancel)
733 edge_to_cancel = loop_edge_to_cancel (loop);
735 if (n_unroll)
737 sbitmap wont_exit;
738 edge e;
739 unsigned i;
740 bool large;
741 vec<edge> to_remove = vNULL;
742 if (ul == UL_SINGLE_ITER)
743 return false;
745 large = tree_estimate_loop_size
746 (loop, exit, edge_to_cancel, &size,
747 PARAM_VALUE (PARAM_MAX_COMPLETELY_PEELED_INSNS));
748 ninsns = size.overall;
749 if (large)
751 if (dump_file && (dump_flags & TDF_DETAILS))
752 fprintf (dump_file, "Not unrolling loop %d: it is too large.\n",
753 loop->num);
754 return false;
757 unr_insns = estimated_unrolled_size (&size, n_unroll);
758 if (dump_file && (dump_flags & TDF_DETAILS))
760 fprintf (dump_file, " Loop size: %d\n", (int) ninsns);
761 fprintf (dump_file, " Estimated size after unrolling: %d\n",
762 (int) unr_insns);
765 /* If the code is going to shrink, we don't need to be extra cautious
766 on guessing if the unrolling is going to be profitable. */
767 if (unr_insns
768 /* If there is IV variable that will become constant, we save
769 one instruction in the loop prologue we do not account
770 otherwise. */
771 <= ninsns + (size.constant_iv != false))
773 /* We unroll only inner loops, because we do not consider it profitable
774 otheriwse. We still can cancel loopback edge of not rolling loop;
775 this is always a good idea. */
776 else if (ul == UL_NO_GROWTH)
778 if (dump_file && (dump_flags & TDF_DETAILS))
779 fprintf (dump_file, "Not unrolling loop %d: size would grow.\n",
780 loop->num);
781 return false;
783 /* Outer loops tend to be less interesting candidates for complete
784 unrolling unless we can do a lot of propagation into the inner loop
785 body. For now we disable outer loop unrolling when the code would
786 grow. */
787 else if (loop->inner)
789 if (dump_file && (dump_flags & TDF_DETAILS))
790 fprintf (dump_file, "Not unrolling loop %d: "
791 "it is not innermost and code would grow.\n",
792 loop->num);
793 return false;
795 /* If there is call on a hot path through the loop, then
796 there is most probably not much to optimize. */
797 else if (size.num_non_pure_calls_on_hot_path)
799 if (dump_file && (dump_flags & TDF_DETAILS))
800 fprintf (dump_file, "Not unrolling loop %d: "
801 "contains call and code would grow.\n",
802 loop->num);
803 return false;
805 /* If there is pure/const call in the function, then we
806 can still optimize the unrolled loop body if it contains
807 some other interesting code than the calls and code
808 storing or cumulating the return value. */
809 else if (size.num_pure_calls_on_hot_path
810 /* One IV increment, one test, one ivtmp store
811 and one useful stmt. That is about minimal loop
812 doing pure call. */
813 && (size.non_call_stmts_on_hot_path
814 <= 3 + size.num_pure_calls_on_hot_path))
816 if (dump_file && (dump_flags & TDF_DETAILS))
817 fprintf (dump_file, "Not unrolling loop %d: "
818 "contains just pure calls and code would grow.\n",
819 loop->num);
820 return false;
822 /* Complette unrolling is major win when control flow is removed and
823 one big basic block is created. If the loop contains control flow
824 the optimization may still be a win because of eliminating the loop
825 overhead but it also may blow the branch predictor tables.
826 Limit number of branches on the hot path through the peeled
827 sequence. */
828 else if (size.num_branches_on_hot_path * (int)n_unroll
829 > PARAM_VALUE (PARAM_MAX_PEEL_BRANCHES))
831 if (dump_file && (dump_flags & TDF_DETAILS))
832 fprintf (dump_file, "Not unrolling loop %d: "
833 " number of branches on hot path in the unrolled sequence"
834 " reach --param max-peel-branches limit.\n",
835 loop->num);
836 return false;
838 else if (unr_insns
839 > (unsigned) PARAM_VALUE (PARAM_MAX_COMPLETELY_PEELED_INSNS))
841 if (dump_file && (dump_flags & TDF_DETAILS))
842 fprintf (dump_file, "Not unrolling loop %d: "
843 "(--param max-completely-peeled-insns limit reached).\n",
844 loop->num);
845 return false;
847 dump_printf_loc (report_flags, locus,
848 "loop turned into non-loop; it never loops.\n");
850 initialize_original_copy_tables ();
851 wont_exit = sbitmap_alloc (n_unroll + 1);
852 bitmap_ones (wont_exit);
853 bitmap_clear_bit (wont_exit, 0);
855 if (!gimple_duplicate_loop_to_header_edge (loop, loop_preheader_edge (loop),
856 n_unroll, wont_exit,
857 exit, &to_remove,
858 DLTHE_FLAG_UPDATE_FREQ
859 | DLTHE_FLAG_COMPLETTE_PEEL))
861 free_original_copy_tables ();
862 free (wont_exit);
863 if (dump_file && (dump_flags & TDF_DETAILS))
864 fprintf (dump_file, "Failed to duplicate the loop\n");
865 return false;
868 FOR_EACH_VEC_ELT (to_remove, i, e)
870 bool ok = remove_path (e);
871 gcc_assert (ok);
874 to_remove.release ();
875 free (wont_exit);
876 free_original_copy_tables ();
880 /* Remove the conditional from the last copy of the loop. */
881 if (edge_to_cancel)
883 gcond *cond = as_a <gcond *> (last_stmt (edge_to_cancel->src));
884 if (edge_to_cancel->flags & EDGE_TRUE_VALUE)
885 gimple_cond_make_false (cond);
886 else
887 gimple_cond_make_true (cond);
888 update_stmt (cond);
889 /* Do not remove the path. Doing so may remove outer loop
890 and confuse bookkeeping code in tree_unroll_loops_completelly. */
893 /* Store the loop for later unlooping and exit removal. */
894 loops_to_unloop.safe_push (loop);
895 loops_to_unloop_nunroll.safe_push (n_unroll);
897 if (dump_enabled_p ())
899 if (!n_unroll)
900 dump_printf_loc (MSG_OPTIMIZED_LOCATIONS | TDF_DETAILS, locus,
901 "loop turned into non-loop; it never loops\n");
902 else
904 dump_printf_loc (MSG_OPTIMIZED_LOCATIONS | TDF_DETAILS, locus,
905 "loop with %d iterations completely unrolled",
906 (int) (n_unroll + 1));
907 if (profile_info)
908 dump_printf (MSG_OPTIMIZED_LOCATIONS | TDF_DETAILS,
909 " (header execution count %d)",
910 (int)loop->header->count);
911 dump_printf (MSG_OPTIMIZED_LOCATIONS | TDF_DETAILS, "\n");
915 if (dump_file && (dump_flags & TDF_DETAILS))
917 if (exit)
918 fprintf (dump_file, "Exit condition of peeled iterations was "
919 "eliminated.\n");
920 if (edge_to_cancel)
921 fprintf (dump_file, "Last iteration exit edge was proved true.\n");
922 else
923 fprintf (dump_file, "Latch of last iteration was marked by "
924 "__builtin_unreachable ().\n");
927 return true;
930 /* Return number of instructions after peeling. */
931 static unsigned HOST_WIDE_INT
932 estimated_peeled_sequence_size (struct loop_size *size,
933 unsigned HOST_WIDE_INT npeel)
935 return MAX (npeel * (HOST_WIDE_INT) (size->overall
936 - size->eliminated_by_peeling), 1);
939 /* If the loop is expected to iterate N times and is
940 small enough, duplicate the loop body N+1 times before
941 the loop itself. This way the hot path will never
942 enter the loop.
943 Parameters are the same as for try_unroll_loops_completely */
945 static bool
946 try_peel_loop (struct loop *loop,
947 edge exit, tree niter,
948 HOST_WIDE_INT maxiter)
950 int npeel;
951 struct loop_size size;
952 int peeled_size;
953 sbitmap wont_exit;
954 unsigned i;
955 vec<edge> to_remove = vNULL;
956 edge e;
958 /* If the iteration bound is known and large, then we can safely eliminate
959 the check in peeled copies. */
960 if (TREE_CODE (niter) != INTEGER_CST)
961 exit = NULL;
963 if (!flag_peel_loops || PARAM_VALUE (PARAM_MAX_PEEL_TIMES) <= 0)
964 return false;
966 /* Peel only innermost loops. */
967 if (loop->inner)
969 if (dump_file)
970 fprintf (dump_file, "Not peeling: outer loop\n");
971 return false;
974 if (!optimize_loop_for_speed_p (loop))
976 if (dump_file)
977 fprintf (dump_file, "Not peeling: cold loop\n");
978 return false;
981 /* Check if there is an estimate on the number of iterations. */
982 npeel = estimated_loop_iterations_int (loop);
983 if (npeel < 0)
985 if (dump_file)
986 fprintf (dump_file, "Not peeling: number of iterations is not "
987 "estimated\n");
988 return false;
990 if (maxiter >= 0 && maxiter <= npeel)
992 if (dump_file)
993 fprintf (dump_file, "Not peeling: upper bound is known so can "
994 "unroll completely\n");
995 return false;
998 /* We want to peel estimated number of iterations + 1 (so we never
999 enter the loop on quick path). Check against PARAM_MAX_PEEL_TIMES
1000 and be sure to avoid overflows. */
1001 if (npeel > PARAM_VALUE (PARAM_MAX_PEEL_TIMES) - 1)
1003 if (dump_file)
1004 fprintf (dump_file, "Not peeling: rolls too much "
1005 "(%i + 1 > --param max-peel-times)\n", npeel);
1006 return false;
1008 npeel++;
1010 /* Check peeled loops size. */
1011 tree_estimate_loop_size (loop, exit, NULL, &size,
1012 PARAM_VALUE (PARAM_MAX_PEELED_INSNS));
1013 if ((peeled_size = estimated_peeled_sequence_size (&size, npeel))
1014 > PARAM_VALUE (PARAM_MAX_PEELED_INSNS))
1016 if (dump_file)
1017 fprintf (dump_file, "Not peeling: peeled sequence size is too large "
1018 "(%i insns > --param max-peel-insns)", peeled_size);
1019 return false;
1022 /* Duplicate possibly eliminating the exits. */
1023 initialize_original_copy_tables ();
1024 wont_exit = sbitmap_alloc (npeel + 1);
1025 bitmap_ones (wont_exit);
1026 bitmap_clear_bit (wont_exit, 0);
1027 if (!gimple_duplicate_loop_to_header_edge (loop, loop_preheader_edge (loop),
1028 npeel, wont_exit,
1029 exit, &to_remove,
1030 DLTHE_FLAG_UPDATE_FREQ
1031 | DLTHE_FLAG_COMPLETTE_PEEL))
1033 free_original_copy_tables ();
1034 free (wont_exit);
1035 return false;
1037 FOR_EACH_VEC_ELT (to_remove, i, e)
1039 bool ok = remove_path (e);
1040 gcc_assert (ok);
1042 free (wont_exit);
1043 free_original_copy_tables ();
1044 if (dump_file && (dump_flags & TDF_DETAILS))
1046 fprintf (dump_file, "Peeled loop %d, %i times.\n",
1047 loop->num, npeel);
1049 if (loop->any_upper_bound)
1050 loop->nb_iterations_upper_bound -= npeel;
1051 loop->nb_iterations_estimate = 0;
1052 /* Make sure to mark loop cold so we do not try to peel it more. */
1053 scale_loop_profile (loop, 1, 0);
1054 loop->header->count = 0;
1055 return true;
1057 /* Adds a canonical induction variable to LOOP if suitable.
1058 CREATE_IV is true if we may create a new iv. UL determines
1059 which loops we are allowed to completely unroll. If TRY_EVAL is true, we try
1060 to determine the number of iterations of a loop by direct evaluation.
1061 Returns true if cfg is changed. */
1063 static bool
1064 canonicalize_loop_induction_variables (struct loop *loop,
1065 bool create_iv, enum unroll_level ul,
1066 bool try_eval)
1068 edge exit = NULL;
1069 tree niter;
1070 HOST_WIDE_INT maxiter;
1071 bool modified = false;
1072 location_t locus = UNKNOWN_LOCATION;
1074 niter = number_of_latch_executions (loop);
1075 exit = single_exit (loop);
1076 if (TREE_CODE (niter) == INTEGER_CST)
1077 locus = gimple_location (last_stmt (exit->src));
1078 else
1080 /* If the loop has more than one exit, try checking all of them
1081 for # of iterations determinable through scev. */
1082 if (!exit)
1083 niter = find_loop_niter (loop, &exit);
1085 /* Finally if everything else fails, try brute force evaluation. */
1086 if (try_eval
1087 && (chrec_contains_undetermined (niter)
1088 || TREE_CODE (niter) != INTEGER_CST))
1089 niter = find_loop_niter_by_eval (loop, &exit);
1091 if (exit)
1092 locus = gimple_location (last_stmt (exit->src));
1094 if (TREE_CODE (niter) != INTEGER_CST)
1095 exit = NULL;
1098 /* We work exceptionally hard here to estimate the bound
1099 by find_loop_niter_by_eval. Be sure to keep it for future. */
1100 if (niter && TREE_CODE (niter) == INTEGER_CST)
1102 record_niter_bound (loop, wi::to_widest (niter),
1103 exit == single_likely_exit (loop), true);
1106 /* Force re-computation of loop bounds so we can remove redundant exits. */
1107 maxiter = max_loop_iterations_int (loop);
1109 if (dump_file && (dump_flags & TDF_DETAILS)
1110 && TREE_CODE (niter) == INTEGER_CST)
1112 fprintf (dump_file, "Loop %d iterates ", loop->num);
1113 print_generic_expr (dump_file, niter, TDF_SLIM);
1114 fprintf (dump_file, " times.\n");
1116 if (dump_file && (dump_flags & TDF_DETAILS)
1117 && maxiter >= 0)
1119 fprintf (dump_file, "Loop %d iterates at most %i times.\n", loop->num,
1120 (int)maxiter);
1123 /* Remove exits that are known to be never taken based on loop bound.
1124 Needs to be called after compilation of max_loop_iterations_int that
1125 populates the loop bounds. */
1126 modified |= remove_redundant_iv_tests (loop);
1128 if (try_unroll_loop_completely (loop, exit, niter, ul, maxiter, locus))
1129 return true;
1131 if (create_iv
1132 && niter && !chrec_contains_undetermined (niter)
1133 && exit && just_once_each_iteration_p (loop, exit->src))
1134 create_canonical_iv (loop, exit, niter);
1136 if (ul == UL_ALL)
1137 modified |= try_peel_loop (loop, exit, niter, maxiter);
1139 return modified;
1142 /* The main entry point of the pass. Adds canonical induction variables
1143 to the suitable loops. */
1145 unsigned int
1146 canonicalize_induction_variables (void)
1148 struct loop *loop;
1149 bool changed = false;
1150 bool irred_invalidated = false;
1151 bitmap loop_closed_ssa_invalidated = BITMAP_ALLOC (NULL);
1153 free_numbers_of_iterations_estimates ();
1154 estimate_numbers_of_iterations ();
1156 FOR_EACH_LOOP (loop, LI_FROM_INNERMOST)
1158 changed |= canonicalize_loop_induction_variables (loop,
1159 true, UL_SINGLE_ITER,
1160 true);
1162 gcc_assert (!need_ssa_update_p (cfun));
1164 unloop_loops (loop_closed_ssa_invalidated, &irred_invalidated);
1165 if (irred_invalidated
1166 && loops_state_satisfies_p (LOOPS_HAVE_MARKED_IRREDUCIBLE_REGIONS))
1167 mark_irreducible_loops ();
1169 /* Clean up the information about numbers of iterations, since brute force
1170 evaluation could reveal new information. */
1171 scev_reset ();
1173 if (!bitmap_empty_p (loop_closed_ssa_invalidated))
1175 gcc_checking_assert (loops_state_satisfies_p (LOOP_CLOSED_SSA));
1176 rewrite_into_loop_closed_ssa (NULL, TODO_update_ssa);
1178 BITMAP_FREE (loop_closed_ssa_invalidated);
1180 if (changed)
1181 return TODO_cleanup_cfg;
1182 return 0;
1185 /* Propagate VAL into all uses of SSA_NAME. */
1187 static void
1188 propagate_into_all_uses (tree ssa_name, tree val)
1190 imm_use_iterator iter;
1191 gimple use_stmt;
1193 FOR_EACH_IMM_USE_STMT (use_stmt, iter, ssa_name)
1195 gimple_stmt_iterator use_stmt_gsi = gsi_for_stmt (use_stmt);
1196 use_operand_p use;
1198 FOR_EACH_IMM_USE_ON_STMT (use, iter)
1199 SET_USE (use, val);
1201 if (is_gimple_assign (use_stmt)
1202 && get_gimple_rhs_class (gimple_assign_rhs_code (use_stmt))
1203 == GIMPLE_SINGLE_RHS)
1205 tree rhs = gimple_assign_rhs1 (use_stmt);
1207 if (TREE_CODE (rhs) == ADDR_EXPR)
1208 recompute_tree_invariant_for_addr_expr (rhs);
1211 fold_stmt_inplace (&use_stmt_gsi);
1212 update_stmt (use_stmt);
1213 maybe_clean_or_replace_eh_stmt (use_stmt, use_stmt);
1217 /* Propagate constant SSA_NAMEs defined in basic block BB. */
1219 static void
1220 propagate_constants_for_unrolling (basic_block bb)
1222 /* Look for degenerate PHI nodes with constant argument. */
1223 for (gphi_iterator gsi = gsi_start_phis (bb); !gsi_end_p (gsi); )
1225 gphi *phi = gsi.phi ();
1226 tree result = gimple_phi_result (phi);
1227 tree arg = gimple_phi_arg_def (phi, 0);
1229 if (gimple_phi_num_args (phi) == 1 && TREE_CODE (arg) == INTEGER_CST)
1231 propagate_into_all_uses (result, arg);
1232 gsi_remove (&gsi, true);
1233 release_ssa_name (result);
1235 else
1236 gsi_next (&gsi);
1239 /* Look for assignments to SSA names with constant RHS. */
1240 for (gimple_stmt_iterator gsi = gsi_start_bb (bb); !gsi_end_p (gsi); )
1242 gimple stmt = gsi_stmt (gsi);
1243 tree lhs;
1245 if (is_gimple_assign (stmt)
1246 && gimple_assign_rhs_code (stmt) == INTEGER_CST
1247 && (lhs = gimple_assign_lhs (stmt), TREE_CODE (lhs) == SSA_NAME)
1248 && !SSA_NAME_OCCURS_IN_ABNORMAL_PHI (lhs))
1250 propagate_into_all_uses (lhs, gimple_assign_rhs1 (stmt));
1251 gsi_remove (&gsi, true);
1252 release_ssa_name (lhs);
1254 else
1255 gsi_next (&gsi);
1259 /* Process loops from innermost to outer, stopping at the innermost
1260 loop we unrolled. */
1262 static bool
1263 tree_unroll_loops_completely_1 (bool may_increase_size, bool unroll_outer,
1264 vec<loop_p, va_heap>& father_stack,
1265 struct loop *loop)
1267 struct loop *loop_father;
1268 bool changed = false;
1269 struct loop *inner;
1270 enum unroll_level ul;
1272 /* Process inner loops first. */
1273 for (inner = loop->inner; inner != NULL; inner = inner->next)
1274 changed |= tree_unroll_loops_completely_1 (may_increase_size,
1275 unroll_outer, father_stack,
1276 inner);
1278 /* If we changed an inner loop we cannot process outer loops in this
1279 iteration because SSA form is not up-to-date. Continue with
1280 siblings of outer loops instead. */
1281 if (changed)
1282 return true;
1284 /* Don't unroll #pragma omp simd loops until the vectorizer
1285 attempts to vectorize those. */
1286 if (loop->force_vectorize)
1287 return false;
1289 /* Try to unroll this loop. */
1290 loop_father = loop_outer (loop);
1291 if (!loop_father)
1292 return false;
1294 if (may_increase_size && optimize_loop_nest_for_speed_p (loop)
1295 /* Unroll outermost loops only if asked to do so or they do
1296 not cause code growth. */
1297 && (unroll_outer || loop_outer (loop_father)))
1298 ul = UL_ALL;
1299 else
1300 ul = UL_NO_GROWTH;
1302 if (canonicalize_loop_induction_variables
1303 (loop, false, ul, !flag_tree_loop_ivcanon))
1305 /* If we'll continue unrolling, we need to propagate constants
1306 within the new basic blocks to fold away induction variable
1307 computations; otherwise, the size might blow up before the
1308 iteration is complete and the IR eventually cleaned up. */
1309 if (loop_outer (loop_father) && !loop_father->aux)
1311 father_stack.safe_push (loop_father);
1312 loop_father->aux = loop_father;
1315 return true;
1318 return false;
1321 /* Unroll LOOPS completely if they iterate just few times. Unless
1322 MAY_INCREASE_SIZE is true, perform the unrolling only if the
1323 size of the code does not increase. */
1325 unsigned int
1326 tree_unroll_loops_completely (bool may_increase_size, bool unroll_outer)
1328 auto_vec<loop_p, 16> father_stack;
1329 bool changed;
1330 int iteration = 0;
1331 bool irred_invalidated = false;
1335 changed = false;
1336 bitmap loop_closed_ssa_invalidated = NULL;
1338 if (loops_state_satisfies_p (LOOP_CLOSED_SSA))
1339 loop_closed_ssa_invalidated = BITMAP_ALLOC (NULL);
1341 free_numbers_of_iterations_estimates ();
1342 estimate_numbers_of_iterations ();
1344 changed = tree_unroll_loops_completely_1 (may_increase_size,
1345 unroll_outer, father_stack,
1346 current_loops->tree_root);
1347 if (changed)
1349 struct loop **iter;
1350 unsigned i;
1352 /* Be sure to skip unlooped loops while procesing father_stack
1353 array. */
1354 FOR_EACH_VEC_ELT (loops_to_unloop, i, iter)
1355 (*iter)->aux = NULL;
1356 FOR_EACH_VEC_ELT (father_stack, i, iter)
1357 if (!(*iter)->aux)
1358 *iter = NULL;
1359 unloop_loops (loop_closed_ssa_invalidated, &irred_invalidated);
1361 /* We can not use TODO_update_ssa_no_phi because VOPS gets confused. */
1362 if (loop_closed_ssa_invalidated
1363 && !bitmap_empty_p (loop_closed_ssa_invalidated))
1364 rewrite_into_loop_closed_ssa (loop_closed_ssa_invalidated,
1365 TODO_update_ssa);
1366 else
1367 update_ssa (TODO_update_ssa);
1369 /* Propagate the constants within the new basic blocks. */
1370 FOR_EACH_VEC_ELT (father_stack, i, iter)
1371 if (*iter)
1373 unsigned j;
1374 basic_block *body = get_loop_body_in_dom_order (*iter);
1375 for (j = 0; j < (*iter)->num_nodes; j++)
1376 propagate_constants_for_unrolling (body[j]);
1377 free (body);
1378 (*iter)->aux = NULL;
1380 father_stack.truncate (0);
1382 /* This will take care of removing completely unrolled loops
1383 from the loop structures so we can continue unrolling now
1384 innermost loops. */
1385 if (cleanup_tree_cfg ())
1386 update_ssa (TODO_update_ssa_only_virtuals);
1388 /* Clean up the information about numbers of iterations, since
1389 complete unrolling might have invalidated it. */
1390 scev_reset ();
1391 #ifdef ENABLE_CHECKING
1392 if (loops_state_satisfies_p (LOOP_CLOSED_SSA))
1393 verify_loop_closed_ssa (true);
1394 #endif
1396 if (loop_closed_ssa_invalidated)
1397 BITMAP_FREE (loop_closed_ssa_invalidated);
1399 while (changed
1400 && ++iteration <= PARAM_VALUE (PARAM_MAX_UNROLL_ITERATIONS));
1402 father_stack.release ();
1404 if (irred_invalidated
1405 && loops_state_satisfies_p (LOOPS_HAVE_MARKED_IRREDUCIBLE_REGIONS))
1406 mark_irreducible_loops ();
1408 return 0;
1411 /* Canonical induction variable creation pass. */
1413 namespace {
1415 const pass_data pass_data_iv_canon =
1417 GIMPLE_PASS, /* type */
1418 "ivcanon", /* name */
1419 OPTGROUP_LOOP, /* optinfo_flags */
1420 TV_TREE_LOOP_IVCANON, /* tv_id */
1421 ( PROP_cfg | PROP_ssa ), /* properties_required */
1422 0, /* properties_provided */
1423 0, /* properties_destroyed */
1424 0, /* todo_flags_start */
1425 0, /* todo_flags_finish */
1428 class pass_iv_canon : public gimple_opt_pass
1430 public:
1431 pass_iv_canon (gcc::context *ctxt)
1432 : gimple_opt_pass (pass_data_iv_canon, ctxt)
1435 /* opt_pass methods: */
1436 virtual bool gate (function *) { return flag_tree_loop_ivcanon != 0; }
1437 virtual unsigned int execute (function *fun);
1439 }; // class pass_iv_canon
1441 unsigned int
1442 pass_iv_canon::execute (function *fun)
1444 if (number_of_loops (fun) <= 1)
1445 return 0;
1447 return canonicalize_induction_variables ();
1450 } // anon namespace
1452 gimple_opt_pass *
1453 make_pass_iv_canon (gcc::context *ctxt)
1455 return new pass_iv_canon (ctxt);
1458 /* Complete unrolling of loops. */
1460 namespace {
1462 const pass_data pass_data_complete_unroll =
1464 GIMPLE_PASS, /* type */
1465 "cunroll", /* name */
1466 OPTGROUP_LOOP, /* optinfo_flags */
1467 TV_COMPLETE_UNROLL, /* tv_id */
1468 ( PROP_cfg | PROP_ssa ), /* properties_required */
1469 0, /* properties_provided */
1470 0, /* properties_destroyed */
1471 0, /* todo_flags_start */
1472 0, /* todo_flags_finish */
1475 class pass_complete_unroll : public gimple_opt_pass
1477 public:
1478 pass_complete_unroll (gcc::context *ctxt)
1479 : gimple_opt_pass (pass_data_complete_unroll, ctxt)
1482 /* opt_pass methods: */
1483 virtual unsigned int execute (function *);
1485 }; // class pass_complete_unroll
1487 unsigned int
1488 pass_complete_unroll::execute (function *fun)
1490 if (number_of_loops (fun) <= 1)
1491 return 0;
1493 return tree_unroll_loops_completely (flag_unroll_loops
1494 || flag_peel_loops
1495 || optimize >= 3, true);
1498 } // anon namespace
1500 gimple_opt_pass *
1501 make_pass_complete_unroll (gcc::context *ctxt)
1503 return new pass_complete_unroll (ctxt);
1506 /* Complete unrolling of inner loops. */
1508 namespace {
1510 const pass_data pass_data_complete_unrolli =
1512 GIMPLE_PASS, /* type */
1513 "cunrolli", /* name */
1514 OPTGROUP_LOOP, /* optinfo_flags */
1515 TV_COMPLETE_UNROLL, /* tv_id */
1516 ( PROP_cfg | PROP_ssa ), /* properties_required */
1517 0, /* properties_provided */
1518 0, /* properties_destroyed */
1519 0, /* todo_flags_start */
1520 0, /* todo_flags_finish */
1523 class pass_complete_unrolli : public gimple_opt_pass
1525 public:
1526 pass_complete_unrolli (gcc::context *ctxt)
1527 : gimple_opt_pass (pass_data_complete_unrolli, ctxt)
1530 /* opt_pass methods: */
1531 virtual bool gate (function *) { return optimize >= 2; }
1532 virtual unsigned int execute (function *);
1534 }; // class pass_complete_unrolli
1536 unsigned int
1537 pass_complete_unrolli::execute (function *fun)
1539 unsigned ret = 0;
1541 loop_optimizer_init (LOOPS_NORMAL
1542 | LOOPS_HAVE_RECORDED_EXITS);
1543 if (number_of_loops (fun) > 1)
1545 scev_initialize ();
1546 ret = tree_unroll_loops_completely (optimize >= 3, false);
1547 free_numbers_of_iterations_estimates ();
1548 scev_finalize ();
1550 loop_optimizer_finalize ();
1552 return ret;
1555 } // anon namespace
1557 gimple_opt_pass *
1558 make_pass_complete_unrolli (gcc::context *ctxt)
1560 return new pass_complete_unrolli (ctxt);