c++: using from enclosing class template [PR105006]
[official-gcc.git] / gcc / tree-vect-loop-manip.cc
bloba7bbc916bbcfc68508540e16c39c51a15f6a5f06
1 /* Vectorizer Specific Loop Manipulations
2 Copyright (C) 2003-2022 Free Software Foundation, Inc.
3 Contributed by Dorit Naishlos <dorit@il.ibm.com>
4 and Ira Rosen <irar@il.ibm.com>
6 This file is part of GCC.
8 GCC is free software; you can redistribute it and/or modify it under
9 the terms of the GNU General Public License as published by the Free
10 Software Foundation; either version 3, or (at your option) any later
11 version.
13 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
14 WARRANTY; without even the implied warranty of MERCHANTABILITY or
15 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
16 for more details.
18 You should have received a copy of the GNU General Public License
19 along with GCC; see the file COPYING3. If not see
20 <http://www.gnu.org/licenses/>. */
22 #include "config.h"
23 #include "system.h"
24 #include "coretypes.h"
25 #include "backend.h"
26 #include "tree.h"
27 #include "gimple.h"
28 #include "cfghooks.h"
29 #include "tree-pass.h"
30 #include "ssa.h"
31 #include "fold-const.h"
32 #include "cfganal.h"
33 #include "gimplify.h"
34 #include "gimple-iterator.h"
35 #include "gimplify-me.h"
36 #include "tree-cfg.h"
37 #include "tree-ssa-loop-manip.h"
38 #include "tree-into-ssa.h"
39 #include "tree-ssa.h"
40 #include "cfgloop.h"
41 #include "tree-scalar-evolution.h"
42 #include "tree-vectorizer.h"
43 #include "tree-ssa-loop-ivopts.h"
44 #include "gimple-fold.h"
45 #include "tree-ssa-loop-niter.h"
46 #include "internal-fn.h"
47 #include "stor-layout.h"
48 #include "optabs-query.h"
49 #include "vec-perm-indices.h"
50 #include "insn-config.h"
51 #include "rtl.h"
52 #include "recog.h"
54 /*************************************************************************
55 Simple Loop Peeling Utilities
57 Utilities to support loop peeling for vectorization purposes.
58 *************************************************************************/
61 /* Renames the use *OP_P. */
63 static void
64 rename_use_op (use_operand_p op_p)
66 tree new_name;
68 if (TREE_CODE (USE_FROM_PTR (op_p)) != SSA_NAME)
69 return;
71 new_name = get_current_def (USE_FROM_PTR (op_p));
73 /* Something defined outside of the loop. */
74 if (!new_name)
75 return;
77 /* An ordinary ssa name defined in the loop. */
79 SET_USE (op_p, new_name);
83 /* Renames the variables in basic block BB. Allow renaming of PHI arguments
84 on edges incoming from outer-block header if RENAME_FROM_OUTER_LOOP is
85 true. */
87 static void
88 rename_variables_in_bb (basic_block bb, bool rename_from_outer_loop)
90 gimple *stmt;
91 use_operand_p use_p;
92 ssa_op_iter iter;
93 edge e;
94 edge_iterator ei;
95 class loop *loop = bb->loop_father;
96 class loop *outer_loop = NULL;
98 if (rename_from_outer_loop)
100 gcc_assert (loop);
101 outer_loop = loop_outer (loop);
104 for (gimple_stmt_iterator gsi = gsi_start_bb (bb); !gsi_end_p (gsi);
105 gsi_next (&gsi))
107 stmt = gsi_stmt (gsi);
108 FOR_EACH_SSA_USE_OPERAND (use_p, stmt, iter, SSA_OP_ALL_USES)
109 rename_use_op (use_p);
112 FOR_EACH_EDGE (e, ei, bb->preds)
114 if (!flow_bb_inside_loop_p (loop, e->src))
116 if (!rename_from_outer_loop)
117 continue;
118 if (e->src != outer_loop->header)
120 if (outer_loop->inner->next)
122 /* If outer_loop has 2 inner loops, allow there to
123 be an extra basic block which decides which of the
124 two loops to use using LOOP_VECTORIZED. */
125 if (!single_pred_p (e->src)
126 || single_pred (e->src) != outer_loop->header)
127 continue;
131 for (gphi_iterator gsi = gsi_start_phis (bb); !gsi_end_p (gsi);
132 gsi_next (&gsi))
133 rename_use_op (PHI_ARG_DEF_PTR_FROM_EDGE (gsi.phi (), e));
138 struct adjust_info
140 tree from, to;
141 basic_block bb;
144 /* A stack of values to be adjusted in debug stmts. We have to
145 process them LIFO, so that the closest substitution applies. If we
146 processed them FIFO, without the stack, we might substitute uses
147 with a PHI DEF that would soon become non-dominant, and when we got
148 to the suitable one, it wouldn't have anything to substitute any
149 more. */
150 static vec<adjust_info, va_heap> adjust_vec;
152 /* Adjust any debug stmts that referenced AI->from values to use the
153 loop-closed AI->to, if the references are dominated by AI->bb and
154 not by the definition of AI->from. */
156 static void
157 adjust_debug_stmts_now (adjust_info *ai)
159 basic_block bbphi = ai->bb;
160 tree orig_def = ai->from;
161 tree new_def = ai->to;
162 imm_use_iterator imm_iter;
163 gimple *stmt;
164 basic_block bbdef = gimple_bb (SSA_NAME_DEF_STMT (orig_def));
166 gcc_assert (dom_info_available_p (CDI_DOMINATORS));
168 /* Adjust any debug stmts that held onto non-loop-closed
169 references. */
170 FOR_EACH_IMM_USE_STMT (stmt, imm_iter, orig_def)
172 use_operand_p use_p;
173 basic_block bbuse;
175 if (!is_gimple_debug (stmt))
176 continue;
178 gcc_assert (gimple_debug_bind_p (stmt));
180 bbuse = gimple_bb (stmt);
182 if ((bbuse == bbphi
183 || dominated_by_p (CDI_DOMINATORS, bbuse, bbphi))
184 && !(bbuse == bbdef
185 || dominated_by_p (CDI_DOMINATORS, bbuse, bbdef)))
187 if (new_def)
188 FOR_EACH_IMM_USE_ON_STMT (use_p, imm_iter)
189 SET_USE (use_p, new_def);
190 else
192 gimple_debug_bind_reset_value (stmt);
193 update_stmt (stmt);
199 /* Adjust debug stmts as scheduled before. */
201 static void
202 adjust_vec_debug_stmts (void)
204 if (!MAY_HAVE_DEBUG_BIND_STMTS)
205 return;
207 gcc_assert (adjust_vec.exists ());
209 while (!adjust_vec.is_empty ())
211 adjust_debug_stmts_now (&adjust_vec.last ());
212 adjust_vec.pop ();
216 /* Adjust any debug stmts that referenced FROM values to use the
217 loop-closed TO, if the references are dominated by BB and not by
218 the definition of FROM. If adjust_vec is non-NULL, adjustments
219 will be postponed until adjust_vec_debug_stmts is called. */
221 static void
222 adjust_debug_stmts (tree from, tree to, basic_block bb)
224 adjust_info ai;
226 if (MAY_HAVE_DEBUG_BIND_STMTS
227 && TREE_CODE (from) == SSA_NAME
228 && ! SSA_NAME_IS_DEFAULT_DEF (from)
229 && ! virtual_operand_p (from))
231 ai.from = from;
232 ai.to = to;
233 ai.bb = bb;
235 if (adjust_vec.exists ())
236 adjust_vec.safe_push (ai);
237 else
238 adjust_debug_stmts_now (&ai);
242 /* Change E's phi arg in UPDATE_PHI to NEW_DEF, and record information
243 to adjust any debug stmts that referenced the old phi arg,
244 presumably non-loop-closed references left over from other
245 transformations. */
247 static void
248 adjust_phi_and_debug_stmts (gimple *update_phi, edge e, tree new_def)
250 tree orig_def = PHI_ARG_DEF_FROM_EDGE (update_phi, e);
252 SET_PHI_ARG_DEF (update_phi, e->dest_idx, new_def);
254 if (MAY_HAVE_DEBUG_BIND_STMTS)
255 adjust_debug_stmts (orig_def, PHI_RESULT (update_phi),
256 gimple_bb (update_phi));
259 /* Define one loop rgroup control CTRL from loop LOOP. INIT_CTRL is the value
260 that the control should have during the first iteration and NEXT_CTRL is the
261 value that it should have on subsequent iterations. */
263 static void
264 vect_set_loop_control (class loop *loop, tree ctrl, tree init_ctrl,
265 tree next_ctrl)
267 gphi *phi = create_phi_node (ctrl, loop->header);
268 add_phi_arg (phi, init_ctrl, loop_preheader_edge (loop), UNKNOWN_LOCATION);
269 add_phi_arg (phi, next_ctrl, loop_latch_edge (loop), UNKNOWN_LOCATION);
272 /* Add SEQ to the end of LOOP's preheader block. */
274 static void
275 add_preheader_seq (class loop *loop, gimple_seq seq)
277 if (seq)
279 edge pe = loop_preheader_edge (loop);
280 basic_block new_bb = gsi_insert_seq_on_edge_immediate (pe, seq);
281 gcc_assert (!new_bb);
285 /* Add SEQ to the beginning of LOOP's header block. */
287 static void
288 add_header_seq (class loop *loop, gimple_seq seq)
290 if (seq)
292 gimple_stmt_iterator gsi = gsi_after_labels (loop->header);
293 gsi_insert_seq_before (&gsi, seq, GSI_SAME_STMT);
297 /* Return true if the target can interleave elements of two vectors.
298 OFFSET is 0 if the first half of the vectors should be interleaved
299 or 1 if the second half should. When returning true, store the
300 associated permutation in INDICES. */
302 static bool
303 interleave_supported_p (vec_perm_indices *indices, tree vectype,
304 unsigned int offset)
306 poly_uint64 nelts = TYPE_VECTOR_SUBPARTS (vectype);
307 poly_uint64 base = exact_div (nelts, 2) * offset;
308 vec_perm_builder sel (nelts, 2, 3);
309 for (unsigned int i = 0; i < 3; ++i)
311 sel.quick_push (base + i);
312 sel.quick_push (base + i + nelts);
314 indices->new_vector (sel, 2, nelts);
315 return can_vec_perm_const_p (TYPE_MODE (vectype), *indices);
318 /* Try to use permutes to define the masks in DEST_RGM using the masks
319 in SRC_RGM, given that the former has twice as many masks as the
320 latter. Return true on success, adding any new statements to SEQ. */
322 static bool
323 vect_maybe_permute_loop_masks (gimple_seq *seq, rgroup_controls *dest_rgm,
324 rgroup_controls *src_rgm)
326 tree src_masktype = src_rgm->type;
327 tree dest_masktype = dest_rgm->type;
328 machine_mode src_mode = TYPE_MODE (src_masktype);
329 insn_code icode1, icode2;
330 if (dest_rgm->max_nscalars_per_iter <= src_rgm->max_nscalars_per_iter
331 && (icode1 = optab_handler (vec_unpacku_hi_optab,
332 src_mode)) != CODE_FOR_nothing
333 && (icode2 = optab_handler (vec_unpacku_lo_optab,
334 src_mode)) != CODE_FOR_nothing)
336 /* Unpacking the source masks gives at least as many mask bits as
337 we need. We can then VIEW_CONVERT any excess bits away. */
338 machine_mode dest_mode = insn_data[icode1].operand[0].mode;
339 gcc_assert (dest_mode == insn_data[icode2].operand[0].mode);
340 tree unpack_masktype = vect_halve_mask_nunits (src_masktype, dest_mode);
341 for (unsigned int i = 0; i < dest_rgm->controls.length (); ++i)
343 tree src = src_rgm->controls[i / 2];
344 tree dest = dest_rgm->controls[i];
345 tree_code code = ((i & 1) == (BYTES_BIG_ENDIAN ? 0 : 1)
346 ? VEC_UNPACK_HI_EXPR
347 : VEC_UNPACK_LO_EXPR);
348 gassign *stmt;
349 if (dest_masktype == unpack_masktype)
350 stmt = gimple_build_assign (dest, code, src);
351 else
353 tree temp = make_ssa_name (unpack_masktype);
354 stmt = gimple_build_assign (temp, code, src);
355 gimple_seq_add_stmt (seq, stmt);
356 stmt = gimple_build_assign (dest, VIEW_CONVERT_EXPR,
357 build1 (VIEW_CONVERT_EXPR,
358 dest_masktype, temp));
360 gimple_seq_add_stmt (seq, stmt);
362 return true;
364 vec_perm_indices indices[2];
365 if (dest_masktype == src_masktype
366 && interleave_supported_p (&indices[0], src_masktype, 0)
367 && interleave_supported_p (&indices[1], src_masktype, 1))
369 /* The destination requires twice as many mask bits as the source, so
370 we can use interleaving permutes to double up the number of bits. */
371 tree masks[2];
372 for (unsigned int i = 0; i < 2; ++i)
373 masks[i] = vect_gen_perm_mask_checked (src_masktype, indices[i]);
374 for (unsigned int i = 0; i < dest_rgm->controls.length (); ++i)
376 tree src = src_rgm->controls[i / 2];
377 tree dest = dest_rgm->controls[i];
378 gimple *stmt = gimple_build_assign (dest, VEC_PERM_EXPR,
379 src, src, masks[i & 1]);
380 gimple_seq_add_stmt (seq, stmt);
382 return true;
384 return false;
387 /* Helper for vect_set_loop_condition_partial_vectors. Generate definitions
388 for all the rgroup controls in RGC and return a control that is nonzero
389 when the loop needs to iterate. Add any new preheader statements to
390 PREHEADER_SEQ. Use LOOP_COND_GSI to insert code before the exit gcond.
392 RGC belongs to loop LOOP. The loop originally iterated NITERS
393 times and has been vectorized according to LOOP_VINFO.
395 If NITERS_SKIP is nonnull, the first iteration of the vectorized loop
396 starts with NITERS_SKIP dummy iterations of the scalar loop before
397 the real work starts. The mask elements for these dummy iterations
398 must be 0, to ensure that the extra iterations do not have an effect.
400 It is known that:
402 NITERS * RGC->max_nscalars_per_iter * RGC->factor
404 does not overflow. However, MIGHT_WRAP_P says whether an induction
405 variable that starts at 0 and has step:
407 VF * RGC->max_nscalars_per_iter * RGC->factor
409 might overflow before hitting a value above:
411 (NITERS + NITERS_SKIP) * RGC->max_nscalars_per_iter * RGC->factor
413 This means that we cannot guarantee that such an induction variable
414 would ever hit a value that produces a set of all-false masks or zero
415 lengths for RGC.
417 Note: the cost of the code generated by this function is modeled
418 by vect_estimate_min_profitable_iters, so changes here may need
419 corresponding changes there. */
421 static tree
422 vect_set_loop_controls_directly (class loop *loop, loop_vec_info loop_vinfo,
423 gimple_seq *preheader_seq,
424 gimple_seq *header_seq,
425 gimple_stmt_iterator loop_cond_gsi,
426 rgroup_controls *rgc, tree niters,
427 tree niters_skip, bool might_wrap_p)
429 tree compare_type = LOOP_VINFO_RGROUP_COMPARE_TYPE (loop_vinfo);
430 tree iv_type = LOOP_VINFO_RGROUP_IV_TYPE (loop_vinfo);
431 bool use_masks_p = LOOP_VINFO_FULLY_MASKED_P (loop_vinfo);
433 tree ctrl_type = rgc->type;
434 unsigned int nitems_per_iter = rgc->max_nscalars_per_iter * rgc->factor;
435 poly_uint64 nitems_per_ctrl = TYPE_VECTOR_SUBPARTS (ctrl_type) * rgc->factor;
436 poly_uint64 vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo);
437 tree length_limit = NULL_TREE;
438 /* For length, we need length_limit to ensure length in range. */
439 if (!use_masks_p)
440 length_limit = build_int_cst (compare_type, nitems_per_ctrl);
442 /* Calculate the maximum number of item values that the rgroup
443 handles in total, the number that it handles for each iteration
444 of the vector loop, and the number that it should skip during the
445 first iteration of the vector loop. */
446 tree nitems_total = niters;
447 tree nitems_step = build_int_cst (iv_type, vf);
448 tree nitems_skip = niters_skip;
449 if (nitems_per_iter != 1)
451 /* We checked before setting LOOP_VINFO_USING_PARTIAL_VECTORS_P that
452 these multiplications don't overflow. */
453 tree compare_factor = build_int_cst (compare_type, nitems_per_iter);
454 tree iv_factor = build_int_cst (iv_type, nitems_per_iter);
455 nitems_total = gimple_build (preheader_seq, MULT_EXPR, compare_type,
456 nitems_total, compare_factor);
457 nitems_step = gimple_build (preheader_seq, MULT_EXPR, iv_type,
458 nitems_step, iv_factor);
459 if (nitems_skip)
460 nitems_skip = gimple_build (preheader_seq, MULT_EXPR, compare_type,
461 nitems_skip, compare_factor);
464 /* Create an induction variable that counts the number of items
465 processed. */
466 tree index_before_incr, index_after_incr;
467 gimple_stmt_iterator incr_gsi;
468 bool insert_after;
469 standard_iv_increment_position (loop, &incr_gsi, &insert_after);
470 create_iv (build_int_cst (iv_type, 0), nitems_step, NULL_TREE, loop,
471 &incr_gsi, insert_after, &index_before_incr, &index_after_incr);
473 tree zero_index = build_int_cst (compare_type, 0);
474 tree test_index, test_limit, first_limit;
475 gimple_stmt_iterator *test_gsi;
476 if (might_wrap_p)
478 /* In principle the loop should stop iterating once the incremented
479 IV reaches a value greater than or equal to:
481 NITEMS_TOTAL +[infinite-prec] NITEMS_SKIP
483 However, there's no guarantee that this addition doesn't overflow
484 the comparison type, or that the IV hits a value above it before
485 wrapping around. We therefore adjust the limit down by one
486 IV step:
488 (NITEMS_TOTAL +[infinite-prec] NITEMS_SKIP)
489 -[infinite-prec] NITEMS_STEP
491 and compare the IV against this limit _before_ incrementing it.
492 Since the comparison type is unsigned, we actually want the
493 subtraction to saturate at zero:
495 (NITEMS_TOTAL +[infinite-prec] NITEMS_SKIP)
496 -[sat] NITEMS_STEP
498 And since NITEMS_SKIP < NITEMS_STEP, we can reassociate this as:
500 NITEMS_TOTAL -[sat] (NITEMS_STEP - NITEMS_SKIP)
502 where the rightmost subtraction can be done directly in
503 COMPARE_TYPE. */
504 test_index = index_before_incr;
505 tree adjust = gimple_convert (preheader_seq, compare_type,
506 nitems_step);
507 if (nitems_skip)
508 adjust = gimple_build (preheader_seq, MINUS_EXPR, compare_type,
509 adjust, nitems_skip);
510 test_limit = gimple_build (preheader_seq, MAX_EXPR, compare_type,
511 nitems_total, adjust);
512 test_limit = gimple_build (preheader_seq, MINUS_EXPR, compare_type,
513 test_limit, adjust);
514 test_gsi = &incr_gsi;
516 /* Get a safe limit for the first iteration. */
517 if (nitems_skip)
519 /* The first vector iteration can handle at most NITEMS_STEP
520 items. NITEMS_STEP <= CONST_LIMIT, and adding
521 NITEMS_SKIP to that cannot overflow. */
522 tree const_limit = build_int_cst (compare_type,
523 LOOP_VINFO_VECT_FACTOR (loop_vinfo)
524 * nitems_per_iter);
525 first_limit = gimple_build (preheader_seq, MIN_EXPR, compare_type,
526 nitems_total, const_limit);
527 first_limit = gimple_build (preheader_seq, PLUS_EXPR, compare_type,
528 first_limit, nitems_skip);
530 else
531 /* For the first iteration it doesn't matter whether the IV hits
532 a value above NITEMS_TOTAL. That only matters for the latch
533 condition. */
534 first_limit = nitems_total;
536 else
538 /* Test the incremented IV, which will always hit a value above
539 the bound before wrapping. */
540 test_index = index_after_incr;
541 test_limit = nitems_total;
542 if (nitems_skip)
543 test_limit = gimple_build (preheader_seq, PLUS_EXPR, compare_type,
544 test_limit, nitems_skip);
545 test_gsi = &loop_cond_gsi;
547 first_limit = test_limit;
550 /* Convert the IV value to the comparison type (either a no-op or
551 a demotion). */
552 gimple_seq test_seq = NULL;
553 test_index = gimple_convert (&test_seq, compare_type, test_index);
554 gsi_insert_seq_before (test_gsi, test_seq, GSI_SAME_STMT);
556 /* Provide a definition of each control in the group. */
557 tree next_ctrl = NULL_TREE;
558 tree ctrl;
559 unsigned int i;
560 FOR_EACH_VEC_ELT_REVERSE (rgc->controls, i, ctrl)
562 /* Previous controls will cover BIAS items. This control covers the
563 next batch. */
564 poly_uint64 bias = nitems_per_ctrl * i;
565 tree bias_tree = build_int_cst (compare_type, bias);
567 /* See whether the first iteration of the vector loop is known
568 to have a full control. */
569 poly_uint64 const_limit;
570 bool first_iteration_full
571 = (poly_int_tree_p (first_limit, &const_limit)
572 && known_ge (const_limit, (i + 1) * nitems_per_ctrl));
574 /* Rather than have a new IV that starts at BIAS and goes up to
575 TEST_LIMIT, prefer to use the same 0-based IV for each control
576 and adjust the bound down by BIAS. */
577 tree this_test_limit = test_limit;
578 if (i != 0)
580 this_test_limit = gimple_build (preheader_seq, MAX_EXPR,
581 compare_type, this_test_limit,
582 bias_tree);
583 this_test_limit = gimple_build (preheader_seq, MINUS_EXPR,
584 compare_type, this_test_limit,
585 bias_tree);
588 /* Create the initial control. First include all items that
589 are within the loop limit. */
590 tree init_ctrl = NULL_TREE;
591 if (!first_iteration_full)
593 tree start, end;
594 if (first_limit == test_limit)
596 /* Use a natural test between zero (the initial IV value)
597 and the loop limit. The "else" block would be valid too,
598 but this choice can avoid the need to load BIAS_TREE into
599 a register. */
600 start = zero_index;
601 end = this_test_limit;
603 else
605 /* FIRST_LIMIT is the maximum number of items handled by the
606 first iteration of the vector loop. Test the portion
607 associated with this control. */
608 start = bias_tree;
609 end = first_limit;
612 if (use_masks_p)
613 init_ctrl = vect_gen_while (preheader_seq, ctrl_type,
614 start, end, "max_mask");
615 else
617 init_ctrl = make_temp_ssa_name (compare_type, NULL, "max_len");
618 gimple_seq seq = vect_gen_len (init_ctrl, start,
619 end, length_limit);
620 gimple_seq_add_seq (preheader_seq, seq);
624 /* Now AND out the bits that are within the number of skipped
625 items. */
626 poly_uint64 const_skip;
627 if (nitems_skip
628 && !(poly_int_tree_p (nitems_skip, &const_skip)
629 && known_le (const_skip, bias)))
631 gcc_assert (use_masks_p);
632 tree unskipped_mask = vect_gen_while_not (preheader_seq, ctrl_type,
633 bias_tree, nitems_skip);
634 if (init_ctrl)
635 init_ctrl = gimple_build (preheader_seq, BIT_AND_EXPR, ctrl_type,
636 init_ctrl, unskipped_mask);
637 else
638 init_ctrl = unskipped_mask;
641 if (!init_ctrl)
643 /* First iteration is full. */
644 if (use_masks_p)
645 init_ctrl = build_minus_one_cst (ctrl_type);
646 else
647 init_ctrl = length_limit;
650 /* Get the control value for the next iteration of the loop. */
651 if (use_masks_p)
653 gimple_seq stmts = NULL;
654 next_ctrl = vect_gen_while (&stmts, ctrl_type, test_index,
655 this_test_limit, "next_mask");
656 gsi_insert_seq_before (test_gsi, stmts, GSI_SAME_STMT);
658 else
660 next_ctrl = make_temp_ssa_name (compare_type, NULL, "next_len");
661 gimple_seq seq = vect_gen_len (next_ctrl, test_index, this_test_limit,
662 length_limit);
663 gsi_insert_seq_before (test_gsi, seq, GSI_SAME_STMT);
666 vect_set_loop_control (loop, ctrl, init_ctrl, next_ctrl);
669 int partial_load_bias = LOOP_VINFO_PARTIAL_LOAD_STORE_BIAS (loop_vinfo);
670 if (partial_load_bias != 0)
672 tree adjusted_len = rgc->bias_adjusted_ctrl;
673 gassign *minus = gimple_build_assign (adjusted_len, PLUS_EXPR,
674 rgc->controls[0],
675 build_int_cst
676 (TREE_TYPE (rgc->controls[0]),
677 partial_load_bias));
678 gimple_seq_add_stmt (header_seq, minus);
681 return next_ctrl;
684 /* Set up the iteration condition and rgroup controls for LOOP, given
685 that LOOP_VINFO_USING_PARTIAL_VECTORS_P is true for the vectorized
686 loop. LOOP_VINFO describes the vectorization of LOOP. NITERS is
687 the number of iterations of the original scalar loop that should be
688 handled by the vector loop. NITERS_MAYBE_ZERO and FINAL_IV are as
689 for vect_set_loop_condition.
691 Insert the branch-back condition before LOOP_COND_GSI and return the
692 final gcond. */
694 static gcond *
695 vect_set_loop_condition_partial_vectors (class loop *loop,
696 loop_vec_info loop_vinfo, tree niters,
697 tree final_iv, bool niters_maybe_zero,
698 gimple_stmt_iterator loop_cond_gsi)
700 gimple_seq preheader_seq = NULL;
701 gimple_seq header_seq = NULL;
703 bool use_masks_p = LOOP_VINFO_FULLY_MASKED_P (loop_vinfo);
704 tree compare_type = LOOP_VINFO_RGROUP_COMPARE_TYPE (loop_vinfo);
705 unsigned int compare_precision = TYPE_PRECISION (compare_type);
706 tree orig_niters = niters;
708 /* Type of the initial value of NITERS. */
709 tree ni_actual_type = TREE_TYPE (niters);
710 unsigned int ni_actual_precision = TYPE_PRECISION (ni_actual_type);
711 tree niters_skip = LOOP_VINFO_MASK_SKIP_NITERS (loop_vinfo);
713 /* Convert NITERS to the same size as the compare. */
714 if (compare_precision > ni_actual_precision
715 && niters_maybe_zero)
717 /* We know that there is always at least one iteration, so if the
718 count is zero then it must have wrapped. Cope with this by
719 subtracting 1 before the conversion and adding 1 to the result. */
720 gcc_assert (TYPE_UNSIGNED (ni_actual_type));
721 niters = gimple_build (&preheader_seq, PLUS_EXPR, ni_actual_type,
722 niters, build_minus_one_cst (ni_actual_type));
723 niters = gimple_convert (&preheader_seq, compare_type, niters);
724 niters = gimple_build (&preheader_seq, PLUS_EXPR, compare_type,
725 niters, build_one_cst (compare_type));
727 else
728 niters = gimple_convert (&preheader_seq, compare_type, niters);
730 /* Iterate over all the rgroups and fill in their controls. We could use
731 the first control from any rgroup for the loop condition; here we
732 arbitrarily pick the last. */
733 tree test_ctrl = NULL_TREE;
734 rgroup_controls *rgc;
735 unsigned int i;
736 auto_vec<rgroup_controls> *controls = use_masks_p
737 ? &LOOP_VINFO_MASKS (loop_vinfo)
738 : &LOOP_VINFO_LENS (loop_vinfo);
739 FOR_EACH_VEC_ELT (*controls, i, rgc)
740 if (!rgc->controls.is_empty ())
742 /* First try using permutes. This adds a single vector
743 instruction to the loop for each mask, but needs no extra
744 loop invariants or IVs. */
745 unsigned int nmasks = i + 1;
746 if (use_masks_p && (nmasks & 1) == 0)
748 rgroup_controls *half_rgc = &(*controls)[nmasks / 2 - 1];
749 if (!half_rgc->controls.is_empty ()
750 && vect_maybe_permute_loop_masks (&header_seq, rgc, half_rgc))
751 continue;
754 /* See whether zero-based IV would ever generate all-false masks
755 or zero length before wrapping around. */
756 bool might_wrap_p = vect_rgroup_iv_might_wrap_p (loop_vinfo, rgc);
758 /* Set up all controls for this group. */
759 test_ctrl = vect_set_loop_controls_directly (loop, loop_vinfo,
760 &preheader_seq,
761 &header_seq,
762 loop_cond_gsi, rgc,
763 niters, niters_skip,
764 might_wrap_p);
767 /* Emit all accumulated statements. */
768 add_preheader_seq (loop, preheader_seq);
769 add_header_seq (loop, header_seq);
771 /* Get a boolean result that tells us whether to iterate. */
772 edge exit_edge = single_exit (loop);
773 tree_code code = (exit_edge->flags & EDGE_TRUE_VALUE) ? EQ_EXPR : NE_EXPR;
774 tree zero_ctrl = build_zero_cst (TREE_TYPE (test_ctrl));
775 gcond *cond_stmt = gimple_build_cond (code, test_ctrl, zero_ctrl,
776 NULL_TREE, NULL_TREE);
777 gsi_insert_before (&loop_cond_gsi, cond_stmt, GSI_SAME_STMT);
779 /* The loop iterates (NITERS - 1) / VF + 1 times.
780 Subtract one from this to get the latch count. */
781 tree step = build_int_cst (compare_type,
782 LOOP_VINFO_VECT_FACTOR (loop_vinfo));
783 tree niters_minus_one = fold_build2 (PLUS_EXPR, compare_type, niters,
784 build_minus_one_cst (compare_type));
785 loop->nb_iterations = fold_build2 (TRUNC_DIV_EXPR, compare_type,
786 niters_minus_one, step);
788 if (final_iv)
790 gassign *assign = gimple_build_assign (final_iv, orig_niters);
791 gsi_insert_on_edge_immediate (single_exit (loop), assign);
794 return cond_stmt;
797 /* Like vect_set_loop_condition, but handle the case in which the vector
798 loop handles exactly VF scalars per iteration. */
800 static gcond *
801 vect_set_loop_condition_normal (class loop *loop, tree niters, tree step,
802 tree final_iv, bool niters_maybe_zero,
803 gimple_stmt_iterator loop_cond_gsi)
805 tree indx_before_incr, indx_after_incr;
806 gcond *cond_stmt;
807 gcond *orig_cond;
808 edge pe = loop_preheader_edge (loop);
809 edge exit_edge = single_exit (loop);
810 gimple_stmt_iterator incr_gsi;
811 bool insert_after;
812 enum tree_code code;
813 tree niters_type = TREE_TYPE (niters);
815 orig_cond = get_loop_exit_condition (loop);
816 gcc_assert (orig_cond);
817 loop_cond_gsi = gsi_for_stmt (orig_cond);
819 tree init, limit;
820 if (!niters_maybe_zero && integer_onep (step))
822 /* In this case we can use a simple 0-based IV:
825 x = 0;
829 x += 1;
831 while (x < NITERS); */
832 code = (exit_edge->flags & EDGE_TRUE_VALUE) ? GE_EXPR : LT_EXPR;
833 init = build_zero_cst (niters_type);
834 limit = niters;
836 else
838 /* The following works for all values of NITERS except 0:
841 x = 0;
845 x += STEP;
847 while (x <= NITERS - STEP);
849 so that the loop continues to iterate if x + STEP - 1 < NITERS
850 but stops if x + STEP - 1 >= NITERS.
852 However, if NITERS is zero, x never hits a value above NITERS - STEP
853 before wrapping around. There are two obvious ways of dealing with
854 this:
856 - start at STEP - 1 and compare x before incrementing it
857 - start at -1 and compare x after incrementing it
859 The latter is simpler and is what we use. The loop in this case
860 looks like:
863 x = -1;
867 x += STEP;
869 while (x < NITERS - STEP);
871 In both cases the loop limit is NITERS - STEP. */
872 gimple_seq seq = NULL;
873 limit = force_gimple_operand (niters, &seq, true, NULL_TREE);
874 limit = gimple_build (&seq, MINUS_EXPR, TREE_TYPE (limit), limit, step);
875 if (seq)
877 basic_block new_bb = gsi_insert_seq_on_edge_immediate (pe, seq);
878 gcc_assert (!new_bb);
880 if (niters_maybe_zero)
882 /* Case C. */
883 code = (exit_edge->flags & EDGE_TRUE_VALUE) ? GE_EXPR : LT_EXPR;
884 init = build_all_ones_cst (niters_type);
886 else
888 /* Case B. */
889 code = (exit_edge->flags & EDGE_TRUE_VALUE) ? GT_EXPR : LE_EXPR;
890 init = build_zero_cst (niters_type);
894 standard_iv_increment_position (loop, &incr_gsi, &insert_after);
895 create_iv (init, step, NULL_TREE, loop,
896 &incr_gsi, insert_after, &indx_before_incr, &indx_after_incr);
897 indx_after_incr = force_gimple_operand_gsi (&loop_cond_gsi, indx_after_incr,
898 true, NULL_TREE, true,
899 GSI_SAME_STMT);
900 limit = force_gimple_operand_gsi (&loop_cond_gsi, limit, true, NULL_TREE,
901 true, GSI_SAME_STMT);
903 cond_stmt = gimple_build_cond (code, indx_after_incr, limit, NULL_TREE,
904 NULL_TREE);
906 gsi_insert_before (&loop_cond_gsi, cond_stmt, GSI_SAME_STMT);
908 /* Record the number of latch iterations. */
909 if (limit == niters)
910 /* Case A: the loop iterates NITERS times. Subtract one to get the
911 latch count. */
912 loop->nb_iterations = fold_build2 (MINUS_EXPR, niters_type, niters,
913 build_int_cst (niters_type, 1));
914 else
915 /* Case B or C: the loop iterates (NITERS - STEP) / STEP + 1 times.
916 Subtract one from this to get the latch count. */
917 loop->nb_iterations = fold_build2 (TRUNC_DIV_EXPR, niters_type,
918 limit, step);
920 if (final_iv)
922 gassign *assign = gimple_build_assign (final_iv, MINUS_EXPR,
923 indx_after_incr, init);
924 gsi_insert_on_edge_immediate (single_exit (loop), assign);
927 return cond_stmt;
930 /* If we're using fully-masked loops, make LOOP iterate:
932 N == (NITERS - 1) / STEP + 1
934 times. When NITERS is zero, this is equivalent to making the loop
935 execute (1 << M) / STEP times, where M is the precision of NITERS.
936 NITERS_MAYBE_ZERO is true if this last case might occur.
938 If we're not using fully-masked loops, make LOOP iterate:
940 N == (NITERS - STEP) / STEP + 1
942 times, where NITERS is known to be outside the range [1, STEP - 1].
943 This is equivalent to making the loop execute NITERS / STEP times
944 when NITERS is nonzero and (1 << M) / STEP times otherwise.
945 NITERS_MAYBE_ZERO again indicates whether this last case might occur.
947 If FINAL_IV is nonnull, it is an SSA name that should be set to
948 N * STEP on exit from the loop.
950 Assumption: the exit-condition of LOOP is the last stmt in the loop. */
952 void
953 vect_set_loop_condition (class loop *loop, loop_vec_info loop_vinfo,
954 tree niters, tree step, tree final_iv,
955 bool niters_maybe_zero)
957 gcond *cond_stmt;
958 gcond *orig_cond = get_loop_exit_condition (loop);
959 gimple_stmt_iterator loop_cond_gsi = gsi_for_stmt (orig_cond);
961 if (loop_vinfo && LOOP_VINFO_USING_PARTIAL_VECTORS_P (loop_vinfo))
962 cond_stmt = vect_set_loop_condition_partial_vectors (loop, loop_vinfo,
963 niters, final_iv,
964 niters_maybe_zero,
965 loop_cond_gsi);
966 else
967 cond_stmt = vect_set_loop_condition_normal (loop, niters, step, final_iv,
968 niters_maybe_zero,
969 loop_cond_gsi);
971 /* Remove old loop exit test. */
972 stmt_vec_info orig_cond_info;
973 if (loop_vinfo
974 && (orig_cond_info = loop_vinfo->lookup_stmt (orig_cond)))
975 loop_vinfo->remove_stmt (orig_cond_info);
976 else
977 gsi_remove (&loop_cond_gsi, true);
979 if (dump_enabled_p ())
980 dump_printf_loc (MSG_NOTE, vect_location, "New loop exit condition: %G",
981 cond_stmt);
984 /* Helper routine of slpeel_tree_duplicate_loop_to_edge_cfg.
985 For all PHI arguments in FROM->dest and TO->dest from those
986 edges ensure that TO->dest PHI arguments have current_def
987 to that in from. */
989 static void
990 slpeel_duplicate_current_defs_from_edges (edge from, edge to)
992 gimple_stmt_iterator gsi_from, gsi_to;
994 for (gsi_from = gsi_start_phis (from->dest),
995 gsi_to = gsi_start_phis (to->dest);
996 !gsi_end_p (gsi_from) && !gsi_end_p (gsi_to);)
998 gimple *from_phi = gsi_stmt (gsi_from);
999 gimple *to_phi = gsi_stmt (gsi_to);
1000 tree from_arg = PHI_ARG_DEF_FROM_EDGE (from_phi, from);
1001 tree to_arg = PHI_ARG_DEF_FROM_EDGE (to_phi, to);
1002 if (virtual_operand_p (from_arg))
1004 gsi_next (&gsi_from);
1005 continue;
1007 if (virtual_operand_p (to_arg))
1009 gsi_next (&gsi_to);
1010 continue;
1012 if (TREE_CODE (from_arg) != SSA_NAME)
1013 gcc_assert (operand_equal_p (from_arg, to_arg, 0));
1014 else if (TREE_CODE (to_arg) == SSA_NAME
1015 && from_arg != to_arg)
1017 if (get_current_def (to_arg) == NULL_TREE)
1019 gcc_assert (types_compatible_p (TREE_TYPE (to_arg),
1020 TREE_TYPE (get_current_def
1021 (from_arg))));
1022 set_current_def (to_arg, get_current_def (from_arg));
1025 gsi_next (&gsi_from);
1026 gsi_next (&gsi_to);
1029 gphi *from_phi = get_virtual_phi (from->dest);
1030 gphi *to_phi = get_virtual_phi (to->dest);
1031 if (from_phi)
1032 set_current_def (PHI_ARG_DEF_FROM_EDGE (to_phi, to),
1033 get_current_def (PHI_ARG_DEF_FROM_EDGE (from_phi, from)));
1037 /* Given LOOP this function generates a new copy of it and puts it
1038 on E which is either the entry or exit of LOOP. If SCALAR_LOOP is
1039 non-NULL, assume LOOP and SCALAR_LOOP are equivalent and copy the
1040 basic blocks from SCALAR_LOOP instead of LOOP, but to either the
1041 entry or exit of LOOP. */
1043 class loop *
1044 slpeel_tree_duplicate_loop_to_edge_cfg (class loop *loop,
1045 class loop *scalar_loop, edge e)
1047 class loop *new_loop;
1048 basic_block *new_bbs, *bbs, *pbbs;
1049 bool at_exit;
1050 bool was_imm_dom;
1051 basic_block exit_dest;
1052 edge exit, new_exit;
1053 bool duplicate_outer_loop = false;
1055 exit = single_exit (loop);
1056 at_exit = (e == exit);
1057 if (!at_exit && e != loop_preheader_edge (loop))
1058 return NULL;
1060 if (scalar_loop == NULL)
1061 scalar_loop = loop;
1063 bbs = XNEWVEC (basic_block, scalar_loop->num_nodes + 1);
1064 pbbs = bbs + 1;
1065 get_loop_body_with_size (scalar_loop, pbbs, scalar_loop->num_nodes);
1066 /* Allow duplication of outer loops. */
1067 if (scalar_loop->inner)
1068 duplicate_outer_loop = true;
1069 /* Check whether duplication is possible. */
1070 if (!can_copy_bbs_p (pbbs, scalar_loop->num_nodes))
1072 free (bbs);
1073 return NULL;
1076 /* Generate new loop structure. */
1077 new_loop = duplicate_loop (scalar_loop, loop_outer (scalar_loop));
1078 duplicate_subloops (scalar_loop, new_loop);
1080 exit_dest = exit->dest;
1081 was_imm_dom = (get_immediate_dominator (CDI_DOMINATORS,
1082 exit_dest) == loop->header ?
1083 true : false);
1085 /* Also copy the pre-header, this avoids jumping through hoops to
1086 duplicate the loop entry PHI arguments. Create an empty
1087 pre-header unconditionally for this. */
1088 basic_block preheader = split_edge (loop_preheader_edge (scalar_loop));
1089 edge entry_e = single_pred_edge (preheader);
1090 bbs[0] = preheader;
1091 new_bbs = XNEWVEC (basic_block, scalar_loop->num_nodes + 1);
1093 exit = single_exit (scalar_loop);
1094 copy_bbs (bbs, scalar_loop->num_nodes + 1, new_bbs,
1095 &exit, 1, &new_exit, NULL,
1096 at_exit ? loop->latch : e->src, true);
1097 exit = single_exit (loop);
1098 basic_block new_preheader = new_bbs[0];
1100 /* Before installing PHI arguments make sure that the edges
1101 into them match that of the scalar loop we analyzed. This
1102 makes sure the SLP tree matches up between the main vectorized
1103 loop and the epilogue vectorized copies. */
1104 if (single_succ_edge (preheader)->dest_idx
1105 != single_succ_edge (new_bbs[0])->dest_idx)
1107 basic_block swap_bb = new_bbs[1];
1108 gcc_assert (EDGE_COUNT (swap_bb->preds) == 2);
1109 std::swap (EDGE_PRED (swap_bb, 0), EDGE_PRED (swap_bb, 1));
1110 EDGE_PRED (swap_bb, 0)->dest_idx = 0;
1111 EDGE_PRED (swap_bb, 1)->dest_idx = 1;
1113 if (duplicate_outer_loop)
1115 class loop *new_inner_loop = get_loop_copy (scalar_loop->inner);
1116 if (loop_preheader_edge (scalar_loop)->dest_idx
1117 != loop_preheader_edge (new_inner_loop)->dest_idx)
1119 basic_block swap_bb = new_inner_loop->header;
1120 gcc_assert (EDGE_COUNT (swap_bb->preds) == 2);
1121 std::swap (EDGE_PRED (swap_bb, 0), EDGE_PRED (swap_bb, 1));
1122 EDGE_PRED (swap_bb, 0)->dest_idx = 0;
1123 EDGE_PRED (swap_bb, 1)->dest_idx = 1;
1127 add_phi_args_after_copy (new_bbs, scalar_loop->num_nodes + 1, NULL);
1129 /* Skip new preheader since it's deleted if copy loop is added at entry. */
1130 for (unsigned i = (at_exit ? 0 : 1); i < scalar_loop->num_nodes + 1; i++)
1131 rename_variables_in_bb (new_bbs[i], duplicate_outer_loop);
1133 if (scalar_loop != loop)
1135 /* If we copied from SCALAR_LOOP rather than LOOP, SSA_NAMEs from
1136 SCALAR_LOOP will have current_def set to SSA_NAMEs in the new_loop,
1137 but LOOP will not. slpeel_update_phi_nodes_for_guard{1,2} expects
1138 the LOOP SSA_NAMEs (on the exit edge and edge from latch to
1139 header) to have current_def set, so copy them over. */
1140 slpeel_duplicate_current_defs_from_edges (single_exit (scalar_loop),
1141 exit);
1142 slpeel_duplicate_current_defs_from_edges (EDGE_SUCC (scalar_loop->latch,
1144 EDGE_SUCC (loop->latch, 0));
1147 if (at_exit) /* Add the loop copy at exit. */
1149 if (scalar_loop != loop)
1151 gphi_iterator gsi;
1152 new_exit = redirect_edge_and_branch (new_exit, exit_dest);
1154 for (gsi = gsi_start_phis (exit_dest); !gsi_end_p (gsi);
1155 gsi_next (&gsi))
1157 gphi *phi = gsi.phi ();
1158 tree orig_arg = PHI_ARG_DEF_FROM_EDGE (phi, e);
1159 location_t orig_locus
1160 = gimple_phi_arg_location_from_edge (phi, e);
1162 add_phi_arg (phi, orig_arg, new_exit, orig_locus);
1165 redirect_edge_and_branch_force (e, new_preheader);
1166 flush_pending_stmts (e);
1167 set_immediate_dominator (CDI_DOMINATORS, new_preheader, e->src);
1168 if (was_imm_dom || duplicate_outer_loop)
1169 set_immediate_dominator (CDI_DOMINATORS, exit_dest, new_exit->src);
1171 /* And remove the non-necessary forwarder again. Keep the other
1172 one so we have a proper pre-header for the loop at the exit edge. */
1173 redirect_edge_pred (single_succ_edge (preheader),
1174 single_pred (preheader));
1175 delete_basic_block (preheader);
1176 set_immediate_dominator (CDI_DOMINATORS, scalar_loop->header,
1177 loop_preheader_edge (scalar_loop)->src);
1179 else /* Add the copy at entry. */
1181 if (scalar_loop != loop)
1183 /* Remove the non-necessary forwarder of scalar_loop again. */
1184 redirect_edge_pred (single_succ_edge (preheader),
1185 single_pred (preheader));
1186 delete_basic_block (preheader);
1187 set_immediate_dominator (CDI_DOMINATORS, scalar_loop->header,
1188 loop_preheader_edge (scalar_loop)->src);
1189 preheader = split_edge (loop_preheader_edge (loop));
1190 entry_e = single_pred_edge (preheader);
1193 redirect_edge_and_branch_force (entry_e, new_preheader);
1194 flush_pending_stmts (entry_e);
1195 set_immediate_dominator (CDI_DOMINATORS, new_preheader, entry_e->src);
1197 redirect_edge_and_branch_force (new_exit, preheader);
1198 flush_pending_stmts (new_exit);
1199 set_immediate_dominator (CDI_DOMINATORS, preheader, new_exit->src);
1201 /* And remove the non-necessary forwarder again. Keep the other
1202 one so we have a proper pre-header for the loop at the exit edge. */
1203 redirect_edge_pred (single_succ_edge (new_preheader),
1204 single_pred (new_preheader));
1205 delete_basic_block (new_preheader);
1206 set_immediate_dominator (CDI_DOMINATORS, new_loop->header,
1207 loop_preheader_edge (new_loop)->src);
1210 if (scalar_loop != loop)
1212 /* Update new_loop->header PHIs, so that on the preheader
1213 edge they are the ones from loop rather than scalar_loop. */
1214 gphi_iterator gsi_orig, gsi_new;
1215 edge orig_e = loop_preheader_edge (loop);
1216 edge new_e = loop_preheader_edge (new_loop);
1218 for (gsi_orig = gsi_start_phis (loop->header),
1219 gsi_new = gsi_start_phis (new_loop->header);
1220 !gsi_end_p (gsi_orig) && !gsi_end_p (gsi_new);
1221 gsi_next (&gsi_orig), gsi_next (&gsi_new))
1223 gphi *orig_phi = gsi_orig.phi ();
1224 gphi *new_phi = gsi_new.phi ();
1225 tree orig_arg = PHI_ARG_DEF_FROM_EDGE (orig_phi, orig_e);
1226 location_t orig_locus
1227 = gimple_phi_arg_location_from_edge (orig_phi, orig_e);
1229 add_phi_arg (new_phi, orig_arg, new_e, orig_locus);
1233 free (new_bbs);
1234 free (bbs);
1236 checking_verify_dominators (CDI_DOMINATORS);
1238 return new_loop;
1242 /* Given the condition expression COND, put it as the last statement of
1243 GUARD_BB; set both edges' probability; set dominator of GUARD_TO to
1244 DOM_BB; return the skip edge. GUARD_TO is the target basic block to
1245 skip the loop. PROBABILITY is the skip edge's probability. Mark the
1246 new edge as irreducible if IRREDUCIBLE_P is true. */
1248 static edge
1249 slpeel_add_loop_guard (basic_block guard_bb, tree cond,
1250 basic_block guard_to, basic_block dom_bb,
1251 profile_probability probability, bool irreducible_p)
1253 gimple_stmt_iterator gsi;
1254 edge new_e, enter_e;
1255 gcond *cond_stmt;
1256 gimple_seq gimplify_stmt_list = NULL;
1258 enter_e = EDGE_SUCC (guard_bb, 0);
1259 enter_e->flags &= ~EDGE_FALLTHRU;
1260 enter_e->flags |= EDGE_FALSE_VALUE;
1261 gsi = gsi_last_bb (guard_bb);
1263 cond = force_gimple_operand_1 (cond, &gimplify_stmt_list, is_gimple_condexpr,
1264 NULL_TREE);
1265 if (gimplify_stmt_list)
1266 gsi_insert_seq_after (&gsi, gimplify_stmt_list, GSI_NEW_STMT);
1268 cond_stmt = gimple_build_cond_from_tree (cond, NULL_TREE, NULL_TREE);
1269 gsi = gsi_last_bb (guard_bb);
1270 gsi_insert_after (&gsi, cond_stmt, GSI_NEW_STMT);
1272 /* Add new edge to connect guard block to the merge/loop-exit block. */
1273 new_e = make_edge (guard_bb, guard_to, EDGE_TRUE_VALUE);
1275 new_e->probability = probability;
1276 if (irreducible_p)
1277 new_e->flags |= EDGE_IRREDUCIBLE_LOOP;
1279 enter_e->probability = probability.invert ();
1280 set_immediate_dominator (CDI_DOMINATORS, guard_to, dom_bb);
1282 /* Split enter_e to preserve LOOPS_HAVE_PREHEADERS. */
1283 if (enter_e->dest->loop_father->header == enter_e->dest)
1284 split_edge (enter_e);
1286 return new_e;
1290 /* This function verifies that the following restrictions apply to LOOP:
1291 (1) it consists of exactly 2 basic blocks - header, and an empty latch
1292 for innermost loop and 5 basic blocks for outer-loop.
1293 (2) it is single entry, single exit
1294 (3) its exit condition is the last stmt in the header
1295 (4) E is the entry/exit edge of LOOP.
1298 bool
1299 slpeel_can_duplicate_loop_p (const class loop *loop, const_edge e)
1301 edge exit_e = single_exit (loop);
1302 edge entry_e = loop_preheader_edge (loop);
1303 gcond *orig_cond = get_loop_exit_condition (loop);
1304 gimple_stmt_iterator loop_exit_gsi = gsi_last_bb (exit_e->src);
1305 unsigned int num_bb = loop->inner? 5 : 2;
1307 /* All loops have an outer scope; the only case loop->outer is NULL is for
1308 the function itself. */
1309 if (!loop_outer (loop)
1310 || loop->num_nodes != num_bb
1311 || !empty_block_p (loop->latch)
1312 || !single_exit (loop)
1313 /* Verify that new loop exit condition can be trivially modified. */
1314 || (!orig_cond || orig_cond != gsi_stmt (loop_exit_gsi))
1315 || (e != exit_e && e != entry_e))
1316 return false;
1318 return true;
1321 /* If the loop has a virtual PHI, but exit bb doesn't, create a virtual PHI
1322 in the exit bb and rename all the uses after the loop. This simplifies
1323 the *guard[12] routines, which assume loop closed SSA form for all PHIs
1324 (but normally loop closed SSA form doesn't require virtual PHIs to be
1325 in the same form). Doing this early simplifies the checking what
1326 uses should be renamed.
1328 If we create a new phi after the loop, return the definition that
1329 applies on entry to the loop, otherwise return null. */
1331 static tree
1332 create_lcssa_for_virtual_phi (class loop *loop)
1334 gphi_iterator gsi;
1335 edge exit_e = single_exit (loop);
1337 for (gsi = gsi_start_phis (loop->header); !gsi_end_p (gsi); gsi_next (&gsi))
1338 if (virtual_operand_p (gimple_phi_result (gsi_stmt (gsi))))
1340 gphi *phi = gsi.phi ();
1341 for (gsi = gsi_start_phis (exit_e->dest);
1342 !gsi_end_p (gsi); gsi_next (&gsi))
1343 if (virtual_operand_p (gimple_phi_result (gsi_stmt (gsi))))
1344 break;
1345 if (gsi_end_p (gsi))
1347 tree new_vop = copy_ssa_name (PHI_RESULT (phi));
1348 gphi *new_phi = create_phi_node (new_vop, exit_e->dest);
1349 tree vop = PHI_ARG_DEF_FROM_EDGE (phi, EDGE_SUCC (loop->latch, 0));
1350 imm_use_iterator imm_iter;
1351 gimple *stmt;
1352 use_operand_p use_p;
1354 SSA_NAME_OCCURS_IN_ABNORMAL_PHI (new_vop)
1355 = SSA_NAME_OCCURS_IN_ABNORMAL_PHI (vop);
1356 add_phi_arg (new_phi, vop, exit_e, UNKNOWN_LOCATION);
1357 gimple_phi_set_result (new_phi, new_vop);
1358 FOR_EACH_IMM_USE_STMT (stmt, imm_iter, vop)
1359 if (stmt != new_phi
1360 && !flow_bb_inside_loop_p (loop, gimple_bb (stmt)))
1361 FOR_EACH_IMM_USE_ON_STMT (use_p, imm_iter)
1362 SET_USE (use_p, new_vop);
1364 return PHI_ARG_DEF_FROM_EDGE (phi, loop_preheader_edge (loop));
1366 break;
1368 return NULL_TREE;
1371 /* Function vect_get_loop_location.
1373 Extract the location of the loop in the source code.
1374 If the loop is not well formed for vectorization, an estimated
1375 location is calculated.
1376 Return the loop location if succeed and NULL if not. */
1378 dump_user_location_t
1379 find_loop_location (class loop *loop)
1381 gimple *stmt = NULL;
1382 basic_block bb;
1383 gimple_stmt_iterator si;
1385 if (!loop)
1386 return dump_user_location_t ();
1388 stmt = get_loop_exit_condition (loop);
1390 if (stmt
1391 && LOCATION_LOCUS (gimple_location (stmt)) > BUILTINS_LOCATION)
1392 return stmt;
1394 /* If we got here the loop is probably not "well formed",
1395 try to estimate the loop location */
1397 if (!loop->header)
1398 return dump_user_location_t ();
1400 bb = loop->header;
1402 for (si = gsi_start_bb (bb); !gsi_end_p (si); gsi_next (&si))
1404 stmt = gsi_stmt (si);
1405 if (LOCATION_LOCUS (gimple_location (stmt)) > BUILTINS_LOCATION)
1406 return stmt;
1409 return dump_user_location_t ();
1412 /* Return true if the phi described by STMT_INFO defines an IV of the
1413 loop to be vectorized. */
1415 static bool
1416 iv_phi_p (stmt_vec_info stmt_info)
1418 gphi *phi = as_a <gphi *> (stmt_info->stmt);
1419 if (virtual_operand_p (PHI_RESULT (phi)))
1420 return false;
1422 if (STMT_VINFO_DEF_TYPE (stmt_info) == vect_reduction_def
1423 || STMT_VINFO_DEF_TYPE (stmt_info) == vect_double_reduction_def)
1424 return false;
1426 return true;
1429 /* Function vect_can_advance_ivs_p
1431 In case the number of iterations that LOOP iterates is unknown at compile
1432 time, an epilog loop will be generated, and the loop induction variables
1433 (IVs) will be "advanced" to the value they are supposed to take just before
1434 the epilog loop. Here we check that the access function of the loop IVs
1435 and the expression that represents the loop bound are simple enough.
1436 These restrictions will be relaxed in the future. */
1438 bool
1439 vect_can_advance_ivs_p (loop_vec_info loop_vinfo)
1441 class loop *loop = LOOP_VINFO_LOOP (loop_vinfo);
1442 basic_block bb = loop->header;
1443 gphi_iterator gsi;
1445 /* Analyze phi functions of the loop header. */
1447 if (dump_enabled_p ())
1448 dump_printf_loc (MSG_NOTE, vect_location, "vect_can_advance_ivs_p:\n");
1449 for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi))
1451 tree evolution_part;
1453 gphi *phi = gsi.phi ();
1454 stmt_vec_info phi_info = loop_vinfo->lookup_stmt (phi);
1455 if (dump_enabled_p ())
1456 dump_printf_loc (MSG_NOTE, vect_location, "Analyze phi: %G",
1457 phi_info->stmt);
1459 /* Skip virtual phi's. The data dependences that are associated with
1460 virtual defs/uses (i.e., memory accesses) are analyzed elsewhere.
1462 Skip reduction phis. */
1463 if (!iv_phi_p (phi_info))
1465 if (dump_enabled_p ())
1466 dump_printf_loc (MSG_NOTE, vect_location,
1467 "reduc or virtual phi. skip.\n");
1468 continue;
1471 /* Analyze the evolution function. */
1473 evolution_part = STMT_VINFO_LOOP_PHI_EVOLUTION_PART (phi_info);
1474 if (evolution_part == NULL_TREE)
1476 if (dump_enabled_p ())
1477 dump_printf (MSG_MISSED_OPTIMIZATION,
1478 "No access function or evolution.\n");
1479 return false;
1482 /* FORNOW: We do not transform initial conditions of IVs
1483 which evolution functions are not invariants in the loop. */
1485 if (!expr_invariant_in_loop_p (loop, evolution_part))
1487 if (dump_enabled_p ())
1488 dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
1489 "evolution not invariant in loop.\n");
1490 return false;
1493 /* FORNOW: We do not transform initial conditions of IVs
1494 which evolution functions are a polynomial of degree >= 2. */
1496 if (tree_is_chrec (evolution_part))
1498 if (dump_enabled_p ())
1499 dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
1500 "evolution is chrec.\n");
1501 return false;
1505 return true;
1509 /* Function vect_update_ivs_after_vectorizer.
1511 "Advance" the induction variables of LOOP to the value they should take
1512 after the execution of LOOP. This is currently necessary because the
1513 vectorizer does not handle induction variables that are used after the
1514 loop. Such a situation occurs when the last iterations of LOOP are
1515 peeled, because:
1516 1. We introduced new uses after LOOP for IVs that were not originally used
1517 after LOOP: the IVs of LOOP are now used by an epilog loop.
1518 2. LOOP is going to be vectorized; this means that it will iterate N/VF
1519 times, whereas the loop IVs should be bumped N times.
1521 Input:
1522 - LOOP - a loop that is going to be vectorized. The last few iterations
1523 of LOOP were peeled.
1524 - NITERS - the number of iterations that LOOP executes (before it is
1525 vectorized). i.e, the number of times the ivs should be bumped.
1526 - UPDATE_E - a successor edge of LOOP->exit that is on the (only) path
1527 coming out from LOOP on which there are uses of the LOOP ivs
1528 (this is the path from LOOP->exit to epilog_loop->preheader).
1530 The new definitions of the ivs are placed in LOOP->exit.
1531 The phi args associated with the edge UPDATE_E in the bb
1532 UPDATE_E->dest are updated accordingly.
1534 Assumption 1: Like the rest of the vectorizer, this function assumes
1535 a single loop exit that has a single predecessor.
1537 Assumption 2: The phi nodes in the LOOP header and in update_bb are
1538 organized in the same order.
1540 Assumption 3: The access function of the ivs is simple enough (see
1541 vect_can_advance_ivs_p). This assumption will be relaxed in the future.
1543 Assumption 4: Exactly one of the successors of LOOP exit-bb is on a path
1544 coming out of LOOP on which the ivs of LOOP are used (this is the path
1545 that leads to the epilog loop; other paths skip the epilog loop). This
1546 path starts with the edge UPDATE_E, and its destination (denoted update_bb)
1547 needs to have its phis updated.
1550 static void
1551 vect_update_ivs_after_vectorizer (loop_vec_info loop_vinfo,
1552 tree niters, edge update_e)
1554 gphi_iterator gsi, gsi1;
1555 class loop *loop = LOOP_VINFO_LOOP (loop_vinfo);
1556 basic_block update_bb = update_e->dest;
1557 basic_block exit_bb = single_exit (loop)->dest;
1559 /* Make sure there exists a single-predecessor exit bb: */
1560 gcc_assert (single_pred_p (exit_bb));
1561 gcc_assert (single_succ_edge (exit_bb) == update_e);
1563 for (gsi = gsi_start_phis (loop->header), gsi1 = gsi_start_phis (update_bb);
1564 !gsi_end_p (gsi) && !gsi_end_p (gsi1);
1565 gsi_next (&gsi), gsi_next (&gsi1))
1567 tree init_expr;
1568 tree step_expr, off;
1569 tree type;
1570 tree var, ni, ni_name;
1571 gimple_stmt_iterator last_gsi;
1573 gphi *phi = gsi.phi ();
1574 gphi *phi1 = gsi1.phi ();
1575 stmt_vec_info phi_info = loop_vinfo->lookup_stmt (phi);
1576 if (dump_enabled_p ())
1577 dump_printf_loc (MSG_NOTE, vect_location,
1578 "vect_update_ivs_after_vectorizer: phi: %G", phi);
1580 /* Skip reduction and virtual phis. */
1581 if (!iv_phi_p (phi_info))
1583 if (dump_enabled_p ())
1584 dump_printf_loc (MSG_NOTE, vect_location,
1585 "reduc or virtual phi. skip.\n");
1586 continue;
1589 type = TREE_TYPE (gimple_phi_result (phi));
1590 step_expr = STMT_VINFO_LOOP_PHI_EVOLUTION_PART (phi_info);
1591 step_expr = unshare_expr (step_expr);
1593 /* FORNOW: We do not support IVs whose evolution function is a polynomial
1594 of degree >= 2 or exponential. */
1595 gcc_assert (!tree_is_chrec (step_expr));
1597 init_expr = PHI_ARG_DEF_FROM_EDGE (phi, loop_preheader_edge (loop));
1599 off = fold_build2 (MULT_EXPR, TREE_TYPE (step_expr),
1600 fold_convert (TREE_TYPE (step_expr), niters),
1601 step_expr);
1602 if (POINTER_TYPE_P (type))
1603 ni = fold_build_pointer_plus (init_expr, off);
1604 else
1605 ni = fold_build2 (PLUS_EXPR, type,
1606 init_expr, fold_convert (type, off));
1608 var = create_tmp_var (type, "tmp");
1610 last_gsi = gsi_last_bb (exit_bb);
1611 gimple_seq new_stmts = NULL;
1612 ni_name = force_gimple_operand (ni, &new_stmts, false, var);
1613 /* Exit_bb shouldn't be empty. */
1614 if (!gsi_end_p (last_gsi))
1615 gsi_insert_seq_after (&last_gsi, new_stmts, GSI_SAME_STMT);
1616 else
1617 gsi_insert_seq_before (&last_gsi, new_stmts, GSI_SAME_STMT);
1619 /* Fix phi expressions in the successor bb. */
1620 adjust_phi_and_debug_stmts (phi1, update_e, ni_name);
1624 /* Return a gimple value containing the misalignment (measured in vector
1625 elements) for the loop described by LOOP_VINFO, i.e. how many elements
1626 it is away from a perfectly aligned address. Add any new statements
1627 to SEQ. */
1629 static tree
1630 get_misalign_in_elems (gimple **seq, loop_vec_info loop_vinfo)
1632 dr_vec_info *dr_info = LOOP_VINFO_UNALIGNED_DR (loop_vinfo);
1633 stmt_vec_info stmt_info = dr_info->stmt;
1634 tree vectype = STMT_VINFO_VECTYPE (stmt_info);
1636 poly_uint64 target_align = DR_TARGET_ALIGNMENT (dr_info);
1637 unsigned HOST_WIDE_INT target_align_c;
1638 tree target_align_minus_1;
1640 bool negative = tree_int_cst_compare (DR_STEP (dr_info->dr),
1641 size_zero_node) < 0;
1642 tree offset = (negative
1643 ? size_int ((-TYPE_VECTOR_SUBPARTS (vectype) + 1)
1644 * TREE_INT_CST_LOW
1645 (TYPE_SIZE_UNIT (TREE_TYPE (vectype))))
1646 : size_zero_node);
1647 tree start_addr = vect_create_addr_base_for_vector_ref (loop_vinfo,
1648 stmt_info, seq,
1649 offset);
1650 tree type = unsigned_type_for (TREE_TYPE (start_addr));
1651 if (target_align.is_constant (&target_align_c))
1652 target_align_minus_1 = build_int_cst (type, target_align_c - 1);
1653 else
1655 tree vla = build_int_cst (type, target_align);
1656 tree vla_align = fold_build2 (BIT_AND_EXPR, type, vla,
1657 fold_build2 (MINUS_EXPR, type,
1658 build_int_cst (type, 0), vla));
1659 target_align_minus_1 = fold_build2 (MINUS_EXPR, type, vla_align,
1660 build_int_cst (type, 1));
1663 HOST_WIDE_INT elem_size
1664 = int_cst_value (TYPE_SIZE_UNIT (TREE_TYPE (vectype)));
1665 tree elem_size_log = build_int_cst (type, exact_log2 (elem_size));
1667 /* Create: misalign_in_bytes = addr & (target_align - 1). */
1668 tree int_start_addr = fold_convert (type, start_addr);
1669 tree misalign_in_bytes = fold_build2 (BIT_AND_EXPR, type, int_start_addr,
1670 target_align_minus_1);
1672 /* Create: misalign_in_elems = misalign_in_bytes / element_size. */
1673 tree misalign_in_elems = fold_build2 (RSHIFT_EXPR, type, misalign_in_bytes,
1674 elem_size_log);
1676 return misalign_in_elems;
1679 /* Function vect_gen_prolog_loop_niters
1681 Generate the number of iterations which should be peeled as prolog for the
1682 loop represented by LOOP_VINFO. It is calculated as the misalignment of
1683 DR - the data reference recorded in LOOP_VINFO_UNALIGNED_DR (LOOP_VINFO).
1684 As a result, after the execution of this loop, the data reference DR will
1685 refer to an aligned location. The following computation is generated:
1687 If the misalignment of DR is known at compile time:
1688 addr_mis = int mis = DR_MISALIGNMENT (dr);
1689 Else, compute address misalignment in bytes:
1690 addr_mis = addr & (target_align - 1)
1692 prolog_niters = ((VF - addr_mis/elem_size)&(VF-1))/step
1694 (elem_size = element type size; an element is the scalar element whose type
1695 is the inner type of the vectype)
1697 The computations will be emitted at the end of BB. We also compute and
1698 store upper bound (included) of the result in BOUND.
1700 When the step of the data-ref in the loop is not 1 (as in interleaved data
1701 and SLP), the number of iterations of the prolog must be divided by the step
1702 (which is equal to the size of interleaved group).
1704 The above formulas assume that VF == number of elements in the vector. This
1705 may not hold when there are multiple-types in the loop.
1706 In this case, for some data-references in the loop the VF does not represent
1707 the number of elements that fit in the vector. Therefore, instead of VF we
1708 use TYPE_VECTOR_SUBPARTS. */
1710 static tree
1711 vect_gen_prolog_loop_niters (loop_vec_info loop_vinfo,
1712 basic_block bb, int *bound)
1714 dr_vec_info *dr_info = LOOP_VINFO_UNALIGNED_DR (loop_vinfo);
1715 tree var;
1716 tree niters_type = TREE_TYPE (LOOP_VINFO_NITERS (loop_vinfo));
1717 gimple_seq stmts = NULL, new_stmts = NULL;
1718 tree iters, iters_name;
1719 stmt_vec_info stmt_info = dr_info->stmt;
1720 tree vectype = STMT_VINFO_VECTYPE (stmt_info);
1721 poly_uint64 target_align = DR_TARGET_ALIGNMENT (dr_info);
1723 if (LOOP_VINFO_PEELING_FOR_ALIGNMENT (loop_vinfo) > 0)
1725 int npeel = LOOP_VINFO_PEELING_FOR_ALIGNMENT (loop_vinfo);
1727 if (dump_enabled_p ())
1728 dump_printf_loc (MSG_NOTE, vect_location,
1729 "known peeling = %d.\n", npeel);
1731 iters = build_int_cst (niters_type, npeel);
1732 *bound = LOOP_VINFO_PEELING_FOR_ALIGNMENT (loop_vinfo);
1734 else
1736 tree misalign_in_elems = get_misalign_in_elems (&stmts, loop_vinfo);
1737 tree type = TREE_TYPE (misalign_in_elems);
1738 HOST_WIDE_INT elem_size
1739 = int_cst_value (TYPE_SIZE_UNIT (TREE_TYPE (vectype)));
1740 /* We only do prolog peeling if the target alignment is known at compile
1741 time. */
1742 poly_uint64 align_in_elems =
1743 exact_div (target_align, elem_size);
1744 tree align_in_elems_minus_1 =
1745 build_int_cst (type, align_in_elems - 1);
1746 tree align_in_elems_tree = build_int_cst (type, align_in_elems);
1748 /* Create: (niters_type) ((align_in_elems - misalign_in_elems)
1749 & (align_in_elems - 1)). */
1750 bool negative = tree_int_cst_compare (DR_STEP (dr_info->dr),
1751 size_zero_node) < 0;
1752 if (negative)
1753 iters = fold_build2 (MINUS_EXPR, type, misalign_in_elems,
1754 align_in_elems_tree);
1755 else
1756 iters = fold_build2 (MINUS_EXPR, type, align_in_elems_tree,
1757 misalign_in_elems);
1758 iters = fold_build2 (BIT_AND_EXPR, type, iters, align_in_elems_minus_1);
1759 iters = fold_convert (niters_type, iters);
1760 unsigned HOST_WIDE_INT align_in_elems_c;
1761 if (align_in_elems.is_constant (&align_in_elems_c))
1762 *bound = align_in_elems_c - 1;
1763 else
1764 *bound = -1;
1767 if (dump_enabled_p ())
1768 dump_printf_loc (MSG_NOTE, vect_location,
1769 "niters for prolog loop: %T\n", iters);
1771 var = create_tmp_var (niters_type, "prolog_loop_niters");
1772 iters_name = force_gimple_operand (iters, &new_stmts, false, var);
1774 if (new_stmts)
1775 gimple_seq_add_seq (&stmts, new_stmts);
1776 if (stmts)
1778 gcc_assert (single_succ_p (bb));
1779 gimple_stmt_iterator gsi = gsi_last_bb (bb);
1780 if (gsi_end_p (gsi))
1781 gsi_insert_seq_before (&gsi, stmts, GSI_SAME_STMT);
1782 else
1783 gsi_insert_seq_after (&gsi, stmts, GSI_SAME_STMT);
1785 return iters_name;
1789 /* Function vect_update_init_of_dr
1791 If CODE is PLUS, the vector loop starts NITERS iterations after the
1792 scalar one, otherwise CODE is MINUS and the vector loop starts NITERS
1793 iterations before the scalar one (using masking to skip inactive
1794 elements). This function updates the information recorded in DR to
1795 account for the difference. Specifically, it updates the OFFSET
1796 field of DR_INFO. */
1798 static void
1799 vect_update_init_of_dr (dr_vec_info *dr_info, tree niters, tree_code code)
1801 struct data_reference *dr = dr_info->dr;
1802 tree offset = dr_info->offset;
1803 if (!offset)
1804 offset = build_zero_cst (sizetype);
1806 niters = fold_build2 (MULT_EXPR, sizetype,
1807 fold_convert (sizetype, niters),
1808 fold_convert (sizetype, DR_STEP (dr)));
1809 offset = fold_build2 (code, sizetype,
1810 fold_convert (sizetype, offset), niters);
1811 dr_info->offset = offset;
1815 /* Function vect_update_inits_of_drs
1817 Apply vect_update_inits_of_dr to all accesses in LOOP_VINFO.
1818 CODE and NITERS are as for vect_update_inits_of_dr. */
1820 void
1821 vect_update_inits_of_drs (loop_vec_info loop_vinfo, tree niters,
1822 tree_code code)
1824 unsigned int i;
1825 vec<data_reference_p> datarefs = LOOP_VINFO_DATAREFS (loop_vinfo);
1826 struct data_reference *dr;
1828 DUMP_VECT_SCOPE ("vect_update_inits_of_dr");
1830 /* Adjust niters to sizetype. We used to insert the stmts on loop preheader
1831 here, but since we might use these niters to update the epilogues niters
1832 and data references we can't insert them here as this definition might not
1833 always dominate its uses. */
1834 if (!types_compatible_p (sizetype, TREE_TYPE (niters)))
1835 niters = fold_convert (sizetype, niters);
1837 FOR_EACH_VEC_ELT (datarefs, i, dr)
1839 dr_vec_info *dr_info = loop_vinfo->lookup_dr (dr);
1840 if (!STMT_VINFO_GATHER_SCATTER_P (dr_info->stmt)
1841 && !STMT_VINFO_SIMD_LANE_ACCESS_P (dr_info->stmt))
1842 vect_update_init_of_dr (dr_info, niters, code);
1846 /* For the information recorded in LOOP_VINFO prepare the loop for peeling
1847 by masking. This involves calculating the number of iterations to
1848 be peeled and then aligning all memory references appropriately. */
1850 void
1851 vect_prepare_for_masked_peels (loop_vec_info loop_vinfo)
1853 tree misalign_in_elems;
1854 tree type = LOOP_VINFO_RGROUP_COMPARE_TYPE (loop_vinfo);
1856 gcc_assert (vect_use_loop_mask_for_alignment_p (loop_vinfo));
1858 /* From the information recorded in LOOP_VINFO get the number of iterations
1859 that need to be skipped via masking. */
1860 if (LOOP_VINFO_PEELING_FOR_ALIGNMENT (loop_vinfo) > 0)
1862 poly_int64 misalign = (LOOP_VINFO_VECT_FACTOR (loop_vinfo)
1863 - LOOP_VINFO_PEELING_FOR_ALIGNMENT (loop_vinfo));
1864 misalign_in_elems = build_int_cst (type, misalign);
1866 else
1868 gimple_seq seq1 = NULL, seq2 = NULL;
1869 misalign_in_elems = get_misalign_in_elems (&seq1, loop_vinfo);
1870 misalign_in_elems = fold_convert (type, misalign_in_elems);
1871 misalign_in_elems = force_gimple_operand (misalign_in_elems,
1872 &seq2, true, NULL_TREE);
1873 gimple_seq_add_seq (&seq1, seq2);
1874 if (seq1)
1876 edge pe = loop_preheader_edge (LOOP_VINFO_LOOP (loop_vinfo));
1877 basic_block new_bb = gsi_insert_seq_on_edge_immediate (pe, seq1);
1878 gcc_assert (!new_bb);
1882 if (dump_enabled_p ())
1883 dump_printf_loc (MSG_NOTE, vect_location,
1884 "misalignment for fully-masked loop: %T\n",
1885 misalign_in_elems);
1887 LOOP_VINFO_MASK_SKIP_NITERS (loop_vinfo) = misalign_in_elems;
1889 vect_update_inits_of_drs (loop_vinfo, misalign_in_elems, MINUS_EXPR);
1892 /* This function builds ni_name = number of iterations. Statements
1893 are emitted on the loop preheader edge. If NEW_VAR_P is not NULL, set
1894 it to TRUE if new ssa_var is generated. */
1896 tree
1897 vect_build_loop_niters (loop_vec_info loop_vinfo, bool *new_var_p)
1899 tree ni = unshare_expr (LOOP_VINFO_NITERS (loop_vinfo));
1900 if (TREE_CODE (ni) == INTEGER_CST)
1901 return ni;
1902 else
1904 tree ni_name, var;
1905 gimple_seq stmts = NULL;
1906 edge pe = loop_preheader_edge (LOOP_VINFO_LOOP (loop_vinfo));
1908 var = create_tmp_var (TREE_TYPE (ni), "niters");
1909 ni_name = force_gimple_operand (ni, &stmts, false, var);
1910 if (stmts)
1912 gsi_insert_seq_on_edge_immediate (pe, stmts);
1913 if (new_var_p != NULL)
1914 *new_var_p = true;
1917 return ni_name;
1921 /* Calculate the number of iterations above which vectorized loop will be
1922 preferred than scalar loop. NITERS_PROLOG is the number of iterations
1923 of prolog loop. If it's integer const, the integer number is also passed
1924 in INT_NITERS_PROLOG. BOUND_PROLOG is the upper bound (inclusive) of the
1925 number of iterations of the prolog loop. BOUND_EPILOG is the corresponding
1926 value for the epilog loop. If CHECK_PROFITABILITY is true, TH is the
1927 threshold below which the scalar (rather than vectorized) loop will be
1928 executed. This function stores the upper bound (inclusive) of the result
1929 in BOUND_SCALAR. */
1931 static tree
1932 vect_gen_scalar_loop_niters (tree niters_prolog, int int_niters_prolog,
1933 int bound_prolog, poly_int64 bound_epilog, int th,
1934 poly_uint64 *bound_scalar,
1935 bool check_profitability)
1937 tree type = TREE_TYPE (niters_prolog);
1938 tree niters = fold_build2 (PLUS_EXPR, type, niters_prolog,
1939 build_int_cst (type, bound_epilog));
1941 *bound_scalar = bound_prolog + bound_epilog;
1942 if (check_profitability)
1944 /* TH indicates the minimum niters of vectorized loop, while we
1945 compute the maximum niters of scalar loop. */
1946 th--;
1947 /* Peeling for constant times. */
1948 if (int_niters_prolog >= 0)
1950 *bound_scalar = upper_bound (int_niters_prolog + bound_epilog, th);
1951 return build_int_cst (type, *bound_scalar);
1953 /* Peeling an unknown number of times. Note that both BOUND_PROLOG
1954 and BOUND_EPILOG are inclusive upper bounds. */
1955 if (known_ge (th, bound_prolog + bound_epilog))
1957 *bound_scalar = th;
1958 return build_int_cst (type, th);
1960 /* Need to do runtime comparison. */
1961 else if (maybe_gt (th, bound_epilog))
1963 *bound_scalar = upper_bound (*bound_scalar, th);
1964 return fold_build2 (MAX_EXPR, type,
1965 build_int_cst (type, th), niters);
1968 return niters;
1971 /* NITERS is the number of times that the original scalar loop executes
1972 after peeling. Work out the maximum number of iterations N that can
1973 be handled by the vectorized form of the loop and then either:
1975 a) set *STEP_VECTOR_PTR to the vectorization factor and generate:
1977 niters_vector = N
1979 b) set *STEP_VECTOR_PTR to one and generate:
1981 niters_vector = N / vf
1983 In both cases, store niters_vector in *NITERS_VECTOR_PTR and add
1984 any new statements on the loop preheader edge. NITERS_NO_OVERFLOW
1985 is true if NITERS doesn't overflow (i.e. if NITERS is always nonzero). */
1987 void
1988 vect_gen_vector_loop_niters (loop_vec_info loop_vinfo, tree niters,
1989 tree *niters_vector_ptr, tree *step_vector_ptr,
1990 bool niters_no_overflow)
1992 tree ni_minus_gap, var;
1993 tree niters_vector, step_vector, type = TREE_TYPE (niters);
1994 poly_uint64 vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo);
1995 edge pe = loop_preheader_edge (LOOP_VINFO_LOOP (loop_vinfo));
1996 tree log_vf = NULL_TREE;
1998 /* If epilogue loop is required because of data accesses with gaps, we
1999 subtract one iteration from the total number of iterations here for
2000 correct calculation of RATIO. */
2001 if (LOOP_VINFO_PEELING_FOR_GAPS (loop_vinfo))
2003 ni_minus_gap = fold_build2 (MINUS_EXPR, type, niters,
2004 build_one_cst (type));
2005 if (!is_gimple_val (ni_minus_gap))
2007 var = create_tmp_var (type, "ni_gap");
2008 gimple *stmts = NULL;
2009 ni_minus_gap = force_gimple_operand (ni_minus_gap, &stmts,
2010 true, var);
2011 gsi_insert_seq_on_edge_immediate (pe, stmts);
2014 else
2015 ni_minus_gap = niters;
2017 unsigned HOST_WIDE_INT const_vf;
2018 if (vf.is_constant (&const_vf)
2019 && !LOOP_VINFO_USING_PARTIAL_VECTORS_P (loop_vinfo))
2021 /* Create: niters >> log2(vf) */
2022 /* If it's known that niters == number of latch executions + 1 doesn't
2023 overflow, we can generate niters >> log2(vf); otherwise we generate
2024 (niters - vf) >> log2(vf) + 1 by using the fact that we know ratio
2025 will be at least one. */
2026 log_vf = build_int_cst (type, exact_log2 (const_vf));
2027 if (niters_no_overflow)
2028 niters_vector = fold_build2 (RSHIFT_EXPR, type, ni_minus_gap, log_vf);
2029 else
2030 niters_vector
2031 = fold_build2 (PLUS_EXPR, type,
2032 fold_build2 (RSHIFT_EXPR, type,
2033 fold_build2 (MINUS_EXPR, type,
2034 ni_minus_gap,
2035 build_int_cst (type, vf)),
2036 log_vf),
2037 build_int_cst (type, 1));
2038 step_vector = build_one_cst (type);
2040 else
2042 niters_vector = ni_minus_gap;
2043 step_vector = build_int_cst (type, vf);
2046 if (!is_gimple_val (niters_vector))
2048 var = create_tmp_var (type, "bnd");
2049 gimple_seq stmts = NULL;
2050 niters_vector = force_gimple_operand (niters_vector, &stmts, true, var);
2051 gsi_insert_seq_on_edge_immediate (pe, stmts);
2052 /* Peeling algorithm guarantees that vector loop bound is at least ONE,
2053 we set range information to make niters analyzer's life easier.
2054 Note the number of latch iteration value can be TYPE_MAX_VALUE so
2055 we have to represent the vector niter TYPE_MAX_VALUE + 1 >> log_vf. */
2056 if (stmts != NULL && log_vf)
2058 if (niters_no_overflow)
2059 set_range_info (niters_vector, VR_RANGE,
2060 wi::one (TYPE_PRECISION (type)),
2061 wi::rshift (wi::max_value (TYPE_PRECISION (type),
2062 TYPE_SIGN (type)),
2063 exact_log2 (const_vf),
2064 TYPE_SIGN (type)));
2065 /* For VF == 1 the vector IV might also overflow so we cannot
2066 assert a minimum value of 1. */
2067 else if (const_vf > 1)
2068 set_range_info (niters_vector, VR_RANGE,
2069 wi::one (TYPE_PRECISION (type)),
2070 wi::rshift (wi::max_value (TYPE_PRECISION (type),
2071 TYPE_SIGN (type))
2072 - (const_vf - 1),
2073 exact_log2 (const_vf), TYPE_SIGN (type))
2074 + 1);
2077 *niters_vector_ptr = niters_vector;
2078 *step_vector_ptr = step_vector;
2080 return;
2083 /* Given NITERS_VECTOR which is the number of iterations for vectorized
2084 loop specified by LOOP_VINFO after vectorization, compute the number
2085 of iterations before vectorization (niters_vector * vf) and store it
2086 to NITERS_VECTOR_MULT_VF_PTR. */
2088 static void
2089 vect_gen_vector_loop_niters_mult_vf (loop_vec_info loop_vinfo,
2090 tree niters_vector,
2091 tree *niters_vector_mult_vf_ptr)
2093 /* We should be using a step_vector of VF if VF is variable. */
2094 int vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo).to_constant ();
2095 class loop *loop = LOOP_VINFO_LOOP (loop_vinfo);
2096 tree type = TREE_TYPE (niters_vector);
2097 tree log_vf = build_int_cst (type, exact_log2 (vf));
2098 basic_block exit_bb = single_exit (loop)->dest;
2100 gcc_assert (niters_vector_mult_vf_ptr != NULL);
2101 tree niters_vector_mult_vf = fold_build2 (LSHIFT_EXPR, type,
2102 niters_vector, log_vf);
2103 if (!is_gimple_val (niters_vector_mult_vf))
2105 tree var = create_tmp_var (type, "niters_vector_mult_vf");
2106 gimple_seq stmts = NULL;
2107 niters_vector_mult_vf = force_gimple_operand (niters_vector_mult_vf,
2108 &stmts, true, var);
2109 gimple_stmt_iterator gsi = gsi_start_bb (exit_bb);
2110 gsi_insert_seq_before (&gsi, stmts, GSI_SAME_STMT);
2112 *niters_vector_mult_vf_ptr = niters_vector_mult_vf;
2115 /* LCSSA_PHI is a lcssa phi of EPILOG loop which is copied from LOOP,
2116 this function searches for the corresponding lcssa phi node in exit
2117 bb of LOOP. If it is found, return the phi result; otherwise return
2118 NULL. */
2120 static tree
2121 find_guard_arg (class loop *loop, class loop *epilog ATTRIBUTE_UNUSED,
2122 gphi *lcssa_phi)
2124 gphi_iterator gsi;
2125 edge e = single_exit (loop);
2127 gcc_assert (single_pred_p (e->dest));
2128 for (gsi = gsi_start_phis (e->dest); !gsi_end_p (gsi); gsi_next (&gsi))
2130 gphi *phi = gsi.phi ();
2131 if (operand_equal_p (PHI_ARG_DEF (phi, 0),
2132 PHI_ARG_DEF (lcssa_phi, 0), 0))
2133 return PHI_RESULT (phi);
2135 return NULL_TREE;
2138 /* Function slpeel_tree_duplicate_loop_to_edge_cfg duplciates FIRST/SECOND
2139 from SECOND/FIRST and puts it at the original loop's preheader/exit
2140 edge, the two loops are arranged as below:
2142 preheader_a:
2143 first_loop:
2144 header_a:
2145 i_1 = PHI<i_0, i_2>;
2147 i_2 = i_1 + 1;
2148 if (cond_a)
2149 goto latch_a;
2150 else
2151 goto between_bb;
2152 latch_a:
2153 goto header_a;
2155 between_bb:
2156 ;; i_x = PHI<i_2>; ;; LCSSA phi node to be created for FIRST,
2158 second_loop:
2159 header_b:
2160 i_3 = PHI<i_0, i_4>; ;; Use of i_0 to be replaced with i_x,
2161 or with i_2 if no LCSSA phi is created
2162 under condition of CREATE_LCSSA_FOR_IV_PHIS.
2164 i_4 = i_3 + 1;
2165 if (cond_b)
2166 goto latch_b;
2167 else
2168 goto exit_bb;
2169 latch_b:
2170 goto header_b;
2172 exit_bb:
2174 This function creates loop closed SSA for the first loop; update the
2175 second loop's PHI nodes by replacing argument on incoming edge with the
2176 result of newly created lcssa PHI nodes. IF CREATE_LCSSA_FOR_IV_PHIS
2177 is false, Loop closed ssa phis will only be created for non-iv phis for
2178 the first loop.
2180 This function assumes exit bb of the first loop is preheader bb of the
2181 second loop, i.e, between_bb in the example code. With PHIs updated,
2182 the second loop will execute rest iterations of the first. */
2184 static void
2185 slpeel_update_phi_nodes_for_loops (loop_vec_info loop_vinfo,
2186 class loop *first, class loop *second,
2187 bool create_lcssa_for_iv_phis)
2189 gphi_iterator gsi_update, gsi_orig;
2190 class loop *loop = LOOP_VINFO_LOOP (loop_vinfo);
2192 edge first_latch_e = EDGE_SUCC (first->latch, 0);
2193 edge second_preheader_e = loop_preheader_edge (second);
2194 basic_block between_bb = single_exit (first)->dest;
2196 gcc_assert (between_bb == second_preheader_e->src);
2197 gcc_assert (single_pred_p (between_bb) && single_succ_p (between_bb));
2198 /* Either the first loop or the second is the loop to be vectorized. */
2199 gcc_assert (loop == first || loop == second);
2201 for (gsi_orig = gsi_start_phis (first->header),
2202 gsi_update = gsi_start_phis (second->header);
2203 !gsi_end_p (gsi_orig) && !gsi_end_p (gsi_update);
2204 gsi_next (&gsi_orig), gsi_next (&gsi_update))
2206 gphi *orig_phi = gsi_orig.phi ();
2207 gphi *update_phi = gsi_update.phi ();
2209 tree arg = PHI_ARG_DEF_FROM_EDGE (orig_phi, first_latch_e);
2210 /* Generate lcssa PHI node for the first loop. */
2211 gphi *vect_phi = (loop == first) ? orig_phi : update_phi;
2212 stmt_vec_info vect_phi_info = loop_vinfo->lookup_stmt (vect_phi);
2213 if (create_lcssa_for_iv_phis || !iv_phi_p (vect_phi_info))
2215 tree new_res = copy_ssa_name (PHI_RESULT (orig_phi));
2216 gphi *lcssa_phi = create_phi_node (new_res, between_bb);
2217 add_phi_arg (lcssa_phi, arg, single_exit (first), UNKNOWN_LOCATION);
2218 arg = new_res;
2221 /* Update PHI node in the second loop by replacing arg on the loop's
2222 incoming edge. */
2223 adjust_phi_and_debug_stmts (update_phi, second_preheader_e, arg);
2226 /* For epilogue peeling we have to make sure to copy all LC PHIs
2227 for correct vectorization of live stmts. */
2228 if (loop == first)
2230 basic_block orig_exit = single_exit (second)->dest;
2231 for (gsi_orig = gsi_start_phis (orig_exit);
2232 !gsi_end_p (gsi_orig); gsi_next (&gsi_orig))
2234 gphi *orig_phi = gsi_orig.phi ();
2235 tree orig_arg = PHI_ARG_DEF (orig_phi, 0);
2236 if (TREE_CODE (orig_arg) != SSA_NAME || virtual_operand_p (orig_arg))
2237 continue;
2239 /* Already created in the above loop. */
2240 if (find_guard_arg (first, second, orig_phi))
2241 continue;
2243 tree new_res = copy_ssa_name (orig_arg);
2244 gphi *lcphi = create_phi_node (new_res, between_bb);
2245 add_phi_arg (lcphi, orig_arg, single_exit (first), UNKNOWN_LOCATION);
2250 /* Function slpeel_add_loop_guard adds guard skipping from the beginning
2251 of SKIP_LOOP to the beginning of UPDATE_LOOP. GUARD_EDGE and MERGE_EDGE
2252 are two pred edges of the merge point before UPDATE_LOOP. The two loops
2253 appear like below:
2255 guard_bb:
2256 if (cond)
2257 goto merge_bb;
2258 else
2259 goto skip_loop;
2261 skip_loop:
2262 header_a:
2263 i_1 = PHI<i_0, i_2>;
2265 i_2 = i_1 + 1;
2266 if (cond_a)
2267 goto latch_a;
2268 else
2269 goto exit_a;
2270 latch_a:
2271 goto header_a;
2273 exit_a:
2274 i_5 = PHI<i_2>;
2276 merge_bb:
2277 ;; PHI (i_x = PHI<i_0, i_5>) to be created at merge point.
2279 update_loop:
2280 header_b:
2281 i_3 = PHI<i_5, i_4>; ;; Use of i_5 to be replaced with i_x.
2283 i_4 = i_3 + 1;
2284 if (cond_b)
2285 goto latch_b;
2286 else
2287 goto exit_bb;
2288 latch_b:
2289 goto header_b;
2291 exit_bb:
2293 This function creates PHI nodes at merge_bb and replaces the use of i_5
2294 in the update_loop's PHI node with the result of new PHI result. */
2296 static void
2297 slpeel_update_phi_nodes_for_guard1 (class loop *skip_loop,
2298 class loop *update_loop,
2299 edge guard_edge, edge merge_edge)
2301 location_t merge_loc, guard_loc;
2302 edge orig_e = loop_preheader_edge (skip_loop);
2303 edge update_e = loop_preheader_edge (update_loop);
2304 gphi_iterator gsi_orig, gsi_update;
2306 for ((gsi_orig = gsi_start_phis (skip_loop->header),
2307 gsi_update = gsi_start_phis (update_loop->header));
2308 !gsi_end_p (gsi_orig) && !gsi_end_p (gsi_update);
2309 gsi_next (&gsi_orig), gsi_next (&gsi_update))
2311 gphi *orig_phi = gsi_orig.phi ();
2312 gphi *update_phi = gsi_update.phi ();
2314 /* Generate new phi node at merge bb of the guard. */
2315 tree new_res = copy_ssa_name (PHI_RESULT (orig_phi));
2316 gphi *new_phi = create_phi_node (new_res, guard_edge->dest);
2318 /* Merge bb has two incoming edges: GUARD_EDGE and MERGE_EDGE. Set the
2319 args in NEW_PHI for these edges. */
2320 tree merge_arg = PHI_ARG_DEF_FROM_EDGE (update_phi, update_e);
2321 tree guard_arg = PHI_ARG_DEF_FROM_EDGE (orig_phi, orig_e);
2322 merge_loc = gimple_phi_arg_location_from_edge (update_phi, update_e);
2323 guard_loc = gimple_phi_arg_location_from_edge (orig_phi, orig_e);
2324 add_phi_arg (new_phi, merge_arg, merge_edge, merge_loc);
2325 add_phi_arg (new_phi, guard_arg, guard_edge, guard_loc);
2327 /* Update phi in UPDATE_PHI. */
2328 adjust_phi_and_debug_stmts (update_phi, update_e, new_res);
2332 /* LOOP and EPILOG are two consecutive loops in CFG and EPILOG is copied
2333 from LOOP. Function slpeel_add_loop_guard adds guard skipping from a
2334 point between the two loops to the end of EPILOG. Edges GUARD_EDGE
2335 and MERGE_EDGE are the two pred edges of merge_bb at the end of EPILOG.
2336 The CFG looks like:
2338 loop:
2339 header_a:
2340 i_1 = PHI<i_0, i_2>;
2342 i_2 = i_1 + 1;
2343 if (cond_a)
2344 goto latch_a;
2345 else
2346 goto exit_a;
2347 latch_a:
2348 goto header_a;
2350 exit_a:
2352 guard_bb:
2353 if (cond)
2354 goto merge_bb;
2355 else
2356 goto epilog_loop;
2358 ;; fall_through_bb
2360 epilog_loop:
2361 header_b:
2362 i_3 = PHI<i_2, i_4>;
2364 i_4 = i_3 + 1;
2365 if (cond_b)
2366 goto latch_b;
2367 else
2368 goto merge_bb;
2369 latch_b:
2370 goto header_b;
2372 merge_bb:
2373 ; PHI node (i_y = PHI<i_2, i_4>) to be created at merge point.
2375 exit_bb:
2376 i_x = PHI<i_4>; ;Use of i_4 to be replaced with i_y in merge_bb.
2378 For each name used out side EPILOG (i.e - for each name that has a lcssa
2379 phi in exit_bb) we create a new PHI in merge_bb. The new PHI has two
2380 args corresponding to GUARD_EDGE and MERGE_EDGE. Arg for MERGE_EDGE is
2381 the arg of the original PHI in exit_bb, arg for GUARD_EDGE is defined
2382 by LOOP and is found in the exit bb of LOOP. Arg of the original PHI
2383 in exit_bb will also be updated. */
2385 static void
2386 slpeel_update_phi_nodes_for_guard2 (class loop *loop, class loop *epilog,
2387 edge guard_edge, edge merge_edge)
2389 gphi_iterator gsi;
2390 basic_block merge_bb = guard_edge->dest;
2392 gcc_assert (single_succ_p (merge_bb));
2393 edge e = single_succ_edge (merge_bb);
2394 basic_block exit_bb = e->dest;
2395 gcc_assert (single_pred_p (exit_bb));
2396 gcc_assert (single_pred (exit_bb) == single_exit (epilog)->dest);
2398 for (gsi = gsi_start_phis (exit_bb); !gsi_end_p (gsi); gsi_next (&gsi))
2400 gphi *update_phi = gsi.phi ();
2401 tree old_arg = PHI_ARG_DEF (update_phi, 0);
2403 tree merge_arg = NULL_TREE;
2405 /* If the old argument is a SSA_NAME use its current_def. */
2406 if (TREE_CODE (old_arg) == SSA_NAME)
2407 merge_arg = get_current_def (old_arg);
2408 /* If it's a constant or doesn't have a current_def, just use the old
2409 argument. */
2410 if (!merge_arg)
2411 merge_arg = old_arg;
2413 tree guard_arg = find_guard_arg (loop, epilog, update_phi);
2414 /* If the var is live after loop but not a reduction, we simply
2415 use the old arg. */
2416 if (!guard_arg)
2417 guard_arg = old_arg;
2419 /* Create new phi node in MERGE_BB: */
2420 tree new_res = copy_ssa_name (PHI_RESULT (update_phi));
2421 gphi *merge_phi = create_phi_node (new_res, merge_bb);
2423 /* MERGE_BB has two incoming edges: GUARD_EDGE and MERGE_EDGE, Set
2424 the two PHI args in merge_phi for these edges. */
2425 add_phi_arg (merge_phi, merge_arg, merge_edge, UNKNOWN_LOCATION);
2426 add_phi_arg (merge_phi, guard_arg, guard_edge, UNKNOWN_LOCATION);
2428 /* Update the original phi in exit_bb. */
2429 adjust_phi_and_debug_stmts (update_phi, e, new_res);
2433 /* EPILOG loop is duplicated from the original loop for vectorizing,
2434 the arg of its loop closed ssa PHI needs to be updated. */
2436 static void
2437 slpeel_update_phi_nodes_for_lcssa (class loop *epilog)
2439 gphi_iterator gsi;
2440 basic_block exit_bb = single_exit (epilog)->dest;
2442 gcc_assert (single_pred_p (exit_bb));
2443 edge e = EDGE_PRED (exit_bb, 0);
2444 for (gsi = gsi_start_phis (exit_bb); !gsi_end_p (gsi); gsi_next (&gsi))
2445 rename_use_op (PHI_ARG_DEF_PTR_FROM_EDGE (gsi.phi (), e));
2448 /* EPILOGUE_VINFO is an epilogue loop that we now know would need to
2449 iterate exactly CONST_NITERS times. Make a final decision about
2450 whether the epilogue loop should be used, returning true if so. */
2452 static bool
2453 vect_update_epilogue_niters (loop_vec_info epilogue_vinfo,
2454 unsigned HOST_WIDE_INT const_niters)
2456 /* Avoid wrap-around when computing const_niters - 1. Also reject
2457 using an epilogue loop for a single scalar iteration, even if
2458 we could in principle implement that using partial vectors. */
2459 unsigned int gap_niters = LOOP_VINFO_PEELING_FOR_GAPS (epilogue_vinfo);
2460 if (const_niters <= gap_niters + 1)
2461 return false;
2463 /* Install the number of iterations. */
2464 tree niters_type = TREE_TYPE (LOOP_VINFO_NITERS (epilogue_vinfo));
2465 tree niters_tree = build_int_cst (niters_type, const_niters);
2466 tree nitersm1_tree = build_int_cst (niters_type, const_niters - 1);
2468 LOOP_VINFO_NITERS (epilogue_vinfo) = niters_tree;
2469 LOOP_VINFO_NITERSM1 (epilogue_vinfo) = nitersm1_tree;
2471 /* Decide what to do if the number of epilogue iterations is not
2472 a multiple of the epilogue loop's vectorization factor. */
2473 return vect_determine_partial_vectors_and_peeling (epilogue_vinfo, true);
2476 /* LOOP_VINFO is an epilogue loop whose corresponding main loop can be skipped.
2477 Return a value that equals:
2479 - MAIN_LOOP_VALUE when LOOP_VINFO is entered from the main loop and
2480 - SKIP_VALUE when the main loop is skipped. */
2482 tree
2483 vect_get_main_loop_result (loop_vec_info loop_vinfo, tree main_loop_value,
2484 tree skip_value)
2486 gcc_assert (loop_vinfo->main_loop_edge);
2488 tree phi_result = make_ssa_name (TREE_TYPE (main_loop_value));
2489 basic_block bb = loop_vinfo->main_loop_edge->dest;
2490 gphi *new_phi = create_phi_node (phi_result, bb);
2491 add_phi_arg (new_phi, main_loop_value, loop_vinfo->main_loop_edge,
2492 UNKNOWN_LOCATION);
2493 add_phi_arg (new_phi, skip_value,
2494 loop_vinfo->skip_main_loop_edge, UNKNOWN_LOCATION);
2495 return phi_result;
2498 /* Function vect_do_peeling.
2500 Input:
2501 - LOOP_VINFO: Represent a loop to be vectorized, which looks like:
2503 preheader:
2504 LOOP:
2505 header_bb:
2506 loop_body
2507 if (exit_loop_cond) goto exit_bb
2508 else goto header_bb
2509 exit_bb:
2511 - NITERS: The number of iterations of the loop.
2512 - NITERSM1: The number of iterations of the loop's latch.
2513 - NITERS_NO_OVERFLOW: No overflow in computing NITERS.
2514 - TH, CHECK_PROFITABILITY: Threshold of niters to vectorize loop if
2515 CHECK_PROFITABILITY is true.
2516 Output:
2517 - *NITERS_VECTOR and *STEP_VECTOR describe how the main loop should
2518 iterate after vectorization; see vect_set_loop_condition for details.
2519 - *NITERS_VECTOR_MULT_VF_VAR is either null or an SSA name that
2520 should be set to the number of scalar iterations handled by the
2521 vector loop. The SSA name is only used on exit from the loop.
2523 This function peels prolog and epilog from the loop, adds guards skipping
2524 PROLOG and EPILOG for various conditions. As a result, the changed CFG
2525 would look like:
2527 guard_bb_1:
2528 if (prefer_scalar_loop) goto merge_bb_1
2529 else goto guard_bb_2
2531 guard_bb_2:
2532 if (skip_prolog) goto merge_bb_2
2533 else goto prolog_preheader
2535 prolog_preheader:
2536 PROLOG:
2537 prolog_header_bb:
2538 prolog_body
2539 if (exit_prolog_cond) goto prolog_exit_bb
2540 else goto prolog_header_bb
2541 prolog_exit_bb:
2543 merge_bb_2:
2545 vector_preheader:
2546 VECTOR LOOP:
2547 vector_header_bb:
2548 vector_body
2549 if (exit_vector_cond) goto vector_exit_bb
2550 else goto vector_header_bb
2551 vector_exit_bb:
2553 guard_bb_3:
2554 if (skip_epilog) goto merge_bb_3
2555 else goto epilog_preheader
2557 merge_bb_1:
2559 epilog_preheader:
2560 EPILOG:
2561 epilog_header_bb:
2562 epilog_body
2563 if (exit_epilog_cond) goto merge_bb_3
2564 else goto epilog_header_bb
2566 merge_bb_3:
2568 Note this function peels prolog and epilog only if it's necessary,
2569 as well as guards.
2570 This function returns the epilogue loop if a decision was made to vectorize
2571 it, otherwise NULL.
2573 The analysis resulting in this epilogue loop's loop_vec_info was performed
2574 in the same vect_analyze_loop call as the main loop's. At that time
2575 vect_analyze_loop constructs a list of accepted loop_vec_info's for lower
2576 vectorization factors than the main loop. This list is stored in the main
2577 loop's loop_vec_info in the 'epilogue_vinfos' member. Everytime we decide to
2578 vectorize the epilogue loop for a lower vectorization factor, the
2579 loop_vec_info sitting at the top of the epilogue_vinfos list is removed,
2580 updated and linked to the epilogue loop. This is later used to vectorize
2581 the epilogue. The reason the loop_vec_info needs updating is that it was
2582 constructed based on the original main loop, and the epilogue loop is a
2583 copy of this loop, so all links pointing to statements in the original loop
2584 need updating. Furthermore, these loop_vec_infos share the
2585 data_reference's records, which will also need to be updated.
2587 TODO: Guard for prefer_scalar_loop should be emitted along with
2588 versioning conditions if loop versioning is needed. */
2591 class loop *
2592 vect_do_peeling (loop_vec_info loop_vinfo, tree niters, tree nitersm1,
2593 tree *niters_vector, tree *step_vector,
2594 tree *niters_vector_mult_vf_var, int th,
2595 bool check_profitability, bool niters_no_overflow,
2596 tree *advance)
2598 edge e, guard_e;
2599 tree type = TREE_TYPE (niters), guard_cond;
2600 basic_block guard_bb, guard_to;
2601 profile_probability prob_prolog, prob_vector, prob_epilog;
2602 int estimated_vf;
2603 int prolog_peeling = 0;
2604 bool vect_epilogues = loop_vinfo->epilogue_vinfos.length () > 0;
2605 bool vect_epilogues_updated_niters = false;
2606 /* We currently do not support prolog peeling if the target alignment is not
2607 known at compile time. 'vect_gen_prolog_loop_niters' depends on the
2608 target alignment being constant. */
2609 dr_vec_info *dr_info = LOOP_VINFO_UNALIGNED_DR (loop_vinfo);
2610 if (dr_info && !DR_TARGET_ALIGNMENT (dr_info).is_constant ())
2611 return NULL;
2613 if (!vect_use_loop_mask_for_alignment_p (loop_vinfo))
2614 prolog_peeling = LOOP_VINFO_PEELING_FOR_ALIGNMENT (loop_vinfo);
2616 poly_uint64 vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo);
2617 poly_uint64 bound_epilog = 0;
2618 if (!LOOP_VINFO_USING_PARTIAL_VECTORS_P (loop_vinfo)
2619 && LOOP_VINFO_PEELING_FOR_NITER (loop_vinfo))
2620 bound_epilog += vf - 1;
2621 if (LOOP_VINFO_PEELING_FOR_GAPS (loop_vinfo))
2622 bound_epilog += 1;
2623 bool epilog_peeling = maybe_ne (bound_epilog, 0U);
2624 poly_uint64 bound_scalar = bound_epilog;
2626 if (!prolog_peeling && !epilog_peeling)
2627 return NULL;
2629 /* Before doing any peeling make sure to reset debug binds outside of
2630 the loop refering to defs not in LC SSA. */
2631 class loop *loop = LOOP_VINFO_LOOP (loop_vinfo);
2632 for (unsigned i = 0; i < loop->num_nodes; ++i)
2634 basic_block bb = LOOP_VINFO_BBS (loop_vinfo)[i];
2635 imm_use_iterator ui;
2636 gimple *use_stmt;
2637 for (gphi_iterator gsi = gsi_start_phis (bb); !gsi_end_p (gsi);
2638 gsi_next (&gsi))
2640 FOR_EACH_IMM_USE_STMT (use_stmt, ui, gimple_phi_result (gsi.phi ()))
2641 if (gimple_debug_bind_p (use_stmt)
2642 && loop != gimple_bb (use_stmt)->loop_father
2643 && !flow_loop_nested_p (loop,
2644 gimple_bb (use_stmt)->loop_father))
2646 gimple_debug_bind_reset_value (use_stmt);
2647 update_stmt (use_stmt);
2650 for (gimple_stmt_iterator gsi = gsi_start_bb (bb); !gsi_end_p (gsi);
2651 gsi_next (&gsi))
2653 ssa_op_iter op_iter;
2654 def_operand_p def_p;
2655 FOR_EACH_SSA_DEF_OPERAND (def_p, gsi_stmt (gsi), op_iter, SSA_OP_DEF)
2656 FOR_EACH_IMM_USE_STMT (use_stmt, ui, DEF_FROM_PTR (def_p))
2657 if (gimple_debug_bind_p (use_stmt)
2658 && loop != gimple_bb (use_stmt)->loop_father
2659 && !flow_loop_nested_p (loop,
2660 gimple_bb (use_stmt)->loop_father))
2662 gimple_debug_bind_reset_value (use_stmt);
2663 update_stmt (use_stmt);
2668 prob_vector = profile_probability::guessed_always ().apply_scale (9, 10);
2669 estimated_vf = vect_vf_for_cost (loop_vinfo);
2670 if (estimated_vf == 2)
2671 estimated_vf = 3;
2672 prob_prolog = prob_epilog = profile_probability::guessed_always ()
2673 .apply_scale (estimated_vf - 1, estimated_vf);
2675 class loop *prolog, *epilog = NULL;
2676 class loop *first_loop = loop;
2677 bool irred_flag = loop_preheader_edge (loop)->flags & EDGE_IRREDUCIBLE_LOOP;
2679 /* We might have a queued need to update virtual SSA form. As we
2680 delete the update SSA machinery below after doing a regular
2681 incremental SSA update during loop copying make sure we don't
2682 lose that fact.
2683 ??? Needing to update virtual SSA form by renaming is unfortunate
2684 but not all of the vectorizer code inserting new loads / stores
2685 properly assigns virtual operands to those statements. */
2686 update_ssa (TODO_update_ssa_only_virtuals);
2688 create_lcssa_for_virtual_phi (loop);
2690 /* If we're vectorizing an epilogue loop, the update_ssa above will
2691 have ensured that the virtual operand is in SSA form throughout the
2692 vectorized main loop. Normally it is possible to trace the updated
2693 vector-stmt vdefs back to scalar-stmt vdefs and vector-stmt vuses
2694 back to scalar-stmt vuses, meaning that the effect of the SSA update
2695 remains local to the main loop. However, there are rare cases in
2696 which the vectorized loop has vdefs even when the original scalar
2697 loop didn't. For example, vectorizing a load with IFN_LOAD_LANES
2698 introduces clobbers of the temporary vector array, which in turn
2699 needs new vdefs. If the scalar loop doesn't write to memory, these
2700 new vdefs will be the only ones in the vector loop.
2702 In that case, update_ssa will have added a new virtual phi to the
2703 main loop, which previously didn't need one. Ensure that we (locally)
2704 maintain LCSSA form for the virtual operand, just as we would have
2705 done if the virtual phi had existed from the outset. This makes it
2706 easier to duplicate the scalar epilogue loop below. */
2707 tree vop_to_rename = NULL_TREE;
2708 if (loop_vec_info orig_loop_vinfo = LOOP_VINFO_ORIG_LOOP_INFO (loop_vinfo))
2710 class loop *orig_loop = LOOP_VINFO_LOOP (orig_loop_vinfo);
2711 vop_to_rename = create_lcssa_for_virtual_phi (orig_loop);
2714 if (MAY_HAVE_DEBUG_BIND_STMTS)
2716 gcc_assert (!adjust_vec.exists ());
2717 adjust_vec.create (32);
2719 initialize_original_copy_tables ();
2721 /* Record the anchor bb at which the guard should be placed if the scalar
2722 loop might be preferred. */
2723 basic_block anchor = loop_preheader_edge (loop)->src;
2725 /* Generate the number of iterations for the prolog loop. We do this here
2726 so that we can also get the upper bound on the number of iterations. */
2727 tree niters_prolog;
2728 int bound_prolog = 0;
2729 if (prolog_peeling)
2730 niters_prolog = vect_gen_prolog_loop_niters (loop_vinfo, anchor,
2731 &bound_prolog);
2732 else
2733 niters_prolog = build_int_cst (type, 0);
2735 loop_vec_info epilogue_vinfo = NULL;
2736 if (vect_epilogues)
2738 epilogue_vinfo = loop_vinfo->epilogue_vinfos[0];
2739 loop_vinfo->epilogue_vinfos.ordered_remove (0);
2742 tree niters_vector_mult_vf = NULL_TREE;
2743 /* Saving NITERs before the loop, as this may be changed by prologue. */
2744 tree before_loop_niters = LOOP_VINFO_NITERS (loop_vinfo);
2745 edge update_e = NULL, skip_e = NULL;
2746 unsigned int lowest_vf = constant_lower_bound (vf);
2747 /* If we know the number of scalar iterations for the main loop we should
2748 check whether after the main loop there are enough iterations left over
2749 for the epilogue. */
2750 if (vect_epilogues
2751 && LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo)
2752 && prolog_peeling >= 0
2753 && known_eq (vf, lowest_vf))
2755 unsigned HOST_WIDE_INT eiters
2756 = (LOOP_VINFO_INT_NITERS (loop_vinfo)
2757 - LOOP_VINFO_PEELING_FOR_GAPS (loop_vinfo));
2759 eiters -= prolog_peeling;
2760 eiters
2761 = eiters % lowest_vf + LOOP_VINFO_PEELING_FOR_GAPS (loop_vinfo);
2763 while (!vect_update_epilogue_niters (epilogue_vinfo, eiters))
2765 delete epilogue_vinfo;
2766 epilogue_vinfo = NULL;
2767 if (loop_vinfo->epilogue_vinfos.length () == 0)
2769 vect_epilogues = false;
2770 break;
2772 epilogue_vinfo = loop_vinfo->epilogue_vinfos[0];
2773 loop_vinfo->epilogue_vinfos.ordered_remove (0);
2775 vect_epilogues_updated_niters = true;
2777 /* Prolog loop may be skipped. */
2778 bool skip_prolog = (prolog_peeling != 0);
2779 /* Skip this loop to epilog when there are not enough iterations to enter this
2780 vectorized loop. If true we should perform runtime checks on the NITERS
2781 to check whether we should skip the current vectorized loop. If we know
2782 the number of scalar iterations we may choose to add a runtime check if
2783 this number "maybe" smaller than the number of iterations required
2784 when we know the number of scalar iterations may potentially
2785 be smaller than the number of iterations required to enter this loop, for
2786 this we use the upper bounds on the prolog and epilog peeling. When we
2787 don't know the number of iterations and don't require versioning it is
2788 because we have asserted that there are enough scalar iterations to enter
2789 the main loop, so this skip is not necessary. When we are versioning then
2790 we only add such a skip if we have chosen to vectorize the epilogue. */
2791 bool skip_vector = (LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo)
2792 ? maybe_lt (LOOP_VINFO_INT_NITERS (loop_vinfo),
2793 bound_prolog + bound_epilog)
2794 : (!LOOP_REQUIRES_VERSIONING (loop_vinfo)
2795 || vect_epilogues));
2796 /* Epilog loop must be executed if the number of iterations for epilog
2797 loop is known at compile time, otherwise we need to add a check at
2798 the end of vector loop and skip to the end of epilog loop. */
2799 bool skip_epilog = (prolog_peeling < 0
2800 || !LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo)
2801 || !vf.is_constant ());
2802 /* PEELING_FOR_GAPS is special because epilog loop must be executed. */
2803 if (LOOP_VINFO_PEELING_FOR_GAPS (loop_vinfo))
2804 skip_epilog = false;
2806 if (skip_vector)
2808 split_edge (loop_preheader_edge (loop));
2810 /* Due to the order in which we peel prolog and epilog, we first
2811 propagate probability to the whole loop. The purpose is to
2812 avoid adjusting probabilities of both prolog and vector loops
2813 separately. Note in this case, the probability of epilog loop
2814 needs to be scaled back later. */
2815 basic_block bb_before_loop = loop_preheader_edge (loop)->src;
2816 if (prob_vector.initialized_p ())
2818 scale_bbs_frequencies (&bb_before_loop, 1, prob_vector);
2819 scale_loop_profile (loop, prob_vector, 0);
2823 dump_user_location_t loop_loc = find_loop_location (loop);
2824 class loop *scalar_loop = LOOP_VINFO_SCALAR_LOOP (loop_vinfo);
2825 if (vect_epilogues)
2826 /* Make sure to set the epilogue's epilogue scalar loop, such that we can
2827 use the original scalar loop as remaining epilogue if necessary. */
2828 LOOP_VINFO_SCALAR_LOOP (epilogue_vinfo)
2829 = LOOP_VINFO_SCALAR_LOOP (loop_vinfo);
2831 if (prolog_peeling)
2833 e = loop_preheader_edge (loop);
2834 if (!slpeel_can_duplicate_loop_p (loop, e))
2836 dump_printf_loc (MSG_MISSED_OPTIMIZATION, loop_loc,
2837 "loop can't be duplicated to preheader edge.\n");
2838 gcc_unreachable ();
2840 /* Peel prolog and put it on preheader edge of loop. */
2841 prolog = slpeel_tree_duplicate_loop_to_edge_cfg (loop, scalar_loop, e);
2842 if (!prolog)
2844 dump_printf_loc (MSG_MISSED_OPTIMIZATION, loop_loc,
2845 "slpeel_tree_duplicate_loop_to_edge_cfg failed.\n");
2846 gcc_unreachable ();
2848 prolog->force_vectorize = false;
2849 slpeel_update_phi_nodes_for_loops (loop_vinfo, prolog, loop, true);
2850 first_loop = prolog;
2851 reset_original_copy_tables ();
2853 /* Update the number of iterations for prolog loop. */
2854 tree step_prolog = build_one_cst (TREE_TYPE (niters_prolog));
2855 vect_set_loop_condition (prolog, NULL, niters_prolog,
2856 step_prolog, NULL_TREE, false);
2858 /* Skip the prolog loop. */
2859 if (skip_prolog)
2861 guard_cond = fold_build2 (EQ_EXPR, boolean_type_node,
2862 niters_prolog, build_int_cst (type, 0));
2863 guard_bb = loop_preheader_edge (prolog)->src;
2864 basic_block bb_after_prolog = loop_preheader_edge (loop)->src;
2865 guard_to = split_edge (loop_preheader_edge (loop));
2866 guard_e = slpeel_add_loop_guard (guard_bb, guard_cond,
2867 guard_to, guard_bb,
2868 prob_prolog.invert (),
2869 irred_flag);
2870 e = EDGE_PRED (guard_to, 0);
2871 e = (e != guard_e ? e : EDGE_PRED (guard_to, 1));
2872 slpeel_update_phi_nodes_for_guard1 (prolog, loop, guard_e, e);
2874 scale_bbs_frequencies (&bb_after_prolog, 1, prob_prolog);
2875 scale_loop_profile (prolog, prob_prolog, bound_prolog);
2878 /* Update init address of DRs. */
2879 vect_update_inits_of_drs (loop_vinfo, niters_prolog, PLUS_EXPR);
2880 /* Update niters for vector loop. */
2881 LOOP_VINFO_NITERS (loop_vinfo)
2882 = fold_build2 (MINUS_EXPR, type, niters, niters_prolog);
2883 LOOP_VINFO_NITERSM1 (loop_vinfo)
2884 = fold_build2 (MINUS_EXPR, type,
2885 LOOP_VINFO_NITERSM1 (loop_vinfo), niters_prolog);
2886 bool new_var_p = false;
2887 niters = vect_build_loop_niters (loop_vinfo, &new_var_p);
2888 /* It's guaranteed that vector loop bound before vectorization is at
2889 least VF, so set range information for newly generated var. */
2890 if (new_var_p)
2891 set_range_info (niters, VR_RANGE,
2892 wi::to_wide (build_int_cst (type, vf)),
2893 wi::to_wide (TYPE_MAX_VALUE (type)));
2895 /* Prolog iterates at most bound_prolog times, latch iterates at
2896 most bound_prolog - 1 times. */
2897 record_niter_bound (prolog, bound_prolog - 1, false, true);
2898 delete_update_ssa ();
2899 adjust_vec_debug_stmts ();
2900 scev_reset ();
2903 if (epilog_peeling)
2905 e = single_exit (loop);
2906 if (!slpeel_can_duplicate_loop_p (loop, e))
2908 dump_printf_loc (MSG_MISSED_OPTIMIZATION, loop_loc,
2909 "loop can't be duplicated to exit edge.\n");
2910 gcc_unreachable ();
2912 /* Peel epilog and put it on exit edge of loop. If we are vectorizing
2913 said epilog then we should use a copy of the main loop as a starting
2914 point. This loop may have already had some preliminary transformations
2915 to allow for more optimal vectorization, for example if-conversion.
2916 If we are not vectorizing the epilog then we should use the scalar loop
2917 as the transformations mentioned above make less or no sense when not
2918 vectorizing. */
2919 epilog = vect_epilogues ? get_loop_copy (loop) : scalar_loop;
2920 if (vop_to_rename)
2922 /* Vectorizing the main loop can sometimes introduce a vdef to
2923 a loop that previously didn't have one; see the comment above
2924 the definition of VOP_TO_RENAME for details. The definition
2925 D that holds on E will then be different from the definition
2926 VOP_TO_RENAME that holds during SCALAR_LOOP, so we need to
2927 rename VOP_TO_RENAME to D when copying the loop.
2929 The virtual operand is in LCSSA form for the main loop,
2930 and no stmt between the main loop and E needs a vdef,
2931 so we know that D is provided by a phi rather than by a
2932 vdef on a normal gimple stmt. */
2933 basic_block vdef_bb = e->src;
2934 gphi *vphi;
2935 while (!(vphi = get_virtual_phi (vdef_bb)))
2936 vdef_bb = get_immediate_dominator (CDI_DOMINATORS, vdef_bb);
2937 gcc_assert (vop_to_rename != gimple_phi_result (vphi));
2938 set_current_def (vop_to_rename, gimple_phi_result (vphi));
2940 epilog = slpeel_tree_duplicate_loop_to_edge_cfg (loop, epilog, e);
2941 if (!epilog)
2943 dump_printf_loc (MSG_MISSED_OPTIMIZATION, loop_loc,
2944 "slpeel_tree_duplicate_loop_to_edge_cfg failed.\n");
2945 gcc_unreachable ();
2947 epilog->force_vectorize = false;
2948 slpeel_update_phi_nodes_for_loops (loop_vinfo, loop, epilog, false);
2950 /* Scalar version loop may be preferred. In this case, add guard
2951 and skip to epilog. Note this only happens when the number of
2952 iterations of loop is unknown at compile time, otherwise this
2953 won't be vectorized. */
2954 if (skip_vector)
2956 /* Additional epilogue iteration is peeled if gap exists. */
2957 tree t = vect_gen_scalar_loop_niters (niters_prolog, prolog_peeling,
2958 bound_prolog, bound_epilog,
2959 th, &bound_scalar,
2960 check_profitability);
2961 /* Build guard against NITERSM1 since NITERS may overflow. */
2962 guard_cond = fold_build2 (LT_EXPR, boolean_type_node, nitersm1, t);
2963 guard_bb = anchor;
2964 guard_to = split_edge (loop_preheader_edge (epilog));
2965 guard_e = slpeel_add_loop_guard (guard_bb, guard_cond,
2966 guard_to, guard_bb,
2967 prob_vector.invert (),
2968 irred_flag);
2969 skip_e = guard_e;
2970 e = EDGE_PRED (guard_to, 0);
2971 e = (e != guard_e ? e : EDGE_PRED (guard_to, 1));
2972 slpeel_update_phi_nodes_for_guard1 (first_loop, epilog, guard_e, e);
2974 /* Simply propagate profile info from guard_bb to guard_to which is
2975 a merge point of control flow. */
2976 guard_to->count = guard_bb->count;
2978 /* Scale probability of epilog loop back.
2979 FIXME: We should avoid scaling down and back up. Profile may
2980 get lost if we scale down to 0. */
2981 basic_block *bbs = get_loop_body (epilog);
2982 for (unsigned int i = 0; i < epilog->num_nodes; i++)
2983 bbs[i]->count = bbs[i]->count.apply_scale
2984 (bbs[i]->count,
2985 bbs[i]->count.apply_probability
2986 (prob_vector));
2987 free (bbs);
2990 basic_block bb_before_epilog = loop_preheader_edge (epilog)->src;
2991 /* If loop is peeled for non-zero constant times, now niters refers to
2992 orig_niters - prolog_peeling, it won't overflow even the orig_niters
2993 overflows. */
2994 niters_no_overflow |= (prolog_peeling > 0);
2995 vect_gen_vector_loop_niters (loop_vinfo, niters,
2996 niters_vector, step_vector,
2997 niters_no_overflow);
2998 if (!integer_onep (*step_vector))
3000 /* On exit from the loop we will have an easy way of calcalating
3001 NITERS_VECTOR / STEP * STEP. Install a dummy definition
3002 until then. */
3003 niters_vector_mult_vf = make_ssa_name (TREE_TYPE (*niters_vector));
3004 SSA_NAME_DEF_STMT (niters_vector_mult_vf) = gimple_build_nop ();
3005 *niters_vector_mult_vf_var = niters_vector_mult_vf;
3007 else
3008 vect_gen_vector_loop_niters_mult_vf (loop_vinfo, *niters_vector,
3009 &niters_vector_mult_vf);
3010 /* Update IVs of original loop as if they were advanced by
3011 niters_vector_mult_vf steps. */
3012 gcc_checking_assert (vect_can_advance_ivs_p (loop_vinfo));
3013 update_e = skip_vector ? e : loop_preheader_edge (epilog);
3014 vect_update_ivs_after_vectorizer (loop_vinfo, niters_vector_mult_vf,
3015 update_e);
3017 if (skip_epilog)
3019 guard_cond = fold_build2 (EQ_EXPR, boolean_type_node,
3020 niters, niters_vector_mult_vf);
3021 guard_bb = single_exit (loop)->dest;
3022 guard_to = split_edge (single_exit (epilog));
3023 guard_e = slpeel_add_loop_guard (guard_bb, guard_cond, guard_to,
3024 skip_vector ? anchor : guard_bb,
3025 prob_epilog.invert (),
3026 irred_flag);
3027 if (vect_epilogues)
3028 epilogue_vinfo->skip_this_loop_edge = guard_e;
3029 slpeel_update_phi_nodes_for_guard2 (loop, epilog, guard_e,
3030 single_exit (epilog));
3031 /* Only need to handle basic block before epilog loop if it's not
3032 the guard_bb, which is the case when skip_vector is true. */
3033 if (guard_bb != bb_before_epilog)
3035 prob_epilog = prob_vector * prob_epilog + prob_vector.invert ();
3037 scale_bbs_frequencies (&bb_before_epilog, 1, prob_epilog);
3039 scale_loop_profile (epilog, prob_epilog, 0);
3041 else
3042 slpeel_update_phi_nodes_for_lcssa (epilog);
3044 unsigned HOST_WIDE_INT bound;
3045 if (bound_scalar.is_constant (&bound))
3047 gcc_assert (bound != 0);
3048 /* -1 to convert loop iterations to latch iterations. */
3049 record_niter_bound (epilog, bound - 1, false, true);
3052 delete_update_ssa ();
3053 adjust_vec_debug_stmts ();
3054 scev_reset ();
3057 if (vect_epilogues)
3059 epilog->aux = epilogue_vinfo;
3060 LOOP_VINFO_LOOP (epilogue_vinfo) = epilog;
3062 loop_constraint_clear (epilog, LOOP_C_INFINITE);
3064 /* We now must calculate the number of NITERS performed by the previous
3065 loop and EPILOGUE_NITERS to be performed by the epilogue. */
3066 tree niters = fold_build2 (PLUS_EXPR, TREE_TYPE (niters_vector_mult_vf),
3067 niters_prolog, niters_vector_mult_vf);
3069 /* If skip_vector we may skip the previous loop, we insert a phi-node to
3070 determine whether we are coming from the previous vectorized loop
3071 using the update_e edge or the skip_vector basic block using the
3072 skip_e edge. */
3073 if (skip_vector)
3075 gcc_assert (update_e != NULL
3076 && skip_e != NULL
3077 && !vect_epilogues_updated_niters);
3078 gphi *new_phi = create_phi_node (make_ssa_name (TREE_TYPE (niters)),
3079 update_e->dest);
3080 tree new_ssa = make_ssa_name (TREE_TYPE (niters));
3081 gimple *stmt = gimple_build_assign (new_ssa, niters);
3082 gimple_stmt_iterator gsi;
3083 if (TREE_CODE (niters_vector_mult_vf) == SSA_NAME
3084 && SSA_NAME_DEF_STMT (niters_vector_mult_vf)->bb != NULL)
3086 gsi = gsi_for_stmt (SSA_NAME_DEF_STMT (niters_vector_mult_vf));
3087 gsi_insert_after (&gsi, stmt, GSI_NEW_STMT);
3089 else
3091 gsi = gsi_last_bb (update_e->src);
3092 gsi_insert_before (&gsi, stmt, GSI_NEW_STMT);
3095 niters = new_ssa;
3096 add_phi_arg (new_phi, niters, update_e, UNKNOWN_LOCATION);
3097 add_phi_arg (new_phi, build_zero_cst (TREE_TYPE (niters)), skip_e,
3098 UNKNOWN_LOCATION);
3099 niters = PHI_RESULT (new_phi);
3100 epilogue_vinfo->main_loop_edge = update_e;
3101 epilogue_vinfo->skip_main_loop_edge = skip_e;
3104 /* Set ADVANCE to the number of iterations performed by the previous
3105 loop and its prologue. */
3106 *advance = niters;
3108 if (!vect_epilogues_updated_niters)
3110 /* Subtract the number of iterations performed by the vectorized loop
3111 from the number of total iterations. */
3112 tree epilogue_niters = fold_build2 (MINUS_EXPR, TREE_TYPE (niters),
3113 before_loop_niters,
3114 niters);
3116 LOOP_VINFO_NITERS (epilogue_vinfo) = epilogue_niters;
3117 LOOP_VINFO_NITERSM1 (epilogue_vinfo)
3118 = fold_build2 (MINUS_EXPR, TREE_TYPE (epilogue_niters),
3119 epilogue_niters,
3120 build_one_cst (TREE_TYPE (epilogue_niters)));
3122 /* Decide what to do if the number of epilogue iterations is not
3123 a multiple of the epilogue loop's vectorization factor.
3124 We should have rejected the loop during the analysis phase
3125 if this fails. */
3126 if (!vect_determine_partial_vectors_and_peeling (epilogue_vinfo,
3127 true))
3128 gcc_unreachable ();
3132 adjust_vec.release ();
3133 free_original_copy_tables ();
3135 return vect_epilogues ? epilog : NULL;
3138 /* Function vect_create_cond_for_niters_checks.
3140 Create a conditional expression that represents the run-time checks for
3141 loop's niter. The loop is guaranteed to terminate if the run-time
3142 checks hold.
3144 Input:
3145 COND_EXPR - input conditional expression. New conditions will be chained
3146 with logical AND operation. If it is NULL, then the function
3147 is used to return the number of alias checks.
3148 LOOP_VINFO - field LOOP_VINFO_MAY_ALIAS_STMTS contains the list of ddrs
3149 to be checked.
3151 Output:
3152 COND_EXPR - conditional expression.
3154 The returned COND_EXPR is the conditional expression to be used in the
3155 if statement that controls which version of the loop gets executed at
3156 runtime. */
3158 static void
3159 vect_create_cond_for_niters_checks (loop_vec_info loop_vinfo, tree *cond_expr)
3161 tree part_cond_expr = LOOP_VINFO_NITERS_ASSUMPTIONS (loop_vinfo);
3163 if (*cond_expr)
3164 *cond_expr = fold_build2 (TRUTH_AND_EXPR, boolean_type_node,
3165 *cond_expr, part_cond_expr);
3166 else
3167 *cond_expr = part_cond_expr;
3170 /* Set *COND_EXPR to a tree that is true when both the original *COND_EXPR
3171 and PART_COND_EXPR are true. Treat a null *COND_EXPR as "true". */
3173 static void
3174 chain_cond_expr (tree *cond_expr, tree part_cond_expr)
3176 if (*cond_expr)
3177 *cond_expr = fold_build2 (TRUTH_AND_EXPR, boolean_type_node,
3178 *cond_expr, part_cond_expr);
3179 else
3180 *cond_expr = part_cond_expr;
3183 /* Function vect_create_cond_for_align_checks.
3185 Create a conditional expression that represents the alignment checks for
3186 all of data references (array element references) whose alignment must be
3187 checked at runtime.
3189 Input:
3190 COND_EXPR - input conditional expression. New conditions will be chained
3191 with logical AND operation.
3192 LOOP_VINFO - two fields of the loop information are used.
3193 LOOP_VINFO_PTR_MASK is the mask used to check the alignment.
3194 LOOP_VINFO_MAY_MISALIGN_STMTS contains the refs to be checked.
3196 Output:
3197 COND_EXPR_STMT_LIST - statements needed to construct the conditional
3198 expression.
3199 The returned value is the conditional expression to be used in the if
3200 statement that controls which version of the loop gets executed at runtime.
3202 The algorithm makes two assumptions:
3203 1) The number of bytes "n" in a vector is a power of 2.
3204 2) An address "a" is aligned if a%n is zero and that this
3205 test can be done as a&(n-1) == 0. For example, for 16
3206 byte vectors the test is a&0xf == 0. */
3208 static void
3209 vect_create_cond_for_align_checks (loop_vec_info loop_vinfo,
3210 tree *cond_expr,
3211 gimple_seq *cond_expr_stmt_list)
3213 const vec<stmt_vec_info> &may_misalign_stmts
3214 = LOOP_VINFO_MAY_MISALIGN_STMTS (loop_vinfo);
3215 stmt_vec_info stmt_info;
3216 int mask = LOOP_VINFO_PTR_MASK (loop_vinfo);
3217 tree mask_cst;
3218 unsigned int i;
3219 tree int_ptrsize_type;
3220 char tmp_name[20];
3221 tree or_tmp_name = NULL_TREE;
3222 tree and_tmp_name;
3223 gimple *and_stmt;
3224 tree ptrsize_zero;
3225 tree part_cond_expr;
3227 /* Check that mask is one less than a power of 2, i.e., mask is
3228 all zeros followed by all ones. */
3229 gcc_assert ((mask != 0) && ((mask & (mask+1)) == 0));
3231 int_ptrsize_type = signed_type_for (ptr_type_node);
3233 /* Create expression (mask & (dr_1 || ... || dr_n)) where dr_i is the address
3234 of the first vector of the i'th data reference. */
3236 FOR_EACH_VEC_ELT (may_misalign_stmts, i, stmt_info)
3238 gimple_seq new_stmt_list = NULL;
3239 tree addr_base;
3240 tree addr_tmp_name;
3241 tree new_or_tmp_name;
3242 gimple *addr_stmt, *or_stmt;
3243 tree vectype = STMT_VINFO_VECTYPE (stmt_info);
3244 bool negative = tree_int_cst_compare
3245 (DR_STEP (STMT_VINFO_DATA_REF (stmt_info)), size_zero_node) < 0;
3246 tree offset = negative
3247 ? size_int ((-TYPE_VECTOR_SUBPARTS (vectype) + 1)
3248 * TREE_INT_CST_LOW (TYPE_SIZE_UNIT (TREE_TYPE (vectype))))
3249 : size_zero_node;
3251 /* create: addr_tmp = (int)(address_of_first_vector) */
3252 addr_base =
3253 vect_create_addr_base_for_vector_ref (loop_vinfo,
3254 stmt_info, &new_stmt_list,
3255 offset);
3256 if (new_stmt_list != NULL)
3257 gimple_seq_add_seq (cond_expr_stmt_list, new_stmt_list);
3259 sprintf (tmp_name, "addr2int%d", i);
3260 addr_tmp_name = make_temp_ssa_name (int_ptrsize_type, NULL, tmp_name);
3261 addr_stmt = gimple_build_assign (addr_tmp_name, NOP_EXPR, addr_base);
3262 gimple_seq_add_stmt (cond_expr_stmt_list, addr_stmt);
3264 /* The addresses are OR together. */
3266 if (or_tmp_name != NULL_TREE)
3268 /* create: or_tmp = or_tmp | addr_tmp */
3269 sprintf (tmp_name, "orptrs%d", i);
3270 new_or_tmp_name = make_temp_ssa_name (int_ptrsize_type, NULL, tmp_name);
3271 or_stmt = gimple_build_assign (new_or_tmp_name, BIT_IOR_EXPR,
3272 or_tmp_name, addr_tmp_name);
3273 gimple_seq_add_stmt (cond_expr_stmt_list, or_stmt);
3274 or_tmp_name = new_or_tmp_name;
3276 else
3277 or_tmp_name = addr_tmp_name;
3279 } /* end for i */
3281 mask_cst = build_int_cst (int_ptrsize_type, mask);
3283 /* create: and_tmp = or_tmp & mask */
3284 and_tmp_name = make_temp_ssa_name (int_ptrsize_type, NULL, "andmask");
3286 and_stmt = gimple_build_assign (and_tmp_name, BIT_AND_EXPR,
3287 or_tmp_name, mask_cst);
3288 gimple_seq_add_stmt (cond_expr_stmt_list, and_stmt);
3290 /* Make and_tmp the left operand of the conditional test against zero.
3291 if and_tmp has a nonzero bit then some address is unaligned. */
3292 ptrsize_zero = build_int_cst (int_ptrsize_type, 0);
3293 part_cond_expr = fold_build2 (EQ_EXPR, boolean_type_node,
3294 and_tmp_name, ptrsize_zero);
3295 chain_cond_expr (cond_expr, part_cond_expr);
3298 /* If LOOP_VINFO_CHECK_UNEQUAL_ADDRS contains <A1, B1>, ..., <An, Bn>,
3299 create a tree representation of: (&A1 != &B1) && ... && (&An != &Bn).
3300 Set *COND_EXPR to a tree that is true when both the original *COND_EXPR
3301 and this new condition are true. Treat a null *COND_EXPR as "true". */
3303 static void
3304 vect_create_cond_for_unequal_addrs (loop_vec_info loop_vinfo, tree *cond_expr)
3306 const vec<vec_object_pair> &pairs
3307 = LOOP_VINFO_CHECK_UNEQUAL_ADDRS (loop_vinfo);
3308 unsigned int i;
3309 vec_object_pair *pair;
3310 FOR_EACH_VEC_ELT (pairs, i, pair)
3312 tree addr1 = build_fold_addr_expr (pair->first);
3313 tree addr2 = build_fold_addr_expr (pair->second);
3314 tree part_cond_expr = fold_build2 (NE_EXPR, boolean_type_node,
3315 addr1, addr2);
3316 chain_cond_expr (cond_expr, part_cond_expr);
3320 /* Create an expression that is true when all lower-bound conditions for
3321 the vectorized loop are met. Chain this condition with *COND_EXPR. */
3323 static void
3324 vect_create_cond_for_lower_bounds (loop_vec_info loop_vinfo, tree *cond_expr)
3326 const vec<vec_lower_bound> &lower_bounds
3327 = LOOP_VINFO_LOWER_BOUNDS (loop_vinfo);
3328 for (unsigned int i = 0; i < lower_bounds.length (); ++i)
3330 tree expr = lower_bounds[i].expr;
3331 tree type = unsigned_type_for (TREE_TYPE (expr));
3332 expr = fold_convert (type, expr);
3333 poly_uint64 bound = lower_bounds[i].min_value;
3334 if (!lower_bounds[i].unsigned_p)
3336 expr = fold_build2 (PLUS_EXPR, type, expr,
3337 build_int_cstu (type, bound - 1));
3338 bound += bound - 1;
3340 tree part_cond_expr = fold_build2 (GE_EXPR, boolean_type_node, expr,
3341 build_int_cstu (type, bound));
3342 chain_cond_expr (cond_expr, part_cond_expr);
3346 /* Function vect_create_cond_for_alias_checks.
3348 Create a conditional expression that represents the run-time checks for
3349 overlapping of address ranges represented by a list of data references
3350 relations passed as input.
3352 Input:
3353 COND_EXPR - input conditional expression. New conditions will be chained
3354 with logical AND operation. If it is NULL, then the function
3355 is used to return the number of alias checks.
3356 LOOP_VINFO - field LOOP_VINFO_MAY_ALIAS_STMTS contains the list of ddrs
3357 to be checked.
3359 Output:
3360 COND_EXPR - conditional expression.
3362 The returned COND_EXPR is the conditional expression to be used in the if
3363 statement that controls which version of the loop gets executed at runtime.
3366 void
3367 vect_create_cond_for_alias_checks (loop_vec_info loop_vinfo, tree * cond_expr)
3369 const vec<dr_with_seg_len_pair_t> &comp_alias_ddrs =
3370 LOOP_VINFO_COMP_ALIAS_DDRS (loop_vinfo);
3372 if (comp_alias_ddrs.is_empty ())
3373 return;
3375 create_runtime_alias_checks (LOOP_VINFO_LOOP (loop_vinfo),
3376 &comp_alias_ddrs, cond_expr);
3377 if (dump_enabled_p ())
3378 dump_printf_loc (MSG_NOTE, vect_location,
3379 "created %u versioning for alias checks.\n",
3380 comp_alias_ddrs.length ());
3384 /* Function vect_loop_versioning.
3386 If the loop has data references that may or may not be aligned or/and
3387 has data reference relations whose independence was not proven then
3388 two versions of the loop need to be generated, one which is vectorized
3389 and one which isn't. A test is then generated to control which of the
3390 loops is executed. The test checks for the alignment of all of the
3391 data references that may or may not be aligned. An additional
3392 sequence of runtime tests is generated for each pairs of DDRs whose
3393 independence was not proven. The vectorized version of loop is
3394 executed only if both alias and alignment tests are passed.
3396 The test generated to check which version of loop is executed
3397 is modified to also check for profitability as indicated by the
3398 cost model threshold TH.
3400 The versioning precondition(s) are placed in *COND_EXPR and
3401 *COND_EXPR_STMT_LIST. */
3403 class loop *
3404 vect_loop_versioning (loop_vec_info loop_vinfo,
3405 gimple *loop_vectorized_call)
3407 class loop *loop = LOOP_VINFO_LOOP (loop_vinfo), *nloop;
3408 class loop *scalar_loop = LOOP_VINFO_SCALAR_LOOP (loop_vinfo);
3409 basic_block condition_bb;
3410 gphi_iterator gsi;
3411 gimple_stmt_iterator cond_exp_gsi;
3412 basic_block merge_bb;
3413 basic_block new_exit_bb;
3414 edge new_exit_e, e;
3415 gphi *orig_phi, *new_phi;
3416 tree cond_expr = NULL_TREE;
3417 gimple_seq cond_expr_stmt_list = NULL;
3418 tree arg;
3419 profile_probability prob = profile_probability::likely ();
3420 gimple_seq gimplify_stmt_list = NULL;
3421 tree scalar_loop_iters = LOOP_VINFO_NITERSM1 (loop_vinfo);
3422 bool version_align = LOOP_REQUIRES_VERSIONING_FOR_ALIGNMENT (loop_vinfo);
3423 bool version_alias = LOOP_REQUIRES_VERSIONING_FOR_ALIAS (loop_vinfo);
3424 bool version_niter = LOOP_REQUIRES_VERSIONING_FOR_NITERS (loop_vinfo);
3425 poly_uint64 versioning_threshold
3426 = LOOP_VINFO_VERSIONING_THRESHOLD (loop_vinfo);
3427 tree version_simd_if_cond
3428 = LOOP_REQUIRES_VERSIONING_FOR_SIMD_IF_COND (loop_vinfo);
3429 unsigned th = LOOP_VINFO_COST_MODEL_THRESHOLD (loop_vinfo);
3431 if (vect_apply_runtime_profitability_check_p (loop_vinfo)
3432 && !ordered_p (th, versioning_threshold))
3433 cond_expr = fold_build2 (GE_EXPR, boolean_type_node, scalar_loop_iters,
3434 build_int_cst (TREE_TYPE (scalar_loop_iters),
3435 th - 1));
3436 if (maybe_ne (versioning_threshold, 0U))
3438 tree expr = fold_build2 (GE_EXPR, boolean_type_node, scalar_loop_iters,
3439 build_int_cst (TREE_TYPE (scalar_loop_iters),
3440 versioning_threshold - 1));
3441 if (cond_expr)
3442 cond_expr = fold_build2 (BIT_AND_EXPR, boolean_type_node,
3443 expr, cond_expr);
3444 else
3445 cond_expr = expr;
3448 if (version_niter)
3449 vect_create_cond_for_niters_checks (loop_vinfo, &cond_expr);
3451 if (cond_expr)
3452 cond_expr = force_gimple_operand_1 (unshare_expr (cond_expr),
3453 &cond_expr_stmt_list,
3454 is_gimple_condexpr, NULL_TREE);
3456 if (version_align)
3457 vect_create_cond_for_align_checks (loop_vinfo, &cond_expr,
3458 &cond_expr_stmt_list);
3460 if (version_alias)
3462 vect_create_cond_for_unequal_addrs (loop_vinfo, &cond_expr);
3463 vect_create_cond_for_lower_bounds (loop_vinfo, &cond_expr);
3464 vect_create_cond_for_alias_checks (loop_vinfo, &cond_expr);
3467 if (version_simd_if_cond)
3469 gcc_assert (dom_info_available_p (CDI_DOMINATORS));
3470 if (flag_checking)
3471 if (basic_block bb
3472 = gimple_bb (SSA_NAME_DEF_STMT (version_simd_if_cond)))
3473 gcc_assert (bb != loop->header
3474 && dominated_by_p (CDI_DOMINATORS, loop->header, bb)
3475 && (scalar_loop == NULL
3476 || (bb != scalar_loop->header
3477 && dominated_by_p (CDI_DOMINATORS,
3478 scalar_loop->header, bb))));
3479 tree zero = build_zero_cst (TREE_TYPE (version_simd_if_cond));
3480 tree c = fold_build2 (NE_EXPR, boolean_type_node,
3481 version_simd_if_cond, zero);
3482 if (cond_expr)
3483 cond_expr = fold_build2 (TRUTH_AND_EXPR, boolean_type_node,
3484 c, cond_expr);
3485 else
3486 cond_expr = c;
3487 if (dump_enabled_p ())
3488 dump_printf_loc (MSG_NOTE, vect_location,
3489 "created versioning for simd if condition check.\n");
3492 cond_expr = force_gimple_operand_1 (unshare_expr (cond_expr),
3493 &gimplify_stmt_list,
3494 is_gimple_condexpr, NULL_TREE);
3495 gimple_seq_add_seq (&cond_expr_stmt_list, gimplify_stmt_list);
3497 /* Compute the outermost loop cond_expr and cond_expr_stmt_list are
3498 invariant in. */
3499 class loop *outermost = outermost_invariant_loop_for_expr (loop, cond_expr);
3500 for (gimple_stmt_iterator gsi = gsi_start (cond_expr_stmt_list);
3501 !gsi_end_p (gsi); gsi_next (&gsi))
3503 gimple *stmt = gsi_stmt (gsi);
3504 update_stmt (stmt);
3505 ssa_op_iter iter;
3506 use_operand_p use_p;
3507 basic_block def_bb;
3508 FOR_EACH_SSA_USE_OPERAND (use_p, stmt, iter, SSA_OP_USE)
3509 if ((def_bb = gimple_bb (SSA_NAME_DEF_STMT (USE_FROM_PTR (use_p))))
3510 && flow_bb_inside_loop_p (outermost, def_bb))
3511 outermost = superloop_at_depth (loop, bb_loop_depth (def_bb) + 1);
3514 /* Search for the outermost loop we can version. Avoid versioning of
3515 non-perfect nests but allow if-conversion versioned loops inside. */
3516 class loop *loop_to_version = loop;
3517 if (flow_loop_nested_p (outermost, loop))
3519 if (dump_enabled_p ())
3520 dump_printf_loc (MSG_NOTE, vect_location,
3521 "trying to apply versioning to outer loop %d\n",
3522 outermost->num);
3523 if (outermost->num == 0)
3524 outermost = superloop_at_depth (loop, 1);
3525 /* And avoid applying versioning on non-perfect nests. */
3526 while (loop_to_version != outermost
3527 && single_exit (loop_outer (loop_to_version))
3528 && (!loop_outer (loop_to_version)->inner->next
3529 || vect_loop_vectorized_call (loop_to_version))
3530 && (!loop_outer (loop_to_version)->inner->next
3531 || !loop_outer (loop_to_version)->inner->next->next))
3532 loop_to_version = loop_outer (loop_to_version);
3535 /* Apply versioning. If there is already a scalar version created by
3536 if-conversion re-use that. Note we cannot re-use the copy of
3537 an if-converted outer-loop when vectorizing the inner loop only. */
3538 gcond *cond;
3539 if ((!loop_to_version->inner || loop == loop_to_version)
3540 && loop_vectorized_call)
3542 gcc_assert (scalar_loop);
3543 condition_bb = gimple_bb (loop_vectorized_call);
3544 cond = as_a <gcond *> (last_stmt (condition_bb));
3545 gimple_cond_set_condition_from_tree (cond, cond_expr);
3546 update_stmt (cond);
3548 if (cond_expr_stmt_list)
3550 cond_exp_gsi = gsi_for_stmt (loop_vectorized_call);
3551 gsi_insert_seq_before (&cond_exp_gsi, cond_expr_stmt_list,
3552 GSI_SAME_STMT);
3555 /* if-conversion uses profile_probability::always () for both paths,
3556 reset the paths probabilities appropriately. */
3557 edge te, fe;
3558 extract_true_false_edges_from_block (condition_bb, &te, &fe);
3559 te->probability = prob;
3560 fe->probability = prob.invert ();
3561 /* We can scale loops counts immediately but have to postpone
3562 scaling the scalar loop because we re-use it during peeling. */
3563 scale_loop_frequencies (loop_to_version, te->probability);
3564 LOOP_VINFO_SCALAR_LOOP_SCALING (loop_vinfo) = fe->probability;
3566 nloop = scalar_loop;
3567 if (dump_enabled_p ())
3568 dump_printf_loc (MSG_NOTE, vect_location,
3569 "reusing %sloop version created by if conversion\n",
3570 loop_to_version != loop ? "outer " : "");
3572 else
3574 if (loop_to_version != loop
3575 && dump_enabled_p ())
3576 dump_printf_loc (MSG_NOTE, vect_location,
3577 "applying loop versioning to outer loop %d\n",
3578 loop_to_version->num);
3580 unsigned orig_pe_idx = loop_preheader_edge (loop)->dest_idx;
3582 initialize_original_copy_tables ();
3583 nloop = loop_version (loop_to_version, cond_expr, &condition_bb,
3584 prob, prob.invert (), prob, prob.invert (), true);
3585 gcc_assert (nloop);
3586 nloop = get_loop_copy (loop);
3588 /* For cycle vectorization with SLP we rely on the PHI arguments
3589 appearing in the same order as the SLP node operands which for the
3590 loop PHI nodes means the preheader edge dest index needs to remain
3591 the same for the analyzed loop which also becomes the vectorized one.
3592 Make it so in case the state after versioning differs by redirecting
3593 the first edge into the header to the same destination which moves
3594 it last. */
3595 if (loop_preheader_edge (loop)->dest_idx != orig_pe_idx)
3597 edge e = EDGE_PRED (loop->header, 0);
3598 ssa_redirect_edge (e, e->dest);
3599 flush_pending_stmts (e);
3601 gcc_assert (loop_preheader_edge (loop)->dest_idx == orig_pe_idx);
3603 /* Kill off IFN_LOOP_VECTORIZED_CALL in the copy, nobody will
3604 reap those otherwise; they also refer to the original
3605 loops. */
3606 class loop *l = loop;
3607 while (gimple *call = vect_loop_vectorized_call (l))
3609 call = SSA_NAME_DEF_STMT (get_current_def (gimple_call_lhs (call)));
3610 fold_loop_internal_call (call, boolean_false_node);
3611 l = loop_outer (l);
3613 free_original_copy_tables ();
3615 if (cond_expr_stmt_list)
3617 cond_exp_gsi = gsi_last_bb (condition_bb);
3618 gsi_insert_seq_before (&cond_exp_gsi, cond_expr_stmt_list,
3619 GSI_SAME_STMT);
3622 /* Loop versioning violates an assumption we try to maintain during
3623 vectorization - that the loop exit block has a single predecessor.
3624 After versioning, the exit block of both loop versions is the same
3625 basic block (i.e. it has two predecessors). Just in order to simplify
3626 following transformations in the vectorizer, we fix this situation
3627 here by adding a new (empty) block on the exit-edge of the loop,
3628 with the proper loop-exit phis to maintain loop-closed-form.
3629 If loop versioning wasn't done from loop, but scalar_loop instead,
3630 merge_bb will have already just a single successor. */
3632 merge_bb = single_exit (loop_to_version)->dest;
3633 if (EDGE_COUNT (merge_bb->preds) >= 2)
3635 gcc_assert (EDGE_COUNT (merge_bb->preds) >= 2);
3636 new_exit_bb = split_edge (single_exit (loop_to_version));
3637 new_exit_e = single_exit (loop_to_version);
3638 e = EDGE_SUCC (new_exit_bb, 0);
3640 for (gsi = gsi_start_phis (merge_bb); !gsi_end_p (gsi);
3641 gsi_next (&gsi))
3643 tree new_res;
3644 orig_phi = gsi.phi ();
3645 new_res = copy_ssa_name (PHI_RESULT (orig_phi));
3646 new_phi = create_phi_node (new_res, new_exit_bb);
3647 arg = PHI_ARG_DEF_FROM_EDGE (orig_phi, e);
3648 add_phi_arg (new_phi, arg, new_exit_e,
3649 gimple_phi_arg_location_from_edge (orig_phi, e));
3650 adjust_phi_and_debug_stmts (orig_phi, e, PHI_RESULT (new_phi));
3654 update_ssa (TODO_update_ssa);
3657 if (version_niter)
3659 /* The versioned loop could be infinite, we need to clear existing
3660 niter information which is copied from the original loop. */
3661 gcc_assert (loop_constraint_set_p (loop, LOOP_C_FINITE));
3662 vect_free_loop_info_assumptions (nloop);
3665 if (LOCATION_LOCUS (vect_location.get_location_t ()) != UNKNOWN_LOCATION
3666 && dump_enabled_p ())
3668 if (version_alias)
3669 dump_printf_loc (MSG_OPTIMIZED_LOCATIONS | MSG_PRIORITY_USER_FACING,
3670 vect_location,
3671 "loop versioned for vectorization because of "
3672 "possible aliasing\n");
3673 if (version_align)
3674 dump_printf_loc (MSG_OPTIMIZED_LOCATIONS | MSG_PRIORITY_USER_FACING,
3675 vect_location,
3676 "loop versioned for vectorization to enhance "
3677 "alignment\n");
3681 return nloop;