typeck.c (cp_truthvalue_conversion): Add tsubst_flags_t parameter and use it in calls...
[official-gcc.git] / gcc / tree-ssa-phiopt.c
blobc2595c85241b65fcfeb3b1a526399d0753b47b7b
1 /* Optimization of PHI nodes by converting them into straightline code.
2 Copyright (C) 2004-2019 Free Software Foundation, Inc.
4 This file is part of GCC.
6 GCC is free software; you can redistribute it and/or modify it
7 under the terms of the GNU General Public License as published by the
8 Free Software Foundation; either version 3, or (at your option) any
9 later version.
11 GCC is distributed in the hope that it will be useful, but WITHOUT
12 ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
13 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
14 for more details.
16 You should have received a copy of the GNU General Public License
17 along with GCC; see the file COPYING3. If not see
18 <http://www.gnu.org/licenses/>. */
20 #include "config.h"
21 #include "system.h"
22 #include "coretypes.h"
23 #include "backend.h"
24 #include "insn-codes.h"
25 #include "rtl.h"
26 #include "tree.h"
27 #include "gimple.h"
28 #include "cfghooks.h"
29 #include "tree-pass.h"
30 #include "ssa.h"
31 #include "optabs-tree.h"
32 #include "insn-config.h"
33 #include "gimple-pretty-print.h"
34 #include "fold-const.h"
35 #include "stor-layout.h"
36 #include "cfganal.h"
37 #include "gimplify.h"
38 #include "gimple-iterator.h"
39 #include "gimplify-me.h"
40 #include "tree-cfg.h"
41 #include "tree-dfa.h"
42 #include "domwalk.h"
43 #include "cfgloop.h"
44 #include "tree-data-ref.h"
45 #include "tree-scalar-evolution.h"
46 #include "tree-inline.h"
47 #include "case-cfn-macros.h"
49 static unsigned int tree_ssa_phiopt_worker (bool, bool, bool);
50 static bool two_value_replacement (basic_block, basic_block, edge, gphi *,
51 tree, tree);
52 static bool conditional_replacement (basic_block, basic_block,
53 edge, edge, gphi *, tree, tree);
54 static gphi *factor_out_conditional_conversion (edge, edge, gphi *, tree, tree,
55 gimple *);
56 static int value_replacement (basic_block, basic_block,
57 edge, edge, gimple *, tree, tree);
58 static bool minmax_replacement (basic_block, basic_block,
59 edge, edge, gimple *, tree, tree);
60 static bool abs_replacement (basic_block, basic_block,
61 edge, edge, gimple *, tree, tree);
62 static bool cond_removal_in_popcount_pattern (basic_block, basic_block,
63 edge, edge, gimple *, tree, tree);
64 static bool cond_store_replacement (basic_block, basic_block, edge, edge,
65 hash_set<tree> *);
66 static bool cond_if_else_store_replacement (basic_block, basic_block, basic_block);
67 static hash_set<tree> * get_non_trapping ();
68 static void replace_phi_edge_with_variable (basic_block, edge, gimple *, tree);
69 static void hoist_adjacent_loads (basic_block, basic_block,
70 basic_block, basic_block);
71 static bool gate_hoist_loads (void);
73 /* This pass tries to transform conditional stores into unconditional
74 ones, enabling further simplifications with the simpler then and else
75 blocks. In particular it replaces this:
77 bb0:
78 if (cond) goto bb2; else goto bb1;
79 bb1:
80 *p = RHS;
81 bb2:
83 with
85 bb0:
86 if (cond) goto bb1; else goto bb2;
87 bb1:
88 condtmp' = *p;
89 bb2:
90 condtmp = PHI <RHS, condtmp'>
91 *p = condtmp;
93 This transformation can only be done under several constraints,
94 documented below. It also replaces:
96 bb0:
97 if (cond) goto bb2; else goto bb1;
98 bb1:
99 *p = RHS1;
100 goto bb3;
101 bb2:
102 *p = RHS2;
103 bb3:
105 with
107 bb0:
108 if (cond) goto bb3; else goto bb1;
109 bb1:
110 bb3:
111 condtmp = PHI <RHS1, RHS2>
112 *p = condtmp; */
114 static unsigned int
115 tree_ssa_cs_elim (void)
117 unsigned todo;
118 /* ??? We are not interested in loop related info, but the following
119 will create it, ICEing as we didn't init loops with pre-headers.
120 An interfacing issue of find_data_references_in_bb. */
121 loop_optimizer_init (LOOPS_NORMAL);
122 scev_initialize ();
123 todo = tree_ssa_phiopt_worker (true, false, false);
124 scev_finalize ();
125 loop_optimizer_finalize ();
126 return todo;
129 /* Return the singleton PHI in the SEQ of PHIs for edges E0 and E1. */
131 static gphi *
132 single_non_singleton_phi_for_edges (gimple_seq seq, edge e0, edge e1)
134 gimple_stmt_iterator i;
135 gphi *phi = NULL;
136 if (gimple_seq_singleton_p (seq))
137 return as_a <gphi *> (gsi_stmt (gsi_start (seq)));
138 for (i = gsi_start (seq); !gsi_end_p (i); gsi_next (&i))
140 gphi *p = as_a <gphi *> (gsi_stmt (i));
141 /* If the PHI arguments are equal then we can skip this PHI. */
142 if (operand_equal_for_phi_arg_p (gimple_phi_arg_def (p, e0->dest_idx),
143 gimple_phi_arg_def (p, e1->dest_idx)))
144 continue;
146 /* If we already have a PHI that has the two edge arguments are
147 different, then return it is not a singleton for these PHIs. */
148 if (phi)
149 return NULL;
151 phi = p;
153 return phi;
156 /* The core routine of conditional store replacement and normal
157 phi optimizations. Both share much of the infrastructure in how
158 to match applicable basic block patterns. DO_STORE_ELIM is true
159 when we want to do conditional store replacement, false otherwise.
160 DO_HOIST_LOADS is true when we want to hoist adjacent loads out
161 of diamond control flow patterns, false otherwise. */
162 static unsigned int
163 tree_ssa_phiopt_worker (bool do_store_elim, bool do_hoist_loads, bool early_p)
165 basic_block bb;
166 basic_block *bb_order;
167 unsigned n, i;
168 bool cfgchanged = false;
169 hash_set<tree> *nontrap = 0;
171 if (do_store_elim)
172 /* Calculate the set of non-trapping memory accesses. */
173 nontrap = get_non_trapping ();
175 /* Search every basic block for COND_EXPR we may be able to optimize.
177 We walk the blocks in order that guarantees that a block with
178 a single predecessor is processed before the predecessor.
179 This ensures that we collapse inner ifs before visiting the
180 outer ones, and also that we do not try to visit a removed
181 block. */
182 bb_order = single_pred_before_succ_order ();
183 n = n_basic_blocks_for_fn (cfun) - NUM_FIXED_BLOCKS;
185 for (i = 0; i < n; i++)
187 gimple *cond_stmt;
188 gphi *phi;
189 basic_block bb1, bb2;
190 edge e1, e2;
191 tree arg0, arg1;
193 bb = bb_order[i];
195 cond_stmt = last_stmt (bb);
196 /* Check to see if the last statement is a GIMPLE_COND. */
197 if (!cond_stmt
198 || gimple_code (cond_stmt) != GIMPLE_COND)
199 continue;
201 e1 = EDGE_SUCC (bb, 0);
202 bb1 = e1->dest;
203 e2 = EDGE_SUCC (bb, 1);
204 bb2 = e2->dest;
206 /* We cannot do the optimization on abnormal edges. */
207 if ((e1->flags & EDGE_ABNORMAL) != 0
208 || (e2->flags & EDGE_ABNORMAL) != 0)
209 continue;
211 /* If either bb1's succ or bb2 or bb2's succ is non NULL. */
212 if (EDGE_COUNT (bb1->succs) == 0
213 || bb2 == NULL
214 || EDGE_COUNT (bb2->succs) == 0)
215 continue;
217 /* Find the bb which is the fall through to the other. */
218 if (EDGE_SUCC (bb1, 0)->dest == bb2)
220 else if (EDGE_SUCC (bb2, 0)->dest == bb1)
222 std::swap (bb1, bb2);
223 std::swap (e1, e2);
225 else if (do_store_elim
226 && EDGE_SUCC (bb1, 0)->dest == EDGE_SUCC (bb2, 0)->dest)
228 basic_block bb3 = EDGE_SUCC (bb1, 0)->dest;
230 if (!single_succ_p (bb1)
231 || (EDGE_SUCC (bb1, 0)->flags & EDGE_FALLTHRU) == 0
232 || !single_succ_p (bb2)
233 || (EDGE_SUCC (bb2, 0)->flags & EDGE_FALLTHRU) == 0
234 || EDGE_COUNT (bb3->preds) != 2)
235 continue;
236 if (cond_if_else_store_replacement (bb1, bb2, bb3))
237 cfgchanged = true;
238 continue;
240 else if (do_hoist_loads
241 && EDGE_SUCC (bb1, 0)->dest == EDGE_SUCC (bb2, 0)->dest)
243 basic_block bb3 = EDGE_SUCC (bb1, 0)->dest;
245 if (!FLOAT_TYPE_P (TREE_TYPE (gimple_cond_lhs (cond_stmt)))
246 && single_succ_p (bb1)
247 && single_succ_p (bb2)
248 && single_pred_p (bb1)
249 && single_pred_p (bb2)
250 && EDGE_COUNT (bb->succs) == 2
251 && EDGE_COUNT (bb3->preds) == 2
252 /* If one edge or the other is dominant, a conditional move
253 is likely to perform worse than the well-predicted branch. */
254 && !predictable_edge_p (EDGE_SUCC (bb, 0))
255 && !predictable_edge_p (EDGE_SUCC (bb, 1)))
256 hoist_adjacent_loads (bb, bb1, bb2, bb3);
257 continue;
259 else
260 continue;
262 e1 = EDGE_SUCC (bb1, 0);
264 /* Make sure that bb1 is just a fall through. */
265 if (!single_succ_p (bb1)
266 || (e1->flags & EDGE_FALLTHRU) == 0)
267 continue;
269 /* Also make sure that bb1 only have one predecessor and that it
270 is bb. */
271 if (!single_pred_p (bb1)
272 || single_pred (bb1) != bb)
273 continue;
275 if (do_store_elim)
277 /* bb1 is the middle block, bb2 the join block, bb the split block,
278 e1 the fallthrough edge from bb1 to bb2. We can't do the
279 optimization if the join block has more than two predecessors. */
280 if (EDGE_COUNT (bb2->preds) > 2)
281 continue;
282 if (cond_store_replacement (bb1, bb2, e1, e2, nontrap))
283 cfgchanged = true;
285 else
287 gimple_seq phis = phi_nodes (bb2);
288 gimple_stmt_iterator gsi;
289 bool candorest = true;
291 /* Value replacement can work with more than one PHI
292 so try that first. */
293 if (!early_p)
294 for (gsi = gsi_start (phis); !gsi_end_p (gsi); gsi_next (&gsi))
296 phi = as_a <gphi *> (gsi_stmt (gsi));
297 arg0 = gimple_phi_arg_def (phi, e1->dest_idx);
298 arg1 = gimple_phi_arg_def (phi, e2->dest_idx);
299 if (value_replacement (bb, bb1, e1, e2, phi, arg0, arg1) == 2)
301 candorest = false;
302 cfgchanged = true;
303 break;
307 if (!candorest)
308 continue;
310 phi = single_non_singleton_phi_for_edges (phis, e1, e2);
311 if (!phi)
312 continue;
314 arg0 = gimple_phi_arg_def (phi, e1->dest_idx);
315 arg1 = gimple_phi_arg_def (phi, e2->dest_idx);
317 /* Something is wrong if we cannot find the arguments in the PHI
318 node. */
319 gcc_assert (arg0 != NULL_TREE && arg1 != NULL_TREE);
321 gphi *newphi = factor_out_conditional_conversion (e1, e2, phi,
322 arg0, arg1,
323 cond_stmt);
324 if (newphi != NULL)
326 phi = newphi;
327 /* factor_out_conditional_conversion may create a new PHI in
328 BB2 and eliminate an existing PHI in BB2. Recompute values
329 that may be affected by that change. */
330 arg0 = gimple_phi_arg_def (phi, e1->dest_idx);
331 arg1 = gimple_phi_arg_def (phi, e2->dest_idx);
332 gcc_assert (arg0 != NULL_TREE && arg1 != NULL_TREE);
335 /* Do the replacement of conditional if it can be done. */
336 if (two_value_replacement (bb, bb1, e2, phi, arg0, arg1))
337 cfgchanged = true;
338 else if (!early_p
339 && conditional_replacement (bb, bb1, e1, e2, phi,
340 arg0, arg1))
341 cfgchanged = true;
342 else if (abs_replacement (bb, bb1, e1, e2, phi, arg0, arg1))
343 cfgchanged = true;
344 else if (!early_p
345 && cond_removal_in_popcount_pattern (bb, bb1, e1, e2,
346 phi, arg0, arg1))
347 cfgchanged = true;
348 else if (minmax_replacement (bb, bb1, e1, e2, phi, arg0, arg1))
349 cfgchanged = true;
353 free (bb_order);
355 if (do_store_elim)
356 delete nontrap;
357 /* If the CFG has changed, we should cleanup the CFG. */
358 if (cfgchanged && do_store_elim)
360 /* In cond-store replacement we have added some loads on edges
361 and new VOPS (as we moved the store, and created a load). */
362 gsi_commit_edge_inserts ();
363 return TODO_cleanup_cfg | TODO_update_ssa_only_virtuals;
365 else if (cfgchanged)
366 return TODO_cleanup_cfg;
367 return 0;
370 /* Replace PHI node element whose edge is E in block BB with variable NEW.
371 Remove the edge from COND_BLOCK which does not lead to BB (COND_BLOCK
372 is known to have two edges, one of which must reach BB). */
374 static void
375 replace_phi_edge_with_variable (basic_block cond_block,
376 edge e, gimple *phi, tree new_tree)
378 basic_block bb = gimple_bb (phi);
379 basic_block block_to_remove;
380 gimple_stmt_iterator gsi;
382 /* Change the PHI argument to new. */
383 SET_USE (PHI_ARG_DEF_PTR (phi, e->dest_idx), new_tree);
385 /* Remove the empty basic block. */
386 if (EDGE_SUCC (cond_block, 0)->dest == bb)
388 EDGE_SUCC (cond_block, 0)->flags |= EDGE_FALLTHRU;
389 EDGE_SUCC (cond_block, 0)->flags &= ~(EDGE_TRUE_VALUE | EDGE_FALSE_VALUE);
390 EDGE_SUCC (cond_block, 0)->probability = profile_probability::always ();
392 block_to_remove = EDGE_SUCC (cond_block, 1)->dest;
394 else
396 EDGE_SUCC (cond_block, 1)->flags |= EDGE_FALLTHRU;
397 EDGE_SUCC (cond_block, 1)->flags
398 &= ~(EDGE_TRUE_VALUE | EDGE_FALSE_VALUE);
399 EDGE_SUCC (cond_block, 1)->probability = profile_probability::always ();
401 block_to_remove = EDGE_SUCC (cond_block, 0)->dest;
403 delete_basic_block (block_to_remove);
405 /* Eliminate the COND_EXPR at the end of COND_BLOCK. */
406 gsi = gsi_last_bb (cond_block);
407 gsi_remove (&gsi, true);
409 if (dump_file && (dump_flags & TDF_DETAILS))
410 fprintf (dump_file,
411 "COND_EXPR in block %d and PHI in block %d converted to straightline code.\n",
412 cond_block->index,
413 bb->index);
416 /* PR66726: Factor conversion out of COND_EXPR. If the arguments of the PHI
417 stmt are CONVERT_STMT, factor out the conversion and perform the conversion
418 to the result of PHI stmt. COND_STMT is the controlling predicate.
419 Return the newly-created PHI, if any. */
421 static gphi *
422 factor_out_conditional_conversion (edge e0, edge e1, gphi *phi,
423 tree arg0, tree arg1, gimple *cond_stmt)
425 gimple *arg0_def_stmt = NULL, *arg1_def_stmt = NULL, *new_stmt;
426 tree new_arg0 = NULL_TREE, new_arg1 = NULL_TREE;
427 tree temp, result;
428 gphi *newphi;
429 gimple_stmt_iterator gsi, gsi_for_def;
430 location_t locus = gimple_location (phi);
431 enum tree_code convert_code;
433 /* Handle only PHI statements with two arguments. TODO: If all
434 other arguments to PHI are INTEGER_CST or if their defining
435 statement have the same unary operation, we can handle more
436 than two arguments too. */
437 if (gimple_phi_num_args (phi) != 2)
438 return NULL;
440 /* First canonicalize to simplify tests. */
441 if (TREE_CODE (arg0) != SSA_NAME)
443 std::swap (arg0, arg1);
444 std::swap (e0, e1);
447 if (TREE_CODE (arg0) != SSA_NAME
448 || (TREE_CODE (arg1) != SSA_NAME
449 && TREE_CODE (arg1) != INTEGER_CST))
450 return NULL;
452 /* Check if arg0 is an SSA_NAME and the stmt which defines arg0 is
453 a conversion. */
454 arg0_def_stmt = SSA_NAME_DEF_STMT (arg0);
455 if (!gimple_assign_cast_p (arg0_def_stmt))
456 return NULL;
458 /* Use the RHS as new_arg0. */
459 convert_code = gimple_assign_rhs_code (arg0_def_stmt);
460 new_arg0 = gimple_assign_rhs1 (arg0_def_stmt);
461 if (convert_code == VIEW_CONVERT_EXPR)
463 new_arg0 = TREE_OPERAND (new_arg0, 0);
464 if (!is_gimple_reg_type (TREE_TYPE (new_arg0)))
465 return NULL;
468 if (TREE_CODE (arg1) == SSA_NAME)
470 /* Check if arg1 is an SSA_NAME and the stmt which defines arg1
471 is a conversion. */
472 arg1_def_stmt = SSA_NAME_DEF_STMT (arg1);
473 if (!is_gimple_assign (arg1_def_stmt)
474 || gimple_assign_rhs_code (arg1_def_stmt) != convert_code)
475 return NULL;
477 /* Use the RHS as new_arg1. */
478 new_arg1 = gimple_assign_rhs1 (arg1_def_stmt);
479 if (convert_code == VIEW_CONVERT_EXPR)
480 new_arg1 = TREE_OPERAND (new_arg1, 0);
482 else
484 /* If arg1 is an INTEGER_CST, fold it to new type. */
485 if (INTEGRAL_TYPE_P (TREE_TYPE (new_arg0))
486 && int_fits_type_p (arg1, TREE_TYPE (new_arg0)))
488 if (gimple_assign_cast_p (arg0_def_stmt))
490 /* For the INTEGER_CST case, we are just moving the
491 conversion from one place to another, which can often
492 hurt as the conversion moves further away from the
493 statement that computes the value. So, perform this
494 only if new_arg0 is an operand of COND_STMT, or
495 if arg0_def_stmt is the only non-debug stmt in
496 its basic block, because then it is possible this
497 could enable further optimizations (minmax replacement
498 etc.). See PR71016. */
499 if (new_arg0 != gimple_cond_lhs (cond_stmt)
500 && new_arg0 != gimple_cond_rhs (cond_stmt)
501 && gimple_bb (arg0_def_stmt) == e0->src)
503 gsi = gsi_for_stmt (arg0_def_stmt);
504 gsi_prev_nondebug (&gsi);
505 if (!gsi_end_p (gsi))
507 if (gassign *assign
508 = dyn_cast <gassign *> (gsi_stmt (gsi)))
510 tree lhs = gimple_assign_lhs (assign);
511 enum tree_code ass_code
512 = gimple_assign_rhs_code (assign);
513 if (ass_code != MAX_EXPR && ass_code != MIN_EXPR)
514 return NULL;
515 if (lhs != gimple_assign_rhs1 (arg0_def_stmt))
516 return NULL;
517 gsi_prev_nondebug (&gsi);
518 if (!gsi_end_p (gsi))
519 return NULL;
521 else
522 return NULL;
524 gsi = gsi_for_stmt (arg0_def_stmt);
525 gsi_next_nondebug (&gsi);
526 if (!gsi_end_p (gsi))
527 return NULL;
529 new_arg1 = fold_convert (TREE_TYPE (new_arg0), arg1);
531 else
532 return NULL;
534 else
535 return NULL;
538 /* If arg0/arg1 have > 1 use, then this transformation actually increases
539 the number of expressions evaluated at runtime. */
540 if (!has_single_use (arg0)
541 || (arg1_def_stmt && !has_single_use (arg1)))
542 return NULL;
544 /* If types of new_arg0 and new_arg1 are different bailout. */
545 if (!types_compatible_p (TREE_TYPE (new_arg0), TREE_TYPE (new_arg1)))
546 return NULL;
548 /* Create a new PHI stmt. */
549 result = PHI_RESULT (phi);
550 temp = make_ssa_name (TREE_TYPE (new_arg0), NULL);
551 newphi = create_phi_node (temp, gimple_bb (phi));
553 if (dump_file && (dump_flags & TDF_DETAILS))
555 fprintf (dump_file, "PHI ");
556 print_generic_expr (dump_file, gimple_phi_result (phi));
557 fprintf (dump_file,
558 " changed to factor conversion out from COND_EXPR.\n");
559 fprintf (dump_file, "New stmt with CAST that defines ");
560 print_generic_expr (dump_file, result);
561 fprintf (dump_file, ".\n");
564 /* Remove the old cast(s) that has single use. */
565 gsi_for_def = gsi_for_stmt (arg0_def_stmt);
566 gsi_remove (&gsi_for_def, true);
567 release_defs (arg0_def_stmt);
569 if (arg1_def_stmt)
571 gsi_for_def = gsi_for_stmt (arg1_def_stmt);
572 gsi_remove (&gsi_for_def, true);
573 release_defs (arg1_def_stmt);
576 add_phi_arg (newphi, new_arg0, e0, locus);
577 add_phi_arg (newphi, new_arg1, e1, locus);
579 /* Create the conversion stmt and insert it. */
580 if (convert_code == VIEW_CONVERT_EXPR)
582 temp = fold_build1 (VIEW_CONVERT_EXPR, TREE_TYPE (result), temp);
583 new_stmt = gimple_build_assign (result, temp);
585 else
586 new_stmt = gimple_build_assign (result, convert_code, temp);
587 gsi = gsi_after_labels (gimple_bb (phi));
588 gsi_insert_before (&gsi, new_stmt, GSI_SAME_STMT);
590 /* Remove the original PHI stmt. */
591 gsi = gsi_for_stmt (phi);
592 gsi_remove (&gsi, true);
593 return newphi;
596 /* Optimize
597 # x_5 in range [cst1, cst2] where cst2 = cst1 + 1
598 if (x_5 op cstN) # where op is == or != and N is 1 or 2
599 goto bb3;
600 else
601 goto bb4;
602 bb3:
603 bb4:
604 # r_6 = PHI<cst3(2), cst4(3)> # where cst3 == cst4 + 1 or cst4 == cst3 + 1
606 to r_6 = x_5 + (min (cst3, cst4) - cst1) or
607 r_6 = (min (cst3, cst4) + cst1) - x_5 depending on op, N and which
608 of cst3 and cst4 is smaller. */
610 static bool
611 two_value_replacement (basic_block cond_bb, basic_block middle_bb,
612 edge e1, gphi *phi, tree arg0, tree arg1)
614 /* Only look for adjacent integer constants. */
615 if (!INTEGRAL_TYPE_P (TREE_TYPE (arg0))
616 || !INTEGRAL_TYPE_P (TREE_TYPE (arg1))
617 || TREE_CODE (arg0) != INTEGER_CST
618 || TREE_CODE (arg1) != INTEGER_CST
619 || (tree_int_cst_lt (arg0, arg1)
620 ? wi::to_widest (arg0) + 1 != wi::to_widest (arg1)
621 : wi::to_widest (arg1) + 1 != wi::to_widest (arg0)))
622 return false;
624 if (!empty_block_p (middle_bb))
625 return false;
627 gimple *stmt = last_stmt (cond_bb);
628 tree lhs = gimple_cond_lhs (stmt);
629 tree rhs = gimple_cond_rhs (stmt);
631 if (TREE_CODE (lhs) != SSA_NAME
632 || !INTEGRAL_TYPE_P (TREE_TYPE (lhs))
633 || TREE_CODE (TREE_TYPE (lhs)) == BOOLEAN_TYPE
634 || TREE_CODE (rhs) != INTEGER_CST)
635 return false;
637 switch (gimple_cond_code (stmt))
639 case EQ_EXPR:
640 case NE_EXPR:
641 break;
642 default:
643 return false;
646 wide_int min, max;
647 if (get_range_info (lhs, &min, &max) != VR_RANGE
648 || min + 1 != max
649 || (wi::to_wide (rhs) != min
650 && wi::to_wide (rhs) != max))
651 return false;
653 /* We need to know which is the true edge and which is the false
654 edge so that we know when to invert the condition below. */
655 edge true_edge, false_edge;
656 extract_true_false_edges_from_block (cond_bb, &true_edge, &false_edge);
657 if ((gimple_cond_code (stmt) == EQ_EXPR)
658 ^ (wi::to_wide (rhs) == max)
659 ^ (e1 == false_edge))
660 std::swap (arg0, arg1);
662 tree type;
663 if (TYPE_PRECISION (TREE_TYPE (lhs)) == TYPE_PRECISION (TREE_TYPE (arg0)))
665 /* Avoid performing the arithmetics in bool type which has different
666 semantics, otherwise prefer unsigned types from the two with
667 the same precision. */
668 if (TREE_CODE (TREE_TYPE (arg0)) == BOOLEAN_TYPE
669 || !TYPE_UNSIGNED (TREE_TYPE (arg0)))
670 type = TREE_TYPE (lhs);
671 else
672 type = TREE_TYPE (arg0);
674 else if (TYPE_PRECISION (TREE_TYPE (lhs)) > TYPE_PRECISION (TREE_TYPE (arg0)))
675 type = TREE_TYPE (lhs);
676 else
677 type = TREE_TYPE (arg0);
679 min = wide_int::from (min, TYPE_PRECISION (type),
680 TYPE_SIGN (TREE_TYPE (lhs)));
681 wide_int a = wide_int::from (wi::to_wide (arg0), TYPE_PRECISION (type),
682 TYPE_SIGN (TREE_TYPE (arg0)));
683 enum tree_code code;
684 wi::overflow_type ovf;
685 if (tree_int_cst_lt (arg0, arg1))
687 code = PLUS_EXPR;
688 a -= min;
689 if (!TYPE_UNSIGNED (type))
691 /* lhs is known to be in range [min, min+1] and we want to add a
692 to it. Check if that operation can overflow for those 2 values
693 and if yes, force unsigned type. */
694 wi::add (min + (wi::neg_p (a) ? 0 : 1), a, SIGNED, &ovf);
695 if (ovf)
696 type = unsigned_type_for (type);
699 else
701 code = MINUS_EXPR;
702 a += min;
703 if (!TYPE_UNSIGNED (type))
705 /* lhs is known to be in range [min, min+1] and we want to subtract
706 it from a. Check if that operation can overflow for those 2
707 values and if yes, force unsigned type. */
708 wi::sub (a, min + (wi::neg_p (min) ? 0 : 1), SIGNED, &ovf);
709 if (ovf)
710 type = unsigned_type_for (type);
714 tree arg = wide_int_to_tree (type, a);
715 gimple_stmt_iterator gsi = gsi_for_stmt (stmt);
716 if (!useless_type_conversion_p (type, TREE_TYPE (lhs)))
717 lhs = gimplify_build1 (&gsi, NOP_EXPR, type, lhs);
718 tree new_rhs;
719 if (code == PLUS_EXPR)
720 new_rhs = gimplify_build2 (&gsi, PLUS_EXPR, type, lhs, arg);
721 else
722 new_rhs = gimplify_build2 (&gsi, MINUS_EXPR, type, arg, lhs);
723 if (!useless_type_conversion_p (TREE_TYPE (arg0), type))
724 new_rhs = gimplify_build1 (&gsi, NOP_EXPR, TREE_TYPE (arg0), new_rhs);
726 replace_phi_edge_with_variable (cond_bb, e1, phi, new_rhs);
728 /* Note that we optimized this PHI. */
729 return true;
732 /* The function conditional_replacement does the main work of doing the
733 conditional replacement. Return true if the replacement is done.
734 Otherwise return false.
735 BB is the basic block where the replacement is going to be done on. ARG0
736 is argument 0 from PHI. Likewise for ARG1. */
738 static bool
739 conditional_replacement (basic_block cond_bb, basic_block middle_bb,
740 edge e0, edge e1, gphi *phi,
741 tree arg0, tree arg1)
743 tree result;
744 gimple *stmt;
745 gassign *new_stmt;
746 tree cond;
747 gimple_stmt_iterator gsi;
748 edge true_edge, false_edge;
749 tree new_var, new_var2;
750 bool neg;
752 /* FIXME: Gimplification of complex type is too hard for now. */
753 /* We aren't prepared to handle vectors either (and it is a question
754 if it would be worthwhile anyway). */
755 if (!(INTEGRAL_TYPE_P (TREE_TYPE (arg0))
756 || POINTER_TYPE_P (TREE_TYPE (arg0)))
757 || !(INTEGRAL_TYPE_P (TREE_TYPE (arg1))
758 || POINTER_TYPE_P (TREE_TYPE (arg1))))
759 return false;
761 /* The PHI arguments have the constants 0 and 1, or 0 and -1, then
762 convert it to the conditional. */
763 if ((integer_zerop (arg0) && integer_onep (arg1))
764 || (integer_zerop (arg1) && integer_onep (arg0)))
765 neg = false;
766 else if ((integer_zerop (arg0) && integer_all_onesp (arg1))
767 || (integer_zerop (arg1) && integer_all_onesp (arg0)))
768 neg = true;
769 else
770 return false;
772 if (!empty_block_p (middle_bb))
773 return false;
775 /* At this point we know we have a GIMPLE_COND with two successors.
776 One successor is BB, the other successor is an empty block which
777 falls through into BB.
779 There is a single PHI node at the join point (BB) and its arguments
780 are constants (0, 1) or (0, -1).
782 So, given the condition COND, and the two PHI arguments, we can
783 rewrite this PHI into non-branching code:
785 dest = (COND) or dest = COND'
787 We use the condition as-is if the argument associated with the
788 true edge has the value one or the argument associated with the
789 false edge as the value zero. Note that those conditions are not
790 the same since only one of the outgoing edges from the GIMPLE_COND
791 will directly reach BB and thus be associated with an argument. */
793 stmt = last_stmt (cond_bb);
794 result = PHI_RESULT (phi);
796 /* To handle special cases like floating point comparison, it is easier and
797 less error-prone to build a tree and gimplify it on the fly though it is
798 less efficient. */
799 cond = fold_build2_loc (gimple_location (stmt),
800 gimple_cond_code (stmt), boolean_type_node,
801 gimple_cond_lhs (stmt), gimple_cond_rhs (stmt));
803 /* We need to know which is the true edge and which is the false
804 edge so that we know when to invert the condition below. */
805 extract_true_false_edges_from_block (cond_bb, &true_edge, &false_edge);
806 if ((e0 == true_edge && integer_zerop (arg0))
807 || (e0 == false_edge && !integer_zerop (arg0))
808 || (e1 == true_edge && integer_zerop (arg1))
809 || (e1 == false_edge && !integer_zerop (arg1)))
810 cond = fold_build1_loc (gimple_location (stmt),
811 TRUTH_NOT_EXPR, TREE_TYPE (cond), cond);
813 if (neg)
815 cond = fold_convert_loc (gimple_location (stmt),
816 TREE_TYPE (result), cond);
817 cond = fold_build1_loc (gimple_location (stmt),
818 NEGATE_EXPR, TREE_TYPE (cond), cond);
821 /* Insert our new statements at the end of conditional block before the
822 COND_STMT. */
823 gsi = gsi_for_stmt (stmt);
824 new_var = force_gimple_operand_gsi (&gsi, cond, true, NULL, true,
825 GSI_SAME_STMT);
827 if (!useless_type_conversion_p (TREE_TYPE (result), TREE_TYPE (new_var)))
829 location_t locus_0, locus_1;
831 new_var2 = make_ssa_name (TREE_TYPE (result));
832 new_stmt = gimple_build_assign (new_var2, CONVERT_EXPR, new_var);
833 gsi_insert_before (&gsi, new_stmt, GSI_SAME_STMT);
834 new_var = new_var2;
836 /* Set the locus to the first argument, unless is doesn't have one. */
837 locus_0 = gimple_phi_arg_location (phi, 0);
838 locus_1 = gimple_phi_arg_location (phi, 1);
839 if (locus_0 == UNKNOWN_LOCATION)
840 locus_0 = locus_1;
841 gimple_set_location (new_stmt, locus_0);
844 replace_phi_edge_with_variable (cond_bb, e1, phi, new_var);
846 /* Note that we optimized this PHI. */
847 return true;
850 /* Update *ARG which is defined in STMT so that it contains the
851 computed value if that seems profitable. Return true if the
852 statement is made dead by that rewriting. */
854 static bool
855 jump_function_from_stmt (tree *arg, gimple *stmt)
857 enum tree_code code = gimple_assign_rhs_code (stmt);
858 if (code == ADDR_EXPR)
860 /* For arg = &p->i transform it to p, if possible. */
861 tree rhs1 = gimple_assign_rhs1 (stmt);
862 poly_int64 offset;
863 tree tem = get_addr_base_and_unit_offset (TREE_OPERAND (rhs1, 0),
864 &offset);
865 if (tem
866 && TREE_CODE (tem) == MEM_REF
867 && known_eq (mem_ref_offset (tem) + offset, 0))
869 *arg = TREE_OPERAND (tem, 0);
870 return true;
873 /* TODO: Much like IPA-CP jump-functions we want to handle constant
874 additions symbolically here, and we'd need to update the comparison
875 code that compares the arg + cst tuples in our caller. For now the
876 code above exactly handles the VEC_BASE pattern from vec.h. */
877 return false;
880 /* RHS is a source argument in a BIT_AND_EXPR which feeds a conditional
881 of the form SSA_NAME NE 0.
883 If RHS is fed by a simple EQ_EXPR comparison of two values, see if
884 the two input values of the EQ_EXPR match arg0 and arg1.
886 If so update *code and return TRUE. Otherwise return FALSE. */
888 static bool
889 rhs_is_fed_for_value_replacement (const_tree arg0, const_tree arg1,
890 enum tree_code *code, const_tree rhs)
892 /* Obviously if RHS is not an SSA_NAME, we can't look at the defining
893 statement. */
894 if (TREE_CODE (rhs) == SSA_NAME)
896 gimple *def1 = SSA_NAME_DEF_STMT (rhs);
898 /* Verify the defining statement has an EQ_EXPR on the RHS. */
899 if (is_gimple_assign (def1) && gimple_assign_rhs_code (def1) == EQ_EXPR)
901 /* Finally verify the source operands of the EQ_EXPR are equal
902 to arg0 and arg1. */
903 tree op0 = gimple_assign_rhs1 (def1);
904 tree op1 = gimple_assign_rhs2 (def1);
905 if ((operand_equal_for_phi_arg_p (arg0, op0)
906 && operand_equal_for_phi_arg_p (arg1, op1))
907 || (operand_equal_for_phi_arg_p (arg0, op1)
908 && operand_equal_for_phi_arg_p (arg1, op0)))
910 /* We will perform the optimization. */
911 *code = gimple_assign_rhs_code (def1);
912 return true;
916 return false;
919 /* Return TRUE if arg0/arg1 are equal to the rhs/lhs or lhs/rhs of COND.
921 Also return TRUE if arg0/arg1 are equal to the source arguments of a
922 an EQ comparison feeding a BIT_AND_EXPR which feeds COND.
924 Return FALSE otherwise. */
926 static bool
927 operand_equal_for_value_replacement (const_tree arg0, const_tree arg1,
928 enum tree_code *code, gimple *cond)
930 gimple *def;
931 tree lhs = gimple_cond_lhs (cond);
932 tree rhs = gimple_cond_rhs (cond);
934 if ((operand_equal_for_phi_arg_p (arg0, lhs)
935 && operand_equal_for_phi_arg_p (arg1, rhs))
936 || (operand_equal_for_phi_arg_p (arg1, lhs)
937 && operand_equal_for_phi_arg_p (arg0, rhs)))
938 return true;
940 /* Now handle more complex case where we have an EQ comparison
941 which feeds a BIT_AND_EXPR which feeds COND.
943 First verify that COND is of the form SSA_NAME NE 0. */
944 if (*code != NE_EXPR || !integer_zerop (rhs)
945 || TREE_CODE (lhs) != SSA_NAME)
946 return false;
948 /* Now ensure that SSA_NAME is set by a BIT_AND_EXPR. */
949 def = SSA_NAME_DEF_STMT (lhs);
950 if (!is_gimple_assign (def) || gimple_assign_rhs_code (def) != BIT_AND_EXPR)
951 return false;
953 /* Now verify arg0/arg1 correspond to the source arguments of an
954 EQ comparison feeding the BIT_AND_EXPR. */
956 tree tmp = gimple_assign_rhs1 (def);
957 if (rhs_is_fed_for_value_replacement (arg0, arg1, code, tmp))
958 return true;
960 tmp = gimple_assign_rhs2 (def);
961 if (rhs_is_fed_for_value_replacement (arg0, arg1, code, tmp))
962 return true;
964 return false;
967 /* Returns true if ARG is a neutral element for operation CODE
968 on the RIGHT side. */
970 static bool
971 neutral_element_p (tree_code code, tree arg, bool right)
973 switch (code)
975 case PLUS_EXPR:
976 case BIT_IOR_EXPR:
977 case BIT_XOR_EXPR:
978 return integer_zerop (arg);
980 case LROTATE_EXPR:
981 case RROTATE_EXPR:
982 case LSHIFT_EXPR:
983 case RSHIFT_EXPR:
984 case MINUS_EXPR:
985 case POINTER_PLUS_EXPR:
986 return right && integer_zerop (arg);
988 case MULT_EXPR:
989 return integer_onep (arg);
991 case TRUNC_DIV_EXPR:
992 case CEIL_DIV_EXPR:
993 case FLOOR_DIV_EXPR:
994 case ROUND_DIV_EXPR:
995 case EXACT_DIV_EXPR:
996 return right && integer_onep (arg);
998 case BIT_AND_EXPR:
999 return integer_all_onesp (arg);
1001 default:
1002 return false;
1006 /* Returns true if ARG is an absorbing element for operation CODE. */
1008 static bool
1009 absorbing_element_p (tree_code code, tree arg, bool right, tree rval)
1011 switch (code)
1013 case BIT_IOR_EXPR:
1014 return integer_all_onesp (arg);
1016 case MULT_EXPR:
1017 case BIT_AND_EXPR:
1018 return integer_zerop (arg);
1020 case LSHIFT_EXPR:
1021 case RSHIFT_EXPR:
1022 case LROTATE_EXPR:
1023 case RROTATE_EXPR:
1024 return !right && integer_zerop (arg);
1026 case TRUNC_DIV_EXPR:
1027 case CEIL_DIV_EXPR:
1028 case FLOOR_DIV_EXPR:
1029 case ROUND_DIV_EXPR:
1030 case EXACT_DIV_EXPR:
1031 case TRUNC_MOD_EXPR:
1032 case CEIL_MOD_EXPR:
1033 case FLOOR_MOD_EXPR:
1034 case ROUND_MOD_EXPR:
1035 return (!right
1036 && integer_zerop (arg)
1037 && tree_single_nonzero_warnv_p (rval, NULL));
1039 default:
1040 return false;
1044 /* The function value_replacement does the main work of doing the value
1045 replacement. Return non-zero if the replacement is done. Otherwise return
1046 0. If we remove the middle basic block, return 2.
1047 BB is the basic block where the replacement is going to be done on. ARG0
1048 is argument 0 from the PHI. Likewise for ARG1. */
1050 static int
1051 value_replacement (basic_block cond_bb, basic_block middle_bb,
1052 edge e0, edge e1, gimple *phi,
1053 tree arg0, tree arg1)
1055 gimple_stmt_iterator gsi;
1056 gimple *cond;
1057 edge true_edge, false_edge;
1058 enum tree_code code;
1059 bool emtpy_or_with_defined_p = true;
1061 /* If the type says honor signed zeros we cannot do this
1062 optimization. */
1063 if (HONOR_SIGNED_ZEROS (arg1))
1064 return 0;
1066 /* If there is a statement in MIDDLE_BB that defines one of the PHI
1067 arguments, then adjust arg0 or arg1. */
1068 gsi = gsi_start_nondebug_after_labels_bb (middle_bb);
1069 while (!gsi_end_p (gsi))
1071 gimple *stmt = gsi_stmt (gsi);
1072 tree lhs;
1073 gsi_next_nondebug (&gsi);
1074 if (!is_gimple_assign (stmt))
1076 if (gimple_code (stmt) != GIMPLE_PREDICT
1077 && gimple_code (stmt) != GIMPLE_NOP)
1078 emtpy_or_with_defined_p = false;
1079 continue;
1081 /* Now try to adjust arg0 or arg1 according to the computation
1082 in the statement. */
1083 lhs = gimple_assign_lhs (stmt);
1084 if (!(lhs == arg0
1085 && jump_function_from_stmt (&arg0, stmt))
1086 || (lhs == arg1
1087 && jump_function_from_stmt (&arg1, stmt)))
1088 emtpy_or_with_defined_p = false;
1091 cond = last_stmt (cond_bb);
1092 code = gimple_cond_code (cond);
1094 /* This transformation is only valid for equality comparisons. */
1095 if (code != NE_EXPR && code != EQ_EXPR)
1096 return 0;
1098 /* We need to know which is the true edge and which is the false
1099 edge so that we know if have abs or negative abs. */
1100 extract_true_false_edges_from_block (cond_bb, &true_edge, &false_edge);
1102 /* At this point we know we have a COND_EXPR with two successors.
1103 One successor is BB, the other successor is an empty block which
1104 falls through into BB.
1106 The condition for the COND_EXPR is known to be NE_EXPR or EQ_EXPR.
1108 There is a single PHI node at the join point (BB) with two arguments.
1110 We now need to verify that the two arguments in the PHI node match
1111 the two arguments to the equality comparison. */
1113 if (operand_equal_for_value_replacement (arg0, arg1, &code, cond))
1115 edge e;
1116 tree arg;
1118 /* For NE_EXPR, we want to build an assignment result = arg where
1119 arg is the PHI argument associated with the true edge. For
1120 EQ_EXPR we want the PHI argument associated with the false edge. */
1121 e = (code == NE_EXPR ? true_edge : false_edge);
1123 /* Unfortunately, E may not reach BB (it may instead have gone to
1124 OTHER_BLOCK). If that is the case, then we want the single outgoing
1125 edge from OTHER_BLOCK which reaches BB and represents the desired
1126 path from COND_BLOCK. */
1127 if (e->dest == middle_bb)
1128 e = single_succ_edge (e->dest);
1130 /* Now we know the incoming edge to BB that has the argument for the
1131 RHS of our new assignment statement. */
1132 if (e0 == e)
1133 arg = arg0;
1134 else
1135 arg = arg1;
1137 /* If the middle basic block was empty or is defining the
1138 PHI arguments and this is a single phi where the args are different
1139 for the edges e0 and e1 then we can remove the middle basic block. */
1140 if (emtpy_or_with_defined_p
1141 && single_non_singleton_phi_for_edges (phi_nodes (gimple_bb (phi)),
1142 e0, e1) == phi)
1144 replace_phi_edge_with_variable (cond_bb, e1, phi, arg);
1145 /* Note that we optimized this PHI. */
1146 return 2;
1148 else
1150 /* Replace the PHI arguments with arg. */
1151 SET_PHI_ARG_DEF (phi, e0->dest_idx, arg);
1152 SET_PHI_ARG_DEF (phi, e1->dest_idx, arg);
1153 if (dump_file && (dump_flags & TDF_DETAILS))
1155 fprintf (dump_file, "PHI ");
1156 print_generic_expr (dump_file, gimple_phi_result (phi));
1157 fprintf (dump_file, " reduced for COND_EXPR in block %d to ",
1158 cond_bb->index);
1159 print_generic_expr (dump_file, arg);
1160 fprintf (dump_file, ".\n");
1162 return 1;
1167 /* Now optimize (x != 0) ? x + y : y to just x + y. */
1168 gsi = gsi_last_nondebug_bb (middle_bb);
1169 if (gsi_end_p (gsi))
1170 return 0;
1172 gimple *assign = gsi_stmt (gsi);
1173 if (!is_gimple_assign (assign)
1174 || gimple_assign_rhs_class (assign) != GIMPLE_BINARY_RHS
1175 || (!INTEGRAL_TYPE_P (TREE_TYPE (arg0))
1176 && !POINTER_TYPE_P (TREE_TYPE (arg0))))
1177 return 0;
1179 /* Punt if there are (degenerate) PHIs in middle_bb, there should not be. */
1180 if (!gimple_seq_empty_p (phi_nodes (middle_bb)))
1181 return 0;
1183 /* Allow up to 2 cheap preparation statements that prepare argument
1184 for assign, e.g.:
1185 if (y_4 != 0)
1186 goto <bb 3>;
1187 else
1188 goto <bb 4>;
1189 <bb 3>:
1190 _1 = (int) y_4;
1191 iftmp.0_6 = x_5(D) r<< _1;
1192 <bb 4>:
1193 # iftmp.0_2 = PHI <iftmp.0_6(3), x_5(D)(2)>
1195 if (y_3(D) == 0)
1196 goto <bb 4>;
1197 else
1198 goto <bb 3>;
1199 <bb 3>:
1200 y_4 = y_3(D) & 31;
1201 _1 = (int) y_4;
1202 _6 = x_5(D) r<< _1;
1203 <bb 4>:
1204 # _2 = PHI <x_5(D)(2), _6(3)> */
1205 gimple *prep_stmt[2] = { NULL, NULL };
1206 int prep_cnt;
1207 for (prep_cnt = 0; ; prep_cnt++)
1209 gsi_prev_nondebug (&gsi);
1210 if (gsi_end_p (gsi))
1211 break;
1213 gimple *g = gsi_stmt (gsi);
1214 if (gimple_code (g) == GIMPLE_LABEL)
1215 break;
1217 if (prep_cnt == 2 || !is_gimple_assign (g))
1218 return 0;
1220 tree lhs = gimple_assign_lhs (g);
1221 tree rhs1 = gimple_assign_rhs1 (g);
1222 use_operand_p use_p;
1223 gimple *use_stmt;
1224 if (TREE_CODE (lhs) != SSA_NAME
1225 || TREE_CODE (rhs1) != SSA_NAME
1226 || !INTEGRAL_TYPE_P (TREE_TYPE (lhs))
1227 || !INTEGRAL_TYPE_P (TREE_TYPE (rhs1))
1228 || !single_imm_use (lhs, &use_p, &use_stmt)
1229 || use_stmt != (prep_cnt ? prep_stmt[prep_cnt - 1] : assign))
1230 return 0;
1231 switch (gimple_assign_rhs_code (g))
1233 CASE_CONVERT:
1234 break;
1235 case PLUS_EXPR:
1236 case BIT_AND_EXPR:
1237 case BIT_IOR_EXPR:
1238 case BIT_XOR_EXPR:
1239 if (TREE_CODE (gimple_assign_rhs2 (g)) != INTEGER_CST)
1240 return 0;
1241 break;
1242 default:
1243 return 0;
1245 prep_stmt[prep_cnt] = g;
1248 /* Only transform if it removes the condition. */
1249 if (!single_non_singleton_phi_for_edges (phi_nodes (gimple_bb (phi)), e0, e1))
1250 return 0;
1252 /* Size-wise, this is always profitable. */
1253 if (optimize_bb_for_speed_p (cond_bb)
1254 /* The special case is useless if it has a low probability. */
1255 && profile_status_for_fn (cfun) != PROFILE_ABSENT
1256 && EDGE_PRED (middle_bb, 0)->probability < profile_probability::even ()
1257 /* If assign is cheap, there is no point avoiding it. */
1258 && estimate_num_insns (bb_seq (middle_bb), &eni_time_weights)
1259 >= 3 * estimate_num_insns (cond, &eni_time_weights))
1260 return 0;
1262 tree lhs = gimple_assign_lhs (assign);
1263 tree rhs1 = gimple_assign_rhs1 (assign);
1264 tree rhs2 = gimple_assign_rhs2 (assign);
1265 enum tree_code code_def = gimple_assign_rhs_code (assign);
1266 tree cond_lhs = gimple_cond_lhs (cond);
1267 tree cond_rhs = gimple_cond_rhs (cond);
1269 /* Propagate the cond_rhs constant through preparation stmts,
1270 make sure UB isn't invoked while doing that. */
1271 for (int i = prep_cnt - 1; i >= 0; --i)
1273 gimple *g = prep_stmt[i];
1274 tree grhs1 = gimple_assign_rhs1 (g);
1275 if (!operand_equal_for_phi_arg_p (cond_lhs, grhs1))
1276 return 0;
1277 cond_lhs = gimple_assign_lhs (g);
1278 cond_rhs = fold_convert (TREE_TYPE (grhs1), cond_rhs);
1279 if (TREE_CODE (cond_rhs) != INTEGER_CST
1280 || TREE_OVERFLOW (cond_rhs))
1281 return 0;
1282 if (gimple_assign_rhs_class (g) == GIMPLE_BINARY_RHS)
1284 cond_rhs = int_const_binop (gimple_assign_rhs_code (g), cond_rhs,
1285 gimple_assign_rhs2 (g));
1286 if (TREE_OVERFLOW (cond_rhs))
1287 return 0;
1289 cond_rhs = fold_convert (TREE_TYPE (cond_lhs), cond_rhs);
1290 if (TREE_CODE (cond_rhs) != INTEGER_CST
1291 || TREE_OVERFLOW (cond_rhs))
1292 return 0;
1295 if (((code == NE_EXPR && e1 == false_edge)
1296 || (code == EQ_EXPR && e1 == true_edge))
1297 && arg0 == lhs
1298 && ((arg1 == rhs1
1299 && operand_equal_for_phi_arg_p (rhs2, cond_lhs)
1300 && neutral_element_p (code_def, cond_rhs, true))
1301 || (arg1 == rhs2
1302 && operand_equal_for_phi_arg_p (rhs1, cond_lhs)
1303 && neutral_element_p (code_def, cond_rhs, false))
1304 || (operand_equal_for_phi_arg_p (arg1, cond_rhs)
1305 && ((operand_equal_for_phi_arg_p (rhs2, cond_lhs)
1306 && absorbing_element_p (code_def, cond_rhs, true, rhs2))
1307 || (operand_equal_for_phi_arg_p (rhs1, cond_lhs)
1308 && absorbing_element_p (code_def,
1309 cond_rhs, false, rhs2))))))
1311 gsi = gsi_for_stmt (cond);
1312 /* Moving ASSIGN might change VR of lhs, e.g. when moving u_6
1313 def-stmt in:
1314 if (n_5 != 0)
1315 goto <bb 3>;
1316 else
1317 goto <bb 4>;
1319 <bb 3>:
1320 # RANGE [0, 4294967294]
1321 u_6 = n_5 + 4294967295;
1323 <bb 4>:
1324 # u_3 = PHI <u_6(3), 4294967295(2)> */
1325 reset_flow_sensitive_info (lhs);
1326 if (INTEGRAL_TYPE_P (TREE_TYPE (lhs)))
1328 /* If available, we can use VR of phi result at least. */
1329 tree phires = gimple_phi_result (phi);
1330 struct range_info_def *phires_range_info
1331 = SSA_NAME_RANGE_INFO (phires);
1332 if (phires_range_info)
1333 duplicate_ssa_name_range_info (lhs, SSA_NAME_RANGE_TYPE (phires),
1334 phires_range_info);
1336 gimple_stmt_iterator gsi_from;
1337 for (int i = prep_cnt - 1; i >= 0; --i)
1339 tree plhs = gimple_assign_lhs (prep_stmt[i]);
1340 reset_flow_sensitive_info (plhs);
1341 gsi_from = gsi_for_stmt (prep_stmt[i]);
1342 gsi_move_before (&gsi_from, &gsi);
1344 gsi_from = gsi_for_stmt (assign);
1345 gsi_move_before (&gsi_from, &gsi);
1346 replace_phi_edge_with_variable (cond_bb, e1, phi, lhs);
1347 return 2;
1350 return 0;
1353 /* The function minmax_replacement does the main work of doing the minmax
1354 replacement. Return true if the replacement is done. Otherwise return
1355 false.
1356 BB is the basic block where the replacement is going to be done on. ARG0
1357 is argument 0 from the PHI. Likewise for ARG1. */
1359 static bool
1360 minmax_replacement (basic_block cond_bb, basic_block middle_bb,
1361 edge e0, edge e1, gimple *phi,
1362 tree arg0, tree arg1)
1364 tree result, type, rhs;
1365 gcond *cond;
1366 gassign *new_stmt;
1367 edge true_edge, false_edge;
1368 enum tree_code cmp, minmax, ass_code;
1369 tree smaller, alt_smaller, larger, alt_larger, arg_true, arg_false;
1370 gimple_stmt_iterator gsi, gsi_from;
1372 type = TREE_TYPE (PHI_RESULT (phi));
1374 /* The optimization may be unsafe due to NaNs. */
1375 if (HONOR_NANS (type) || HONOR_SIGNED_ZEROS (type))
1376 return false;
1378 cond = as_a <gcond *> (last_stmt (cond_bb));
1379 cmp = gimple_cond_code (cond);
1380 rhs = gimple_cond_rhs (cond);
1382 /* Turn EQ/NE of extreme values to order comparisons. */
1383 if ((cmp == NE_EXPR || cmp == EQ_EXPR)
1384 && TREE_CODE (rhs) == INTEGER_CST)
1386 if (wi::eq_p (wi::to_wide (rhs), wi::min_value (TREE_TYPE (rhs))))
1388 cmp = (cmp == EQ_EXPR) ? LT_EXPR : GE_EXPR;
1389 rhs = wide_int_to_tree (TREE_TYPE (rhs),
1390 wi::min_value (TREE_TYPE (rhs)) + 1);
1392 else if (wi::eq_p (wi::to_wide (rhs), wi::max_value (TREE_TYPE (rhs))))
1394 cmp = (cmp == EQ_EXPR) ? GT_EXPR : LE_EXPR;
1395 rhs = wide_int_to_tree (TREE_TYPE (rhs),
1396 wi::max_value (TREE_TYPE (rhs)) - 1);
1400 /* This transformation is only valid for order comparisons. Record which
1401 operand is smaller/larger if the result of the comparison is true. */
1402 alt_smaller = NULL_TREE;
1403 alt_larger = NULL_TREE;
1404 if (cmp == LT_EXPR || cmp == LE_EXPR)
1406 smaller = gimple_cond_lhs (cond);
1407 larger = rhs;
1408 /* If we have smaller < CST it is equivalent to smaller <= CST-1.
1409 Likewise smaller <= CST is equivalent to smaller < CST+1. */
1410 if (TREE_CODE (larger) == INTEGER_CST)
1412 if (cmp == LT_EXPR)
1414 wi::overflow_type overflow;
1415 wide_int alt = wi::sub (wi::to_wide (larger), 1,
1416 TYPE_SIGN (TREE_TYPE (larger)),
1417 &overflow);
1418 if (! overflow)
1419 alt_larger = wide_int_to_tree (TREE_TYPE (larger), alt);
1421 else
1423 wi::overflow_type overflow;
1424 wide_int alt = wi::add (wi::to_wide (larger), 1,
1425 TYPE_SIGN (TREE_TYPE (larger)),
1426 &overflow);
1427 if (! overflow)
1428 alt_larger = wide_int_to_tree (TREE_TYPE (larger), alt);
1432 else if (cmp == GT_EXPR || cmp == GE_EXPR)
1434 smaller = rhs;
1435 larger = gimple_cond_lhs (cond);
1436 /* If we have larger > CST it is equivalent to larger >= CST+1.
1437 Likewise larger >= CST is equivalent to larger > CST-1. */
1438 if (TREE_CODE (smaller) == INTEGER_CST)
1440 wi::overflow_type overflow;
1441 if (cmp == GT_EXPR)
1443 wide_int alt = wi::add (wi::to_wide (smaller), 1,
1444 TYPE_SIGN (TREE_TYPE (smaller)),
1445 &overflow);
1446 if (! overflow)
1447 alt_smaller = wide_int_to_tree (TREE_TYPE (smaller), alt);
1449 else
1451 wide_int alt = wi::sub (wi::to_wide (smaller), 1,
1452 TYPE_SIGN (TREE_TYPE (smaller)),
1453 &overflow);
1454 if (! overflow)
1455 alt_smaller = wide_int_to_tree (TREE_TYPE (smaller), alt);
1459 else
1460 return false;
1462 /* We need to know which is the true edge and which is the false
1463 edge so that we know if have abs or negative abs. */
1464 extract_true_false_edges_from_block (cond_bb, &true_edge, &false_edge);
1466 /* Forward the edges over the middle basic block. */
1467 if (true_edge->dest == middle_bb)
1468 true_edge = EDGE_SUCC (true_edge->dest, 0);
1469 if (false_edge->dest == middle_bb)
1470 false_edge = EDGE_SUCC (false_edge->dest, 0);
1472 if (true_edge == e0)
1474 gcc_assert (false_edge == e1);
1475 arg_true = arg0;
1476 arg_false = arg1;
1478 else
1480 gcc_assert (false_edge == e0);
1481 gcc_assert (true_edge == e1);
1482 arg_true = arg1;
1483 arg_false = arg0;
1486 if (empty_block_p (middle_bb))
1488 if ((operand_equal_for_phi_arg_p (arg_true, smaller)
1489 || (alt_smaller
1490 && operand_equal_for_phi_arg_p (arg_true, alt_smaller)))
1491 && (operand_equal_for_phi_arg_p (arg_false, larger)
1492 || (alt_larger
1493 && operand_equal_for_phi_arg_p (arg_true, alt_larger))))
1495 /* Case
1497 if (smaller < larger)
1498 rslt = smaller;
1499 else
1500 rslt = larger; */
1501 minmax = MIN_EXPR;
1503 else if ((operand_equal_for_phi_arg_p (arg_false, smaller)
1504 || (alt_smaller
1505 && operand_equal_for_phi_arg_p (arg_false, alt_smaller)))
1506 && (operand_equal_for_phi_arg_p (arg_true, larger)
1507 || (alt_larger
1508 && operand_equal_for_phi_arg_p (arg_true, alt_larger))))
1509 minmax = MAX_EXPR;
1510 else
1511 return false;
1513 else
1515 /* Recognize the following case, assuming d <= u:
1517 if (a <= u)
1518 b = MAX (a, d);
1519 x = PHI <b, u>
1521 This is equivalent to
1523 b = MAX (a, d);
1524 x = MIN (b, u); */
1526 gimple *assign = last_and_only_stmt (middle_bb);
1527 tree lhs, op0, op1, bound;
1529 if (!assign
1530 || gimple_code (assign) != GIMPLE_ASSIGN)
1531 return false;
1533 lhs = gimple_assign_lhs (assign);
1534 ass_code = gimple_assign_rhs_code (assign);
1535 if (ass_code != MAX_EXPR && ass_code != MIN_EXPR)
1536 return false;
1537 op0 = gimple_assign_rhs1 (assign);
1538 op1 = gimple_assign_rhs2 (assign);
1540 if (true_edge->src == middle_bb)
1542 /* We got here if the condition is true, i.e., SMALLER < LARGER. */
1543 if (!operand_equal_for_phi_arg_p (lhs, arg_true))
1544 return false;
1546 if (operand_equal_for_phi_arg_p (arg_false, larger)
1547 || (alt_larger
1548 && operand_equal_for_phi_arg_p (arg_false, alt_larger)))
1550 /* Case
1552 if (smaller < larger)
1554 r' = MAX_EXPR (smaller, bound)
1556 r = PHI <r', larger> --> to be turned to MIN_EXPR. */
1557 if (ass_code != MAX_EXPR)
1558 return false;
1560 minmax = MIN_EXPR;
1561 if (operand_equal_for_phi_arg_p (op0, smaller)
1562 || (alt_smaller
1563 && operand_equal_for_phi_arg_p (op0, alt_smaller)))
1564 bound = op1;
1565 else if (operand_equal_for_phi_arg_p (op1, smaller)
1566 || (alt_smaller
1567 && operand_equal_for_phi_arg_p (op1, alt_smaller)))
1568 bound = op0;
1569 else
1570 return false;
1572 /* We need BOUND <= LARGER. */
1573 if (!integer_nonzerop (fold_build2 (LE_EXPR, boolean_type_node,
1574 bound, larger)))
1575 return false;
1577 else if (operand_equal_for_phi_arg_p (arg_false, smaller)
1578 || (alt_smaller
1579 && operand_equal_for_phi_arg_p (arg_false, alt_smaller)))
1581 /* Case
1583 if (smaller < larger)
1585 r' = MIN_EXPR (larger, bound)
1587 r = PHI <r', smaller> --> to be turned to MAX_EXPR. */
1588 if (ass_code != MIN_EXPR)
1589 return false;
1591 minmax = MAX_EXPR;
1592 if (operand_equal_for_phi_arg_p (op0, larger)
1593 || (alt_larger
1594 && operand_equal_for_phi_arg_p (op0, alt_larger)))
1595 bound = op1;
1596 else if (operand_equal_for_phi_arg_p (op1, larger)
1597 || (alt_larger
1598 && operand_equal_for_phi_arg_p (op1, alt_larger)))
1599 bound = op0;
1600 else
1601 return false;
1603 /* We need BOUND >= SMALLER. */
1604 if (!integer_nonzerop (fold_build2 (GE_EXPR, boolean_type_node,
1605 bound, smaller)))
1606 return false;
1608 else
1609 return false;
1611 else
1613 /* We got here if the condition is false, i.e., SMALLER > LARGER. */
1614 if (!operand_equal_for_phi_arg_p (lhs, arg_false))
1615 return false;
1617 if (operand_equal_for_phi_arg_p (arg_true, larger)
1618 || (alt_larger
1619 && operand_equal_for_phi_arg_p (arg_true, alt_larger)))
1621 /* Case
1623 if (smaller > larger)
1625 r' = MIN_EXPR (smaller, bound)
1627 r = PHI <r', larger> --> to be turned to MAX_EXPR. */
1628 if (ass_code != MIN_EXPR)
1629 return false;
1631 minmax = MAX_EXPR;
1632 if (operand_equal_for_phi_arg_p (op0, smaller)
1633 || (alt_smaller
1634 && operand_equal_for_phi_arg_p (op0, alt_smaller)))
1635 bound = op1;
1636 else if (operand_equal_for_phi_arg_p (op1, smaller)
1637 || (alt_smaller
1638 && operand_equal_for_phi_arg_p (op1, alt_smaller)))
1639 bound = op0;
1640 else
1641 return false;
1643 /* We need BOUND >= LARGER. */
1644 if (!integer_nonzerop (fold_build2 (GE_EXPR, boolean_type_node,
1645 bound, larger)))
1646 return false;
1648 else if (operand_equal_for_phi_arg_p (arg_true, smaller)
1649 || (alt_smaller
1650 && operand_equal_for_phi_arg_p (arg_true, alt_smaller)))
1652 /* Case
1654 if (smaller > larger)
1656 r' = MAX_EXPR (larger, bound)
1658 r = PHI <r', smaller> --> to be turned to MIN_EXPR. */
1659 if (ass_code != MAX_EXPR)
1660 return false;
1662 minmax = MIN_EXPR;
1663 if (operand_equal_for_phi_arg_p (op0, larger))
1664 bound = op1;
1665 else if (operand_equal_for_phi_arg_p (op1, larger))
1666 bound = op0;
1667 else
1668 return false;
1670 /* We need BOUND <= SMALLER. */
1671 if (!integer_nonzerop (fold_build2 (LE_EXPR, boolean_type_node,
1672 bound, smaller)))
1673 return false;
1675 else
1676 return false;
1679 /* Move the statement from the middle block. */
1680 gsi = gsi_last_bb (cond_bb);
1681 gsi_from = gsi_last_nondebug_bb (middle_bb);
1682 reset_flow_sensitive_info (SINGLE_SSA_TREE_OPERAND (gsi_stmt (gsi_from),
1683 SSA_OP_DEF));
1684 gsi_move_before (&gsi_from, &gsi);
1687 /* Create an SSA var to hold the min/max result. If we're the only
1688 things setting the target PHI, then we can clone the PHI
1689 variable. Otherwise we must create a new one. */
1690 result = PHI_RESULT (phi);
1691 if (EDGE_COUNT (gimple_bb (phi)->preds) == 2)
1692 result = duplicate_ssa_name (result, NULL);
1693 else
1694 result = make_ssa_name (TREE_TYPE (result));
1696 /* Emit the statement to compute min/max. */
1697 new_stmt = gimple_build_assign (result, minmax, arg0, arg1);
1698 gsi = gsi_last_bb (cond_bb);
1699 gsi_insert_before (&gsi, new_stmt, GSI_NEW_STMT);
1701 replace_phi_edge_with_variable (cond_bb, e1, phi, result);
1703 return true;
1706 /* Convert
1708 <bb 2>
1709 if (b_4(D) != 0)
1710 goto <bb 3>
1711 else
1712 goto <bb 4>
1714 <bb 3>
1715 _2 = (unsigned long) b_4(D);
1716 _9 = __builtin_popcountl (_2);
1718 _9 = __builtin_popcountl (b_4(D));
1720 <bb 4>
1721 c_12 = PHI <0(2), _9(3)>
1723 Into
1724 <bb 2>
1725 _2 = (unsigned long) b_4(D);
1726 _9 = __builtin_popcountl (_2);
1728 _9 = __builtin_popcountl (b_4(D));
1730 <bb 4>
1731 c_12 = PHI <_9(2)>
1734 static bool
1735 cond_removal_in_popcount_pattern (basic_block cond_bb, basic_block middle_bb,
1736 edge e1, edge e2,
1737 gimple *phi, tree arg0, tree arg1)
1739 gimple *cond;
1740 gimple_stmt_iterator gsi, gsi_from;
1741 gimple *popcount;
1742 gimple *cast = NULL;
1743 tree lhs, arg;
1745 /* Check that
1746 _2 = (unsigned long) b_4(D);
1747 _9 = __builtin_popcountl (_2);
1749 _9 = __builtin_popcountl (b_4(D));
1750 are the only stmts in the middle_bb. */
1752 gsi = gsi_start_nondebug_after_labels_bb (middle_bb);
1753 if (gsi_end_p (gsi))
1754 return false;
1755 cast = gsi_stmt (gsi);
1756 gsi_next_nondebug (&gsi);
1757 if (!gsi_end_p (gsi))
1759 popcount = gsi_stmt (gsi);
1760 gsi_next_nondebug (&gsi);
1761 if (!gsi_end_p (gsi))
1762 return false;
1764 else
1766 popcount = cast;
1767 cast = NULL;
1770 /* Check that we have a popcount builtin. */
1771 if (!is_gimple_call (popcount))
1772 return false;
1773 combined_fn cfn = gimple_call_combined_fn (popcount);
1774 switch (cfn)
1776 CASE_CFN_POPCOUNT:
1777 break;
1778 default:
1779 return false;
1782 arg = gimple_call_arg (popcount, 0);
1783 lhs = gimple_get_lhs (popcount);
1785 if (cast)
1787 /* We have a cast stmt feeding popcount builtin. */
1788 /* Check that we have a cast prior to that. */
1789 if (gimple_code (cast) != GIMPLE_ASSIGN
1790 || !CONVERT_EXPR_CODE_P (gimple_assign_rhs_code (cast)))
1791 return false;
1792 /* Result of the cast stmt is the argument to the builtin. */
1793 if (arg != gimple_assign_lhs (cast))
1794 return false;
1795 arg = gimple_assign_rhs1 (cast);
1798 cond = last_stmt (cond_bb);
1800 /* Cond_bb has a check for b_4 [!=|==] 0 before calling the popcount
1801 builtin. */
1802 if (gimple_code (cond) != GIMPLE_COND
1803 || (gimple_cond_code (cond) != NE_EXPR
1804 && gimple_cond_code (cond) != EQ_EXPR)
1805 || !integer_zerop (gimple_cond_rhs (cond))
1806 || arg != gimple_cond_lhs (cond))
1807 return false;
1809 /* Canonicalize. */
1810 if ((e2->flags & EDGE_TRUE_VALUE
1811 && gimple_cond_code (cond) == NE_EXPR)
1812 || (e1->flags & EDGE_TRUE_VALUE
1813 && gimple_cond_code (cond) == EQ_EXPR))
1815 std::swap (arg0, arg1);
1816 std::swap (e1, e2);
1819 /* Check PHI arguments. */
1820 if (lhs != arg0 || !integer_zerop (arg1))
1821 return false;
1823 /* And insert the popcount builtin and cast stmt before the cond_bb. */
1824 gsi = gsi_last_bb (cond_bb);
1825 if (cast)
1827 gsi_from = gsi_for_stmt (cast);
1828 gsi_move_before (&gsi_from, &gsi);
1829 reset_flow_sensitive_info (gimple_get_lhs (cast));
1831 gsi_from = gsi_for_stmt (popcount);
1832 gsi_move_before (&gsi_from, &gsi);
1833 reset_flow_sensitive_info (gimple_get_lhs (popcount));
1835 /* Now update the PHI and remove unneeded bbs. */
1836 replace_phi_edge_with_variable (cond_bb, e2, phi, lhs);
1837 return true;
1840 /* The function absolute_replacement does the main work of doing the absolute
1841 replacement. Return true if the replacement is done. Otherwise return
1842 false.
1843 bb is the basic block where the replacement is going to be done on. arg0
1844 is argument 0 from the phi. Likewise for arg1. */
1846 static bool
1847 abs_replacement (basic_block cond_bb, basic_block middle_bb,
1848 edge e0 ATTRIBUTE_UNUSED, edge e1,
1849 gimple *phi, tree arg0, tree arg1)
1851 tree result;
1852 gassign *new_stmt;
1853 gimple *cond;
1854 gimple_stmt_iterator gsi;
1855 edge true_edge, false_edge;
1856 gimple *assign;
1857 edge e;
1858 tree rhs, lhs;
1859 bool negate;
1860 enum tree_code cond_code;
1862 /* If the type says honor signed zeros we cannot do this
1863 optimization. */
1864 if (HONOR_SIGNED_ZEROS (arg1))
1865 return false;
1867 /* OTHER_BLOCK must have only one executable statement which must have the
1868 form arg0 = -arg1 or arg1 = -arg0. */
1870 assign = last_and_only_stmt (middle_bb);
1871 /* If we did not find the proper negation assignment, then we cannot
1872 optimize. */
1873 if (assign == NULL)
1874 return false;
1876 /* If we got here, then we have found the only executable statement
1877 in OTHER_BLOCK. If it is anything other than arg = -arg1 or
1878 arg1 = -arg0, then we cannot optimize. */
1879 if (gimple_code (assign) != GIMPLE_ASSIGN)
1880 return false;
1882 lhs = gimple_assign_lhs (assign);
1884 if (gimple_assign_rhs_code (assign) != NEGATE_EXPR)
1885 return false;
1887 rhs = gimple_assign_rhs1 (assign);
1889 /* The assignment has to be arg0 = -arg1 or arg1 = -arg0. */
1890 if (!(lhs == arg0 && rhs == arg1)
1891 && !(lhs == arg1 && rhs == arg0))
1892 return false;
1894 cond = last_stmt (cond_bb);
1895 result = PHI_RESULT (phi);
1897 /* Only relationals comparing arg[01] against zero are interesting. */
1898 cond_code = gimple_cond_code (cond);
1899 if (cond_code != GT_EXPR && cond_code != GE_EXPR
1900 && cond_code != LT_EXPR && cond_code != LE_EXPR)
1901 return false;
1903 /* Make sure the conditional is arg[01] OP y. */
1904 if (gimple_cond_lhs (cond) != rhs)
1905 return false;
1907 if (FLOAT_TYPE_P (TREE_TYPE (gimple_cond_rhs (cond)))
1908 ? real_zerop (gimple_cond_rhs (cond))
1909 : integer_zerop (gimple_cond_rhs (cond)))
1911 else
1912 return false;
1914 /* We need to know which is the true edge and which is the false
1915 edge so that we know if have abs or negative abs. */
1916 extract_true_false_edges_from_block (cond_bb, &true_edge, &false_edge);
1918 /* For GT_EXPR/GE_EXPR, if the true edge goes to OTHER_BLOCK, then we
1919 will need to negate the result. Similarly for LT_EXPR/LE_EXPR if
1920 the false edge goes to OTHER_BLOCK. */
1921 if (cond_code == GT_EXPR || cond_code == GE_EXPR)
1922 e = true_edge;
1923 else
1924 e = false_edge;
1926 if (e->dest == middle_bb)
1927 negate = true;
1928 else
1929 negate = false;
1931 /* If the code negates only iff positive then make sure to not
1932 introduce undefined behavior when negating or computing the absolute.
1933 ??? We could use range info if present to check for arg1 == INT_MIN. */
1934 if (negate
1935 && (ANY_INTEGRAL_TYPE_P (TREE_TYPE (arg1))
1936 && ! TYPE_OVERFLOW_WRAPS (TREE_TYPE (arg1))))
1937 return false;
1939 result = duplicate_ssa_name (result, NULL);
1941 if (negate)
1942 lhs = make_ssa_name (TREE_TYPE (result));
1943 else
1944 lhs = result;
1946 /* Build the modify expression with abs expression. */
1947 new_stmt = gimple_build_assign (lhs, ABS_EXPR, rhs);
1949 gsi = gsi_last_bb (cond_bb);
1950 gsi_insert_before (&gsi, new_stmt, GSI_NEW_STMT);
1952 if (negate)
1954 /* Get the right GSI. We want to insert after the recently
1955 added ABS_EXPR statement (which we know is the first statement
1956 in the block. */
1957 new_stmt = gimple_build_assign (result, NEGATE_EXPR, lhs);
1959 gsi_insert_after (&gsi, new_stmt, GSI_NEW_STMT);
1962 replace_phi_edge_with_variable (cond_bb, e1, phi, result);
1964 /* Note that we optimized this PHI. */
1965 return true;
1968 /* Auxiliary functions to determine the set of memory accesses which
1969 can't trap because they are preceded by accesses to the same memory
1970 portion. We do that for MEM_REFs, so we only need to track
1971 the SSA_NAME of the pointer indirectly referenced. The algorithm
1972 simply is a walk over all instructions in dominator order. When
1973 we see an MEM_REF we determine if we've already seen a same
1974 ref anywhere up to the root of the dominator tree. If we do the
1975 current access can't trap. If we don't see any dominating access
1976 the current access might trap, but might also make later accesses
1977 non-trapping, so we remember it. We need to be careful with loads
1978 or stores, for instance a load might not trap, while a store would,
1979 so if we see a dominating read access this doesn't mean that a later
1980 write access would not trap. Hence we also need to differentiate the
1981 type of access(es) seen.
1983 ??? We currently are very conservative and assume that a load might
1984 trap even if a store doesn't (write-only memory). This probably is
1985 overly conservative. */
1987 /* A hash-table of SSA_NAMEs, and in which basic block an MEM_REF
1988 through it was seen, which would constitute a no-trap region for
1989 same accesses. */
1990 struct name_to_bb
1992 unsigned int ssa_name_ver;
1993 unsigned int phase;
1994 bool store;
1995 HOST_WIDE_INT offset, size;
1996 basic_block bb;
1999 /* Hashtable helpers. */
2001 struct ssa_names_hasher : free_ptr_hash <name_to_bb>
2003 static inline hashval_t hash (const name_to_bb *);
2004 static inline bool equal (const name_to_bb *, const name_to_bb *);
2007 /* Used for quick clearing of the hash-table when we see calls.
2008 Hash entries with phase < nt_call_phase are invalid. */
2009 static unsigned int nt_call_phase;
2011 /* The hash function. */
2013 inline hashval_t
2014 ssa_names_hasher::hash (const name_to_bb *n)
2016 return n->ssa_name_ver ^ (((hashval_t) n->store) << 31)
2017 ^ (n->offset << 6) ^ (n->size << 3);
2020 /* The equality function of *P1 and *P2. */
2022 inline bool
2023 ssa_names_hasher::equal (const name_to_bb *n1, const name_to_bb *n2)
2025 return n1->ssa_name_ver == n2->ssa_name_ver
2026 && n1->store == n2->store
2027 && n1->offset == n2->offset
2028 && n1->size == n2->size;
2031 class nontrapping_dom_walker : public dom_walker
2033 public:
2034 nontrapping_dom_walker (cdi_direction direction, hash_set<tree> *ps)
2035 : dom_walker (direction), m_nontrapping (ps), m_seen_ssa_names (128) {}
2037 virtual edge before_dom_children (basic_block);
2038 virtual void after_dom_children (basic_block);
2040 private:
2042 /* We see the expression EXP in basic block BB. If it's an interesting
2043 expression (an MEM_REF through an SSA_NAME) possibly insert the
2044 expression into the set NONTRAP or the hash table of seen expressions.
2045 STORE is true if this expression is on the LHS, otherwise it's on
2046 the RHS. */
2047 void add_or_mark_expr (basic_block, tree, bool);
2049 hash_set<tree> *m_nontrapping;
2051 /* The hash table for remembering what we've seen. */
2052 hash_table<ssa_names_hasher> m_seen_ssa_names;
2055 /* Called by walk_dominator_tree, when entering the block BB. */
2056 edge
2057 nontrapping_dom_walker::before_dom_children (basic_block bb)
2059 edge e;
2060 edge_iterator ei;
2061 gimple_stmt_iterator gsi;
2063 /* If we haven't seen all our predecessors, clear the hash-table. */
2064 FOR_EACH_EDGE (e, ei, bb->preds)
2065 if ((((size_t)e->src->aux) & 2) == 0)
2067 nt_call_phase++;
2068 break;
2071 /* Mark this BB as being on the path to dominator root and as visited. */
2072 bb->aux = (void*)(1 | 2);
2074 /* And walk the statements in order. */
2075 for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
2077 gimple *stmt = gsi_stmt (gsi);
2079 if ((gimple_code (stmt) == GIMPLE_ASM && gimple_vdef (stmt))
2080 || (is_gimple_call (stmt)
2081 && (!nonfreeing_call_p (stmt) || !nonbarrier_call_p (stmt))))
2082 nt_call_phase++;
2083 else if (gimple_assign_single_p (stmt) && !gimple_has_volatile_ops (stmt))
2085 add_or_mark_expr (bb, gimple_assign_lhs (stmt), true);
2086 add_or_mark_expr (bb, gimple_assign_rhs1 (stmt), false);
2089 return NULL;
2092 /* Called by walk_dominator_tree, when basic block BB is exited. */
2093 void
2094 nontrapping_dom_walker::after_dom_children (basic_block bb)
2096 /* This BB isn't on the path to dominator root anymore. */
2097 bb->aux = (void*)2;
2100 /* We see the expression EXP in basic block BB. If it's an interesting
2101 expression (an MEM_REF through an SSA_NAME) possibly insert the
2102 expression into the set NONTRAP or the hash table of seen expressions.
2103 STORE is true if this expression is on the LHS, otherwise it's on
2104 the RHS. */
2105 void
2106 nontrapping_dom_walker::add_or_mark_expr (basic_block bb, tree exp, bool store)
2108 HOST_WIDE_INT size;
2110 if (TREE_CODE (exp) == MEM_REF
2111 && TREE_CODE (TREE_OPERAND (exp, 0)) == SSA_NAME
2112 && tree_fits_shwi_p (TREE_OPERAND (exp, 1))
2113 && (size = int_size_in_bytes (TREE_TYPE (exp))) > 0)
2115 tree name = TREE_OPERAND (exp, 0);
2116 struct name_to_bb map;
2117 name_to_bb **slot;
2118 struct name_to_bb *n2bb;
2119 basic_block found_bb = 0;
2121 /* Try to find the last seen MEM_REF through the same
2122 SSA_NAME, which can trap. */
2123 map.ssa_name_ver = SSA_NAME_VERSION (name);
2124 map.phase = 0;
2125 map.bb = 0;
2126 map.store = store;
2127 map.offset = tree_to_shwi (TREE_OPERAND (exp, 1));
2128 map.size = size;
2130 slot = m_seen_ssa_names.find_slot (&map, INSERT);
2131 n2bb = *slot;
2132 if (n2bb && n2bb->phase >= nt_call_phase)
2133 found_bb = n2bb->bb;
2135 /* If we've found a trapping MEM_REF, _and_ it dominates EXP
2136 (it's in a basic block on the path from us to the dominator root)
2137 then we can't trap. */
2138 if (found_bb && (((size_t)found_bb->aux) & 1) == 1)
2140 m_nontrapping->add (exp);
2142 else
2144 /* EXP might trap, so insert it into the hash table. */
2145 if (n2bb)
2147 n2bb->phase = nt_call_phase;
2148 n2bb->bb = bb;
2150 else
2152 n2bb = XNEW (struct name_to_bb);
2153 n2bb->ssa_name_ver = SSA_NAME_VERSION (name);
2154 n2bb->phase = nt_call_phase;
2155 n2bb->bb = bb;
2156 n2bb->store = store;
2157 n2bb->offset = map.offset;
2158 n2bb->size = size;
2159 *slot = n2bb;
2165 /* This is the entry point of gathering non trapping memory accesses.
2166 It will do a dominator walk over the whole function, and it will
2167 make use of the bb->aux pointers. It returns a set of trees
2168 (the MEM_REFs itself) which can't trap. */
2169 static hash_set<tree> *
2170 get_non_trapping (void)
2172 nt_call_phase = 0;
2173 hash_set<tree> *nontrap = new hash_set<tree>;
2174 /* We're going to do a dominator walk, so ensure that we have
2175 dominance information. */
2176 calculate_dominance_info (CDI_DOMINATORS);
2178 nontrapping_dom_walker (CDI_DOMINATORS, nontrap)
2179 .walk (cfun->cfg->x_entry_block_ptr);
2181 clear_aux_for_blocks ();
2182 return nontrap;
2185 /* Do the main work of conditional store replacement. We already know
2186 that the recognized pattern looks like so:
2188 split:
2189 if (cond) goto MIDDLE_BB; else goto JOIN_BB (edge E1)
2190 MIDDLE_BB:
2191 something
2192 fallthrough (edge E0)
2193 JOIN_BB:
2194 some more
2196 We check that MIDDLE_BB contains only one store, that that store
2197 doesn't trap (not via NOTRAP, but via checking if an access to the same
2198 memory location dominates us, or the store is to a local addressable
2199 object) and that the store has a "simple" RHS. */
2201 static bool
2202 cond_store_replacement (basic_block middle_bb, basic_block join_bb,
2203 edge e0, edge e1, hash_set<tree> *nontrap)
2205 gimple *assign = last_and_only_stmt (middle_bb);
2206 tree lhs, rhs, name, name2;
2207 gphi *newphi;
2208 gassign *new_stmt;
2209 gimple_stmt_iterator gsi;
2210 location_t locus;
2212 /* Check if middle_bb contains of only one store. */
2213 if (!assign
2214 || !gimple_assign_single_p (assign)
2215 || gimple_has_volatile_ops (assign))
2216 return false;
2218 /* And no PHI nodes so all uses in the single stmt are also
2219 available where we insert to. */
2220 if (!gimple_seq_empty_p (phi_nodes (middle_bb)))
2221 return false;
2223 locus = gimple_location (assign);
2224 lhs = gimple_assign_lhs (assign);
2225 rhs = gimple_assign_rhs1 (assign);
2226 if ((TREE_CODE (lhs) != MEM_REF
2227 && TREE_CODE (lhs) != ARRAY_REF
2228 && TREE_CODE (lhs) != COMPONENT_REF)
2229 || !is_gimple_reg_type (TREE_TYPE (lhs)))
2230 return false;
2232 /* Prove that we can move the store down. We could also check
2233 TREE_THIS_NOTRAP here, but in that case we also could move stores,
2234 whose value is not available readily, which we want to avoid. */
2235 if (!nontrap->contains (lhs))
2237 /* If LHS is a local variable without address-taken, we could
2238 always safely move down the store. */
2239 tree base = get_base_address (lhs);
2240 if (!auto_var_p (base) || TREE_ADDRESSABLE (base))
2241 return false;
2244 /* Now we've checked the constraints, so do the transformation:
2245 1) Remove the single store. */
2246 gsi = gsi_for_stmt (assign);
2247 unlink_stmt_vdef (assign);
2248 gsi_remove (&gsi, true);
2249 release_defs (assign);
2251 /* Make both store and load use alias-set zero as we have to
2252 deal with the case of the store being a conditional change
2253 of the dynamic type. */
2254 lhs = unshare_expr (lhs);
2255 tree *basep = &lhs;
2256 while (handled_component_p (*basep))
2257 basep = &TREE_OPERAND (*basep, 0);
2258 if (TREE_CODE (*basep) == MEM_REF
2259 || TREE_CODE (*basep) == TARGET_MEM_REF)
2260 TREE_OPERAND (*basep, 1)
2261 = fold_convert (ptr_type_node, TREE_OPERAND (*basep, 1));
2262 else
2263 *basep = build2 (MEM_REF, TREE_TYPE (*basep),
2264 build_fold_addr_expr (*basep),
2265 build_zero_cst (ptr_type_node));
2267 /* 2) Insert a load from the memory of the store to the temporary
2268 on the edge which did not contain the store. */
2269 name = make_temp_ssa_name (TREE_TYPE (lhs), NULL, "cstore");
2270 new_stmt = gimple_build_assign (name, lhs);
2271 gimple_set_location (new_stmt, locus);
2272 gsi_insert_on_edge (e1, new_stmt);
2274 /* 3) Create a PHI node at the join block, with one argument
2275 holding the old RHS, and the other holding the temporary
2276 where we stored the old memory contents. */
2277 name2 = make_temp_ssa_name (TREE_TYPE (lhs), NULL, "cstore");
2278 newphi = create_phi_node (name2, join_bb);
2279 add_phi_arg (newphi, rhs, e0, locus);
2280 add_phi_arg (newphi, name, e1, locus);
2282 lhs = unshare_expr (lhs);
2283 new_stmt = gimple_build_assign (lhs, PHI_RESULT (newphi));
2285 /* 4) Insert that PHI node. */
2286 gsi = gsi_after_labels (join_bb);
2287 if (gsi_end_p (gsi))
2289 gsi = gsi_last_bb (join_bb);
2290 gsi_insert_after (&gsi, new_stmt, GSI_NEW_STMT);
2292 else
2293 gsi_insert_before (&gsi, new_stmt, GSI_NEW_STMT);
2295 if (dump_file && (dump_flags & TDF_DETAILS))
2297 fprintf (dump_file, "\nConditional store replacement happened!");
2298 fprintf (dump_file, "\nReplaced the store with a load.");
2299 fprintf (dump_file, "\nInserted a new PHI statement in joint block:\n");
2300 print_gimple_stmt (dump_file, new_stmt, 0, TDF_VOPS|TDF_MEMSYMS);
2303 return true;
2306 /* Do the main work of conditional store replacement. */
2308 static bool
2309 cond_if_else_store_replacement_1 (basic_block then_bb, basic_block else_bb,
2310 basic_block join_bb, gimple *then_assign,
2311 gimple *else_assign)
2313 tree lhs_base, lhs, then_rhs, else_rhs, name;
2314 location_t then_locus, else_locus;
2315 gimple_stmt_iterator gsi;
2316 gphi *newphi;
2317 gassign *new_stmt;
2319 if (then_assign == NULL
2320 || !gimple_assign_single_p (then_assign)
2321 || gimple_clobber_p (then_assign)
2322 || gimple_has_volatile_ops (then_assign)
2323 || else_assign == NULL
2324 || !gimple_assign_single_p (else_assign)
2325 || gimple_clobber_p (else_assign)
2326 || gimple_has_volatile_ops (else_assign))
2327 return false;
2329 lhs = gimple_assign_lhs (then_assign);
2330 if (!is_gimple_reg_type (TREE_TYPE (lhs))
2331 || !operand_equal_p (lhs, gimple_assign_lhs (else_assign), 0))
2332 return false;
2334 lhs_base = get_base_address (lhs);
2335 if (lhs_base == NULL_TREE
2336 || (!DECL_P (lhs_base) && TREE_CODE (lhs_base) != MEM_REF))
2337 return false;
2339 then_rhs = gimple_assign_rhs1 (then_assign);
2340 else_rhs = gimple_assign_rhs1 (else_assign);
2341 then_locus = gimple_location (then_assign);
2342 else_locus = gimple_location (else_assign);
2344 /* Now we've checked the constraints, so do the transformation:
2345 1) Remove the stores. */
2346 gsi = gsi_for_stmt (then_assign);
2347 unlink_stmt_vdef (then_assign);
2348 gsi_remove (&gsi, true);
2349 release_defs (then_assign);
2351 gsi = gsi_for_stmt (else_assign);
2352 unlink_stmt_vdef (else_assign);
2353 gsi_remove (&gsi, true);
2354 release_defs (else_assign);
2356 /* 2) Create a PHI node at the join block, with one argument
2357 holding the old RHS, and the other holding the temporary
2358 where we stored the old memory contents. */
2359 name = make_temp_ssa_name (TREE_TYPE (lhs), NULL, "cstore");
2360 newphi = create_phi_node (name, join_bb);
2361 add_phi_arg (newphi, then_rhs, EDGE_SUCC (then_bb, 0), then_locus);
2362 add_phi_arg (newphi, else_rhs, EDGE_SUCC (else_bb, 0), else_locus);
2364 new_stmt = gimple_build_assign (lhs, PHI_RESULT (newphi));
2366 /* 3) Insert that PHI node. */
2367 gsi = gsi_after_labels (join_bb);
2368 if (gsi_end_p (gsi))
2370 gsi = gsi_last_bb (join_bb);
2371 gsi_insert_after (&gsi, new_stmt, GSI_NEW_STMT);
2373 else
2374 gsi_insert_before (&gsi, new_stmt, GSI_NEW_STMT);
2376 return true;
2379 /* Return the single store in BB with VDEF or NULL if there are
2380 other stores in the BB or loads following the store. */
2382 static gimple *
2383 single_trailing_store_in_bb (basic_block bb, tree vdef)
2385 if (SSA_NAME_IS_DEFAULT_DEF (vdef))
2386 return NULL;
2387 gimple *store = SSA_NAME_DEF_STMT (vdef);
2388 if (gimple_bb (store) != bb
2389 || gimple_code (store) == GIMPLE_PHI)
2390 return NULL;
2392 /* Verify there is no other store in this BB. */
2393 if (!SSA_NAME_IS_DEFAULT_DEF (gimple_vuse (store))
2394 && gimple_bb (SSA_NAME_DEF_STMT (gimple_vuse (store))) == bb
2395 && gimple_code (SSA_NAME_DEF_STMT (gimple_vuse (store))) != GIMPLE_PHI)
2396 return NULL;
2398 /* Verify there is no load or store after the store. */
2399 use_operand_p use_p;
2400 imm_use_iterator imm_iter;
2401 FOR_EACH_IMM_USE_FAST (use_p, imm_iter, gimple_vdef (store))
2402 if (USE_STMT (use_p) != store
2403 && gimple_bb (USE_STMT (use_p)) == bb)
2404 return NULL;
2406 return store;
2409 /* Conditional store replacement. We already know
2410 that the recognized pattern looks like so:
2412 split:
2413 if (cond) goto THEN_BB; else goto ELSE_BB (edge E1)
2414 THEN_BB:
2416 X = Y;
2418 goto JOIN_BB;
2419 ELSE_BB:
2421 X = Z;
2423 fallthrough (edge E0)
2424 JOIN_BB:
2425 some more
2427 We check that it is safe to sink the store to JOIN_BB by verifying that
2428 there are no read-after-write or write-after-write dependencies in
2429 THEN_BB and ELSE_BB. */
2431 static bool
2432 cond_if_else_store_replacement (basic_block then_bb, basic_block else_bb,
2433 basic_block join_bb)
2435 vec<data_reference_p> then_datarefs, else_datarefs;
2436 vec<ddr_p> then_ddrs, else_ddrs;
2437 gimple *then_store, *else_store;
2438 bool found, ok = false, res;
2439 struct data_dependence_relation *ddr;
2440 data_reference_p then_dr, else_dr;
2441 int i, j;
2442 tree then_lhs, else_lhs;
2443 basic_block blocks[3];
2445 /* Handle the case with single store in THEN_BB and ELSE_BB. That is
2446 cheap enough to always handle as it allows us to elide dependence
2447 checking. */
2448 gphi *vphi = NULL;
2449 for (gphi_iterator si = gsi_start_phis (join_bb); !gsi_end_p (si);
2450 gsi_next (&si))
2451 if (virtual_operand_p (gimple_phi_result (si.phi ())))
2453 vphi = si.phi ();
2454 break;
2456 if (!vphi)
2457 return false;
2458 tree then_vdef = PHI_ARG_DEF_FROM_EDGE (vphi, single_succ_edge (then_bb));
2459 tree else_vdef = PHI_ARG_DEF_FROM_EDGE (vphi, single_succ_edge (else_bb));
2460 gimple *then_assign = single_trailing_store_in_bb (then_bb, then_vdef);
2461 if (then_assign)
2463 gimple *else_assign = single_trailing_store_in_bb (else_bb, else_vdef);
2464 if (else_assign)
2465 return cond_if_else_store_replacement_1 (then_bb, else_bb, join_bb,
2466 then_assign, else_assign);
2469 /* If either vectorization or if-conversion is disabled then do
2470 not sink any stores. */
2471 if (param_max_stores_to_sink == 0
2472 || (!flag_tree_loop_vectorize && !flag_tree_slp_vectorize)
2473 || !flag_tree_loop_if_convert)
2474 return false;
2476 /* Find data references. */
2477 then_datarefs.create (1);
2478 else_datarefs.create (1);
2479 if ((find_data_references_in_bb (NULL, then_bb, &then_datarefs)
2480 == chrec_dont_know)
2481 || !then_datarefs.length ()
2482 || (find_data_references_in_bb (NULL, else_bb, &else_datarefs)
2483 == chrec_dont_know)
2484 || !else_datarefs.length ())
2486 free_data_refs (then_datarefs);
2487 free_data_refs (else_datarefs);
2488 return false;
2491 /* Find pairs of stores with equal LHS. */
2492 auto_vec<gimple *, 1> then_stores, else_stores;
2493 FOR_EACH_VEC_ELT (then_datarefs, i, then_dr)
2495 if (DR_IS_READ (then_dr))
2496 continue;
2498 then_store = DR_STMT (then_dr);
2499 then_lhs = gimple_get_lhs (then_store);
2500 if (then_lhs == NULL_TREE)
2501 continue;
2502 found = false;
2504 FOR_EACH_VEC_ELT (else_datarefs, j, else_dr)
2506 if (DR_IS_READ (else_dr))
2507 continue;
2509 else_store = DR_STMT (else_dr);
2510 else_lhs = gimple_get_lhs (else_store);
2511 if (else_lhs == NULL_TREE)
2512 continue;
2514 if (operand_equal_p (then_lhs, else_lhs, 0))
2516 found = true;
2517 break;
2521 if (!found)
2522 continue;
2524 then_stores.safe_push (then_store);
2525 else_stores.safe_push (else_store);
2528 /* No pairs of stores found. */
2529 if (!then_stores.length ()
2530 || then_stores.length () > (unsigned) param_max_stores_to_sink)
2532 free_data_refs (then_datarefs);
2533 free_data_refs (else_datarefs);
2534 return false;
2537 /* Compute and check data dependencies in both basic blocks. */
2538 then_ddrs.create (1);
2539 else_ddrs.create (1);
2540 if (!compute_all_dependences (then_datarefs, &then_ddrs,
2541 vNULL, false)
2542 || !compute_all_dependences (else_datarefs, &else_ddrs,
2543 vNULL, false))
2545 free_dependence_relations (then_ddrs);
2546 free_dependence_relations (else_ddrs);
2547 free_data_refs (then_datarefs);
2548 free_data_refs (else_datarefs);
2549 return false;
2551 blocks[0] = then_bb;
2552 blocks[1] = else_bb;
2553 blocks[2] = join_bb;
2554 renumber_gimple_stmt_uids_in_blocks (blocks, 3);
2556 /* Check that there are no read-after-write or write-after-write dependencies
2557 in THEN_BB. */
2558 FOR_EACH_VEC_ELT (then_ddrs, i, ddr)
2560 struct data_reference *dra = DDR_A (ddr);
2561 struct data_reference *drb = DDR_B (ddr);
2563 if (DDR_ARE_DEPENDENT (ddr) != chrec_known
2564 && ((DR_IS_READ (dra) && DR_IS_WRITE (drb)
2565 && gimple_uid (DR_STMT (dra)) > gimple_uid (DR_STMT (drb)))
2566 || (DR_IS_READ (drb) && DR_IS_WRITE (dra)
2567 && gimple_uid (DR_STMT (drb)) > gimple_uid (DR_STMT (dra)))
2568 || (DR_IS_WRITE (dra) && DR_IS_WRITE (drb))))
2570 free_dependence_relations (then_ddrs);
2571 free_dependence_relations (else_ddrs);
2572 free_data_refs (then_datarefs);
2573 free_data_refs (else_datarefs);
2574 return false;
2578 /* Check that there are no read-after-write or write-after-write dependencies
2579 in ELSE_BB. */
2580 FOR_EACH_VEC_ELT (else_ddrs, i, ddr)
2582 struct data_reference *dra = DDR_A (ddr);
2583 struct data_reference *drb = DDR_B (ddr);
2585 if (DDR_ARE_DEPENDENT (ddr) != chrec_known
2586 && ((DR_IS_READ (dra) && DR_IS_WRITE (drb)
2587 && gimple_uid (DR_STMT (dra)) > gimple_uid (DR_STMT (drb)))
2588 || (DR_IS_READ (drb) && DR_IS_WRITE (dra)
2589 && gimple_uid (DR_STMT (drb)) > gimple_uid (DR_STMT (dra)))
2590 || (DR_IS_WRITE (dra) && DR_IS_WRITE (drb))))
2592 free_dependence_relations (then_ddrs);
2593 free_dependence_relations (else_ddrs);
2594 free_data_refs (then_datarefs);
2595 free_data_refs (else_datarefs);
2596 return false;
2600 /* Sink stores with same LHS. */
2601 FOR_EACH_VEC_ELT (then_stores, i, then_store)
2603 else_store = else_stores[i];
2604 res = cond_if_else_store_replacement_1 (then_bb, else_bb, join_bb,
2605 then_store, else_store);
2606 ok = ok || res;
2609 free_dependence_relations (then_ddrs);
2610 free_dependence_relations (else_ddrs);
2611 free_data_refs (then_datarefs);
2612 free_data_refs (else_datarefs);
2614 return ok;
2617 /* Return TRUE if STMT has a VUSE whose corresponding VDEF is in BB. */
2619 static bool
2620 local_mem_dependence (gimple *stmt, basic_block bb)
2622 tree vuse = gimple_vuse (stmt);
2623 gimple *def;
2625 if (!vuse)
2626 return false;
2628 def = SSA_NAME_DEF_STMT (vuse);
2629 return (def && gimple_bb (def) == bb);
2632 /* Given a "diamond" control-flow pattern where BB0 tests a condition,
2633 BB1 and BB2 are "then" and "else" blocks dependent on this test,
2634 and BB3 rejoins control flow following BB1 and BB2, look for
2635 opportunities to hoist loads as follows. If BB3 contains a PHI of
2636 two loads, one each occurring in BB1 and BB2, and the loads are
2637 provably of adjacent fields in the same structure, then move both
2638 loads into BB0. Of course this can only be done if there are no
2639 dependencies preventing such motion.
2641 One of the hoisted loads will always be speculative, so the
2642 transformation is currently conservative:
2644 - The fields must be strictly adjacent.
2645 - The two fields must occupy a single memory block that is
2646 guaranteed to not cross a page boundary.
2648 The last is difficult to prove, as such memory blocks should be
2649 aligned on the minimum of the stack alignment boundary and the
2650 alignment guaranteed by heap allocation interfaces. Thus we rely
2651 on a parameter for the alignment value.
2653 Provided a good value is used for the last case, the first
2654 restriction could possibly be relaxed. */
2656 static void
2657 hoist_adjacent_loads (basic_block bb0, basic_block bb1,
2658 basic_block bb2, basic_block bb3)
2660 int param_align = param_l1_cache_line_size;
2661 unsigned param_align_bits = (unsigned) (param_align * BITS_PER_UNIT);
2662 gphi_iterator gsi;
2664 /* Walk the phis in bb3 looking for an opportunity. We are looking
2665 for phis of two SSA names, one each of which is defined in bb1 and
2666 bb2. */
2667 for (gsi = gsi_start_phis (bb3); !gsi_end_p (gsi); gsi_next (&gsi))
2669 gphi *phi_stmt = gsi.phi ();
2670 gimple *def1, *def2;
2671 tree arg1, arg2, ref1, ref2, field1, field2;
2672 tree tree_offset1, tree_offset2, tree_size2, next;
2673 int offset1, offset2, size2;
2674 unsigned align1;
2675 gimple_stmt_iterator gsi2;
2676 basic_block bb_for_def1, bb_for_def2;
2678 if (gimple_phi_num_args (phi_stmt) != 2
2679 || virtual_operand_p (gimple_phi_result (phi_stmt)))
2680 continue;
2682 arg1 = gimple_phi_arg_def (phi_stmt, 0);
2683 arg2 = gimple_phi_arg_def (phi_stmt, 1);
2685 if (TREE_CODE (arg1) != SSA_NAME
2686 || TREE_CODE (arg2) != SSA_NAME
2687 || SSA_NAME_IS_DEFAULT_DEF (arg1)
2688 || SSA_NAME_IS_DEFAULT_DEF (arg2))
2689 continue;
2691 def1 = SSA_NAME_DEF_STMT (arg1);
2692 def2 = SSA_NAME_DEF_STMT (arg2);
2694 if ((gimple_bb (def1) != bb1 || gimple_bb (def2) != bb2)
2695 && (gimple_bb (def2) != bb1 || gimple_bb (def1) != bb2))
2696 continue;
2698 /* Check the mode of the arguments to be sure a conditional move
2699 can be generated for it. */
2700 if (optab_handler (movcc_optab, TYPE_MODE (TREE_TYPE (arg1)))
2701 == CODE_FOR_nothing)
2702 continue;
2704 /* Both statements must be assignments whose RHS is a COMPONENT_REF. */
2705 if (!gimple_assign_single_p (def1)
2706 || !gimple_assign_single_p (def2)
2707 || gimple_has_volatile_ops (def1)
2708 || gimple_has_volatile_ops (def2))
2709 continue;
2711 ref1 = gimple_assign_rhs1 (def1);
2712 ref2 = gimple_assign_rhs1 (def2);
2714 if (TREE_CODE (ref1) != COMPONENT_REF
2715 || TREE_CODE (ref2) != COMPONENT_REF)
2716 continue;
2718 /* The zeroth operand of the two component references must be
2719 identical. It is not sufficient to compare get_base_address of
2720 the two references, because this could allow for different
2721 elements of the same array in the two trees. It is not safe to
2722 assume that the existence of one array element implies the
2723 existence of a different one. */
2724 if (!operand_equal_p (TREE_OPERAND (ref1, 0), TREE_OPERAND (ref2, 0), 0))
2725 continue;
2727 field1 = TREE_OPERAND (ref1, 1);
2728 field2 = TREE_OPERAND (ref2, 1);
2730 /* Check for field adjacency, and ensure field1 comes first. */
2731 for (next = DECL_CHAIN (field1);
2732 next && TREE_CODE (next) != FIELD_DECL;
2733 next = DECL_CHAIN (next))
2736 if (next != field2)
2738 for (next = DECL_CHAIN (field2);
2739 next && TREE_CODE (next) != FIELD_DECL;
2740 next = DECL_CHAIN (next))
2743 if (next != field1)
2744 continue;
2746 std::swap (field1, field2);
2747 std::swap (def1, def2);
2750 bb_for_def1 = gimple_bb (def1);
2751 bb_for_def2 = gimple_bb (def2);
2753 /* Check for proper alignment of the first field. */
2754 tree_offset1 = bit_position (field1);
2755 tree_offset2 = bit_position (field2);
2756 tree_size2 = DECL_SIZE (field2);
2758 if (!tree_fits_uhwi_p (tree_offset1)
2759 || !tree_fits_uhwi_p (tree_offset2)
2760 || !tree_fits_uhwi_p (tree_size2))
2761 continue;
2763 offset1 = tree_to_uhwi (tree_offset1);
2764 offset2 = tree_to_uhwi (tree_offset2);
2765 size2 = tree_to_uhwi (tree_size2);
2766 align1 = DECL_ALIGN (field1) % param_align_bits;
2768 if (offset1 % BITS_PER_UNIT != 0)
2769 continue;
2771 /* For profitability, the two field references should fit within
2772 a single cache line. */
2773 if (align1 + offset2 - offset1 + size2 > param_align_bits)
2774 continue;
2776 /* The two expressions cannot be dependent upon vdefs defined
2777 in bb1/bb2. */
2778 if (local_mem_dependence (def1, bb_for_def1)
2779 || local_mem_dependence (def2, bb_for_def2))
2780 continue;
2782 /* The conditions are satisfied; hoist the loads from bb1 and bb2 into
2783 bb0. We hoist the first one first so that a cache miss is handled
2784 efficiently regardless of hardware cache-fill policy. */
2785 gsi2 = gsi_for_stmt (def1);
2786 gsi_move_to_bb_end (&gsi2, bb0);
2787 gsi2 = gsi_for_stmt (def2);
2788 gsi_move_to_bb_end (&gsi2, bb0);
2790 if (dump_file && (dump_flags & TDF_DETAILS))
2792 fprintf (dump_file,
2793 "\nHoisting adjacent loads from %d and %d into %d: \n",
2794 bb_for_def1->index, bb_for_def2->index, bb0->index);
2795 print_gimple_stmt (dump_file, def1, 0, TDF_VOPS|TDF_MEMSYMS);
2796 print_gimple_stmt (dump_file, def2, 0, TDF_VOPS|TDF_MEMSYMS);
2801 /* Determine whether we should attempt to hoist adjacent loads out of
2802 diamond patterns in pass_phiopt. Always hoist loads if
2803 -fhoist-adjacent-loads is specified and the target machine has
2804 both a conditional move instruction and a defined cache line size. */
2806 static bool
2807 gate_hoist_loads (void)
2809 return (flag_hoist_adjacent_loads == 1
2810 && param_l1_cache_line_size
2811 && HAVE_conditional_move);
2814 /* This pass tries to replaces an if-then-else block with an
2815 assignment. We have four kinds of transformations. Some of these
2816 transformations are also performed by the ifcvt RTL optimizer.
2818 Conditional Replacement
2819 -----------------------
2821 This transformation, implemented in conditional_replacement,
2822 replaces
2824 bb0:
2825 if (cond) goto bb2; else goto bb1;
2826 bb1:
2827 bb2:
2828 x = PHI <0 (bb1), 1 (bb0), ...>;
2830 with
2832 bb0:
2833 x' = cond;
2834 goto bb2;
2835 bb2:
2836 x = PHI <x' (bb0), ...>;
2838 We remove bb1 as it becomes unreachable. This occurs often due to
2839 gimplification of conditionals.
2841 Value Replacement
2842 -----------------
2844 This transformation, implemented in value_replacement, replaces
2846 bb0:
2847 if (a != b) goto bb2; else goto bb1;
2848 bb1:
2849 bb2:
2850 x = PHI <a (bb1), b (bb0), ...>;
2852 with
2854 bb0:
2855 bb2:
2856 x = PHI <b (bb0), ...>;
2858 This opportunity can sometimes occur as a result of other
2859 optimizations.
2862 Another case caught by value replacement looks like this:
2864 bb0:
2865 t1 = a == CONST;
2866 t2 = b > c;
2867 t3 = t1 & t2;
2868 if (t3 != 0) goto bb1; else goto bb2;
2869 bb1:
2870 bb2:
2871 x = PHI (CONST, a)
2873 Gets replaced with:
2874 bb0:
2875 bb2:
2876 t1 = a == CONST;
2877 t2 = b > c;
2878 t3 = t1 & t2;
2879 x = a;
2881 ABS Replacement
2882 ---------------
2884 This transformation, implemented in abs_replacement, replaces
2886 bb0:
2887 if (a >= 0) goto bb2; else goto bb1;
2888 bb1:
2889 x = -a;
2890 bb2:
2891 x = PHI <x (bb1), a (bb0), ...>;
2893 with
2895 bb0:
2896 x' = ABS_EXPR< a >;
2897 bb2:
2898 x = PHI <x' (bb0), ...>;
2900 MIN/MAX Replacement
2901 -------------------
2903 This transformation, minmax_replacement replaces
2905 bb0:
2906 if (a <= b) goto bb2; else goto bb1;
2907 bb1:
2908 bb2:
2909 x = PHI <b (bb1), a (bb0), ...>;
2911 with
2913 bb0:
2914 x' = MIN_EXPR (a, b)
2915 bb2:
2916 x = PHI <x' (bb0), ...>;
2918 A similar transformation is done for MAX_EXPR.
2921 This pass also performs a fifth transformation of a slightly different
2922 flavor.
2924 Factor conversion in COND_EXPR
2925 ------------------------------
2927 This transformation factors the conversion out of COND_EXPR with
2928 factor_out_conditional_conversion.
2930 For example:
2931 if (a <= CST) goto <bb 3>; else goto <bb 4>;
2932 <bb 3>:
2933 tmp = (int) a;
2934 <bb 4>:
2935 tmp = PHI <tmp, CST>
2937 Into:
2938 if (a <= CST) goto <bb 3>; else goto <bb 4>;
2939 <bb 3>:
2940 <bb 4>:
2941 a = PHI <a, CST>
2942 tmp = (int) a;
2944 Adjacent Load Hoisting
2945 ----------------------
2947 This transformation replaces
2949 bb0:
2950 if (...) goto bb2; else goto bb1;
2951 bb1:
2952 x1 = (<expr>).field1;
2953 goto bb3;
2954 bb2:
2955 x2 = (<expr>).field2;
2956 bb3:
2957 # x = PHI <x1, x2>;
2959 with
2961 bb0:
2962 x1 = (<expr>).field1;
2963 x2 = (<expr>).field2;
2964 if (...) goto bb2; else goto bb1;
2965 bb1:
2966 goto bb3;
2967 bb2:
2968 bb3:
2969 # x = PHI <x1, x2>;
2971 The purpose of this transformation is to enable generation of conditional
2972 move instructions such as Intel CMOVE or PowerPC ISEL. Because one of
2973 the loads is speculative, the transformation is restricted to very
2974 specific cases to avoid introducing a page fault. We are looking for
2975 the common idiom:
2977 if (...)
2978 x = y->left;
2979 else
2980 x = y->right;
2982 where left and right are typically adjacent pointers in a tree structure. */
2984 namespace {
2986 const pass_data pass_data_phiopt =
2988 GIMPLE_PASS, /* type */
2989 "phiopt", /* name */
2990 OPTGROUP_NONE, /* optinfo_flags */
2991 TV_TREE_PHIOPT, /* tv_id */
2992 ( PROP_cfg | PROP_ssa ), /* properties_required */
2993 0, /* properties_provided */
2994 0, /* properties_destroyed */
2995 0, /* todo_flags_start */
2996 0, /* todo_flags_finish */
2999 class pass_phiopt : public gimple_opt_pass
3001 public:
3002 pass_phiopt (gcc::context *ctxt)
3003 : gimple_opt_pass (pass_data_phiopt, ctxt), early_p (false)
3006 /* opt_pass methods: */
3007 opt_pass * clone () { return new pass_phiopt (m_ctxt); }
3008 void set_pass_param (unsigned n, bool param)
3010 gcc_assert (n == 0);
3011 early_p = param;
3013 virtual bool gate (function *) { return flag_ssa_phiopt; }
3014 virtual unsigned int execute (function *)
3016 return tree_ssa_phiopt_worker (false,
3017 !early_p ? gate_hoist_loads () : false,
3018 early_p);
3021 private:
3022 bool early_p;
3023 }; // class pass_phiopt
3025 } // anon namespace
3027 gimple_opt_pass *
3028 make_pass_phiopt (gcc::context *ctxt)
3030 return new pass_phiopt (ctxt);
3033 namespace {
3035 const pass_data pass_data_cselim =
3037 GIMPLE_PASS, /* type */
3038 "cselim", /* name */
3039 OPTGROUP_NONE, /* optinfo_flags */
3040 TV_TREE_PHIOPT, /* tv_id */
3041 ( PROP_cfg | PROP_ssa ), /* properties_required */
3042 0, /* properties_provided */
3043 0, /* properties_destroyed */
3044 0, /* todo_flags_start */
3045 0, /* todo_flags_finish */
3048 class pass_cselim : public gimple_opt_pass
3050 public:
3051 pass_cselim (gcc::context *ctxt)
3052 : gimple_opt_pass (pass_data_cselim, ctxt)
3055 /* opt_pass methods: */
3056 virtual bool gate (function *) { return flag_tree_cselim; }
3057 virtual unsigned int execute (function *) { return tree_ssa_cs_elim (); }
3059 }; // class pass_cselim
3061 } // anon namespace
3063 gimple_opt_pass *
3064 make_pass_cselim (gcc::context *ctxt)
3066 return new pass_cselim (ctxt);