1 /* Thread edges through blocks and update the control flow and SSA graphs.
2 Copyright (C) 2004-2015 Free Software Foundation, Inc.
4 This file is part of GCC.
6 GCC is free software; you can redistribute it and/or modify
7 it under the terms of the GNU General Public License as published by
8 the Free Software Foundation; either version 3, or (at your option)
11 GCC is distributed in the hope that it will be useful,
12 but WITHOUT ANY WARRANTY; without even the implied warranty of
13 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 GNU General Public License for more details.
16 You should have received a copy of the GNU General Public License
17 along with GCC; see the file COPYING3. If not see
18 <http://www.gnu.org/licenses/>. */
22 #include "coretypes.h"
27 #include "fold-const.h"
31 #include "hard-reg-set.h"
33 #include "dominance.h"
36 #include "basic-block.h"
37 #include "tree-ssa-alias.h"
38 #include "internal-fn.h"
39 #include "gimple-expr.h"
41 #include "gimple-iterator.h"
42 #include "gimple-ssa.h"
43 #include "tree-phinodes.h"
45 #include "tree-ssa-threadupdate.h"
46 #include "ssa-iterators.h"
51 #include "tree-pass.h"
53 /* Given a block B, update the CFG and SSA graph to reflect redirecting
54 one or more in-edges to B to instead reach the destination of an
55 out-edge from B while preserving any side effects in B.
57 i.e., given A->B and B->C, change A->B to be A->C yet still preserve the
58 side effects of executing B.
60 1. Make a copy of B (including its outgoing edges and statements). Call
61 the copy B'. Note B' has no incoming edges or PHIs at this time.
63 2. Remove the control statement at the end of B' and all outgoing edges
66 3. Add a new argument to each PHI in C with the same value as the existing
67 argument associated with edge B->C. Associate the new PHI arguments
70 4. For each PHI in B, find or create a PHI in B' with an identical
71 PHI_RESULT. Add an argument to the PHI in B' which has the same
72 value as the PHI in B associated with the edge A->B. Associate
73 the new argument in the PHI in B' with the edge A->B.
75 5. Change the edge A->B to A->B'.
77 5a. This automatically deletes any PHI arguments associated with the
80 5b. This automatically associates each new argument added in step 4
83 6. Repeat for other incoming edges into B.
85 7. Put the duplicated resources in B and all the B' blocks into SSA form.
87 Note that block duplication can be minimized by first collecting the
88 set of unique destination blocks that the incoming edges should
91 We reduce the number of edges and statements we create by not copying all
92 the outgoing edges and the control statement in step #1. We instead create
93 a template block without the outgoing edges and duplicate the template.
95 Another case this code handles is threading through a "joiner" block. In
96 this case, we do not know the destination of the joiner block, but one
97 of the outgoing edges from the joiner block leads to a threadable path. This
98 case largely works as outlined above, except the duplicate of the joiner
99 block still contains a full set of outgoing edges and its control statement.
100 We just redirect one of its outgoing edges to our jump threading path. */
103 /* Steps #5 and #6 of the above algorithm are best implemented by walking
104 all the incoming edges which thread to the same destination edge at
105 the same time. That avoids lots of table lookups to get information
106 for the destination edge.
108 To realize that implementation we create a list of incoming edges
109 which thread to the same outgoing edge. Thus to implement steps
110 #5 and #6 we traverse our hash table of outgoing edge information.
111 For each entry we walk the list of incoming edges which thread to
112 the current outgoing edge. */
120 /* Main data structure recording information regarding B's duplicate
123 /* We need to efficiently record the unique thread destinations of this
124 block and specific information associated with those destinations. We
125 may have many incoming edges threaded to the same outgoing edge. This
126 can be naturally implemented with a hash table. */
128 struct redirection_data
: typed_free_remove
<redirection_data
>
130 /* We support wiring up two block duplicates in a jump threading path.
132 One is a normal block copy where we remove the control statement
133 and wire up its single remaining outgoing edge to the thread path.
135 The other is a joiner block where we leave the control statement
136 in place, but wire one of the outgoing edges to a thread path.
138 In theory we could have multiple block duplicates in a jump
139 threading path, but I haven't tried that.
141 The duplicate blocks appear in this array in the same order in
142 which they appear in the jump thread path. */
143 basic_block dup_blocks
[2];
145 /* The jump threading path. */
146 vec
<jump_thread_edge
*> *path
;
148 /* A list of incoming edges which we want to thread to the
150 struct el
*incoming_edges
;
152 /* hash_table support. */
153 typedef redirection_data
*value_type
;
154 typedef redirection_data
*compare_type
;
155 static inline hashval_t
hash (const redirection_data
*);
156 static inline int equal (const redirection_data
*, const redirection_data
*);
159 /* Dump a jump threading path, including annotations about each
163 dump_jump_thread_path (FILE *dump_file
, vec
<jump_thread_edge
*> path
,
167 " %s%s jump thread: (%d, %d) incoming edge; ",
168 (registering
? "Registering" : "Cancelling"),
169 (path
[0]->type
== EDGE_FSM_THREAD
? " FSM": ""),
170 path
[0]->e
->src
->index
, path
[0]->e
->dest
->index
);
172 for (unsigned int i
= 1; i
< path
.length (); i
++)
174 /* We can get paths with a NULL edge when the final destination
175 of a jump thread turns out to be a constant address. We dump
176 those paths when debugging, so we have to be prepared for that
178 if (path
[i
]->e
== NULL
)
181 if (path
[i
]->type
== EDGE_COPY_SRC_JOINER_BLOCK
)
182 fprintf (dump_file
, " (%d, %d) joiner; ",
183 path
[i
]->e
->src
->index
, path
[i
]->e
->dest
->index
);
184 if (path
[i
]->type
== EDGE_COPY_SRC_BLOCK
)
185 fprintf (dump_file
, " (%d, %d) normal;",
186 path
[i
]->e
->src
->index
, path
[i
]->e
->dest
->index
);
187 if (path
[i
]->type
== EDGE_NO_COPY_SRC_BLOCK
)
188 fprintf (dump_file
, " (%d, %d) nocopy;",
189 path
[i
]->e
->src
->index
, path
[i
]->e
->dest
->index
);
190 if (path
[0]->type
== EDGE_FSM_THREAD
)
191 fprintf (dump_file
, " (%d, %d) ",
192 path
[i
]->e
->src
->index
, path
[i
]->e
->dest
->index
);
194 fputc ('\n', dump_file
);
197 /* Simple hashing function. For any given incoming edge E, we're going
198 to be most concerned with the final destination of its jump thread
199 path. So hash on the block index of the final edge in the path. */
202 redirection_data::hash (const redirection_data
*p
)
204 vec
<jump_thread_edge
*> *path
= p
->path
;
205 return path
->last ()->e
->dest
->index
;
208 /* Given two hash table entries, return true if they have the same
209 jump threading path. */
211 redirection_data::equal (const redirection_data
*p1
, const redirection_data
*p2
)
213 vec
<jump_thread_edge
*> *path1
= p1
->path
;
214 vec
<jump_thread_edge
*> *path2
= p2
->path
;
216 if (path1
->length () != path2
->length ())
219 for (unsigned int i
= 1; i
< path1
->length (); i
++)
221 if ((*path1
)[i
]->type
!= (*path2
)[i
]->type
222 || (*path1
)[i
]->e
!= (*path2
)[i
]->e
)
229 /* Data structure of information to pass to hash table traversal routines. */
230 struct ssa_local_info_t
232 /* The current block we are working on. */
235 /* We only create a template block for the first duplicated block in a
236 jump threading path as we may need many duplicates of that block.
238 The second duplicate block in a path is specific to that path. Creating
239 and sharing a template for that block is considerably more difficult. */
240 basic_block template_block
;
242 /* TRUE if we thread one or more jumps, FALSE otherwise. */
245 /* Blocks duplicated for the thread. */
246 bitmap duplicate_blocks
;
249 /* Passes which use the jump threading code register jump threading
250 opportunities as they are discovered. We keep the registered
251 jump threading opportunities in this vector as edge pairs
252 (original_edge, target_edge). */
253 static vec
<vec
<jump_thread_edge
*> *> paths
;
255 /* When we start updating the CFG for threading, data necessary for jump
256 threading is attached to the AUX field for the incoming edge. Use these
257 macros to access the underlying structure attached to the AUX field. */
258 #define THREAD_PATH(E) ((vec<jump_thread_edge *> *)(E)->aux)
260 /* Jump threading statistics. */
262 struct thread_stats_d
264 unsigned long num_threaded_edges
;
267 struct thread_stats_d thread_stats
;
270 /* Remove the last statement in block BB if it is a control statement
271 Also remove all outgoing edges except the edge which reaches DEST_BB.
272 If DEST_BB is NULL, then remove all outgoing edges. */
275 remove_ctrl_stmt_and_useless_edges (basic_block bb
, basic_block dest_bb
)
277 gimple_stmt_iterator gsi
;
281 gsi
= gsi_last_bb (bb
);
283 /* If the duplicate ends with a control statement, then remove it.
285 Note that if we are duplicating the template block rather than the
286 original basic block, then the duplicate might not have any real
290 && (gimple_code (gsi_stmt (gsi
)) == GIMPLE_COND
291 || gimple_code (gsi_stmt (gsi
)) == GIMPLE_GOTO
292 || gimple_code (gsi_stmt (gsi
)) == GIMPLE_SWITCH
))
293 gsi_remove (&gsi
, true);
295 for (ei
= ei_start (bb
->succs
); (e
= ei_safe_edge (ei
)); )
297 if (e
->dest
!= dest_bb
)
304 /* Create a duplicate of BB. Record the duplicate block in an array
305 indexed by COUNT stored in RD. */
308 create_block_for_threading (basic_block bb
,
309 struct redirection_data
*rd
,
311 bitmap
*duplicate_blocks
)
316 /* We can use the generic block duplication code and simply remove
317 the stuff we do not need. */
318 rd
->dup_blocks
[count
] = duplicate_block (bb
, NULL
, NULL
);
320 FOR_EACH_EDGE (e
, ei
, rd
->dup_blocks
[count
]->succs
)
323 /* Zero out the profile, since the block is unreachable for now. */
324 rd
->dup_blocks
[count
]->frequency
= 0;
325 rd
->dup_blocks
[count
]->count
= 0;
326 if (duplicate_blocks
)
327 bitmap_set_bit (*duplicate_blocks
, rd
->dup_blocks
[count
]->index
);
330 /* Main data structure to hold information for duplicates of BB. */
332 static hash_table
<redirection_data
> *redirection_data
;
334 /* Given an outgoing edge E lookup and return its entry in our hash table.
336 If INSERT is true, then we insert the entry into the hash table if
337 it is not already present. INCOMING_EDGE is added to the list of incoming
338 edges associated with E in the hash table. */
340 static struct redirection_data
*
341 lookup_redirection_data (edge e
, enum insert_option insert
)
343 struct redirection_data
**slot
;
344 struct redirection_data
*elt
;
345 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
347 /* Build a hash table element so we can see if E is already
349 elt
= XNEW (struct redirection_data
);
351 elt
->dup_blocks
[0] = NULL
;
352 elt
->dup_blocks
[1] = NULL
;
353 elt
->incoming_edges
= NULL
;
355 slot
= redirection_data
->find_slot (elt
, insert
);
357 /* This will only happen if INSERT is false and the entry is not
358 in the hash table. */
365 /* This will only happen if E was not in the hash table and
370 elt
->incoming_edges
= XNEW (struct el
);
371 elt
->incoming_edges
->e
= e
;
372 elt
->incoming_edges
->next
= NULL
;
375 /* E was in the hash table. */
378 /* Free ELT as we do not need it anymore, we will extract the
379 relevant entry from the hash table itself. */
382 /* Get the entry stored in the hash table. */
385 /* If insertion was requested, then we need to add INCOMING_EDGE
386 to the list of incoming edges associated with E. */
389 struct el
*el
= XNEW (struct el
);
390 el
->next
= elt
->incoming_edges
;
392 elt
->incoming_edges
= el
;
399 /* Similar to copy_phi_args, except that the PHI arg exists, it just
400 does not have a value associated with it. */
403 copy_phi_arg_into_existing_phi (edge src_e
, edge tgt_e
)
405 int src_idx
= src_e
->dest_idx
;
406 int tgt_idx
= tgt_e
->dest_idx
;
408 /* Iterate over each PHI in e->dest. */
409 for (gphi_iterator gsi
= gsi_start_phis (src_e
->dest
),
410 gsi2
= gsi_start_phis (tgt_e
->dest
);
412 gsi_next (&gsi
), gsi_next (&gsi2
))
414 gphi
*src_phi
= gsi
.phi ();
415 gphi
*dest_phi
= gsi2
.phi ();
416 tree val
= gimple_phi_arg_def (src_phi
, src_idx
);
417 source_location locus
= gimple_phi_arg_location (src_phi
, src_idx
);
419 SET_PHI_ARG_DEF (dest_phi
, tgt_idx
, val
);
420 gimple_phi_arg_set_location (dest_phi
, tgt_idx
, locus
);
424 /* Given ssa_name DEF, backtrack jump threading PATH from node IDX
425 to see if it has constant value in a flow sensitive manner. Set
426 LOCUS to location of the constant phi arg and return the value.
427 Return DEF directly if either PATH or idx is ZERO. */
430 get_value_locus_in_path (tree def
, vec
<jump_thread_edge
*> *path
,
431 basic_block bb
, int idx
, source_location
*locus
)
437 if (path
== NULL
|| idx
== 0)
440 def_phi
= dyn_cast
<gphi
*> (SSA_NAME_DEF_STMT (def
));
444 def_bb
= gimple_bb (def_phi
);
445 /* Don't propagate loop invariants into deeper loops. */
446 if (!def_bb
|| bb_loop_depth (def_bb
) < bb_loop_depth (bb
))
449 /* Backtrack jump threading path from IDX to see if def has constant
451 for (int j
= idx
- 1; j
>= 0; j
--)
453 edge e
= (*path
)[j
]->e
;
454 if (e
->dest
== def_bb
)
456 arg
= gimple_phi_arg_def (def_phi
, e
->dest_idx
);
457 if (is_gimple_min_invariant (arg
))
459 *locus
= gimple_phi_arg_location (def_phi
, e
->dest_idx
);
469 /* For each PHI in BB, copy the argument associated with SRC_E to TGT_E.
470 Try to backtrack jump threading PATH from node IDX to see if the arg
471 has constant value, copy constant value instead of argument itself
475 copy_phi_args (basic_block bb
, edge src_e
, edge tgt_e
,
476 vec
<jump_thread_edge
*> *path
, int idx
)
479 int src_indx
= src_e
->dest_idx
;
481 for (gsi
= gsi_start_phis (bb
); !gsi_end_p (gsi
); gsi_next (&gsi
))
483 gphi
*phi
= gsi
.phi ();
484 tree def
= gimple_phi_arg_def (phi
, src_indx
);
485 source_location locus
= gimple_phi_arg_location (phi
, src_indx
);
487 if (TREE_CODE (def
) == SSA_NAME
488 && !virtual_operand_p (gimple_phi_result (phi
)))
489 def
= get_value_locus_in_path (def
, path
, bb
, idx
, &locus
);
491 add_phi_arg (phi
, def
, tgt_e
, locus
);
495 /* We have recently made a copy of ORIG_BB, including its outgoing
496 edges. The copy is NEW_BB. Every PHI node in every direct successor of
497 ORIG_BB has a new argument associated with edge from NEW_BB to the
498 successor. Initialize the PHI argument so that it is equal to the PHI
499 argument associated with the edge from ORIG_BB to the successor.
500 PATH and IDX are used to check if the new PHI argument has constant
501 value in a flow sensitive manner. */
504 update_destination_phis (basic_block orig_bb
, basic_block new_bb
,
505 vec
<jump_thread_edge
*> *path
, int idx
)
510 FOR_EACH_EDGE (e
, ei
, orig_bb
->succs
)
512 edge e2
= find_edge (new_bb
, e
->dest
);
513 copy_phi_args (e
->dest
, e
, e2
, path
, idx
);
517 /* Given a duplicate block and its single destination (both stored
518 in RD). Create an edge between the duplicate and its single
521 Add an additional argument to any PHI nodes at the single
522 destination. IDX is the start node in jump threading path
523 we start to check to see if the new PHI argument has constant
524 value along the jump threading path. */
527 create_edge_and_update_destination_phis (struct redirection_data
*rd
,
528 basic_block bb
, int idx
)
530 edge e
= make_edge (bb
, rd
->path
->last ()->e
->dest
, EDGE_FALLTHRU
);
532 rescan_loop_exit (e
, true, false);
533 e
->probability
= REG_BR_PROB_BASE
;
534 e
->count
= bb
->count
;
536 /* We used to copy the thread path here. That was added in 2007
537 and dutifully updated through the representation changes in 2013.
539 In 2013 we added code to thread from an interior node through
540 the backedge to another interior node. That runs after the code
541 to thread through loop headers from outside the loop.
543 The latter may delete edges in the CFG, including those
544 which appeared in the jump threading path we copied here. Thus
545 we'd end up using a dangling pointer.
547 After reviewing the 2007/2011 code, I can't see how anything
548 depended on copying the AUX field and clearly copying the jump
549 threading path is problematical due to embedded edge pointers.
550 It has been removed. */
553 /* If there are any PHI nodes at the destination of the outgoing edge
554 from the duplicate block, then we will need to add a new argument
555 to them. The argument should have the same value as the argument
556 associated with the outgoing edge stored in RD. */
557 copy_phi_args (e
->dest
, rd
->path
->last ()->e
, e
, rd
->path
, idx
);
560 /* Look through PATH beginning at START and return TRUE if there are
561 any additional blocks that need to be duplicated. Otherwise,
564 any_remaining_duplicated_blocks (vec
<jump_thread_edge
*> *path
,
567 for (unsigned int i
= start
+ 1; i
< path
->length (); i
++)
569 if ((*path
)[i
]->type
== EDGE_COPY_SRC_JOINER_BLOCK
570 || (*path
)[i
]->type
== EDGE_COPY_SRC_BLOCK
)
577 /* Compute the amount of profile count/frequency coming into the jump threading
578 path stored in RD that we are duplicating, returned in PATH_IN_COUNT_PTR and
579 PATH_IN_FREQ_PTR, as well as the amount of counts flowing out of the
580 duplicated path, returned in PATH_OUT_COUNT_PTR. LOCAL_INFO is used to
581 identify blocks duplicated for jump threading, which have duplicated
582 edges that need to be ignored in the analysis. Return true if path contains
583 a joiner, false otherwise.
585 In the non-joiner case, this is straightforward - all the counts/frequency
586 flowing into the jump threading path should flow through the duplicated
587 block and out of the duplicated path.
589 In the joiner case, it is very tricky. Some of the counts flowing into
590 the original path go offpath at the joiner. The problem is that while
591 we know how much total count goes off-path in the original control flow,
592 we don't know how many of the counts corresponding to just the jump
593 threading path go offpath at the joiner.
595 For example, assume we have the following control flow and identified
596 jump threading paths:
615 Jump threading paths: A -> J -> Son -> D (path 1)
616 C -> J -> Son -> E (path 2)
618 Note that the control flow could be more complicated:
619 - Each jump threading path may have more than one incoming edge. I.e. A and
620 Ea could represent multiple incoming blocks/edges that are included in
622 - There could be EDGE_NO_COPY_SRC_BLOCK edges after the joiner (either
623 before or after the "normal" copy block). These are not duplicated onto
624 the jump threading path, as they are single-successor.
625 - Any of the blocks along the path may have other incoming edges that
626 are not part of any jump threading path, but add profile counts along
629 In the aboe example, after all jump threading is complete, we will
630 end up with the following control flow:
639 Eona/ \ ---/---\-------- \Eonc
644 \___________ / \ _____/
649 The main issue to notice here is that when we are processing path 1
650 (A->J->Son->D) we need to figure out the outgoing edge weights to
651 the duplicated edges Ja->Sona and Ja->Soff, while ensuring that the
652 sum of the incoming weights to D remain Ed. The problem with simply
653 assuming that Ja (and Jc when processing path 2) has the same outgoing
654 probabilities to its successors as the original block J, is that after
655 all paths are processed and other edges/counts removed (e.g. none
656 of Ec will reach D after processing path 2), we may end up with not
657 enough count flowing along duplicated edge Sona->D.
659 Therefore, in the case of a joiner, we keep track of all counts
660 coming in along the current path, as well as from predecessors not
661 on any jump threading path (Eb in the above example). While we
662 first assume that the duplicated Eona for Ja->Sona has the same
663 probability as the original, we later compensate for other jump
664 threading paths that may eliminate edges. We do that by keep track
665 of all counts coming into the original path that are not in a jump
666 thread (Eb in the above example, but as noted earlier, there could
667 be other predecessors incoming to the path at various points, such
668 as at Son). Call this cumulative non-path count coming into the path
669 before D as Enonpath. We then ensure that the count from Sona->D is as at
670 least as big as (Ed - Enonpath), but no bigger than the minimum
671 weight along the jump threading path. The probabilities of both the
672 original and duplicated joiner block J and Ja will be adjusted
673 accordingly after the updates. */
676 compute_path_counts (struct redirection_data
*rd
,
677 ssa_local_info_t
*local_info
,
678 gcov_type
*path_in_count_ptr
,
679 gcov_type
*path_out_count_ptr
,
680 int *path_in_freq_ptr
)
682 edge e
= rd
->incoming_edges
->e
;
683 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
684 edge elast
= path
->last ()->e
;
685 gcov_type nonpath_count
= 0;
686 bool has_joiner
= false;
687 gcov_type path_in_count
= 0;
688 int path_in_freq
= 0;
690 /* Start by accumulating incoming edge counts to the path's first bb
691 into a couple buckets:
692 path_in_count: total count of incoming edges that flow into the
694 nonpath_count: total count of incoming edges that are not
695 flowing along *any* path. These are the counts
696 that will still flow along the original path after
697 all path duplication is done by potentially multiple
698 calls to this routine.
699 (any other incoming edge counts are for a different jump threading
700 path that will be handled by a later call to this routine.)
701 To make this easier, start by recording all incoming edges that flow into
702 the current path in a bitmap. We could add up the path's incoming edge
703 counts here, but we still need to walk all the first bb's incoming edges
704 below to add up the counts of the other edges not included in this jump
706 struct el
*next
, *el
;
707 bitmap in_edge_srcs
= BITMAP_ALLOC (NULL
);
708 for (el
= rd
->incoming_edges
; el
; el
= next
)
711 bitmap_set_bit (in_edge_srcs
, el
->e
->src
->index
);
715 FOR_EACH_EDGE (ein
, ei
, e
->dest
->preds
)
717 vec
<jump_thread_edge
*> *ein_path
= THREAD_PATH (ein
);
718 /* Simply check the incoming edge src against the set captured above. */
720 && bitmap_bit_p (in_edge_srcs
, (*ein_path
)[0]->e
->src
->index
))
722 /* It is necessary but not sufficient that the last path edges
723 are identical. There may be different paths that share the
724 same last path edge in the case where the last edge has a nocopy
726 gcc_assert (ein_path
->last ()->e
== elast
);
727 path_in_count
+= ein
->count
;
728 path_in_freq
+= EDGE_FREQUENCY (ein
);
732 /* Keep track of the incoming edges that are not on any jump-threading
733 path. These counts will still flow out of original path after all
734 jump threading is complete. */
735 nonpath_count
+= ein
->count
;
739 /* This is needed due to insane incoming frequencies. */
740 if (path_in_freq
> BB_FREQ_MAX
)
741 path_in_freq
= BB_FREQ_MAX
;
743 BITMAP_FREE (in_edge_srcs
);
745 /* Now compute the fraction of the total count coming into the first
746 path bb that is from the current threading path. */
747 gcov_type total_count
= e
->dest
->count
;
748 /* Handle incoming profile insanities. */
749 if (total_count
< path_in_count
)
750 path_in_count
= total_count
;
751 int onpath_scale
= GCOV_COMPUTE_SCALE (path_in_count
, total_count
);
753 /* Walk the entire path to do some more computation in order to estimate
754 how much of the path_in_count will flow out of the duplicated threading
755 path. In the non-joiner case this is straightforward (it should be
756 the same as path_in_count, although we will handle incoming profile
757 insanities by setting it equal to the minimum count along the path).
759 In the joiner case, we need to estimate how much of the path_in_count
760 will stay on the threading path after the joiner's conditional branch.
761 We don't really know for sure how much of the counts
762 associated with this path go to each successor of the joiner, but we'll
763 estimate based on the fraction of the total count coming into the path
764 bb was from the threading paths (computed above in onpath_scale).
765 Afterwards, we will need to do some fixup to account for other threading
766 paths and possible profile insanities.
768 In order to estimate the joiner case's counts we also need to update
769 nonpath_count with any additional counts coming into the path. Other
770 blocks along the path may have additional predecessors from outside
772 gcov_type path_out_count
= path_in_count
;
773 gcov_type min_path_count
= path_in_count
;
774 for (unsigned int i
= 1; i
< path
->length (); i
++)
776 edge epath
= (*path
)[i
]->e
;
777 gcov_type cur_count
= epath
->count
;
778 if ((*path
)[i
]->type
== EDGE_COPY_SRC_JOINER_BLOCK
)
781 cur_count
= apply_probability (cur_count
, onpath_scale
);
783 /* In the joiner case we need to update nonpath_count for any edges
784 coming into the path that will contribute to the count flowing
785 into the path successor. */
786 if (has_joiner
&& epath
!= elast
)
788 /* Look for other incoming edges after joiner. */
789 FOR_EACH_EDGE (ein
, ei
, epath
->dest
->preds
)
792 /* Ignore in edges from blocks we have duplicated for a
793 threading path, which have duplicated edge counts until
794 they are redirected by an invocation of this routine. */
795 && !bitmap_bit_p (local_info
->duplicate_blocks
,
797 nonpath_count
+= ein
->count
;
800 if (cur_count
< path_out_count
)
801 path_out_count
= cur_count
;
802 if (epath
->count
< min_path_count
)
803 min_path_count
= epath
->count
;
806 /* We computed path_out_count above assuming that this path targeted
807 the joiner's on-path successor with the same likelihood as it
808 reached the joiner. However, other thread paths through the joiner
809 may take a different path through the normal copy source block
810 (i.e. they have a different elast), meaning that they do not
811 contribute any counts to this path's elast. As a result, it may
812 turn out that this path must have more count flowing to the on-path
813 successor of the joiner. Essentially, all of this path's elast
814 count must be contributed by this path and any nonpath counts
815 (since any path through the joiner with a different elast will not
816 include a copy of this elast in its duplicated path).
817 So ensure that this path's path_out_count is at least the
818 difference between elast->count and nonpath_count. Otherwise the edge
819 counts after threading will not be sane. */
820 if (has_joiner
&& path_out_count
< elast
->count
- nonpath_count
)
822 path_out_count
= elast
->count
- nonpath_count
;
823 /* But neither can we go above the minimum count along the path
824 we are duplicating. This can be an issue due to profile
825 insanities coming in to this pass. */
826 if (path_out_count
> min_path_count
)
827 path_out_count
= min_path_count
;
830 *path_in_count_ptr
= path_in_count
;
831 *path_out_count_ptr
= path_out_count
;
832 *path_in_freq_ptr
= path_in_freq
;
837 /* Update the counts and frequencies for both an original path
838 edge EPATH and its duplicate EDUP. The duplicate source block
839 will get a count/frequency of PATH_IN_COUNT and PATH_IN_FREQ,
840 and the duplicate edge EDUP will have a count of PATH_OUT_COUNT. */
842 update_profile (edge epath
, edge edup
, gcov_type path_in_count
,
843 gcov_type path_out_count
, int path_in_freq
)
846 /* First update the duplicated block's count / frequency. */
849 basic_block dup_block
= edup
->src
;
850 gcc_assert (dup_block
->count
== 0);
851 gcc_assert (dup_block
->frequency
== 0);
852 dup_block
->count
= path_in_count
;
853 dup_block
->frequency
= path_in_freq
;
856 /* Now update the original block's count and frequency in the
857 opposite manner - remove the counts/freq that will flow
858 into the duplicated block. Handle underflow due to precision/
860 epath
->src
->count
-= path_in_count
;
861 if (epath
->src
->count
< 0)
862 epath
->src
->count
= 0;
863 epath
->src
->frequency
-= path_in_freq
;
864 if (epath
->src
->frequency
< 0)
865 epath
->src
->frequency
= 0;
867 /* Next update this path edge's original and duplicated counts. We know
868 that the duplicated path will have path_out_count flowing
869 out of it (in the joiner case this is the count along the duplicated path
870 out of the duplicated joiner). This count can then be removed from the
871 original path edge. */
873 edup
->count
= path_out_count
;
874 epath
->count
-= path_out_count
;
875 gcc_assert (epath
->count
>= 0);
879 /* The duplicate and original joiner blocks may end up with different
880 probabilities (different from both the original and from each other).
881 Recompute the probabilities here once we have updated the edge
882 counts and frequencies. */
885 recompute_probabilities (basic_block bb
)
889 FOR_EACH_EDGE (esucc
, ei
, bb
->succs
)
894 /* Prevent overflow computation due to insane profiles. */
895 if (esucc
->count
< bb
->count
)
896 esucc
->probability
= GCOV_COMPUTE_SCALE (esucc
->count
,
899 /* Can happen with missing/guessed probabilities, since we
900 may determine that more is flowing along duplicated
901 path than joiner succ probabilities allowed.
902 Counts and freqs will be insane after jump threading,
903 at least make sure probability is sane or we will
904 get a flow verification error.
905 Not much we can do to make counts/freqs sane without
906 redoing the profile estimation. */
907 esucc
->probability
= REG_BR_PROB_BASE
;
912 /* Update the counts of the original and duplicated edges from a joiner
913 that go off path, given that we have already determined that the
914 duplicate joiner DUP_BB has incoming count PATH_IN_COUNT and
915 outgoing count along the path PATH_OUT_COUNT. The original (on-)path
916 edge from joiner is EPATH. */
919 update_joiner_offpath_counts (edge epath
, basic_block dup_bb
,
920 gcov_type path_in_count
,
921 gcov_type path_out_count
)
923 /* Compute the count that currently flows off path from the joiner.
924 In other words, the total count of joiner's out edges other than
925 epath. Compute this by walking the successors instead of
926 subtracting epath's count from the joiner bb count, since there
927 are sometimes slight insanities where the total out edge count is
928 larger than the bb count (possibly due to rounding/truncation
930 gcov_type total_orig_off_path_count
= 0;
933 FOR_EACH_EDGE (enonpath
, ei
, epath
->src
->succs
)
935 if (enonpath
== epath
)
937 total_orig_off_path_count
+= enonpath
->count
;
940 /* For the path that we are duplicating, the amount that will flow
941 off path from the duplicated joiner is the delta between the
942 path's cumulative in count and the portion of that count we
943 estimated above as flowing from the joiner along the duplicated
945 gcov_type total_dup_off_path_count
= path_in_count
- path_out_count
;
947 /* Now do the actual updates of the off-path edges. */
948 FOR_EACH_EDGE (enonpath
, ei
, epath
->src
->succs
)
950 /* Look for edges going off of the threading path. */
951 if (enonpath
== epath
)
954 /* Find the corresponding edge out of the duplicated joiner. */
955 edge enonpathdup
= find_edge (dup_bb
, enonpath
->dest
);
956 gcc_assert (enonpathdup
);
958 /* We can't use the original probability of the joiner's out
959 edges, since the probabilities of the original branch
960 and the duplicated branches may vary after all threading is
961 complete. But apportion the duplicated joiner's off-path
962 total edge count computed earlier (total_dup_off_path_count)
963 among the duplicated off-path edges based on their original
964 ratio to the full off-path count (total_orig_off_path_count).
966 int scale
= GCOV_COMPUTE_SCALE (enonpath
->count
,
967 total_orig_off_path_count
);
968 /* Give the duplicated offpath edge a portion of the duplicated
970 enonpathdup
->count
= apply_scale (scale
,
971 total_dup_off_path_count
);
972 /* Now update the original offpath edge count, handling underflow
973 due to rounding errors. */
974 enonpath
->count
-= enonpathdup
->count
;
975 if (enonpath
->count
< 0)
981 /* Check if the paths through RD all have estimated frequencies but zero
982 profile counts. This is more accurate than checking the entry block
983 for a zero profile count, since profile insanities sometimes creep in. */
986 estimated_freqs_path (struct redirection_data
*rd
)
988 edge e
= rd
->incoming_edges
->e
;
989 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
992 bool non_zero_freq
= false;
993 FOR_EACH_EDGE (ein
, ei
, e
->dest
->preds
)
997 non_zero_freq
|= ein
->src
->frequency
!= 0;
1000 for (unsigned int i
= 1; i
< path
->length (); i
++)
1002 edge epath
= (*path
)[i
]->e
;
1003 if (epath
->src
->count
)
1005 non_zero_freq
|= epath
->src
->frequency
!= 0;
1007 FOR_EACH_EDGE (esucc
, ei
, epath
->src
->succs
)
1011 non_zero_freq
|= esucc
->src
->frequency
!= 0;
1014 return non_zero_freq
;
1018 /* Invoked for routines that have guessed frequencies and no profile
1019 counts to record the block and edge frequencies for paths through RD
1020 in the profile count fields of those blocks and edges. This is because
1021 ssa_fix_duplicate_block_edges incrementally updates the block and
1022 edge counts as edges are redirected, and it is difficult to do that
1023 for edge frequencies which are computed on the fly from the source
1024 block frequency and probability. When a block frequency is updated
1025 its outgoing edge frequencies are affected and become difficult to
1029 freqs_to_counts_path (struct redirection_data
*rd
)
1031 edge e
= rd
->incoming_edges
->e
;
1032 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
1035 FOR_EACH_EDGE (ein
, ei
, e
->dest
->preds
)
1037 /* Scale up the frequency by REG_BR_PROB_BASE, to avoid rounding
1038 errors applying the probability when the frequencies are very
1040 ein
->count
= apply_probability (ein
->src
->frequency
* REG_BR_PROB_BASE
,
1044 for (unsigned int i
= 1; i
< path
->length (); i
++)
1046 edge epath
= (*path
)[i
]->e
;
1048 /* Scale up the frequency by REG_BR_PROB_BASE, to avoid rounding
1049 errors applying the edge probability when the frequencies are very
1051 epath
->src
->count
= epath
->src
->frequency
* REG_BR_PROB_BASE
;
1052 FOR_EACH_EDGE (esucc
, ei
, epath
->src
->succs
)
1053 esucc
->count
= apply_probability (esucc
->src
->count
,
1054 esucc
->probability
);
1059 /* For routines that have guessed frequencies and no profile counts, where we
1060 used freqs_to_counts_path to record block and edge frequencies for paths
1061 through RD, we clear the counts after completing all updates for RD.
1062 The updates in ssa_fix_duplicate_block_edges are based off the count fields,
1063 but the block frequencies and edge probabilities were updated as well,
1064 so we can simply clear the count fields. */
1067 clear_counts_path (struct redirection_data
*rd
)
1069 edge e
= rd
->incoming_edges
->e
;
1070 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
1073 FOR_EACH_EDGE (ein
, ei
, e
->dest
->preds
)
1076 /* First clear counts along original path. */
1077 for (unsigned int i
= 1; i
< path
->length (); i
++)
1079 edge epath
= (*path
)[i
]->e
;
1080 FOR_EACH_EDGE (esucc
, ei
, epath
->src
->succs
)
1082 epath
->src
->count
= 0;
1084 /* Also need to clear the counts along duplicated path. */
1085 for (unsigned int i
= 0; i
< 2; i
++)
1087 basic_block dup
= rd
->dup_blocks
[i
];
1090 FOR_EACH_EDGE (esucc
, ei
, dup
->succs
)
1096 /* Wire up the outgoing edges from the duplicate blocks and
1097 update any PHIs as needed. Also update the profile counts
1098 on the original and duplicate blocks and edges. */
1100 ssa_fix_duplicate_block_edges (struct redirection_data
*rd
,
1101 ssa_local_info_t
*local_info
)
1103 bool multi_incomings
= (rd
->incoming_edges
->next
!= NULL
);
1104 edge e
= rd
->incoming_edges
->e
;
1105 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
1106 edge elast
= path
->last ()->e
;
1107 gcov_type path_in_count
= 0;
1108 gcov_type path_out_count
= 0;
1109 int path_in_freq
= 0;
1111 /* This routine updates profile counts, frequencies, and probabilities
1112 incrementally. Since it is difficult to do the incremental updates
1113 using frequencies/probabilities alone, for routines without profile
1114 data we first take a snapshot of the existing block and edge frequencies
1115 by copying them into the empty profile count fields. These counts are
1116 then used to do the incremental updates, and cleared at the end of this
1117 routine. If the function is marked as having a profile, we still check
1118 to see if the paths through RD are using estimated frequencies because
1119 the routine had zero profile counts. */
1120 bool do_freqs_to_counts
= (profile_status_for_fn (cfun
) != PROFILE_READ
1121 || estimated_freqs_path (rd
));
1122 if (do_freqs_to_counts
)
1123 freqs_to_counts_path (rd
);
1125 /* First determine how much profile count to move from original
1126 path to the duplicate path. This is tricky in the presence of
1127 a joiner (see comments for compute_path_counts), where some portion
1128 of the path's counts will flow off-path from the joiner. In the
1129 non-joiner case the path_in_count and path_out_count should be the
1131 bool has_joiner
= compute_path_counts (rd
, local_info
,
1132 &path_in_count
, &path_out_count
,
1135 int cur_path_freq
= path_in_freq
;
1136 for (unsigned int count
= 0, i
= 1; i
< path
->length (); i
++)
1138 edge epath
= (*path
)[i
]->e
;
1140 /* If we were threading through an joiner block, then we want
1141 to keep its control statement and redirect an outgoing edge.
1142 Else we want to remove the control statement & edges, then create
1143 a new outgoing edge. In both cases we may need to update PHIs. */
1144 if ((*path
)[i
]->type
== EDGE_COPY_SRC_JOINER_BLOCK
)
1149 gcc_assert (has_joiner
);
1151 /* This updates the PHIs at the destination of the duplicate
1152 block. Pass 0 instead of i if we are threading a path which
1153 has multiple incoming edges. */
1154 update_destination_phis (local_info
->bb
, rd
->dup_blocks
[count
],
1155 path
, multi_incomings
? 0 : i
);
1157 /* Find the edge from the duplicate block to the block we're
1158 threading through. That's the edge we want to redirect. */
1159 victim
= find_edge (rd
->dup_blocks
[count
], (*path
)[i
]->e
->dest
);
1161 /* If there are no remaining blocks on the path to duplicate,
1162 then redirect VICTIM to the final destination of the jump
1164 if (!any_remaining_duplicated_blocks (path
, i
))
1166 e2
= redirect_edge_and_branch (victim
, elast
->dest
);
1167 /* If we redirected the edge, then we need to copy PHI arguments
1168 at the target. If the edge already existed (e2 != victim
1169 case), then the PHIs in the target already have the correct
1172 copy_phi_args (e2
->dest
, elast
, e2
,
1173 path
, multi_incomings
? 0 : i
);
1177 /* Redirect VICTIM to the next duplicated block in the path. */
1178 e2
= redirect_edge_and_branch (victim
, rd
->dup_blocks
[count
+ 1]);
1180 /* We need to update the PHIs in the next duplicated block. We
1181 want the new PHI args to have the same value as they had
1182 in the source of the next duplicate block.
1184 Thus, we need to know which edge we traversed into the
1185 source of the duplicate. Furthermore, we may have
1186 traversed many edges to reach the source of the duplicate.
1188 Walk through the path starting at element I until we
1189 hit an edge marked with EDGE_COPY_SRC_BLOCK. We want
1190 the edge from the prior element. */
1191 for (unsigned int j
= i
+ 1; j
< path
->length (); j
++)
1193 if ((*path
)[j
]->type
== EDGE_COPY_SRC_BLOCK
)
1195 copy_phi_arg_into_existing_phi ((*path
)[j
- 1]->e
, e2
);
1201 /* Update the counts and frequency of both the original block
1202 and path edge, and the duplicates. The path duplicate's
1203 incoming count and frequency are the totals for all edges
1204 incoming to this jump threading path computed earlier.
1205 And we know that the duplicated path will have path_out_count
1206 flowing out of it (i.e. along the duplicated path out of the
1207 duplicated joiner). */
1208 update_profile (epath
, e2
, path_in_count
, path_out_count
,
1211 /* Next we need to update the counts of the original and duplicated
1212 edges from the joiner that go off path. */
1213 update_joiner_offpath_counts (epath
, e2
->src
, path_in_count
,
1216 /* Finally, we need to set the probabilities on the duplicated
1217 edges out of the duplicated joiner (e2->src). The probabilities
1218 along the original path will all be updated below after we finish
1219 processing the whole path. */
1220 recompute_probabilities (e2
->src
);
1222 /* Record the frequency flowing to the downstream duplicated
1224 cur_path_freq
= EDGE_FREQUENCY (e2
);
1226 else if ((*path
)[i
]->type
== EDGE_COPY_SRC_BLOCK
)
1228 remove_ctrl_stmt_and_useless_edges (rd
->dup_blocks
[count
], NULL
);
1229 create_edge_and_update_destination_phis (rd
, rd
->dup_blocks
[count
],
1230 multi_incomings
? 0 : i
);
1232 single_succ_edge (rd
->dup_blocks
[1])->aux
= NULL
;
1234 /* Update the counts and frequency of both the original block
1235 and path edge, and the duplicates. Since we are now after
1236 any joiner that may have existed on the path, the count
1237 flowing along the duplicated threaded path is path_out_count.
1238 If we didn't have a joiner, then cur_path_freq was the sum
1239 of the total frequencies along all incoming edges to the
1240 thread path (path_in_freq). If we had a joiner, it would have
1241 been updated at the end of that handling to the edge frequency
1242 along the duplicated joiner path edge. */
1243 update_profile (epath
, EDGE_SUCC (rd
->dup_blocks
[count
], 0),
1244 path_out_count
, path_out_count
,
1249 /* No copy case. In this case we don't have an equivalent block
1250 on the duplicated thread path to update, but we do need
1251 to remove the portion of the counts/freqs that were moved
1252 to the duplicated path from the counts/freqs flowing through
1253 this block on the original path. Since all the no-copy edges
1254 are after any joiner, the removed count is the same as
1257 If we didn't have a joiner, then cur_path_freq was the sum
1258 of the total frequencies along all incoming edges to the
1259 thread path (path_in_freq). If we had a joiner, it would have
1260 been updated at the end of that handling to the edge frequency
1261 along the duplicated joiner path edge. */
1262 update_profile (epath
, NULL
, path_out_count
, path_out_count
,
1266 /* Increment the index into the duplicated path when we processed
1267 a duplicated block. */
1268 if ((*path
)[i
]->type
== EDGE_COPY_SRC_JOINER_BLOCK
1269 || (*path
)[i
]->type
== EDGE_COPY_SRC_BLOCK
)
1275 /* Now walk orig blocks and update their probabilities, since the
1276 counts and freqs should be updated properly by above loop. */
1277 for (unsigned int i
= 1; i
< path
->length (); i
++)
1279 edge epath
= (*path
)[i
]->e
;
1280 recompute_probabilities (epath
->src
);
1283 /* Done with all profile and frequency updates, clear counts if they
1285 if (do_freqs_to_counts
)
1286 clear_counts_path (rd
);
1289 /* Hash table traversal callback routine to create duplicate blocks. */
1292 ssa_create_duplicates (struct redirection_data
**slot
,
1293 ssa_local_info_t
*local_info
)
1295 struct redirection_data
*rd
= *slot
;
1297 /* The second duplicated block in a jump threading path is specific
1298 to the path. So it gets stored in RD rather than in LOCAL_DATA.
1300 Each time we're called, we have to look through the path and see
1301 if a second block needs to be duplicated.
1303 Note the search starts with the third edge on the path. The first
1304 edge is the incoming edge, the second edge always has its source
1305 duplicated. Thus we start our search with the third edge. */
1306 vec
<jump_thread_edge
*> *path
= rd
->path
;
1307 for (unsigned int i
= 2; i
< path
->length (); i
++)
1309 if ((*path
)[i
]->type
== EDGE_COPY_SRC_BLOCK
1310 || (*path
)[i
]->type
== EDGE_COPY_SRC_JOINER_BLOCK
)
1312 create_block_for_threading ((*path
)[i
]->e
->src
, rd
, 1,
1313 &local_info
->duplicate_blocks
);
1318 /* Create a template block if we have not done so already. Otherwise
1319 use the template to create a new block. */
1320 if (local_info
->template_block
== NULL
)
1322 create_block_for_threading ((*path
)[1]->e
->src
, rd
, 0,
1323 &local_info
->duplicate_blocks
);
1324 local_info
->template_block
= rd
->dup_blocks
[0];
1326 /* We do not create any outgoing edges for the template. We will
1327 take care of that in a later traversal. That way we do not
1328 create edges that are going to just be deleted. */
1332 create_block_for_threading (local_info
->template_block
, rd
, 0,
1333 &local_info
->duplicate_blocks
);
1335 /* Go ahead and wire up outgoing edges and update PHIs for the duplicate
1337 ssa_fix_duplicate_block_edges (rd
, local_info
);
1340 /* Keep walking the hash table. */
1344 /* We did not create any outgoing edges for the template block during
1345 block creation. This hash table traversal callback creates the
1346 outgoing edge for the template block. */
1349 ssa_fixup_template_block (struct redirection_data
**slot
,
1350 ssa_local_info_t
*local_info
)
1352 struct redirection_data
*rd
= *slot
;
1354 /* If this is the template block halt the traversal after updating
1357 If we were threading through an joiner block, then we want
1358 to keep its control statement and redirect an outgoing edge.
1359 Else we want to remove the control statement & edges, then create
1360 a new outgoing edge. In both cases we may need to update PHIs. */
1361 if (rd
->dup_blocks
[0] && rd
->dup_blocks
[0] == local_info
->template_block
)
1363 ssa_fix_duplicate_block_edges (rd
, local_info
);
1370 /* Hash table traversal callback to redirect each incoming edge
1371 associated with this hash table element to its new destination. */
1374 ssa_redirect_edges (struct redirection_data
**slot
,
1375 ssa_local_info_t
*local_info
)
1377 struct redirection_data
*rd
= *slot
;
1378 struct el
*next
, *el
;
1380 /* Walk over all the incoming edges associated associated with this
1381 hash table entry. */
1382 for (el
= rd
->incoming_edges
; el
; el
= next
)
1385 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
1387 /* Go ahead and free this element from the list. Doing this now
1388 avoids the need for another list walk when we destroy the hash
1393 thread_stats
.num_threaded_edges
++;
1395 if (rd
->dup_blocks
[0])
1399 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
1400 fprintf (dump_file
, " Threaded jump %d --> %d to %d\n",
1401 e
->src
->index
, e
->dest
->index
, rd
->dup_blocks
[0]->index
);
1403 /* If we redirect a loop latch edge cancel its loop. */
1404 if (e
->src
== e
->src
->loop_father
->latch
)
1405 mark_loop_for_removal (e
->src
->loop_father
);
1407 /* Redirect the incoming edge (possibly to the joiner block) to the
1408 appropriate duplicate block. */
1409 e2
= redirect_edge_and_branch (e
, rd
->dup_blocks
[0]);
1410 gcc_assert (e
== e2
);
1411 flush_pending_stmts (e2
);
1414 /* Go ahead and clear E->aux. It's not needed anymore and failure
1415 to clear it will cause all kinds of unpleasant problems later. */
1416 delete_jump_thread_path (path
);
1421 /* Indicate that we actually threaded one or more jumps. */
1422 if (rd
->incoming_edges
)
1423 local_info
->jumps_threaded
= true;
1428 /* Return true if this block has no executable statements other than
1429 a simple ctrl flow instruction. When the number of outgoing edges
1430 is one, this is equivalent to a "forwarder" block. */
1433 redirection_block_p (basic_block bb
)
1435 gimple_stmt_iterator gsi
;
1437 /* Advance to the first executable statement. */
1438 gsi
= gsi_start_bb (bb
);
1439 while (!gsi_end_p (gsi
)
1440 && (gimple_code (gsi_stmt (gsi
)) == GIMPLE_LABEL
1441 || is_gimple_debug (gsi_stmt (gsi
))
1442 || gimple_nop_p (gsi_stmt (gsi
))
1443 || gimple_clobber_p (gsi_stmt (gsi
))))
1446 /* Check if this is an empty block. */
1447 if (gsi_end_p (gsi
))
1450 /* Test that we've reached the terminating control statement. */
1451 return gsi_stmt (gsi
)
1452 && (gimple_code (gsi_stmt (gsi
)) == GIMPLE_COND
1453 || gimple_code (gsi_stmt (gsi
)) == GIMPLE_GOTO
1454 || gimple_code (gsi_stmt (gsi
)) == GIMPLE_SWITCH
);
1457 /* BB is a block which ends with a COND_EXPR or SWITCH_EXPR and when BB
1458 is reached via one or more specific incoming edges, we know which
1459 outgoing edge from BB will be traversed.
1461 We want to redirect those incoming edges to the target of the
1462 appropriate outgoing edge. Doing so avoids a conditional branch
1463 and may expose new optimization opportunities. Note that we have
1464 to update dominator tree and SSA graph after such changes.
1466 The key to keeping the SSA graph update manageable is to duplicate
1467 the side effects occurring in BB so that those side effects still
1468 occur on the paths which bypass BB after redirecting edges.
1470 We accomplish this by creating duplicates of BB and arranging for
1471 the duplicates to unconditionally pass control to one specific
1472 successor of BB. We then revector the incoming edges into BB to
1473 the appropriate duplicate of BB.
1475 If NOLOOP_ONLY is true, we only perform the threading as long as it
1476 does not affect the structure of the loops in a nontrivial way.
1478 If JOINERS is true, then thread through joiner blocks as well. */
1481 thread_block_1 (basic_block bb
, bool noloop_only
, bool joiners
)
1483 /* E is an incoming edge into BB that we may or may not want to
1484 redirect to a duplicate of BB. */
1487 ssa_local_info_t local_info
;
1489 local_info
.duplicate_blocks
= BITMAP_ALLOC (NULL
);
1491 /* To avoid scanning a linear array for the element we need we instead
1492 use a hash table. For normal code there should be no noticeable
1493 difference. However, if we have a block with a large number of
1494 incoming and outgoing edges such linear searches can get expensive. */
1496 = new hash_table
<struct redirection_data
> (EDGE_COUNT (bb
->succs
));
1498 /* Record each unique threaded destination into a hash table for
1499 efficient lookups. */
1500 FOR_EACH_EDGE (e
, ei
, bb
->preds
)
1505 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
1507 if (((*path
)[1]->type
== EDGE_COPY_SRC_JOINER_BLOCK
&& !joiners
)
1508 || ((*path
)[1]->type
== EDGE_COPY_SRC_BLOCK
&& joiners
))
1511 e2
= path
->last ()->e
;
1512 if (!e2
|| noloop_only
)
1514 /* If NOLOOP_ONLY is true, we only allow threading through the
1515 header of a loop to exit edges. */
1517 /* One case occurs when there was loop header buried in a jump
1518 threading path that crosses loop boundaries. We do not try
1519 and thread this elsewhere, so just cancel the jump threading
1520 request by clearing the AUX field now. */
1521 if ((bb
->loop_father
!= e2
->src
->loop_father
1522 && !loop_exit_edge_p (e2
->src
->loop_father
, e2
))
1523 || (e2
->src
->loop_father
!= e2
->dest
->loop_father
1524 && !loop_exit_edge_p (e2
->src
->loop_father
, e2
)))
1526 /* Since this case is not handled by our special code
1527 to thread through a loop header, we must explicitly
1528 cancel the threading request here. */
1529 delete_jump_thread_path (path
);
1534 /* Another case occurs when trying to thread through our
1535 own loop header, possibly from inside the loop. We will
1536 thread these later. */
1538 for (i
= 1; i
< path
->length (); i
++)
1540 if ((*path
)[i
]->e
->src
== bb
->loop_father
->header
1541 && (!loop_exit_edge_p (bb
->loop_father
, e2
)
1542 || (*path
)[1]->type
== EDGE_COPY_SRC_JOINER_BLOCK
))
1546 if (i
!= path
->length ())
1550 /* Insert the outgoing edge into the hash table if it is not
1551 already in the hash table. */
1552 lookup_redirection_data (e
, INSERT
);
1555 /* We do not update dominance info. */
1556 free_dominance_info (CDI_DOMINATORS
);
1558 /* We know we only thread through the loop header to loop exits.
1559 Let the basic block duplication hook know we are not creating
1560 a multiple entry loop. */
1562 && bb
== bb
->loop_father
->header
)
1563 set_loop_copy (bb
->loop_father
, loop_outer (bb
->loop_father
));
1565 /* Now create duplicates of BB.
1567 Note that for a block with a high outgoing degree we can waste
1568 a lot of time and memory creating and destroying useless edges.
1570 So we first duplicate BB and remove the control structure at the
1571 tail of the duplicate as well as all outgoing edges from the
1572 duplicate. We then use that duplicate block as a template for
1573 the rest of the duplicates. */
1574 local_info
.template_block
= NULL
;
1576 local_info
.jumps_threaded
= false;
1577 redirection_data
->traverse
<ssa_local_info_t
*, ssa_create_duplicates
>
1580 /* The template does not have an outgoing edge. Create that outgoing
1581 edge and update PHI nodes as the edge's target as necessary.
1583 We do this after creating all the duplicates to avoid creating
1584 unnecessary edges. */
1585 redirection_data
->traverse
<ssa_local_info_t
*, ssa_fixup_template_block
>
1588 /* The hash table traversals above created the duplicate blocks (and the
1589 statements within the duplicate blocks). This loop creates PHI nodes for
1590 the duplicated blocks and redirects the incoming edges into BB to reach
1591 the duplicates of BB. */
1592 redirection_data
->traverse
<ssa_local_info_t
*, ssa_redirect_edges
>
1595 /* Done with this block. Clear REDIRECTION_DATA. */
1596 delete redirection_data
;
1597 redirection_data
= NULL
;
1600 && bb
== bb
->loop_father
->header
)
1601 set_loop_copy (bb
->loop_father
, NULL
);
1603 BITMAP_FREE (local_info
.duplicate_blocks
);
1604 local_info
.duplicate_blocks
= NULL
;
1606 /* Indicate to our caller whether or not any jumps were threaded. */
1607 return local_info
.jumps_threaded
;
1610 /* Wrapper for thread_block_1 so that we can first handle jump
1611 thread paths which do not involve copying joiner blocks, then
1612 handle jump thread paths which have joiner blocks.
1614 By doing things this way we can be as aggressive as possible and
1615 not worry that copying a joiner block will create a jump threading
1619 thread_block (basic_block bb
, bool noloop_only
)
1622 retval
= thread_block_1 (bb
, noloop_only
, false);
1623 retval
|= thread_block_1 (bb
, noloop_only
, true);
1628 /* Threads edge E through E->dest to the edge THREAD_TARGET (E). Returns the
1629 copy of E->dest created during threading, or E->dest if it was not necessary
1630 to copy it (E is its single predecessor). */
1633 thread_single_edge (edge e
)
1635 basic_block bb
= e
->dest
;
1636 struct redirection_data rd
;
1637 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
1638 edge eto
= (*path
)[1]->e
;
1640 delete_jump_thread_path (path
);
1643 thread_stats
.num_threaded_edges
++;
1645 if (single_pred_p (bb
))
1647 /* If BB has just a single predecessor, we should only remove the
1648 control statements at its end, and successors except for ETO. */
1649 remove_ctrl_stmt_and_useless_edges (bb
, eto
->dest
);
1651 /* And fixup the flags on the single remaining edge. */
1652 eto
->flags
&= ~(EDGE_TRUE_VALUE
| EDGE_FALSE_VALUE
| EDGE_ABNORMAL
);
1653 eto
->flags
|= EDGE_FALLTHRU
;
1658 /* Otherwise, we need to create a copy. */
1659 if (e
->dest
== eto
->src
)
1660 update_bb_profile_for_threading (bb
, EDGE_FREQUENCY (e
), e
->count
, eto
);
1662 vec
<jump_thread_edge
*> *npath
= new vec
<jump_thread_edge
*> ();
1663 jump_thread_edge
*x
= new jump_thread_edge (e
, EDGE_START_JUMP_THREAD
);
1664 npath
->safe_push (x
);
1666 x
= new jump_thread_edge (eto
, EDGE_COPY_SRC_BLOCK
);
1667 npath
->safe_push (x
);
1670 create_block_for_threading (bb
, &rd
, 0, NULL
);
1671 remove_ctrl_stmt_and_useless_edges (rd
.dup_blocks
[0], NULL
);
1672 create_edge_and_update_destination_phis (&rd
, rd
.dup_blocks
[0], 0);
1674 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
1675 fprintf (dump_file
, " Threaded jump %d --> %d to %d\n",
1676 e
->src
->index
, e
->dest
->index
, rd
.dup_blocks
[0]->index
);
1678 rd
.dup_blocks
[0]->count
= e
->count
;
1679 rd
.dup_blocks
[0]->frequency
= EDGE_FREQUENCY (e
);
1680 single_succ_edge (rd
.dup_blocks
[0])->count
= e
->count
;
1681 redirect_edge_and_branch (e
, rd
.dup_blocks
[0]);
1682 flush_pending_stmts (e
);
1684 delete_jump_thread_path (npath
);
1685 return rd
.dup_blocks
[0];
1688 /* Callback for dfs_enumerate_from. Returns true if BB is different
1689 from STOP and DBDS_CE_STOP. */
1691 static basic_block dbds_ce_stop
;
1693 dbds_continue_enumeration_p (const_basic_block bb
, const void *stop
)
1695 return (bb
!= (const_basic_block
) stop
1696 && bb
!= dbds_ce_stop
);
1699 /* Evaluates the dominance relationship of latch of the LOOP and BB, and
1700 returns the state. */
1704 /* BB does not dominate latch of the LOOP. */
1705 DOMST_NONDOMINATING
,
1706 /* The LOOP is broken (there is no path from the header to its latch. */
1708 /* BB dominates the latch of the LOOP. */
1712 static enum bb_dom_status
1713 determine_bb_domination_status (struct loop
*loop
, basic_block bb
)
1715 basic_block
*bblocks
;
1716 unsigned nblocks
, i
;
1717 bool bb_reachable
= false;
1721 /* This function assumes BB is a successor of LOOP->header.
1722 If that is not the case return DOMST_NONDOMINATING which
1727 FOR_EACH_EDGE (e
, ei
, bb
->preds
)
1729 if (e
->src
== loop
->header
)
1737 return DOMST_NONDOMINATING
;
1740 if (bb
== loop
->latch
)
1741 return DOMST_DOMINATING
;
1743 /* Check that BB dominates LOOP->latch, and that it is back-reachable
1746 bblocks
= XCNEWVEC (basic_block
, loop
->num_nodes
);
1747 dbds_ce_stop
= loop
->header
;
1748 nblocks
= dfs_enumerate_from (loop
->latch
, 1, dbds_continue_enumeration_p
,
1749 bblocks
, loop
->num_nodes
, bb
);
1750 for (i
= 0; i
< nblocks
; i
++)
1751 FOR_EACH_EDGE (e
, ei
, bblocks
[i
]->preds
)
1753 if (e
->src
== loop
->header
)
1756 return DOMST_NONDOMINATING
;
1759 bb_reachable
= true;
1763 return (bb_reachable
? DOMST_DOMINATING
: DOMST_LOOP_BROKEN
);
1766 /* Return true if BB is part of the new pre-header that is created
1767 when threading the latch to DATA. */
1770 def_split_header_continue_p (const_basic_block bb
, const void *data
)
1772 const_basic_block new_header
= (const_basic_block
) data
;
1773 const struct loop
*l
;
1775 if (bb
== new_header
1776 || loop_depth (bb
->loop_father
) < loop_depth (new_header
->loop_father
))
1778 for (l
= bb
->loop_father
; l
; l
= loop_outer (l
))
1779 if (l
== new_header
->loop_father
)
1784 /* Thread jumps through the header of LOOP. Returns true if cfg changes.
1785 If MAY_PEEL_LOOP_HEADERS is false, we avoid threading from entry edges
1786 to the inside of the loop. */
1789 thread_through_loop_header (struct loop
*loop
, bool may_peel_loop_headers
)
1791 basic_block header
= loop
->header
;
1792 edge e
, tgt_edge
, latch
= loop_latch_edge (loop
);
1794 basic_block tgt_bb
, atgt_bb
;
1795 enum bb_dom_status domst
;
1797 /* We have already threaded through headers to exits, so all the threading
1798 requests now are to the inside of the loop. We need to avoid creating
1799 irreducible regions (i.e., loops with more than one entry block), and
1800 also loop with several latch edges, or new subloops of the loop (although
1801 there are cases where it might be appropriate, it is difficult to decide,
1802 and doing it wrongly may confuse other optimizers).
1804 We could handle more general cases here. However, the intention is to
1805 preserve some information about the loop, which is impossible if its
1806 structure changes significantly, in a way that is not well understood.
1807 Thus we only handle few important special cases, in which also updating
1808 of the loop-carried information should be feasible:
1810 1) Propagation of latch edge to a block that dominates the latch block
1811 of a loop. This aims to handle the following idiom:
1822 After threading the latch edge, this becomes
1833 The original header of the loop is moved out of it, and we may thread
1834 the remaining edges through it without further constraints.
1836 2) All entry edges are propagated to a single basic block that dominates
1837 the latch block of the loop. This aims to handle the following idiom
1838 (normally created for "for" loops):
1861 /* Threading through the header won't improve the code if the header has just
1863 if (single_succ_p (header
))
1866 /* If we threaded the latch using a joiner block, we cancel the
1867 threading opportunity out of an abundance of caution. However,
1868 still allow threading from outside to inside the loop. */
1871 vec
<jump_thread_edge
*> *path
= THREAD_PATH (latch
);
1872 if ((*path
)[1]->type
== EDGE_COPY_SRC_JOINER_BLOCK
)
1874 delete_jump_thread_path (path
);
1881 vec
<jump_thread_edge
*> *path
= THREAD_PATH (latch
);
1882 tgt_edge
= (*path
)[1]->e
;
1883 tgt_bb
= tgt_edge
->dest
;
1885 else if (!may_peel_loop_headers
1886 && !redirection_block_p (loop
->header
))
1892 FOR_EACH_EDGE (e
, ei
, header
->preds
)
1899 /* If latch is not threaded, and there is a header
1900 edge that is not threaded, we would create loop
1901 with multiple entries. */
1905 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
1907 if ((*path
)[1]->type
== EDGE_COPY_SRC_JOINER_BLOCK
)
1909 tgt_edge
= (*path
)[1]->e
;
1910 atgt_bb
= tgt_edge
->dest
;
1913 /* Two targets of threading would make us create loop
1914 with multiple entries. */
1915 else if (tgt_bb
!= atgt_bb
)
1921 /* There are no threading requests. */
1925 /* Redirecting to empty loop latch is useless. */
1926 if (tgt_bb
== loop
->latch
1927 && empty_block_p (loop
->latch
))
1931 /* The target block must dominate the loop latch, otherwise we would be
1932 creating a subloop. */
1933 domst
= determine_bb_domination_status (loop
, tgt_bb
);
1934 if (domst
== DOMST_NONDOMINATING
)
1936 if (domst
== DOMST_LOOP_BROKEN
)
1938 /* If the loop ceased to exist, mark it as such, and thread through its
1940 mark_loop_for_removal (loop
);
1941 return thread_block (header
, false);
1944 if (tgt_bb
->loop_father
->header
== tgt_bb
)
1946 /* If the target of the threading is a header of a subloop, we need
1947 to create a preheader for it, so that the headers of the two loops
1949 if (EDGE_COUNT (tgt_bb
->preds
) > 2)
1951 tgt_bb
= create_preheader (tgt_bb
->loop_father
, 0);
1952 gcc_assert (tgt_bb
!= NULL
);
1955 tgt_bb
= split_edge (tgt_edge
);
1960 basic_block
*bblocks
;
1961 unsigned nblocks
, i
;
1963 /* First handle the case latch edge is redirected. We are copying
1964 the loop header but not creating a multiple entry loop. Make the
1965 cfg manipulation code aware of that fact. */
1966 set_loop_copy (loop
, loop
);
1967 loop
->latch
= thread_single_edge (latch
);
1968 set_loop_copy (loop
, NULL
);
1969 gcc_assert (single_succ (loop
->latch
) == tgt_bb
);
1970 loop
->header
= tgt_bb
;
1972 /* Remove the new pre-header blocks from our loop. */
1973 bblocks
= XCNEWVEC (basic_block
, loop
->num_nodes
);
1974 nblocks
= dfs_enumerate_from (header
, 0, def_split_header_continue_p
,
1975 bblocks
, loop
->num_nodes
, tgt_bb
);
1976 for (i
= 0; i
< nblocks
; i
++)
1977 if (bblocks
[i
]->loop_father
== loop
)
1979 remove_bb_from_loops (bblocks
[i
]);
1980 add_bb_to_loop (bblocks
[i
], loop_outer (loop
));
1984 /* If the new header has multiple latches mark it so. */
1985 FOR_EACH_EDGE (e
, ei
, loop
->header
->preds
)
1986 if (e
->src
->loop_father
== loop
1987 && e
->src
!= loop
->latch
)
1990 loops_state_set (LOOPS_MAY_HAVE_MULTIPLE_LATCHES
);
1993 /* Cancel remaining threading requests that would make the
1994 loop a multiple entry loop. */
1995 FOR_EACH_EDGE (e
, ei
, header
->preds
)
2002 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
2003 e2
= path
->last ()->e
;
2005 if (e
->src
->loop_father
!= e2
->dest
->loop_father
2006 && e2
->dest
!= loop
->header
)
2008 delete_jump_thread_path (path
);
2013 /* Thread the remaining edges through the former header. */
2014 thread_block (header
, false);
2018 basic_block new_preheader
;
2020 /* Now consider the case entry edges are redirected to the new entry
2021 block. Remember one entry edge, so that we can find the new
2022 preheader (its destination after threading). */
2023 FOR_EACH_EDGE (e
, ei
, header
->preds
)
2029 /* The duplicate of the header is the new preheader of the loop. Ensure
2030 that it is placed correctly in the loop hierarchy. */
2031 set_loop_copy (loop
, loop_outer (loop
));
2033 thread_block (header
, false);
2034 set_loop_copy (loop
, NULL
);
2035 new_preheader
= e
->dest
;
2037 /* Create the new latch block. This is always necessary, as the latch
2038 must have only a single successor, but the original header had at
2039 least two successors. */
2041 mfb_kj_edge
= single_succ_edge (new_preheader
);
2042 loop
->header
= mfb_kj_edge
->dest
;
2043 latch
= make_forwarder_block (tgt_bb
, mfb_keep_just
, NULL
);
2044 loop
->header
= latch
->dest
;
2045 loop
->latch
= latch
->src
;
2051 /* We failed to thread anything. Cancel the requests. */
2052 FOR_EACH_EDGE (e
, ei
, header
->preds
)
2054 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
2058 delete_jump_thread_path (path
);
2065 /* E1 and E2 are edges into the same basic block. Return TRUE if the
2066 PHI arguments associated with those edges are equal or there are no
2067 PHI arguments, otherwise return FALSE. */
2070 phi_args_equal_on_edges (edge e1
, edge e2
)
2073 int indx1
= e1
->dest_idx
;
2074 int indx2
= e2
->dest_idx
;
2076 for (gsi
= gsi_start_phis (e1
->dest
); !gsi_end_p (gsi
); gsi_next (&gsi
))
2078 gphi
*phi
= gsi
.phi ();
2080 if (!operand_equal_p (gimple_phi_arg_def (phi
, indx1
),
2081 gimple_phi_arg_def (phi
, indx2
), 0))
2087 /* Walk through the registered jump threads and convert them into a
2088 form convenient for this pass.
2090 Any block which has incoming edges threaded to outgoing edges
2091 will have its entry in THREADED_BLOCK set.
2093 Any threaded edge will have its new outgoing edge stored in the
2094 original edge's AUX field.
2096 This form avoids the need to walk all the edges in the CFG to
2097 discover blocks which need processing and avoids unnecessary
2098 hash table lookups to map from threaded edge to new target. */
2101 mark_threaded_blocks (bitmap threaded_blocks
)
2105 bitmap tmp
= BITMAP_ALLOC (NULL
);
2110 /* It is possible to have jump threads in which one is a subpath
2111 of the other. ie, (A, B), (B, C), (C, D) where B is a joiner
2112 block and (B, C), (C, D) where no joiner block exists.
2114 When this occurs ignore the jump thread request with the joiner
2115 block. It's totally subsumed by the simpler jump thread request.
2117 This results in less block copying, simpler CFGs. More importantly,
2118 when we duplicate the joiner block, B, in this case we will create
2119 a new threading opportunity that we wouldn't be able to optimize
2120 until the next jump threading iteration.
2122 So first convert the jump thread requests which do not require a
2124 for (i
= 0; i
< paths
.length (); i
++)
2126 vec
<jump_thread_edge
*> *path
= paths
[i
];
2128 if ((*path
)[1]->type
!= EDGE_COPY_SRC_JOINER_BLOCK
)
2130 edge e
= (*path
)[0]->e
;
2131 e
->aux
= (void *)path
;
2132 bitmap_set_bit (tmp
, e
->dest
->index
);
2136 /* Now iterate again, converting cases where we want to thread
2137 through a joiner block, but only if no other edge on the path
2138 already has a jump thread attached to it. We do this in two passes,
2139 to avoid situations where the order in the paths vec can hide overlapping
2140 threads (the path is recorded on the incoming edge, so we would miss
2141 cases where the second path starts at a downstream edge on the same
2142 path). First record all joiner paths, deleting any in the unexpected
2143 case where there is already a path for that incoming edge. */
2144 for (i
= 0; i
< paths
.length (); i
++)
2146 vec
<jump_thread_edge
*> *path
= paths
[i
];
2148 if ((*path
)[1]->type
== EDGE_COPY_SRC_JOINER_BLOCK
)
2150 /* Attach the path to the starting edge if none is yet recorded. */
2151 if ((*path
)[0]->e
->aux
== NULL
)
2153 (*path
)[0]->e
->aux
= path
;
2157 paths
.unordered_remove (i
);
2158 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
2159 dump_jump_thread_path (dump_file
, *path
, false);
2160 delete_jump_thread_path (path
);
2164 /* Second, look for paths that have any other jump thread attached to
2165 them, and either finish converting them or cancel them. */
2166 for (i
= 0; i
< paths
.length (); i
++)
2168 vec
<jump_thread_edge
*> *path
= paths
[i
];
2169 edge e
= (*path
)[0]->e
;
2171 if ((*path
)[1]->type
== EDGE_COPY_SRC_JOINER_BLOCK
&& e
->aux
== path
)
2174 for (j
= 1; j
< path
->length (); j
++)
2175 if ((*path
)[j
]->e
->aux
!= NULL
)
2178 /* If we iterated through the entire path without exiting the loop,
2179 then we are good to go, record it. */
2180 if (j
== path
->length ())
2181 bitmap_set_bit (tmp
, e
->dest
->index
);
2185 paths
.unordered_remove (i
);
2186 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
2187 dump_jump_thread_path (dump_file
, *path
, false);
2188 delete_jump_thread_path (path
);
2193 /* If optimizing for size, only thread through block if we don't have
2194 to duplicate it or it's an otherwise empty redirection block. */
2195 if (optimize_function_for_size_p (cfun
))
2197 EXECUTE_IF_SET_IN_BITMAP (tmp
, 0, i
, bi
)
2199 bb
= BASIC_BLOCK_FOR_FN (cfun
, i
);
2200 if (EDGE_COUNT (bb
->preds
) > 1
2201 && !redirection_block_p (bb
))
2203 FOR_EACH_EDGE (e
, ei
, bb
->preds
)
2207 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
2208 delete_jump_thread_path (path
);
2214 bitmap_set_bit (threaded_blocks
, i
);
2218 bitmap_copy (threaded_blocks
, tmp
);
2220 /* Look for jump threading paths which cross multiple loop headers.
2222 The code to thread through loop headers will change the CFG in ways
2223 that break assumptions made by the loop optimization code.
2225 We don't want to blindly cancel the requests. We can instead do better
2226 by trimming off the end of the jump thread path. */
2227 EXECUTE_IF_SET_IN_BITMAP (tmp
, 0, i
, bi
)
2229 basic_block bb
= BASIC_BLOCK_FOR_FN (cfun
, i
);
2230 FOR_EACH_EDGE (e
, ei
, bb
->preds
)
2234 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
2236 for (unsigned int i
= 0, crossed_headers
= 0;
2237 i
< path
->length ();
2240 basic_block dest
= (*path
)[i
]->e
->dest
;
2241 crossed_headers
+= (dest
== dest
->loop_father
->header
);
2242 if (crossed_headers
> 1)
2244 /* Trim from entry I onwards. */
2245 for (unsigned int j
= i
; j
< path
->length (); j
++)
2249 /* Now that we've truncated the path, make sure
2250 what's left is still valid. We need at least
2251 two edges on the path and the last edge can not
2252 be a joiner. This should never happen, but let's
2254 if (path
->length () < 2
2255 || (path
->last ()->type
2256 == EDGE_COPY_SRC_JOINER_BLOCK
))
2258 delete_jump_thread_path (path
);
2268 /* If we have a joiner block (J) which has two successors S1 and S2 and
2269 we are threading though S1 and the final destination of the thread
2270 is S2, then we must verify that any PHI nodes in S2 have the same
2271 PHI arguments for the edge J->S2 and J->S1->...->S2.
2273 We used to detect this prior to registering the jump thread, but
2274 that prohibits propagation of edge equivalences into non-dominated
2275 PHI nodes as the equivalency test might occur before propagation.
2277 This must also occur after we truncate any jump threading paths
2278 as this scenario may only show up after truncation.
2280 This works for now, but will need improvement as part of the FSA
2283 Note since we've moved the thread request data to the edges,
2284 we have to iterate on those rather than the threaded_edges vector. */
2285 EXECUTE_IF_SET_IN_BITMAP (tmp
, 0, i
, bi
)
2287 bb
= BASIC_BLOCK_FOR_FN (cfun
, i
);
2288 FOR_EACH_EDGE (e
, ei
, bb
->preds
)
2292 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
2293 bool have_joiner
= ((*path
)[1]->type
== EDGE_COPY_SRC_JOINER_BLOCK
);
2297 basic_block joiner
= e
->dest
;
2298 edge final_edge
= path
->last ()->e
;
2299 basic_block final_dest
= final_edge
->dest
;
2300 edge e2
= find_edge (joiner
, final_dest
);
2302 if (e2
&& !phi_args_equal_on_edges (e2
, final_edge
))
2304 delete_jump_thread_path (path
);
2316 /* Return TRUE if BB ends with a switch statement or a computed goto.
2317 Otherwise return false. */
2319 bb_ends_with_multiway_branch (basic_block bb ATTRIBUTE_UNUSED
)
2321 gimple stmt
= last_stmt (bb
);
2322 if (stmt
&& gimple_code (stmt
) == GIMPLE_SWITCH
)
2324 if (stmt
&& gimple_code (stmt
) == GIMPLE_GOTO
2325 && TREE_CODE (gimple_goto_dest (stmt
)) == SSA_NAME
)
2330 /* Verify that the REGION is a valid jump thread. A jump thread is a special
2331 case of SEME Single Entry Multiple Exits region in which all nodes in the
2332 REGION have exactly one incoming edge. The only exception is the first block
2333 that may not have been connected to the rest of the cfg yet. */
2336 verify_jump_thread (basic_block
*region
, unsigned n_region
)
2338 for (unsigned i
= 0; i
< n_region
; i
++)
2339 gcc_assert (EDGE_COUNT (region
[i
]->preds
) <= 1);
2342 /* Return true when BB is one of the first N items in BBS. */
2345 bb_in_bbs (basic_block bb
, basic_block
*bbs
, int n
)
2347 for (int i
= 0; i
< n
; i
++)
2354 /* Duplicates a jump-thread path of N_REGION basic blocks.
2355 The ENTRY edge is redirected to the duplicate of the region.
2357 Remove the last conditional statement in the last basic block in the REGION,
2358 and create a single fallthru edge pointing to the same destination as the
2361 The new basic blocks are stored to REGION_COPY in the same order as they had
2362 in REGION, provided that REGION_COPY is not NULL.
2364 Returns false if it is unable to copy the region, true otherwise. */
2367 duplicate_thread_path (edge entry
, edge exit
,
2368 basic_block
*region
, unsigned n_region
,
2369 basic_block
*region_copy
)
2372 bool free_region_copy
= false;
2373 struct loop
*loop
= entry
->dest
->loop_father
;
2376 int total_freq
= 0, entry_freq
= 0;
2377 gcov_type total_count
= 0, entry_count
= 0;
2379 if (!can_copy_bbs_p (region
, n_region
))
2382 /* Some sanity checking. Note that we do not check for all possible
2383 missuses of the functions. I.e. if you ask to copy something weird,
2384 it will work, but the state of structures probably will not be
2386 for (i
= 0; i
< n_region
; i
++)
2388 /* We do not handle subloops, i.e. all the blocks must belong to the
2390 if (region
[i
]->loop_father
!= loop
)
2394 initialize_original_copy_tables ();
2396 set_loop_copy (loop
, loop
);
2400 region_copy
= XNEWVEC (basic_block
, n_region
);
2401 free_region_copy
= true;
2404 if (entry
->dest
->count
)
2406 total_count
= entry
->dest
->count
;
2407 entry_count
= entry
->count
;
2408 /* Fix up corner cases, to avoid division by zero or creation of negative
2410 if (entry_count
> total_count
)
2411 entry_count
= total_count
;
2415 total_freq
= entry
->dest
->frequency
;
2416 entry_freq
= EDGE_FREQUENCY (entry
);
2417 /* Fix up corner cases, to avoid division by zero or creation of negative
2419 if (total_freq
== 0)
2421 else if (entry_freq
> total_freq
)
2422 entry_freq
= total_freq
;
2425 copy_bbs (region
, n_region
, region_copy
, &exit
, 1, &exit_copy
, loop
,
2426 split_edge_bb_loc (entry
), false);
2428 /* Fix up: copy_bbs redirects all edges pointing to copied blocks. The
2429 following code ensures that all the edges exiting the jump-thread path are
2430 redirected back to the original code: these edges are exceptions
2431 invalidating the property that is propagated by executing all the blocks of
2432 the jump-thread path in order. */
2434 for (i
= 0; i
< n_region
; i
++)
2438 basic_block bb
= region_copy
[i
];
2440 if (single_succ_p (bb
))
2442 /* Make sure the successor is the next node in the path. */
2443 gcc_assert (i
+ 1 == n_region
2444 || region_copy
[i
+ 1] == single_succ_edge (bb
)->dest
);
2448 /* Special case the last block on the path: make sure that it does not
2449 jump back on the copied path. */
2450 if (i
+ 1 == n_region
)
2452 FOR_EACH_EDGE (e
, ei
, bb
->succs
)
2453 if (bb_in_bbs (e
->dest
, region_copy
, n_region
- 1))
2455 basic_block orig
= get_bb_original (e
->dest
);
2457 redirect_edge_and_branch_force (e
, orig
);
2462 /* Redirect all other edges jumping to non-adjacent blocks back to the
2464 FOR_EACH_EDGE (e
, ei
, bb
->succs
)
2465 if (region_copy
[i
+ 1] != e
->dest
)
2467 basic_block orig
= get_bb_original (e
->dest
);
2469 redirect_edge_and_branch_force (e
, orig
);
2475 scale_bbs_frequencies_gcov_type (region
, n_region
,
2476 total_count
- entry_count
,
2478 scale_bbs_frequencies_gcov_type (region_copy
, n_region
, entry_count
,
2483 scale_bbs_frequencies_int (region
, n_region
, total_freq
- entry_freq
,
2485 scale_bbs_frequencies_int (region_copy
, n_region
, entry_freq
, total_freq
);
2488 #ifdef ENABLE_CHECKING
2489 verify_jump_thread (region_copy
, n_region
);
2492 /* Remove the last branch in the jump thread path. */
2493 remove_ctrl_stmt_and_useless_edges (region_copy
[n_region
- 1], exit
->dest
);
2494 edge e
= make_edge (region_copy
[n_region
- 1], exit
->dest
, EDGE_FALLTHRU
);
2497 rescan_loop_exit (e
, true, false);
2498 e
->probability
= REG_BR_PROB_BASE
;
2499 e
->count
= region_copy
[n_region
- 1]->count
;
2502 /* Redirect the entry and add the phi node arguments. */
2503 if (entry
->dest
== loop
->header
)
2504 mark_loop_for_removal (loop
);
2505 redirected
= redirect_edge_and_branch (entry
, get_bb_copy (entry
->dest
));
2506 gcc_assert (redirected
!= NULL
);
2507 flush_pending_stmts (entry
);
2509 /* Add the other PHI node arguments. */
2510 add_phi_args_after_copy (region_copy
, n_region
, NULL
);
2512 if (free_region_copy
)
2515 free_original_copy_tables ();
2519 /* Return true when PATH is a valid jump-thread path. */
2522 valid_jump_thread_path (vec
<jump_thread_edge
*> *path
)
2524 unsigned len
= path
->length ();
2526 /* Check that the path is connected. */
2527 for (unsigned int j
= 0; j
< len
- 1; j
++)
2528 if ((*path
)[j
]->e
->dest
!= (*path
)[j
+1]->e
->src
)
2534 /* Walk through all blocks and thread incoming edges to the appropriate
2535 outgoing edge for each edge pair recorded in THREADED_EDGES.
2537 It is the caller's responsibility to fix the dominance information
2538 and rewrite duplicated SSA_NAMEs back into SSA form.
2540 If MAY_PEEL_LOOP_HEADERS is false, we avoid threading edges through
2541 loop headers if it does not simplify the loop.
2543 Returns true if one or more edges were threaded, false otherwise. */
2546 thread_through_all_blocks (bool may_peel_loop_headers
)
2548 bool retval
= false;
2551 bitmap threaded_blocks
;
2554 if (!paths
.exists ())
2557 threaded_blocks
= BITMAP_ALLOC (NULL
);
2558 memset (&thread_stats
, 0, sizeof (thread_stats
));
2560 /* Jump-thread all FSM threads before other jump-threads. */
2561 for (i
= 0; i
< paths
.length ();)
2563 vec
<jump_thread_edge
*> *path
= paths
[i
];
2564 edge entry
= (*path
)[0]->e
;
2566 /* Only code-generate FSM jump-threads in this loop. */
2567 if ((*path
)[0]->type
!= EDGE_FSM_THREAD
)
2573 /* Do not jump-thread twice from the same block. */
2574 if (bitmap_bit_p (threaded_blocks
, entry
->src
->index
)
2575 /* Verify that the jump thread path is still valid: a
2576 previous jump-thread may have changed the CFG, and
2577 invalidated the current path. */
2578 || !valid_jump_thread_path (path
))
2580 /* Remove invalid FSM jump-thread paths. */
2581 delete_jump_thread_path (path
);
2582 paths
.unordered_remove (i
);
2586 unsigned len
= path
->length ();
2587 edge exit
= (*path
)[len
- 1]->e
;
2588 basic_block
*region
= XNEWVEC (basic_block
, len
- 1);
2590 for (unsigned int j
= 0; j
< len
- 1; j
++)
2591 region
[j
] = (*path
)[j
]->e
->dest
;
2593 if (duplicate_thread_path (entry
, exit
, region
, len
- 1, NULL
))
2595 /* We do not update dominance info. */
2596 free_dominance_info (CDI_DOMINATORS
);
2597 bitmap_set_bit (threaded_blocks
, entry
->src
->index
);
2601 delete_jump_thread_path (path
);
2602 paths
.unordered_remove (i
);
2605 /* Remove from PATHS all the jump-threads starting with an edge already
2607 for (i
= 0; i
< paths
.length ();)
2609 vec
<jump_thread_edge
*> *path
= paths
[i
];
2610 edge entry
= (*path
)[0]->e
;
2612 /* Do not jump-thread twice from the same block. */
2613 if (bitmap_bit_p (threaded_blocks
, entry
->src
->index
))
2615 delete_jump_thread_path (path
);
2616 paths
.unordered_remove (i
);
2622 bitmap_clear (threaded_blocks
);
2624 mark_threaded_blocks (threaded_blocks
);
2626 initialize_original_copy_tables ();
2628 /* First perform the threading requests that do not affect
2630 EXECUTE_IF_SET_IN_BITMAP (threaded_blocks
, 0, i
, bi
)
2632 basic_block bb
= BASIC_BLOCK_FOR_FN (cfun
, i
);
2634 if (EDGE_COUNT (bb
->preds
) > 0)
2635 retval
|= thread_block (bb
, true);
2638 /* Then perform the threading through loop headers. We start with the
2639 innermost loop, so that the changes in cfg we perform won't affect
2640 further threading. */
2641 FOR_EACH_LOOP (loop
, LI_FROM_INNERMOST
)
2644 || !bitmap_bit_p (threaded_blocks
, loop
->header
->index
))
2647 retval
|= thread_through_loop_header (loop
, may_peel_loop_headers
);
2650 /* Any jump threading paths that are still attached to edges at this
2651 point must be one of two cases.
2653 First, we could have a jump threading path which went from outside
2654 a loop to inside a loop that was ignored because a prior jump thread
2655 across a backedge was realized (which indirectly causes the loop
2656 above to ignore the latter thread). We can detect these because the
2657 loop structures will be different and we do not currently try to
2660 Second, we could be threading across a backedge to a point within the
2661 same loop. This occurrs for the FSA/FSM optimization and we would
2662 like to optimize it. However, we have to be very careful as this
2663 may completely scramble the loop structures, with the result being
2664 irreducible loops causing us to throw away our loop structure.
2666 As a compromise for the latter case, if the thread path ends in
2667 a block where the last statement is a multiway branch, then go
2668 ahead and thread it, else ignore it. */
2671 FOR_EACH_BB_FN (bb
, cfun
)
2673 /* If we do end up threading here, we can remove elements from
2674 BB->preds. Thus we can not use the FOR_EACH_EDGE iterator. */
2675 for (edge_iterator ei
= ei_start (bb
->preds
);
2676 (e
= ei_safe_edge (ei
));)
2679 vec
<jump_thread_edge
*> *path
= THREAD_PATH (e
);
2681 /* Case 1, threading from outside to inside the loop
2682 after we'd already threaded through the header. */
2683 if ((*path
)[0]->e
->dest
->loop_father
2684 != path
->last ()->e
->src
->loop_father
)
2686 delete_jump_thread_path (path
);
2690 else if (bb_ends_with_multiway_branch (path
->last ()->e
->src
))
2692 /* The code to thread through loop headers may have
2693 split a block with jump threads attached to it.
2695 We can identify this with a disjoint jump threading
2696 path. If found, just remove it. */
2697 for (unsigned int i
= 0; i
< path
->length () - 1; i
++)
2698 if ((*path
)[i
]->e
->dest
!= (*path
)[i
+ 1]->e
->src
)
2700 delete_jump_thread_path (path
);
2706 /* Our path is still valid, thread it. */
2709 if (thread_block ((*path
)[0]->e
->dest
, false))
2713 delete_jump_thread_path (path
);
2721 delete_jump_thread_path (path
);
2730 statistics_counter_event (cfun
, "Jumps threaded",
2731 thread_stats
.num_threaded_edges
);
2733 free_original_copy_tables ();
2735 BITMAP_FREE (threaded_blocks
);
2736 threaded_blocks
= NULL
;
2740 loops_state_set (LOOPS_NEED_FIXUP
);
2745 /* Delete the jump threading path PATH. We have to explcitly delete
2746 each entry in the vector, then the container. */
2749 delete_jump_thread_path (vec
<jump_thread_edge
*> *path
)
2751 for (unsigned int i
= 0; i
< path
->length (); i
++)
2757 /* Register a jump threading opportunity. We queue up all the jump
2758 threading opportunities discovered by a pass and update the CFG
2759 and SSA form all at once.
2761 E is the edge we can thread, E2 is the new target edge, i.e., we
2762 are effectively recording that E->dest can be changed to E2->dest
2763 after fixing the SSA graph. */
2766 register_jump_thread (vec
<jump_thread_edge
*> *path
)
2768 if (!dbg_cnt (registered_jump_thread
))
2770 delete_jump_thread_path (path
);
2774 /* First make sure there are no NULL outgoing edges on the jump threading
2775 path. That can happen for jumping to a constant address. */
2776 for (unsigned int i
= 0; i
< path
->length (); i
++)
2777 if ((*path
)[i
]->e
== NULL
)
2779 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
2782 "Found NULL edge in jump threading path. Cancelling jump thread:\n");
2783 dump_jump_thread_path (dump_file
, *path
, false);
2786 delete_jump_thread_path (path
);
2790 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
2791 dump_jump_thread_path (dump_file
, *path
, true);
2793 if (!paths
.exists ())
2796 paths
.safe_push (path
);