1 /* Callgraph transformations to handle inlining
2 Copyright (C) 2003-2017 Free Software Foundation, Inc.
3 Contributed by Jan Hubicka
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify it under
8 the terms of the GNU General Public License as published by the Free
9 Software Foundation; either version 3, or (at your option) any later
12 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
21 /* The inline decisions are stored in callgraph in "inline plan" and
24 To mark given call inline, use inline_call function.
25 The function marks the edge inlinable and, if necessary, produces
26 virtual clone in the callgraph representing the new copy of callee's
29 The inline plan is applied on given function body by inline_transform. */
33 #include "coretypes.h"
37 #include "alloc-pool.h"
38 #include "tree-pass.h"
41 #include "symbol-summary.h"
44 #include "ipa-fnsummary.h"
45 #include "ipa-inline.h"
46 #include "tree-inline.h"
49 #include "basic-block.h"
52 int nfunctions_inlined
;
54 /* Scale counts of NODE edges by NUM/DEN. */
57 update_noncloned_counts (struct cgraph_node
*node
,
58 profile_count num
, profile_count den
)
60 struct cgraph_edge
*e
;
62 profile_count::adjust_for_ipa_scaling (&num
, &den
);
64 for (e
= node
->callees
; e
; e
= e
->next_callee
)
66 if (!e
->inline_failed
)
67 update_noncloned_counts (e
->callee
, num
, den
);
68 e
->count
= e
->count
.apply_scale (num
, den
);
70 for (e
= node
->indirect_calls
; e
; e
= e
->next_callee
)
71 e
->count
= e
->count
.apply_scale (num
, den
);
72 node
->count
= node
->count
.apply_scale (num
, den
);
75 /* We removed or are going to remove the last call to NODE.
76 Return true if we can and want proactively remove the NODE now.
77 This is important to do, since we want inliner to know when offline
78 copy of function was removed. */
81 can_remove_node_now_p_1 (struct cgraph_node
*node
, struct cgraph_edge
*e
)
85 FOR_EACH_ALIAS (node
, ref
)
87 cgraph_node
*alias
= dyn_cast
<cgraph_node
*> (ref
->referring
);
88 if ((alias
->callers
&& alias
->callers
!= e
)
89 || !can_remove_node_now_p_1 (alias
, e
))
92 /* FIXME: When address is taken of DECL_EXTERNAL function we still
93 can remove its offline copy, but we would need to keep unanalyzed node in
94 the callgraph so references can point to it.
96 Also for comdat group we can ignore references inside a group as we
97 want to prove the group as a whole to be dead. */
98 return (!node
->address_taken
99 && node
->can_remove_if_no_direct_calls_and_refs_p ()
100 /* Inlining might enable more devirtualizing, so we want to remove
101 those only after all devirtualizable virtual calls are processed.
102 Lacking may edges in callgraph we just preserve them post
104 && (!DECL_VIRTUAL_P (node
->decl
)
105 || !opt_for_fn (node
->decl
, flag_devirtualize
))
106 /* During early inlining some unanalyzed cgraph nodes might be in the
107 callgraph and they might reffer the function in question. */
108 && !cgraph_new_nodes
.exists ());
111 /* We are going to eliminate last direct call to NODE (or alias of it) via edge E.
112 Verify that the NODE can be removed from unit and if it is contained in comdat
113 group that the whole comdat group is removable. */
116 can_remove_node_now_p (struct cgraph_node
*node
, struct cgraph_edge
*e
)
118 struct cgraph_node
*next
;
119 if (!can_remove_node_now_p_1 (node
, e
))
122 /* When we see same comdat group, we need to be sure that all
123 items can be removed. */
124 if (!node
->same_comdat_group
|| !node
->externally_visible
)
126 for (next
= dyn_cast
<cgraph_node
*> (node
->same_comdat_group
);
127 next
!= node
; next
= dyn_cast
<cgraph_node
*> (next
->same_comdat_group
))
131 if ((next
->callers
&& next
->callers
!= e
)
132 || !can_remove_node_now_p_1 (next
, e
))
138 /* Return true if NODE is a master clone with non-inline clones. */
141 master_clone_with_noninline_clones_p (struct cgraph_node
*node
)
146 for (struct cgraph_node
*n
= node
->clones
; n
; n
= n
->next_sibling_clone
)
147 if (n
->decl
!= node
->decl
)
153 /* E is expected to be an edge being inlined. Clone destination node of
154 the edge and redirect it to the new clone.
155 DUPLICATE is used for bookkeeping on whether we are actually creating new
156 clones or re-using node originally representing out-of-line function call.
157 By default the offline copy is removed, when it appears dead after inlining.
158 UPDATE_ORIGINAL prevents this transformation.
159 If OVERALL_SIZE is non-NULL, the size is updated to reflect the
163 clone_inlined_nodes (struct cgraph_edge
*e
, bool duplicate
,
164 bool update_original
, int *overall_size
)
166 struct cgraph_node
*inlining_into
;
167 struct cgraph_edge
*next
;
169 if (e
->caller
->global
.inlined_to
)
170 inlining_into
= e
->caller
->global
.inlined_to
;
172 inlining_into
= e
->caller
;
176 /* We may eliminate the need for out-of-line copy to be output.
177 In that case just go ahead and re-use it. This is not just an
178 memory optimization. Making offline copy of fuction disappear
179 from the program will improve future decisions on inlining. */
180 if (!e
->callee
->callers
->next_caller
181 /* Recursive inlining never wants the master clone to
184 && can_remove_node_now_p (e
->callee
, e
)
185 /* We cannot overwrite a master clone with non-inline clones
186 until after these clones are materialized. */
187 && !master_clone_with_noninline_clones_p (e
->callee
))
189 /* TODO: When callee is in a comdat group, we could remove all of it,
190 including all inline clones inlined into it. That would however
191 need small function inlining to register edge removal hook to
192 maintain the priority queue.
194 For now we keep the ohter functions in the group in program until
195 cgraph_remove_unreachable_functions gets rid of them. */
196 gcc_assert (!e
->callee
->global
.inlined_to
);
197 e
->callee
->remove_from_same_comdat_group ();
198 if (e
->callee
->definition
199 && inline_account_function_p (e
->callee
))
201 gcc_assert (!e
->callee
->alias
);
203 *overall_size
-= ipa_fn_summaries
->get (e
->callee
)->size
;
204 nfunctions_inlined
++;
207 e
->callee
->externally_visible
= false;
208 update_noncloned_counts (e
->callee
, e
->count
, e
->callee
->count
);
210 dump_callgraph_transformation (e
->callee
, inlining_into
,
215 struct cgraph_node
*n
;
217 n
= e
->callee
->create_clone (e
->callee
->decl
,
219 update_original
, vNULL
, true,
222 n
->used_as_abstract_origin
= e
->callee
->used_as_abstract_origin
;
223 e
->redirect_callee (n
);
227 e
->callee
->remove_from_same_comdat_group ();
229 e
->callee
->global
.inlined_to
= inlining_into
;
231 /* Recursively clone all bodies. */
232 for (e
= e
->callee
->callees
; e
; e
= next
)
234 next
= e
->next_callee
;
235 if (!e
->inline_failed
)
236 clone_inlined_nodes (e
, duplicate
, update_original
, overall_size
);
240 /* Check all speculations in N and resolve them if they seems useless. */
243 check_speculations (cgraph_node
*n
)
245 bool speculation_removed
= false;
248 for (cgraph_edge
*e
= n
->callees
; e
; e
= next
)
250 next
= e
->next_callee
;
251 if (e
->speculative
&& !speculation_useful_p (e
, true))
253 e
->resolve_speculation (NULL
);
254 speculation_removed
= true;
256 else if (!e
->inline_failed
)
257 speculation_removed
|= check_speculations (e
->callee
);
259 return speculation_removed
;
262 /* Mark all call graph edges coming out of NODE and all nodes that have been
263 inlined to it as in_polymorphic_cdtor. */
266 mark_all_inlined_calls_cdtor (cgraph_node
*node
)
268 for (cgraph_edge
*cs
= node
->callees
; cs
; cs
= cs
->next_callee
)
270 cs
->in_polymorphic_cdtor
= true;
271 if (!cs
->inline_failed
)
272 mark_all_inlined_calls_cdtor (cs
->callee
);
274 for (cgraph_edge
*cs
= node
->indirect_calls
; cs
; cs
= cs
->next_callee
)
275 cs
->in_polymorphic_cdtor
= true;
279 /* Mark edge E as inlined and update callgraph accordingly. UPDATE_ORIGINAL
280 specify whether profile of original function should be updated. If any new
281 indirect edges are discovered in the process, add them to NEW_EDGES, unless
282 it is NULL. If UPDATE_OVERALL_SUMMARY is false, do not bother to recompute overall
283 size of caller after inlining. Caller is required to eventually do it via
284 ipa_update_overall_fn_summary.
285 If callee_removed is non-NULL, set it to true if we removed callee node.
287 Return true iff any new callgraph edges were discovered as a
288 result of inlining. */
291 inline_call (struct cgraph_edge
*e
, bool update_original
,
292 vec
<cgraph_edge
*> *new_edges
,
293 int *overall_size
, bool update_overall_summary
,
294 bool *callee_removed
)
296 int old_size
= 0, new_size
= 0;
297 struct cgraph_node
*to
= NULL
;
298 struct cgraph_edge
*curr
= e
;
299 struct cgraph_node
*callee
= e
->callee
->ultimate_alias_target ();
300 bool new_edges_found
= false;
302 int estimated_growth
= 0;
303 if (! update_overall_summary
)
304 estimated_growth
= estimate_edge_growth (e
);
305 /* This is used only for assert bellow. */
307 bool predicated
= inline_edge_summary (e
)->predicate
!= NULL
;
310 /* Don't inline inlined edges. */
311 gcc_assert (e
->inline_failed
);
312 /* Don't even think of inlining inline clone. */
313 gcc_assert (!callee
->global
.inlined_to
);
316 if (to
->global
.inlined_to
)
317 to
= to
->global
.inlined_to
;
318 if (to
->thunk
.thunk_p
)
320 struct cgraph_node
*target
= to
->callees
->callee
;
322 to
->get_untransformed_body ();
323 to
->expand_thunk (false, true);
324 /* When thunk is instrumented we may have multiple callees. */
325 for (e
= to
->callees
; e
&& e
->callee
!= target
; e
= e
->next_callee
)
331 e
->inline_failed
= CIF_OK
;
332 DECL_POSSIBLY_INLINED (callee
->decl
) = true;
334 if (DECL_FUNCTION_PERSONALITY (callee
->decl
))
335 DECL_FUNCTION_PERSONALITY (to
->decl
)
336 = DECL_FUNCTION_PERSONALITY (callee
->decl
);
338 bool reload_optimization_node
= false;
339 if (!opt_for_fn (callee
->decl
, flag_strict_aliasing
)
340 && opt_for_fn (to
->decl
, flag_strict_aliasing
))
342 struct gcc_options opts
= global_options
;
344 cl_optimization_restore (&opts
, opts_for_fn (to
->decl
));
345 opts
.x_flag_strict_aliasing
= false;
347 fprintf (dump_file
, "Dropping flag_strict_aliasing on %s\n",
349 DECL_FUNCTION_SPECIFIC_OPTIMIZATION (to
->decl
)
350 = build_optimization_node (&opts
);
351 reload_optimization_node
= true;
354 ipa_fn_summary
*caller_info
= ipa_fn_summaries
->get (to
);
355 ipa_fn_summary
*callee_info
= ipa_fn_summaries
->get (callee
);
356 if (!caller_info
->fp_expressions
&& callee_info
->fp_expressions
)
358 caller_info
->fp_expressions
= true;
359 if (opt_for_fn (callee
->decl
, flag_rounding_math
)
360 != opt_for_fn (to
->decl
, flag_rounding_math
)
361 || opt_for_fn (callee
->decl
, flag_trapping_math
)
362 != opt_for_fn (to
->decl
, flag_trapping_math
)
363 || opt_for_fn (callee
->decl
, flag_unsafe_math_optimizations
)
364 != opt_for_fn (to
->decl
, flag_unsafe_math_optimizations
)
365 || opt_for_fn (callee
->decl
, flag_finite_math_only
)
366 != opt_for_fn (to
->decl
, flag_finite_math_only
)
367 || opt_for_fn (callee
->decl
, flag_signaling_nans
)
368 != opt_for_fn (to
->decl
, flag_signaling_nans
)
369 || opt_for_fn (callee
->decl
, flag_cx_limited_range
)
370 != opt_for_fn (to
->decl
, flag_cx_limited_range
)
371 || opt_for_fn (callee
->decl
, flag_signed_zeros
)
372 != opt_for_fn (to
->decl
, flag_signed_zeros
)
373 || opt_for_fn (callee
->decl
, flag_associative_math
)
374 != opt_for_fn (to
->decl
, flag_associative_math
)
375 || opt_for_fn (callee
->decl
, flag_reciprocal_math
)
376 != opt_for_fn (to
->decl
, flag_reciprocal_math
)
377 || opt_for_fn (callee
->decl
, flag_fp_int_builtin_inexact
)
378 != opt_for_fn (to
->decl
, flag_fp_int_builtin_inexact
)
379 || opt_for_fn (callee
->decl
, flag_errno_math
)
380 != opt_for_fn (to
->decl
, flag_errno_math
))
382 struct gcc_options opts
= global_options
;
384 cl_optimization_restore (&opts
, opts_for_fn (to
->decl
));
385 opts
.x_flag_rounding_math
386 = opt_for_fn (callee
->decl
, flag_rounding_math
);
387 opts
.x_flag_trapping_math
388 = opt_for_fn (callee
->decl
, flag_trapping_math
);
389 opts
.x_flag_unsafe_math_optimizations
390 = opt_for_fn (callee
->decl
, flag_unsafe_math_optimizations
);
391 opts
.x_flag_finite_math_only
392 = opt_for_fn (callee
->decl
, flag_finite_math_only
);
393 opts
.x_flag_signaling_nans
394 = opt_for_fn (callee
->decl
, flag_signaling_nans
);
395 opts
.x_flag_cx_limited_range
396 = opt_for_fn (callee
->decl
, flag_cx_limited_range
);
397 opts
.x_flag_signed_zeros
398 = opt_for_fn (callee
->decl
, flag_signed_zeros
);
399 opts
.x_flag_associative_math
400 = opt_for_fn (callee
->decl
, flag_associative_math
);
401 opts
.x_flag_reciprocal_math
402 = opt_for_fn (callee
->decl
, flag_reciprocal_math
);
403 opts
.x_flag_fp_int_builtin_inexact
404 = opt_for_fn (callee
->decl
, flag_fp_int_builtin_inexact
);
405 opts
.x_flag_errno_math
406 = opt_for_fn (callee
->decl
, flag_errno_math
);
408 fprintf (dump_file
, "Copying FP flags from %s to %s\n",
409 callee
->dump_name (), to
->dump_name ());
410 DECL_FUNCTION_SPECIFIC_OPTIMIZATION (to
->decl
)
411 = build_optimization_node (&opts
);
412 reload_optimization_node
= true;
416 /* Reload global optimization flags. */
417 if (reload_optimization_node
&& DECL_STRUCT_FUNCTION (to
->decl
) == cfun
)
418 set_cfun (cfun
, true);
420 /* If aliases are involved, redirect edge to the actual destination and
421 possibly remove the aliases. */
422 if (e
->callee
!= callee
)
424 struct cgraph_node
*alias
= e
->callee
, *next_alias
;
425 e
->redirect_callee (callee
);
426 while (alias
&& alias
!= callee
)
429 && can_remove_node_now_p (alias
,
430 !e
->next_caller
&& !e
->prev_caller
? e
: NULL
))
432 next_alias
= alias
->get_alias_target ();
435 *callee_removed
= true;
443 clone_inlined_nodes (e
, true, update_original
, overall_size
);
445 gcc_assert (curr
->callee
->global
.inlined_to
== to
);
447 old_size
= ipa_fn_summaries
->get (to
)->size
;
448 ipa_merge_fn_summary_after_inlining (e
);
449 if (e
->in_polymorphic_cdtor
)
450 mark_all_inlined_calls_cdtor (e
->callee
);
451 if (opt_for_fn (e
->caller
->decl
, optimize
))
452 new_edges_found
= ipa_propagate_indirect_call_infos (curr
, new_edges
);
453 check_speculations (e
->callee
);
454 if (update_overall_summary
)
455 ipa_update_overall_fn_summary (to
);
457 /* Update self size by the estimate so overall function growth limits
458 work for further inlining into this function. Before inlining
459 the function we inlined to again we expect the caller to update
460 the overall summary. */
461 ipa_fn_summaries
->get (to
)->size
+= estimated_growth
;
462 new_size
= ipa_fn_summaries
->get (to
)->size
;
464 if (callee
->calls_comdat_local
)
465 to
->calls_comdat_local
= true;
466 else if (to
->calls_comdat_local
&& callee
->comdat_local_p ())
468 struct cgraph_edge
*se
= to
->callees
;
469 for (; se
; se
= se
->next_callee
)
470 if (se
->inline_failed
&& se
->callee
->comdat_local_p ())
473 to
->calls_comdat_local
= false;
476 /* FIXME: This assert suffers from roundoff errors, disable it for GCC 5
477 and revisit it after conversion to sreals in GCC 6.
480 /* Verify that estimated growth match real growth. Allow off-by-one
481 error due to ipa_fn_summary::size_scale roudoff errors. */
482 gcc_assert (!update_overall_summary
|| !overall_size
|| new_edges_found
483 || abs (estimated_growth
- (new_size
- old_size
)) <= 1
484 || speculation_removed
485 /* FIXME: a hack. Edges with false predicate are accounted
486 wrong, we should remove them from callgraph. */
490 /* Account the change of overall unit size; external functions will be
491 removed and are thus not accounted. */
492 if (overall_size
&& inline_account_function_p (to
))
493 *overall_size
+= new_size
- old_size
;
496 /* This must happen after ipa_merge_fn_summary_after_inlining that rely on jump
497 functions of callee to not be updated. */
498 return new_edges_found
;
502 /* Copy function body of NODE and redirect all inline clones to it.
503 This is done before inline plan is applied to NODE when there are
504 still some inline clones if it.
506 This is necessary because inline decisions are not really transitive
507 and the other inline clones may have different bodies. */
509 static struct cgraph_node
*
510 save_inline_function_body (struct cgraph_node
*node
)
512 struct cgraph_node
*first_clone
, *n
;
515 fprintf (dump_file
, "\nSaving body of %s for later reuse\n",
518 gcc_assert (node
== cgraph_node::get (node
->decl
));
520 /* first_clone will be turned into real function. */
521 first_clone
= node
->clones
;
523 /* Arrange first clone to not be thunk as those do not have bodies. */
524 if (first_clone
->thunk
.thunk_p
)
526 while (first_clone
->thunk
.thunk_p
)
527 first_clone
= first_clone
->next_sibling_clone
;
528 first_clone
->prev_sibling_clone
->next_sibling_clone
529 = first_clone
->next_sibling_clone
;
530 if (first_clone
->next_sibling_clone
)
531 first_clone
->next_sibling_clone
->prev_sibling_clone
532 = first_clone
->prev_sibling_clone
;
533 first_clone
->next_sibling_clone
= node
->clones
;
534 first_clone
->prev_sibling_clone
= NULL
;
535 node
->clones
->prev_sibling_clone
= first_clone
;
536 node
->clones
= first_clone
;
538 first_clone
->decl
= copy_node (node
->decl
);
539 first_clone
->decl
->decl_with_vis
.symtab_node
= first_clone
;
540 gcc_assert (first_clone
== cgraph_node::get (first_clone
->decl
));
542 /* Now reshape the clone tree, so all other clones descends from
544 if (first_clone
->next_sibling_clone
)
546 for (n
= first_clone
->next_sibling_clone
; n
->next_sibling_clone
;
547 n
= n
->next_sibling_clone
)
548 n
->clone_of
= first_clone
;
549 n
->clone_of
= first_clone
;
550 n
->next_sibling_clone
= first_clone
->clones
;
551 if (first_clone
->clones
)
552 first_clone
->clones
->prev_sibling_clone
= n
;
553 first_clone
->clones
= first_clone
->next_sibling_clone
;
554 first_clone
->next_sibling_clone
->prev_sibling_clone
= NULL
;
555 first_clone
->next_sibling_clone
= NULL
;
556 gcc_assert (!first_clone
->prev_sibling_clone
);
558 first_clone
->clone_of
= NULL
;
560 /* Now node in question has no clones. */
563 /* Inline clones share decl with the function they are cloned
564 from. Walk the whole clone tree and redirect them all to the
566 if (first_clone
->clones
)
567 for (n
= first_clone
->clones
; n
!= first_clone
;)
569 gcc_assert (n
->decl
== node
->decl
);
570 n
->decl
= first_clone
->decl
;
573 else if (n
->next_sibling_clone
)
574 n
= n
->next_sibling_clone
;
577 while (n
!= first_clone
&& !n
->next_sibling_clone
)
579 if (n
!= first_clone
)
580 n
= n
->next_sibling_clone
;
584 /* Copy the OLD_VERSION_NODE function tree to the new version. */
585 tree_function_versioning (node
->decl
, first_clone
->decl
,
586 NULL
, true, NULL
, false,
589 /* The function will be short lived and removed after we inline all the clones,
590 but make it internal so we won't confuse ourself. */
591 DECL_EXTERNAL (first_clone
->decl
) = 0;
592 TREE_PUBLIC (first_clone
->decl
) = 0;
593 DECL_COMDAT (first_clone
->decl
) = 0;
594 first_clone
->ipa_transforms_to_apply
.release ();
596 /* When doing recursive inlining, the clone may become unnecessary.
597 This is possible i.e. in the case when the recursive function is proved to be
598 non-throwing and the recursion happens only in the EH landing pad.
599 We can not remove the clone until we are done with saving the body.
601 if (!first_clone
->callers
)
603 first_clone
->remove_symbol_and_inline_clones ();
606 else if (flag_checking
)
607 first_clone
->verify ();
612 /* Return true when function body of DECL still needs to be kept around
615 preserve_function_body_p (struct cgraph_node
*node
)
617 gcc_assert (symtab
->global_info_ready
);
618 gcc_assert (!node
->alias
&& !node
->thunk
.thunk_p
);
620 /* Look if there is any non-thunk clone around. */
621 for (node
= node
->clones
; node
; node
= node
->next_sibling_clone
)
622 if (!node
->thunk
.thunk_p
)
627 /* Apply inline plan to function. */
630 inline_transform (struct cgraph_node
*node
)
632 unsigned int todo
= 0;
633 struct cgraph_edge
*e
, *next
;
634 bool has_inline
= false;
636 /* FIXME: Currently the pass manager is adding inline transform more than
637 once to some clones. This needs revisiting after WPA cleanups. */
638 if (cfun
->after_inlining
)
641 /* We might need the body of this function so that we can expand
642 it inline somewhere else. */
643 if (preserve_function_body_p (node
))
644 save_inline_function_body (node
);
646 for (e
= node
->callees
; e
; e
= next
)
648 if (!e
->inline_failed
)
650 next
= e
->next_callee
;
651 e
->redirect_call_stmt_to_callee ();
653 node
->remove_all_references ();
655 timevar_push (TV_INTEGRATION
);
656 if (node
->callees
&& (opt_for_fn (node
->decl
, optimize
) || has_inline
))
658 profile_count num
= node
->count
;
659 profile_count den
= ENTRY_BLOCK_PTR_FOR_FN (cfun
)->count
;
660 bool scale
= num
.initialized_p () && den
.ipa_p ()
661 && (den
.nonzero_p () || num
== profile_count::zero ())
662 && !(num
== den
.ipa ());
667 fprintf (dump_file
, "Applying count scale ");
668 num
.dump (dump_file
);
669 fprintf (dump_file
, "/");
670 den
.dump (dump_file
);
671 fprintf (dump_file
, "\n");
675 FOR_ALL_BB_FN (bb
, cfun
)
676 if (num
== profile_count::zero ())
677 bb
->count
= bb
->count
.global0 ();
679 bb
->count
= bb
->count
.apply_scale (num
, den
);
680 ENTRY_BLOCK_PTR_FOR_FN (cfun
)->count
= node
->count
;
682 todo
= optimize_inline_calls (current_function_decl
);
684 timevar_pop (TV_INTEGRATION
);
686 cfun
->always_inline_functions_inlined
= true;
687 cfun
->after_inlining
= true;
688 todo
|= execute_fixup_cfg ();
690 if (!(todo
& TODO_update_ssa_any
))
691 /* Redirecting edges might lead to a need for vops to be recomputed. */
692 todo
|= TODO_update_ssa_only_virtuals
;