1 /* Function splitting pass
2 Copyright (C) 2010, 2011, 2012
3 Free Software Foundation, Inc.
4 Contributed by Jan Hubicka <jh@suse.cz>
6 This file is part of GCC.
8 GCC is free software; you can redistribute it and/or modify it under
9 the terms of the GNU General Public License as published by the Free
10 Software Foundation; either version 3, or (at your option) any later
13 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
14 WARRANTY; without even the implied warranty of MERCHANTABILITY or
15 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
18 You should have received a copy of the GNU General Public License
19 along with GCC; see the file COPYING3. If not see
20 <http://www.gnu.org/licenses/>. */
22 /* The purpose of this pass is to split function bodies to improve
23 inlining. I.e. for function of the form:
48 When func becomes inlinable and when cheap_test is often true, inlining func,
49 but not fund.part leads to performance improvement similar as inlining
50 original func while the code size growth is smaller.
52 The pass is organized in three stages:
53 1) Collect local info about basic block into BB_INFO structure and
54 compute function body estimated size and time.
55 2) Via DFS walk find all possible basic blocks where we can split
57 3) If split point is found, split at the specified BB by creating a clone
58 and updating function to call it.
60 The decisions what functions to split are in execute_split_functions
63 There are several possible future improvements for this pass including:
65 1) Splitting to break up large functions
66 2) Splitting to reduce stack frame usage
67 3) Allow split part of function to use values computed in the header part.
68 The values needs to be passed to split function, perhaps via same
69 interface as for nested functions or as argument.
70 4) Support for simple rematerialization. I.e. when split part use
71 value computed in header from function parameter in very cheap way, we
72 can just recompute it.
73 5) Support splitting of nested functions.
74 6) Support non-SSA arguments.
75 7) There is nothing preventing us from producing multiple parts of single function
76 when needed or splitting also the parts. */
80 #include "coretypes.h"
85 #include "tree-flow.h"
86 #include "tree-pass.h"
88 #include "diagnostic.h"
89 #include "tree-dump.h"
90 #include "tree-inline.h"
92 #include "gimple-pretty-print.h"
93 #include "ipa-inline.h"
95 /* Per basic block info. */
103 DEF_VEC_ALLOC_O(bb_info
,heap
);
105 static VEC(bb_info
, heap
) *bb_info_vec
;
107 /* Description of split point. */
111 /* Size of the partitions. */
112 unsigned int header_time
, header_size
, split_time
, split_size
;
114 /* SSA names that need to be passed into spit function. */
115 bitmap ssa_names_to_pass
;
117 /* Basic block where we split (that will become entry point of new function. */
118 basic_block entry_bb
;
120 /* Basic blocks we are splitting away. */
123 /* True when return value is computed on split part and thus it needs
125 bool split_part_set_retval
;
128 /* Best split point found. */
130 struct split_point best_split_point
;
132 /* Set of basic blocks that are not allowed to dominate a split point. */
134 static bitmap forbidden_dominators
;
136 static tree
find_retval (basic_block return_bb
);
138 /* Callback for walk_stmt_load_store_addr_ops. If T is non-SSA automatic
139 variable, check it if it is present in bitmap passed via DATA. */
142 test_nonssa_use (gimple stmt ATTRIBUTE_UNUSED
, tree t
, void *data
)
144 t
= get_base_address (t
);
146 if (!t
|| is_gimple_reg (t
))
149 if (TREE_CODE (t
) == PARM_DECL
150 || (TREE_CODE (t
) == VAR_DECL
151 && auto_var_in_fn_p (t
, current_function_decl
))
152 || TREE_CODE (t
) == RESULT_DECL
153 || TREE_CODE (t
) == LABEL_DECL
)
154 return bitmap_bit_p ((bitmap
)data
, DECL_UID (t
));
156 /* For DECL_BY_REFERENCE, the return value is actually a pointer. We want
157 to pretend that the value pointed to is actual result decl. */
158 if ((TREE_CODE (t
) == MEM_REF
|| INDIRECT_REF_P (t
))
159 && TREE_CODE (TREE_OPERAND (t
, 0)) == SSA_NAME
160 && SSA_NAME_VAR (TREE_OPERAND (t
, 0))
161 && TREE_CODE (SSA_NAME_VAR (TREE_OPERAND (t
, 0))) == RESULT_DECL
162 && DECL_BY_REFERENCE (DECL_RESULT (current_function_decl
)))
164 bitmap_bit_p ((bitmap
)data
,
165 DECL_UID (DECL_RESULT (current_function_decl
)));
170 /* Dump split point CURRENT. */
173 dump_split_point (FILE * file
, struct split_point
*current
)
176 "Split point at BB %i\n"
177 " header time: %i header size: %i\n"
178 " split time: %i split size: %i\n bbs: ",
179 current
->entry_bb
->index
, current
->header_time
,
180 current
->header_size
, current
->split_time
, current
->split_size
);
181 dump_bitmap (file
, current
->split_bbs
);
182 fprintf (file
, " SSA names to pass: ");
183 dump_bitmap (file
, current
->ssa_names_to_pass
);
186 /* Look for all BBs in header that might lead to the split part and verify
187 that they are not defining any non-SSA var used by the split part.
188 Parameters are the same as for consider_split. */
191 verify_non_ssa_vars (struct split_point
*current
, bitmap non_ssa_vars
,
192 basic_block return_bb
)
194 bitmap seen
= BITMAP_ALLOC (NULL
);
195 VEC (basic_block
,heap
) *worklist
= NULL
;
200 FOR_EACH_EDGE (e
, ei
, current
->entry_bb
->preds
)
201 if (e
->src
!= ENTRY_BLOCK_PTR
202 && !bitmap_bit_p (current
->split_bbs
, e
->src
->index
))
204 VEC_safe_push (basic_block
, heap
, worklist
, e
->src
);
205 bitmap_set_bit (seen
, e
->src
->index
);
208 while (!VEC_empty (basic_block
, worklist
))
210 gimple_stmt_iterator bsi
;
211 basic_block bb
= VEC_pop (basic_block
, worklist
);
213 FOR_EACH_EDGE (e
, ei
, bb
->preds
)
214 if (e
->src
!= ENTRY_BLOCK_PTR
215 && bitmap_set_bit (seen
, e
->src
->index
))
217 gcc_checking_assert (!bitmap_bit_p (current
->split_bbs
,
219 VEC_safe_push (basic_block
, heap
, worklist
, e
->src
);
221 for (bsi
= gsi_start_bb (bb
); !gsi_end_p (bsi
); gsi_next (&bsi
))
223 gimple stmt
= gsi_stmt (bsi
);
224 if (is_gimple_debug (stmt
))
226 if (walk_stmt_load_store_addr_ops
227 (stmt
, non_ssa_vars
, test_nonssa_use
, test_nonssa_use
,
233 if (gimple_code (stmt
) == GIMPLE_LABEL
234 && test_nonssa_use (stmt
, gimple_label_label (stmt
),
241 for (bsi
= gsi_start_phis (bb
); !gsi_end_p (bsi
); gsi_next (&bsi
))
243 if (walk_stmt_load_store_addr_ops
244 (gsi_stmt (bsi
), non_ssa_vars
, test_nonssa_use
, test_nonssa_use
,
251 FOR_EACH_EDGE (e
, ei
, bb
->succs
)
253 if (e
->dest
!= return_bb
)
255 for (bsi
= gsi_start_phis (return_bb
); !gsi_end_p (bsi
);
258 gimple stmt
= gsi_stmt (bsi
);
259 tree op
= gimple_phi_arg_def (stmt
, e
->dest_idx
);
261 if (virtual_operand_p (gimple_phi_result (stmt
)))
263 if (TREE_CODE (op
) != SSA_NAME
264 && test_nonssa_use (stmt
, op
, non_ssa_vars
))
274 VEC_free (basic_block
, heap
, worklist
);
278 /* If STMT is a call, check the callee against a list of forbidden
279 predicate functions. If a match is found, look for uses of the
280 call result in condition statements that compare against zero.
281 For each such use, find the block targeted by the condition
282 statement for the nonzero result, and set the bit for this block
283 in the forbidden dominators bitmap. The purpose of this is to avoid
284 selecting a split point where we are likely to lose the chance
285 to optimize away an unused function call. */
288 check_forbidden_calls (gimple stmt
)
290 imm_use_iterator use_iter
;
294 /* At the moment, __builtin_constant_p is the only forbidden
295 predicate function call (see PR49642). */
296 if (!gimple_call_builtin_p (stmt
, BUILT_IN_CONSTANT_P
))
299 lhs
= gimple_call_lhs (stmt
);
301 if (!lhs
|| TREE_CODE (lhs
) != SSA_NAME
)
304 FOR_EACH_IMM_USE_FAST (use_p
, use_iter
, lhs
)
307 basic_block use_bb
, forbidden_bb
;
309 edge true_edge
, false_edge
;
310 gimple use_stmt
= USE_STMT (use_p
);
312 if (gimple_code (use_stmt
) != GIMPLE_COND
)
315 /* Assuming canonical form for GIMPLE_COND here, with constant
316 in second position. */
317 op1
= gimple_cond_rhs (use_stmt
);
318 code
= gimple_cond_code (use_stmt
);
319 use_bb
= gimple_bb (use_stmt
);
321 extract_true_false_edges_from_block (use_bb
, &true_edge
, &false_edge
);
323 /* We're only interested in comparisons that distinguish
324 unambiguously from zero. */
325 if (!integer_zerop (op1
) || code
== LE_EXPR
|| code
== GE_EXPR
)
329 forbidden_bb
= false_edge
->dest
;
331 forbidden_bb
= true_edge
->dest
;
333 bitmap_set_bit (forbidden_dominators
, forbidden_bb
->index
);
337 /* If BB is dominated by any block in the forbidden dominators set,
338 return TRUE; else FALSE. */
341 dominated_by_forbidden (basic_block bb
)
346 EXECUTE_IF_SET_IN_BITMAP (forbidden_dominators
, 1, dom_bb
, bi
)
348 if (dominated_by_p (CDI_DOMINATORS
, bb
, BASIC_BLOCK (dom_bb
)))
355 /* We found an split_point CURRENT. NON_SSA_VARS is bitmap of all non ssa
356 variables used and RETURN_BB is return basic block.
357 See if we can split function here. */
360 consider_split (struct split_point
*current
, bitmap non_ssa_vars
,
361 basic_block return_bb
)
364 unsigned int num_args
= 0;
365 unsigned int call_overhead
;
368 gimple_stmt_iterator bsi
;
370 int incoming_freq
= 0;
373 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
374 dump_split_point (dump_file
, current
);
376 FOR_EACH_EDGE (e
, ei
, current
->entry_bb
->preds
)
377 if (!bitmap_bit_p (current
->split_bbs
, e
->src
->index
))
378 incoming_freq
+= EDGE_FREQUENCY (e
);
380 /* Do not split when we would end up calling function anyway. */
382 >= (ENTRY_BLOCK_PTR
->frequency
383 * PARAM_VALUE (PARAM_PARTIAL_INLINING_ENTRY_PROBABILITY
) / 100))
385 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
387 " Refused: incoming frequency is too large.\n");
391 if (!current
->header_size
)
393 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
394 fprintf (dump_file
, " Refused: header empty\n");
398 /* Verify that PHI args on entry are either virtual or all their operands
399 incoming from header are the same. */
400 for (bsi
= gsi_start_phis (current
->entry_bb
); !gsi_end_p (bsi
); gsi_next (&bsi
))
402 gimple stmt
= gsi_stmt (bsi
);
405 if (virtual_operand_p (gimple_phi_result (stmt
)))
407 for (i
= 0; i
< gimple_phi_num_args (stmt
); i
++)
409 edge e
= gimple_phi_arg_edge (stmt
, i
);
410 if (!bitmap_bit_p (current
->split_bbs
, e
->src
->index
))
412 tree edge_val
= gimple_phi_arg_def (stmt
, i
);
413 if (val
&& edge_val
!= val
)
415 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
417 " Refused: entry BB has PHI with multiple variants\n");
426 /* See what argument we will pass to the split function and compute
428 call_overhead
= eni_size_weights
.call_cost
;
429 for (parm
= DECL_ARGUMENTS (current_function_decl
); parm
;
430 parm
= DECL_CHAIN (parm
))
432 if (!is_gimple_reg (parm
))
434 if (bitmap_bit_p (non_ssa_vars
, DECL_UID (parm
)))
436 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
438 " Refused: need to pass non-ssa param values\n");
444 tree ddef
= ssa_default_def (cfun
, parm
);
446 && bitmap_bit_p (current
->ssa_names_to_pass
,
447 SSA_NAME_VERSION (ddef
)))
449 if (!VOID_TYPE_P (TREE_TYPE (parm
)))
450 call_overhead
+= estimate_move_cost (TREE_TYPE (parm
));
455 if (!VOID_TYPE_P (TREE_TYPE (current_function_decl
)))
456 call_overhead
+= estimate_move_cost (TREE_TYPE (current_function_decl
));
458 if (current
->split_size
<= call_overhead
)
460 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
462 " Refused: split size is smaller than call overhead\n");
465 if (current
->header_size
+ call_overhead
466 >= (unsigned int)(DECL_DECLARED_INLINE_P (current_function_decl
)
467 ? MAX_INLINE_INSNS_SINGLE
468 : MAX_INLINE_INSNS_AUTO
))
470 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
472 " Refused: header size is too large for inline candidate\n");
476 /* FIXME: we currently can pass only SSA function parameters to the split
477 arguments. Once parm_adjustment infrastructure is supported by cloning,
478 we can pass more than that. */
479 if (num_args
!= bitmap_count_bits (current
->ssa_names_to_pass
))
482 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
484 " Refused: need to pass non-param values\n");
488 /* When there are non-ssa vars used in the split region, see if they
489 are used in the header region. If so, reject the split.
490 FIXME: we can use nested function support to access both. */
491 if (!bitmap_empty_p (non_ssa_vars
)
492 && !verify_non_ssa_vars (current
, non_ssa_vars
, return_bb
))
494 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
496 " Refused: split part has non-ssa uses\n");
500 /* If the split point is dominated by a forbidden block, reject
502 if (!bitmap_empty_p (forbidden_dominators
)
503 && dominated_by_forbidden (current
->entry_bb
))
505 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
507 " Refused: split point dominated by forbidden block\n");
511 /* See if retval used by return bb is computed by header or split part.
512 When it is computed by split part, we need to produce return statement
513 in the split part and add code to header to pass it around.
515 This is bit tricky to test:
516 1) When there is no return_bb or no return value, we always pass
518 2) Invariants are always computed by caller.
519 3) For SSA we need to look if defining statement is in header or split part
520 4) For non-SSA we need to look where the var is computed. */
521 retval
= find_retval (return_bb
);
523 current
->split_part_set_retval
= true;
524 else if (is_gimple_min_invariant (retval
))
525 current
->split_part_set_retval
= false;
526 /* Special case is value returned by reference we record as if it was non-ssa
527 set to result_decl. */
528 else if (TREE_CODE (retval
) == SSA_NAME
529 && SSA_NAME_VAR (retval
)
530 && TREE_CODE (SSA_NAME_VAR (retval
)) == RESULT_DECL
531 && DECL_BY_REFERENCE (DECL_RESULT (current_function_decl
)))
532 current
->split_part_set_retval
533 = bitmap_bit_p (non_ssa_vars
, DECL_UID (SSA_NAME_VAR (retval
)));
534 else if (TREE_CODE (retval
) == SSA_NAME
)
535 current
->split_part_set_retval
536 = (!SSA_NAME_IS_DEFAULT_DEF (retval
)
537 && (bitmap_bit_p (current
->split_bbs
,
538 gimple_bb (SSA_NAME_DEF_STMT (retval
))->index
)
539 || gimple_bb (SSA_NAME_DEF_STMT (retval
)) == return_bb
));
540 else if (TREE_CODE (retval
) == PARM_DECL
)
541 current
->split_part_set_retval
= false;
542 else if (TREE_CODE (retval
) == VAR_DECL
543 || TREE_CODE (retval
) == RESULT_DECL
)
544 current
->split_part_set_retval
545 = bitmap_bit_p (non_ssa_vars
, DECL_UID (retval
));
547 current
->split_part_set_retval
= true;
549 /* split_function fixes up at most one PHI non-virtual PHI node in return_bb,
550 for the return value. If there are other PHIs, give up. */
551 if (return_bb
!= EXIT_BLOCK_PTR
)
553 gimple_stmt_iterator psi
;
555 for (psi
= gsi_start_phis (return_bb
); !gsi_end_p (psi
); gsi_next (&psi
))
556 if (!virtual_operand_p (gimple_phi_result (gsi_stmt (psi
)))
558 && current
->split_part_set_retval
559 && TREE_CODE (retval
) == SSA_NAME
560 && !DECL_BY_REFERENCE (DECL_RESULT (current_function_decl
))
561 && SSA_NAME_DEF_STMT (retval
) == gsi_stmt (psi
)))
563 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
565 " Refused: return bb has extra PHIs\n");
570 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
571 fprintf (dump_file
, " Accepted!\n");
573 /* At the moment chose split point with lowest frequency and that leaves
574 out smallest size of header.
575 In future we might re-consider this heuristics. */
576 if (!best_split_point
.split_bbs
577 || best_split_point
.entry_bb
->frequency
> current
->entry_bb
->frequency
578 || (best_split_point
.entry_bb
->frequency
== current
->entry_bb
->frequency
579 && best_split_point
.split_size
< current
->split_size
))
582 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
583 fprintf (dump_file
, " New best split point!\n");
584 if (best_split_point
.ssa_names_to_pass
)
586 BITMAP_FREE (best_split_point
.ssa_names_to_pass
);
587 BITMAP_FREE (best_split_point
.split_bbs
);
589 best_split_point
= *current
;
590 best_split_point
.ssa_names_to_pass
= BITMAP_ALLOC (NULL
);
591 bitmap_copy (best_split_point
.ssa_names_to_pass
,
592 current
->ssa_names_to_pass
);
593 best_split_point
.split_bbs
= BITMAP_ALLOC (NULL
);
594 bitmap_copy (best_split_point
.split_bbs
, current
->split_bbs
);
598 /* Return basic block containing RETURN statement. We allow basic blocks
602 but return_bb can not be more complex than this.
603 If nothing is found, return EXIT_BLOCK_PTR.
605 When there are multiple RETURN statement, chose one with return value,
606 since that one is more likely shared by multiple code paths.
608 Return BB is special, because for function splitting it is the only
609 basic block that is duplicated in between header and split part of the
612 TODO: We might support multiple return blocks. */
615 find_return_bb (void)
618 basic_block return_bb
= EXIT_BLOCK_PTR
;
619 gimple_stmt_iterator bsi
;
620 bool found_return
= false;
621 tree retval
= NULL_TREE
;
623 if (!single_pred_p (EXIT_BLOCK_PTR
))
626 e
= single_pred_edge (EXIT_BLOCK_PTR
);
627 for (bsi
= gsi_last_bb (e
->src
); !gsi_end_p (bsi
); gsi_prev (&bsi
))
629 gimple stmt
= gsi_stmt (bsi
);
630 if (gimple_code (stmt
) == GIMPLE_LABEL
631 || is_gimple_debug (stmt
)
632 || gimple_clobber_p (stmt
))
634 else if (gimple_code (stmt
) == GIMPLE_ASSIGN
636 && gimple_assign_single_p (stmt
)
637 && (auto_var_in_fn_p (gimple_assign_rhs1 (stmt
),
638 current_function_decl
)
639 || is_gimple_min_invariant (gimple_assign_rhs1 (stmt
)))
640 && retval
== gimple_assign_lhs (stmt
))
642 else if (gimple_code (stmt
) == GIMPLE_RETURN
)
645 retval
= gimple_return_retval (stmt
);
650 if (gsi_end_p (bsi
) && found_return
)
656 /* Given return basic block RETURN_BB, see where return value is really
659 find_retval (basic_block return_bb
)
661 gimple_stmt_iterator bsi
;
662 for (bsi
= gsi_start_bb (return_bb
); !gsi_end_p (bsi
); gsi_next (&bsi
))
663 if (gimple_code (gsi_stmt (bsi
)) == GIMPLE_RETURN
)
664 return gimple_return_retval (gsi_stmt (bsi
));
665 else if (gimple_code (gsi_stmt (bsi
)) == GIMPLE_ASSIGN
666 && !gimple_clobber_p (gsi_stmt (bsi
)))
667 return gimple_assign_rhs1 (gsi_stmt (bsi
));
671 /* Callback for walk_stmt_load_store_addr_ops. If T is non-SSA automatic
672 variable, mark it as used in bitmap passed via DATA.
673 Return true when access to T prevents splitting the function. */
676 mark_nonssa_use (gimple stmt ATTRIBUTE_UNUSED
, tree t
, void *data
)
678 t
= get_base_address (t
);
680 if (!t
|| is_gimple_reg (t
))
683 /* At present we can't pass non-SSA arguments to split function.
684 FIXME: this can be relaxed by passing references to arguments. */
685 if (TREE_CODE (t
) == PARM_DECL
)
687 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
689 "Cannot split: use of non-ssa function parameter.\n");
693 if ((TREE_CODE (t
) == VAR_DECL
694 && auto_var_in_fn_p (t
, current_function_decl
))
695 || TREE_CODE (t
) == RESULT_DECL
696 || TREE_CODE (t
) == LABEL_DECL
)
697 bitmap_set_bit ((bitmap
)data
, DECL_UID (t
));
699 /* For DECL_BY_REFERENCE, the return value is actually a pointer. We want
700 to pretend that the value pointed to is actual result decl. */
701 if ((TREE_CODE (t
) == MEM_REF
|| INDIRECT_REF_P (t
))
702 && TREE_CODE (TREE_OPERAND (t
, 0)) == SSA_NAME
703 && SSA_NAME_VAR (TREE_OPERAND (t
, 0))
704 && TREE_CODE (SSA_NAME_VAR (TREE_OPERAND (t
, 0))) == RESULT_DECL
705 && DECL_BY_REFERENCE (DECL_RESULT (current_function_decl
)))
707 bitmap_bit_p ((bitmap
)data
,
708 DECL_UID (DECL_RESULT (current_function_decl
)));
713 /* Compute local properties of basic block BB we collect when looking for
714 split points. We look for ssa defs and store them in SET_SSA_NAMES,
715 for ssa uses and store them in USED_SSA_NAMES and for any non-SSA automatic
716 vars stored in NON_SSA_VARS.
718 When BB has edge to RETURN_BB, collect uses in RETURN_BB too.
720 Return false when BB contains something that prevents it from being put into
724 visit_bb (basic_block bb
, basic_block return_bb
,
725 bitmap set_ssa_names
, bitmap used_ssa_names
,
728 gimple_stmt_iterator bsi
;
731 bool can_split
= true;
733 for (bsi
= gsi_start_bb (bb
); !gsi_end_p (bsi
); gsi_next (&bsi
))
735 gimple stmt
= gsi_stmt (bsi
);
740 if (is_gimple_debug (stmt
))
743 if (gimple_clobber_p (stmt
))
746 /* FIXME: We can split regions containing EH. We can not however
747 split RESX, EH_DISPATCH and EH_POINTER referring to same region
748 into different partitions. This would require tracking of
749 EH regions and checking in consider_split_point if they
750 are not used elsewhere. */
751 if (gimple_code (stmt
) == GIMPLE_RESX
)
753 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
754 fprintf (dump_file
, "Cannot split: resx.\n");
757 if (gimple_code (stmt
) == GIMPLE_EH_DISPATCH
)
759 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
760 fprintf (dump_file
, "Cannot split: eh dispatch.\n");
764 /* Check builtins that prevent splitting. */
765 if (gimple_code (stmt
) == GIMPLE_CALL
766 && (decl
= gimple_call_fndecl (stmt
)) != NULL_TREE
767 && DECL_BUILT_IN (decl
)
768 && DECL_BUILT_IN_CLASS (decl
) == BUILT_IN_NORMAL
)
769 switch (DECL_FUNCTION_CODE (decl
))
771 /* FIXME: once we will allow passing non-parm values to split part,
772 we need to be sure to handle correct builtin_stack_save and
773 builtin_stack_restore. At the moment we are safe; there is no
774 way to store builtin_stack_save result in non-SSA variable
775 since all calls to those are compiler generated. */
777 case BUILT_IN_APPLY_ARGS
:
778 case BUILT_IN_VA_START
:
779 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
781 "Cannot split: builtin_apply and va_start.\n");
784 case BUILT_IN_EH_POINTER
:
785 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
786 fprintf (dump_file
, "Cannot split: builtin_eh_pointer.\n");
793 FOR_EACH_SSA_TREE_OPERAND (op
, stmt
, iter
, SSA_OP_DEF
)
794 bitmap_set_bit (set_ssa_names
, SSA_NAME_VERSION (op
));
795 FOR_EACH_SSA_TREE_OPERAND (op
, stmt
, iter
, SSA_OP_USE
)
796 bitmap_set_bit (used_ssa_names
, SSA_NAME_VERSION (op
));
797 can_split
&= !walk_stmt_load_store_addr_ops (stmt
, non_ssa_vars
,
802 for (bsi
= gsi_start_phis (bb
); !gsi_end_p (bsi
); gsi_next (&bsi
))
804 gimple stmt
= gsi_stmt (bsi
);
807 if (virtual_operand_p (gimple_phi_result (stmt
)))
809 bitmap_set_bit (set_ssa_names
,
810 SSA_NAME_VERSION (gimple_phi_result (stmt
)));
811 for (i
= 0; i
< gimple_phi_num_args (stmt
); i
++)
813 tree op
= gimple_phi_arg_def (stmt
, i
);
814 if (TREE_CODE (op
) == SSA_NAME
)
815 bitmap_set_bit (used_ssa_names
, SSA_NAME_VERSION (op
));
817 can_split
&= !walk_stmt_load_store_addr_ops (stmt
, non_ssa_vars
,
822 /* Record also uses coming from PHI operand in return BB. */
823 FOR_EACH_EDGE (e
, ei
, bb
->succs
)
824 if (e
->dest
== return_bb
)
826 for (bsi
= gsi_start_phis (return_bb
); !gsi_end_p (bsi
); gsi_next (&bsi
))
828 gimple stmt
= gsi_stmt (bsi
);
829 tree op
= gimple_phi_arg_def (stmt
, e
->dest_idx
);
831 if (virtual_operand_p (gimple_phi_result (stmt
)))
833 if (TREE_CODE (op
) == SSA_NAME
)
834 bitmap_set_bit (used_ssa_names
, SSA_NAME_VERSION (op
));
836 can_split
&= !mark_nonssa_use (stmt
, op
, non_ssa_vars
);
842 /* Stack entry for recursive DFS walk in find_split_point. */
846 /* Basic block we are examining. */
849 /* SSA names set and used by the BB and all BBs reachable
850 from it via DFS walk. */
851 bitmap set_ssa_names
, used_ssa_names
;
854 /* All BBS visited from this BB via DFS walk. */
857 /* Last examined edge in DFS walk. Since we walk unoriented graph,
858 the value is up to sum of incoming and outgoing edges of BB. */
859 unsigned int edge_num
;
861 /* Stack entry index of earliest BB reachable from current BB
862 or any BB visited later in DFS walk. */
865 /* Overall time and size of all BBs reached from this BB in DFS walk. */
866 int overall_time
, overall_size
;
868 /* When false we can not split on this BB. */
871 DEF_VEC_O(stack_entry
);
872 DEF_VEC_ALLOC_O(stack_entry
,heap
);
875 /* Find all articulations and call consider_split on them.
876 OVERALL_TIME and OVERALL_SIZE is time and size of the function.
878 We perform basic algorithm for finding an articulation in a graph
879 created from CFG by considering it to be an unoriented graph.
881 The articulation is discovered via DFS walk. We collect earliest
882 basic block on stack that is reachable via backward edge. Articulation
883 is any basic block such that there is no backward edge bypassing it.
884 To reduce stack usage we maintain heap allocated stack in STACK vector.
885 AUX pointer of BB is set to index it appears in the stack or -1 once
886 it is visited and popped off the stack.
888 The algorithm finds articulation after visiting the whole component
889 reachable by it. This makes it convenient to collect information about
890 the component used by consider_split. */
893 find_split_points (int overall_time
, int overall_size
)
896 VEC(stack_entry
, heap
) *stack
= NULL
;
898 basic_block return_bb
= find_return_bb ();
899 struct split_point current
;
901 current
.header_time
= overall_time
;
902 current
.header_size
= overall_size
;
903 current
.split_time
= 0;
904 current
.split_size
= 0;
905 current
.ssa_names_to_pass
= BITMAP_ALLOC (NULL
);
907 first
.bb
= ENTRY_BLOCK_PTR
;
909 first
.overall_time
= 0;
910 first
.overall_size
= 0;
911 first
.earliest
= INT_MAX
;
912 first
.set_ssa_names
= 0;
913 first
.used_ssa_names
= 0;
914 first
.bbs_visited
= 0;
915 VEC_safe_push (stack_entry
, heap
, stack
, &first
);
916 ENTRY_BLOCK_PTR
->aux
= (void *)(intptr_t)-1;
918 while (!VEC_empty (stack_entry
, stack
))
920 stack_entry
*entry
= &VEC_last (stack_entry
, stack
);
922 /* We are walking an acyclic graph, so edge_num counts
923 succ and pred edges together. However when considering
924 articulation, we want to have processed everything reachable
925 from articulation but nothing that reaches into it. */
926 if (entry
->edge_num
== EDGE_COUNT (entry
->bb
->succs
)
927 && entry
->bb
!= ENTRY_BLOCK_PTR
)
929 int pos
= VEC_length (stack_entry
, stack
);
930 entry
->can_split
&= visit_bb (entry
->bb
, return_bb
,
931 entry
->set_ssa_names
,
932 entry
->used_ssa_names
,
933 entry
->non_ssa_vars
);
934 if (pos
<= entry
->earliest
&& !entry
->can_split
935 && dump_file
&& (dump_flags
& TDF_DETAILS
))
937 "found articulation at bb %i but can not split\n",
939 if (pos
<= entry
->earliest
&& entry
->can_split
)
941 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
942 fprintf (dump_file
, "found articulation at bb %i\n",
944 current
.entry_bb
= entry
->bb
;
945 current
.ssa_names_to_pass
= BITMAP_ALLOC (NULL
);
946 bitmap_and_compl (current
.ssa_names_to_pass
,
947 entry
->used_ssa_names
, entry
->set_ssa_names
);
948 current
.header_time
= overall_time
- entry
->overall_time
;
949 current
.header_size
= overall_size
- entry
->overall_size
;
950 current
.split_time
= entry
->overall_time
;
951 current
.split_size
= entry
->overall_size
;
952 current
.split_bbs
= entry
->bbs_visited
;
953 consider_split (¤t
, entry
->non_ssa_vars
, return_bb
);
954 BITMAP_FREE (current
.ssa_names_to_pass
);
957 /* Do actual DFS walk. */
959 < (EDGE_COUNT (entry
->bb
->succs
)
960 + EDGE_COUNT (entry
->bb
->preds
)))
964 if (entry
->edge_num
< EDGE_COUNT (entry
->bb
->succs
))
966 e
= EDGE_SUCC (entry
->bb
, entry
->edge_num
);
971 e
= EDGE_PRED (entry
->bb
, entry
->edge_num
972 - EDGE_COUNT (entry
->bb
->succs
));
978 /* New BB to visit, push it to the stack. */
979 if (dest
!= return_bb
&& dest
!= EXIT_BLOCK_PTR
982 stack_entry new_entry
;
985 new_entry
.edge_num
= 0;
986 new_entry
.overall_time
987 = VEC_index (bb_info
, bb_info_vec
, dest
->index
).time
;
988 new_entry
.overall_size
989 = VEC_index (bb_info
, bb_info_vec
, dest
->index
).size
;
990 new_entry
.earliest
= INT_MAX
;
991 new_entry
.set_ssa_names
= BITMAP_ALLOC (NULL
);
992 new_entry
.used_ssa_names
= BITMAP_ALLOC (NULL
);
993 new_entry
.bbs_visited
= BITMAP_ALLOC (NULL
);
994 new_entry
.non_ssa_vars
= BITMAP_ALLOC (NULL
);
995 new_entry
.can_split
= true;
996 bitmap_set_bit (new_entry
.bbs_visited
, dest
->index
);
997 VEC_safe_push (stack_entry
, heap
, stack
, &new_entry
);
998 dest
->aux
= (void *)(intptr_t)VEC_length (stack_entry
, stack
);
1000 /* Back edge found, record the earliest point. */
1001 else if ((intptr_t)dest
->aux
> 0
1002 && (intptr_t)dest
->aux
< entry
->earliest
)
1003 entry
->earliest
= (intptr_t)dest
->aux
;
1005 /* We are done with examining the edges. Pop off the value from stack
1006 and merge stuff we accumulate during the walk. */
1007 else if (entry
->bb
!= ENTRY_BLOCK_PTR
)
1009 stack_entry
*prev
= &VEC_index (stack_entry
, stack
,
1010 VEC_length (stack_entry
, stack
) - 2);
1012 entry
->bb
->aux
= (void *)(intptr_t)-1;
1013 prev
->can_split
&= entry
->can_split
;
1014 if (prev
->set_ssa_names
)
1016 bitmap_ior_into (prev
->set_ssa_names
, entry
->set_ssa_names
);
1017 bitmap_ior_into (prev
->used_ssa_names
, entry
->used_ssa_names
);
1018 bitmap_ior_into (prev
->bbs_visited
, entry
->bbs_visited
);
1019 bitmap_ior_into (prev
->non_ssa_vars
, entry
->non_ssa_vars
);
1021 if (prev
->earliest
> entry
->earliest
)
1022 prev
->earliest
= entry
->earliest
;
1023 prev
->overall_time
+= entry
->overall_time
;
1024 prev
->overall_size
+= entry
->overall_size
;
1025 BITMAP_FREE (entry
->set_ssa_names
);
1026 BITMAP_FREE (entry
->used_ssa_names
);
1027 BITMAP_FREE (entry
->bbs_visited
);
1028 BITMAP_FREE (entry
->non_ssa_vars
);
1029 VEC_pop (stack_entry
, stack
);
1032 VEC_pop (stack_entry
, stack
);
1034 ENTRY_BLOCK_PTR
->aux
= NULL
;
1037 VEC_free (stack_entry
, heap
, stack
);
1038 BITMAP_FREE (current
.ssa_names_to_pass
);
1041 /* Split function at SPLIT_POINT. */
1044 split_function (struct split_point
*split_point
)
1046 VEC (tree
, heap
) *args_to_pass
= NULL
;
1047 bitmap args_to_skip
;
1050 struct cgraph_node
*node
, *cur_node
= cgraph_get_node (current_function_decl
);
1051 basic_block return_bb
= find_return_bb ();
1052 basic_block call_bb
;
1053 gimple_stmt_iterator gsi
;
1057 tree retval
= NULL
, real_retval
= NULL
;
1058 bool split_part_return_p
= false;
1059 gimple last_stmt
= NULL
;
1065 fprintf (dump_file
, "\n\nSplitting function at:\n");
1066 dump_split_point (dump_file
, split_point
);
1069 if (cur_node
->local
.can_change_signature
)
1070 args_to_skip
= BITMAP_ALLOC (NULL
);
1072 args_to_skip
= NULL
;
1074 /* Collect the parameters of new function and args_to_skip bitmap. */
1075 for (parm
= DECL_ARGUMENTS (current_function_decl
);
1076 parm
; parm
= DECL_CHAIN (parm
), num
++)
1078 && (!is_gimple_reg (parm
)
1079 || (ddef
= ssa_default_def (cfun
, parm
)) == NULL_TREE
1080 || !bitmap_bit_p (split_point
->ssa_names_to_pass
,
1081 SSA_NAME_VERSION (ddef
))))
1082 bitmap_set_bit (args_to_skip
, num
);
1085 /* This parm might not have been used up to now, but is going to be
1086 used, hence register it. */
1087 if (is_gimple_reg (parm
))
1088 arg
= get_or_create_ssa_default_def (cfun
, parm
);
1092 if (!useless_type_conversion_p (DECL_ARG_TYPE (parm
), TREE_TYPE (arg
)))
1093 arg
= fold_convert (DECL_ARG_TYPE (parm
), arg
);
1094 VEC_safe_push (tree
, heap
, args_to_pass
, arg
);
1097 /* See if the split function will return. */
1098 FOR_EACH_EDGE (e
, ei
, return_bb
->preds
)
1099 if (bitmap_bit_p (split_point
->split_bbs
, e
->src
->index
))
1102 split_part_return_p
= true;
1104 /* Add return block to what will become the split function.
1105 We do not return; no return block is needed. */
1106 if (!split_part_return_p
)
1108 /* We have no return block, so nothing is needed. */
1109 else if (return_bb
== EXIT_BLOCK_PTR
)
1111 /* When we do not want to return value, we need to construct
1112 new return block with empty return statement.
1113 FIXME: Once we are able to change return type, we should change function
1114 to return void instead of just outputting function with undefined return
1115 value. For structures this affects quality of codegen. */
1116 else if (!split_point
->split_part_set_retval
1117 && find_retval (return_bb
))
1119 bool redirected
= true;
1120 basic_block new_return_bb
= create_basic_block (NULL
, 0, return_bb
);
1121 gimple_stmt_iterator gsi
= gsi_start_bb (new_return_bb
);
1122 gsi_insert_after (&gsi
, gimple_build_return (NULL
), GSI_NEW_STMT
);
1126 FOR_EACH_EDGE (e
, ei
, return_bb
->preds
)
1127 if (bitmap_bit_p (split_point
->split_bbs
, e
->src
->index
))
1129 new_return_bb
->count
+= e
->count
;
1130 new_return_bb
->frequency
+= EDGE_FREQUENCY (e
);
1131 redirect_edge_and_branch (e
, new_return_bb
);
1136 e
= make_edge (new_return_bb
, EXIT_BLOCK_PTR
, 0);
1137 e
->probability
= REG_BR_PROB_BASE
;
1138 e
->count
= new_return_bb
->count
;
1139 bitmap_set_bit (split_point
->split_bbs
, new_return_bb
->index
);
1141 /* When we pass around the value, use existing return block. */
1143 bitmap_set_bit (split_point
->split_bbs
, return_bb
->index
);
1145 /* If RETURN_BB has virtual operand PHIs, they must be removed and the
1146 virtual operand marked for renaming as we change the CFG in a way that
1147 tree-inline is not able to compensate for.
1149 Note this can happen whether or not we have a return value. If we have
1150 a return value, then RETURN_BB may have PHIs for real operands too. */
1151 if (return_bb
!= EXIT_BLOCK_PTR
)
1154 for (gsi
= gsi_start_phis (return_bb
); !gsi_end_p (gsi
);)
1156 gimple stmt
= gsi_stmt (gsi
);
1157 if (!virtual_operand_p (gimple_phi_result (stmt
)))
1162 mark_virtual_phi_result_for_renaming (stmt
);
1163 remove_phi_node (&gsi
, true);
1166 /* In reality we have to rename the reaching definition of the
1167 virtual operand at return_bb as we will eventually release it
1168 when we remove the code region we outlined.
1169 So we have to rename all immediate virtual uses of that region
1170 if we didn't see a PHI definition yet. */
1171 /* ??? In real reality we want to set the reaching vdef of the
1172 entry of the SESE region as the vuse of the call and the reaching
1173 vdef of the exit of the SESE region as the vdef of the call. */
1175 for (gsi
= gsi_start_bb (return_bb
); !gsi_end_p (gsi
); gsi_next (&gsi
))
1177 gimple stmt
= gsi_stmt (gsi
);
1178 if (gimple_vuse (stmt
))
1180 gimple_set_vuse (stmt
, NULL_TREE
);
1183 if (gimple_vdef (stmt
))
1188 /* Now create the actual clone. */
1189 rebuild_cgraph_edges ();
1190 node
= cgraph_function_versioning (cur_node
, NULL
, NULL
, args_to_skip
,
1191 !split_part_return_p
,
1192 split_point
->split_bbs
,
1193 split_point
->entry_bb
, "part");
1194 /* For usual cloning it is enough to clear builtin only when signature
1195 changes. For partial inlining we however can not expect the part
1196 of builtin implementation to have same semantic as the whole. */
1197 if (DECL_BUILT_IN (node
->symbol
.decl
))
1199 DECL_BUILT_IN_CLASS (node
->symbol
.decl
) = NOT_BUILT_IN
;
1200 DECL_FUNCTION_CODE (node
->symbol
.decl
) = (enum built_in_function
) 0;
1202 cgraph_node_remove_callees (cur_node
);
1203 if (!split_part_return_p
)
1204 TREE_THIS_VOLATILE (node
->symbol
.decl
) = 1;
1206 dump_function_to_file (node
->symbol
.decl
, dump_file
, dump_flags
);
1208 /* Create the basic block we place call into. It is the entry basic block
1209 split after last label. */
1210 call_bb
= split_point
->entry_bb
;
1211 for (gsi
= gsi_start_bb (call_bb
); !gsi_end_p (gsi
);)
1212 if (gimple_code (gsi_stmt (gsi
)) == GIMPLE_LABEL
)
1214 last_stmt
= gsi_stmt (gsi
);
1219 e
= split_block (split_point
->entry_bb
, last_stmt
);
1222 /* Produce the call statement. */
1223 gsi
= gsi_last_bb (call_bb
);
1224 FOR_EACH_VEC_ELT (tree
, args_to_pass
, i
, arg
)
1225 if (!is_gimple_val (arg
))
1227 arg
= force_gimple_operand_gsi (&gsi
, arg
, true, NULL_TREE
,
1228 false, GSI_CONTINUE_LINKING
);
1229 VEC_replace (tree
, args_to_pass
, i
, arg
);
1231 call
= gimple_build_call_vec (node
->symbol
.decl
, args_to_pass
);
1232 gimple_set_block (call
, DECL_INITIAL (current_function_decl
));
1233 VEC_free (tree
, heap
, args_to_pass
);
1235 /* We avoid address being taken on any variable used by split part,
1236 so return slot optimization is always possible. Moreover this is
1237 required to make DECL_BY_REFERENCE work. */
1238 if (aggregate_value_p (DECL_RESULT (current_function_decl
),
1239 TREE_TYPE (current_function_decl
)))
1240 gimple_call_set_return_slot_opt (call
, true);
1242 /* Update return value. This is bit tricky. When we do not return,
1243 do nothing. When we return we might need to update return_bb
1244 or produce a new return statement. */
1245 if (!split_part_return_p
)
1246 gsi_insert_after (&gsi
, call
, GSI_NEW_STMT
);
1249 e
= make_edge (call_bb
, return_bb
,
1250 return_bb
== EXIT_BLOCK_PTR
? 0 : EDGE_FALLTHRU
);
1251 e
->count
= call_bb
->count
;
1252 e
->probability
= REG_BR_PROB_BASE
;
1254 /* If there is return basic block, see what value we need to store
1255 return value into and put call just before it. */
1256 if (return_bb
!= EXIT_BLOCK_PTR
)
1258 real_retval
= retval
= find_retval (return_bb
);
1260 if (real_retval
&& split_point
->split_part_set_retval
)
1262 gimple_stmt_iterator psi
;
1264 /* See if we need new SSA_NAME for the result.
1265 When DECL_BY_REFERENCE is true, retval is actually pointer to
1266 return value and it is constant in whole function. */
1267 if (TREE_CODE (retval
) == SSA_NAME
1268 && !DECL_BY_REFERENCE (DECL_RESULT (current_function_decl
)))
1270 retval
= copy_ssa_name (retval
, call
);
1272 /* See if there is PHI defining return value. */
1273 for (psi
= gsi_start_phis (return_bb
);
1274 !gsi_end_p (psi
); gsi_next (&psi
))
1275 if (!virtual_operand_p (gimple_phi_result (gsi_stmt (psi
))))
1278 /* When there is PHI, just update its value. */
1279 if (TREE_CODE (retval
) == SSA_NAME
1280 && !gsi_end_p (psi
))
1281 add_phi_arg (gsi_stmt (psi
), retval
, e
, UNKNOWN_LOCATION
);
1282 /* Otherwise update the return BB itself.
1283 find_return_bb allows at most one assignment to return value,
1284 so update first statement. */
1287 gimple_stmt_iterator bsi
;
1288 for (bsi
= gsi_start_bb (return_bb
); !gsi_end_p (bsi
);
1290 if (gimple_code (gsi_stmt (bsi
)) == GIMPLE_RETURN
)
1292 gimple_return_set_retval (gsi_stmt (bsi
), retval
);
1295 else if (gimple_code (gsi_stmt (bsi
)) == GIMPLE_ASSIGN
1296 && !gimple_clobber_p (gsi_stmt (bsi
)))
1298 gimple_assign_set_rhs1 (gsi_stmt (bsi
), retval
);
1301 update_stmt (gsi_stmt (bsi
));
1304 if (DECL_BY_REFERENCE (DECL_RESULT (current_function_decl
)))
1306 gimple_call_set_lhs (call
, build_simple_mem_ref (retval
));
1307 gsi_insert_after (&gsi
, call
, GSI_NEW_STMT
);
1312 restype
= TREE_TYPE (DECL_RESULT (current_function_decl
));
1313 gsi_insert_after (&gsi
, call
, GSI_NEW_STMT
);
1314 if (!useless_type_conversion_p (TREE_TYPE (retval
), restype
))
1317 tree tem
= create_tmp_reg (restype
, NULL
);
1318 tem
= make_ssa_name (tem
, call
);
1319 cpy
= gimple_build_assign_with_ops (NOP_EXPR
, retval
,
1321 gsi_insert_after (&gsi
, cpy
, GSI_NEW_STMT
);
1324 gimple_call_set_lhs (call
, retval
);
1329 gsi_insert_after (&gsi
, call
, GSI_NEW_STMT
);
1331 /* We don't use return block (there is either no return in function or
1332 multiple of them). So create new basic block with return statement.
1337 if (split_point
->split_part_set_retval
1338 && !VOID_TYPE_P (TREE_TYPE (TREE_TYPE (current_function_decl
))))
1340 retval
= DECL_RESULT (current_function_decl
);
1342 /* We use temporary register to hold value when aggregate_value_p
1343 is false. Similarly for DECL_BY_REFERENCE we must avoid extra
1345 if (!aggregate_value_p (retval
, TREE_TYPE (current_function_decl
))
1346 && !DECL_BY_REFERENCE (retval
))
1347 retval
= create_tmp_reg (TREE_TYPE (retval
), NULL
);
1348 if (is_gimple_reg (retval
))
1350 /* When returning by reference, there is only one SSA name
1351 assigned to RESULT_DECL (that is pointer to return value).
1352 Look it up or create new one if it is missing. */
1353 if (DECL_BY_REFERENCE (retval
))
1354 retval
= get_or_create_ssa_default_def (cfun
, retval
);
1355 /* Otherwise produce new SSA name for return value. */
1357 retval
= make_ssa_name (retval
, call
);
1359 if (DECL_BY_REFERENCE (DECL_RESULT (current_function_decl
)))
1360 gimple_call_set_lhs (call
, build_simple_mem_ref (retval
));
1362 gimple_call_set_lhs (call
, retval
);
1364 gsi_insert_after (&gsi
, call
, GSI_NEW_STMT
);
1365 ret
= gimple_build_return (retval
);
1366 gsi_insert_after (&gsi
, ret
, GSI_NEW_STMT
);
1369 free_dominance_info (CDI_DOMINATORS
);
1370 free_dominance_info (CDI_POST_DOMINATORS
);
1371 compute_inline_parameters (node
, true);
1374 /* Execute function splitting pass. */
1377 execute_split_functions (void)
1379 gimple_stmt_iterator bsi
;
1381 int overall_time
= 0, overall_size
= 0;
1383 struct cgraph_node
*node
= cgraph_get_node (current_function_decl
);
1385 if (flags_from_decl_or_type (current_function_decl
)
1386 & (ECF_NORETURN
|ECF_MALLOC
))
1389 fprintf (dump_file
, "Not splitting: noreturn/malloc function.\n");
1392 if (MAIN_NAME_P (DECL_NAME (current_function_decl
)))
1395 fprintf (dump_file
, "Not splitting: main function.\n");
1398 /* This can be relaxed; function might become inlinable after splitting
1399 away the uninlinable part. */
1400 if (inline_edge_summary_vec
&& !inline_summary (node
)->inlinable
)
1403 fprintf (dump_file
, "Not splitting: not inlinable.\n");
1406 if (DECL_DISREGARD_INLINE_LIMITS (node
->symbol
.decl
))
1409 fprintf (dump_file
, "Not splitting: disregarding inline limits.\n");
1412 /* This can be relaxed; most of versioning tests actually prevents
1414 if (!tree_versionable_function_p (current_function_decl
))
1417 fprintf (dump_file
, "Not splitting: not versionable.\n");
1420 /* FIXME: we could support this. */
1421 if (DECL_STRUCT_FUNCTION (current_function_decl
)->static_chain_decl
)
1424 fprintf (dump_file
, "Not splitting: nested function.\n");
1428 /* See if it makes sense to try to split.
1429 It makes sense to split if we inline, that is if we have direct calls to
1430 handle or direct calls are possibly going to appear as result of indirect
1431 inlining or LTO. Also handle -fprofile-generate as LTO to allow non-LTO
1432 training for LTO -fprofile-use build.
1434 Note that we are not completely conservative about disqualifying functions
1435 called once. It is possible that the caller is called more then once and
1436 then inlining would still benefit. */
1437 if ((!node
->callers
|| !node
->callers
->next_caller
)
1438 && !node
->symbol
.address_taken
1439 && (!flag_lto
|| !node
->symbol
.externally_visible
))
1442 fprintf (dump_file
, "Not splitting: not called directly "
1443 "or called once.\n");
1447 /* FIXME: We can actually split if splitting reduces call overhead. */
1448 if (!flag_inline_small_functions
1449 && !DECL_DECLARED_INLINE_P (current_function_decl
))
1452 fprintf (dump_file
, "Not splitting: not autoinlining and function"
1453 " is not inline.\n");
1457 /* Initialize bitmap to track forbidden calls. */
1458 forbidden_dominators
= BITMAP_ALLOC (NULL
);
1459 calculate_dominance_info (CDI_DOMINATORS
);
1461 /* Compute local info about basic blocks and determine function size/time. */
1462 VEC_safe_grow_cleared (bb_info
, heap
, bb_info_vec
, last_basic_block
+ 1);
1463 memset (&best_split_point
, 0, sizeof (best_split_point
));
1468 int freq
= compute_call_stmt_bb_frequency (current_function_decl
, bb
);
1470 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
1471 fprintf (dump_file
, "Basic block %i\n", bb
->index
);
1473 for (bsi
= gsi_start_bb (bb
); !gsi_end_p (bsi
); gsi_next (&bsi
))
1475 int this_time
, this_size
;
1476 gimple stmt
= gsi_stmt (bsi
);
1478 this_size
= estimate_num_insns (stmt
, &eni_size_weights
);
1479 this_time
= estimate_num_insns (stmt
, &eni_time_weights
) * freq
;
1482 check_forbidden_calls (stmt
);
1484 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
1486 fprintf (dump_file
, " freq:%6i size:%3i time:%3i ",
1487 freq
, this_size
, this_time
);
1488 print_gimple_stmt (dump_file
, stmt
, 0, 0);
1491 overall_time
+= time
;
1492 overall_size
+= size
;
1493 VEC_index (bb_info
, bb_info_vec
, bb
->index
).time
= time
;
1494 VEC_index (bb_info
, bb_info_vec
, bb
->index
).size
= size
;
1496 find_split_points (overall_time
, overall_size
);
1497 if (best_split_point
.split_bbs
)
1499 split_function (&best_split_point
);
1500 BITMAP_FREE (best_split_point
.ssa_names_to_pass
);
1501 BITMAP_FREE (best_split_point
.split_bbs
);
1502 todo
= TODO_update_ssa
| TODO_cleanup_cfg
;
1504 BITMAP_FREE (forbidden_dominators
);
1505 VEC_free (bb_info
, heap
, bb_info_vec
);
1510 /* Gate function splitting pass. When doing profile feedback, we want
1511 to execute the pass after profiling is read. So disable one in
1512 early optimization. */
1515 gate_split_functions (void)
1517 return (flag_partial_inlining
1518 && !profile_arc_flag
&& !flag_branch_probabilities
);
1521 struct gimple_opt_pass pass_split_functions
=
1525 "fnsplit", /* name */
1526 gate_split_functions
, /* gate */
1527 execute_split_functions
, /* execute */
1530 0, /* static_pass_number */
1531 TV_IPA_FNSPLIT
, /* tv_id */
1532 PROP_cfg
, /* properties_required */
1533 0, /* properties_provided */
1534 0, /* properties_destroyed */
1535 0, /* todo_flags_start */
1536 TODO_verify_all
/* todo_flags_finish */
1540 /* Gate feedback driven function splitting pass.
1541 We don't need to split when profiling at all, we are producing
1542 lousy code anyway. */
1545 gate_feedback_split_functions (void)
1547 return (flag_partial_inlining
1548 && flag_branch_probabilities
);
1551 /* Execute function splitting pass. */
1554 execute_feedback_split_functions (void)
1556 unsigned int retval
= execute_split_functions ();
1558 retval
|= TODO_rebuild_cgraph_edges
;
1562 struct gimple_opt_pass pass_feedback_split_functions
=
1566 "feedback_fnsplit", /* name */
1567 gate_feedback_split_functions
, /* gate */
1568 execute_feedback_split_functions
, /* execute */
1571 0, /* static_pass_number */
1572 TV_IPA_FNSPLIT
, /* tv_id */
1573 PROP_cfg
, /* properties_required */
1574 0, /* properties_provided */
1575 0, /* properties_destroyed */
1576 0, /* todo_flags_start */
1577 TODO_verify_all
/* todo_flags_finish */