1 /* Pass computing data for optimizing stdarg functions.
2 Copyright (C) 2004, 2005, 2007, 2008, 2009, 2010, 2011
3 Free Software Foundation, Inc.
4 Contributed by Jakub Jelinek <jakub@redhat.com>
6 This file is part of GCC.
8 GCC is free software; you can redistribute it and/or modify
9 it under the terms of the GNU General Public License as published by
10 the Free Software Foundation; either version 3, or (at your option)
13 GCC is distributed in the hope that it will be useful,
14 but WITHOUT ANY WARRANTY; without even the implied warranty of
15 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 GNU General Public License for more details.
18 You should have received a copy of the GNU General Public License
19 along with GCC; see the file COPYING3. If not see
20 <http://www.gnu.org/licenses/>. */
24 #include "coretypes.h"
28 #include "langhooks.h"
29 #include "gimple-pretty-print.h"
31 #include "tree-flow.h"
32 #include "tree-pass.h"
33 #include "tree-stdarg.h"
35 /* A simple pass that attempts to optimize stdarg functions on architectures
36 that need to save register arguments to stack on entry to stdarg functions.
37 If the function doesn't use any va_start macros, no registers need to
38 be saved. If va_start macros are used, the va_list variables don't escape
39 the function, it is only necessary to save registers that will be used
40 in va_arg macros. E.g. if va_arg is only used with integral types
41 in the function, floating point registers don't need to be saved, etc. */
44 /* Return true if basic block VA_ARG_BB is dominated by VA_START_BB and
45 is executed at most as many times as VA_START_BB. */
48 reachable_at_most_once (basic_block va_arg_bb
, basic_block va_start_bb
)
50 vec
<edge
> stack
= vec
<edge
>();
56 if (va_arg_bb
== va_start_bb
)
59 if (! dominated_by_p (CDI_DOMINATORS
, va_arg_bb
, va_start_bb
))
62 visited
= sbitmap_alloc (last_basic_block
);
63 bitmap_clear (visited
);
66 FOR_EACH_EDGE (e
, ei
, va_arg_bb
->preds
)
69 while (! stack
.is_empty ())
76 if (e
->flags
& EDGE_COMPLEX
)
82 if (src
== va_start_bb
)
85 /* va_arg_bb can be executed more times than va_start_bb. */
92 gcc_assert (src
!= ENTRY_BLOCK_PTR
);
94 if (! bitmap_bit_p (visited
, src
->index
))
96 bitmap_set_bit (visited
, src
->index
);
97 FOR_EACH_EDGE (e
, ei
, src
->preds
)
103 sbitmap_free (visited
);
108 /* For statement COUNTER = RHS, if RHS is COUNTER + constant,
109 return constant, otherwise return (unsigned HOST_WIDE_INT) -1.
110 GPR_P is true if this is GPR counter. */
112 static unsigned HOST_WIDE_INT
113 va_list_counter_bump (struct stdarg_info
*si
, tree counter
, tree rhs
,
118 unsigned HOST_WIDE_INT ret
= 0, val
, counter_val
;
119 unsigned int max_size
;
121 if (si
->offsets
== NULL
)
125 si
->offsets
= XNEWVEC (int, num_ssa_names
);
126 for (i
= 0; i
< num_ssa_names
; ++i
)
130 counter_val
= gpr_p
? cfun
->va_list_gpr_size
: cfun
->va_list_fpr_size
;
131 max_size
= gpr_p
? VA_LIST_MAX_GPR_SIZE
: VA_LIST_MAX_FPR_SIZE
;
132 orig_lhs
= lhs
= rhs
;
135 enum tree_code rhs_code
;
138 if (si
->offsets
[SSA_NAME_VERSION (lhs
)] != -1)
140 if (counter_val
>= max_size
)
146 ret
-= counter_val
- si
->offsets
[SSA_NAME_VERSION (lhs
)];
150 stmt
= SSA_NAME_DEF_STMT (lhs
);
152 if (!is_gimple_assign (stmt
) || gimple_assign_lhs (stmt
) != lhs
)
153 return (unsigned HOST_WIDE_INT
) -1;
155 rhs_code
= gimple_assign_rhs_code (stmt
);
156 rhs1
= gimple_assign_rhs1 (stmt
);
157 if ((get_gimple_rhs_class (rhs_code
) == GIMPLE_SINGLE_RHS
158 || gimple_assign_cast_p (stmt
))
159 && TREE_CODE (rhs1
) == SSA_NAME
)
165 if ((rhs_code
== POINTER_PLUS_EXPR
166 || rhs_code
== PLUS_EXPR
)
167 && TREE_CODE (rhs1
) == SSA_NAME
168 && host_integerp (gimple_assign_rhs2 (stmt
), 1))
170 ret
+= tree_low_cst (gimple_assign_rhs2 (stmt
), 1);
175 if (rhs_code
== ADDR_EXPR
176 && TREE_CODE (TREE_OPERAND (rhs1
, 0)) == MEM_REF
177 && TREE_CODE (TREE_OPERAND (TREE_OPERAND (rhs1
, 0), 0)) == SSA_NAME
178 && host_integerp (TREE_OPERAND (TREE_OPERAND (rhs1
, 0), 1), 1))
180 ret
+= tree_low_cst (TREE_OPERAND (TREE_OPERAND (rhs1
, 0), 1), 1);
181 lhs
= TREE_OPERAND (TREE_OPERAND (rhs1
, 0), 0);
185 if (get_gimple_rhs_class (rhs_code
) != GIMPLE_SINGLE_RHS
)
186 return (unsigned HOST_WIDE_INT
) -1;
188 rhs
= gimple_assign_rhs1 (stmt
);
189 if (TREE_CODE (counter
) != TREE_CODE (rhs
))
190 return (unsigned HOST_WIDE_INT
) -1;
192 if (TREE_CODE (counter
) == COMPONENT_REF
)
194 if (get_base_address (counter
) != get_base_address (rhs
)
195 || TREE_CODE (TREE_OPERAND (rhs
, 1)) != FIELD_DECL
196 || TREE_OPERAND (counter
, 1) != TREE_OPERAND (rhs
, 1))
197 return (unsigned HOST_WIDE_INT
) -1;
199 else if (counter
!= rhs
)
200 return (unsigned HOST_WIDE_INT
) -1;
206 val
= ret
+ counter_val
;
209 enum tree_code rhs_code
;
212 if (si
->offsets
[SSA_NAME_VERSION (lhs
)] != -1)
216 si
->offsets
[SSA_NAME_VERSION (lhs
)] = max_size
;
218 si
->offsets
[SSA_NAME_VERSION (lhs
)] = val
;
220 stmt
= SSA_NAME_DEF_STMT (lhs
);
222 rhs_code
= gimple_assign_rhs_code (stmt
);
223 rhs1
= gimple_assign_rhs1 (stmt
);
224 if ((get_gimple_rhs_class (rhs_code
) == GIMPLE_SINGLE_RHS
225 || gimple_assign_cast_p (stmt
))
226 && TREE_CODE (rhs1
) == SSA_NAME
)
232 if ((rhs_code
== POINTER_PLUS_EXPR
233 || rhs_code
== PLUS_EXPR
)
234 && TREE_CODE (rhs1
) == SSA_NAME
235 && host_integerp (gimple_assign_rhs2 (stmt
), 1))
237 val
-= tree_low_cst (gimple_assign_rhs2 (stmt
), 1);
242 if (rhs_code
== ADDR_EXPR
243 && TREE_CODE (TREE_OPERAND (rhs1
, 0)) == MEM_REF
244 && TREE_CODE (TREE_OPERAND (TREE_OPERAND (rhs1
, 0), 0)) == SSA_NAME
245 && host_integerp (TREE_OPERAND (TREE_OPERAND (rhs1
, 0), 1), 1))
247 val
-= tree_low_cst (TREE_OPERAND (TREE_OPERAND (rhs1
, 0), 1), 1);
248 lhs
= TREE_OPERAND (TREE_OPERAND (rhs1
, 0), 0);
259 /* Called by walk_tree to look for references to va_list variables. */
262 find_va_list_reference (tree
*tp
, int *walk_subtrees ATTRIBUTE_UNUSED
,
265 bitmap va_list_vars
= (bitmap
) ((struct walk_stmt_info
*) data
)->info
;
268 if (TREE_CODE (var
) == SSA_NAME
)
270 if (bitmap_bit_p (va_list_vars
, SSA_NAME_VERSION (var
)))
273 else if (TREE_CODE (var
) == VAR_DECL
)
275 if (bitmap_bit_p (va_list_vars
, DECL_UID (var
) + num_ssa_names
))
283 /* Helper function of va_list_counter_struct_op. Compute
284 cfun->va_list_{g,f}pr_size. AP is a va_list GPR/FPR counter,
285 if WRITE_P is true, seen in AP = VAR, otherwise seen in VAR = AP
286 statement. GPR_P is true if AP is a GPR counter, false if it is
290 va_list_counter_op (struct stdarg_info
*si
, tree ap
, tree var
, bool gpr_p
,
293 unsigned HOST_WIDE_INT increment
;
295 if (si
->compute_sizes
< 0)
297 si
->compute_sizes
= 0;
298 if (si
->va_start_count
== 1
299 && reachable_at_most_once (si
->bb
, si
->va_start_bb
))
300 si
->compute_sizes
= 1;
302 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
304 "bb%d will %sbe executed at most once for each va_start "
305 "in bb%d\n", si
->bb
->index
, si
->compute_sizes
? "" : "not ",
306 si
->va_start_bb
->index
);
311 && (increment
= va_list_counter_bump (si
, ap
, var
, gpr_p
)) + 1 > 1)
313 if (gpr_p
&& cfun
->va_list_gpr_size
+ increment
< VA_LIST_MAX_GPR_SIZE
)
315 cfun
->va_list_gpr_size
+= increment
;
319 if (!gpr_p
&& cfun
->va_list_fpr_size
+ increment
< VA_LIST_MAX_FPR_SIZE
)
321 cfun
->va_list_fpr_size
+= increment
;
326 if (write_p
|| !si
->compute_sizes
)
329 cfun
->va_list_gpr_size
= VA_LIST_MAX_GPR_SIZE
;
331 cfun
->va_list_fpr_size
= VA_LIST_MAX_FPR_SIZE
;
336 /* If AP is a va_list GPR/FPR counter, compute cfun->va_list_{g,f}pr_size.
337 If WRITE_P is true, AP has been seen in AP = VAR assignment, if WRITE_P
338 is false, AP has been seen in VAR = AP assignment.
339 Return true if the AP = VAR (resp. VAR = AP) statement is a recognized
340 va_arg operation that doesn't cause the va_list variable to escape
344 va_list_counter_struct_op (struct stdarg_info
*si
, tree ap
, tree var
,
349 if (TREE_CODE (ap
) != COMPONENT_REF
350 || TREE_CODE (TREE_OPERAND (ap
, 1)) != FIELD_DECL
)
353 if (TREE_CODE (var
) != SSA_NAME
354 || bitmap_bit_p (si
->va_list_vars
, SSA_NAME_VERSION (var
)))
357 base
= get_base_address (ap
);
358 if (TREE_CODE (base
) != VAR_DECL
359 || !bitmap_bit_p (si
->va_list_vars
, DECL_UID (base
) + num_ssa_names
))
362 if (TREE_OPERAND (ap
, 1) == va_list_gpr_counter_field
)
363 va_list_counter_op (si
, ap
, var
, true, write_p
);
364 else if (TREE_OPERAND (ap
, 1) == va_list_fpr_counter_field
)
365 va_list_counter_op (si
, ap
, var
, false, write_p
);
371 /* Check for TEM = AP. Return true if found and the caller shouldn't
372 search for va_list references in the statement. */
375 va_list_ptr_read (struct stdarg_info
*si
, tree ap
, tree tem
)
377 if (TREE_CODE (ap
) != VAR_DECL
378 || !bitmap_bit_p (si
->va_list_vars
, DECL_UID (ap
) + num_ssa_names
))
381 if (TREE_CODE (tem
) != SSA_NAME
382 || bitmap_bit_p (si
->va_list_vars
, SSA_NAME_VERSION (tem
)))
385 if (si
->compute_sizes
< 0)
387 si
->compute_sizes
= 0;
388 if (si
->va_start_count
== 1
389 && reachable_at_most_once (si
->bb
, si
->va_start_bb
))
390 si
->compute_sizes
= 1;
392 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
394 "bb%d will %sbe executed at most once for each va_start "
395 "in bb%d\n", si
->bb
->index
, si
->compute_sizes
? "" : "not ",
396 si
->va_start_bb
->index
);
399 /* For void * or char * va_list types, there is just one counter.
400 If va_arg is used in a loop, we don't know how many registers need
402 if (! si
->compute_sizes
)
405 if (va_list_counter_bump (si
, ap
, tem
, true) == (unsigned HOST_WIDE_INT
) -1)
408 /* Note the temporary, as we need to track whether it doesn't escape
409 the current function. */
410 bitmap_set_bit (si
->va_list_escape_vars
, SSA_NAME_VERSION (tem
));
420 sequence and update cfun->va_list_gpr_size. Return true if found. */
423 va_list_ptr_write (struct stdarg_info
*si
, tree ap
, tree tem2
)
425 unsigned HOST_WIDE_INT increment
;
427 if (TREE_CODE (ap
) != VAR_DECL
428 || !bitmap_bit_p (si
->va_list_vars
, DECL_UID (ap
) + num_ssa_names
))
431 if (TREE_CODE (tem2
) != SSA_NAME
432 || bitmap_bit_p (si
->va_list_vars
, SSA_NAME_VERSION (tem2
)))
435 if (si
->compute_sizes
<= 0)
438 increment
= va_list_counter_bump (si
, ap
, tem2
, true);
439 if (increment
+ 1 <= 1)
442 if (cfun
->va_list_gpr_size
+ increment
< VA_LIST_MAX_GPR_SIZE
)
443 cfun
->va_list_gpr_size
+= increment
;
445 cfun
->va_list_gpr_size
= VA_LIST_MAX_GPR_SIZE
;
451 /* If RHS is X, (some type *) X or X + CST for X a temporary variable
452 containing value of some va_list variable plus optionally some constant,
453 either set si->va_list_escapes or add LHS to si->va_list_escape_vars,
454 depending whether LHS is a function local temporary. */
457 check_va_list_escapes (struct stdarg_info
*si
, tree lhs
, tree rhs
)
459 if (! POINTER_TYPE_P (TREE_TYPE (rhs
)))
462 if (TREE_CODE (rhs
) == SSA_NAME
)
464 if (! bitmap_bit_p (si
->va_list_escape_vars
, SSA_NAME_VERSION (rhs
)))
467 else if (TREE_CODE (rhs
) == ADDR_EXPR
468 && TREE_CODE (TREE_OPERAND (rhs
, 0)) == MEM_REF
469 && TREE_CODE (TREE_OPERAND (TREE_OPERAND (rhs
, 0), 0)) == SSA_NAME
)
471 tree ptr
= TREE_OPERAND (TREE_OPERAND (rhs
, 0), 0);
472 if (! bitmap_bit_p (si
->va_list_escape_vars
, SSA_NAME_VERSION (ptr
)))
478 if (TREE_CODE (lhs
) != SSA_NAME
)
480 si
->va_list_escapes
= true;
484 if (si
->compute_sizes
< 0)
486 si
->compute_sizes
= 0;
487 if (si
->va_start_count
== 1
488 && reachable_at_most_once (si
->bb
, si
->va_start_bb
))
489 si
->compute_sizes
= 1;
491 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
493 "bb%d will %sbe executed at most once for each va_start "
494 "in bb%d\n", si
->bb
->index
, si
->compute_sizes
? "" : "not ",
495 si
->va_start_bb
->index
);
498 /* For void * or char * va_list types, there is just one counter.
499 If va_arg is used in a loop, we don't know how many registers need
501 if (! si
->compute_sizes
)
503 si
->va_list_escapes
= true;
507 if (va_list_counter_bump (si
, si
->va_start_ap
, lhs
, true)
508 == (unsigned HOST_WIDE_INT
) -1)
510 si
->va_list_escapes
= true;
514 bitmap_set_bit (si
->va_list_escape_vars
, SSA_NAME_VERSION (lhs
));
518 /* Check all uses of temporaries from si->va_list_escape_vars bitmap.
519 Return true if va_list might be escaping. */
522 check_all_va_list_escapes (struct stdarg_info
*si
)
528 gimple_stmt_iterator i
;
530 for (i
= gsi_start_bb (bb
); !gsi_end_p (i
); gsi_next (&i
))
532 gimple stmt
= gsi_stmt (i
);
536 if (is_gimple_debug (stmt
))
539 FOR_EACH_SSA_TREE_OPERAND (use
, stmt
, iter
, SSA_OP_ALL_USES
)
541 if (! bitmap_bit_p (si
->va_list_escape_vars
,
542 SSA_NAME_VERSION (use
)))
545 if (is_gimple_assign (stmt
))
547 tree rhs
= gimple_assign_rhs1 (stmt
);
548 enum tree_code rhs_code
= gimple_assign_rhs_code (stmt
);
551 if (rhs_code
== MEM_REF
552 && TREE_OPERAND (rhs
, 0) == use
553 && TYPE_SIZE_UNIT (TREE_TYPE (rhs
))
554 && host_integerp (TYPE_SIZE_UNIT (TREE_TYPE (rhs
)), 1)
555 && si
->offsets
[SSA_NAME_VERSION (use
)] != -1)
557 unsigned HOST_WIDE_INT gpr_size
;
558 tree access_size
= TYPE_SIZE_UNIT (TREE_TYPE (rhs
));
560 gpr_size
= si
->offsets
[SSA_NAME_VERSION (use
)]
561 + tree_low_cst (TREE_OPERAND (rhs
, 1), 0)
562 + tree_low_cst (access_size
, 1);
563 if (gpr_size
>= VA_LIST_MAX_GPR_SIZE
)
564 cfun
->va_list_gpr_size
= VA_LIST_MAX_GPR_SIZE
;
565 else if (gpr_size
> cfun
->va_list_gpr_size
)
566 cfun
->va_list_gpr_size
= gpr_size
;
570 /* va_arg sequences may contain
571 other_ap_temp = ap_temp;
572 other_ap_temp = ap_temp + constant;
573 other_ap_temp = (some_type *) ap_temp;
577 && ((rhs_code
== POINTER_PLUS_EXPR
578 && (TREE_CODE (gimple_assign_rhs2 (stmt
))
580 || gimple_assign_cast_p (stmt
)
581 || (get_gimple_rhs_class (rhs_code
)
582 == GIMPLE_SINGLE_RHS
)))
584 tree lhs
= gimple_assign_lhs (stmt
);
586 if (TREE_CODE (lhs
) == SSA_NAME
587 && bitmap_bit_p (si
->va_list_escape_vars
,
588 SSA_NAME_VERSION (lhs
)))
591 if (TREE_CODE (lhs
) == VAR_DECL
592 && bitmap_bit_p (si
->va_list_vars
,
593 DECL_UID (lhs
) + num_ssa_names
))
596 else if (rhs_code
== ADDR_EXPR
597 && TREE_CODE (TREE_OPERAND (rhs
, 0)) == MEM_REF
598 && TREE_OPERAND (TREE_OPERAND (rhs
, 0), 0) == use
)
600 tree lhs
= gimple_assign_lhs (stmt
);
602 if (bitmap_bit_p (si
->va_list_escape_vars
,
603 SSA_NAME_VERSION (lhs
)))
608 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
610 fputs ("va_list escapes in ", dump_file
);
611 print_gimple_stmt (dump_file
, stmt
, 0, dump_flags
);
612 fputc ('\n', dump_file
);
623 /* Return true if this optimization pass should be done.
624 It makes only sense for stdarg functions. */
627 gate_optimize_stdarg (void)
629 /* This optimization is only for stdarg functions. */
630 return cfun
->stdarg
!= 0;
634 /* Entry point to the stdarg optimization pass. */
637 execute_optimize_stdarg (void)
640 bool va_list_escapes
= false;
641 bool va_list_simple_ptr
;
642 struct stdarg_info si
;
643 struct walk_stmt_info wi
;
644 const char *funcname
= NULL
;
647 cfun
->va_list_gpr_size
= 0;
648 cfun
->va_list_fpr_size
= 0;
649 memset (&si
, 0, sizeof (si
));
650 si
.va_list_vars
= BITMAP_ALLOC (NULL
);
651 si
.va_list_escape_vars
= BITMAP_ALLOC (NULL
);
654 funcname
= lang_hooks
.decl_printable_name (current_function_decl
, 2);
656 cfun_va_list
= targetm
.fn_abi_va_list (cfun
->decl
);
657 va_list_simple_ptr
= POINTER_TYPE_P (cfun_va_list
)
658 && (TREE_TYPE (cfun_va_list
) == void_type_node
659 || TREE_TYPE (cfun_va_list
) == char_type_node
);
660 gcc_assert (is_gimple_reg_type (cfun_va_list
) == va_list_simple_ptr
);
664 gimple_stmt_iterator i
;
666 for (i
= gsi_start_bb (bb
); !gsi_end_p (i
); gsi_next (&i
))
668 gimple stmt
= gsi_stmt (i
);
671 if (!is_gimple_call (stmt
))
674 callee
= gimple_call_fndecl (stmt
);
676 || DECL_BUILT_IN_CLASS (callee
) != BUILT_IN_NORMAL
)
679 switch (DECL_FUNCTION_CODE (callee
))
681 case BUILT_IN_VA_START
:
683 /* If old style builtins are used, don't optimize anything. */
684 case BUILT_IN_SAVEREGS
:
685 case BUILT_IN_NEXT_ARG
:
686 va_list_escapes
= true;
693 ap
= gimple_call_arg (stmt
, 0);
695 if (TREE_CODE (ap
) != ADDR_EXPR
)
697 va_list_escapes
= true;
700 ap
= TREE_OPERAND (ap
, 0);
701 if (TREE_CODE (ap
) == ARRAY_REF
)
703 if (! integer_zerop (TREE_OPERAND (ap
, 1)))
705 va_list_escapes
= true;
708 ap
= TREE_OPERAND (ap
, 0);
710 if (TYPE_MAIN_VARIANT (TREE_TYPE (ap
))
711 != TYPE_MAIN_VARIANT (targetm
.fn_abi_va_list (cfun
->decl
))
712 || TREE_CODE (ap
) != VAR_DECL
)
714 va_list_escapes
= true;
718 if (is_global_var (ap
))
720 va_list_escapes
= true;
724 bitmap_set_bit (si
.va_list_vars
, DECL_UID (ap
) + num_ssa_names
);
726 /* VA_START_BB and VA_START_AP will be only used if there is just
727 one va_start in the function. */
736 /* If there were no va_start uses in the function, there is no need to
738 if (si
.va_start_count
== 0)
741 /* If some va_list arguments weren't local, we can't optimize. */
745 /* For void * or char * va_list, something useful can be done only
746 if there is just one va_start. */
747 if (va_list_simple_ptr
&& si
.va_start_count
> 1)
749 va_list_escapes
= true;
753 /* For struct * va_list, if the backend didn't tell us what the counter fields
754 are, there is nothing more we can do. */
755 if (!va_list_simple_ptr
756 && va_list_gpr_counter_field
== NULL_TREE
757 && va_list_fpr_counter_field
== NULL_TREE
)
759 va_list_escapes
= true;
763 /* For void * or char * va_list there is just one counter
764 (va_list itself). Use VA_LIST_GPR_SIZE for it. */
765 if (va_list_simple_ptr
)
766 cfun
->va_list_fpr_size
= VA_LIST_MAX_FPR_SIZE
;
768 calculate_dominance_info (CDI_DOMINATORS
);
769 memset (&wi
, 0, sizeof (wi
));
770 wi
.info
= si
.va_list_vars
;
774 gimple_stmt_iterator i
;
776 si
.compute_sizes
= -1;
779 /* For va_list_simple_ptr, we have to check PHI nodes too. We treat
780 them as assignments for the purpose of escape analysis. This is
781 not needed for non-simple va_list because virtual phis don't perform
782 any real data movement. */
783 if (va_list_simple_ptr
)
789 for (i
= gsi_start_phis (bb
); !gsi_end_p (i
); gsi_next (&i
))
791 gimple phi
= gsi_stmt (i
);
792 lhs
= PHI_RESULT (phi
);
794 if (virtual_operand_p (lhs
))
797 FOR_EACH_PHI_ARG (uop
, phi
, soi
, SSA_OP_USE
)
799 rhs
= USE_FROM_PTR (uop
);
800 if (va_list_ptr_read (&si
, rhs
, lhs
))
802 else if (va_list_ptr_write (&si
, lhs
, rhs
))
805 check_va_list_escapes (&si
, lhs
, rhs
);
807 if (si
.va_list_escapes
)
809 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
811 fputs ("va_list escapes in ", dump_file
);
812 print_gimple_stmt (dump_file
, phi
, 0, dump_flags
);
813 fputc ('\n', dump_file
);
815 va_list_escapes
= true;
821 for (i
= gsi_start_bb (bb
);
822 !gsi_end_p (i
) && !va_list_escapes
;
825 gimple stmt
= gsi_stmt (i
);
827 /* Don't look at __builtin_va_{start,end}, they are ok. */
828 if (is_gimple_call (stmt
))
830 tree callee
= gimple_call_fndecl (stmt
);
833 && DECL_BUILT_IN_CLASS (callee
) == BUILT_IN_NORMAL
834 && (DECL_FUNCTION_CODE (callee
) == BUILT_IN_VA_START
835 || DECL_FUNCTION_CODE (callee
) == BUILT_IN_VA_END
))
839 if (is_gimple_assign (stmt
))
841 tree lhs
= gimple_assign_lhs (stmt
);
842 tree rhs
= gimple_assign_rhs1 (stmt
);
844 if (va_list_simple_ptr
)
846 if (get_gimple_rhs_class (gimple_assign_rhs_code (stmt
))
847 == GIMPLE_SINGLE_RHS
)
849 /* Check for ap ={v} {}. */
850 if (TREE_CLOBBER_P (rhs
))
853 /* Check for tem = ap. */
854 else if (va_list_ptr_read (&si
, rhs
, lhs
))
857 /* Check for the last insn in:
862 else if (va_list_ptr_write (&si
, lhs
, rhs
))
866 if ((gimple_assign_rhs_code (stmt
) == POINTER_PLUS_EXPR
867 && TREE_CODE (gimple_assign_rhs2 (stmt
)) == INTEGER_CST
)
868 || CONVERT_EXPR_CODE_P (gimple_assign_rhs_code (stmt
))
869 || (get_gimple_rhs_class (gimple_assign_rhs_code (stmt
))
870 == GIMPLE_SINGLE_RHS
))
871 check_va_list_escapes (&si
, lhs
, rhs
);
875 if (get_gimple_rhs_class (gimple_assign_rhs_code (stmt
))
876 == GIMPLE_SINGLE_RHS
)
878 /* Check for ap ={v} {}. */
879 if (TREE_CLOBBER_P (rhs
))
882 /* Check for ap[0].field = temp. */
883 else if (va_list_counter_struct_op (&si
, lhs
, rhs
, true))
886 /* Check for temp = ap[0].field. */
887 else if (va_list_counter_struct_op (&si
, rhs
, lhs
,
892 /* Do any architecture specific checking. */
893 if (targetm
.stdarg_optimize_hook
894 && targetm
.stdarg_optimize_hook (&si
, stmt
))
898 else if (is_gimple_debug (stmt
))
901 /* All other uses of va_list are either va_copy (that is not handled
902 in this optimization), taking address of va_list variable or
903 passing va_list to other functions (in that case va_list might
904 escape the function and therefore va_start needs to set it up
905 fully), or some unexpected use of va_list. None of these should
906 happen in a gimplified VA_ARG_EXPR. */
907 if (si
.va_list_escapes
908 || walk_gimple_op (stmt
, find_va_list_reference
, &wi
))
910 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
912 fputs ("va_list escapes in ", dump_file
);
913 print_gimple_stmt (dump_file
, stmt
, 0, dump_flags
);
914 fputc ('\n', dump_file
);
916 va_list_escapes
= true;
924 if (! va_list_escapes
925 && va_list_simple_ptr
926 && ! bitmap_empty_p (si
.va_list_escape_vars
)
927 && check_all_va_list_escapes (&si
))
928 va_list_escapes
= true;
933 cfun
->va_list_gpr_size
= VA_LIST_MAX_GPR_SIZE
;
934 cfun
->va_list_fpr_size
= VA_LIST_MAX_FPR_SIZE
;
936 BITMAP_FREE (si
.va_list_vars
);
937 BITMAP_FREE (si
.va_list_escape_vars
);
941 fprintf (dump_file
, "%s: va_list escapes %d, needs to save ",
942 funcname
, (int) va_list_escapes
);
943 if (cfun
->va_list_gpr_size
>= VA_LIST_MAX_GPR_SIZE
)
944 fputs ("all", dump_file
);
946 fprintf (dump_file
, "%d", cfun
->va_list_gpr_size
);
947 fputs (" GPR units and ", dump_file
);
948 if (cfun
->va_list_fpr_size
>= VA_LIST_MAX_FPR_SIZE
)
949 fputs ("all", dump_file
);
951 fprintf (dump_file
, "%d", cfun
->va_list_fpr_size
);
952 fputs (" FPR units.\n", dump_file
);
958 struct gimple_opt_pass pass_stdarg
=
963 OPTGROUP_NONE
, /* optinfo_flags */
964 gate_optimize_stdarg
, /* gate */
965 execute_optimize_stdarg
, /* execute */
968 0, /* static_pass_number */
970 PROP_cfg
| PROP_ssa
, /* properties_required */
971 0, /* properties_provided */
972 0, /* properties_destroyed */
973 0, /* todo_flags_start */
974 0 /* todo_flags_finish */