1 /* __builtin_object_size (ptr, object_size_type) computation
2 Copyright (C) 2004-2017 Free Software Foundation, Inc.
3 Contributed by Jakub Jelinek <jakub@redhat.com>
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License as published by
9 the Free Software Foundation; either version 3, or (at your option)
12 GCC is distributed in the hope that it will be useful,
13 but WITHOUT ANY WARRANTY; without even the implied warranty of
14 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 GNU General Public License for more details.
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
23 #include "coretypes.h"
27 #include "tree-pass.h"
29 #include "gimple-pretty-print.h"
30 #include "fold-const.h"
31 #include "tree-object-size.h"
32 #include "gimple-fold.h"
33 #include "gimple-iterator.h"
35 #include "stringpool.h"
38 struct object_size_info
43 bitmap visited
, reexamine
;
45 unsigned int *stack
, *tos
;
48 static const unsigned HOST_WIDE_INT unknown
[4] = {
55 static tree
compute_object_offset (const_tree
, const_tree
);
56 static bool addr_object_size (struct object_size_info
*,
57 const_tree
, int, unsigned HOST_WIDE_INT
*);
58 static unsigned HOST_WIDE_INT
alloc_object_size (const gcall
*, int);
59 static tree
pass_through_call (const gcall
*);
60 static void collect_object_sizes_for (struct object_size_info
*, tree
);
61 static void expr_object_size (struct object_size_info
*, tree
, tree
);
62 static bool merge_object_sizes (struct object_size_info
*, tree
, tree
,
63 unsigned HOST_WIDE_INT
);
64 static bool plus_stmt_object_size (struct object_size_info
*, tree
, gimple
*);
65 static bool cond_expr_object_size (struct object_size_info
*, tree
, gimple
*);
66 static void init_offset_limit (void);
67 static void check_for_plus_in_loops (struct object_size_info
*, tree
);
68 static void check_for_plus_in_loops_1 (struct object_size_info
*, tree
,
71 /* object_sizes[0] is upper bound for number of bytes till the end of
73 object_sizes[1] is upper bound for number of bytes till the end of
74 the subobject (innermost array or field with address taken).
75 object_sizes[2] is lower bound for number of bytes till the end of
76 the object and object_sizes[3] lower bound for subobject. */
77 static vec
<unsigned HOST_WIDE_INT
> object_sizes
[4];
79 /* Bitmaps what object sizes have been computed already. */
80 static bitmap computed
[4];
82 /* Maximum value of offset we consider to be addition. */
83 static unsigned HOST_WIDE_INT offset_limit
;
86 /* Initialize OFFSET_LIMIT variable. */
88 init_offset_limit (void)
90 if (tree_fits_uhwi_p (TYPE_MAX_VALUE (sizetype
)))
91 offset_limit
= tree_to_uhwi (TYPE_MAX_VALUE (sizetype
));
98 /* Compute offset of EXPR within VAR. Return error_mark_node
102 compute_object_offset (const_tree expr
, const_tree var
)
104 enum tree_code code
= PLUS_EXPR
;
108 return size_zero_node
;
110 switch (TREE_CODE (expr
))
113 base
= compute_object_offset (TREE_OPERAND (expr
, 0), var
);
114 if (base
== error_mark_node
)
117 t
= TREE_OPERAND (expr
, 1);
118 off
= size_binop (PLUS_EXPR
, DECL_FIELD_OFFSET (t
),
119 size_int (tree_to_uhwi (DECL_FIELD_BIT_OFFSET (t
))
125 case VIEW_CONVERT_EXPR
:
126 case NON_LVALUE_EXPR
:
127 return compute_object_offset (TREE_OPERAND (expr
, 0), var
);
130 base
= compute_object_offset (TREE_OPERAND (expr
, 0), var
);
131 if (base
== error_mark_node
)
134 off
= TYPE_SIZE_UNIT (TREE_TYPE (expr
));
138 base
= compute_object_offset (TREE_OPERAND (expr
, 0), var
);
139 if (base
== error_mark_node
)
142 t
= TREE_OPERAND (expr
, 1);
143 tree low_bound
, unit_size
;
144 low_bound
= array_ref_low_bound (CONST_CAST_TREE (expr
));
145 unit_size
= array_ref_element_size (CONST_CAST_TREE (expr
));
146 if (! integer_zerop (low_bound
))
147 t
= fold_build2 (MINUS_EXPR
, TREE_TYPE (t
), t
, low_bound
);
148 if (TREE_CODE (t
) == INTEGER_CST
&& tree_int_cst_sgn (t
) < 0)
151 t
= fold_build1 (NEGATE_EXPR
, TREE_TYPE (t
), t
);
153 t
= fold_convert (sizetype
, t
);
154 off
= size_binop (MULT_EXPR
, unit_size
, t
);
158 gcc_assert (TREE_CODE (TREE_OPERAND (expr
, 0)) == ADDR_EXPR
);
159 return wide_int_to_tree (sizetype
, mem_ref_offset (expr
));
162 return error_mark_node
;
165 return size_binop (code
, base
, off
);
169 /* Compute __builtin_object_size for PTR, which is a ADDR_EXPR.
170 OBJECT_SIZE_TYPE is the second argument from __builtin_object_size.
171 If unknown, return unknown[object_size_type]. */
174 addr_object_size (struct object_size_info
*osi
, const_tree ptr
,
175 int object_size_type
, unsigned HOST_WIDE_INT
*psize
)
177 tree pt_var
, pt_var_size
= NULL_TREE
, var_size
, bytes
;
179 gcc_assert (TREE_CODE (ptr
) == ADDR_EXPR
);
181 /* Set to unknown and overwrite just before returning if the size
182 could be determined. */
183 *psize
= unknown
[object_size_type
];
185 pt_var
= TREE_OPERAND (ptr
, 0);
186 while (handled_component_p (pt_var
))
187 pt_var
= TREE_OPERAND (pt_var
, 0);
190 && TREE_CODE (pt_var
) == MEM_REF
)
192 unsigned HOST_WIDE_INT sz
;
194 if (!osi
|| (object_size_type
& 1) != 0
195 || TREE_CODE (TREE_OPERAND (pt_var
, 0)) != SSA_NAME
)
197 compute_builtin_object_size (TREE_OPERAND (pt_var
, 0),
198 object_size_type
& ~1, &sz
);
202 tree var
= TREE_OPERAND (pt_var
, 0);
204 collect_object_sizes_for (osi
, var
);
205 if (bitmap_bit_p (computed
[object_size_type
],
206 SSA_NAME_VERSION (var
)))
207 sz
= object_sizes
[object_size_type
][SSA_NAME_VERSION (var
)];
209 sz
= unknown
[object_size_type
];
211 if (sz
!= unknown
[object_size_type
])
213 offset_int dsz
= wi::sub (sz
, mem_ref_offset (pt_var
));
216 else if (wi::fits_uhwi_p (dsz
))
219 sz
= unknown
[object_size_type
];
222 if (sz
!= unknown
[object_size_type
] && sz
< offset_limit
)
223 pt_var_size
= size_int (sz
);
227 && tree_fits_uhwi_p (DECL_SIZE_UNIT (pt_var
))
228 && tree_to_uhwi (DECL_SIZE_UNIT (pt_var
)) < offset_limit
)
229 pt_var_size
= DECL_SIZE_UNIT (pt_var
);
231 && TREE_CODE (pt_var
) == STRING_CST
232 && TYPE_SIZE_UNIT (TREE_TYPE (pt_var
))
233 && tree_fits_uhwi_p (TYPE_SIZE_UNIT (TREE_TYPE (pt_var
)))
234 && tree_to_uhwi (TYPE_SIZE_UNIT (TREE_TYPE (pt_var
)))
236 pt_var_size
= TYPE_SIZE_UNIT (TREE_TYPE (pt_var
));
240 if (pt_var
!= TREE_OPERAND (ptr
, 0))
244 if (object_size_type
& 1)
246 var
= TREE_OPERAND (ptr
, 0);
249 && TREE_CODE (var
) != BIT_FIELD_REF
250 && TREE_CODE (var
) != COMPONENT_REF
251 && TREE_CODE (var
) != ARRAY_REF
252 && TREE_CODE (var
) != ARRAY_RANGE_REF
253 && TREE_CODE (var
) != REALPART_EXPR
254 && TREE_CODE (var
) != IMAGPART_EXPR
)
255 var
= TREE_OPERAND (var
, 0);
256 if (var
!= pt_var
&& TREE_CODE (var
) == ARRAY_REF
)
257 var
= TREE_OPERAND (var
, 0);
258 if (! TYPE_SIZE_UNIT (TREE_TYPE (var
))
259 || ! tree_fits_uhwi_p (TYPE_SIZE_UNIT (TREE_TYPE (var
)))
261 && tree_int_cst_lt (pt_var_size
,
262 TYPE_SIZE_UNIT (TREE_TYPE (var
)))))
264 else if (var
!= pt_var
&& TREE_CODE (pt_var
) == MEM_REF
)
267 /* For &X->fld, compute object size only if fld isn't the last
268 field, as struct { int i; char c[1]; } is often used instead
269 of flexible array member. */
270 while (v
&& v
!= pt_var
)
271 switch (TREE_CODE (v
))
274 if (TYPE_SIZE_UNIT (TREE_TYPE (TREE_OPERAND (v
, 0)))
275 && TREE_CODE (TREE_OPERAND (v
, 1)) == INTEGER_CST
)
278 = TYPE_DOMAIN (TREE_TYPE (TREE_OPERAND (v
, 0)));
280 && TYPE_MAX_VALUE (domain
)
281 && TREE_CODE (TYPE_MAX_VALUE (domain
))
283 && tree_int_cst_lt (TREE_OPERAND (v
, 1),
284 TYPE_MAX_VALUE (domain
)))
290 v
= TREE_OPERAND (v
, 0);
297 if (TREE_CODE (TREE_TYPE (v
)) != ARRAY_TYPE
)
302 while (v
!= pt_var
&& TREE_CODE (v
) == COMPONENT_REF
)
303 if (TREE_CODE (TREE_TYPE (TREE_OPERAND (v
, 0)))
305 && TREE_CODE (TREE_TYPE (TREE_OPERAND (v
, 0)))
309 v
= TREE_OPERAND (v
, 0);
310 if (TREE_CODE (v
) == COMPONENT_REF
311 && TREE_CODE (TREE_TYPE (TREE_OPERAND (v
, 0)))
314 tree fld_chain
= DECL_CHAIN (TREE_OPERAND (v
, 1));
315 for (; fld_chain
; fld_chain
= DECL_CHAIN (fld_chain
))
316 if (TREE_CODE (fld_chain
) == FIELD_DECL
)
324 v
= TREE_OPERAND (v
, 0);
326 while (v
!= pt_var
&& TREE_CODE (v
) == COMPONENT_REF
)
327 if (TREE_CODE (TREE_TYPE (TREE_OPERAND (v
, 0)))
329 && TREE_CODE (TREE_TYPE (TREE_OPERAND (v
, 0)))
333 v
= TREE_OPERAND (v
, 0);
351 var_size
= TYPE_SIZE_UNIT (TREE_TYPE (var
));
352 else if (!pt_var_size
)
355 var_size
= pt_var_size
;
356 bytes
= compute_object_offset (TREE_OPERAND (ptr
, 0), var
);
357 if (bytes
!= error_mark_node
)
359 if (TREE_CODE (bytes
) == INTEGER_CST
360 && tree_int_cst_lt (var_size
, bytes
))
361 bytes
= size_zero_node
;
363 bytes
= size_binop (MINUS_EXPR
, var_size
, bytes
);
367 && TREE_CODE (pt_var
) == MEM_REF
368 && bytes
!= error_mark_node
)
370 tree bytes2
= compute_object_offset (TREE_OPERAND (ptr
, 0), pt_var
);
371 if (bytes2
!= error_mark_node
)
373 if (TREE_CODE (bytes2
) == INTEGER_CST
374 && tree_int_cst_lt (pt_var_size
, bytes2
))
375 bytes2
= size_zero_node
;
377 bytes2
= size_binop (MINUS_EXPR
, pt_var_size
, bytes2
);
378 bytes
= size_binop (MIN_EXPR
, bytes
, bytes2
);
382 else if (!pt_var_size
)
387 if (tree_fits_uhwi_p (bytes
))
389 *psize
= tree_to_uhwi (bytes
);
397 /* Compute __builtin_object_size for CALL, which is a GIMPLE_CALL.
398 Handles various allocation calls. OBJECT_SIZE_TYPE is the second
399 argument from __builtin_object_size. If unknown, return
400 unknown[object_size_type]. */
402 static unsigned HOST_WIDE_INT
403 alloc_object_size (const gcall
*call
, int object_size_type
)
405 tree callee
, bytes
= NULL_TREE
;
407 int arg1
= -1, arg2
= -1;
409 gcc_assert (is_gimple_call (call
));
411 callee
= gimple_call_fndecl (call
);
413 return unknown
[object_size_type
];
415 alloc_size
= lookup_attribute ("alloc_size",
416 TYPE_ATTRIBUTES (TREE_TYPE (callee
)));
417 if (alloc_size
&& TREE_VALUE (alloc_size
))
419 tree p
= TREE_VALUE (alloc_size
);
421 arg1
= TREE_INT_CST_LOW (TREE_VALUE (p
))-1;
423 arg2
= TREE_INT_CST_LOW (TREE_VALUE (TREE_CHAIN (p
)))-1;
426 if (DECL_BUILT_IN_CLASS (callee
) == BUILT_IN_NORMAL
)
427 switch (DECL_FUNCTION_CODE (callee
))
429 case BUILT_IN_CALLOC
:
432 case BUILT_IN_MALLOC
:
433 CASE_BUILT_IN_ALLOCA
:
439 if (arg1
< 0 || arg1
>= (int)gimple_call_num_args (call
)
440 || TREE_CODE (gimple_call_arg (call
, arg1
)) != INTEGER_CST
442 && (arg2
>= (int)gimple_call_num_args (call
)
443 || TREE_CODE (gimple_call_arg (call
, arg2
)) != INTEGER_CST
)))
444 return unknown
[object_size_type
];
447 bytes
= size_binop (MULT_EXPR
,
448 fold_convert (sizetype
, gimple_call_arg (call
, arg1
)),
449 fold_convert (sizetype
, gimple_call_arg (call
, arg2
)));
451 bytes
= fold_convert (sizetype
, gimple_call_arg (call
, arg1
));
453 if (bytes
&& tree_fits_uhwi_p (bytes
))
454 return tree_to_uhwi (bytes
);
456 return unknown
[object_size_type
];
460 /* If object size is propagated from one of function's arguments directly
461 to its return value, return that argument for GIMPLE_CALL statement CALL.
462 Otherwise return NULL. */
465 pass_through_call (const gcall
*call
)
467 unsigned rf
= gimple_call_return_flags (call
);
468 if (rf
& ERF_RETURNS_ARG
)
470 unsigned argnum
= rf
& ERF_RETURN_ARG_MASK
;
471 if (argnum
< gimple_call_num_args (call
))
472 return gimple_call_arg (call
, argnum
);
475 /* __builtin_assume_aligned is intentionally not marked RET1. */
476 if (gimple_call_builtin_p (call
, BUILT_IN_ASSUME_ALIGNED
))
477 return gimple_call_arg (call
, 0);
483 /* Compute __builtin_object_size value for PTR and set *PSIZE to
484 the resulting value. OBJECT_SIZE_TYPE is the second argument
485 to __builtin_object_size. Return true on success and false
486 when the object size could not be determined. */
489 compute_builtin_object_size (tree ptr
, int object_size_type
,
490 unsigned HOST_WIDE_INT
*psize
)
492 gcc_assert (object_size_type
>= 0 && object_size_type
<= 3);
494 /* Set to unknown and overwrite just before returning if the size
495 could be determined. */
496 *psize
= unknown
[object_size_type
];
499 init_offset_limit ();
501 if (TREE_CODE (ptr
) == ADDR_EXPR
)
502 return addr_object_size (NULL
, ptr
, object_size_type
, psize
);
504 if (TREE_CODE (ptr
) != SSA_NAME
505 || !POINTER_TYPE_P (TREE_TYPE (ptr
)))
508 if (computed
[object_size_type
] == NULL
)
510 if (optimize
|| object_size_type
& 1)
513 /* When not optimizing, rather than failing, make a small effort
514 to determine the object size without the full benefit of
515 the (costly) computation below. */
516 gimple
*def
= SSA_NAME_DEF_STMT (ptr
);
517 if (gimple_code (def
) == GIMPLE_ASSIGN
)
519 tree_code code
= gimple_assign_rhs_code (def
);
520 if (code
== POINTER_PLUS_EXPR
)
522 tree offset
= gimple_assign_rhs2 (def
);
523 ptr
= gimple_assign_rhs1 (def
);
525 if (tree_fits_shwi_p (offset
)
526 && compute_builtin_object_size (ptr
, object_size_type
, psize
))
528 /* Return zero when the offset is out of bounds. */
529 unsigned HOST_WIDE_INT off
= tree_to_shwi (offset
);
530 *psize
= off
< *psize
? *psize
- off
: 0;
538 if (!bitmap_bit_p (computed
[object_size_type
], SSA_NAME_VERSION (ptr
)))
540 struct object_size_info osi
;
544 if (num_ssa_names
> object_sizes
[object_size_type
].length ())
545 object_sizes
[object_size_type
].safe_grow (num_ssa_names
);
548 fprintf (dump_file
, "Computing %s %sobject size for ",
549 (object_size_type
& 2) ? "minimum" : "maximum",
550 (object_size_type
& 1) ? "sub" : "");
551 print_generic_expr (dump_file
, ptr
, dump_flags
);
552 fprintf (dump_file
, ":\n");
555 osi
.visited
= BITMAP_ALLOC (NULL
);
556 osi
.reexamine
= BITMAP_ALLOC (NULL
);
557 osi
.object_size_type
= object_size_type
;
562 /* First pass: walk UD chains, compute object sizes that
563 can be computed. osi.reexamine bitmap at the end will
564 contain what variables were found in dependency cycles
565 and therefore need to be reexamined. */
568 collect_object_sizes_for (&osi
, ptr
);
570 /* Second pass: keep recomputing object sizes of variables
571 that need reexamination, until no object sizes are
572 increased or all object sizes are computed. */
573 if (! bitmap_empty_p (osi
.reexamine
))
575 bitmap reexamine
= BITMAP_ALLOC (NULL
);
577 /* If looking for minimum instead of maximum object size,
578 detect cases where a pointer is increased in a loop.
579 Although even without this detection pass 2 would eventually
580 terminate, it could take a long time. If a pointer is
581 increasing this way, we need to assume 0 object size.
582 E.g. p = &buf[0]; while (cond) p = p + 4; */
583 if (object_size_type
& 2)
585 osi
.depths
= XCNEWVEC (unsigned int, num_ssa_names
);
586 osi
.stack
= XNEWVEC (unsigned int, num_ssa_names
);
589 /* collect_object_sizes_for is changing
590 osi.reexamine bitmap, so iterate over a copy. */
591 bitmap_copy (reexamine
, osi
.reexamine
);
592 EXECUTE_IF_SET_IN_BITMAP (reexamine
, 0, i
, bi
)
593 if (bitmap_bit_p (osi
.reexamine
, i
))
594 check_for_plus_in_loops (&osi
, ssa_name (i
));
607 /* collect_object_sizes_for is changing
608 osi.reexamine bitmap, so iterate over a copy. */
609 bitmap_copy (reexamine
, osi
.reexamine
);
610 EXECUTE_IF_SET_IN_BITMAP (reexamine
, 0, i
, bi
)
611 if (bitmap_bit_p (osi
.reexamine
, i
))
613 collect_object_sizes_for (&osi
, ssa_name (i
));
614 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
616 fprintf (dump_file
, "Reexamining ");
617 print_generic_expr (dump_file
, ssa_name (i
),
619 fprintf (dump_file
, "\n");
625 BITMAP_FREE (reexamine
);
627 EXECUTE_IF_SET_IN_BITMAP (osi
.reexamine
, 0, i
, bi
)
628 bitmap_set_bit (computed
[object_size_type
], i
);
630 /* Debugging dumps. */
633 EXECUTE_IF_SET_IN_BITMAP (osi
.visited
, 0, i
, bi
)
634 if (object_sizes
[object_size_type
][i
]
635 != unknown
[object_size_type
])
637 print_generic_expr (dump_file
, ssa_name (i
),
640 ": %s %sobject size "
641 HOST_WIDE_INT_PRINT_UNSIGNED
"\n",
642 (object_size_type
& 2) ? "minimum" : "maximum",
643 (object_size_type
& 1) ? "sub" : "",
644 object_sizes
[object_size_type
][i
]);
648 BITMAP_FREE (osi
.reexamine
);
649 BITMAP_FREE (osi
.visited
);
652 *psize
= object_sizes
[object_size_type
][SSA_NAME_VERSION (ptr
)];
653 return *psize
!= unknown
[object_size_type
];
656 /* Compute object_sizes for PTR, defined to VALUE, which is not an SSA_NAME. */
659 expr_object_size (struct object_size_info
*osi
, tree ptr
, tree value
)
661 int object_size_type
= osi
->object_size_type
;
662 unsigned int varno
= SSA_NAME_VERSION (ptr
);
663 unsigned HOST_WIDE_INT bytes
;
665 gcc_assert (object_sizes
[object_size_type
][varno
]
666 != unknown
[object_size_type
]);
667 gcc_assert (osi
->pass
== 0);
669 if (TREE_CODE (value
) == WITH_SIZE_EXPR
)
670 value
= TREE_OPERAND (value
, 0);
672 /* Pointer variables should have been handled by merge_object_sizes. */
673 gcc_assert (TREE_CODE (value
) != SSA_NAME
674 || !POINTER_TYPE_P (TREE_TYPE (value
)));
676 if (TREE_CODE (value
) == ADDR_EXPR
)
677 addr_object_size (osi
, value
, object_size_type
, &bytes
);
679 bytes
= unknown
[object_size_type
];
681 if ((object_size_type
& 2) == 0)
683 if (object_sizes
[object_size_type
][varno
] < bytes
)
684 object_sizes
[object_size_type
][varno
] = bytes
;
688 if (object_sizes
[object_size_type
][varno
] > bytes
)
689 object_sizes
[object_size_type
][varno
] = bytes
;
694 /* Compute object_sizes for PTR, defined to the result of a call. */
697 call_object_size (struct object_size_info
*osi
, tree ptr
, gcall
*call
)
699 int object_size_type
= osi
->object_size_type
;
700 unsigned int varno
= SSA_NAME_VERSION (ptr
);
701 unsigned HOST_WIDE_INT bytes
;
703 gcc_assert (is_gimple_call (call
));
705 gcc_assert (object_sizes
[object_size_type
][varno
]
706 != unknown
[object_size_type
]);
707 gcc_assert (osi
->pass
== 0);
709 bytes
= alloc_object_size (call
, object_size_type
);
711 if ((object_size_type
& 2) == 0)
713 if (object_sizes
[object_size_type
][varno
] < bytes
)
714 object_sizes
[object_size_type
][varno
] = bytes
;
718 if (object_sizes
[object_size_type
][varno
] > bytes
)
719 object_sizes
[object_size_type
][varno
] = bytes
;
724 /* Compute object_sizes for PTR, defined to an unknown value. */
727 unknown_object_size (struct object_size_info
*osi
, tree ptr
)
729 int object_size_type
= osi
->object_size_type
;
730 unsigned int varno
= SSA_NAME_VERSION (ptr
);
731 unsigned HOST_WIDE_INT bytes
;
733 gcc_assert (object_sizes
[object_size_type
][varno
]
734 != unknown
[object_size_type
]);
735 gcc_assert (osi
->pass
== 0);
737 bytes
= unknown
[object_size_type
];
739 if ((object_size_type
& 2) == 0)
741 if (object_sizes
[object_size_type
][varno
] < bytes
)
742 object_sizes
[object_size_type
][varno
] = bytes
;
746 if (object_sizes
[object_size_type
][varno
] > bytes
)
747 object_sizes
[object_size_type
][varno
] = bytes
;
752 /* Merge object sizes of ORIG + OFFSET into DEST. Return true if
753 the object size might need reexamination later. */
756 merge_object_sizes (struct object_size_info
*osi
, tree dest
, tree orig
,
757 unsigned HOST_WIDE_INT offset
)
759 int object_size_type
= osi
->object_size_type
;
760 unsigned int varno
= SSA_NAME_VERSION (dest
);
761 unsigned HOST_WIDE_INT orig_bytes
;
763 if (object_sizes
[object_size_type
][varno
] == unknown
[object_size_type
])
765 if (offset
>= offset_limit
)
767 object_sizes
[object_size_type
][varno
] = unknown
[object_size_type
];
772 collect_object_sizes_for (osi
, orig
);
774 orig_bytes
= object_sizes
[object_size_type
][SSA_NAME_VERSION (orig
)];
775 if (orig_bytes
!= unknown
[object_size_type
])
776 orig_bytes
= (offset
> orig_bytes
)
777 ? HOST_WIDE_INT_0U
: orig_bytes
- offset
;
779 if ((object_size_type
& 2) == 0)
781 if (object_sizes
[object_size_type
][varno
] < orig_bytes
)
783 object_sizes
[object_size_type
][varno
] = orig_bytes
;
789 if (object_sizes
[object_size_type
][varno
] > orig_bytes
)
791 object_sizes
[object_size_type
][varno
] = orig_bytes
;
795 return bitmap_bit_p (osi
->reexamine
, SSA_NAME_VERSION (orig
));
799 /* Compute object_sizes for VAR, defined to the result of an assignment
800 with operator POINTER_PLUS_EXPR. Return true if the object size might
801 need reexamination later. */
804 plus_stmt_object_size (struct object_size_info
*osi
, tree var
, gimple
*stmt
)
806 int object_size_type
= osi
->object_size_type
;
807 unsigned int varno
= SSA_NAME_VERSION (var
);
808 unsigned HOST_WIDE_INT bytes
;
811 if (gimple_assign_rhs_code (stmt
) == POINTER_PLUS_EXPR
)
813 op0
= gimple_assign_rhs1 (stmt
);
814 op1
= gimple_assign_rhs2 (stmt
);
816 else if (gimple_assign_rhs_code (stmt
) == ADDR_EXPR
)
818 tree rhs
= TREE_OPERAND (gimple_assign_rhs1 (stmt
), 0);
819 gcc_assert (TREE_CODE (rhs
) == MEM_REF
);
820 op0
= TREE_OPERAND (rhs
, 0);
821 op1
= TREE_OPERAND (rhs
, 1);
826 if (object_sizes
[object_size_type
][varno
] == unknown
[object_size_type
])
829 /* Handle PTR + OFFSET here. */
830 if (TREE_CODE (op1
) == INTEGER_CST
831 && (TREE_CODE (op0
) == SSA_NAME
832 || TREE_CODE (op0
) == ADDR_EXPR
))
834 if (! tree_fits_uhwi_p (op1
))
835 bytes
= unknown
[object_size_type
];
836 else if (TREE_CODE (op0
) == SSA_NAME
)
837 return merge_object_sizes (osi
, var
, op0
, tree_to_uhwi (op1
));
840 unsigned HOST_WIDE_INT off
= tree_to_uhwi (op1
);
842 /* op0 will be ADDR_EXPR here. */
843 addr_object_size (osi
, op0
, object_size_type
, &bytes
);
844 if (bytes
== unknown
[object_size_type
])
846 else if (off
> offset_limit
)
847 bytes
= unknown
[object_size_type
];
848 else if (off
> bytes
)
855 bytes
= unknown
[object_size_type
];
857 if ((object_size_type
& 2) == 0)
859 if (object_sizes
[object_size_type
][varno
] < bytes
)
860 object_sizes
[object_size_type
][varno
] = bytes
;
864 if (object_sizes
[object_size_type
][varno
] > bytes
)
865 object_sizes
[object_size_type
][varno
] = bytes
;
871 /* Compute object_sizes for VAR, defined at STMT, which is
872 a COND_EXPR. Return true if the object size might need reexamination
876 cond_expr_object_size (struct object_size_info
*osi
, tree var
, gimple
*stmt
)
879 int object_size_type
= osi
->object_size_type
;
880 unsigned int varno
= SSA_NAME_VERSION (var
);
881 bool reexamine
= false;
883 gcc_assert (gimple_assign_rhs_code (stmt
) == COND_EXPR
);
885 if (object_sizes
[object_size_type
][varno
] == unknown
[object_size_type
])
888 then_
= gimple_assign_rhs2 (stmt
);
889 else_
= gimple_assign_rhs3 (stmt
);
891 if (TREE_CODE (then_
) == SSA_NAME
)
892 reexamine
|= merge_object_sizes (osi
, var
, then_
, 0);
894 expr_object_size (osi
, var
, then_
);
896 if (TREE_CODE (else_
) == SSA_NAME
)
897 reexamine
|= merge_object_sizes (osi
, var
, else_
, 0);
899 expr_object_size (osi
, var
, else_
);
904 /* Compute object sizes for VAR.
905 For ADDR_EXPR an object size is the number of remaining bytes
906 to the end of the object (where what is considered an object depends on
907 OSI->object_size_type).
908 For allocation GIMPLE_CALL like malloc or calloc object size is the size
910 For POINTER_PLUS_EXPR where second operand is a constant integer,
911 object size is object size of the first operand minus the constant.
912 If the constant is bigger than the number of remaining bytes until the
913 end of the object, object size is 0, but if it is instead a pointer
914 subtraction, object size is unknown[object_size_type].
915 To differentiate addition from subtraction, ADDR_EXPR returns
916 unknown[object_size_type] for all objects bigger than half of the address
917 space, and constants less than half of the address space are considered
918 addition, while bigger constants subtraction.
919 For a memcpy like GIMPLE_CALL that always returns one of its arguments, the
920 object size is object size of that argument.
921 Otherwise, object size is the maximum of object sizes of variables
922 that it might be set to. */
925 collect_object_sizes_for (struct object_size_info
*osi
, tree var
)
927 int object_size_type
= osi
->object_size_type
;
928 unsigned int varno
= SSA_NAME_VERSION (var
);
932 if (bitmap_bit_p (computed
[object_size_type
], varno
))
937 if (bitmap_set_bit (osi
->visited
, varno
))
939 object_sizes
[object_size_type
][varno
]
940 = (object_size_type
& 2) ? -1 : 0;
944 /* Found a dependency loop. Mark the variable for later
946 bitmap_set_bit (osi
->reexamine
, varno
);
947 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
949 fprintf (dump_file
, "Found a dependency loop at ");
950 print_generic_expr (dump_file
, var
, dump_flags
);
951 fprintf (dump_file
, "\n");
957 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
959 fprintf (dump_file
, "Visiting use-def links for ");
960 print_generic_expr (dump_file
, var
, dump_flags
);
961 fprintf (dump_file
, "\n");
964 stmt
= SSA_NAME_DEF_STMT (var
);
967 switch (gimple_code (stmt
))
971 tree rhs
= gimple_assign_rhs1 (stmt
);
972 if (gimple_assign_rhs_code (stmt
) == POINTER_PLUS_EXPR
973 || (gimple_assign_rhs_code (stmt
) == ADDR_EXPR
974 && TREE_CODE (TREE_OPERAND (rhs
, 0)) == MEM_REF
))
975 reexamine
= plus_stmt_object_size (osi
, var
, stmt
);
976 else if (gimple_assign_rhs_code (stmt
) == COND_EXPR
)
977 reexamine
= cond_expr_object_size (osi
, var
, stmt
);
978 else if (gimple_assign_single_p (stmt
)
979 || gimple_assign_unary_nop_p (stmt
))
981 if (TREE_CODE (rhs
) == SSA_NAME
982 && POINTER_TYPE_P (TREE_TYPE (rhs
)))
983 reexamine
= merge_object_sizes (osi
, var
, rhs
, 0);
985 expr_object_size (osi
, var
, rhs
);
988 unknown_object_size (osi
, var
);
994 gcall
*call_stmt
= as_a
<gcall
*> (stmt
);
995 tree arg
= pass_through_call (call_stmt
);
998 if (TREE_CODE (arg
) == SSA_NAME
999 && POINTER_TYPE_P (TREE_TYPE (arg
)))
1000 reexamine
= merge_object_sizes (osi
, var
, arg
, 0);
1002 expr_object_size (osi
, var
, arg
);
1005 call_object_size (osi
, var
, call_stmt
);
1010 /* Pointers defined by __asm__ statements can point anywhere. */
1011 object_sizes
[object_size_type
][varno
] = unknown
[object_size_type
];
1015 if (SSA_NAME_VAR (var
)
1016 && TREE_CODE (SSA_NAME_VAR (var
)) == PARM_DECL
)
1017 expr_object_size (osi
, var
, SSA_NAME_VAR (var
));
1019 /* Uninitialized SSA names point nowhere. */
1020 object_sizes
[object_size_type
][varno
] = unknown
[object_size_type
];
1027 for (i
= 0; i
< gimple_phi_num_args (stmt
); i
++)
1029 tree rhs
= gimple_phi_arg (stmt
, i
)->def
;
1031 if (object_sizes
[object_size_type
][varno
]
1032 == unknown
[object_size_type
])
1035 if (TREE_CODE (rhs
) == SSA_NAME
)
1036 reexamine
|= merge_object_sizes (osi
, var
, rhs
, 0);
1037 else if (osi
->pass
== 0)
1038 expr_object_size (osi
, var
, rhs
);
1048 || object_sizes
[object_size_type
][varno
] == unknown
[object_size_type
])
1050 bitmap_set_bit (computed
[object_size_type
], varno
);
1051 bitmap_clear_bit (osi
->reexamine
, varno
);
1055 bitmap_set_bit (osi
->reexamine
, varno
);
1056 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
1058 fprintf (dump_file
, "Need to reexamine ");
1059 print_generic_expr (dump_file
, var
, dump_flags
);
1060 fprintf (dump_file
, "\n");
1066 /* Helper function for check_for_plus_in_loops. Called recursively
1070 check_for_plus_in_loops_1 (struct object_size_info
*osi
, tree var
,
1073 gimple
*stmt
= SSA_NAME_DEF_STMT (var
);
1074 unsigned int varno
= SSA_NAME_VERSION (var
);
1076 if (osi
->depths
[varno
])
1078 if (osi
->depths
[varno
] != depth
)
1082 /* Found a loop involving pointer addition. */
1083 for (sp
= osi
->tos
; sp
> osi
->stack
; )
1086 bitmap_clear_bit (osi
->reexamine
, *sp
);
1087 bitmap_set_bit (computed
[osi
->object_size_type
], *sp
);
1088 object_sizes
[osi
->object_size_type
][*sp
] = 0;
1095 else if (! bitmap_bit_p (osi
->reexamine
, varno
))
1098 osi
->depths
[varno
] = depth
;
1099 *osi
->tos
++ = varno
;
1101 switch (gimple_code (stmt
))
1106 if ((gimple_assign_single_p (stmt
)
1107 || gimple_assign_unary_nop_p (stmt
))
1108 && TREE_CODE (gimple_assign_rhs1 (stmt
)) == SSA_NAME
)
1110 tree rhs
= gimple_assign_rhs1 (stmt
);
1112 check_for_plus_in_loops_1 (osi
, rhs
, depth
);
1114 else if (gimple_assign_rhs_code (stmt
) == POINTER_PLUS_EXPR
)
1116 tree basevar
= gimple_assign_rhs1 (stmt
);
1117 tree cst
= gimple_assign_rhs2 (stmt
);
1119 gcc_assert (TREE_CODE (cst
) == INTEGER_CST
);
1121 check_for_plus_in_loops_1 (osi
, basevar
,
1122 depth
+ !integer_zerop (cst
));
1131 gcall
*call_stmt
= as_a
<gcall
*> (stmt
);
1132 tree arg
= pass_through_call (call_stmt
);
1135 if (TREE_CODE (arg
) == SSA_NAME
)
1136 check_for_plus_in_loops_1 (osi
, arg
, depth
);
1147 for (i
= 0; i
< gimple_phi_num_args (stmt
); i
++)
1149 tree rhs
= gimple_phi_arg (stmt
, i
)->def
;
1151 if (TREE_CODE (rhs
) == SSA_NAME
)
1152 check_for_plus_in_loops_1 (osi
, rhs
, depth
);
1161 osi
->depths
[varno
] = 0;
1166 /* Check if some pointer we are computing object size of is being increased
1167 within a loop. If yes, assume all the SSA variables participating in
1168 that loop have minimum object sizes 0. */
1171 check_for_plus_in_loops (struct object_size_info
*osi
, tree var
)
1173 gimple
*stmt
= SSA_NAME_DEF_STMT (var
);
1175 /* NOTE: In the pre-tuples code, we handled a CALL_EXPR here,
1176 and looked for a POINTER_PLUS_EXPR in the pass-through
1177 argument, if any. In GIMPLE, however, such an expression
1178 is not a valid call operand. */
1180 if (is_gimple_assign (stmt
)
1181 && gimple_assign_rhs_code (stmt
) == POINTER_PLUS_EXPR
)
1183 tree basevar
= gimple_assign_rhs1 (stmt
);
1184 tree cst
= gimple_assign_rhs2 (stmt
);
1186 gcc_assert (TREE_CODE (cst
) == INTEGER_CST
);
1188 if (integer_zerop (cst
))
1191 osi
->depths
[SSA_NAME_VERSION (basevar
)] = 1;
1192 *osi
->tos
++ = SSA_NAME_VERSION (basevar
);
1193 check_for_plus_in_loops_1 (osi
, var
, 2);
1194 osi
->depths
[SSA_NAME_VERSION (basevar
)] = 0;
1200 /* Initialize data structures for the object size computation. */
1203 init_object_sizes (void)
1205 int object_size_type
;
1210 for (object_size_type
= 0; object_size_type
<= 3; object_size_type
++)
1212 object_sizes
[object_size_type
].safe_grow (num_ssa_names
);
1213 computed
[object_size_type
] = BITMAP_ALLOC (NULL
);
1216 init_offset_limit ();
1220 /* Destroy data structures after the object size computation. */
1223 fini_object_sizes (void)
1225 int object_size_type
;
1227 for (object_size_type
= 0; object_size_type
<= 3; object_size_type
++)
1229 object_sizes
[object_size_type
].release ();
1230 BITMAP_FREE (computed
[object_size_type
]);
1235 /* Simple pass to optimize all __builtin_object_size () builtins. */
1239 const pass_data pass_data_object_sizes
=
1241 GIMPLE_PASS
, /* type */
1243 OPTGROUP_NONE
, /* optinfo_flags */
1244 TV_NONE
, /* tv_id */
1245 ( PROP_cfg
| PROP_ssa
), /* properties_required */
1246 0, /* properties_provided */
1247 0, /* properties_destroyed */
1248 0, /* todo_flags_start */
1249 0, /* todo_flags_finish */
1252 class pass_object_sizes
: public gimple_opt_pass
1255 pass_object_sizes (gcc::context
*ctxt
)
1256 : gimple_opt_pass (pass_data_object_sizes
, ctxt
), insert_min_max_p (false)
1259 /* opt_pass methods: */
1260 opt_pass
* clone () { return new pass_object_sizes (m_ctxt
); }
1261 void set_pass_param (unsigned int n
, bool param
)
1263 gcc_assert (n
== 0);
1264 insert_min_max_p
= param
;
1266 virtual unsigned int execute (function
*);
1269 /* Determines whether the pass instance creates MIN/MAX_EXPRs. */
1270 bool insert_min_max_p
;
1271 }; // class pass_object_sizes
1273 /* Dummy valueize function. */
1276 do_valueize (tree t
)
1282 pass_object_sizes::execute (function
*fun
)
1285 FOR_EACH_BB_FN (bb
, fun
)
1287 gimple_stmt_iterator i
;
1288 for (i
= gsi_start_bb (bb
); !gsi_end_p (i
); gsi_next (&i
))
1291 gimple
*call
= gsi_stmt (i
);
1292 if (!gimple_call_builtin_p (call
, BUILT_IN_OBJECT_SIZE
))
1295 init_object_sizes ();
1297 /* If insert_min_max_p, only attempt to fold
1298 __builtin_object_size (x, 1) and __builtin_object_size (x, 3),
1299 and rather than folding the builtin to the constant if any,
1300 create a MIN_EXPR or MAX_EXPR of the __builtin_object_size
1301 call result and the computed constant. */
1302 if (insert_min_max_p
)
1304 tree ost
= gimple_call_arg (call
, 1);
1305 if (tree_fits_uhwi_p (ost
))
1307 unsigned HOST_WIDE_INT object_size_type
= tree_to_uhwi (ost
);
1308 tree ptr
= gimple_call_arg (call
, 0);
1309 tree lhs
= gimple_call_lhs (call
);
1310 if ((object_size_type
== 1 || object_size_type
== 3)
1311 && (TREE_CODE (ptr
) == ADDR_EXPR
1312 || TREE_CODE (ptr
) == SSA_NAME
)
1315 tree type
= TREE_TYPE (lhs
);
1316 unsigned HOST_WIDE_INT bytes
;
1317 if (compute_builtin_object_size (ptr
, object_size_type
,
1319 && wi::fits_to_tree_p (bytes
, type
))
1321 tree tem
= make_ssa_name (type
);
1322 gimple_call_set_lhs (call
, tem
);
1324 = object_size_type
== 1 ? MIN_EXPR
: MAX_EXPR
;
1325 tree cst
= build_int_cstu (type
, bytes
);
1327 = gimple_build_assign (lhs
, code
, tem
, cst
);
1328 gsi_insert_after (&i
, g
, GSI_NEW_STMT
);
1336 tree lhs
= gimple_call_lhs (call
);
1340 result
= gimple_fold_stmt_to_constant (call
, do_valueize
);
1343 tree ost
= gimple_call_arg (call
, 1);
1345 if (tree_fits_uhwi_p (ost
))
1347 unsigned HOST_WIDE_INT object_size_type
= tree_to_uhwi (ost
);
1349 if (object_size_type
< 2)
1350 result
= fold_convert (size_type_node
,
1351 integer_minus_one_node
);
1352 else if (object_size_type
< 4)
1353 result
= build_zero_cst (size_type_node
);
1360 gcc_assert (TREE_CODE (result
) == INTEGER_CST
);
1362 if (dump_file
&& (dump_flags
& TDF_DETAILS
))
1364 fprintf (dump_file
, "Simplified\n ");
1365 print_gimple_stmt (dump_file
, call
, 0, dump_flags
);
1366 fprintf (dump_file
, " to ");
1367 print_generic_expr (dump_file
, result
);
1368 fprintf (dump_file
, "\n");
1371 /* Propagate into all uses and fold those stmts. */
1372 replace_uses_by (lhs
, result
);
1376 fini_object_sizes ();
1383 make_pass_object_sizes (gcc::context
*ctxt
)
1385 return new pass_object_sizes (ctxt
);