1 /* Data flow functions for trees.
2 Copyright (C) 2001-2023 Free Software Foundation, Inc.
3 Contributed by Diego Novillo <dnovillo@redhat.com>
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License as published by
9 the Free Software Foundation; either version 3, or (at your option)
12 GCC is distributed in the hope that it will be useful,
13 but WITHOUT ANY WARRANTY; without even the implied warranty of
14 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 GNU General Public License for more details.
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
23 #include "coretypes.h"
28 #include "tree-pass.h"
30 #include "tree-pretty-print.h"
31 #include "fold-const.h"
32 #include "stor-layout.h"
33 #include "langhooks.h"
34 #include "gimple-iterator.h"
35 #include "gimple-walk.h"
37 #include "gimple-range.h"
39 /* Build and maintain data flow information for trees. */
41 /* Counters used to display DFA and SSA statistics. */
48 size_t max_num_phi_args
;
54 /* Local functions. */
55 static void collect_dfa_stats (struct dfa_stats_d
*);
58 /*---------------------------------------------------------------------------
59 Dataflow analysis (DFA) routines
60 ---------------------------------------------------------------------------*/
62 /* Renumber the gimple stmt uids in one block. The caller is responsible
63 of calling set_gimple_stmt_max_uid (fun, 0) at some point. */
66 renumber_gimple_stmt_uids_in_block (struct function
*fun
, basic_block bb
)
68 gimple_stmt_iterator bsi
;
69 for (bsi
= gsi_start_phis (bb
); !gsi_end_p (bsi
); gsi_next (&bsi
))
71 gimple
*stmt
= gsi_stmt (bsi
);
72 gimple_set_uid (stmt
, inc_gimple_stmt_max_uid (fun
));
74 for (bsi
= gsi_start_bb (bb
); !gsi_end_p (bsi
); gsi_next (&bsi
))
76 gimple
*stmt
= gsi_stmt (bsi
);
77 gimple_set_uid (stmt
, inc_gimple_stmt_max_uid (fun
));
81 /* Renumber all of the gimple stmt uids. */
84 renumber_gimple_stmt_uids (struct function
*fun
)
88 set_gimple_stmt_max_uid (fun
, 0);
89 FOR_ALL_BB_FN (bb
, fun
)
90 renumber_gimple_stmt_uids_in_block (fun
, bb
);
93 /* Like renumber_gimple_stmt_uids, but only do work on the basic blocks
94 in BLOCKS, of which there are N_BLOCKS. Also renumbers PHIs. */
97 renumber_gimple_stmt_uids_in_blocks (basic_block
*blocks
, int n_blocks
)
101 set_gimple_stmt_max_uid (cfun
, 0);
102 for (i
= 0; i
< n_blocks
; i
++)
103 renumber_gimple_stmt_uids_in_block (cfun
, blocks
[i
]);
108 /*---------------------------------------------------------------------------
110 ---------------------------------------------------------------------------*/
112 /* Dump variable VAR and its may-aliases to FILE. */
115 dump_variable (FILE *file
, tree var
)
117 if (TREE_CODE (var
) == SSA_NAME
)
119 if (POINTER_TYPE_P (TREE_TYPE (var
)))
120 dump_points_to_info_for (file
, var
);
121 var
= SSA_NAME_VAR (var
);
124 if (var
== NULL_TREE
)
126 fprintf (file
, "<nil>");
130 print_generic_expr (file
, var
, dump_flags
);
132 fprintf (file
, ", UID D.%u", (unsigned) DECL_UID (var
));
133 if (DECL_PT_UID (var
) != DECL_UID (var
))
134 fprintf (file
, ", PT-UID D.%u", (unsigned) DECL_PT_UID (var
));
136 fprintf (file
, ", ");
137 print_generic_expr (file
, TREE_TYPE (var
), dump_flags
);
139 if (TREE_ADDRESSABLE (var
))
140 fprintf (file
, ", is addressable");
142 if (is_global_var (var
))
143 fprintf (file
, ", is global");
145 if (TREE_THIS_VOLATILE (var
))
146 fprintf (file
, ", is volatile");
148 if (cfun
&& ssa_default_def (cfun
, var
))
150 fprintf (file
, ", default def: ");
151 print_generic_expr (file
, ssa_default_def (cfun
, var
), dump_flags
);
154 if (DECL_INITIAL (var
))
156 fprintf (file
, ", initial: ");
157 print_generic_expr (file
, DECL_INITIAL (var
), dump_flags
);
160 fprintf (file
, "\n");
164 /* Dump variable VAR and its may-aliases to stderr. */
167 debug_variable (tree var
)
169 dump_variable (stderr
, var
);
173 /* Dump various DFA statistics to FILE. */
176 dump_dfa_stats (FILE *file
)
178 struct dfa_stats_d dfa_stats
;
180 unsigned long size
, total
= 0;
181 const char * const fmt_str
= "%-30s%-13s%12s\n";
182 const char * const fmt_str_1
= "%-30s%13lu" PRsa (11) "\n";
183 const char * const fmt_str_3
= "%-43s" PRsa (11) "\n";
185 = lang_hooks
.decl_printable_name (current_function_decl
, 2);
187 collect_dfa_stats (&dfa_stats
);
189 fprintf (file
, "\nDFA Statistics for %s\n\n", funcname
);
191 fprintf (file
, "---------------------------------------------------------\n");
192 fprintf (file
, fmt_str
, "", " Number of ", "Memory");
193 fprintf (file
, fmt_str
, "", " instances ", "used ");
194 fprintf (file
, "---------------------------------------------------------\n");
196 size
= dfa_stats
.num_uses
* sizeof (tree
*);
198 fprintf (file
, fmt_str_1
, "USE operands", dfa_stats
.num_uses
,
201 size
= dfa_stats
.num_defs
* sizeof (tree
*);
203 fprintf (file
, fmt_str_1
, "DEF operands", dfa_stats
.num_defs
,
206 size
= dfa_stats
.num_vuses
* sizeof (tree
*);
208 fprintf (file
, fmt_str_1
, "VUSE operands", dfa_stats
.num_vuses
,
211 size
= dfa_stats
.num_vdefs
* sizeof (tree
*);
213 fprintf (file
, fmt_str_1
, "VDEF operands", dfa_stats
.num_vdefs
,
216 size
= dfa_stats
.num_phis
* sizeof (struct gphi
);
218 fprintf (file
, fmt_str_1
, "PHI nodes", dfa_stats
.num_phis
,
221 size
= dfa_stats
.num_phi_args
* sizeof (struct phi_arg_d
);
223 fprintf (file
, fmt_str_1
, "PHI arguments", dfa_stats
.num_phi_args
,
226 fprintf (file
, "---------------------------------------------------------\n");
227 fprintf (file
, fmt_str_3
, "Total memory used by DFA/SSA data",
228 SIZE_AMOUNT (total
));
229 fprintf (file
, "---------------------------------------------------------\n");
230 fprintf (file
, "\n");
232 if (dfa_stats
.num_phis
)
233 fprintf (file
, "Average number of arguments per PHI node: %.1f (max: %ld)\n",
234 (float) dfa_stats
.num_phi_args
/ (float) dfa_stats
.num_phis
,
235 (long) dfa_stats
.max_num_phi_args
);
237 fprintf (file
, "\n");
241 /* Dump DFA statistics on stderr. */
244 debug_dfa_stats (void)
246 dump_dfa_stats (stderr
);
250 /* Collect DFA statistics and store them in the structure pointed to by
254 collect_dfa_stats (struct dfa_stats_d
*dfa_stats_p ATTRIBUTE_UNUSED
)
258 gcc_assert (dfa_stats_p
);
260 memset ((void *)dfa_stats_p
, 0, sizeof (struct dfa_stats_d
));
262 /* Walk all the statements in the function counting references. */
263 FOR_EACH_BB_FN (bb
, cfun
)
265 for (gphi_iterator si
= gsi_start_phis (bb
); !gsi_end_p (si
);
268 gphi
*phi
= si
.phi ();
269 dfa_stats_p
->num_phis
++;
270 dfa_stats_p
->num_phi_args
+= gimple_phi_num_args (phi
);
271 if (gimple_phi_num_args (phi
) > dfa_stats_p
->max_num_phi_args
)
272 dfa_stats_p
->max_num_phi_args
= gimple_phi_num_args (phi
);
275 for (gimple_stmt_iterator si
= gsi_start_bb (bb
); !gsi_end_p (si
);
278 gimple
*stmt
= gsi_stmt (si
);
279 dfa_stats_p
->num_defs
+= NUM_SSA_OPERANDS (stmt
, SSA_OP_DEF
);
280 dfa_stats_p
->num_uses
+= NUM_SSA_OPERANDS (stmt
, SSA_OP_USE
);
281 dfa_stats_p
->num_vdefs
+= gimple_vdef (stmt
) ? 1 : 0;
282 dfa_stats_p
->num_vuses
+= gimple_vuse (stmt
) ? 1 : 0;
288 /*---------------------------------------------------------------------------
289 Miscellaneous helpers
290 ---------------------------------------------------------------------------*/
292 /* Lookup VAR UID in the default_defs hashtable and return the associated
296 ssa_default_def (struct function
*fn
, tree var
)
298 struct tree_decl_minimal ind
;
299 struct tree_ssa_name in
;
300 gcc_assert (VAR_P (var
)
301 || TREE_CODE (var
) == PARM_DECL
302 || TREE_CODE (var
) == RESULT_DECL
);
304 /* Always NULL_TREE for rtl function dumps. */
309 ind
.uid
= DECL_UID (var
);
310 return DEFAULT_DEFS (fn
)->find_with_hash ((tree
)&in
, DECL_UID (var
));
313 /* Insert the pair VAR's UID, DEF into the default_defs hashtable
317 set_ssa_default_def (struct function
*fn
, tree var
, tree def
)
319 struct tree_decl_minimal ind
;
320 struct tree_ssa_name in
;
322 gcc_assert (VAR_P (var
)
323 || TREE_CODE (var
) == PARM_DECL
324 || TREE_CODE (var
) == RESULT_DECL
);
326 ind
.uid
= DECL_UID (var
);
329 tree
*loc
= DEFAULT_DEFS (fn
)->find_slot_with_hash ((tree
)&in
,
334 SSA_NAME_IS_DEFAULT_DEF (*(tree
*)loc
) = false;
335 DEFAULT_DEFS (fn
)->clear_slot (loc
);
339 gcc_assert (TREE_CODE (def
) == SSA_NAME
&& SSA_NAME_VAR (def
) == var
);
340 tree
*loc
= DEFAULT_DEFS (fn
)->find_slot_with_hash ((tree
)&in
,
341 DECL_UID (var
), INSERT
);
343 /* Default definition might be changed by tail call optimization. */
345 SSA_NAME_IS_DEFAULT_DEF (*loc
) = false;
347 /* Mark DEF as the default definition for VAR. */
349 SSA_NAME_IS_DEFAULT_DEF (def
) = true;
352 /* Retrieve or create a default definition for VAR. */
355 get_or_create_ssa_default_def (struct function
*fn
, tree var
)
357 tree ddef
= ssa_default_def (fn
, var
);
358 if (ddef
== NULL_TREE
)
360 ddef
= make_ssa_name_fn (fn
, var
, gimple_build_nop ());
361 set_ssa_default_def (fn
, var
, ddef
);
367 /* If EXP is a handled component reference for a structure, return the
368 base variable. The access range is delimited by bit positions *POFFSET and
369 *POFFSET + *PMAX_SIZE. The access size is *PSIZE bits. If either
370 *PSIZE or *PMAX_SIZE is -1, they could not be determined. If *PSIZE
371 and *PMAX_SIZE are equal, the access is non-variable. If *PREVERSE is
372 true, the storage order of the reference is reversed. */
375 get_ref_base_and_extent (tree exp
, poly_int64_pod
*poffset
,
376 poly_int64_pod
*psize
,
377 poly_int64_pod
*pmax_size
,
380 poly_offset_int bitsize
= -1;
381 poly_offset_int maxsize
;
382 tree size_tree
= NULL_TREE
;
383 poly_offset_int bit_offset
= 0;
384 bool seen_variable_array_ref
= false;
386 /* First get the final access size and the storage order from just the
387 outermost expression. */
388 if (TREE_CODE (exp
) == COMPONENT_REF
)
389 size_tree
= DECL_SIZE (TREE_OPERAND (exp
, 1));
390 else if (TREE_CODE (exp
) == BIT_FIELD_REF
)
391 size_tree
= TREE_OPERAND (exp
, 1);
392 else if (TREE_CODE (exp
) == WITH_SIZE_EXPR
)
394 size_tree
= TREE_OPERAND (exp
, 1);
395 exp
= TREE_OPERAND (exp
, 0);
397 else if (!VOID_TYPE_P (TREE_TYPE (exp
)))
399 machine_mode mode
= TYPE_MODE (TREE_TYPE (exp
));
401 size_tree
= TYPE_SIZE (TREE_TYPE (exp
));
403 bitsize
= GET_MODE_BITSIZE (mode
);
405 if (size_tree
!= NULL_TREE
406 && poly_int_tree_p (size_tree
))
407 bitsize
= wi::to_poly_offset (size_tree
);
409 *preverse
= reverse_storage_order_for_component_p (exp
);
411 /* Initially, maxsize is the same as the accessed element size.
412 In the following it will only grow (or become -1). */
415 /* Compute cumulative bit-offset for nested component-refs and array-refs,
416 and find the ultimate containing object. */
419 switch (TREE_CODE (exp
))
422 bit_offset
+= wi::to_poly_offset (TREE_OPERAND (exp
, 2));
427 tree field
= TREE_OPERAND (exp
, 1);
428 tree this_offset
= component_ref_field_offset (exp
);
430 if (this_offset
&& poly_int_tree_p (this_offset
))
432 poly_offset_int woffset
= (wi::to_poly_offset (this_offset
)
433 << LOG2_BITS_PER_UNIT
);
434 woffset
+= wi::to_offset (DECL_FIELD_BIT_OFFSET (field
));
435 bit_offset
+= woffset
;
437 /* If we had seen a variable array ref already and we just
438 referenced the last field of a struct or a union member
439 then we have to adjust maxsize by the padding at the end
441 if (seen_variable_array_ref
)
443 tree stype
= TREE_TYPE (TREE_OPERAND (exp
, 0));
444 tree next
= DECL_CHAIN (field
);
445 while (next
&& TREE_CODE (next
) != FIELD_DECL
)
446 next
= DECL_CHAIN (next
);
448 || TREE_CODE (stype
) != RECORD_TYPE
)
450 tree fsize
= DECL_SIZE (field
);
451 tree ssize
= TYPE_SIZE (stype
);
453 || !poly_int_tree_p (fsize
)
455 || !poly_int_tree_p (ssize
))
457 else if (known_size_p (maxsize
))
460 = (wi::to_poly_offset (ssize
)
461 - wi::to_poly_offset (fsize
));
466 /* An component ref with an adjacent field up in the
467 structure hierarchy constrains the size of any variable
468 array ref lower in the access hierarchy. */
470 seen_variable_array_ref
= false;
475 tree csize
= TYPE_SIZE (TREE_TYPE (TREE_OPERAND (exp
, 0)));
476 /* We need to adjust maxsize to the whole structure bitsize.
477 But we can subtract any constant offset seen so far,
478 because that would get us out of the structure otherwise. */
479 if (known_size_p (maxsize
)
481 && poly_int_tree_p (csize
))
482 maxsize
= wi::to_poly_offset (csize
) - bit_offset
;
490 case ARRAY_RANGE_REF
:
492 tree index
= TREE_OPERAND (exp
, 1);
493 tree low_bound
, unit_size
;
495 /* If the resulting bit-offset is constant, track it. */
496 if (poly_int_tree_p (index
)
497 && (low_bound
= array_ref_low_bound (exp
),
498 poly_int_tree_p (low_bound
))
499 && (unit_size
= array_ref_element_size (exp
),
500 TREE_CODE (unit_size
) == INTEGER_CST
))
502 poly_offset_int woffset
503 = wi::sext (wi::to_poly_offset (index
)
504 - wi::to_poly_offset (low_bound
),
505 TYPE_PRECISION (sizetype
));
506 woffset
*= wi::to_offset (unit_size
);
507 woffset
<<= LOG2_BITS_PER_UNIT
;
508 bit_offset
+= woffset
;
510 /* An array ref with a constant index up in the structure
511 hierarchy will constrain the size of any variable array ref
512 lower in the access hierarchy. */
513 seen_variable_array_ref
= false;
517 tree asize
= TYPE_SIZE (TREE_TYPE (TREE_OPERAND (exp
, 0)));
518 /* We need to adjust maxsize to the whole array bitsize.
519 But we can subtract any constant offset seen so far,
520 because that would get us outside of the array otherwise. */
521 if (known_size_p (maxsize
)
523 && poly_int_tree_p (asize
))
524 maxsize
= wi::to_poly_offset (asize
) - bit_offset
;
528 /* Remember that we have seen an array ref with a variable
530 seen_variable_array_ref
= true;
535 query
= get_range_query (cfun
);
537 query
= get_global_range_query ();
539 if (TREE_CODE (index
) == SSA_NAME
540 && (low_bound
= array_ref_low_bound (exp
),
541 poly_int_tree_p (low_bound
))
542 && (unit_size
= array_ref_element_size (exp
),
543 TREE_CODE (unit_size
) == INTEGER_CST
)
544 && query
->range_of_expr (vr
, index
)
545 && vr
.kind () == VR_RANGE
)
547 wide_int min
= vr
.lower_bound ();
548 wide_int max
= vr
.upper_bound ();
549 poly_offset_int lbound
= wi::to_poly_offset (low_bound
);
550 /* Try to constrain maxsize with range information. */
552 = offset_int::from (max
, TYPE_SIGN (TREE_TYPE (index
)));
553 if (known_lt (lbound
, omax
))
555 poly_offset_int rmaxsize
;
556 rmaxsize
= (omax
- lbound
+ 1)
557 * wi::to_offset (unit_size
) << LOG2_BITS_PER_UNIT
;
558 if (!known_size_p (maxsize
)
559 || known_lt (rmaxsize
, maxsize
))
561 /* If we know an upper bound below the declared
562 one this is no longer variable. */
563 if (known_size_p (maxsize
))
564 seen_variable_array_ref
= false;
568 /* Try to adjust bit_offset with range information. */
570 = offset_int::from (min
, TYPE_SIGN (TREE_TYPE (index
)));
571 if (known_le (lbound
, omin
))
573 poly_offset_int woffset
574 = wi::sext (omin
- lbound
,
575 TYPE_PRECISION (sizetype
));
576 woffset
*= wi::to_offset (unit_size
);
577 woffset
<<= LOG2_BITS_PER_UNIT
;
578 bit_offset
+= woffset
;
579 if (known_size_p (maxsize
))
591 bit_offset
+= bitsize
;
594 case VIEW_CONVERT_EXPR
:
598 /* Via the variable index or index2 we can reach the
599 whole object. Still hand back the decl here. */
600 if (TREE_CODE (TMR_BASE (exp
)) == ADDR_EXPR
601 && (TMR_INDEX (exp
) || TMR_INDEX2 (exp
)))
603 exp
= TREE_OPERAND (TMR_BASE (exp
), 0);
610 /* We need to deal with variable arrays ending structures such as
611 struct { int length; int a[1]; } x; x.a[d]
612 struct { struct { int a; int b; } a[1]; } x; x.a[d].a
613 struct { struct { int a[1]; } a[1]; } x; x.a[0][d], x.a[d][0]
614 struct { int len; union { int a[1]; struct X x; } u; } x; x.u.a[d]
615 where we do not know maxsize for variable index accesses to
616 the array. The simplest way to conservatively deal with this
617 is to punt in the case that offset + maxsize reaches the
618 base type boundary. This needs to include possible trailing
619 padding that is there for alignment purposes. */
620 if (seen_variable_array_ref
621 && known_size_p (maxsize
)
622 && (TYPE_SIZE (TREE_TYPE (exp
)) == NULL_TREE
623 || !poly_int_tree_p (TYPE_SIZE (TREE_TYPE (exp
)))
625 (bit_offset
+ maxsize
,
626 wi::to_poly_offset (TYPE_SIZE (TREE_TYPE (exp
)))))))
629 /* Hand back the decl for MEM[&decl, off]. */
630 if (TREE_CODE (TREE_OPERAND (exp
, 0)) == ADDR_EXPR
)
632 if (integer_zerop (TREE_OPERAND (exp
, 1)))
633 exp
= TREE_OPERAND (TREE_OPERAND (exp
, 0), 0);
636 poly_offset_int off
= mem_ref_offset (exp
);
637 off
<<= LOG2_BITS_PER_UNIT
;
640 if (off
.to_shwi (&off_hwi
))
642 bit_offset
= off_hwi
;
643 exp
= TREE_OPERAND (TREE_OPERAND (exp
, 0), 0);
653 exp
= TREE_OPERAND (exp
, 0);
657 if (!bitsize
.to_shwi (psize
) || maybe_lt (*psize
, 0))
666 /* ??? Due to negative offsets in ARRAY_REF we can end up with
667 negative bit_offset here. We might want to store a zero offset
669 if (!bit_offset
.to_shwi (poffset
))
677 /* In case of a decl or constant base object we can do better. */
682 && ((flag_unconstrained_commons
&& DECL_COMMON (exp
))
683 || (DECL_EXTERNAL (exp
) && seen_variable_array_ref
)))
685 tree sz_tree
= TYPE_SIZE (TREE_TYPE (exp
));
686 /* If size is unknown, or we have read to the end, assume there
687 may be more to the structure than we are told. */
688 if (TREE_CODE (TREE_TYPE (exp
)) == ARRAY_TYPE
689 || (seen_variable_array_ref
690 && (sz_tree
== NULL_TREE
691 || !poly_int_tree_p (sz_tree
)
692 || maybe_eq (bit_offset
+ maxsize
,
693 wi::to_poly_offset (sz_tree
)))))
696 /* If maxsize is unknown adjust it according to the size of the
698 else if (!known_size_p (maxsize
)
700 && poly_int_tree_p (DECL_SIZE (exp
)))
701 maxsize
= wi::to_poly_offset (DECL_SIZE (exp
)) - bit_offset
;
703 else if (CONSTANT_CLASS_P (exp
))
705 /* If maxsize is unknown adjust it according to the size of the
706 base type constant. */
707 if (!known_size_p (maxsize
)
708 && TYPE_SIZE (TREE_TYPE (exp
))
709 && poly_int_tree_p (TYPE_SIZE (TREE_TYPE (exp
))))
710 maxsize
= (wi::to_poly_offset (TYPE_SIZE (TREE_TYPE (exp
)))
714 if (!maxsize
.to_shwi (pmax_size
)
715 || maybe_lt (*pmax_size
, 0)
716 || !endpoint_representable_p (*poffset
, *pmax_size
))
719 /* Punt if *POFFSET + *PSIZE overflows in HOST_WIDE_INT, the callers don't
720 check for such overflows individually and assume it works. */
721 if (!endpoint_representable_p (*poffset
, *psize
))
733 /* Like get_ref_base_and_extent, but for cases in which we only care
734 about constant-width accesses at constant offsets. Return null
735 if the access is anything else. */
738 get_ref_base_and_extent_hwi (tree exp
, HOST_WIDE_INT
*poffset
,
739 HOST_WIDE_INT
*psize
, bool *preverse
)
741 poly_int64 offset
, size
, max_size
;
742 HOST_WIDE_INT const_offset
, const_size
;
744 tree decl
= get_ref_base_and_extent (exp
, &offset
, &size
, &max_size
,
746 if (!offset
.is_constant (&const_offset
)
747 || !size
.is_constant (&const_size
)
749 || !known_size_p (max_size
)
750 || maybe_ne (max_size
, const_size
))
753 *poffset
= const_offset
;
759 /* Returns the base object and a constant BITS_PER_UNIT offset in *POFFSET that
760 denotes the starting address of the memory access EXP.
761 Returns NULL_TREE if the offset is not constant or any component
762 is not BITS_PER_UNIT-aligned.
763 VALUEIZE if non-NULL is used to valueize SSA names. It should return
764 its argument or a constant if the argument is known to be constant. */
767 get_addr_base_and_unit_offset_1 (tree exp
, poly_int64_pod
*poffset
,
768 tree (*valueize
) (tree
))
770 poly_int64 byte_offset
= 0;
772 /* Compute cumulative byte-offset for nested component-refs and array-refs,
773 and find the ultimate containing object. */
776 switch (TREE_CODE (exp
))
780 poly_int64 this_byte_offset
;
781 poly_uint64 this_bit_offset
;
782 if (!poly_int_tree_p (TREE_OPERAND (exp
, 2), &this_bit_offset
)
783 || !multiple_p (this_bit_offset
, BITS_PER_UNIT
,
786 byte_offset
+= this_byte_offset
;
792 tree field
= TREE_OPERAND (exp
, 1);
793 tree this_offset
= component_ref_field_offset (exp
);
794 poly_int64 hthis_offset
;
797 || !poly_int_tree_p (this_offset
, &hthis_offset
)
798 || (TREE_INT_CST_LOW (DECL_FIELD_BIT_OFFSET (field
))
802 hthis_offset
+= (TREE_INT_CST_LOW (DECL_FIELD_BIT_OFFSET (field
))
804 byte_offset
+= hthis_offset
;
809 case ARRAY_RANGE_REF
:
811 tree index
= TREE_OPERAND (exp
, 1);
812 tree low_bound
, unit_size
;
815 && TREE_CODE (index
) == SSA_NAME
)
816 index
= (*valueize
) (index
);
817 if (!poly_int_tree_p (index
))
819 low_bound
= array_ref_low_bound (exp
);
821 && TREE_CODE (low_bound
) == SSA_NAME
)
822 low_bound
= (*valueize
) (low_bound
);
823 if (!poly_int_tree_p (low_bound
))
825 unit_size
= array_ref_element_size (exp
);
826 if (TREE_CODE (unit_size
) != INTEGER_CST
)
829 /* If the resulting bit-offset is constant, track it. */
830 poly_offset_int woffset
831 = wi::sext (wi::to_poly_offset (index
)
832 - wi::to_poly_offset (low_bound
),
833 TYPE_PRECISION (sizetype
));
834 woffset
*= wi::to_offset (unit_size
);
835 byte_offset
+= woffset
.force_shwi ();
843 byte_offset
+= TREE_INT_CST_LOW (TYPE_SIZE_UNIT (TREE_TYPE (exp
)));
846 case VIEW_CONVERT_EXPR
:
851 tree base
= TREE_OPERAND (exp
, 0);
853 && TREE_CODE (base
) == SSA_NAME
)
854 base
= (*valueize
) (base
);
856 /* Hand back the decl for MEM[&decl, off]. */
857 if (TREE_CODE (base
) == ADDR_EXPR
)
859 if (!integer_zerop (TREE_OPERAND (exp
, 1)))
861 poly_offset_int off
= mem_ref_offset (exp
);
862 byte_offset
+= off
.force_shwi ();
864 exp
= TREE_OPERAND (base
, 0);
871 tree base
= TREE_OPERAND (exp
, 0);
873 && TREE_CODE (base
) == SSA_NAME
)
874 base
= (*valueize
) (base
);
876 /* Hand back the decl for MEM[&decl, off]. */
877 if (TREE_CODE (base
) == ADDR_EXPR
)
879 if (TMR_INDEX (exp
) || TMR_INDEX2 (exp
))
881 if (!integer_zerop (TMR_OFFSET (exp
)))
883 poly_offset_int off
= mem_ref_offset (exp
);
884 byte_offset
+= off
.force_shwi ();
886 exp
= TREE_OPERAND (base
, 0);
895 exp
= TREE_OPERAND (exp
, 0);
899 *poffset
= byte_offset
;
903 /* Returns the base object and a constant BITS_PER_UNIT offset in *POFFSET that
904 denotes the starting address of the memory access EXP.
905 Returns NULL_TREE if the offset is not constant or any component
906 is not BITS_PER_UNIT-aligned. */
909 get_addr_base_and_unit_offset (tree exp
, poly_int64_pod
*poffset
)
911 return get_addr_base_and_unit_offset_1 (exp
, poffset
, NULL
);
914 /* Returns true if STMT references an SSA_NAME that has
915 SSA_NAME_OCCURS_IN_ABNORMAL_PHI set, otherwise false. */
918 stmt_references_abnormal_ssa_name (gimple
*stmt
)
923 FOR_EACH_SSA_USE_OPERAND (use_p
, stmt
, oi
, SSA_OP_USE
)
925 if (SSA_NAME_OCCURS_IN_ABNORMAL_PHI (USE_FROM_PTR (use_p
)))
932 /* If STMT takes any abnormal PHI values as input, replace them with
936 replace_abnormal_ssa_names (gimple
*stmt
)
941 FOR_EACH_SSA_USE_OPERAND (use_p
, stmt
, oi
, SSA_OP_USE
)
943 tree op
= USE_FROM_PTR (use_p
);
944 if (TREE_CODE (op
) == SSA_NAME
&& SSA_NAME_OCCURS_IN_ABNORMAL_PHI (op
))
946 gimple_stmt_iterator gsi
= gsi_for_stmt (stmt
);
947 tree new_name
= make_ssa_name (TREE_TYPE (op
));
948 gassign
*assign
= gimple_build_assign (new_name
, op
);
949 gsi_insert_before (&gsi
, assign
, GSI_SAME_STMT
);
950 SET_USE (use_p
, new_name
);
955 /* Pair of tree and a sorting index, for dump_enumerated_decls. */
956 struct GTY(()) numbered_tree
963 /* Compare two declarations references by their DECL_UID / sequence number.
967 compare_decls_by_uid (const void *pa
, const void *pb
)
969 const numbered_tree
*nt_a
= ((const numbered_tree
*)pa
);
970 const numbered_tree
*nt_b
= ((const numbered_tree
*)pb
);
972 if (DECL_UID (nt_a
->t
) != DECL_UID (nt_b
->t
))
973 return DECL_UID (nt_a
->t
) - DECL_UID (nt_b
->t
);
974 return nt_a
->num
- nt_b
->num
;
977 /* Called via walk_gimple_stmt / walk_gimple_op by dump_enumerated_decls. */
979 dump_enumerated_decls_push (tree
*tp
, int *walk_subtrees
, void *data
)
981 struct walk_stmt_info
*wi
= (struct walk_stmt_info
*) data
;
982 vec
<numbered_tree
> *list
= (vec
<numbered_tree
> *) wi
->info
;
988 nt
.num
= list
->length ();
989 list
->safe_push (nt
);
994 /* Find all the declarations used by the current function, sort them by uid,
995 and emit the sorted list. Each declaration is tagged with a sequence
996 number indicating when it was found during statement / tree walking,
997 so that TDF_NOUID comparisons of anonymous declarations are still
998 meaningful. Where a declaration was encountered more than once, we
999 emit only the sequence number of the first encounter.
1000 FILE is the dump file where to output the list and FLAGS is as in
1001 print_generic_expr. */
1003 dump_enumerated_decls (FILE *file
, dump_flags_t flags
)
1009 struct walk_stmt_info wi
;
1010 auto_vec
<numbered_tree
, 40> decl_list
;
1012 memset (&wi
, '\0', sizeof (wi
));
1013 wi
.info
= (void *) &decl_list
;
1014 FOR_EACH_BB_FN (bb
, cfun
)
1016 gimple_stmt_iterator gsi
;
1018 for (gsi
= gsi_start_bb (bb
); !gsi_end_p (gsi
); gsi_next (&gsi
))
1019 if (!is_gimple_debug (gsi_stmt (gsi
)))
1020 walk_gimple_stmt (&gsi
, NULL
, dump_enumerated_decls_push
, &wi
);
1022 decl_list
.qsort (compare_decls_by_uid
);
1023 if (decl_list
.length ())
1027 tree last
= NULL_TREE
;
1029 fprintf (file
, "Declarations used by %s, sorted by DECL_UID:\n",
1030 current_function_name ());
1031 FOR_EACH_VEC_ELT (decl_list
, ix
, ntp
)
1035 fprintf (file
, "%d: ", ntp
->num
);
1036 print_generic_decl (file
, ntp
->t
, flags
);
1037 fprintf (file
, "\n");