EnumSet*.class: Regenerate
[official-gcc.git] / gcc / tree-ssa-alias.c
blob2dc23513af9a68e5b70d2e1c54a0bc02cad10c7a
1 /* Alias analysis for trees.
2 Copyright (C) 2004, 2005, 2006, 2007 Free Software Foundation, Inc.
3 Contributed by Diego Novillo <dnovillo@redhat.com>
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License as published by
9 the Free Software Foundation; either version 3, or (at your option)
10 any later version.
12 GCC is distributed in the hope that it will be useful,
13 but WITHOUT ANY WARRANTY; without even the implied warranty of
14 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 GNU General Public License for more details.
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
21 #include "config.h"
22 #include "system.h"
23 #include "coretypes.h"
24 #include "tm.h"
25 #include "tree.h"
26 #include "rtl.h"
27 #include "tm_p.h"
28 #include "hard-reg-set.h"
29 #include "basic-block.h"
30 #include "timevar.h"
31 #include "expr.h"
32 #include "ggc.h"
33 #include "langhooks.h"
34 #include "flags.h"
35 #include "function.h"
36 #include "diagnostic.h"
37 #include "tree-dump.h"
38 #include "tree-gimple.h"
39 #include "tree-flow.h"
40 #include "tree-inline.h"
41 #include "tree-pass.h"
42 #include "tree-ssa-structalias.h"
43 #include "convert.h"
44 #include "params.h"
45 #include "ipa-type-escape.h"
46 #include "vec.h"
47 #include "bitmap.h"
48 #include "vecprim.h"
49 #include "pointer-set.h"
50 #include "alloc-pool.h"
52 /* Broad overview of how aliasing works:
54 First we compute points-to sets, which is done in
55 tree-ssa-structalias.c
57 During points-to set constraint finding, a bunch of little bits of
58 information is collected.
59 This is not done because it is necessary for points-to, but because
60 points-to has to walk every statement anyway. The function performing
61 this collecting is update_alias_info.
63 Bits update_alias_info collects include:
64 1. Directly escaping variables and variables whose value escapes
65 (using is_escape_site). This is the set of variables and values that
66 escape prior to transitive closure of the clobbers.
67 2. The set of variables dereferenced on the LHS (into
68 dereferenced_ptr_stores)
69 3. The set of variables dereferenced on the RHS (into
70 dereferenced_ptr_loads)
71 4. The set of all pointers we saw.
72 5. The number of loads and stores for each variable
73 6. The number of statements touching memory
74 7. The set of address taken variables.
77 #1 is computed by a combination of is_escape_site, and counting the
78 number of uses/deref operators. This function properly accounts for
79 situations like &ptr->field, which is *not* a dereference.
81 After points-to sets are computed, the sets themselves still
82 contain points-to specific variables, such as a variable that says
83 the pointer points to anything, a variable that says the pointer
84 points to readonly memory, etc.
86 These are eliminated in a later phase, as we will see.
88 The rest of the phases are located in tree-ssa-alias.c
90 The next phase after points-to set computation is called
91 "setup_pointers_and_addressables"
93 This pass does 3 main things:
95 1. All variables that can have TREE_ADDRESSABLE removed safely (IE
96 non-globals whose address is not taken), have TREE_ADDRESSABLE
97 removed.
98 2. All variables that may be aliased (which is the set of addressable
99 variables and globals) at all, are marked for renaming, and have
100 symbol memory tags created for them.
101 3. All variables which are stored into have their SMT's added to
102 written vars.
105 After this function is run, all variables that will ever have an
106 SMT, have one, though its aliases are not filled in.
108 The next phase is to compute flow-insensitive aliasing, which in
109 our case, is a misnomer. it is really computing aliasing that
110 requires no transitive closure to be correct. In particular, it
111 uses stack vs non-stack, TBAA, etc, to determine whether two
112 symbols could *ever* alias . This phase works by going through all
113 the pointers we collected during update_alias_info, and for every
114 addressable variable in the program, seeing if they alias. If so,
115 the addressable variable is added to the symbol memory tag for the
116 pointer.
118 As part of this, we handle symbol memory tags that conflict but
119 have no aliases in common, by forcing them to have a symbol in
120 common (through unioning alias sets or adding one as an alias of
121 the other), or by adding one as an alias of another. The case of
122 conflicts with no aliases in common occurs mainly due to aliasing
123 we cannot see. In particular, it generally means we have a load
124 through a pointer whose value came from outside the function.
125 Without an addressable symbol to point to, they would get the wrong
126 answer.
128 After flow insensitive aliasing is computed, we compute name tags
129 (called compute_flow_sensitive_info). We walk each pointer we
130 collected and see if it has a usable points-to set. If so, we
131 generate a name tag using that pointer, and make an alias bitmap for
132 it. Name tags are shared between all things with the same alias
133 bitmap. The alias bitmap will be translated from what points-to
134 computed. In particular, the "anything" variable in points-to will be
135 transformed into a pruned set of SMT's and their aliases that
136 compute_flow_insensitive_aliasing computed.
137 Note that since 4.3, every pointer that points-to computed a solution for
138 will get a name tag (whereas before 4.3, only those whose set did
139 *not* include the anything variable would). At the point where name
140 tags are all assigned, symbol memory tags are dead, and could be
141 deleted, *except* on global variables. Global variables still use
142 symbol memory tags as of right now.
144 After name tags are computed, the set of clobbered variables is
145 transitively closed. In particular, we compute the set of clobbered
146 variables based on the initial set of clobbers, plus the aliases of
147 pointers which either escape, or have their value escape.
149 After this, maybe_create_global_var is run, which handles a corner
150 case where we have no call clobbered variables, but have pure and
151 non-pure functions.
153 Staring at this function, I now remember it is a hack for the fact
154 that we do not mark all globals in the program as call clobbered for a
155 function unless they are actually used in that function. Instead, we
156 only mark the set that is actually clobbered. As a result, you can
157 end up with situations where you have no call clobbered vars set.
159 After maybe_create_global_var, we set pointers with the REF_ALL flag
160 to have alias sets that include all clobbered
161 memory tags and variables.
163 After this, memory partitioning is computed (by the function
164 compute_memory_partitions) and alias sets are reworked accordingly.
166 Lastly, we delete partitions with no symbols, and clean up after
167 ourselves. */
169 /* Structure to map a variable to its alias set. */
170 struct alias_map_d
172 /* Variable and its alias set. */
173 tree var;
174 alias_set_type set;
178 /* Counters used to display statistics on alias analysis. */
179 struct alias_stats_d
181 unsigned int alias_queries;
182 unsigned int alias_mayalias;
183 unsigned int alias_noalias;
184 unsigned int simple_queries;
185 unsigned int simple_resolved;
186 unsigned int tbaa_queries;
187 unsigned int tbaa_resolved;
188 unsigned int structnoaddress_queries;
189 unsigned int structnoaddress_resolved;
193 /* Local variables. */
194 static struct alias_stats_d alias_stats;
195 static bitmap_obstack alias_bitmap_obstack;
197 /* Local functions. */
198 static void compute_flow_insensitive_aliasing (struct alias_info *);
199 static void finalize_ref_all_pointers (struct alias_info *);
200 static void dump_alias_stats (FILE *);
201 static bool may_alias_p (tree, alias_set_type, tree, alias_set_type, bool);
202 static tree create_memory_tag (tree type, bool is_type_tag);
203 static tree get_smt_for (tree, struct alias_info *);
204 static tree get_nmt_for (tree);
205 static void add_may_alias (tree, tree);
206 static struct alias_info *init_alias_info (void);
207 static void delete_alias_info (struct alias_info *);
208 static void compute_flow_sensitive_aliasing (struct alias_info *);
209 static void setup_pointers_and_addressables (struct alias_info *);
210 static void create_global_var (void);
211 static void maybe_create_global_var (void);
212 static void set_pt_anything (tree);
214 void debug_mp_info (VEC(mem_sym_stats_t,heap) *);
216 static alloc_pool mem_sym_stats_pool;
218 /* Return memory reference stats for symbol VAR. Create a new slot in
219 cfun->gimple_df->mem_sym_stats if needed. */
221 static struct mem_sym_stats_d *
222 get_mem_sym_stats_for (tree var)
224 void **slot;
225 struct mem_sym_stats_d *stats;
226 struct pointer_map_t *map = gimple_mem_ref_stats (cfun)->mem_sym_stats;
228 gcc_assert (map);
230 slot = pointer_map_insert (map, var);
231 if (*slot == NULL)
233 stats = pool_alloc (mem_sym_stats_pool);
234 memset (stats, 0, sizeof (*stats));
235 stats->var = var;
236 *slot = (void *) stats;
238 else
239 stats = (struct mem_sym_stats_d *) *slot;
241 return stats;
245 /* Set MPT to be the memory partition associated with symbol SYM. */
247 static inline void
248 set_memory_partition (tree sym, tree mpt)
250 #if defined ENABLE_CHECKING
251 if (mpt)
252 gcc_assert (TREE_CODE (mpt) == MEMORY_PARTITION_TAG
253 && !is_gimple_reg (sym));
254 #endif
256 var_ann (sym)->mpt = mpt;
257 if (mpt)
259 if (MPT_SYMBOLS (mpt) == NULL)
260 MPT_SYMBOLS (mpt) = BITMAP_ALLOC (&alias_bitmap_obstack);
262 bitmap_set_bit (MPT_SYMBOLS (mpt), DECL_UID (sym));
264 /* MPT inherits the call-clobbering attributes from SYM. */
265 if (is_call_clobbered (sym))
267 MTAG_GLOBAL (mpt) = 1;
268 mark_call_clobbered (mpt, ESCAPE_IS_GLOBAL);
274 /* Mark variable VAR as being non-addressable. */
276 static void
277 mark_non_addressable (tree var)
279 tree mpt;
281 if (!TREE_ADDRESSABLE (var))
282 return;
284 mpt = memory_partition (var);
286 if (!MTAG_P (var))
287 var_ann (var)->call_clobbered = false;
289 bitmap_clear_bit (gimple_call_clobbered_vars (cfun), DECL_UID (var));
290 TREE_ADDRESSABLE (var) = 0;
292 if (mpt)
294 /* Note that it's possible for a symbol to have an associated
295 MPT and the MPT have a NULL empty set. During
296 init_alias_info, all MPTs get their sets cleared out, but the
297 symbols still point to the old MPTs that used to hold them.
298 This is done so that compute_memory_partitions can now which
299 symbols are losing or changing partitions and mark them for
300 renaming. */
301 if (MPT_SYMBOLS (mpt))
302 bitmap_clear_bit (MPT_SYMBOLS (mpt), DECL_UID (var));
303 set_memory_partition (var, NULL_TREE);
308 /* qsort comparison function to sort type/name tags by DECL_UID. */
310 static int
311 sort_tags_by_id (const void *pa, const void *pb)
313 const_tree const a = *(const_tree const *)pa;
314 const_tree const b = *(const_tree const *)pb;
316 return DECL_UID (a) - DECL_UID (b);
319 /* Initialize WORKLIST to contain those memory tags that are marked call
320 clobbered. Initialized WORKLIST2 to contain the reasons these
321 memory tags escaped. */
323 static void
324 init_transitive_clobber_worklist (VEC (tree, heap) **worklist,
325 VEC (int, heap) **worklist2,
326 bitmap on_worklist)
328 referenced_var_iterator rvi;
329 tree curr;
331 FOR_EACH_REFERENCED_VAR (curr, rvi)
333 if (MTAG_P (curr) && is_call_clobbered (curr))
335 VEC_safe_push (tree, heap, *worklist, curr);
336 VEC_safe_push (int, heap, *worklist2,
337 var_ann (curr)->escape_mask);
338 bitmap_set_bit (on_worklist, DECL_UID (curr));
343 /* Add ALIAS to WORKLIST (and the reason for escaping REASON to WORKLIST2) if
344 ALIAS is not already marked call clobbered, and is a memory
345 tag. */
347 static void
348 add_to_worklist (tree alias, VEC (tree, heap) **worklist,
349 VEC (int, heap) **worklist2, int reason,
350 bitmap on_worklist)
352 if (MTAG_P (alias) && !is_call_clobbered (alias)
353 && !bitmap_bit_p (on_worklist, DECL_UID (alias)))
355 VEC_safe_push (tree, heap, *worklist, alias);
356 VEC_safe_push (int, heap, *worklist2, reason);
357 bitmap_set_bit (on_worklist, DECL_UID (alias));
361 /* Mark aliases of TAG as call clobbered, and place any tags on the
362 alias list that were not already call clobbered on WORKLIST. */
364 static void
365 mark_aliases_call_clobbered (tree tag, VEC (tree, heap) **worklist,
366 VEC (int, heap) **worklist2,
367 bitmap on_worklist)
369 bitmap aliases;
370 bitmap_iterator bi;
371 unsigned int i;
372 tree entry;
373 var_ann_t ta = var_ann (tag);
375 if (!MTAG_P (tag))
376 return;
377 aliases = may_aliases (tag);
378 if (!aliases)
379 return;
381 EXECUTE_IF_SET_IN_BITMAP (aliases, 0, i, bi)
383 entry = referenced_var (i);
384 /* If you clobber one part of a structure, you
385 clobber the entire thing. While this does not make
386 the world a particularly nice place, it is necessary
387 in order to allow C/C++ tricks that involve
388 pointer arithmetic to work. */
389 if (TREE_CODE (entry) == STRUCT_FIELD_TAG)
391 subvar_t svars;
392 svars = get_subvars_for_var (SFT_PARENT_VAR (entry));
393 for (; svars; svars = svars->next)
394 if (!unmodifiable_var_p (entry))
395 mark_call_clobbered (svars->var, ta->escape_mask);
397 else if (!unmodifiable_var_p (entry))
399 add_to_worklist (entry, worklist, worklist2, ta->escape_mask,
400 on_worklist);
401 mark_call_clobbered (entry, ta->escape_mask);
406 /* Tags containing global vars need to be marked as global.
407 Tags containing call clobbered vars need to be marked as call
408 clobbered. */
410 static void
411 compute_tag_properties (void)
413 referenced_var_iterator rvi;
414 tree tag;
415 bool changed = true;
416 VEC (tree, heap) *taglist = NULL;
418 FOR_EACH_REFERENCED_VAR (tag, rvi)
420 if (!MTAG_P (tag) || TREE_CODE (tag) == STRUCT_FIELD_TAG)
421 continue;
422 VEC_safe_push (tree, heap, taglist, tag);
425 /* We sort the taglist by DECL_UID, for two reasons.
426 1. To get a sequential ordering to make the bitmap accesses
427 faster.
428 2. Because of the way we compute aliases, it's more likely that
429 an earlier tag is included in a later tag, and this will reduce
430 the number of iterations.
432 If we had a real tag graph, we would just topo-order it and be
433 done with it. */
434 qsort (VEC_address (tree, taglist),
435 VEC_length (tree, taglist),
436 sizeof (tree),
437 sort_tags_by_id);
439 /* Go through each tag not marked as global, and if it aliases
440 global vars, mark it global.
442 If the tag contains call clobbered vars, mark it call
443 clobbered.
445 This loop iterates because tags may appear in the may-aliases
446 list of other tags when we group. */
448 while (changed)
450 unsigned int k;
452 changed = false;
453 for (k = 0; VEC_iterate (tree, taglist, k, tag); k++)
455 bitmap ma;
456 bitmap_iterator bi;
457 unsigned int i;
458 tree entry;
459 bool tagcc = is_call_clobbered (tag);
460 bool tagglobal = MTAG_GLOBAL (tag);
462 if (tagcc && tagglobal)
463 continue;
465 ma = may_aliases (tag);
466 if (!ma)
467 continue;
469 EXECUTE_IF_SET_IN_BITMAP (ma, 0, i, bi)
471 entry = referenced_var (i);
472 /* Call clobbered entries cause the tag to be marked
473 call clobbered. */
474 if (!tagcc && is_call_clobbered (entry))
476 mark_call_clobbered (tag, var_ann (entry)->escape_mask);
477 tagcc = true;
478 changed = true;
481 /* Global vars cause the tag to be marked global. */
482 if (!tagglobal && is_global_var (entry))
484 MTAG_GLOBAL (tag) = true;
485 changed = true;
486 tagglobal = true;
489 /* Early exit once both global and cc are set, since the
490 loop can't do any more than that. */
491 if (tagcc && tagglobal)
492 break;
496 VEC_free (tree, heap, taglist);
499 /* Set up the initial variable clobbers and globalness.
500 When this function completes, only tags whose aliases need to be
501 clobbered will be set clobbered. Tags clobbered because they
502 contain call clobbered vars are handled in compute_tag_properties. */
504 static void
505 set_initial_properties (struct alias_info *ai)
507 unsigned int i;
508 referenced_var_iterator rvi;
509 tree var;
510 tree ptr;
512 FOR_EACH_REFERENCED_VAR (var, rvi)
514 if (is_global_var (var)
515 && (!var_can_have_subvars (var)
516 || get_subvars_for_var (var) == NULL))
518 if (!unmodifiable_var_p (var))
519 mark_call_clobbered (var, ESCAPE_IS_GLOBAL);
521 else if (TREE_CODE (var) == PARM_DECL
522 && gimple_default_def (cfun, var)
523 && POINTER_TYPE_P (TREE_TYPE (var)))
525 tree def = gimple_default_def (cfun, var);
526 get_ptr_info (def)->value_escapes_p = 1;
527 get_ptr_info (def)->escape_mask |= ESCAPE_IS_PARM;
531 for (i = 0; VEC_iterate (tree, ai->processed_ptrs, i, ptr); i++)
533 struct ptr_info_def *pi = SSA_NAME_PTR_INFO (ptr);
534 tree tag = symbol_mem_tag (SSA_NAME_VAR (ptr));
536 if (pi->value_escapes_p)
538 /* If PTR escapes then its associated memory tags and
539 pointed-to variables are call-clobbered. */
540 if (pi->name_mem_tag)
541 mark_call_clobbered (pi->name_mem_tag, pi->escape_mask);
543 if (tag)
544 mark_call_clobbered (tag, pi->escape_mask);
546 if (pi->pt_vars)
548 bitmap_iterator bi;
549 unsigned int j;
550 EXECUTE_IF_SET_IN_BITMAP (pi->pt_vars, 0, j, bi)
552 tree alias = referenced_var (j);
554 /* If you clobber one part of a structure, you
555 clobber the entire thing. While this does not make
556 the world a particularly nice place, it is necessary
557 in order to allow C/C++ tricks that involve
558 pointer arithmetic to work. */
559 if (TREE_CODE (alias) == STRUCT_FIELD_TAG)
561 subvar_t svars;
562 svars = get_subvars_for_var (SFT_PARENT_VAR (alias));
563 for (; svars; svars = svars->next)
564 if (!unmodifiable_var_p (alias))
565 mark_call_clobbered (svars->var, pi->escape_mask);
567 else if (!unmodifiable_var_p (alias))
568 mark_call_clobbered (alias, pi->escape_mask);
573 /* If the name tag is call clobbered, so is the symbol tag
574 associated with the base VAR_DECL. */
575 if (pi->name_mem_tag
576 && tag
577 && is_call_clobbered (pi->name_mem_tag))
578 mark_call_clobbered (tag, pi->escape_mask);
580 /* Name tags and symbol tags that we don't know where they point
581 to, might point to global memory, and thus, are clobbered.
583 FIXME: This is not quite right. They should only be
584 clobbered if value_escapes_p is true, regardless of whether
585 they point to global memory or not.
586 So removing this code and fixing all the bugs would be nice.
587 It is the cause of a bunch of clobbering. */
588 if ((pi->pt_global_mem || pi->pt_anything)
589 && pi->is_dereferenced && pi->name_mem_tag)
591 mark_call_clobbered (pi->name_mem_tag, ESCAPE_IS_GLOBAL);
592 MTAG_GLOBAL (pi->name_mem_tag) = true;
595 if ((pi->pt_global_mem || pi->pt_anything)
596 && pi->is_dereferenced
597 && tag)
599 mark_call_clobbered (tag, ESCAPE_IS_GLOBAL);
600 MTAG_GLOBAL (tag) = true;
605 /* Compute which variables need to be marked call clobbered because
606 their tag is call clobbered, and which tags need to be marked
607 global because they contain global variables. */
609 static void
610 compute_call_clobbered (struct alias_info *ai)
612 VEC (tree, heap) *worklist = NULL;
613 VEC (int,heap) *worklist2 = NULL;
614 bitmap on_worklist;
616 timevar_push (TV_CALL_CLOBBER);
617 on_worklist = BITMAP_ALLOC (NULL);
619 set_initial_properties (ai);
620 init_transitive_clobber_worklist (&worklist, &worklist2, on_worklist);
621 while (VEC_length (tree, worklist) != 0)
623 tree curr = VEC_pop (tree, worklist);
624 int reason = VEC_pop (int, worklist2);
626 bitmap_clear_bit (on_worklist, DECL_UID (curr));
627 mark_call_clobbered (curr, reason);
628 mark_aliases_call_clobbered (curr, &worklist, &worklist2,
629 on_worklist);
631 VEC_free (tree, heap, worklist);
632 VEC_free (int, heap, worklist2);
633 BITMAP_FREE (on_worklist);
634 compute_tag_properties ();
635 timevar_pop (TV_CALL_CLOBBER);
639 /* Dump memory partition information to FILE. */
641 static void
642 dump_memory_partitions (FILE *file)
644 unsigned i, npart;
645 unsigned long nsyms;
646 tree mpt;
648 fprintf (file, "\nMemory partitions\n\n");
649 for (i = 0, npart = 0, nsyms = 0;
650 VEC_iterate (tree, gimple_ssa_operands (cfun)->mpt_table, i, mpt);
651 i++)
653 if (mpt)
655 bitmap syms = MPT_SYMBOLS (mpt);
656 unsigned long n = (syms) ? bitmap_count_bits (syms) : 0;
658 fprintf (file, "#%u: ", i);
659 print_generic_expr (file, mpt, 0);
660 fprintf (file, ": %lu elements: ", n);
661 dump_decl_set (file, syms);
662 npart++;
663 nsyms += n;
667 fprintf (file, "\n%u memory partitions holding %lu symbols\n", npart, nsyms);
671 /* Dump memory partition information to stderr. */
673 void
674 debug_memory_partitions (void)
676 dump_memory_partitions (stderr);
680 /* Return true if memory partitioning is required given the memory
681 reference estimates in STATS. */
683 static inline bool
684 need_to_partition_p (struct mem_ref_stats_d *stats)
686 long num_vops = stats->num_vuses + stats->num_vdefs;
687 long avg_vops = CEIL (num_vops, stats->num_mem_stmts);
688 return (num_vops > (long) MAX_ALIASED_VOPS
689 && avg_vops > (long) AVG_ALIASED_VOPS);
693 /* Count the actual number of virtual operators in CFUN. Note that
694 this is only meaningful after virtual operands have been populated,
695 so it should be invoked at the end of compute_may_aliases.
697 The number of virtual operators are stored in *NUM_VDEFS_P and
698 *NUM_VUSES_P, the number of partitioned symbols in
699 *NUM_PARTITIONED_P and the number of unpartitioned symbols in
700 *NUM_UNPARTITIONED_P.
702 If any of these pointers is NULL the corresponding count is not
703 computed. */
705 static void
706 count_mem_refs (long *num_vuses_p, long *num_vdefs_p,
707 long *num_partitioned_p, long *num_unpartitioned_p)
709 block_stmt_iterator bsi;
710 basic_block bb;
711 long num_vdefs, num_vuses, num_partitioned, num_unpartitioned;
712 referenced_var_iterator rvi;
713 tree sym;
715 num_vuses = num_vdefs = num_partitioned = num_unpartitioned = 0;
717 if (num_vuses_p || num_vdefs_p)
718 FOR_EACH_BB (bb)
719 for (bsi = bsi_start (bb); !bsi_end_p (bsi); bsi_next (&bsi))
721 tree stmt = bsi_stmt (bsi);
722 if (stmt_references_memory_p (stmt))
724 num_vuses += NUM_SSA_OPERANDS (stmt, SSA_OP_VUSE);
725 num_vdefs += NUM_SSA_OPERANDS (stmt, SSA_OP_VDEF);
729 if (num_partitioned_p || num_unpartitioned_p)
730 FOR_EACH_REFERENCED_VAR (sym, rvi)
732 if (is_gimple_reg (sym))
733 continue;
735 if (memory_partition (sym))
736 num_partitioned++;
737 else
738 num_unpartitioned++;
741 if (num_vdefs_p)
742 *num_vdefs_p = num_vdefs;
744 if (num_vuses_p)
745 *num_vuses_p = num_vuses;
747 if (num_partitioned_p)
748 *num_partitioned_p = num_partitioned;
750 if (num_unpartitioned_p)
751 *num_unpartitioned_p = num_unpartitioned;
755 /* Dump memory reference stats for function CFUN to FILE. */
757 void
758 dump_mem_ref_stats (FILE *file)
760 long actual_num_vuses, actual_num_vdefs;
761 long num_partitioned, num_unpartitioned;
762 struct mem_ref_stats_d *stats;
764 stats = gimple_mem_ref_stats (cfun);
766 count_mem_refs (&actual_num_vuses, &actual_num_vdefs, &num_partitioned,
767 &num_unpartitioned);
769 fprintf (file, "\nMemory reference statistics for %s\n\n",
770 lang_hooks.decl_printable_name (current_function_decl, 2));
772 fprintf (file, "Number of memory statements: %ld\n",
773 stats->num_mem_stmts);
774 fprintf (file, "Number of call sites: %ld\n",
775 stats->num_call_sites);
776 fprintf (file, "Number of pure/const call sites: %ld\n",
777 stats->num_pure_const_call_sites);
778 fprintf (file, "Number of asm sites: %ld\n",
779 stats->num_asm_sites);
780 fprintf (file, "Estimated number of loads: %ld (%ld/stmt)\n",
781 stats->num_vuses,
782 (stats->num_mem_stmts)
783 ? CEIL (stats->num_vuses, stats->num_mem_stmts)
784 : 0);
785 fprintf (file, "Actual number of loads: %ld (%ld/stmt)\n",
786 actual_num_vuses,
787 (stats->num_mem_stmts)
788 ? CEIL (actual_num_vuses, stats->num_mem_stmts)
789 : 0);
791 if (actual_num_vuses > stats->num_vuses + (stats->num_vuses / 25))
792 fprintf (file, "\t(warning: estimation is lower by more than 25%%)\n");
794 fprintf (file, "Estimated number of stores: %ld (%ld/stmt)\n",
795 stats->num_vdefs,
796 (stats->num_mem_stmts)
797 ? CEIL (stats->num_vdefs, stats->num_mem_stmts)
798 : 0);
799 fprintf (file, "Actual number of stores: %ld (%ld/stmt)\n",
800 actual_num_vdefs,
801 (stats->num_mem_stmts)
802 ? CEIL (actual_num_vdefs, stats->num_mem_stmts)
803 : 0);
805 if (actual_num_vdefs > stats->num_vdefs + (stats->num_vdefs / 25))
806 fprintf (file, "\t(warning: estimation is lower by more than 25%%)\n");
808 fprintf (file, "Partitioning thresholds: MAX = %d AVG = %d "
809 "(%sNEED TO PARTITION)\n", MAX_ALIASED_VOPS, AVG_ALIASED_VOPS,
810 stats->num_mem_stmts && need_to_partition_p (stats) ? "" : "NO ");
811 fprintf (file, "Number of partitioned symbols: %ld\n", num_partitioned);
812 fprintf (file, "Number of unpartitioned symbols: %ld\n", num_unpartitioned);
816 /* Dump memory reference stats for function FN to stderr. */
818 void
819 debug_mem_ref_stats (void)
821 dump_mem_ref_stats (stderr);
825 /* Dump memory reference stats for variable VAR to FILE. */
827 static void
828 dump_mem_sym_stats (FILE *file, tree var)
830 mem_sym_stats_t stats = mem_sym_stats (cfun, var);
832 if (stats == NULL)
833 return;
835 fprintf (file, "read frequency: %6ld, write frequency: %6ld, "
836 "direct reads: %3ld, direct writes: %3ld, "
837 "indirect reads: %4ld, indirect writes: %4ld, symbol: ",
838 stats->frequency_reads, stats->frequency_writes,
839 stats->num_direct_reads, stats->num_direct_writes,
840 stats->num_indirect_reads, stats->num_indirect_writes);
841 print_generic_expr (file, stats->var, 0);
842 fprintf (file, ", tags: ");
843 dump_decl_set (file, stats->parent_tags);
847 /* Dump memory reference stats for variable VAR to stderr. */
849 void
850 debug_mem_sym_stats (tree var)
852 dump_mem_sym_stats (stderr, var);
856 /* Dump memory reference stats for all memory symbols to FILE. */
858 static void
859 dump_all_mem_sym_stats (FILE *file)
861 referenced_var_iterator rvi;
862 tree sym;
864 FOR_EACH_REFERENCED_VAR (sym, rvi)
866 if (is_gimple_reg (sym))
867 continue;
869 dump_mem_sym_stats (file, sym);
874 /* Dump memory reference stats for all memory symbols to stderr. */
876 void
877 debug_all_mem_sym_stats (void)
879 dump_all_mem_sym_stats (stderr);
883 /* Dump the MP_INFO array to FILE. */
885 static void
886 dump_mp_info (FILE *file, VEC(mem_sym_stats_t,heap) *mp_info)
888 unsigned i;
889 mem_sym_stats_t mp_p;
891 for (i = 0; VEC_iterate (mem_sym_stats_t, mp_info, i, mp_p); i++)
892 if (!mp_p->partitioned_p)
893 dump_mem_sym_stats (file, mp_p->var);
897 /* Dump the MP_INFO array to stderr. */
899 void
900 debug_mp_info (VEC(mem_sym_stats_t,heap) *mp_info)
902 dump_mp_info (stderr, mp_info);
906 /* Update memory reference stats for symbol VAR in statement STMT.
907 NUM_DIRECT_READS and NUM_DIRECT_WRITES specify the number of times
908 that VAR is read/written in STMT (indirect reads/writes are not
909 recorded by this function, see compute_memory_partitions). */
911 void
912 update_mem_sym_stats_from_stmt (tree var, tree stmt, long num_direct_reads,
913 long num_direct_writes)
915 mem_sym_stats_t stats;
917 gcc_assert (num_direct_reads >= 0 && num_direct_writes >= 0);
919 stats = get_mem_sym_stats_for (var);
921 stats->num_direct_reads += num_direct_reads;
922 stats->frequency_reads += ((long) bb_for_stmt (stmt)->frequency
923 * num_direct_reads);
925 stats->num_direct_writes += num_direct_writes;
926 stats->frequency_writes += ((long) bb_for_stmt (stmt)->frequency
927 * num_direct_writes);
931 /* The list is sorted by increasing partitioning score (PSCORE).
932 This score is computed such that symbols with high scores are
933 those that are least likely to be partitioned. Given a symbol
934 MP->VAR, PSCORE(S) is the result of the following weighted sum
936 PSCORE(S) = FW * 64 + FR * 32
937 + DW * 16 + DR * 8
938 + IW * 4 + IR * 2
939 + NO_ALIAS
941 where
943 FW Execution frequency of writes to S
944 FR Execution frequency of reads from S
945 DW Number of direct writes to S
946 DR Number of direct reads from S
947 IW Number of indirect writes to S
948 IR Number of indirect reads from S
949 NO_ALIAS State of the NO_ALIAS* flags
951 The basic idea here is that symbols that are frequently
952 written-to in hot paths of the code are the last to be considered
953 for partitioning. */
955 static inline long
956 pscore (mem_sym_stats_t mp)
958 return mp->frequency_writes * 64 + mp->frequency_reads * 32
959 + mp->num_direct_writes * 16 + mp->num_direct_reads * 8
960 + mp->num_indirect_writes * 4 + mp->num_indirect_reads * 2
961 + var_ann (mp->var)->noalias_state;
965 /* Given two MP_INFO entries MP1 and MP2, return -1 if MP1->VAR should
966 be partitioned before MP2->VAR, 0 if they are the same or 1 if
967 MP1->VAR should be partitioned after MP2->VAR. */
969 static inline int
970 compare_mp_info_entries (mem_sym_stats_t mp1, mem_sym_stats_t mp2)
972 long pscore1 = pscore (mp1);
973 long pscore2 = pscore (mp2);
975 if (pscore1 < pscore2)
976 return -1;
977 else if (pscore1 > pscore2)
978 return 1;
979 else
980 return 0;
984 /* Comparison routine for qsort. The list is sorted by increasing
985 partitioning score (PSCORE). This score is computed such that
986 symbols with high scores are those that are least likely to be
987 partitioned. */
989 static int
990 mp_info_cmp (const void *p, const void *q)
992 mem_sym_stats_t e1 = *((const mem_sym_stats_t *) p);
993 mem_sym_stats_t e2 = *((const mem_sym_stats_t *) q);
994 return compare_mp_info_entries (e1, e2);
998 /* Sort the array of reference counts used to compute memory partitions.
999 Elements are sorted in ascending order of execution frequency and
1000 descending order of virtual operators needed. */
1002 static inline void
1003 sort_mp_info (VEC(mem_sym_stats_t,heap) *list)
1005 unsigned num = VEC_length (mem_sym_stats_t, list);
1007 if (num < 2)
1008 return;
1010 if (num == 2)
1012 if (compare_mp_info_entries (VEC_index (mem_sym_stats_t, list, 0),
1013 VEC_index (mem_sym_stats_t, list, 1)) > 0)
1015 /* Swap elements if they are in the wrong order. */
1016 mem_sym_stats_t tmp = VEC_index (mem_sym_stats_t, list, 0);
1017 VEC_replace (mem_sym_stats_t, list, 0,
1018 VEC_index (mem_sym_stats_t, list, 1));
1019 VEC_replace (mem_sym_stats_t, list, 1, tmp);
1022 return;
1025 /* There are 3 or more elements, call qsort. */
1026 qsort (VEC_address (mem_sym_stats_t, list),
1027 VEC_length (mem_sym_stats_t, list),
1028 sizeof (mem_sym_stats_t),
1029 mp_info_cmp);
1033 /* Return the memory partition tag (MPT) associated with memory
1034 symbol SYM. */
1036 static tree
1037 get_mpt_for (tree sym)
1039 tree mpt;
1041 /* Don't create a new tag unnecessarily. */
1042 mpt = memory_partition (sym);
1043 if (mpt == NULL_TREE)
1045 mpt = create_tag_raw (MEMORY_PARTITION_TAG, TREE_TYPE (sym), "MPT");
1046 TREE_ADDRESSABLE (mpt) = 0;
1047 add_referenced_var (mpt);
1048 VEC_safe_push (tree, heap, gimple_ssa_operands (cfun)->mpt_table, mpt);
1049 gcc_assert (MPT_SYMBOLS (mpt) == NULL);
1050 set_memory_partition (sym, mpt);
1053 return mpt;
1057 /* Add MP_P->VAR to a memory partition and return the partition. */
1059 static tree
1060 find_partition_for (mem_sym_stats_t mp_p)
1062 unsigned i;
1063 VEC(tree,heap) *mpt_table;
1064 tree mpt;
1066 mpt_table = gimple_ssa_operands (cfun)->mpt_table;
1067 mpt = NULL_TREE;
1069 /* Find an existing partition for MP_P->VAR. */
1070 for (i = 0; VEC_iterate (tree, mpt_table, i, mpt); i++)
1072 mem_sym_stats_t mpt_stats;
1074 /* If MPT does not have any symbols yet, use it. */
1075 if (MPT_SYMBOLS (mpt) == NULL)
1076 break;
1078 /* Otherwise, see if MPT has common parent tags with MP_P->VAR,
1079 but avoid grouping clobbered variables with non-clobbered
1080 variables (otherwise, this tends to creates a single memory
1081 partition because other call-clobbered variables may have
1082 common parent tags with non-clobbered ones). */
1083 mpt_stats = get_mem_sym_stats_for (mpt);
1084 if (mp_p->parent_tags
1085 && mpt_stats->parent_tags
1086 && is_call_clobbered (mpt) == is_call_clobbered (mp_p->var)
1087 && bitmap_intersect_p (mpt_stats->parent_tags, mp_p->parent_tags))
1088 break;
1090 /* If no common parent tags are found, see if both MPT and
1091 MP_P->VAR are call-clobbered. */
1092 if (is_call_clobbered (mpt) && is_call_clobbered (mp_p->var))
1093 break;
1096 if (mpt == NULL_TREE)
1097 mpt = get_mpt_for (mp_p->var);
1098 else
1099 set_memory_partition (mp_p->var, mpt);
1101 mp_p->partitioned_p = true;
1103 mark_sym_for_renaming (mp_p->var);
1104 mark_sym_for_renaming (mpt);
1106 return mpt;
1110 /* Rewrite the alias set for TAG to use the newly created partitions.
1111 If TAG is NULL, rewrite the set of call-clobbered variables.
1112 NEW_ALIASES is a scratch bitmap to build the new set of aliases for
1113 TAG. */
1115 static void
1116 rewrite_alias_set_for (tree tag, bitmap new_aliases)
1118 bitmap_iterator bi;
1119 unsigned i;
1120 tree mpt, sym;
1122 EXECUTE_IF_SET_IN_BITMAP (MTAG_ALIASES (tag), 0, i, bi)
1124 sym = referenced_var (i);
1125 mpt = memory_partition (sym);
1126 if (mpt)
1127 bitmap_set_bit (new_aliases, DECL_UID (mpt));
1128 else
1129 bitmap_set_bit (new_aliases, DECL_UID (sym));
1132 /* Rebuild the may-alias array for TAG. */
1133 bitmap_copy (MTAG_ALIASES (tag), new_aliases);
1137 /* Determine how many virtual operands can be saved by partitioning
1138 MP_P->VAR into MPT. When a symbol S is thrown inside a partition
1139 P, every virtual operand that used to reference S will now
1140 reference P. Whether it reduces the number of virtual operands
1141 depends on:
1143 1- Direct references to S are never saved. Instead of the virtual
1144 operand to S, we will now have a virtual operand to P.
1146 2- Indirect references to S are reduced only for those memory tags
1147 holding S that already had other symbols partitioned into P.
1148 For instance, if a memory tag T has the alias set { a b S c },
1149 the first time we partition S into P, the alias set will become
1150 { a b P c }, so no virtual operands will be saved. However, if
1151 we now partition symbol 'c' into P, then the alias set for T
1152 will become { a b P }, so we will be saving one virtual operand
1153 for every indirect reference to 'c'.
1155 3- Is S is call-clobbered, we save as many virtual operands as
1156 call/asm sites exist in the code, but only if other
1157 call-clobbered symbols have been grouped into P. The first
1158 call-clobbered symbol that we group does not produce any
1159 savings.
1161 MEM_REF_STATS points to CFUN's memory reference information. */
1163 static void
1164 estimate_vop_reduction (struct mem_ref_stats_d *mem_ref_stats,
1165 mem_sym_stats_t mp_p, tree mpt)
1167 unsigned i;
1168 bitmap_iterator bi;
1169 mem_sym_stats_t mpt_stats;
1171 /* We should only get symbols with indirect references here. */
1172 gcc_assert (mp_p->num_indirect_reads > 0 || mp_p->num_indirect_writes > 0);
1174 /* Note that the only statistics we keep for MPT is the set of
1175 parent tags to know which memory tags have had alias members
1176 partitioned, and the indicator has_call_clobbered_vars.
1177 Reference counts are not important for MPT. */
1178 mpt_stats = get_mem_sym_stats_for (mpt);
1180 /* Traverse all the parent tags for MP_P->VAR. For every tag T, if
1181 partition P is already grouping aliases of T, then reduce the
1182 number of virtual operands by the number of direct references
1183 to T. */
1184 if (mp_p->parent_tags)
1186 if (mpt_stats->parent_tags == NULL)
1187 mpt_stats->parent_tags = BITMAP_ALLOC (&alias_bitmap_obstack);
1189 EXECUTE_IF_SET_IN_BITMAP (mp_p->parent_tags, 0, i, bi)
1191 if (bitmap_bit_p (mpt_stats->parent_tags, i))
1193 /* Partition MPT is already partitioning symbols in the
1194 alias set for TAG. This means that we are now saving
1195 1 virtual operand for every direct reference to TAG. */
1196 tree tag = referenced_var (i);
1197 mem_sym_stats_t tag_stats = mem_sym_stats (cfun, tag);
1198 mem_ref_stats->num_vuses -= tag_stats->num_direct_reads;
1199 mem_ref_stats->num_vdefs -= tag_stats->num_direct_writes;
1201 else
1203 /* This is the first symbol in tag I's alias set that is
1204 being grouped under MPT. We will not save any
1205 virtual operands this time, but record that MPT is
1206 grouping a symbol from TAG's alias set so that the
1207 next time we get the savings. */
1208 bitmap_set_bit (mpt_stats->parent_tags, i);
1213 /* If MP_P->VAR is call-clobbered, and MPT is already grouping
1214 call-clobbered symbols, then we will save as many virtual
1215 operands as asm/call sites there are. */
1216 if (is_call_clobbered (mp_p->var))
1218 if (mpt_stats->has_call_clobbered_vars)
1219 mem_ref_stats->num_vdefs -= mem_ref_stats->num_call_sites
1220 + mem_ref_stats->num_asm_sites;
1221 else
1222 mpt_stats->has_call_clobbered_vars = true;
1227 /* Helper for compute_memory_partitions. Transfer reference counts
1228 from pointers to their pointed-to sets. Counters for pointers were
1229 computed by update_alias_info. MEM_REF_STATS points to CFUN's
1230 memory reference information. */
1232 static void
1233 update_reference_counts (struct mem_ref_stats_d *mem_ref_stats)
1235 unsigned i;
1236 bitmap_iterator bi;
1237 mem_sym_stats_t sym_stats;
1239 for (i = 1; i < num_ssa_names; i++)
1241 tree ptr;
1242 struct ptr_info_def *pi;
1244 ptr = ssa_name (i);
1245 if (ptr
1246 && POINTER_TYPE_P (TREE_TYPE (ptr))
1247 && (pi = SSA_NAME_PTR_INFO (ptr)) != NULL
1248 && pi->is_dereferenced)
1250 unsigned j;
1251 bitmap_iterator bj;
1252 tree tag;
1253 mem_sym_stats_t ptr_stats, tag_stats;
1255 /* If PTR has flow-sensitive points-to information, use
1256 PTR's name tag, otherwise use the symbol tag associated
1257 with PTR's symbol. */
1258 if (pi->name_mem_tag)
1259 tag = pi->name_mem_tag;
1260 else
1261 tag = symbol_mem_tag (SSA_NAME_VAR (ptr));
1263 ptr_stats = get_mem_sym_stats_for (ptr);
1264 tag_stats = get_mem_sym_stats_for (tag);
1266 /* TAG has as many direct references as dereferences we
1267 found for its parent pointer. */
1268 tag_stats->num_direct_reads += ptr_stats->num_direct_reads;
1269 tag_stats->num_direct_writes += ptr_stats->num_direct_writes;
1271 /* All the dereferences of pointer PTR are considered direct
1272 references to PTR's memory tag (TAG). In turn,
1273 references to TAG will become virtual operands for every
1274 symbol in TAG's alias set. So, for every symbol ALIAS in
1275 TAG's alias set, add as many indirect references to ALIAS
1276 as direct references there are for TAG. */
1277 if (MTAG_ALIASES (tag))
1278 EXECUTE_IF_SET_IN_BITMAP (MTAG_ALIASES (tag), 0, j, bj)
1280 tree alias = referenced_var (j);
1281 sym_stats = get_mem_sym_stats_for (alias);
1283 /* All the direct references to TAG are indirect references
1284 to ALIAS. */
1285 sym_stats->num_indirect_reads += ptr_stats->num_direct_reads;
1286 sym_stats->num_indirect_writes += ptr_stats->num_direct_writes;
1287 sym_stats->frequency_reads += ptr_stats->frequency_reads;
1288 sym_stats->frequency_writes += ptr_stats->frequency_writes;
1290 /* Indicate that TAG is one of ALIAS's parent tags. */
1291 if (sym_stats->parent_tags == NULL)
1292 sym_stats->parent_tags = BITMAP_ALLOC (&alias_bitmap_obstack);
1293 bitmap_set_bit (sym_stats->parent_tags, DECL_UID (tag));
1298 /* Call-clobbered symbols are indirectly written at every
1299 call/asm site. */
1300 EXECUTE_IF_SET_IN_BITMAP (gimple_call_clobbered_vars (cfun), 0, i, bi)
1302 tree sym = referenced_var (i);
1303 sym_stats = get_mem_sym_stats_for (sym);
1304 sym_stats->num_indirect_writes += mem_ref_stats->num_call_sites
1305 + mem_ref_stats->num_asm_sites;
1308 /* Addressable symbols are indirectly written at some ASM sites.
1309 Since only ASM sites that clobber memory actually affect
1310 addressable symbols, this is an over-estimation. */
1311 EXECUTE_IF_SET_IN_BITMAP (gimple_addressable_vars (cfun), 0, i, bi)
1313 tree sym = referenced_var (i);
1314 sym_stats = get_mem_sym_stats_for (sym);
1315 sym_stats->num_indirect_writes += mem_ref_stats->num_asm_sites;
1320 /* Helper for compute_memory_partitions. Add all memory symbols to
1321 *MP_INFO_P and compute the initial estimate for the total number of
1322 virtual operands needed. MEM_REF_STATS points to CFUN's memory
1323 reference information. On exit, *TAGS_P will contain the list of
1324 memory tags whose alias set need to be rewritten after
1325 partitioning. */
1327 static void
1328 build_mp_info (struct mem_ref_stats_d *mem_ref_stats,
1329 VEC(mem_sym_stats_t,heap) **mp_info_p,
1330 VEC(tree,heap) **tags_p)
1332 tree var;
1333 referenced_var_iterator rvi;
1335 FOR_EACH_REFERENCED_VAR (var, rvi)
1337 mem_sym_stats_t sym_stats;
1338 tree old_mpt;
1340 /* We are only interested in memory symbols other than MPTs. */
1341 if (is_gimple_reg (var) || TREE_CODE (var) == MEMORY_PARTITION_TAG)
1342 continue;
1344 /* Collect memory tags into the TAGS array so that we can
1345 rewrite their alias sets after partitioning. */
1346 if (MTAG_P (var) && MTAG_ALIASES (var))
1347 VEC_safe_push (tree, heap, *tags_p, var);
1349 /* Since we are going to re-compute partitions, any symbols that
1350 used to belong to a partition must be detached from it and
1351 marked for renaming. */
1352 if ((old_mpt = memory_partition (var)) != NULL)
1354 mark_sym_for_renaming (old_mpt);
1355 set_memory_partition (var, NULL_TREE);
1356 mark_sym_for_renaming (var);
1359 sym_stats = get_mem_sym_stats_for (var);
1361 /* Add VAR's reference info to MP_INFO. Note that the only
1362 symbols that make sense to partition are those that have
1363 indirect references. If a symbol S is always directly
1364 referenced, partitioning it will not reduce the number of
1365 virtual operators. The only symbols that are profitable to
1366 partition are those that belong to alias sets and/or are
1367 call-clobbered. */
1368 if (sym_stats->num_indirect_reads > 0
1369 || sym_stats->num_indirect_writes > 0)
1370 VEC_safe_push (mem_sym_stats_t, heap, *mp_info_p, sym_stats);
1372 /* Update the number of estimated VOPS. Note that direct
1373 references to memory tags are always counted as indirect
1374 references to their alias set members, so if a memory tag has
1375 aliases, do not count its direct references to avoid double
1376 accounting. */
1377 if (!MTAG_P (var) || !MTAG_ALIASES (var))
1379 mem_ref_stats->num_vuses += sym_stats->num_direct_reads;
1380 mem_ref_stats->num_vdefs += sym_stats->num_direct_writes;
1383 mem_ref_stats->num_vuses += sym_stats->num_indirect_reads;
1384 mem_ref_stats->num_vdefs += sym_stats->num_indirect_writes;
1389 /* Compute memory partitions. A memory partition (MPT) is an
1390 arbitrary grouping of memory symbols, such that references to one
1391 member of the group is considered a reference to all the members of
1392 the group.
1394 As opposed to alias sets in memory tags, the grouping into
1395 partitions is completely arbitrary and only done to reduce the
1396 number of virtual operands. The only rule that needs to be
1397 observed when creating memory partitions is that given two memory
1398 partitions MPT.i and MPT.j, they must not contain symbols in
1399 common.
1401 Memory partitions are used when putting the program into Memory-SSA
1402 form. In particular, in Memory-SSA PHI nodes are not computed for
1403 individual memory symbols. They are computed for memory
1404 partitions. This reduces the amount of PHI nodes in the SSA graph
1405 at the expense of precision (i.e., it makes unrelated stores affect
1406 each other).
1408 However, it is possible to increase precision by changing this
1409 partitioning scheme. For instance, if the partitioning scheme is
1410 such that get_mpt_for is the identity function (that is,
1411 get_mpt_for (s) = s), this will result in ultimate precision at the
1412 expense of huge SSA webs.
1414 At the other extreme, a partitioning scheme that groups all the
1415 symbols in the same set results in minimal SSA webs and almost
1416 total loss of precision.
1418 There partitioning heuristic uses three parameters to decide the
1419 order in which symbols are processed. The list of symbols is
1420 sorted so that symbols that are more likely to be partitioned are
1421 near the top of the list:
1423 - Execution frequency. If a memory references is in a frequently
1424 executed code path, grouping it into a partition may block useful
1425 transformations and cause sub-optimal code generation. So, the
1426 partition heuristic tries to avoid grouping symbols with high
1427 execution frequency scores. Execution frequency is taken
1428 directly from the basic blocks where every reference is made (see
1429 update_mem_sym_stats_from_stmt), which in turn uses the
1430 profile guided machinery, so if the program is compiled with PGO
1431 enabled, more accurate partitioning decisions will be made.
1433 - Number of references. Symbols with few references in the code,
1434 are partitioned before symbols with many references.
1436 - NO_ALIAS attributes. Symbols with any of the NO_ALIAS*
1437 attributes are partitioned after symbols marked MAY_ALIAS.
1439 Once the list is sorted, the partitioning proceeds as follows:
1441 1- For every symbol S in MP_INFO, create a new memory partition MP,
1442 if necessary. To avoid memory partitions that contain symbols
1443 from non-conflicting alias sets, memory partitions are
1444 associated to the memory tag that holds S in its alias set. So,
1445 when looking for a memory partition for S, the memory partition
1446 associated with one of the memory tags holding S is chosen. If
1447 none exists, a new one is created.
1449 2- Add S to memory partition MP.
1451 3- Reduce by 1 the number of VOPS for every memory tag holding S.
1453 4- If the total number of VOPS is less than MAX_ALIASED_VOPS or the
1454 average number of VOPS per statement is less than
1455 AVG_ALIASED_VOPS, stop. Otherwise, go to the next symbol in the
1456 list. */
1458 static void
1459 compute_memory_partitions (void)
1461 tree tag;
1462 unsigned i;
1463 mem_sym_stats_t mp_p;
1464 VEC(mem_sym_stats_t,heap) *mp_info;
1465 bitmap new_aliases;
1466 VEC(tree,heap) *tags;
1467 struct mem_ref_stats_d *mem_ref_stats;
1468 int prev_max_aliased_vops;
1470 mem_ref_stats = gimple_mem_ref_stats (cfun);
1471 gcc_assert (mem_ref_stats->num_vuses == 0 && mem_ref_stats->num_vdefs == 0);
1473 if (mem_ref_stats->num_mem_stmts == 0)
1474 return;
1476 timevar_push (TV_MEMORY_PARTITIONING);
1478 mp_info = NULL;
1479 tags = NULL;
1480 prev_max_aliased_vops = MAX_ALIASED_VOPS;
1482 /* Since we clearly cannot lower the number of virtual operators
1483 below the total number of memory statements in the function, we
1484 may need to adjust MAX_ALIASED_VOPS beforehand. */
1485 if (MAX_ALIASED_VOPS < mem_ref_stats->num_mem_stmts)
1486 MAX_ALIASED_VOPS = mem_ref_stats->num_mem_stmts;
1488 /* Update reference stats for all the pointed-to variables and
1489 memory tags. */
1490 update_reference_counts (mem_ref_stats);
1492 /* Add all the memory symbols to MP_INFO. */
1493 build_mp_info (mem_ref_stats, &mp_info, &tags);
1495 /* No partitions required if we are below the threshold. */
1496 if (!need_to_partition_p (mem_ref_stats))
1498 if (dump_file)
1499 fprintf (dump_file, "\nMemory partitioning NOT NEEDED for %s\n",
1500 get_name (current_function_decl));
1501 goto done;
1504 /* Sort the MP_INFO array so that symbols that should be partitioned
1505 first are near the top of the list. */
1506 sort_mp_info (mp_info);
1508 if (dump_file)
1510 fprintf (dump_file, "\nMemory partitioning NEEDED for %s\n\n",
1511 get_name (current_function_decl));
1512 fprintf (dump_file, "Memory symbol references before partitioning:\n");
1513 dump_mp_info (dump_file, mp_info);
1516 /* Create partitions for variables in MP_INFO until we have enough
1517 to lower the total number of VOPS below MAX_ALIASED_VOPS or if
1518 the average number of VOPS per statement is below
1519 AVG_ALIASED_VOPS. */
1520 for (i = 0; VEC_iterate (mem_sym_stats_t, mp_info, i, mp_p); i++)
1522 tree mpt;
1524 /* If we are below the threshold, stop. */
1525 if (!need_to_partition_p (mem_ref_stats))
1526 break;
1528 mpt = find_partition_for (mp_p);
1529 estimate_vop_reduction (mem_ref_stats, mp_p, mpt);
1532 /* After partitions have been created, rewrite alias sets to use
1533 them instead of the original symbols. This way, if the alias set
1534 was computed as { a b c d e f }, and the subset { b e f } was
1535 grouped into partition MPT.3, then the new alias set for the tag
1536 will be { a c d MPT.3 }.
1538 Note that this is not strictly necessary. The operand scanner
1539 will always check if a symbol belongs to a partition when adding
1540 virtual operands. However, by reducing the size of the alias
1541 sets to be scanned, the work needed inside the operand scanner is
1542 significantly reduced. */
1543 new_aliases = BITMAP_ALLOC (&alias_bitmap_obstack);
1545 for (i = 0; VEC_iterate (tree, tags, i, tag); i++)
1547 rewrite_alias_set_for (tag, new_aliases);
1548 bitmap_clear (new_aliases);
1551 BITMAP_FREE (new_aliases);
1553 if (dump_file)
1555 fprintf (dump_file, "\nMemory symbol references after partitioning:\n");
1556 dump_mp_info (dump_file, mp_info);
1559 done:
1560 /* Free allocated memory. */
1561 VEC_free (mem_sym_stats_t, heap, mp_info);
1562 VEC_free (tree, heap, tags);
1564 MAX_ALIASED_VOPS = prev_max_aliased_vops;
1566 timevar_pop (TV_MEMORY_PARTITIONING);
1570 /* Compute may-alias information for every variable referenced in function
1571 FNDECL.
1573 Alias analysis proceeds in 3 main phases:
1575 1- Points-to and escape analysis.
1577 This phase walks the use-def chains in the SSA web looking for three
1578 things:
1580 * Assignments of the form P_i = &VAR
1581 * Assignments of the form P_i = malloc()
1582 * Pointers and ADDR_EXPR that escape the current function.
1584 The concept of 'escaping' is the same one used in the Java world. When
1585 a pointer or an ADDR_EXPR escapes, it means that it has been exposed
1586 outside of the current function. So, assignment to global variables,
1587 function arguments and returning a pointer are all escape sites, as are
1588 conversions between pointers and integers.
1590 This is where we are currently limited. Since not everything is renamed
1591 into SSA, we lose track of escape properties when a pointer is stashed
1592 inside a field in a structure, for instance. In those cases, we are
1593 assuming that the pointer does escape.
1595 We use escape analysis to determine whether a variable is
1596 call-clobbered. Simply put, if an ADDR_EXPR escapes, then the variable
1597 is call-clobbered. If a pointer P_i escapes, then all the variables
1598 pointed-to by P_i (and its memory tag) also escape.
1600 2- Compute flow-sensitive aliases
1602 We have two classes of memory tags. Memory tags associated with the
1603 pointed-to data type of the pointers in the program. These tags are
1604 called "symbol memory tag" (SMT). The other class are those associated
1605 with SSA_NAMEs, called "name memory tag" (NMT). The basic idea is that
1606 when adding operands for an INDIRECT_REF *P_i, we will first check
1607 whether P_i has a name tag, if it does we use it, because that will have
1608 more precise aliasing information. Otherwise, we use the standard symbol
1609 tag.
1611 In this phase, we go through all the pointers we found in points-to
1612 analysis and create alias sets for the name memory tags associated with
1613 each pointer P_i. If P_i escapes, we mark call-clobbered the variables
1614 it points to and its tag.
1617 3- Compute flow-insensitive aliases
1619 This pass will compare the alias set of every symbol memory tag and
1620 every addressable variable found in the program. Given a symbol
1621 memory tag SMT and an addressable variable V. If the alias sets of
1622 SMT and V conflict (as computed by may_alias_p), then V is marked
1623 as an alias tag and added to the alias set of SMT.
1625 For instance, consider the following function:
1627 foo (int i)
1629 int *p, a, b;
1631 if (i > 10)
1632 p = &a;
1633 else
1634 p = &b;
1636 *p = 3;
1637 a = b + 2;
1638 return *p;
1641 After aliasing analysis has finished, the symbol memory tag for pointer
1642 'p' will have two aliases, namely variables 'a' and 'b'. Every time
1643 pointer 'p' is dereferenced, we want to mark the operation as a
1644 potential reference to 'a' and 'b'.
1646 foo (int i)
1648 int *p, a, b;
1650 if (i_2 > 10)
1651 p_4 = &a;
1652 else
1653 p_6 = &b;
1654 # p_1 = PHI <p_4(1), p_6(2)>;
1656 # a_7 = VDEF <a_3>;
1657 # b_8 = VDEF <b_5>;
1658 *p_1 = 3;
1660 # a_9 = VDEF <a_7>
1661 # VUSE <b_8>
1662 a_9 = b_8 + 2;
1664 # VUSE <a_9>;
1665 # VUSE <b_8>;
1666 return *p_1;
1669 In certain cases, the list of may aliases for a pointer may grow too
1670 large. This may cause an explosion in the number of virtual operands
1671 inserted in the code. Resulting in increased memory consumption and
1672 compilation time.
1674 When the number of virtual operands needed to represent aliased
1675 loads and stores grows too large (configurable with option --param
1676 max-aliased-vops and --param avg-aliased-vops), alias sets are
1677 grouped to avoid severe compile-time slow downs and memory
1678 consumption. See compute_memory_partitions. */
1680 unsigned int
1681 compute_may_aliases (void)
1683 struct alias_info *ai;
1685 timevar_push (TV_TREE_MAY_ALIAS);
1687 memset (&alias_stats, 0, sizeof (alias_stats));
1689 /* Initialize aliasing information. */
1690 ai = init_alias_info ();
1692 /* For each pointer P_i, determine the sets of variables that P_i may
1693 point-to. For every addressable variable V, determine whether the
1694 address of V escapes the current function, making V call-clobbered
1695 (i.e., whether &V is stored in a global variable or if its passed as a
1696 function call argument). */
1697 compute_points_to_sets (ai);
1699 /* Collect all pointers and addressable variables, compute alias sets,
1700 create memory tags for pointers and promote variables whose address is
1701 not needed anymore. */
1702 setup_pointers_and_addressables (ai);
1704 /* Compute type-based flow-insensitive aliasing for all the type
1705 memory tags. */
1706 compute_flow_insensitive_aliasing (ai);
1708 /* Compute flow-sensitive, points-to based aliasing for all the name
1709 memory tags. */
1710 compute_flow_sensitive_aliasing (ai);
1712 /* Compute call clobbering information. */
1713 compute_call_clobbered (ai);
1715 /* If the program makes no reference to global variables, but it
1716 contains a mixture of pure and non-pure functions, then we need
1717 to create use-def and def-def links between these functions to
1718 avoid invalid transformations on them. */
1719 maybe_create_global_var ();
1721 /* If the program contains ref-all pointers, finalize may-alias information
1722 for them. This pass needs to be run after call-clobbering information
1723 has been computed. */
1724 if (ai->ref_all_symbol_mem_tag)
1725 finalize_ref_all_pointers (ai);
1727 /* Compute memory partitions for every memory variable. */
1728 compute_memory_partitions ();
1730 /* Remove partitions with no symbols. Partitions may end up with an
1731 empty MPT_SYMBOLS set if a previous round of alias analysis
1732 needed to partition more symbols. Since we don't need those
1733 partitions anymore, remove them to free up the space. */
1735 tree mpt;
1736 unsigned i;
1737 VEC(tree,heap) *mpt_table;
1739 mpt_table = gimple_ssa_operands (cfun)->mpt_table;
1740 i = 0;
1741 while (i < VEC_length (tree, mpt_table))
1743 mpt = VEC_index (tree, mpt_table, i);
1744 if (MPT_SYMBOLS (mpt) == NULL)
1745 VEC_unordered_remove (tree, mpt_table, i);
1746 else
1747 i++;
1751 /* Populate all virtual operands and newly promoted register operands. */
1753 block_stmt_iterator bsi;
1754 basic_block bb;
1755 FOR_EACH_BB (bb)
1756 for (bsi = bsi_start (bb); !bsi_end_p (bsi); bsi_next (&bsi))
1757 update_stmt_if_modified (bsi_stmt (bsi));
1760 /* Debugging dumps. */
1761 if (dump_file)
1763 dump_mem_ref_stats (dump_file);
1764 dump_alias_info (dump_file);
1765 dump_points_to_info (dump_file);
1767 if (dump_flags & TDF_STATS)
1768 dump_alias_stats (dump_file);
1770 if (dump_flags & TDF_DETAILS)
1771 dump_referenced_vars (dump_file);
1774 /* Report strict aliasing violations. */
1775 strict_aliasing_warning_backend ();
1777 /* Deallocate memory used by aliasing data structures. */
1778 delete_alias_info (ai);
1780 if (need_ssa_update_p ())
1781 update_ssa (TODO_update_ssa);
1783 timevar_pop (TV_TREE_MAY_ALIAS);
1785 return 0;
1788 /* Data structure used to count the number of dereferences to PTR
1789 inside an expression. */
1790 struct count_ptr_d
1792 tree ptr;
1793 unsigned count;
1797 /* Helper for count_uses_and_derefs. Called by walk_tree to look for
1798 (ALIGN/MISALIGNED_)INDIRECT_REF nodes for the pointer passed in DATA. */
1800 static tree
1801 count_ptr_derefs (tree *tp, int *walk_subtrees, void *data)
1803 struct count_ptr_d *count_p = (struct count_ptr_d *) data;
1805 /* Do not walk inside ADDR_EXPR nodes. In the expression &ptr->fld,
1806 pointer 'ptr' is *not* dereferenced, it is simply used to compute
1807 the address of 'fld' as 'ptr + offsetof(fld)'. */
1808 if (TREE_CODE (*tp) == ADDR_EXPR)
1810 *walk_subtrees = 0;
1811 return NULL_TREE;
1814 if (INDIRECT_REF_P (*tp) && TREE_OPERAND (*tp, 0) == count_p->ptr)
1815 count_p->count++;
1817 return NULL_TREE;
1821 /* Count the number of direct and indirect uses for pointer PTR in
1822 statement STMT. The number of direct uses is stored in
1823 *NUM_USES_P. Indirect references are counted separately depending
1824 on whether they are store or load operations. The counts are
1825 stored in *NUM_STORES_P and *NUM_LOADS_P. */
1827 void
1828 count_uses_and_derefs (tree ptr, tree stmt, unsigned *num_uses_p,
1829 unsigned *num_loads_p, unsigned *num_stores_p)
1831 ssa_op_iter i;
1832 tree use;
1834 *num_uses_p = 0;
1835 *num_loads_p = 0;
1836 *num_stores_p = 0;
1838 /* Find out the total number of uses of PTR in STMT. */
1839 FOR_EACH_SSA_TREE_OPERAND (use, stmt, i, SSA_OP_USE)
1840 if (use == ptr)
1841 (*num_uses_p)++;
1843 /* Now count the number of indirect references to PTR. This is
1844 truly awful, but we don't have much choice. There are no parent
1845 pointers inside INDIRECT_REFs, so an expression like
1846 '*x_1 = foo (x_1, *x_1)' needs to be traversed piece by piece to
1847 find all the indirect and direct uses of x_1 inside. The only
1848 shortcut we can take is the fact that GIMPLE only allows
1849 INDIRECT_REFs inside the expressions below. */
1850 if (TREE_CODE (stmt) == GIMPLE_MODIFY_STMT
1851 || (TREE_CODE (stmt) == RETURN_EXPR
1852 && TREE_CODE (TREE_OPERAND (stmt, 0)) == GIMPLE_MODIFY_STMT)
1853 || TREE_CODE (stmt) == ASM_EXPR
1854 || TREE_CODE (stmt) == CALL_EXPR)
1856 tree lhs, rhs;
1858 if (TREE_CODE (stmt) == GIMPLE_MODIFY_STMT)
1860 lhs = GIMPLE_STMT_OPERAND (stmt, 0);
1861 rhs = GIMPLE_STMT_OPERAND (stmt, 1);
1863 else if (TREE_CODE (stmt) == RETURN_EXPR)
1865 tree e = TREE_OPERAND (stmt, 0);
1866 lhs = GIMPLE_STMT_OPERAND (e, 0);
1867 rhs = GIMPLE_STMT_OPERAND (e, 1);
1869 else if (TREE_CODE (stmt) == ASM_EXPR)
1871 lhs = ASM_OUTPUTS (stmt);
1872 rhs = ASM_INPUTS (stmt);
1874 else
1876 lhs = NULL_TREE;
1877 rhs = stmt;
1880 if (lhs
1881 && (TREE_CODE (lhs) == TREE_LIST
1882 || EXPR_P (lhs)
1883 || GIMPLE_STMT_P (lhs)))
1885 struct count_ptr_d count;
1886 count.ptr = ptr;
1887 count.count = 0;
1888 walk_tree (&lhs, count_ptr_derefs, &count, NULL);
1889 *num_stores_p = count.count;
1892 if (rhs
1893 && (TREE_CODE (rhs) == TREE_LIST
1894 || EXPR_P (rhs)
1895 || GIMPLE_STMT_P (rhs)))
1897 struct count_ptr_d count;
1898 count.ptr = ptr;
1899 count.count = 0;
1900 walk_tree (&rhs, count_ptr_derefs, &count, NULL);
1901 *num_loads_p = count.count;
1905 gcc_assert (*num_uses_p >= *num_loads_p + *num_stores_p);
1908 /* Remove memory references stats for function FN. */
1910 void
1911 delete_mem_ref_stats (struct function *fn)
1913 if (gimple_mem_ref_stats (fn)->mem_sym_stats)
1915 free_alloc_pool (mem_sym_stats_pool);
1916 pointer_map_destroy (gimple_mem_ref_stats (fn)->mem_sym_stats);
1918 gimple_mem_ref_stats (fn)->mem_sym_stats = NULL;
1922 /* Initialize memory reference stats. */
1924 static void
1925 init_mem_ref_stats (void)
1927 struct mem_ref_stats_d *mem_ref_stats = gimple_mem_ref_stats (cfun);
1929 mem_sym_stats_pool = create_alloc_pool ("Mem sym stats",
1930 sizeof (struct mem_sym_stats_d),
1931 100);
1932 memset (mem_ref_stats, 0, sizeof (struct mem_ref_stats_d));
1933 mem_ref_stats->mem_sym_stats = pointer_map_create ();
1937 /* Helper for init_alias_info. Reset existing aliasing information. */
1939 static void
1940 reset_alias_info (void)
1942 referenced_var_iterator rvi;
1943 tree var;
1944 unsigned i;
1945 bitmap active_nmts, all_nmts;
1947 /* Clear the set of addressable variables. We do not need to clear
1948 the TREE_ADDRESSABLE bit on every symbol because we are going to
1949 re-compute addressability here. */
1950 bitmap_clear (gimple_addressable_vars (cfun));
1952 active_nmts = BITMAP_ALLOC (&alias_bitmap_obstack);
1953 all_nmts = BITMAP_ALLOC (&alias_bitmap_obstack);
1955 /* Clear flow-insensitive alias information from each symbol. */
1956 FOR_EACH_REFERENCED_VAR (var, rvi)
1958 if (is_gimple_reg (var))
1959 continue;
1961 if (MTAG_P (var))
1962 MTAG_ALIASES (var) = NULL;
1964 /* Memory partition information will be computed from scratch. */
1965 if (TREE_CODE (var) == MEMORY_PARTITION_TAG)
1966 MPT_SYMBOLS (var) = NULL;
1968 /* Collect all the name tags to determine if we have any
1969 orphaned that need to be removed from the IL. A name tag
1970 will be orphaned if it is not associated with any active SSA
1971 name. */
1972 if (TREE_CODE (var) == NAME_MEMORY_TAG)
1973 bitmap_set_bit (all_nmts, DECL_UID (var));
1975 /* Since we are about to re-discover call-clobbered
1976 variables, clear the call-clobbered flag. Variables that
1977 are intrinsically call-clobbered (globals, local statics,
1978 etc) will not be marked by the aliasing code, so we can't
1979 remove them from CALL_CLOBBERED_VARS.
1981 NB: STRUCT_FIELDS are still call clobbered if they are for a
1982 global variable, so we *don't* clear their call clobberedness
1983 just because they are tags, though we will clear it if they
1984 aren't for global variables. */
1985 if (TREE_CODE (var) == NAME_MEMORY_TAG
1986 || TREE_CODE (var) == SYMBOL_MEMORY_TAG
1987 || TREE_CODE (var) == MEMORY_PARTITION_TAG
1988 || !is_global_var (var))
1989 clear_call_clobbered (var);
1992 /* Clear flow-sensitive points-to information from each SSA name. */
1993 for (i = 1; i < num_ssa_names; i++)
1995 tree name = ssa_name (i);
1997 if (!name || !POINTER_TYPE_P (TREE_TYPE (name)))
1998 continue;
2000 if (SSA_NAME_PTR_INFO (name))
2002 struct ptr_info_def *pi = SSA_NAME_PTR_INFO (name);
2004 /* Clear all the flags but keep the name tag to
2005 avoid creating new temporaries unnecessarily. If
2006 this pointer is found to point to a subset or
2007 superset of its former points-to set, then a new
2008 tag will need to be created in create_name_tags. */
2009 pi->pt_anything = 0;
2010 pi->pt_null = 0;
2011 pi->value_escapes_p = 0;
2012 pi->is_dereferenced = 0;
2013 if (pi->pt_vars)
2014 bitmap_clear (pi->pt_vars);
2016 /* Add NAME's name tag to the set of active tags. */
2017 if (pi->name_mem_tag)
2018 bitmap_set_bit (active_nmts, DECL_UID (pi->name_mem_tag));
2022 /* Name memory tags that are no longer associated with an SSA name
2023 are considered stale and should be removed from the IL. All the
2024 name tags that are in the set ALL_NMTS but not in ACTIVE_NMTS are
2025 considered stale and marked for renaming. */
2026 bitmap_and_compl_into (all_nmts, active_nmts);
2027 mark_set_for_renaming (all_nmts);
2029 BITMAP_FREE (all_nmts);
2030 BITMAP_FREE (active_nmts);
2034 /* Initialize the data structures used for alias analysis. */
2036 static struct alias_info *
2037 init_alias_info (void)
2039 struct alias_info *ai;
2040 referenced_var_iterator rvi;
2041 tree var;
2043 ai = XCNEW (struct alias_info);
2044 ai->ssa_names_visited = sbitmap_alloc (num_ssa_names);
2045 sbitmap_zero (ai->ssa_names_visited);
2046 ai->processed_ptrs = VEC_alloc (tree, heap, 50);
2047 ai->written_vars = pointer_set_create ();
2048 ai->dereferenced_ptrs_store = pointer_set_create ();
2049 ai->dereferenced_ptrs_load = pointer_set_create ();
2051 /* Clear out all memory reference stats. */
2052 init_mem_ref_stats ();
2054 /* If aliases have been computed before, clear existing information. */
2055 if (gimple_aliases_computed_p (cfun))
2056 reset_alias_info ();
2057 else
2059 /* If this is the first time we compute aliasing information,
2060 every non-register symbol will need to be put into SSA form
2061 (the initial SSA form only operates on GIMPLE registers). */
2062 FOR_EACH_REFERENCED_VAR (var, rvi)
2063 if (!is_gimple_reg (var))
2064 mark_sym_for_renaming (var);
2067 /* Next time, we will need to reset alias information. */
2068 cfun->gimple_df->aliases_computed_p = true;
2069 if (alias_bitmap_obstack.elements != NULL)
2070 bitmap_obstack_release (&alias_bitmap_obstack);
2071 bitmap_obstack_initialize (&alias_bitmap_obstack);
2073 return ai;
2077 /* Deallocate memory used by alias analysis. */
2079 static void
2080 delete_alias_info (struct alias_info *ai)
2082 size_t i;
2084 sbitmap_free (ai->ssa_names_visited);
2086 VEC_free (tree, heap, ai->processed_ptrs);
2088 for (i = 0; i < ai->num_addressable_vars; i++)
2089 free (ai->addressable_vars[i]);
2090 free (ai->addressable_vars);
2092 for (i = 0; i < ai->num_pointers; i++)
2093 free (ai->pointers[i]);
2094 free (ai->pointers);
2096 pointer_set_destroy (ai->written_vars);
2097 pointer_set_destroy (ai->dereferenced_ptrs_store);
2098 pointer_set_destroy (ai->dereferenced_ptrs_load);
2099 free (ai);
2101 delete_mem_ref_stats (cfun);
2102 delete_points_to_sets ();
2106 /* Used for hashing to identify pointer infos with identical
2107 pt_vars bitmaps. */
2109 static int
2110 eq_ptr_info (const void *p1, const void *p2)
2112 const struct ptr_info_def *n1 = (const struct ptr_info_def *) p1;
2113 const struct ptr_info_def *n2 = (const struct ptr_info_def *) p2;
2114 return bitmap_equal_p (n1->pt_vars, n2->pt_vars);
2117 static hashval_t
2118 ptr_info_hash (const void *p)
2120 const struct ptr_info_def *n = (const struct ptr_info_def *) p;
2121 return bitmap_hash (n->pt_vars);
2125 /* Create name tags for all the pointers that have been dereferenced.
2126 We only create a name tag for a pointer P if P is found to point to
2127 a set of variables (so that we can alias them to *P) or if it is
2128 the result of a call to malloc (which means that P cannot point to
2129 anything else nor alias any other variable).
2131 If two pointers P and Q point to the same set of variables, they
2132 are assigned the same name tag. */
2134 static void
2135 create_name_tags (void)
2137 size_t i;
2138 VEC (tree, heap) *with_ptvars = NULL;
2139 tree ptr;
2140 htab_t ptr_hash;
2142 /* Collect the list of pointers with a non-empty points to set. */
2143 for (i = 1; i < num_ssa_names; i++)
2145 tree ptr = ssa_name (i);
2146 struct ptr_info_def *pi;
2148 if (!ptr
2149 || !POINTER_TYPE_P (TREE_TYPE (ptr))
2150 || !SSA_NAME_PTR_INFO (ptr))
2151 continue;
2153 pi = SSA_NAME_PTR_INFO (ptr);
2155 if (pi->pt_anything || !pi->is_dereferenced)
2157 /* No name tags for pointers that have not been
2158 dereferenced or point to an arbitrary location. */
2159 pi->name_mem_tag = NULL_TREE;
2160 continue;
2163 /* Set pt_anything on the pointers without pt_vars filled in so
2164 that they are assigned a symbol tag. */
2165 if (pi->pt_vars && !bitmap_empty_p (pi->pt_vars))
2166 VEC_safe_push (tree, heap, with_ptvars, ptr);
2167 else
2168 set_pt_anything (ptr);
2171 /* If we didn't find any pointers with pt_vars set, we're done. */
2172 if (!with_ptvars)
2173 return;
2175 ptr_hash = htab_create (10, ptr_info_hash, eq_ptr_info, NULL);
2177 /* Now go through the pointers with pt_vars, and find a name tag
2178 with the same pt_vars as this pointer, or create one if one
2179 doesn't exist. */
2180 for (i = 0; VEC_iterate (tree, with_ptvars, i, ptr); i++)
2182 struct ptr_info_def *pi = SSA_NAME_PTR_INFO (ptr);
2183 tree old_name_tag = pi->name_mem_tag;
2184 struct ptr_info_def **slot;
2186 /* If PTR points to a set of variables, check if we don't
2187 have another pointer Q with the same points-to set before
2188 creating a tag. If so, use Q's tag instead of creating a
2189 new one.
2191 This is important for not creating unnecessary symbols
2192 and also for copy propagation. If we ever need to
2193 propagate PTR into Q or vice-versa, we would run into
2194 problems if they both had different name tags because
2195 they would have different SSA version numbers (which
2196 would force us to take the name tags in and out of SSA). */
2197 slot = (struct ptr_info_def **) htab_find_slot (ptr_hash, pi, INSERT);
2198 if (*slot)
2199 pi->name_mem_tag = (*slot)->name_mem_tag;
2200 else
2202 *slot = pi;
2204 /* If we didn't find a pointer with the same points-to set
2205 as PTR, create a new name tag if needed. */
2206 if (pi->name_mem_tag == NULL_TREE)
2207 pi->name_mem_tag = get_nmt_for (ptr);
2210 /* If the new name tag computed for PTR is different than
2211 the old name tag that it used to have, then the old tag
2212 needs to be removed from the IL, so we mark it for
2213 renaming. */
2214 if (old_name_tag && old_name_tag != pi->name_mem_tag)
2215 mark_sym_for_renaming (old_name_tag);
2217 /* Inherit volatility from the pointed-to type. */
2218 TREE_THIS_VOLATILE (pi->name_mem_tag)
2219 |= TREE_THIS_VOLATILE (TREE_TYPE (TREE_TYPE (ptr)));
2221 /* Mark the new name tag for renaming. */
2222 mark_sym_for_renaming (pi->name_mem_tag);
2225 htab_delete (ptr_hash);
2227 VEC_free (tree, heap, with_ptvars);
2231 /* Union the alias set SET into the may-aliases for TAG. */
2233 static void
2234 union_alias_set_into (tree tag, bitmap set)
2236 bitmap ma = MTAG_ALIASES (tag);
2238 if (bitmap_empty_p (set))
2239 return;
2241 if (!ma)
2242 ma = MTAG_ALIASES (tag) = BITMAP_ALLOC (&alias_bitmap_obstack);
2243 bitmap_ior_into (ma, set);
2247 /* For every pointer P_i in AI->PROCESSED_PTRS, create may-alias sets for
2248 the name memory tag (NMT) associated with P_i. If P_i escapes, then its
2249 name tag and the variables it points-to are call-clobbered. Finally, if
2250 P_i escapes and we could not determine where it points to, then all the
2251 variables in the same alias set as *P_i are marked call-clobbered. This
2252 is necessary because we must assume that P_i may take the address of any
2253 variable in the same alias set. */
2255 static void
2256 compute_flow_sensitive_aliasing (struct alias_info *ai)
2258 size_t i;
2259 tree ptr;
2261 timevar_push (TV_FLOW_SENSITIVE);
2262 set_used_smts ();
2264 for (i = 0; VEC_iterate (tree, ai->processed_ptrs, i, ptr); i++)
2266 if (!find_what_p_points_to (ptr))
2267 set_pt_anything (ptr);
2270 create_name_tags ();
2272 for (i = 0; VEC_iterate (tree, ai->processed_ptrs, i, ptr); i++)
2274 struct ptr_info_def *pi = SSA_NAME_PTR_INFO (ptr);
2275 tree tag = symbol_mem_tag (SSA_NAME_VAR (ptr));
2277 /* Set up aliasing information for PTR's name memory tag (if it has
2278 one). Note that only pointers that have been dereferenced will
2279 have a name memory tag. */
2280 if (pi->name_mem_tag && pi->pt_vars)
2282 if (!bitmap_empty_p (pi->pt_vars))
2284 union_alias_set_into (pi->name_mem_tag, pi->pt_vars);
2285 union_alias_set_into (tag, pi->pt_vars);
2286 bitmap_clear_bit (MTAG_ALIASES (tag), DECL_UID (tag));
2288 /* It may be the case that this the tag uid was the only
2289 bit we had set in the aliases list, and in this case,
2290 we don't want to keep an empty bitmap, as this
2291 asserts in tree-ssa-operands.c . */
2292 if (bitmap_empty_p (MTAG_ALIASES (tag)))
2293 BITMAP_FREE (MTAG_ALIASES (tag));
2297 timevar_pop (TV_FLOW_SENSITIVE);
2301 /* Return TRUE if at least one symbol in TAG2's alias set is also
2302 present in TAG1's alias set. */
2304 static bool
2305 have_common_aliases_p (bitmap tag1aliases, bitmap tag2aliases)
2308 /* This is the old behavior of have_common_aliases_p, which is to
2309 return false if both sets are empty, or one set is and the other
2310 isn't. */
2311 if ((tag1aliases == NULL && tag2aliases != NULL)
2312 || (tag2aliases == NULL && tag1aliases != NULL)
2313 || (tag1aliases == NULL && tag2aliases == NULL))
2314 return false;
2316 return bitmap_intersect_p (tag1aliases, tag2aliases);
2319 /* Compute type-based alias sets. Traverse all the pointers and
2320 addressable variables found in setup_pointers_and_addressables.
2322 For every pointer P in AI->POINTERS and addressable variable V in
2323 AI->ADDRESSABLE_VARS, add V to the may-alias sets of P's symbol
2324 memory tag (SMT) if their alias sets conflict. V is then marked as
2325 an aliased symbol so that the operand scanner knows that statements
2326 containing V have aliased operands. */
2328 static void
2329 compute_flow_insensitive_aliasing (struct alias_info *ai)
2331 size_t i;
2333 timevar_push (TV_FLOW_INSENSITIVE);
2334 /* For every pointer P, determine which addressable variables may alias
2335 with P's symbol memory tag. */
2336 for (i = 0; i < ai->num_pointers; i++)
2338 size_t j;
2339 struct alias_map_d *p_map = ai->pointers[i];
2340 tree tag = symbol_mem_tag (p_map->var);
2341 tree var;
2343 /* Call-clobbering information is not finalized yet at this point. */
2344 if (PTR_IS_REF_ALL (p_map->var))
2345 continue;
2347 for (j = 0; j < ai->num_addressable_vars; j++)
2349 struct alias_map_d *v_map;
2350 var_ann_t v_ann;
2351 bool tag_stored_p, var_stored_p;
2353 v_map = ai->addressable_vars[j];
2354 var = v_map->var;
2355 v_ann = var_ann (var);
2357 /* Skip memory tags and variables that have never been
2358 written to. We also need to check if the variables are
2359 call-clobbered because they may be overwritten by
2360 function calls. */
2361 tag_stored_p = pointer_set_contains (ai->written_vars, tag)
2362 || is_call_clobbered (tag);
2363 var_stored_p = pointer_set_contains (ai->written_vars, var)
2364 || is_call_clobbered (var);
2365 if (!tag_stored_p && !var_stored_p)
2366 continue;
2368 if (may_alias_p (p_map->var, p_map->set, var, v_map->set, false))
2370 /* We should never have a var with subvars here, because
2371 they shouldn't get into the set of addressable vars */
2372 gcc_assert (!var_can_have_subvars (var)
2373 || get_subvars_for_var (var) == NULL);
2375 /* Add VAR to TAG's may-aliases set. */
2376 add_may_alias (tag, var);
2381 /* Since this analysis is based exclusively on symbols, it fails to
2382 handle cases where two pointers P and Q have different memory
2383 tags with conflicting alias set numbers but no aliased symbols in
2384 common.
2386 For example, suppose that we have two memory tags SMT.1 and SMT.2
2387 such that
2389 may-aliases (SMT.1) = { a }
2390 may-aliases (SMT.2) = { b }
2392 and the alias set number of SMT.1 conflicts with that of SMT.2.
2393 Since they don't have symbols in common, loads and stores from
2394 SMT.1 and SMT.2 will seem independent of each other, which will
2395 lead to the optimizers making invalid transformations (see
2396 testsuite/gcc.c-torture/execute/pr15262-[12].c).
2398 To avoid this problem, we do a final traversal of AI->POINTERS
2399 looking for pairs of pointers that have no aliased symbols in
2400 common and yet have conflicting alias set numbers. */
2401 for (i = 0; i < ai->num_pointers; i++)
2403 size_t j;
2404 struct alias_map_d *p_map1 = ai->pointers[i];
2405 tree tag1 = symbol_mem_tag (p_map1->var);
2406 bitmap may_aliases1 = MTAG_ALIASES (tag1);
2408 if (PTR_IS_REF_ALL (p_map1->var))
2409 continue;
2411 for (j = i + 1; j < ai->num_pointers; j++)
2413 struct alias_map_d *p_map2 = ai->pointers[j];
2414 tree tag2 = symbol_mem_tag (p_map2->var);
2415 bitmap may_aliases2 = may_aliases (tag2);
2417 if (PTR_IS_REF_ALL (p_map2->var))
2418 continue;
2420 /* If the pointers may not point to each other, do nothing. */
2421 if (!may_alias_p (p_map1->var, p_map1->set, tag2, p_map2->set, true))
2422 continue;
2424 /* The two pointers may alias each other. If they already have
2425 symbols in common, do nothing. */
2426 if (have_common_aliases_p (may_aliases1, may_aliases2))
2427 continue;
2429 if (may_aliases2 && !bitmap_empty_p (may_aliases2))
2431 union_alias_set_into (tag1, may_aliases2);
2433 else
2435 /* Since TAG2 does not have any aliases of its own, add
2436 TAG2 itself to the alias set of TAG1. */
2437 add_may_alias (tag1, tag2);
2442 timevar_pop (TV_FLOW_INSENSITIVE);
2446 /* Finalize may-alias information for ref-all pointers. Traverse all
2447 the addressable variables found in setup_pointers_and_addressables.
2449 If flow-sensitive alias analysis has attached a name memory tag to
2450 a ref-all pointer, we will use it for the dereferences because that
2451 will have more precise aliasing information. But if there is no
2452 name tag, we will use a special symbol tag that aliases all the
2453 call-clobbered addressable variables. */
2455 static void
2456 finalize_ref_all_pointers (struct alias_info *ai)
2458 size_t i;
2460 /* First add the real call-clobbered variables. */
2461 for (i = 0; i < ai->num_addressable_vars; i++)
2463 tree var = ai->addressable_vars[i]->var;
2464 if (is_call_clobbered (var))
2465 add_may_alias (ai->ref_all_symbol_mem_tag, var);
2468 /* Then add the call-clobbered pointer memory tags. See
2469 compute_flow_insensitive_aliasing for the rationale. */
2470 for (i = 0; i < ai->num_pointers; i++)
2472 tree ptr = ai->pointers[i]->var, tag;
2473 /* Avoid adding to self and clean up. */
2474 if (PTR_IS_REF_ALL (ptr))
2476 struct ptr_info_def *pi = get_ptr_info (ptr);
2477 if (pi->is_dereferenced)
2478 pi->pt_anything = 0;
2479 continue;
2481 tag = symbol_mem_tag (ptr);
2482 if (is_call_clobbered (tag))
2483 add_may_alias (ai->ref_all_symbol_mem_tag, tag);
2489 /* Create a new alias set entry for VAR in AI->ADDRESSABLE_VARS. */
2491 static void
2492 create_alias_map_for (tree var, struct alias_info *ai)
2494 struct alias_map_d *alias_map;
2495 alias_map = XCNEW (struct alias_map_d);
2496 alias_map->var = var;
2497 alias_map->set = get_alias_set (var);
2498 ai->addressable_vars[ai->num_addressable_vars++] = alias_map;
2502 /* Create memory tags for all the dereferenced pointers and build the
2503 ADDRESSABLE_VARS and POINTERS arrays used for building the may-alias
2504 sets. Based on the address escape and points-to information collected
2505 earlier, this pass will also clear the TREE_ADDRESSABLE flag from those
2506 variables whose address is not needed anymore. */
2508 static void
2509 setup_pointers_and_addressables (struct alias_info *ai)
2511 size_t num_addressable_vars, num_pointers;
2512 referenced_var_iterator rvi;
2513 tree var;
2514 VEC (tree, heap) *varvec = NULL;
2515 safe_referenced_var_iterator srvi;
2517 /* Size up the arrays ADDRESSABLE_VARS and POINTERS. */
2518 num_addressable_vars = num_pointers = 0;
2520 FOR_EACH_REFERENCED_VAR (var, rvi)
2522 if (may_be_aliased (var))
2523 num_addressable_vars++;
2525 if (POINTER_TYPE_P (TREE_TYPE (var)))
2527 /* Since we don't keep track of volatile variables, assume that
2528 these pointers are used in indirect store operations. */
2529 if (TREE_THIS_VOLATILE (var))
2530 pointer_set_insert (ai->dereferenced_ptrs_store, var);
2532 num_pointers++;
2536 /* Create ADDRESSABLE_VARS and POINTERS. Note that these arrays are
2537 always going to be slightly bigger than we actually need them
2538 because some TREE_ADDRESSABLE variables will be marked
2539 non-addressable below and only pointers with unique symbol tags are
2540 going to be added to POINTERS. */
2541 ai->addressable_vars = XCNEWVEC (struct alias_map_d *, num_addressable_vars);
2542 ai->pointers = XCNEWVEC (struct alias_map_d *, num_pointers);
2543 ai->num_addressable_vars = 0;
2544 ai->num_pointers = 0;
2546 FOR_EACH_REFERENCED_VAR_SAFE (var, varvec, srvi)
2548 subvar_t svars;
2550 /* Name memory tags already have flow-sensitive aliasing
2551 information, so they need not be processed by
2552 compute_flow_insensitive_aliasing. Similarly, symbol memory
2553 tags are already accounted for when we process their
2554 associated pointer.
2556 Structure fields, on the other hand, have to have some of this
2557 information processed for them, but it's pointless to mark them
2558 non-addressable (since they are fake variables anyway). */
2559 if (MTAG_P (var) && TREE_CODE (var) != STRUCT_FIELD_TAG)
2560 continue;
2562 /* Remove the ADDRESSABLE flag from every addressable variable whose
2563 address is not needed anymore. This is caused by the propagation
2564 of ADDR_EXPR constants into INDIRECT_REF expressions and the
2565 removal of dead pointer assignments done by the early scalar
2566 cleanup passes. */
2567 if (TREE_ADDRESSABLE (var))
2569 if (!bitmap_bit_p (gimple_addressable_vars (cfun), DECL_UID (var))
2570 && TREE_CODE (var) != RESULT_DECL
2571 && !is_global_var (var))
2573 bool okay_to_mark = true;
2575 /* Since VAR is now a regular GIMPLE register, we will need
2576 to rename VAR into SSA afterwards. */
2577 mark_sym_for_renaming (var);
2579 /* If VAR can have sub-variables, and any of its
2580 sub-variables has its address taken, then we cannot
2581 remove the addressable flag from VAR. */
2582 if (var_can_have_subvars (var)
2583 && (svars = get_subvars_for_var (var)))
2585 subvar_t sv;
2587 for (sv = svars; sv; sv = sv->next)
2589 if (bitmap_bit_p (gimple_addressable_vars (cfun),
2590 DECL_UID (sv->var)))
2591 okay_to_mark = false;
2592 mark_sym_for_renaming (sv->var);
2596 /* The address of VAR is not needed, remove the
2597 addressable bit, so that it can be optimized as a
2598 regular variable. */
2599 if (okay_to_mark)
2601 /* The memory partition holding VAR will no longer
2602 contain VAR, and statements referencing it will need
2603 to be updated. */
2604 if (memory_partition (var))
2605 mark_sym_for_renaming (memory_partition (var));
2607 mark_non_addressable (var);
2612 /* Global variables and addressable locals may be aliased. Create an
2613 entry in ADDRESSABLE_VARS for VAR. */
2614 if (may_be_aliased (var))
2616 if (!var_can_have_subvars (var)
2617 || get_subvars_for_var (var) == NULL)
2618 create_alias_map_for (var, ai);
2620 mark_sym_for_renaming (var);
2623 /* Add pointer variables that have been dereferenced to the POINTERS
2624 array and create a symbol memory tag for them. */
2625 if (POINTER_TYPE_P (TREE_TYPE (var)))
2627 if ((pointer_set_contains (ai->dereferenced_ptrs_store, var)
2628 || pointer_set_contains (ai->dereferenced_ptrs_load, var)))
2630 tree tag, old_tag;
2631 var_ann_t t_ann;
2633 /* If pointer VAR still doesn't have a memory tag
2634 associated with it, create it now or re-use an
2635 existing one. */
2636 tag = get_smt_for (var, ai);
2637 t_ann = var_ann (tag);
2639 /* The symbol tag will need to be renamed into SSA
2640 afterwards. Note that we cannot do this inside
2641 get_smt_for because aliasing may run multiple times
2642 and we only create symbol tags the first time. */
2643 mark_sym_for_renaming (tag);
2645 /* Similarly, if pointer VAR used to have another type
2646 tag, we will need to process it in the renamer to
2647 remove the stale virtual operands. */
2648 old_tag = symbol_mem_tag (var);
2649 if (old_tag)
2650 mark_sym_for_renaming (old_tag);
2652 /* Associate the tag with pointer VAR. */
2653 set_symbol_mem_tag (var, tag);
2655 /* If pointer VAR has been used in a store operation,
2656 then its memory tag must be marked as written-to. */
2657 if (pointer_set_contains (ai->dereferenced_ptrs_store, var))
2658 pointer_set_insert (ai->written_vars, tag);
2660 else
2662 /* The pointer has not been dereferenced. If it had a
2663 symbol memory tag, remove it and mark the old tag for
2664 renaming to remove it out of the IL. */
2665 tree tag = symbol_mem_tag (var);
2666 if (tag)
2668 mark_sym_for_renaming (tag);
2669 set_symbol_mem_tag (var, NULL_TREE);
2675 VEC_free (tree, heap, varvec);
2679 /* Determine whether to use .GLOBAL_VAR to model call clobbering
2680 semantics. If the function makes no references to global
2681 variables and contains at least one call to a non-pure function,
2682 then we need to mark the side-effects of the call using .GLOBAL_VAR
2683 to represent all possible global memory referenced by the callee. */
2685 static void
2686 maybe_create_global_var (void)
2688 /* No need to create it, if we have one already. */
2689 if (gimple_global_var (cfun) == NULL_TREE)
2691 struct mem_ref_stats_d *stats = gimple_mem_ref_stats (cfun);
2693 /* Create .GLOBAL_VAR if there are no call-clobbered
2694 variables and the program contains a mixture of pure/const
2695 and regular function calls. This is to avoid the problem
2696 described in PR 20115:
2698 int X;
2699 int func_pure (void) { return X; }
2700 int func_non_pure (int a) { X += a; }
2701 int foo ()
2703 int a = func_pure ();
2704 func_non_pure (a);
2705 a = func_pure ();
2706 return a;
2709 Since foo() has no call-clobbered variables, there is
2710 no relationship between the calls to func_pure and
2711 func_non_pure. Since func_pure has no side-effects, value
2712 numbering optimizations elide the second call to func_pure.
2713 So, if we have some pure/const and some regular calls in the
2714 program we create .GLOBAL_VAR to avoid missing these
2715 relations. */
2716 if (bitmap_count_bits (gimple_call_clobbered_vars (cfun)) == 0
2717 && stats->num_call_sites > 0
2718 && stats->num_pure_const_call_sites > 0
2719 && stats->num_call_sites > stats->num_pure_const_call_sites)
2720 create_global_var ();
2725 /* Return TRUE if pointer PTR may point to variable VAR.
2727 MEM_ALIAS_SET is the alias set for the memory location pointed-to by PTR
2728 This is needed because when checking for type conflicts we are
2729 interested in the alias set of the memory location pointed-to by
2730 PTR. The alias set of PTR itself is irrelevant.
2732 VAR_ALIAS_SET is the alias set for VAR. */
2734 static bool
2735 may_alias_p (tree ptr, alias_set_type mem_alias_set,
2736 tree var, alias_set_type var_alias_set,
2737 bool alias_set_only)
2739 tree mem;
2741 alias_stats.alias_queries++;
2742 alias_stats.simple_queries++;
2744 /* By convention, a variable cannot alias itself. */
2745 mem = symbol_mem_tag (ptr);
2746 if (mem == var)
2748 alias_stats.alias_noalias++;
2749 alias_stats.simple_resolved++;
2750 return false;
2753 /* If -fargument-noalias-global is > 2, pointer arguments may
2754 not point to anything else. */
2755 if (flag_argument_noalias > 2 && TREE_CODE (ptr) == PARM_DECL)
2757 alias_stats.alias_noalias++;
2758 alias_stats.simple_resolved++;
2759 return false;
2762 /* If -fargument-noalias-global is > 1, pointer arguments may
2763 not point to global variables. */
2764 if (flag_argument_noalias > 1 && is_global_var (var)
2765 && TREE_CODE (ptr) == PARM_DECL)
2767 alias_stats.alias_noalias++;
2768 alias_stats.simple_resolved++;
2769 return false;
2772 /* If either MEM or VAR is a read-only global and the other one
2773 isn't, then PTR cannot point to VAR. */
2774 if ((unmodifiable_var_p (mem) && !unmodifiable_var_p (var))
2775 || (unmodifiable_var_p (var) && !unmodifiable_var_p (mem)))
2777 alias_stats.alias_noalias++;
2778 alias_stats.simple_resolved++;
2779 return false;
2782 gcc_assert (TREE_CODE (mem) == SYMBOL_MEMORY_TAG);
2784 if (!DECL_NO_TBAA_P (ptr))
2786 alias_stats.tbaa_queries++;
2788 /* If the alias sets don't conflict then MEM cannot alias VAR. */
2789 if (!alias_sets_conflict_p (mem_alias_set, var_alias_set))
2791 alias_stats.alias_noalias++;
2792 alias_stats.tbaa_resolved++;
2793 return false;
2796 /* If VAR is a record or union type, PTR cannot point into VAR
2797 unless there is some explicit address operation in the
2798 program that can reference a field of the type pointed-to by
2799 PTR. This also assumes that the types of both VAR and PTR
2800 are contained within the compilation unit, and that there is
2801 no fancy addressing arithmetic associated with any of the
2802 types involved. */
2803 if (mem_alias_set != 0 && var_alias_set != 0)
2805 tree ptr_type = TREE_TYPE (ptr);
2806 tree var_type = TREE_TYPE (var);
2808 /* The star count is -1 if the type at the end of the
2809 pointer_to chain is not a record or union type. */
2810 if ((!alias_set_only) &&
2811 ipa_type_escape_star_count_of_interesting_type (var_type) >= 0)
2813 int ptr_star_count = 0;
2815 /* ipa_type_escape_star_count_of_interesting_type is a
2816 little too restrictive for the pointer type, need to
2817 allow pointers to primitive types as long as those
2818 types cannot be pointers to everything. */
2819 while (POINTER_TYPE_P (ptr_type))
2821 /* Strip the *s off. */
2822 ptr_type = TREE_TYPE (ptr_type);
2823 ptr_star_count++;
2826 /* There does not appear to be a better test to see if
2827 the pointer type was one of the pointer to everything
2828 types. */
2829 if (ptr_star_count > 0)
2831 alias_stats.structnoaddress_queries++;
2832 if (ipa_type_escape_field_does_not_clobber_p (var_type,
2833 TREE_TYPE (ptr)))
2835 alias_stats.structnoaddress_resolved++;
2836 alias_stats.alias_noalias++;
2837 return false;
2840 else if (ptr_star_count == 0)
2842 /* If PTR_TYPE was not really a pointer to type, it cannot
2843 alias. */
2844 alias_stats.structnoaddress_queries++;
2845 alias_stats.structnoaddress_resolved++;
2846 alias_stats.alias_noalias++;
2847 return false;
2853 alias_stats.alias_mayalias++;
2854 return true;
2858 /* Add ALIAS to the set of variables that may alias VAR. */
2860 static void
2861 add_may_alias (tree var, tree alias)
2863 /* Don't allow self-referential aliases. */
2864 gcc_assert (var != alias);
2866 /* ALIAS must be addressable if it's being added to an alias set. */
2867 #if 1
2868 TREE_ADDRESSABLE (alias) = 1;
2869 #else
2870 gcc_assert (may_be_aliased (alias));
2871 #endif
2873 /* VAR must be a symbol or a name tag. */
2874 gcc_assert (TREE_CODE (var) == SYMBOL_MEMORY_TAG
2875 || TREE_CODE (var) == NAME_MEMORY_TAG);
2877 if (MTAG_ALIASES (var) == NULL)
2878 MTAG_ALIASES (var) = BITMAP_ALLOC (&alias_bitmap_obstack);
2880 bitmap_set_bit (MTAG_ALIASES (var), DECL_UID (alias));
2884 /* Mark pointer PTR as pointing to an arbitrary memory location. */
2886 static void
2887 set_pt_anything (tree ptr)
2889 struct ptr_info_def *pi = get_ptr_info (ptr);
2891 pi->pt_anything = 1;
2892 pi->pt_vars = NULL;
2894 /* The pointer used to have a name tag, but we now found it pointing
2895 to an arbitrary location. The name tag needs to be renamed and
2896 disassociated from PTR. */
2897 if (pi->name_mem_tag)
2899 mark_sym_for_renaming (pi->name_mem_tag);
2900 pi->name_mem_tag = NULL_TREE;
2905 /* Return true if STMT is an "escape" site from the current function. Escape
2906 sites those statements which might expose the address of a variable
2907 outside the current function. STMT is an escape site iff:
2909 1- STMT is a function call, or
2910 2- STMT is an __asm__ expression, or
2911 3- STMT is an assignment to a non-local variable, or
2912 4- STMT is a return statement.
2914 Return the type of escape site found, if we found one, or NO_ESCAPE
2915 if none. */
2917 enum escape_type
2918 is_escape_site (tree stmt)
2920 tree call = get_call_expr_in (stmt);
2921 if (call != NULL_TREE)
2923 if (!TREE_SIDE_EFFECTS (call))
2924 return ESCAPE_TO_PURE_CONST;
2926 return ESCAPE_TO_CALL;
2928 else if (TREE_CODE (stmt) == ASM_EXPR)
2929 return ESCAPE_TO_ASM;
2930 else if (TREE_CODE (stmt) == GIMPLE_MODIFY_STMT)
2932 tree lhs = GIMPLE_STMT_OPERAND (stmt, 0);
2934 /* Get to the base of _REF nodes. */
2935 if (TREE_CODE (lhs) != SSA_NAME)
2936 lhs = get_base_address (lhs);
2938 /* If we couldn't recognize the LHS of the assignment, assume that it
2939 is a non-local store. */
2940 if (lhs == NULL_TREE)
2941 return ESCAPE_UNKNOWN;
2943 if (TREE_CODE (GIMPLE_STMT_OPERAND (stmt, 1)) == NOP_EXPR
2944 || TREE_CODE (GIMPLE_STMT_OPERAND (stmt, 1)) == CONVERT_EXPR
2945 || TREE_CODE (GIMPLE_STMT_OPERAND (stmt, 1)) == VIEW_CONVERT_EXPR)
2947 tree from
2948 = TREE_TYPE (TREE_OPERAND (GIMPLE_STMT_OPERAND (stmt, 1), 0));
2949 tree to = TREE_TYPE (GIMPLE_STMT_OPERAND (stmt, 1));
2951 /* If the RHS is a conversion between a pointer and an integer, the
2952 pointer escapes since we can't track the integer. */
2953 if (POINTER_TYPE_P (from) && !POINTER_TYPE_P (to))
2954 return ESCAPE_BAD_CAST;
2956 /* Same if the RHS is a conversion between a regular pointer and a
2957 ref-all pointer since we can't track the SMT of the former. */
2958 if (POINTER_TYPE_P (from) && !TYPE_REF_CAN_ALIAS_ALL (from)
2959 && POINTER_TYPE_P (to) && TYPE_REF_CAN_ALIAS_ALL (to))
2960 return ESCAPE_BAD_CAST;
2963 /* If the LHS is an SSA name, it can't possibly represent a non-local
2964 memory store. */
2965 if (TREE_CODE (lhs) == SSA_NAME)
2966 return NO_ESCAPE;
2968 /* FIXME: LHS is not an SSA_NAME. Even if it's an assignment to a
2969 local variables we cannot be sure if it will escape, because we
2970 don't have information about objects not in SSA form. Need to
2971 implement something along the lines of
2973 J.-D. Choi, M. Gupta, M. J. Serrano, V. C. Sreedhar, and S. P.
2974 Midkiff, ``Escape analysis for java,'' in Proceedings of the
2975 Conference on Object-Oriented Programming Systems, Languages, and
2976 Applications (OOPSLA), pp. 1-19, 1999. */
2977 return ESCAPE_STORED_IN_GLOBAL;
2979 else if (TREE_CODE (stmt) == RETURN_EXPR)
2980 return ESCAPE_TO_RETURN;
2982 return NO_ESCAPE;
2985 /* Create a new memory tag of type TYPE.
2986 Does NOT push it into the current binding. */
2988 tree
2989 create_tag_raw (enum tree_code code, tree type, const char *prefix)
2991 tree tmp_var;
2993 tmp_var = build_decl (code, create_tmp_var_name (prefix), type);
2995 /* Make the variable writable. */
2996 TREE_READONLY (tmp_var) = 0;
2998 /* It doesn't start out global. */
2999 MTAG_GLOBAL (tmp_var) = 0;
3000 TREE_STATIC (tmp_var) = 0;
3001 TREE_USED (tmp_var) = 1;
3003 return tmp_var;
3006 /* Create a new memory tag of type TYPE. If IS_TYPE_TAG is true, the tag
3007 is considered to represent all the pointers whose pointed-to types are
3008 in the same alias set class. Otherwise, the tag represents a single
3009 SSA_NAME pointer variable. */
3011 static tree
3012 create_memory_tag (tree type, bool is_type_tag)
3014 tree tag = create_tag_raw (is_type_tag ? SYMBOL_MEMORY_TAG : NAME_MEMORY_TAG,
3015 type, (is_type_tag) ? "SMT" : "NMT");
3017 /* By default, memory tags are local variables. Alias analysis will
3018 determine whether they should be considered globals. */
3019 DECL_CONTEXT (tag) = current_function_decl;
3021 /* Memory tags are by definition addressable. */
3022 TREE_ADDRESSABLE (tag) = 1;
3024 set_symbol_mem_tag (tag, NULL_TREE);
3026 /* Add the tag to the symbol table. */
3027 add_referenced_var (tag);
3029 return tag;
3033 /* Create a name memory tag to represent a specific SSA_NAME pointer P_i.
3034 This is used if P_i has been found to point to a specific set of
3035 variables or to a non-aliased memory location like the address returned
3036 by malloc functions. */
3038 static tree
3039 get_nmt_for (tree ptr)
3041 struct ptr_info_def *pi = get_ptr_info (ptr);
3042 tree tag = pi->name_mem_tag;
3044 if (tag == NULL_TREE)
3045 tag = create_memory_tag (TREE_TYPE (TREE_TYPE (ptr)), false);
3046 return tag;
3050 /* Return the symbol memory tag associated to pointer PTR. A memory
3051 tag is an artificial variable that represents the memory location
3052 pointed-to by PTR. It is used to model the effects of pointer
3053 de-references on addressable variables.
3055 AI points to the data gathered during alias analysis. This
3056 function populates the array AI->POINTERS. */
3058 static tree
3059 get_smt_for (tree ptr, struct alias_info *ai)
3061 size_t i;
3062 tree tag;
3063 tree tag_type = TREE_TYPE (TREE_TYPE (ptr));
3064 alias_set_type tag_set = get_alias_set (tag_type);
3066 /* We use a unique memory tag for all the ref-all pointers. */
3067 if (PTR_IS_REF_ALL (ptr))
3069 if (!ai->ref_all_symbol_mem_tag)
3070 ai->ref_all_symbol_mem_tag = create_memory_tag (void_type_node, true);
3071 return ai->ref_all_symbol_mem_tag;
3074 /* To avoid creating unnecessary memory tags, only create one memory tag
3075 per alias set class. Note that it may be tempting to group
3076 memory tags based on conflicting alias sets instead of
3077 equivalence. That would be wrong because alias sets are not
3078 necessarily transitive (as demonstrated by the libstdc++ test
3079 23_containers/vector/cons/4.cc). Given three alias sets A, B, C
3080 such that conflicts (A, B) == true and conflicts (A, C) == true,
3081 it does not necessarily follow that conflicts (B, C) == true. */
3082 for (i = 0, tag = NULL_TREE; i < ai->num_pointers; i++)
3084 struct alias_map_d *curr = ai->pointers[i];
3085 tree curr_tag = symbol_mem_tag (curr->var);
3086 if (tag_set == curr->set)
3088 tag = curr_tag;
3089 break;
3093 /* If VAR cannot alias with any of the existing memory tags, create a new
3094 tag for PTR and add it to the POINTERS array. */
3095 if (tag == NULL_TREE)
3097 struct alias_map_d *alias_map;
3099 /* If PTR did not have a symbol tag already, create a new SMT.*
3100 artificial variable representing the memory location
3101 pointed-to by PTR. */
3102 tag = symbol_mem_tag (ptr);
3103 if (tag == NULL_TREE)
3104 tag = create_memory_tag (tag_type, true);
3106 /* Add PTR to the POINTERS array. Note that we are not interested in
3107 PTR's alias set. Instead, we cache the alias set for the memory that
3108 PTR points to. */
3109 alias_map = XCNEW (struct alias_map_d);
3110 alias_map->var = ptr;
3111 alias_map->set = tag_set;
3112 ai->pointers[ai->num_pointers++] = alias_map;
3115 /* If the pointed-to type is volatile, so is the tag. */
3116 TREE_THIS_VOLATILE (tag) |= TREE_THIS_VOLATILE (tag_type);
3118 /* Make sure that the symbol tag has the same alias set as the
3119 pointed-to type. */
3120 gcc_assert (tag_set == get_alias_set (tag));
3122 return tag;
3126 /* Create GLOBAL_VAR, an artificial global variable to act as a
3127 representative of all the variables that may be clobbered by function
3128 calls. */
3130 static void
3131 create_global_var (void)
3133 tree global_var = build_decl (VAR_DECL, get_identifier (".GLOBAL_VAR"),
3134 void_type_node);
3135 DECL_ARTIFICIAL (global_var) = 1;
3136 TREE_READONLY (global_var) = 0;
3137 DECL_EXTERNAL (global_var) = 1;
3138 TREE_STATIC (global_var) = 1;
3139 TREE_USED (global_var) = 1;
3140 DECL_CONTEXT (global_var) = NULL_TREE;
3141 TREE_THIS_VOLATILE (global_var) = 0;
3142 TREE_ADDRESSABLE (global_var) = 0;
3144 create_var_ann (global_var);
3145 mark_call_clobbered (global_var, ESCAPE_UNKNOWN);
3146 add_referenced_var (global_var);
3147 mark_sym_for_renaming (global_var);
3148 cfun->gimple_df->global_var = global_var;
3152 /* Dump alias statistics on FILE. */
3154 static void
3155 dump_alias_stats (FILE *file)
3157 const char *funcname
3158 = lang_hooks.decl_printable_name (current_function_decl, 2);
3159 fprintf (file, "\nAlias statistics for %s\n\n", funcname);
3160 fprintf (file, "Total alias queries:\t%u\n", alias_stats.alias_queries);
3161 fprintf (file, "Total alias mayalias results:\t%u\n",
3162 alias_stats.alias_mayalias);
3163 fprintf (file, "Total alias noalias results:\t%u\n",
3164 alias_stats.alias_noalias);
3165 fprintf (file, "Total simple queries:\t%u\n",
3166 alias_stats.simple_queries);
3167 fprintf (file, "Total simple resolved:\t%u\n",
3168 alias_stats.simple_resolved);
3169 fprintf (file, "Total TBAA queries:\t%u\n",
3170 alias_stats.tbaa_queries);
3171 fprintf (file, "Total TBAA resolved:\t%u\n",
3172 alias_stats.tbaa_resolved);
3173 fprintf (file, "Total non-addressable structure type queries:\t%u\n",
3174 alias_stats.structnoaddress_queries);
3175 fprintf (file, "Total non-addressable structure type resolved:\t%u\n",
3176 alias_stats.structnoaddress_resolved);
3180 /* Dump alias information on FILE. */
3182 void
3183 dump_alias_info (FILE *file)
3185 size_t i;
3186 const char *funcname
3187 = lang_hooks.decl_printable_name (current_function_decl, 2);
3188 referenced_var_iterator rvi;
3189 tree var;
3191 fprintf (file, "\nAlias information for %s\n\n", funcname);
3193 dump_memory_partitions (file);
3195 fprintf (file, "\nFlow-insensitive alias information for %s\n\n", funcname);
3197 fprintf (file, "Aliased symbols\n\n");
3199 FOR_EACH_REFERENCED_VAR (var, rvi)
3201 if (may_be_aliased (var))
3202 dump_variable (file, var);
3205 fprintf (file, "\nDereferenced pointers\n\n");
3207 FOR_EACH_REFERENCED_VAR (var, rvi)
3208 if (symbol_mem_tag (var))
3209 dump_variable (file, var);
3211 fprintf (file, "\nSymbol memory tags\n\n");
3213 FOR_EACH_REFERENCED_VAR (var, rvi)
3215 if (TREE_CODE (var) == SYMBOL_MEMORY_TAG)
3216 dump_variable (file, var);
3219 fprintf (file, "\n\nFlow-sensitive alias information for %s\n\n", funcname);
3221 fprintf (file, "SSA_NAME pointers\n\n");
3222 for (i = 1; i < num_ssa_names; i++)
3224 tree ptr = ssa_name (i);
3225 struct ptr_info_def *pi;
3227 if (ptr == NULL_TREE)
3228 continue;
3230 pi = SSA_NAME_PTR_INFO (ptr);
3231 if (!SSA_NAME_IN_FREE_LIST (ptr)
3232 && pi
3233 && pi->name_mem_tag)
3234 dump_points_to_info_for (file, ptr);
3237 fprintf (file, "\nName memory tags\n\n");
3239 FOR_EACH_REFERENCED_VAR (var, rvi)
3241 if (TREE_CODE (var) == NAME_MEMORY_TAG)
3242 dump_variable (file, var);
3245 fprintf (file, "\n");
3249 /* Dump alias information on stderr. */
3251 void
3252 debug_alias_info (void)
3254 dump_alias_info (stderr);
3258 /* Return the alias information associated with pointer T. It creates a
3259 new instance if none existed. */
3261 struct ptr_info_def *
3262 get_ptr_info (tree t)
3264 struct ptr_info_def *pi;
3266 gcc_assert (POINTER_TYPE_P (TREE_TYPE (t)));
3268 pi = SSA_NAME_PTR_INFO (t);
3269 if (pi == NULL)
3271 pi = GGC_CNEW (struct ptr_info_def);
3272 SSA_NAME_PTR_INFO (t) = pi;
3275 return pi;
3279 /* Dump points-to information for SSA_NAME PTR into FILE. */
3281 void
3282 dump_points_to_info_for (FILE *file, tree ptr)
3284 struct ptr_info_def *pi = SSA_NAME_PTR_INFO (ptr);
3286 print_generic_expr (file, ptr, dump_flags);
3288 if (pi)
3290 if (pi->name_mem_tag)
3292 fprintf (file, ", name memory tag: ");
3293 print_generic_expr (file, pi->name_mem_tag, dump_flags);
3296 if (pi->is_dereferenced)
3297 fprintf (file, ", is dereferenced");
3299 if (pi->value_escapes_p)
3300 fprintf (file, ", its value escapes");
3302 if (pi->pt_anything)
3303 fprintf (file, ", points-to anything");
3305 if (pi->pt_null)
3306 fprintf (file, ", points-to NULL");
3308 if (pi->pt_vars)
3310 fprintf (file, ", points-to vars: ");
3311 dump_decl_set (file, pi->pt_vars);
3315 fprintf (file, "\n");
3319 /* Dump points-to information for VAR into stderr. */
3321 void
3322 debug_points_to_info_for (tree var)
3324 dump_points_to_info_for (stderr, var);
3328 /* Dump points-to information into FILE. NOTE: This function is slow, as
3329 it needs to traverse the whole CFG looking for pointer SSA_NAMEs. */
3331 void
3332 dump_points_to_info (FILE *file)
3334 basic_block bb;
3335 block_stmt_iterator si;
3336 ssa_op_iter iter;
3337 const char *fname =
3338 lang_hooks.decl_printable_name (current_function_decl, 2);
3339 referenced_var_iterator rvi;
3340 tree var;
3342 fprintf (file, "\n\nPointed-to sets for pointers in %s\n\n", fname);
3344 /* First dump points-to information for the default definitions of
3345 pointer variables. This is necessary because default definitions are
3346 not part of the code. */
3347 FOR_EACH_REFERENCED_VAR (var, rvi)
3349 if (POINTER_TYPE_P (TREE_TYPE (var)))
3351 tree def = gimple_default_def (cfun, var);
3352 if (def)
3353 dump_points_to_info_for (file, def);
3357 /* Dump points-to information for every pointer defined in the program. */
3358 FOR_EACH_BB (bb)
3360 tree phi;
3362 for (phi = phi_nodes (bb); phi; phi = PHI_CHAIN (phi))
3364 tree ptr = PHI_RESULT (phi);
3365 if (POINTER_TYPE_P (TREE_TYPE (ptr)))
3366 dump_points_to_info_for (file, ptr);
3369 for (si = bsi_start (bb); !bsi_end_p (si); bsi_next (&si))
3371 tree stmt = bsi_stmt (si);
3372 tree def;
3373 FOR_EACH_SSA_TREE_OPERAND (def, stmt, iter, SSA_OP_DEF)
3374 if (TREE_CODE (def) == SSA_NAME
3375 && POINTER_TYPE_P (TREE_TYPE (def)))
3376 dump_points_to_info_for (file, def);
3380 fprintf (file, "\n");
3384 /* Dump points-to info pointed to by PTO into STDERR. */
3386 void
3387 debug_points_to_info (void)
3389 dump_points_to_info (stderr);
3392 /* Dump to FILE the list of variables that may be aliasing VAR. */
3394 void
3395 dump_may_aliases_for (FILE *file, tree var)
3397 bitmap aliases;
3399 aliases = MTAG_ALIASES (var);
3400 if (aliases)
3402 bitmap_iterator bi;
3403 unsigned int i;
3404 tree al;
3406 fprintf (file, "{ ");
3407 EXECUTE_IF_SET_IN_BITMAP (aliases, 0, i, bi)
3409 al = referenced_var (i);
3410 print_generic_expr (file, al, dump_flags);
3411 fprintf (file, " ");
3413 fprintf (file, "}");
3418 /* Dump to stderr the list of variables that may be aliasing VAR. */
3420 void
3421 debug_may_aliases_for (tree var)
3423 dump_may_aliases_for (stderr, var);
3427 /* Return true if VAR may be aliased. */
3429 bool
3430 may_be_aliased (tree var)
3432 /* Obviously. */
3433 if (TREE_ADDRESSABLE (var))
3434 return true;
3436 /* Globally visible variables can have their addresses taken by other
3437 translation units. */
3438 if (MTAG_P (var)
3439 && (MTAG_GLOBAL (var) || TREE_PUBLIC (var)))
3440 return true;
3441 else if (!MTAG_P (var)
3442 && (DECL_EXTERNAL (var) || TREE_PUBLIC (var)))
3443 return true;
3445 /* Automatic variables can't have their addresses escape any other
3446 way. This must be after the check for global variables, as
3447 extern declarations do not have TREE_STATIC set. */
3448 if (!TREE_STATIC (var))
3449 return false;
3451 /* If we're in unit-at-a-time mode, then we must have seen all
3452 occurrences of address-of operators, and so we can trust
3453 TREE_ADDRESSABLE. Otherwise we can only be sure the variable
3454 isn't addressable if it's local to the current function. */
3455 if (flag_unit_at_a_time)
3456 return false;
3458 if (decl_function_context (var) == current_function_decl)
3459 return false;
3461 return true;
3464 /* The following is based on code in add_stmt_operand to ensure that the
3465 same defs/uses/vdefs/vuses will be found after replacing a reference
3466 to var (or ARRAY_REF to var) with an INDIRECT_REF to ptr whose value
3467 is the address of var. Return a memtag for the ptr, after adding the
3468 proper may_aliases to it (which are the aliases of var, if it has any,
3469 or var itself). */
3471 static tree
3472 add_may_alias_for_new_tag (tree tag, tree var)
3474 bitmap aliases = NULL;
3476 if (MTAG_P (var))
3477 aliases = may_aliases (var);
3479 /* Case 1: |aliases| == 1 */
3480 if (aliases && bitmap_count_bits (aliases) == 1)
3482 tree ali = referenced_var (bitmap_first_set_bit (aliases));
3483 if (TREE_CODE (ali) == SYMBOL_MEMORY_TAG)
3484 return ali;
3487 /* Case 2: |aliases| == 0 */
3488 if (aliases == NULL)
3489 add_may_alias (tag, var);
3490 else
3492 /* Case 3: |aliases| > 1 */
3493 union_alias_set_into (tag, aliases);
3495 return tag;
3498 /* Create a new symbol tag for PTR. Construct the may-alias list of this type
3499 tag so that it has the aliasing of VAR, or of the relevant subvars of VAR
3500 according to the location accessed by EXPR.
3502 Note, the set of aliases represented by the new symbol tag are not marked
3503 for renaming. */
3505 void
3506 new_type_alias (tree ptr, tree var, tree expr)
3508 tree tag_type = TREE_TYPE (TREE_TYPE (ptr));
3509 tree tag;
3510 subvar_t svars;
3511 tree ali = NULL_TREE;
3512 HOST_WIDE_INT offset, size, maxsize;
3513 tree ref;
3514 VEC (tree, heap) *overlaps = NULL;
3515 subvar_t sv;
3516 unsigned int len;
3518 gcc_assert (symbol_mem_tag (ptr) == NULL_TREE);
3519 gcc_assert (!MTAG_P (var));
3521 ref = get_ref_base_and_extent (expr, &offset, &size, &maxsize);
3522 gcc_assert (ref);
3524 tag = create_memory_tag (tag_type, true);
3525 set_symbol_mem_tag (ptr, tag);
3527 /* Add VAR to the may-alias set of PTR's new symbol tag. If VAR has
3528 subvars, add the subvars to the tag instead of the actual var. */
3529 if (var_can_have_subvars (ref)
3530 && (svars = get_subvars_for_var (ref)))
3532 for (sv = svars; sv; sv = sv->next)
3534 bool exact;
3536 if (overlap_subvar (offset, maxsize, sv->var, &exact))
3537 VEC_safe_push (tree, heap, overlaps, sv->var);
3539 gcc_assert (overlaps != NULL);
3541 else if (var_can_have_subvars (var)
3542 && (svars = get_subvars_for_var (var)))
3544 /* If the REF is not a direct access to VAR (e.g., it is a dereference
3545 of a pointer), we should scan the virtual operands of REF the same
3546 way as tree-ssa-operands do. At the moment, this is somewhat
3547 difficult, so we just give up and add all the subvars of VAR.
3548 On mem-ssa branch, the scanning for virtual operands have been
3549 split from the rest of tree-ssa-operands, so it should be much
3550 easier to fix this problem correctly once mem-ssa is merged. */
3551 for (sv = svars; sv; sv = sv->next)
3552 VEC_safe_push (tree, heap, overlaps, sv->var);
3554 gcc_assert (overlaps != NULL);
3556 else
3557 ali = add_may_alias_for_new_tag (tag, var);
3559 len = VEC_length (tree, overlaps);
3560 if (len > 0)
3562 if (dump_file && (dump_flags & TDF_DETAILS))
3563 fprintf (dump_file, "\nnumber of overlapping subvars = %u\n", len);
3565 if (len == 1)
3566 ali = add_may_alias_for_new_tag (tag, VEC_index (tree, overlaps, 0));
3567 else if (len > 1)
3569 unsigned int k;
3570 tree sv_var;
3572 for (k = 0; VEC_iterate (tree, overlaps, k, sv_var); k++)
3574 ali = add_may_alias_for_new_tag (tag, sv_var);
3576 if (ali != tag)
3578 /* Can happen only if 'Case 1' of add_may_alias_for_new_tag
3579 took place. Since more than one svar was found, we add
3580 'ali' as one of the may_aliases of the new tag. */
3581 add_may_alias (tag, ali);
3582 ali = tag;
3586 VEC_free (tree, heap, overlaps);
3589 set_symbol_mem_tag (ptr, ali);
3590 TREE_READONLY (tag) = TREE_READONLY (var);
3591 MTAG_GLOBAL (tag) = is_global_var (var);
3594 /* This represents the used range of a variable. */
3596 typedef struct used_part
3598 HOST_WIDE_INT minused;
3599 HOST_WIDE_INT maxused;
3600 /* True if we have an explicit use/def of some portion of this variable,
3601 even if it is all of it. i.e. a.b = 5 or temp = a.b. */
3602 bool explicit_uses;
3603 /* True if we have an implicit use/def of some portion of this
3604 variable. Implicit uses occur when we can't tell what part we
3605 are referencing, and have to make conservative assumptions. */
3606 bool implicit_uses;
3607 /* True if the structure is only written to or taken its address. */
3608 bool write_only;
3609 } *used_part_t;
3611 /* An array of used_part structures, indexed by variable uid. */
3613 static htab_t used_portions;
3615 struct used_part_map
3617 unsigned int uid;
3618 used_part_t to;
3621 /* Return true if the uid in the two used part maps are equal. */
3623 static int
3624 used_part_map_eq (const void *va, const void *vb)
3626 const struct used_part_map *a = (const struct used_part_map *) va;
3627 const struct used_part_map *b = (const struct used_part_map *) vb;
3628 return (a->uid == b->uid);
3631 /* Hash a from uid in a used_part_map. */
3633 static unsigned int
3634 used_part_map_hash (const void *item)
3636 return ((const struct used_part_map *)item)->uid;
3639 /* Free a used part map element. */
3641 static void
3642 free_used_part_map (void *item)
3644 free (((struct used_part_map *)item)->to);
3645 free (item);
3648 /* Lookup a used_part structure for a UID. */
3650 static used_part_t
3651 up_lookup (unsigned int uid)
3653 struct used_part_map *h, in;
3654 in.uid = uid;
3655 h = (struct used_part_map *) htab_find_with_hash (used_portions, &in, uid);
3656 if (!h)
3657 return NULL;
3658 return h->to;
3661 /* Insert the pair UID, TO into the used part hashtable. */
3663 static void
3664 up_insert (unsigned int uid, used_part_t to)
3666 struct used_part_map *h;
3667 void **loc;
3669 h = XNEW (struct used_part_map);
3670 h->uid = uid;
3671 h->to = to;
3672 loc = htab_find_slot_with_hash (used_portions, h,
3673 uid, INSERT);
3674 if (*loc != NULL)
3675 free (*loc);
3676 *(struct used_part_map **) loc = h;
3680 /* Given a variable uid, UID, get or create the entry in the used portions
3681 table for the variable. */
3683 static used_part_t
3684 get_or_create_used_part_for (size_t uid)
3686 used_part_t up;
3687 if ((up = up_lookup (uid)) == NULL)
3689 up = XCNEW (struct used_part);
3690 up->minused = INT_MAX;
3691 up->maxused = 0;
3692 up->explicit_uses = false;
3693 up->implicit_uses = false;
3694 up->write_only = true;
3697 return up;
3701 /* Create and return a structure sub-variable for field type FIELD at
3702 offset OFFSET, with size SIZE, of variable VAR. If ALIAS_SET not
3703 -1 this field is non-addressable and we should use this alias set
3704 with this field. */
3706 static tree
3707 create_sft (tree var, tree field, unsigned HOST_WIDE_INT offset,
3708 unsigned HOST_WIDE_INT size, alias_set_type alias_set)
3710 tree subvar = create_tag_raw (STRUCT_FIELD_TAG, field, "SFT");
3712 /* We need to copy the various flags from VAR to SUBVAR, so that
3713 they are is_global_var iff the original variable was. */
3714 DECL_CONTEXT (subvar) = DECL_CONTEXT (var);
3715 MTAG_GLOBAL (subvar) = DECL_EXTERNAL (var);
3716 TREE_PUBLIC (subvar) = TREE_PUBLIC (var);
3717 TREE_STATIC (subvar) = TREE_STATIC (var);
3718 TREE_READONLY (subvar) = TREE_READONLY (var);
3719 TREE_ADDRESSABLE (subvar) = TREE_ADDRESSABLE (var);
3721 /* Add the new variable to REFERENCED_VARS. */
3722 set_symbol_mem_tag (subvar, NULL);
3723 add_referenced_var (subvar);
3724 SFT_PARENT_VAR (subvar) = var;
3725 SFT_OFFSET (subvar) = offset;
3726 SFT_SIZE (subvar) = size;
3727 SFT_ALIAS_SET (subvar) = alias_set;
3728 return subvar;
3732 /* Given an aggregate VAR, create the subvariables that represent its
3733 fields. */
3735 static void
3736 create_overlap_variables_for (tree var)
3738 VEC(fieldoff_s,heap) *fieldstack = NULL;
3739 used_part_t up;
3740 size_t uid = DECL_UID (var);
3742 up = up_lookup (uid);
3743 if (!up
3744 || up->write_only)
3745 return;
3747 push_fields_onto_fieldstack (TREE_TYPE (var), &fieldstack, 0, NULL,
3748 TREE_TYPE (var));
3749 if (VEC_length (fieldoff_s, fieldstack) != 0)
3751 subvar_t *subvars;
3752 fieldoff_s *fo;
3753 bool notokay = false;
3754 int fieldcount = 0;
3755 int i;
3756 HOST_WIDE_INT lastfooffset = -1;
3757 HOST_WIDE_INT lastfosize = -1;
3758 tree lastfotype = NULL_TREE;
3760 /* Not all fields have DECL_SIZE set, and those that don't, we don't
3761 know their size, and thus, can't handle.
3762 The same is true of fields with DECL_SIZE that is not an integer
3763 constant (such as variable sized fields).
3764 Fields with offsets which are not constant will have an offset < 0
3765 We *could* handle fields that are constant sized arrays, but
3766 currently don't. Doing so would require some extra changes to
3767 tree-ssa-operands.c. */
3769 for (i = 0; VEC_iterate (fieldoff_s, fieldstack, i, fo); i++)
3771 if (!fo->size
3772 || TREE_CODE (fo->size) != INTEGER_CST
3773 || fo->offset < 0)
3775 notokay = true;
3776 break;
3778 fieldcount++;
3781 /* The current heuristic we use is as follows:
3782 If the variable has no used portions in this function, no
3783 structure vars are created for it.
3784 Otherwise,
3785 If the variable has less than SALIAS_MAX_IMPLICIT_FIELDS,
3786 we always create structure vars for them.
3787 If the variable has more than SALIAS_MAX_IMPLICIT_FIELDS, and
3788 some explicit uses, we create structure vars for them.
3789 If the variable has more than SALIAS_MAX_IMPLICIT_FIELDS, and
3790 no explicit uses, we do not create structure vars for them.
3793 if (fieldcount >= SALIAS_MAX_IMPLICIT_FIELDS
3794 && !up->explicit_uses)
3796 if (dump_file && (dump_flags & TDF_DETAILS))
3798 fprintf (dump_file, "Variable ");
3799 print_generic_expr (dump_file, var, 0);
3800 fprintf (dump_file, " has no explicit uses in this function, and is > SALIAS_MAX_IMPLICIT_FIELDS, so skipping\n");
3802 notokay = true;
3805 /* Bail out, if we can't create overlap variables. */
3806 if (notokay)
3808 VEC_free (fieldoff_s, heap, fieldstack);
3809 return;
3812 /* Otherwise, create the variables. */
3813 subvars = lookup_subvars_for_var (var);
3815 sort_fieldstack (fieldstack);
3817 for (i = VEC_length (fieldoff_s, fieldstack);
3818 VEC_iterate (fieldoff_s, fieldstack, --i, fo);)
3820 subvar_t sv;
3821 HOST_WIDE_INT fosize;
3822 tree currfotype;
3824 fosize = TREE_INT_CST_LOW (fo->size);
3825 currfotype = fo->type;
3827 /* If this field isn't in the used portion,
3828 or it has the exact same offset and size as the last
3829 field, skip it. Note that we always need the field at
3830 offset 0 so we can properly handle pointers to the
3831 structure. */
3833 if ((fo->offset != 0
3834 && ((fo->offset <= up->minused
3835 && fo->offset + fosize <= up->minused)
3836 || fo->offset >= up->maxused))
3837 || (fo->offset == lastfooffset
3838 && fosize == lastfosize
3839 && currfotype == lastfotype))
3840 continue;
3841 sv = GGC_NEW (struct subvar);
3842 sv->next = *subvars;
3843 sv->var =
3844 create_sft (var, fo->type, fo->offset, fosize, fo->alias_set);
3846 if (dump_file)
3848 fprintf (dump_file, "structure field tag %s created for var %s",
3849 get_name (sv->var), get_name (var));
3850 fprintf (dump_file, " offset " HOST_WIDE_INT_PRINT_DEC,
3851 SFT_OFFSET (sv->var));
3852 fprintf (dump_file, " size " HOST_WIDE_INT_PRINT_DEC,
3853 SFT_SIZE (sv->var));
3854 fprintf (dump_file, "\n");
3857 lastfotype = currfotype;
3858 lastfooffset = fo->offset;
3859 lastfosize = fosize;
3860 *subvars = sv;
3863 /* Once we have created subvars, the original is no longer call
3864 clobbered on its own. Its call clobbered status depends
3865 completely on the call clobbered status of the subvars.
3867 add_referenced_var in the above loop will take care of
3868 marking subvars of global variables as call clobbered for us
3869 to start, since they are global as well. */
3870 clear_call_clobbered (var);
3873 VEC_free (fieldoff_s, heap, fieldstack);
3877 /* Find the conservative answer to the question of what portions of what
3878 structures are used by this statement. We assume that if we have a
3879 component ref with a known size + offset, that we only need that part
3880 of the structure. For unknown cases, or cases where we do something
3881 to the whole structure, we assume we need to create fields for the
3882 entire structure. */
3884 static tree
3885 find_used_portions (tree *tp, int *walk_subtrees, void *lhs_p)
3887 switch (TREE_CODE (*tp))
3889 case GIMPLE_MODIFY_STMT:
3890 /* Recurse manually here to track whether the use is in the
3891 LHS of an assignment. */
3892 find_used_portions (&GIMPLE_STMT_OPERAND (*tp, 0), walk_subtrees, tp);
3893 return find_used_portions (&GIMPLE_STMT_OPERAND (*tp, 1),
3894 walk_subtrees, NULL);
3895 case REALPART_EXPR:
3896 case IMAGPART_EXPR:
3897 case COMPONENT_REF:
3898 case ARRAY_REF:
3900 HOST_WIDE_INT bitsize;
3901 HOST_WIDE_INT bitmaxsize;
3902 HOST_WIDE_INT bitpos;
3903 tree ref;
3904 ref = get_ref_base_and_extent (*tp, &bitpos, &bitsize, &bitmaxsize);
3905 if (DECL_P (ref)
3906 && var_can_have_subvars (ref)
3907 && bitmaxsize != -1)
3909 size_t uid = DECL_UID (ref);
3910 used_part_t up;
3912 up = get_or_create_used_part_for (uid);
3914 if (bitpos <= up->minused)
3915 up->minused = bitpos;
3916 if ((bitpos + bitmaxsize >= up->maxused))
3917 up->maxused = bitpos + bitmaxsize;
3919 if (bitsize == bitmaxsize)
3920 up->explicit_uses = true;
3921 else
3922 up->implicit_uses = true;
3923 if (!lhs_p)
3924 up->write_only = false;
3925 up_insert (uid, up);
3927 *walk_subtrees = 0;
3928 return NULL_TREE;
3931 break;
3932 /* This is here to make sure we mark the entire base variable as used
3933 when you take its address. Because our used portion analysis is
3934 simple, we aren't looking at casts or pointer arithmetic to see what
3935 happens when you take the address. */
3936 case ADDR_EXPR:
3938 tree var = get_base_address (TREE_OPERAND (*tp, 0));
3940 if (var
3941 && DECL_P (var)
3942 && DECL_SIZE (var)
3943 && var_can_have_subvars (var)
3944 && TREE_CODE (DECL_SIZE (var)) == INTEGER_CST)
3946 used_part_t up;
3947 size_t uid = DECL_UID (var);
3949 up = get_or_create_used_part_for (uid);
3951 up->minused = 0;
3952 up->maxused = TREE_INT_CST_LOW (DECL_SIZE (var));
3953 up->implicit_uses = true;
3954 if (!lhs_p)
3955 up->write_only = false;
3957 up_insert (uid, up);
3958 *walk_subtrees = 0;
3959 return NULL_TREE;
3962 break;
3963 case CALL_EXPR:
3965 int i;
3966 int nargs = call_expr_nargs (*tp);
3967 for (i = 0; i < nargs; i++)
3969 tree *arg = &CALL_EXPR_ARG (*tp, i);
3970 if (TREE_CODE (*arg) == ADDR_EXPR)
3971 find_used_portions (arg, walk_subtrees, NULL);
3973 *walk_subtrees = 0;
3974 return NULL_TREE;
3976 case VAR_DECL:
3977 case PARM_DECL:
3978 case RESULT_DECL:
3980 tree var = *tp;
3981 if (DECL_SIZE (var)
3982 && var_can_have_subvars (var)
3983 && TREE_CODE (DECL_SIZE (var)) == INTEGER_CST)
3985 used_part_t up;
3986 size_t uid = DECL_UID (var);
3988 up = get_or_create_used_part_for (uid);
3990 up->minused = 0;
3991 up->maxused = TREE_INT_CST_LOW (DECL_SIZE (var));
3992 up->implicit_uses = true;
3994 up_insert (uid, up);
3995 *walk_subtrees = 0;
3996 return NULL_TREE;
3999 break;
4001 default:
4002 break;
4005 return NULL_TREE;
4008 /* Create structure field variables for structures used in this function. */
4010 static unsigned int
4011 create_structure_vars (void)
4013 basic_block bb;
4014 safe_referenced_var_iterator rvi;
4015 VEC (tree, heap) *varvec = NULL;
4016 tree var;
4018 used_portions = htab_create (10, used_part_map_hash, used_part_map_eq,
4019 free_used_part_map);
4021 FOR_EACH_BB (bb)
4023 block_stmt_iterator bsi;
4024 tree phi;
4026 for (phi = phi_nodes (bb); phi; phi = PHI_CHAIN (phi))
4028 use_operand_p use;
4029 ssa_op_iter iter;
4031 FOR_EACH_PHI_ARG (use, phi, iter, SSA_OP_USE)
4033 tree op = USE_FROM_PTR (use);
4034 walk_tree_without_duplicates (&op, find_used_portions,
4035 NULL);
4039 for (bsi = bsi_start (bb); !bsi_end_p (bsi); bsi_next (&bsi))
4041 walk_tree_without_duplicates (bsi_stmt_ptr (bsi),
4042 find_used_portions,
4043 NULL);
4046 FOR_EACH_REFERENCED_VAR_SAFE (var, varvec, rvi)
4048 /* The C++ FE creates vars without DECL_SIZE set, for some reason. */
4049 if (var
4050 && DECL_SIZE (var)
4051 && var_can_have_subvars (var)
4052 && !MTAG_P (var)
4053 && TREE_CODE (DECL_SIZE (var)) == INTEGER_CST)
4054 create_overlap_variables_for (var);
4056 htab_delete (used_portions);
4057 VEC_free (tree, heap, varvec);
4059 /* Update SSA operands of statements mentioning variables we split. */
4060 if (gimple_in_ssa_p (cfun))
4061 FOR_EACH_BB (bb)
4063 block_stmt_iterator bsi;
4064 for (bsi = bsi_start (bb); !bsi_end_p (bsi); bsi_next (&bsi))
4066 tree stmt = bsi_stmt (bsi);
4067 bool update = false;
4068 unsigned int i;
4069 bitmap_iterator bi;
4071 if (STORED_SYMS (stmt))
4072 EXECUTE_IF_SET_IN_BITMAP (STORED_SYMS (stmt), 0, i, bi)
4074 tree sym = referenced_var_lookup (i);
4075 if (get_subvars_for_var (sym))
4077 update = true;
4078 break;
4082 if (LOADED_SYMS (stmt) && !update)
4083 EXECUTE_IF_SET_IN_BITMAP (LOADED_SYMS (stmt), 0, i, bi)
4085 tree sym = referenced_var_lookup (i);
4086 if (get_subvars_for_var (sym))
4088 update = true;
4089 break;
4093 if (stmt_ann (stmt)->addresses_taken && !update)
4094 EXECUTE_IF_SET_IN_BITMAP (stmt_ann (stmt)->addresses_taken,
4095 0, i, bi)
4097 tree sym = referenced_var_lookup (i);
4098 if (get_subvars_for_var (sym))
4100 update = true;
4101 break;
4105 if (update)
4106 update_stmt (stmt);
4110 return TODO_rebuild_alias;
4113 static bool
4114 gate_structure_vars (void)
4116 return flag_tree_salias != 0;
4119 struct tree_opt_pass pass_create_structure_vars =
4121 "salias", /* name */
4122 gate_structure_vars, /* gate */
4123 create_structure_vars, /* execute */
4124 NULL, /* sub */
4125 NULL, /* next */
4126 0, /* static_pass_number */
4127 0, /* tv_id */
4128 PROP_cfg, /* properties_required */
4129 0, /* properties_provided */
4130 0, /* properties_destroyed */
4131 0, /* todo_flags_start */
4132 TODO_dump_func, /* todo_flags_finish */
4133 0 /* letter */
4136 /* Reset the call_clobbered flags on our referenced vars. In
4137 theory, this only needs to be done for globals. */
4139 static unsigned int
4140 reset_cc_flags (void)
4142 tree var;
4143 referenced_var_iterator rvi;
4145 FOR_EACH_REFERENCED_VAR (var, rvi)
4146 var_ann (var)->call_clobbered = false;
4147 return 0;
4150 struct tree_opt_pass pass_reset_cc_flags =
4152 NULL, /* name */
4153 NULL, /* gate */
4154 reset_cc_flags, /* execute */
4155 NULL, /* sub */
4156 NULL, /* next */
4157 0, /* static_pass_number */
4158 0, /* tv_id */
4159 PROP_referenced_vars |PROP_cfg, /* properties_required */
4160 0, /* properties_provided */
4161 0, /* properties_destroyed */
4162 0, /* todo_flags_start */
4163 0, /* todo_flags_finish */
4164 0 /* letter */