2014-04-14 Martin Jambor <mjambor@suse.cz>
[official-gcc.git] / gcc / tree-sra.c
blob91286b44842319ed5584565bb9a149b6a4fb83db
1 /* Scalar Replacement of Aggregates (SRA) converts some structure
2 references into scalar references, exposing them to the scalar
3 optimizers.
4 Copyright (C) 2008-2014 Free Software Foundation, Inc.
5 Contributed by Martin Jambor <mjambor@suse.cz>
7 This file is part of GCC.
9 GCC is free software; you can redistribute it and/or modify it under
10 the terms of the GNU General Public License as published by the Free
11 Software Foundation; either version 3, or (at your option) any later
12 version.
14 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
15 WARRANTY; without even the implied warranty of MERCHANTABILITY or
16 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 for more details.
19 You should have received a copy of the GNU General Public License
20 along with GCC; see the file COPYING3. If not see
21 <http://www.gnu.org/licenses/>. */
23 /* This file implements Scalar Reduction of Aggregates (SRA). SRA is run
24 twice, once in the early stages of compilation (early SRA) and once in the
25 late stages (late SRA). The aim of both is to turn references to scalar
26 parts of aggregates into uses of independent scalar variables.
28 The two passes are nearly identical, the only difference is that early SRA
29 does not scalarize unions which are used as the result in a GIMPLE_RETURN
30 statement because together with inlining this can lead to weird type
31 conversions.
33 Both passes operate in four stages:
35 1. The declarations that have properties which make them candidates for
36 scalarization are identified in function find_var_candidates(). The
37 candidates are stored in candidate_bitmap.
39 2. The function body is scanned. In the process, declarations which are
40 used in a manner that prevent their scalarization are removed from the
41 candidate bitmap. More importantly, for every access into an aggregate,
42 an access structure (struct access) is created by create_access() and
43 stored in a vector associated with the aggregate. Among other
44 information, the aggregate declaration, the offset and size of the access
45 and its type are stored in the structure.
47 On a related note, assign_link structures are created for every assign
48 statement between candidate aggregates and attached to the related
49 accesses.
51 3. The vectors of accesses are analyzed. They are first sorted according to
52 their offset and size and then scanned for partially overlapping accesses
53 (i.e. those which overlap but one is not entirely within another). Such
54 an access disqualifies the whole aggregate from being scalarized.
56 If there is no such inhibiting overlap, a representative access structure
57 is chosen for every unique combination of offset and size. Afterwards,
58 the pass builds a set of trees from these structures, in which children
59 of an access are within their parent (in terms of offset and size).
61 Then accesses are propagated whenever possible (i.e. in cases when it
62 does not create a partially overlapping access) across assign_links from
63 the right hand side to the left hand side.
65 Then the set of trees for each declaration is traversed again and those
66 accesses which should be replaced by a scalar are identified.
68 4. The function is traversed again, and for every reference into an
69 aggregate that has some component which is about to be scalarized,
70 statements are amended and new statements are created as necessary.
71 Finally, if a parameter got scalarized, the scalar replacements are
72 initialized with values from respective parameter aggregates. */
74 #include "config.h"
75 #include "system.h"
76 #include "coretypes.h"
77 #include "hash-table.h"
78 #include "alloc-pool.h"
79 #include "tm.h"
80 #include "tree.h"
81 #include "pointer-set.h"
82 #include "basic-block.h"
83 #include "tree-ssa-alias.h"
84 #include "internal-fn.h"
85 #include "tree-eh.h"
86 #include "gimple-expr.h"
87 #include "is-a.h"
88 #include "gimple.h"
89 #include "stor-layout.h"
90 #include "gimplify.h"
91 #include "gimple-iterator.h"
92 #include "gimplify-me.h"
93 #include "gimple-walk.h"
94 #include "bitmap.h"
95 #include "gimple-ssa.h"
96 #include "tree-cfg.h"
97 #include "tree-phinodes.h"
98 #include "ssa-iterators.h"
99 #include "stringpool.h"
100 #include "tree-ssanames.h"
101 #include "expr.h"
102 #include "tree-dfa.h"
103 #include "tree-ssa.h"
104 #include "tree-pass.h"
105 #include "ipa-prop.h"
106 #include "statistics.h"
107 #include "params.h"
108 #include "target.h"
109 #include "flags.h"
110 #include "dbgcnt.h"
111 #include "tree-inline.h"
112 #include "gimple-pretty-print.h"
113 #include "ipa-inline.h"
114 #include "ipa-utils.h"
116 /* Enumeration of all aggregate reductions we can do. */
117 enum sra_mode { SRA_MODE_EARLY_IPA, /* early call regularization */
118 SRA_MODE_EARLY_INTRA, /* early intraprocedural SRA */
119 SRA_MODE_INTRA }; /* late intraprocedural SRA */
121 /* Global variable describing which aggregate reduction we are performing at
122 the moment. */
123 static enum sra_mode sra_mode;
125 struct assign_link;
127 /* ACCESS represents each access to an aggregate variable (as a whole or a
128 part). It can also represent a group of accesses that refer to exactly the
129 same fragment of an aggregate (i.e. those that have exactly the same offset
130 and size). Such representatives for a single aggregate, once determined,
131 are linked in a linked list and have the group fields set.
133 Moreover, when doing intraprocedural SRA, a tree is built from those
134 representatives (by the means of first_child and next_sibling pointers), in
135 which all items in a subtree are "within" the root, i.e. their offset is
136 greater or equal to offset of the root and offset+size is smaller or equal
137 to offset+size of the root. Children of an access are sorted by offset.
139 Note that accesses to parts of vector and complex number types always
140 represented by an access to the whole complex number or a vector. It is a
141 duty of the modifying functions to replace them appropriately. */
143 struct access
145 /* Values returned by `get_ref_base_and_extent' for each component reference
146 If EXPR isn't a component reference just set `BASE = EXPR', `OFFSET = 0',
147 `SIZE = TREE_SIZE (TREE_TYPE (expr))'. */
148 HOST_WIDE_INT offset;
149 HOST_WIDE_INT size;
150 tree base;
152 /* Expression. It is context dependent so do not use it to create new
153 expressions to access the original aggregate. See PR 42154 for a
154 testcase. */
155 tree expr;
156 /* Type. */
157 tree type;
159 /* The statement this access belongs to. */
160 gimple stmt;
162 /* Next group representative for this aggregate. */
163 struct access *next_grp;
165 /* Pointer to the group representative. Pointer to itself if the struct is
166 the representative. */
167 struct access *group_representative;
169 /* If this access has any children (in terms of the definition above), this
170 points to the first one. */
171 struct access *first_child;
173 /* In intraprocedural SRA, pointer to the next sibling in the access tree as
174 described above. In IPA-SRA this is a pointer to the next access
175 belonging to the same group (having the same representative). */
176 struct access *next_sibling;
178 /* Pointers to the first and last element in the linked list of assign
179 links. */
180 struct assign_link *first_link, *last_link;
182 /* Pointer to the next access in the work queue. */
183 struct access *next_queued;
185 /* Replacement variable for this access "region." Never to be accessed
186 directly, always only by the means of get_access_replacement() and only
187 when grp_to_be_replaced flag is set. */
188 tree replacement_decl;
190 /* Is this particular access write access? */
191 unsigned write : 1;
193 /* Is this access an access to a non-addressable field? */
194 unsigned non_addressable : 1;
196 /* Is this access currently in the work queue? */
197 unsigned grp_queued : 1;
199 /* Does this group contain a write access? This flag is propagated down the
200 access tree. */
201 unsigned grp_write : 1;
203 /* Does this group contain a read access? This flag is propagated down the
204 access tree. */
205 unsigned grp_read : 1;
207 /* Does this group contain a read access that comes from an assignment
208 statement? This flag is propagated down the access tree. */
209 unsigned grp_assignment_read : 1;
211 /* Does this group contain a write access that comes from an assignment
212 statement? This flag is propagated down the access tree. */
213 unsigned grp_assignment_write : 1;
215 /* Does this group contain a read access through a scalar type? This flag is
216 not propagated in the access tree in any direction. */
217 unsigned grp_scalar_read : 1;
219 /* Does this group contain a write access through a scalar type? This flag
220 is not propagated in the access tree in any direction. */
221 unsigned grp_scalar_write : 1;
223 /* Is this access an artificial one created to scalarize some record
224 entirely? */
225 unsigned grp_total_scalarization : 1;
227 /* Other passes of the analysis use this bit to make function
228 analyze_access_subtree create scalar replacements for this group if
229 possible. */
230 unsigned grp_hint : 1;
232 /* Is the subtree rooted in this access fully covered by scalar
233 replacements? */
234 unsigned grp_covered : 1;
236 /* If set to true, this access and all below it in an access tree must not be
237 scalarized. */
238 unsigned grp_unscalarizable_region : 1;
240 /* Whether data have been written to parts of the aggregate covered by this
241 access which is not to be scalarized. This flag is propagated up in the
242 access tree. */
243 unsigned grp_unscalarized_data : 1;
245 /* Does this access and/or group contain a write access through a
246 BIT_FIELD_REF? */
247 unsigned grp_partial_lhs : 1;
249 /* Set when a scalar replacement should be created for this variable. */
250 unsigned grp_to_be_replaced : 1;
252 /* Set when we want a replacement for the sole purpose of having it in
253 generated debug statements. */
254 unsigned grp_to_be_debug_replaced : 1;
256 /* Should TREE_NO_WARNING of a replacement be set? */
257 unsigned grp_no_warning : 1;
259 /* Is it possible that the group refers to data which might be (directly or
260 otherwise) modified? */
261 unsigned grp_maybe_modified : 1;
263 /* Set when this is a representative of a pointer to scalar (i.e. by
264 reference) parameter which we consider for turning into a plain scalar
265 (i.e. a by value parameter). */
266 unsigned grp_scalar_ptr : 1;
268 /* Set when we discover that this pointer is not safe to dereference in the
269 caller. */
270 unsigned grp_not_necessarilly_dereferenced : 1;
273 typedef struct access *access_p;
276 /* Alloc pool for allocating access structures. */
277 static alloc_pool access_pool;
279 /* A structure linking lhs and rhs accesses from an aggregate assignment. They
280 are used to propagate subaccesses from rhs to lhs as long as they don't
281 conflict with what is already there. */
282 struct assign_link
284 struct access *lacc, *racc;
285 struct assign_link *next;
288 /* Alloc pool for allocating assign link structures. */
289 static alloc_pool link_pool;
291 /* Base (tree) -> Vector (vec<access_p> *) map. */
292 static struct pointer_map_t *base_access_vec;
294 /* Candidate hash table helpers. */
296 struct uid_decl_hasher : typed_noop_remove <tree_node>
298 typedef tree_node value_type;
299 typedef tree_node compare_type;
300 static inline hashval_t hash (const value_type *);
301 static inline bool equal (const value_type *, const compare_type *);
304 /* Hash a tree in a uid_decl_map. */
306 inline hashval_t
307 uid_decl_hasher::hash (const value_type *item)
309 return item->decl_minimal.uid;
312 /* Return true if the DECL_UID in both trees are equal. */
314 inline bool
315 uid_decl_hasher::equal (const value_type *a, const compare_type *b)
317 return (a->decl_minimal.uid == b->decl_minimal.uid);
320 /* Set of candidates. */
321 static bitmap candidate_bitmap;
322 static hash_table <uid_decl_hasher> candidates;
324 /* For a candidate UID return the candidates decl. */
326 static inline tree
327 candidate (unsigned uid)
329 tree_node t;
330 t.decl_minimal.uid = uid;
331 return candidates.find_with_hash (&t, static_cast <hashval_t> (uid));
334 /* Bitmap of candidates which we should try to entirely scalarize away and
335 those which cannot be (because they are and need be used as a whole). */
336 static bitmap should_scalarize_away_bitmap, cannot_scalarize_away_bitmap;
338 /* Obstack for creation of fancy names. */
339 static struct obstack name_obstack;
341 /* Head of a linked list of accesses that need to have its subaccesses
342 propagated to their assignment counterparts. */
343 static struct access *work_queue_head;
345 /* Number of parameters of the analyzed function when doing early ipa SRA. */
346 static int func_param_count;
348 /* scan_function sets the following to true if it encounters a call to
349 __builtin_apply_args. */
350 static bool encountered_apply_args;
352 /* Set by scan_function when it finds a recursive call. */
353 static bool encountered_recursive_call;
355 /* Set by scan_function when it finds a recursive call with less actual
356 arguments than formal parameters.. */
357 static bool encountered_unchangable_recursive_call;
359 /* This is a table in which for each basic block and parameter there is a
360 distance (offset + size) in that parameter which is dereferenced and
361 accessed in that BB. */
362 static HOST_WIDE_INT *bb_dereferences;
363 /* Bitmap of BBs that can cause the function to "stop" progressing by
364 returning, throwing externally, looping infinitely or calling a function
365 which might abort etc.. */
366 static bitmap final_bbs;
368 /* Representative of no accesses at all. */
369 static struct access no_accesses_representant;
371 /* Predicate to test the special value. */
373 static inline bool
374 no_accesses_p (struct access *access)
376 return access == &no_accesses_representant;
379 /* Dump contents of ACCESS to file F in a human friendly way. If GRP is true,
380 representative fields are dumped, otherwise those which only describe the
381 individual access are. */
383 static struct
385 /* Number of processed aggregates is readily available in
386 analyze_all_variable_accesses and so is not stored here. */
388 /* Number of created scalar replacements. */
389 int replacements;
391 /* Number of times sra_modify_expr or sra_modify_assign themselves changed an
392 expression. */
393 int exprs;
395 /* Number of statements created by generate_subtree_copies. */
396 int subtree_copies;
398 /* Number of statements created by load_assign_lhs_subreplacements. */
399 int subreplacements;
401 /* Number of times sra_modify_assign has deleted a statement. */
402 int deleted;
404 /* Number of times sra_modify_assign has to deal with subaccesses of LHS and
405 RHS reparately due to type conversions or nonexistent matching
406 references. */
407 int separate_lhs_rhs_handling;
409 /* Number of parameters that were removed because they were unused. */
410 int deleted_unused_parameters;
412 /* Number of scalars passed as parameters by reference that have been
413 converted to be passed by value. */
414 int scalar_by_ref_to_by_val;
416 /* Number of aggregate parameters that were replaced by one or more of their
417 components. */
418 int aggregate_params_reduced;
420 /* Numbber of components created when splitting aggregate parameters. */
421 int param_reductions_created;
422 } sra_stats;
424 static void
425 dump_access (FILE *f, struct access *access, bool grp)
427 fprintf (f, "access { ");
428 fprintf (f, "base = (%d)'", DECL_UID (access->base));
429 print_generic_expr (f, access->base, 0);
430 fprintf (f, "', offset = " HOST_WIDE_INT_PRINT_DEC, access->offset);
431 fprintf (f, ", size = " HOST_WIDE_INT_PRINT_DEC, access->size);
432 fprintf (f, ", expr = ");
433 print_generic_expr (f, access->expr, 0);
434 fprintf (f, ", type = ");
435 print_generic_expr (f, access->type, 0);
436 if (grp)
437 fprintf (f, ", grp_read = %d, grp_write = %d, grp_assignment_read = %d, "
438 "grp_assignment_write = %d, grp_scalar_read = %d, "
439 "grp_scalar_write = %d, grp_total_scalarization = %d, "
440 "grp_hint = %d, grp_covered = %d, "
441 "grp_unscalarizable_region = %d, grp_unscalarized_data = %d, "
442 "grp_partial_lhs = %d, grp_to_be_replaced = %d, "
443 "grp_to_be_debug_replaced = %d, grp_maybe_modified = %d, "
444 "grp_not_necessarilly_dereferenced = %d\n",
445 access->grp_read, access->grp_write, access->grp_assignment_read,
446 access->grp_assignment_write, access->grp_scalar_read,
447 access->grp_scalar_write, access->grp_total_scalarization,
448 access->grp_hint, access->grp_covered,
449 access->grp_unscalarizable_region, access->grp_unscalarized_data,
450 access->grp_partial_lhs, access->grp_to_be_replaced,
451 access->grp_to_be_debug_replaced, access->grp_maybe_modified,
452 access->grp_not_necessarilly_dereferenced);
453 else
454 fprintf (f, ", write = %d, grp_total_scalarization = %d, "
455 "grp_partial_lhs = %d\n",
456 access->write, access->grp_total_scalarization,
457 access->grp_partial_lhs);
460 /* Dump a subtree rooted in ACCESS to file F, indent by LEVEL. */
462 static void
463 dump_access_tree_1 (FILE *f, struct access *access, int level)
467 int i;
469 for (i = 0; i < level; i++)
470 fputs ("* ", dump_file);
472 dump_access (f, access, true);
474 if (access->first_child)
475 dump_access_tree_1 (f, access->first_child, level + 1);
477 access = access->next_sibling;
479 while (access);
482 /* Dump all access trees for a variable, given the pointer to the first root in
483 ACCESS. */
485 static void
486 dump_access_tree (FILE *f, struct access *access)
488 for (; access; access = access->next_grp)
489 dump_access_tree_1 (f, access, 0);
492 /* Return true iff ACC is non-NULL and has subaccesses. */
494 static inline bool
495 access_has_children_p (struct access *acc)
497 return acc && acc->first_child;
500 /* Return true iff ACC is (partly) covered by at least one replacement. */
502 static bool
503 access_has_replacements_p (struct access *acc)
505 struct access *child;
506 if (acc->grp_to_be_replaced)
507 return true;
508 for (child = acc->first_child; child; child = child->next_sibling)
509 if (access_has_replacements_p (child))
510 return true;
511 return false;
514 /* Return a vector of pointers to accesses for the variable given in BASE or
515 NULL if there is none. */
517 static vec<access_p> *
518 get_base_access_vector (tree base)
520 void **slot;
522 slot = pointer_map_contains (base_access_vec, base);
523 if (!slot)
524 return NULL;
525 else
526 return *(vec<access_p> **) slot;
529 /* Find an access with required OFFSET and SIZE in a subtree of accesses rooted
530 in ACCESS. Return NULL if it cannot be found. */
532 static struct access *
533 find_access_in_subtree (struct access *access, HOST_WIDE_INT offset,
534 HOST_WIDE_INT size)
536 while (access && (access->offset != offset || access->size != size))
538 struct access *child = access->first_child;
540 while (child && (child->offset + child->size <= offset))
541 child = child->next_sibling;
542 access = child;
545 return access;
548 /* Return the first group representative for DECL or NULL if none exists. */
550 static struct access *
551 get_first_repr_for_decl (tree base)
553 vec<access_p> *access_vec;
555 access_vec = get_base_access_vector (base);
556 if (!access_vec)
557 return NULL;
559 return (*access_vec)[0];
562 /* Find an access representative for the variable BASE and given OFFSET and
563 SIZE. Requires that access trees have already been built. Return NULL if
564 it cannot be found. */
566 static struct access *
567 get_var_base_offset_size_access (tree base, HOST_WIDE_INT offset,
568 HOST_WIDE_INT size)
570 struct access *access;
572 access = get_first_repr_for_decl (base);
573 while (access && (access->offset + access->size <= offset))
574 access = access->next_grp;
575 if (!access)
576 return NULL;
578 return find_access_in_subtree (access, offset, size);
581 /* Add LINK to the linked list of assign links of RACC. */
582 static void
583 add_link_to_rhs (struct access *racc, struct assign_link *link)
585 gcc_assert (link->racc == racc);
587 if (!racc->first_link)
589 gcc_assert (!racc->last_link);
590 racc->first_link = link;
592 else
593 racc->last_link->next = link;
595 racc->last_link = link;
596 link->next = NULL;
599 /* Move all link structures in their linked list in OLD_RACC to the linked list
600 in NEW_RACC. */
601 static void
602 relink_to_new_repr (struct access *new_racc, struct access *old_racc)
604 if (!old_racc->first_link)
606 gcc_assert (!old_racc->last_link);
607 return;
610 if (new_racc->first_link)
612 gcc_assert (!new_racc->last_link->next);
613 gcc_assert (!old_racc->last_link || !old_racc->last_link->next);
615 new_racc->last_link->next = old_racc->first_link;
616 new_racc->last_link = old_racc->last_link;
618 else
620 gcc_assert (!new_racc->last_link);
622 new_racc->first_link = old_racc->first_link;
623 new_racc->last_link = old_racc->last_link;
625 old_racc->first_link = old_racc->last_link = NULL;
628 /* Add ACCESS to the work queue (which is actually a stack). */
630 static void
631 add_access_to_work_queue (struct access *access)
633 if (!access->grp_queued)
635 gcc_assert (!access->next_queued);
636 access->next_queued = work_queue_head;
637 access->grp_queued = 1;
638 work_queue_head = access;
642 /* Pop an access from the work queue, and return it, assuming there is one. */
644 static struct access *
645 pop_access_from_work_queue (void)
647 struct access *access = work_queue_head;
649 work_queue_head = access->next_queued;
650 access->next_queued = NULL;
651 access->grp_queued = 0;
652 return access;
656 /* Allocate necessary structures. */
658 static void
659 sra_initialize (void)
661 candidate_bitmap = BITMAP_ALLOC (NULL);
662 candidates.create (vec_safe_length (cfun->local_decls) / 2);
663 should_scalarize_away_bitmap = BITMAP_ALLOC (NULL);
664 cannot_scalarize_away_bitmap = BITMAP_ALLOC (NULL);
665 gcc_obstack_init (&name_obstack);
666 access_pool = create_alloc_pool ("SRA accesses", sizeof (struct access), 16);
667 link_pool = create_alloc_pool ("SRA links", sizeof (struct assign_link), 16);
668 base_access_vec = pointer_map_create ();
669 memset (&sra_stats, 0, sizeof (sra_stats));
670 encountered_apply_args = false;
671 encountered_recursive_call = false;
672 encountered_unchangable_recursive_call = false;
675 /* Hook fed to pointer_map_traverse, deallocate stored vectors. */
677 static bool
678 delete_base_accesses (const void *key ATTRIBUTE_UNUSED, void **value,
679 void *data ATTRIBUTE_UNUSED)
681 vec<access_p> *access_vec = (vec<access_p> *) *value;
682 vec_free (access_vec);
683 return true;
686 /* Deallocate all general structures. */
688 static void
689 sra_deinitialize (void)
691 BITMAP_FREE (candidate_bitmap);
692 candidates.dispose ();
693 BITMAP_FREE (should_scalarize_away_bitmap);
694 BITMAP_FREE (cannot_scalarize_away_bitmap);
695 free_alloc_pool (access_pool);
696 free_alloc_pool (link_pool);
697 obstack_free (&name_obstack, NULL);
699 pointer_map_traverse (base_access_vec, delete_base_accesses, NULL);
700 pointer_map_destroy (base_access_vec);
703 /* Remove DECL from candidates for SRA and write REASON to the dump file if
704 there is one. */
705 static void
706 disqualify_candidate (tree decl, const char *reason)
708 if (bitmap_clear_bit (candidate_bitmap, DECL_UID (decl)))
709 candidates.clear_slot (candidates.find_slot_with_hash (decl,
710 DECL_UID (decl),
711 NO_INSERT));
713 if (dump_file && (dump_flags & TDF_DETAILS))
715 fprintf (dump_file, "! Disqualifying ");
716 print_generic_expr (dump_file, decl, 0);
717 fprintf (dump_file, " - %s\n", reason);
721 /* Return true iff the type contains a field or an element which does not allow
722 scalarization. */
724 static bool
725 type_internals_preclude_sra_p (tree type, const char **msg)
727 tree fld;
728 tree et;
730 switch (TREE_CODE (type))
732 case RECORD_TYPE:
733 case UNION_TYPE:
734 case QUAL_UNION_TYPE:
735 for (fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
736 if (TREE_CODE (fld) == FIELD_DECL)
738 tree ft = TREE_TYPE (fld);
740 if (TREE_THIS_VOLATILE (fld))
742 *msg = "volatile structure field";
743 return true;
745 if (!DECL_FIELD_OFFSET (fld))
747 *msg = "no structure field offset";
748 return true;
750 if (!DECL_SIZE (fld))
752 *msg = "zero structure field size";
753 return true;
755 if (!tree_fits_uhwi_p (DECL_FIELD_OFFSET (fld)))
757 *msg = "structure field offset not fixed";
758 return true;
760 if (!tree_fits_uhwi_p (DECL_SIZE (fld)))
762 *msg = "structure field size not fixed";
763 return true;
765 if (!tree_fits_shwi_p (bit_position (fld)))
767 *msg = "structure field size too big";
768 return true;
770 if (AGGREGATE_TYPE_P (ft)
771 && int_bit_position (fld) % BITS_PER_UNIT != 0)
773 *msg = "structure field is bit field";
774 return true;
777 if (AGGREGATE_TYPE_P (ft) && type_internals_preclude_sra_p (ft, msg))
778 return true;
781 return false;
783 case ARRAY_TYPE:
784 et = TREE_TYPE (type);
786 if (TYPE_VOLATILE (et))
788 *msg = "element type is volatile";
789 return true;
792 if (AGGREGATE_TYPE_P (et) && type_internals_preclude_sra_p (et, msg))
793 return true;
795 return false;
797 default:
798 return false;
802 /* If T is an SSA_NAME, return NULL if it is not a default def or return its
803 base variable if it is. Return T if it is not an SSA_NAME. */
805 static tree
806 get_ssa_base_param (tree t)
808 if (TREE_CODE (t) == SSA_NAME)
810 if (SSA_NAME_IS_DEFAULT_DEF (t))
811 return SSA_NAME_VAR (t);
812 else
813 return NULL_TREE;
815 return t;
818 /* Mark a dereference of BASE of distance DIST in a basic block tht STMT
819 belongs to, unless the BB has already been marked as a potentially
820 final. */
822 static void
823 mark_parm_dereference (tree base, HOST_WIDE_INT dist, gimple stmt)
825 basic_block bb = gimple_bb (stmt);
826 int idx, parm_index = 0;
827 tree parm;
829 if (bitmap_bit_p (final_bbs, bb->index))
830 return;
832 for (parm = DECL_ARGUMENTS (current_function_decl);
833 parm && parm != base;
834 parm = DECL_CHAIN (parm))
835 parm_index++;
837 gcc_assert (parm_index < func_param_count);
839 idx = bb->index * func_param_count + parm_index;
840 if (bb_dereferences[idx] < dist)
841 bb_dereferences[idx] = dist;
844 /* Allocate an access structure for BASE, OFFSET and SIZE, clear it, fill in
845 the three fields. Also add it to the vector of accesses corresponding to
846 the base. Finally, return the new access. */
848 static struct access *
849 create_access_1 (tree base, HOST_WIDE_INT offset, HOST_WIDE_INT size)
851 vec<access_p> *v;
852 struct access *access;
853 void **slot;
855 access = (struct access *) pool_alloc (access_pool);
856 memset (access, 0, sizeof (struct access));
857 access->base = base;
858 access->offset = offset;
859 access->size = size;
861 slot = pointer_map_contains (base_access_vec, base);
862 if (slot)
863 v = (vec<access_p> *) *slot;
864 else
865 vec_alloc (v, 32);
867 v->safe_push (access);
869 *((vec<access_p> **)
870 pointer_map_insert (base_access_vec, base)) = v;
872 return access;
875 /* Create and insert access for EXPR. Return created access, or NULL if it is
876 not possible. */
878 static struct access *
879 create_access (tree expr, gimple stmt, bool write)
881 struct access *access;
882 HOST_WIDE_INT offset, size, max_size;
883 tree base = expr;
884 bool ptr, unscalarizable_region = false;
886 base = get_ref_base_and_extent (expr, &offset, &size, &max_size);
888 if (sra_mode == SRA_MODE_EARLY_IPA
889 && TREE_CODE (base) == MEM_REF)
891 base = get_ssa_base_param (TREE_OPERAND (base, 0));
892 if (!base)
893 return NULL;
894 ptr = true;
896 else
897 ptr = false;
899 if (!DECL_P (base) || !bitmap_bit_p (candidate_bitmap, DECL_UID (base)))
900 return NULL;
902 if (sra_mode == SRA_MODE_EARLY_IPA)
904 if (size < 0 || size != max_size)
906 disqualify_candidate (base, "Encountered a variable sized access.");
907 return NULL;
909 if (TREE_CODE (expr) == COMPONENT_REF
910 && DECL_BIT_FIELD (TREE_OPERAND (expr, 1)))
912 disqualify_candidate (base, "Encountered a bit-field access.");
913 return NULL;
915 gcc_checking_assert ((offset % BITS_PER_UNIT) == 0);
917 if (ptr)
918 mark_parm_dereference (base, offset + size, stmt);
920 else
922 if (size != max_size)
924 size = max_size;
925 unscalarizable_region = true;
927 if (size < 0)
929 disqualify_candidate (base, "Encountered an unconstrained access.");
930 return NULL;
934 access = create_access_1 (base, offset, size);
935 access->expr = expr;
936 access->type = TREE_TYPE (expr);
937 access->write = write;
938 access->grp_unscalarizable_region = unscalarizable_region;
939 access->stmt = stmt;
941 if (TREE_CODE (expr) == COMPONENT_REF
942 && DECL_NONADDRESSABLE_P (TREE_OPERAND (expr, 1)))
943 access->non_addressable = 1;
945 return access;
949 /* Return true iff TYPE is a RECORD_TYPE with fields that are either of gimple
950 register types or (recursively) records with only these two kinds of fields.
951 It also returns false if any of these records contains a bit-field. */
953 static bool
954 type_consists_of_records_p (tree type)
956 tree fld;
958 if (TREE_CODE (type) != RECORD_TYPE)
959 return false;
961 for (fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
962 if (TREE_CODE (fld) == FIELD_DECL)
964 tree ft = TREE_TYPE (fld);
966 if (DECL_BIT_FIELD (fld))
967 return false;
969 if (!is_gimple_reg_type (ft)
970 && !type_consists_of_records_p (ft))
971 return false;
974 return true;
977 /* Create total_scalarization accesses for all scalar type fields in DECL that
978 must be of a RECORD_TYPE conforming to type_consists_of_records_p. BASE
979 must be the top-most VAR_DECL representing the variable, OFFSET must be the
980 offset of DECL within BASE. REF must be the memory reference expression for
981 the given decl. */
983 static void
984 completely_scalarize_record (tree base, tree decl, HOST_WIDE_INT offset,
985 tree ref)
987 tree fld, decl_type = TREE_TYPE (decl);
989 for (fld = TYPE_FIELDS (decl_type); fld; fld = DECL_CHAIN (fld))
990 if (TREE_CODE (fld) == FIELD_DECL)
992 HOST_WIDE_INT pos = offset + int_bit_position (fld);
993 tree ft = TREE_TYPE (fld);
994 tree nref = build3 (COMPONENT_REF, TREE_TYPE (fld), ref, fld,
995 NULL_TREE);
997 if (is_gimple_reg_type (ft))
999 struct access *access;
1000 HOST_WIDE_INT size;
1002 size = tree_to_uhwi (DECL_SIZE (fld));
1003 access = create_access_1 (base, pos, size);
1004 access->expr = nref;
1005 access->type = ft;
1006 access->grp_total_scalarization = 1;
1007 /* Accesses for intraprocedural SRA can have their stmt NULL. */
1009 else
1010 completely_scalarize_record (base, fld, pos, nref);
1014 /* Create total_scalarization accesses for all scalar type fields in VAR and
1015 for VAR a a whole. VAR must be of a RECORD_TYPE conforming to
1016 type_consists_of_records_p. */
1018 static void
1019 completely_scalarize_var (tree var)
1021 HOST_WIDE_INT size = tree_to_uhwi (DECL_SIZE (var));
1022 struct access *access;
1024 access = create_access_1 (var, 0, size);
1025 access->expr = var;
1026 access->type = TREE_TYPE (var);
1027 access->grp_total_scalarization = 1;
1029 completely_scalarize_record (var, var, 0, var);
1032 /* Return true if REF has an VIEW_CONVERT_EXPR somewhere in it. */
1034 static inline bool
1035 contains_view_convert_expr_p (const_tree ref)
1037 while (handled_component_p (ref))
1039 if (TREE_CODE (ref) == VIEW_CONVERT_EXPR)
1040 return true;
1041 ref = TREE_OPERAND (ref, 0);
1044 return false;
1047 /* Search the given tree for a declaration by skipping handled components and
1048 exclude it from the candidates. */
1050 static void
1051 disqualify_base_of_expr (tree t, const char *reason)
1053 t = get_base_address (t);
1054 if (sra_mode == SRA_MODE_EARLY_IPA
1055 && TREE_CODE (t) == MEM_REF)
1056 t = get_ssa_base_param (TREE_OPERAND (t, 0));
1058 if (t && DECL_P (t))
1059 disqualify_candidate (t, reason);
1062 /* Scan expression EXPR and create access structures for all accesses to
1063 candidates for scalarization. Return the created access or NULL if none is
1064 created. */
1066 static struct access *
1067 build_access_from_expr_1 (tree expr, gimple stmt, bool write)
1069 struct access *ret = NULL;
1070 bool partial_ref;
1072 if (TREE_CODE (expr) == BIT_FIELD_REF
1073 || TREE_CODE (expr) == IMAGPART_EXPR
1074 || TREE_CODE (expr) == REALPART_EXPR)
1076 expr = TREE_OPERAND (expr, 0);
1077 partial_ref = true;
1079 else
1080 partial_ref = false;
1082 /* We need to dive through V_C_Es in order to get the size of its parameter
1083 and not the result type. Ada produces such statements. We are also
1084 capable of handling the topmost V_C_E but not any of those buried in other
1085 handled components. */
1086 if (TREE_CODE (expr) == VIEW_CONVERT_EXPR)
1087 expr = TREE_OPERAND (expr, 0);
1089 if (contains_view_convert_expr_p (expr))
1091 disqualify_base_of_expr (expr, "V_C_E under a different handled "
1092 "component.");
1093 return NULL;
1096 switch (TREE_CODE (expr))
1098 case MEM_REF:
1099 if (TREE_CODE (TREE_OPERAND (expr, 0)) != ADDR_EXPR
1100 && sra_mode != SRA_MODE_EARLY_IPA)
1101 return NULL;
1102 /* fall through */
1103 case VAR_DECL:
1104 case PARM_DECL:
1105 case RESULT_DECL:
1106 case COMPONENT_REF:
1107 case ARRAY_REF:
1108 case ARRAY_RANGE_REF:
1109 ret = create_access (expr, stmt, write);
1110 break;
1112 default:
1113 break;
1116 if (write && partial_ref && ret)
1117 ret->grp_partial_lhs = 1;
1119 return ret;
1122 /* Scan expression EXPR and create access structures for all accesses to
1123 candidates for scalarization. Return true if any access has been inserted.
1124 STMT must be the statement from which the expression is taken, WRITE must be
1125 true if the expression is a store and false otherwise. */
1127 static bool
1128 build_access_from_expr (tree expr, gimple stmt, bool write)
1130 struct access *access;
1132 access = build_access_from_expr_1 (expr, stmt, write);
1133 if (access)
1135 /* This means the aggregate is accesses as a whole in a way other than an
1136 assign statement and thus cannot be removed even if we had a scalar
1137 replacement for everything. */
1138 if (cannot_scalarize_away_bitmap)
1139 bitmap_set_bit (cannot_scalarize_away_bitmap, DECL_UID (access->base));
1140 return true;
1142 return false;
1145 /* Disqualify LHS and RHS for scalarization if STMT must end its basic block in
1146 modes in which it matters, return true iff they have been disqualified. RHS
1147 may be NULL, in that case ignore it. If we scalarize an aggregate in
1148 intra-SRA we may need to add statements after each statement. This is not
1149 possible if a statement unconditionally has to end the basic block. */
1150 static bool
1151 disqualify_ops_if_throwing_stmt (gimple stmt, tree lhs, tree rhs)
1153 if ((sra_mode == SRA_MODE_EARLY_INTRA || sra_mode == SRA_MODE_INTRA)
1154 && (stmt_can_throw_internal (stmt) || stmt_ends_bb_p (stmt)))
1156 disqualify_base_of_expr (lhs, "LHS of a throwing stmt.");
1157 if (rhs)
1158 disqualify_base_of_expr (rhs, "RHS of a throwing stmt.");
1159 return true;
1161 return false;
1164 /* Scan expressions occurring in STMT, create access structures for all accesses
1165 to candidates for scalarization and remove those candidates which occur in
1166 statements or expressions that prevent them from being split apart. Return
1167 true if any access has been inserted. */
1169 static bool
1170 build_accesses_from_assign (gimple stmt)
1172 tree lhs, rhs;
1173 struct access *lacc, *racc;
1175 if (!gimple_assign_single_p (stmt)
1176 /* Scope clobbers don't influence scalarization. */
1177 || gimple_clobber_p (stmt))
1178 return false;
1180 lhs = gimple_assign_lhs (stmt);
1181 rhs = gimple_assign_rhs1 (stmt);
1183 if (disqualify_ops_if_throwing_stmt (stmt, lhs, rhs))
1184 return false;
1186 racc = build_access_from_expr_1 (rhs, stmt, false);
1187 lacc = build_access_from_expr_1 (lhs, stmt, true);
1189 if (lacc)
1190 lacc->grp_assignment_write = 1;
1192 if (racc)
1194 racc->grp_assignment_read = 1;
1195 if (should_scalarize_away_bitmap && !gimple_has_volatile_ops (stmt)
1196 && !is_gimple_reg_type (racc->type))
1197 bitmap_set_bit (should_scalarize_away_bitmap, DECL_UID (racc->base));
1200 if (lacc && racc
1201 && (sra_mode == SRA_MODE_EARLY_INTRA || sra_mode == SRA_MODE_INTRA)
1202 && !lacc->grp_unscalarizable_region
1203 && !racc->grp_unscalarizable_region
1204 && AGGREGATE_TYPE_P (TREE_TYPE (lhs))
1205 && lacc->size == racc->size
1206 && useless_type_conversion_p (lacc->type, racc->type))
1208 struct assign_link *link;
1210 link = (struct assign_link *) pool_alloc (link_pool);
1211 memset (link, 0, sizeof (struct assign_link));
1213 link->lacc = lacc;
1214 link->racc = racc;
1216 add_link_to_rhs (racc, link);
1219 return lacc || racc;
1222 /* Callback of walk_stmt_load_store_addr_ops visit_addr used to determine
1223 GIMPLE_ASM operands with memory constrains which cannot be scalarized. */
1225 static bool
1226 asm_visit_addr (gimple, tree op, tree, void *)
1228 op = get_base_address (op);
1229 if (op
1230 && DECL_P (op))
1231 disqualify_candidate (op, "Non-scalarizable GIMPLE_ASM operand.");
1233 return false;
1236 /* Return true iff callsite CALL has at least as many actual arguments as there
1237 are formal parameters of the function currently processed by IPA-SRA and
1238 that their types match. */
1240 static inline bool
1241 callsite_arguments_match_p (gimple call)
1243 if (gimple_call_num_args (call) < (unsigned) func_param_count)
1244 return false;
1246 tree parm;
1247 int i;
1248 for (parm = DECL_ARGUMENTS (current_function_decl), i = 0;
1249 parm;
1250 parm = DECL_CHAIN (parm), i++)
1252 tree arg = gimple_call_arg (call, i);
1253 if (!useless_type_conversion_p (TREE_TYPE (parm), TREE_TYPE (arg)))
1254 return false;
1256 return true;
1259 /* Scan function and look for interesting expressions and create access
1260 structures for them. Return true iff any access is created. */
1262 static bool
1263 scan_function (void)
1265 basic_block bb;
1266 bool ret = false;
1268 FOR_EACH_BB_FN (bb, cfun)
1270 gimple_stmt_iterator gsi;
1271 for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
1273 gimple stmt = gsi_stmt (gsi);
1274 tree t;
1275 unsigned i;
1277 if (final_bbs && stmt_can_throw_external (stmt))
1278 bitmap_set_bit (final_bbs, bb->index);
1279 switch (gimple_code (stmt))
1281 case GIMPLE_RETURN:
1282 t = gimple_return_retval (stmt);
1283 if (t != NULL_TREE)
1284 ret |= build_access_from_expr (t, stmt, false);
1285 if (final_bbs)
1286 bitmap_set_bit (final_bbs, bb->index);
1287 break;
1289 case GIMPLE_ASSIGN:
1290 ret |= build_accesses_from_assign (stmt);
1291 break;
1293 case GIMPLE_CALL:
1294 for (i = 0; i < gimple_call_num_args (stmt); i++)
1295 ret |= build_access_from_expr (gimple_call_arg (stmt, i),
1296 stmt, false);
1298 if (sra_mode == SRA_MODE_EARLY_IPA)
1300 tree dest = gimple_call_fndecl (stmt);
1301 int flags = gimple_call_flags (stmt);
1303 if (dest)
1305 if (DECL_BUILT_IN_CLASS (dest) == BUILT_IN_NORMAL
1306 && DECL_FUNCTION_CODE (dest) == BUILT_IN_APPLY_ARGS)
1307 encountered_apply_args = true;
1308 if (recursive_call_p (current_function_decl, dest))
1310 encountered_recursive_call = true;
1311 if (!callsite_arguments_match_p (stmt))
1312 encountered_unchangable_recursive_call = true;
1316 if (final_bbs
1317 && (flags & (ECF_CONST | ECF_PURE)) == 0)
1318 bitmap_set_bit (final_bbs, bb->index);
1321 t = gimple_call_lhs (stmt);
1322 if (t && !disqualify_ops_if_throwing_stmt (stmt, t, NULL))
1323 ret |= build_access_from_expr (t, stmt, true);
1324 break;
1326 case GIMPLE_ASM:
1327 walk_stmt_load_store_addr_ops (stmt, NULL, NULL, NULL,
1328 asm_visit_addr);
1329 if (final_bbs)
1330 bitmap_set_bit (final_bbs, bb->index);
1332 for (i = 0; i < gimple_asm_ninputs (stmt); i++)
1334 t = TREE_VALUE (gimple_asm_input_op (stmt, i));
1335 ret |= build_access_from_expr (t, stmt, false);
1337 for (i = 0; i < gimple_asm_noutputs (stmt); i++)
1339 t = TREE_VALUE (gimple_asm_output_op (stmt, i));
1340 ret |= build_access_from_expr (t, stmt, true);
1342 break;
1344 default:
1345 break;
1350 return ret;
1353 /* Helper of QSORT function. There are pointers to accesses in the array. An
1354 access is considered smaller than another if it has smaller offset or if the
1355 offsets are the same but is size is bigger. */
1357 static int
1358 compare_access_positions (const void *a, const void *b)
1360 const access_p *fp1 = (const access_p *) a;
1361 const access_p *fp2 = (const access_p *) b;
1362 const access_p f1 = *fp1;
1363 const access_p f2 = *fp2;
1365 if (f1->offset != f2->offset)
1366 return f1->offset < f2->offset ? -1 : 1;
1368 if (f1->size == f2->size)
1370 if (f1->type == f2->type)
1371 return 0;
1372 /* Put any non-aggregate type before any aggregate type. */
1373 else if (!is_gimple_reg_type (f1->type)
1374 && is_gimple_reg_type (f2->type))
1375 return 1;
1376 else if (is_gimple_reg_type (f1->type)
1377 && !is_gimple_reg_type (f2->type))
1378 return -1;
1379 /* Put any complex or vector type before any other scalar type. */
1380 else if (TREE_CODE (f1->type) != COMPLEX_TYPE
1381 && TREE_CODE (f1->type) != VECTOR_TYPE
1382 && (TREE_CODE (f2->type) == COMPLEX_TYPE
1383 || TREE_CODE (f2->type) == VECTOR_TYPE))
1384 return 1;
1385 else if ((TREE_CODE (f1->type) == COMPLEX_TYPE
1386 || TREE_CODE (f1->type) == VECTOR_TYPE)
1387 && TREE_CODE (f2->type) != COMPLEX_TYPE
1388 && TREE_CODE (f2->type) != VECTOR_TYPE)
1389 return -1;
1390 /* Put the integral type with the bigger precision first. */
1391 else if (INTEGRAL_TYPE_P (f1->type)
1392 && INTEGRAL_TYPE_P (f2->type))
1393 return TYPE_PRECISION (f2->type) - TYPE_PRECISION (f1->type);
1394 /* Put any integral type with non-full precision last. */
1395 else if (INTEGRAL_TYPE_P (f1->type)
1396 && (TREE_INT_CST_LOW (TYPE_SIZE (f1->type))
1397 != TYPE_PRECISION (f1->type)))
1398 return 1;
1399 else if (INTEGRAL_TYPE_P (f2->type)
1400 && (TREE_INT_CST_LOW (TYPE_SIZE (f2->type))
1401 != TYPE_PRECISION (f2->type)))
1402 return -1;
1403 /* Stabilize the sort. */
1404 return TYPE_UID (f1->type) - TYPE_UID (f2->type);
1407 /* We want the bigger accesses first, thus the opposite operator in the next
1408 line: */
1409 return f1->size > f2->size ? -1 : 1;
1413 /* Append a name of the declaration to the name obstack. A helper function for
1414 make_fancy_name. */
1416 static void
1417 make_fancy_decl_name (tree decl)
1419 char buffer[32];
1421 tree name = DECL_NAME (decl);
1422 if (name)
1423 obstack_grow (&name_obstack, IDENTIFIER_POINTER (name),
1424 IDENTIFIER_LENGTH (name));
1425 else
1427 sprintf (buffer, "D%u", DECL_UID (decl));
1428 obstack_grow (&name_obstack, buffer, strlen (buffer));
1432 /* Helper for make_fancy_name. */
1434 static void
1435 make_fancy_name_1 (tree expr)
1437 char buffer[32];
1438 tree index;
1440 if (DECL_P (expr))
1442 make_fancy_decl_name (expr);
1443 return;
1446 switch (TREE_CODE (expr))
1448 case COMPONENT_REF:
1449 make_fancy_name_1 (TREE_OPERAND (expr, 0));
1450 obstack_1grow (&name_obstack, '$');
1451 make_fancy_decl_name (TREE_OPERAND (expr, 1));
1452 break;
1454 case ARRAY_REF:
1455 make_fancy_name_1 (TREE_OPERAND (expr, 0));
1456 obstack_1grow (&name_obstack, '$');
1457 /* Arrays with only one element may not have a constant as their
1458 index. */
1459 index = TREE_OPERAND (expr, 1);
1460 if (TREE_CODE (index) != INTEGER_CST)
1461 break;
1462 sprintf (buffer, HOST_WIDE_INT_PRINT_DEC, TREE_INT_CST_LOW (index));
1463 obstack_grow (&name_obstack, buffer, strlen (buffer));
1464 break;
1466 case ADDR_EXPR:
1467 make_fancy_name_1 (TREE_OPERAND (expr, 0));
1468 break;
1470 case MEM_REF:
1471 make_fancy_name_1 (TREE_OPERAND (expr, 0));
1472 if (!integer_zerop (TREE_OPERAND (expr, 1)))
1474 obstack_1grow (&name_obstack, '$');
1475 sprintf (buffer, HOST_WIDE_INT_PRINT_DEC,
1476 TREE_INT_CST_LOW (TREE_OPERAND (expr, 1)));
1477 obstack_grow (&name_obstack, buffer, strlen (buffer));
1479 break;
1481 case BIT_FIELD_REF:
1482 case REALPART_EXPR:
1483 case IMAGPART_EXPR:
1484 gcc_unreachable (); /* we treat these as scalars. */
1485 break;
1486 default:
1487 break;
1491 /* Create a human readable name for replacement variable of ACCESS. */
1493 static char *
1494 make_fancy_name (tree expr)
1496 make_fancy_name_1 (expr);
1497 obstack_1grow (&name_obstack, '\0');
1498 return XOBFINISH (&name_obstack, char *);
1501 /* Construct a MEM_REF that would reference a part of aggregate BASE of type
1502 EXP_TYPE at the given OFFSET. If BASE is something for which
1503 get_addr_base_and_unit_offset returns NULL, gsi must be non-NULL and is used
1504 to insert new statements either before or below the current one as specified
1505 by INSERT_AFTER. This function is not capable of handling bitfields.
1507 BASE must be either a declaration or a memory reference that has correct
1508 alignment ifformation embeded in it (e.g. a pre-existing one in SRA). */
1510 tree
1511 build_ref_for_offset (location_t loc, tree base, HOST_WIDE_INT offset,
1512 tree exp_type, gimple_stmt_iterator *gsi,
1513 bool insert_after)
1515 tree prev_base = base;
1516 tree off;
1517 tree mem_ref;
1518 HOST_WIDE_INT base_offset;
1519 unsigned HOST_WIDE_INT misalign;
1520 unsigned int align;
1522 gcc_checking_assert (offset % BITS_PER_UNIT == 0);
1523 get_object_alignment_1 (base, &align, &misalign);
1524 base = get_addr_base_and_unit_offset (base, &base_offset);
1526 /* get_addr_base_and_unit_offset returns NULL for references with a variable
1527 offset such as array[var_index]. */
1528 if (!base)
1530 gimple stmt;
1531 tree tmp, addr;
1533 gcc_checking_assert (gsi);
1534 tmp = make_ssa_name (build_pointer_type (TREE_TYPE (prev_base)), NULL);
1535 addr = build_fold_addr_expr (unshare_expr (prev_base));
1536 STRIP_USELESS_TYPE_CONVERSION (addr);
1537 stmt = gimple_build_assign (tmp, addr);
1538 gimple_set_location (stmt, loc);
1539 if (insert_after)
1540 gsi_insert_after (gsi, stmt, GSI_NEW_STMT);
1541 else
1542 gsi_insert_before (gsi, stmt, GSI_SAME_STMT);
1544 off = build_int_cst (reference_alias_ptr_type (prev_base),
1545 offset / BITS_PER_UNIT);
1546 base = tmp;
1548 else if (TREE_CODE (base) == MEM_REF)
1550 off = build_int_cst (TREE_TYPE (TREE_OPERAND (base, 1)),
1551 base_offset + offset / BITS_PER_UNIT);
1552 off = int_const_binop (PLUS_EXPR, TREE_OPERAND (base, 1), off);
1553 base = unshare_expr (TREE_OPERAND (base, 0));
1555 else
1557 off = build_int_cst (reference_alias_ptr_type (base),
1558 base_offset + offset / BITS_PER_UNIT);
1559 base = build_fold_addr_expr (unshare_expr (base));
1562 misalign = (misalign + offset) & (align - 1);
1563 if (misalign != 0)
1564 align = (misalign & -misalign);
1565 if (align < TYPE_ALIGN (exp_type))
1566 exp_type = build_aligned_type (exp_type, align);
1568 mem_ref = fold_build2_loc (loc, MEM_REF, exp_type, base, off);
1569 if (TREE_THIS_VOLATILE (prev_base))
1570 TREE_THIS_VOLATILE (mem_ref) = 1;
1571 if (TREE_SIDE_EFFECTS (prev_base))
1572 TREE_SIDE_EFFECTS (mem_ref) = 1;
1573 return mem_ref;
1576 /* Construct a memory reference to a part of an aggregate BASE at the given
1577 OFFSET and of the same type as MODEL. In case this is a reference to a
1578 bit-field, the function will replicate the last component_ref of model's
1579 expr to access it. GSI and INSERT_AFTER have the same meaning as in
1580 build_ref_for_offset. */
1582 static tree
1583 build_ref_for_model (location_t loc, tree base, HOST_WIDE_INT offset,
1584 struct access *model, gimple_stmt_iterator *gsi,
1585 bool insert_after)
1587 if (TREE_CODE (model->expr) == COMPONENT_REF
1588 && DECL_BIT_FIELD (TREE_OPERAND (model->expr, 1)))
1590 /* This access represents a bit-field. */
1591 tree t, exp_type, fld = TREE_OPERAND (model->expr, 1);
1593 offset -= int_bit_position (fld);
1594 exp_type = TREE_TYPE (TREE_OPERAND (model->expr, 0));
1595 t = build_ref_for_offset (loc, base, offset, exp_type, gsi, insert_after);
1596 return fold_build3_loc (loc, COMPONENT_REF, TREE_TYPE (fld), t, fld,
1597 NULL_TREE);
1599 else
1600 return build_ref_for_offset (loc, base, offset, model->type,
1601 gsi, insert_after);
1604 /* Attempt to build a memory reference that we could but into a gimple
1605 debug_bind statement. Similar to build_ref_for_model but punts if it has to
1606 create statements and return s NULL instead. This function also ignores
1607 alignment issues and so its results should never end up in non-debug
1608 statements. */
1610 static tree
1611 build_debug_ref_for_model (location_t loc, tree base, HOST_WIDE_INT offset,
1612 struct access *model)
1614 HOST_WIDE_INT base_offset;
1615 tree off;
1617 if (TREE_CODE (model->expr) == COMPONENT_REF
1618 && DECL_BIT_FIELD (TREE_OPERAND (model->expr, 1)))
1619 return NULL_TREE;
1621 base = get_addr_base_and_unit_offset (base, &base_offset);
1622 if (!base)
1623 return NULL_TREE;
1624 if (TREE_CODE (base) == MEM_REF)
1626 off = build_int_cst (TREE_TYPE (TREE_OPERAND (base, 1)),
1627 base_offset + offset / BITS_PER_UNIT);
1628 off = int_const_binop (PLUS_EXPR, TREE_OPERAND (base, 1), off);
1629 base = unshare_expr (TREE_OPERAND (base, 0));
1631 else
1633 off = build_int_cst (reference_alias_ptr_type (base),
1634 base_offset + offset / BITS_PER_UNIT);
1635 base = build_fold_addr_expr (unshare_expr (base));
1638 return fold_build2_loc (loc, MEM_REF, model->type, base, off);
1641 /* Construct a memory reference consisting of component_refs and array_refs to
1642 a part of an aggregate *RES (which is of type TYPE). The requested part
1643 should have type EXP_TYPE at be the given OFFSET. This function might not
1644 succeed, it returns true when it does and only then *RES points to something
1645 meaningful. This function should be used only to build expressions that we
1646 might need to present to user (e.g. in warnings). In all other situations,
1647 build_ref_for_model or build_ref_for_offset should be used instead. */
1649 static bool
1650 build_user_friendly_ref_for_offset (tree *res, tree type, HOST_WIDE_INT offset,
1651 tree exp_type)
1653 while (1)
1655 tree fld;
1656 tree tr_size, index, minidx;
1657 HOST_WIDE_INT el_size;
1659 if (offset == 0 && exp_type
1660 && types_compatible_p (exp_type, type))
1661 return true;
1663 switch (TREE_CODE (type))
1665 case UNION_TYPE:
1666 case QUAL_UNION_TYPE:
1667 case RECORD_TYPE:
1668 for (fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
1670 HOST_WIDE_INT pos, size;
1671 tree tr_pos, expr, *expr_ptr;
1673 if (TREE_CODE (fld) != FIELD_DECL)
1674 continue;
1676 tr_pos = bit_position (fld);
1677 if (!tr_pos || !tree_fits_uhwi_p (tr_pos))
1678 continue;
1679 pos = tree_to_uhwi (tr_pos);
1680 gcc_assert (TREE_CODE (type) == RECORD_TYPE || pos == 0);
1681 tr_size = DECL_SIZE (fld);
1682 if (!tr_size || !tree_fits_uhwi_p (tr_size))
1683 continue;
1684 size = tree_to_uhwi (tr_size);
1685 if (size == 0)
1687 if (pos != offset)
1688 continue;
1690 else if (pos > offset || (pos + size) <= offset)
1691 continue;
1693 expr = build3 (COMPONENT_REF, TREE_TYPE (fld), *res, fld,
1694 NULL_TREE);
1695 expr_ptr = &expr;
1696 if (build_user_friendly_ref_for_offset (expr_ptr, TREE_TYPE (fld),
1697 offset - pos, exp_type))
1699 *res = expr;
1700 return true;
1703 return false;
1705 case ARRAY_TYPE:
1706 tr_size = TYPE_SIZE (TREE_TYPE (type));
1707 if (!tr_size || !tree_fits_uhwi_p (tr_size))
1708 return false;
1709 el_size = tree_to_uhwi (tr_size);
1711 minidx = TYPE_MIN_VALUE (TYPE_DOMAIN (type));
1712 if (TREE_CODE (minidx) != INTEGER_CST || el_size == 0)
1713 return false;
1714 index = build_int_cst (TYPE_DOMAIN (type), offset / el_size);
1715 if (!integer_zerop (minidx))
1716 index = int_const_binop (PLUS_EXPR, index, minidx);
1717 *res = build4 (ARRAY_REF, TREE_TYPE (type), *res, index,
1718 NULL_TREE, NULL_TREE);
1719 offset = offset % el_size;
1720 type = TREE_TYPE (type);
1721 break;
1723 default:
1724 if (offset != 0)
1725 return false;
1727 if (exp_type)
1728 return false;
1729 else
1730 return true;
1735 /* Return true iff TYPE is stdarg va_list type. */
1737 static inline bool
1738 is_va_list_type (tree type)
1740 return TYPE_MAIN_VARIANT (type) == TYPE_MAIN_VARIANT (va_list_type_node);
1743 /* Print message to dump file why a variable was rejected. */
1745 static void
1746 reject (tree var, const char *msg)
1748 if (dump_file && (dump_flags & TDF_DETAILS))
1750 fprintf (dump_file, "Rejected (%d): %s: ", DECL_UID (var), msg);
1751 print_generic_expr (dump_file, var, 0);
1752 fprintf (dump_file, "\n");
1756 /* Return true if VAR is a candidate for SRA. */
1758 static bool
1759 maybe_add_sra_candidate (tree var)
1761 tree type = TREE_TYPE (var);
1762 const char *msg;
1763 tree_node **slot;
1765 if (!AGGREGATE_TYPE_P (type))
1767 reject (var, "not aggregate");
1768 return false;
1770 if (needs_to_live_in_memory (var))
1772 reject (var, "needs to live in memory");
1773 return false;
1775 if (TREE_THIS_VOLATILE (var))
1777 reject (var, "is volatile");
1778 return false;
1780 if (!COMPLETE_TYPE_P (type))
1782 reject (var, "has incomplete type");
1783 return false;
1785 if (!tree_fits_uhwi_p (TYPE_SIZE (type)))
1787 reject (var, "type size not fixed");
1788 return false;
1790 if (tree_to_uhwi (TYPE_SIZE (type)) == 0)
1792 reject (var, "type size is zero");
1793 return false;
1795 if (type_internals_preclude_sra_p (type, &msg))
1797 reject (var, msg);
1798 return false;
1800 if (/* Fix for PR 41089. tree-stdarg.c needs to have va_lists intact but
1801 we also want to schedule it rather late. Thus we ignore it in
1802 the early pass. */
1803 (sra_mode == SRA_MODE_EARLY_INTRA
1804 && is_va_list_type (type)))
1806 reject (var, "is va_list");
1807 return false;
1810 bitmap_set_bit (candidate_bitmap, DECL_UID (var));
1811 slot = candidates.find_slot_with_hash (var, DECL_UID (var), INSERT);
1812 *slot = var;
1814 if (dump_file && (dump_flags & TDF_DETAILS))
1816 fprintf (dump_file, "Candidate (%d): ", DECL_UID (var));
1817 print_generic_expr (dump_file, var, 0);
1818 fprintf (dump_file, "\n");
1821 return true;
1824 /* The very first phase of intraprocedural SRA. It marks in candidate_bitmap
1825 those with type which is suitable for scalarization. */
1827 static bool
1828 find_var_candidates (void)
1830 tree var, parm;
1831 unsigned int i;
1832 bool ret = false;
1834 for (parm = DECL_ARGUMENTS (current_function_decl);
1835 parm;
1836 parm = DECL_CHAIN (parm))
1837 ret |= maybe_add_sra_candidate (parm);
1839 FOR_EACH_LOCAL_DECL (cfun, i, var)
1841 if (TREE_CODE (var) != VAR_DECL)
1842 continue;
1844 ret |= maybe_add_sra_candidate (var);
1847 return ret;
1850 /* Sort all accesses for the given variable, check for partial overlaps and
1851 return NULL if there are any. If there are none, pick a representative for
1852 each combination of offset and size and create a linked list out of them.
1853 Return the pointer to the first representative and make sure it is the first
1854 one in the vector of accesses. */
1856 static struct access *
1857 sort_and_splice_var_accesses (tree var)
1859 int i, j, access_count;
1860 struct access *res, **prev_acc_ptr = &res;
1861 vec<access_p> *access_vec;
1862 bool first = true;
1863 HOST_WIDE_INT low = -1, high = 0;
1865 access_vec = get_base_access_vector (var);
1866 if (!access_vec)
1867 return NULL;
1868 access_count = access_vec->length ();
1870 /* Sort by <OFFSET, SIZE>. */
1871 access_vec->qsort (compare_access_positions);
1873 i = 0;
1874 while (i < access_count)
1876 struct access *access = (*access_vec)[i];
1877 bool grp_write = access->write;
1878 bool grp_read = !access->write;
1879 bool grp_scalar_write = access->write
1880 && is_gimple_reg_type (access->type);
1881 bool grp_scalar_read = !access->write
1882 && is_gimple_reg_type (access->type);
1883 bool grp_assignment_read = access->grp_assignment_read;
1884 bool grp_assignment_write = access->grp_assignment_write;
1885 bool multiple_scalar_reads = false;
1886 bool total_scalarization = access->grp_total_scalarization;
1887 bool grp_partial_lhs = access->grp_partial_lhs;
1888 bool first_scalar = is_gimple_reg_type (access->type);
1889 bool unscalarizable_region = access->grp_unscalarizable_region;
1891 if (first || access->offset >= high)
1893 first = false;
1894 low = access->offset;
1895 high = access->offset + access->size;
1897 else if (access->offset > low && access->offset + access->size > high)
1898 return NULL;
1899 else
1900 gcc_assert (access->offset >= low
1901 && access->offset + access->size <= high);
1903 j = i + 1;
1904 while (j < access_count)
1906 struct access *ac2 = (*access_vec)[j];
1907 if (ac2->offset != access->offset || ac2->size != access->size)
1908 break;
1909 if (ac2->write)
1911 grp_write = true;
1912 grp_scalar_write = (grp_scalar_write
1913 || is_gimple_reg_type (ac2->type));
1915 else
1917 grp_read = true;
1918 if (is_gimple_reg_type (ac2->type))
1920 if (grp_scalar_read)
1921 multiple_scalar_reads = true;
1922 else
1923 grp_scalar_read = true;
1926 grp_assignment_read |= ac2->grp_assignment_read;
1927 grp_assignment_write |= ac2->grp_assignment_write;
1928 grp_partial_lhs |= ac2->grp_partial_lhs;
1929 unscalarizable_region |= ac2->grp_unscalarizable_region;
1930 total_scalarization |= ac2->grp_total_scalarization;
1931 relink_to_new_repr (access, ac2);
1933 /* If there are both aggregate-type and scalar-type accesses with
1934 this combination of size and offset, the comparison function
1935 should have put the scalars first. */
1936 gcc_assert (first_scalar || !is_gimple_reg_type (ac2->type));
1937 ac2->group_representative = access;
1938 j++;
1941 i = j;
1943 access->group_representative = access;
1944 access->grp_write = grp_write;
1945 access->grp_read = grp_read;
1946 access->grp_scalar_read = grp_scalar_read;
1947 access->grp_scalar_write = grp_scalar_write;
1948 access->grp_assignment_read = grp_assignment_read;
1949 access->grp_assignment_write = grp_assignment_write;
1950 access->grp_hint = multiple_scalar_reads || total_scalarization;
1951 access->grp_total_scalarization = total_scalarization;
1952 access->grp_partial_lhs = grp_partial_lhs;
1953 access->grp_unscalarizable_region = unscalarizable_region;
1954 if (access->first_link)
1955 add_access_to_work_queue (access);
1957 *prev_acc_ptr = access;
1958 prev_acc_ptr = &access->next_grp;
1961 gcc_assert (res == (*access_vec)[0]);
1962 return res;
1965 /* Create a variable for the given ACCESS which determines the type, name and a
1966 few other properties. Return the variable declaration and store it also to
1967 ACCESS->replacement. */
1969 static tree
1970 create_access_replacement (struct access *access)
1972 tree repl;
1974 if (access->grp_to_be_debug_replaced)
1976 repl = create_tmp_var_raw (access->type, NULL);
1977 DECL_CONTEXT (repl) = current_function_decl;
1979 else
1980 repl = create_tmp_var (access->type, "SR");
1981 if (TREE_CODE (access->type) == COMPLEX_TYPE
1982 || TREE_CODE (access->type) == VECTOR_TYPE)
1984 if (!access->grp_partial_lhs)
1985 DECL_GIMPLE_REG_P (repl) = 1;
1987 else if (access->grp_partial_lhs
1988 && is_gimple_reg_type (access->type))
1989 TREE_ADDRESSABLE (repl) = 1;
1991 DECL_SOURCE_LOCATION (repl) = DECL_SOURCE_LOCATION (access->base);
1992 DECL_ARTIFICIAL (repl) = 1;
1993 DECL_IGNORED_P (repl) = DECL_IGNORED_P (access->base);
1995 if (DECL_NAME (access->base)
1996 && !DECL_IGNORED_P (access->base)
1997 && !DECL_ARTIFICIAL (access->base))
1999 char *pretty_name = make_fancy_name (access->expr);
2000 tree debug_expr = unshare_expr_without_location (access->expr), d;
2001 bool fail = false;
2003 DECL_NAME (repl) = get_identifier (pretty_name);
2004 obstack_free (&name_obstack, pretty_name);
2006 /* Get rid of any SSA_NAMEs embedded in debug_expr,
2007 as DECL_DEBUG_EXPR isn't considered when looking for still
2008 used SSA_NAMEs and thus they could be freed. All debug info
2009 generation cares is whether something is constant or variable
2010 and that get_ref_base_and_extent works properly on the
2011 expression. It cannot handle accesses at a non-constant offset
2012 though, so just give up in those cases. */
2013 for (d = debug_expr;
2014 !fail && (handled_component_p (d) || TREE_CODE (d) == MEM_REF);
2015 d = TREE_OPERAND (d, 0))
2016 switch (TREE_CODE (d))
2018 case ARRAY_REF:
2019 case ARRAY_RANGE_REF:
2020 if (TREE_OPERAND (d, 1)
2021 && TREE_CODE (TREE_OPERAND (d, 1)) != INTEGER_CST)
2022 fail = true;
2023 if (TREE_OPERAND (d, 3)
2024 && TREE_CODE (TREE_OPERAND (d, 3)) != INTEGER_CST)
2025 fail = true;
2026 /* FALLTHRU */
2027 case COMPONENT_REF:
2028 if (TREE_OPERAND (d, 2)
2029 && TREE_CODE (TREE_OPERAND (d, 2)) != INTEGER_CST)
2030 fail = true;
2031 break;
2032 case MEM_REF:
2033 if (TREE_CODE (TREE_OPERAND (d, 0)) != ADDR_EXPR)
2034 fail = true;
2035 else
2036 d = TREE_OPERAND (d, 0);
2037 break;
2038 default:
2039 break;
2041 if (!fail)
2043 SET_DECL_DEBUG_EXPR (repl, debug_expr);
2044 DECL_HAS_DEBUG_EXPR_P (repl) = 1;
2046 if (access->grp_no_warning)
2047 TREE_NO_WARNING (repl) = 1;
2048 else
2049 TREE_NO_WARNING (repl) = TREE_NO_WARNING (access->base);
2051 else
2052 TREE_NO_WARNING (repl) = 1;
2054 if (dump_file)
2056 if (access->grp_to_be_debug_replaced)
2058 fprintf (dump_file, "Created a debug-only replacement for ");
2059 print_generic_expr (dump_file, access->base, 0);
2060 fprintf (dump_file, " offset: %u, size: %u\n",
2061 (unsigned) access->offset, (unsigned) access->size);
2063 else
2065 fprintf (dump_file, "Created a replacement for ");
2066 print_generic_expr (dump_file, access->base, 0);
2067 fprintf (dump_file, " offset: %u, size: %u: ",
2068 (unsigned) access->offset, (unsigned) access->size);
2069 print_generic_expr (dump_file, repl, 0);
2070 fprintf (dump_file, "\n");
2073 sra_stats.replacements++;
2075 return repl;
2078 /* Return ACCESS scalar replacement, create it if it does not exist yet. */
2080 static inline tree
2081 get_access_replacement (struct access *access)
2083 gcc_checking_assert (access->replacement_decl);
2084 return access->replacement_decl;
2088 /* Build a subtree of accesses rooted in *ACCESS, and move the pointer in the
2089 linked list along the way. Stop when *ACCESS is NULL or the access pointed
2090 to it is not "within" the root. Return false iff some accesses partially
2091 overlap. */
2093 static bool
2094 build_access_subtree (struct access **access)
2096 struct access *root = *access, *last_child = NULL;
2097 HOST_WIDE_INT limit = root->offset + root->size;
2099 *access = (*access)->next_grp;
2100 while (*access && (*access)->offset + (*access)->size <= limit)
2102 if (!last_child)
2103 root->first_child = *access;
2104 else
2105 last_child->next_sibling = *access;
2106 last_child = *access;
2108 if (!build_access_subtree (access))
2109 return false;
2112 if (*access && (*access)->offset < limit)
2113 return false;
2115 return true;
2118 /* Build a tree of access representatives, ACCESS is the pointer to the first
2119 one, others are linked in a list by the next_grp field. Return false iff
2120 some accesses partially overlap. */
2122 static bool
2123 build_access_trees (struct access *access)
2125 while (access)
2127 struct access *root = access;
2129 if (!build_access_subtree (&access))
2130 return false;
2131 root->next_grp = access;
2133 return true;
2136 /* Return true if expr contains some ARRAY_REFs into a variable bounded
2137 array. */
2139 static bool
2140 expr_with_var_bounded_array_refs_p (tree expr)
2142 while (handled_component_p (expr))
2144 if (TREE_CODE (expr) == ARRAY_REF
2145 && !tree_fits_shwi_p (array_ref_low_bound (expr)))
2146 return true;
2147 expr = TREE_OPERAND (expr, 0);
2149 return false;
2152 /* Analyze the subtree of accesses rooted in ROOT, scheduling replacements when
2153 both seeming beneficial and when ALLOW_REPLACEMENTS allows it. Also set all
2154 sorts of access flags appropriately along the way, notably always set
2155 grp_read and grp_assign_read according to MARK_READ and grp_write when
2156 MARK_WRITE is true.
2158 Creating a replacement for a scalar access is considered beneficial if its
2159 grp_hint is set (this means we are either attempting total scalarization or
2160 there is more than one direct read access) or according to the following
2161 table:
2163 Access written to through a scalar type (once or more times)
2165 | Written to in an assignment statement
2167 | | Access read as scalar _once_
2168 | | |
2169 | | | Read in an assignment statement
2170 | | | |
2171 | | | | Scalarize Comment
2172 -----------------------------------------------------------------------------
2173 0 0 0 0 No access for the scalar
2174 0 0 0 1 No access for the scalar
2175 0 0 1 0 No Single read - won't help
2176 0 0 1 1 No The same case
2177 0 1 0 0 No access for the scalar
2178 0 1 0 1 No access for the scalar
2179 0 1 1 0 Yes s = *g; return s.i;
2180 0 1 1 1 Yes The same case as above
2181 1 0 0 0 No Won't help
2182 1 0 0 1 Yes s.i = 1; *g = s;
2183 1 0 1 0 Yes s.i = 5; g = s.i;
2184 1 0 1 1 Yes The same case as above
2185 1 1 0 0 No Won't help.
2186 1 1 0 1 Yes s.i = 1; *g = s;
2187 1 1 1 0 Yes s = *g; return s.i;
2188 1 1 1 1 Yes Any of the above yeses */
2190 static bool
2191 analyze_access_subtree (struct access *root, struct access *parent,
2192 bool allow_replacements)
2194 struct access *child;
2195 HOST_WIDE_INT limit = root->offset + root->size;
2196 HOST_WIDE_INT covered_to = root->offset;
2197 bool scalar = is_gimple_reg_type (root->type);
2198 bool hole = false, sth_created = false;
2200 if (parent)
2202 if (parent->grp_read)
2203 root->grp_read = 1;
2204 if (parent->grp_assignment_read)
2205 root->grp_assignment_read = 1;
2206 if (parent->grp_write)
2207 root->grp_write = 1;
2208 if (parent->grp_assignment_write)
2209 root->grp_assignment_write = 1;
2210 if (parent->grp_total_scalarization)
2211 root->grp_total_scalarization = 1;
2214 if (root->grp_unscalarizable_region)
2215 allow_replacements = false;
2217 if (allow_replacements && expr_with_var_bounded_array_refs_p (root->expr))
2218 allow_replacements = false;
2220 for (child = root->first_child; child; child = child->next_sibling)
2222 hole |= covered_to < child->offset;
2223 sth_created |= analyze_access_subtree (child, root,
2224 allow_replacements && !scalar);
2226 root->grp_unscalarized_data |= child->grp_unscalarized_data;
2227 root->grp_total_scalarization &= child->grp_total_scalarization;
2228 if (child->grp_covered)
2229 covered_to += child->size;
2230 else
2231 hole = true;
2234 if (allow_replacements && scalar && !root->first_child
2235 && (root->grp_hint
2236 || ((root->grp_scalar_read || root->grp_assignment_read)
2237 && (root->grp_scalar_write || root->grp_assignment_write))))
2239 /* Always create access replacements that cover the whole access.
2240 For integral types this means the precision has to match.
2241 Avoid assumptions based on the integral type kind, too. */
2242 if (INTEGRAL_TYPE_P (root->type)
2243 && (TREE_CODE (root->type) != INTEGER_TYPE
2244 || TYPE_PRECISION (root->type) != root->size)
2245 /* But leave bitfield accesses alone. */
2246 && (TREE_CODE (root->expr) != COMPONENT_REF
2247 || !DECL_BIT_FIELD (TREE_OPERAND (root->expr, 1))))
2249 tree rt = root->type;
2250 gcc_assert ((root->offset % BITS_PER_UNIT) == 0
2251 && (root->size % BITS_PER_UNIT) == 0);
2252 root->type = build_nonstandard_integer_type (root->size,
2253 TYPE_UNSIGNED (rt));
2254 root->expr = build_ref_for_offset (UNKNOWN_LOCATION,
2255 root->base, root->offset,
2256 root->type, NULL, false);
2258 if (dump_file && (dump_flags & TDF_DETAILS))
2260 fprintf (dump_file, "Changing the type of a replacement for ");
2261 print_generic_expr (dump_file, root->base, 0);
2262 fprintf (dump_file, " offset: %u, size: %u ",
2263 (unsigned) root->offset, (unsigned) root->size);
2264 fprintf (dump_file, " to an integer.\n");
2268 root->grp_to_be_replaced = 1;
2269 root->replacement_decl = create_access_replacement (root);
2270 sth_created = true;
2271 hole = false;
2273 else
2275 if (allow_replacements
2276 && scalar && !root->first_child
2277 && (root->grp_scalar_write || root->grp_assignment_write)
2278 && !bitmap_bit_p (cannot_scalarize_away_bitmap,
2279 DECL_UID (root->base)))
2281 gcc_checking_assert (!root->grp_scalar_read
2282 && !root->grp_assignment_read);
2283 sth_created = true;
2284 if (MAY_HAVE_DEBUG_STMTS)
2286 root->grp_to_be_debug_replaced = 1;
2287 root->replacement_decl = create_access_replacement (root);
2291 if (covered_to < limit)
2292 hole = true;
2293 if (scalar)
2294 root->grp_total_scalarization = 0;
2297 if (!hole || root->grp_total_scalarization)
2298 root->grp_covered = 1;
2299 else if (root->grp_write || TREE_CODE (root->base) == PARM_DECL)
2300 root->grp_unscalarized_data = 1; /* not covered and written to */
2301 return sth_created;
2304 /* Analyze all access trees linked by next_grp by the means of
2305 analyze_access_subtree. */
2306 static bool
2307 analyze_access_trees (struct access *access)
2309 bool ret = false;
2311 while (access)
2313 if (analyze_access_subtree (access, NULL, true))
2314 ret = true;
2315 access = access->next_grp;
2318 return ret;
2321 /* Return true iff a potential new child of LACC at offset OFFSET and with size
2322 SIZE would conflict with an already existing one. If exactly such a child
2323 already exists in LACC, store a pointer to it in EXACT_MATCH. */
2325 static bool
2326 child_would_conflict_in_lacc (struct access *lacc, HOST_WIDE_INT norm_offset,
2327 HOST_WIDE_INT size, struct access **exact_match)
2329 struct access *child;
2331 for (child = lacc->first_child; child; child = child->next_sibling)
2333 if (child->offset == norm_offset && child->size == size)
2335 *exact_match = child;
2336 return true;
2339 if (child->offset < norm_offset + size
2340 && child->offset + child->size > norm_offset)
2341 return true;
2344 return false;
2347 /* Create a new child access of PARENT, with all properties just like MODEL
2348 except for its offset and with its grp_write false and grp_read true.
2349 Return the new access or NULL if it cannot be created. Note that this access
2350 is created long after all splicing and sorting, it's not located in any
2351 access vector and is automatically a representative of its group. */
2353 static struct access *
2354 create_artificial_child_access (struct access *parent, struct access *model,
2355 HOST_WIDE_INT new_offset)
2357 struct access *access;
2358 struct access **child;
2359 tree expr = parent->base;
2361 gcc_assert (!model->grp_unscalarizable_region);
2363 access = (struct access *) pool_alloc (access_pool);
2364 memset (access, 0, sizeof (struct access));
2365 if (!build_user_friendly_ref_for_offset (&expr, TREE_TYPE (expr), new_offset,
2366 model->type))
2368 access->grp_no_warning = true;
2369 expr = build_ref_for_model (EXPR_LOCATION (parent->base), parent->base,
2370 new_offset, model, NULL, false);
2373 access->base = parent->base;
2374 access->expr = expr;
2375 access->offset = new_offset;
2376 access->size = model->size;
2377 access->type = model->type;
2378 access->grp_write = true;
2379 access->grp_read = false;
2381 child = &parent->first_child;
2382 while (*child && (*child)->offset < new_offset)
2383 child = &(*child)->next_sibling;
2385 access->next_sibling = *child;
2386 *child = access;
2388 return access;
2392 /* Propagate all subaccesses of RACC across an assignment link to LACC. Return
2393 true if any new subaccess was created. Additionally, if RACC is a scalar
2394 access but LACC is not, change the type of the latter, if possible. */
2396 static bool
2397 propagate_subaccesses_across_link (struct access *lacc, struct access *racc)
2399 struct access *rchild;
2400 HOST_WIDE_INT norm_delta = lacc->offset - racc->offset;
2401 bool ret = false;
2403 if (is_gimple_reg_type (lacc->type)
2404 || lacc->grp_unscalarizable_region
2405 || racc->grp_unscalarizable_region)
2406 return false;
2408 if (is_gimple_reg_type (racc->type))
2410 if (!lacc->first_child && !racc->first_child)
2412 tree t = lacc->base;
2414 lacc->type = racc->type;
2415 if (build_user_friendly_ref_for_offset (&t, TREE_TYPE (t),
2416 lacc->offset, racc->type))
2417 lacc->expr = t;
2418 else
2420 lacc->expr = build_ref_for_model (EXPR_LOCATION (lacc->base),
2421 lacc->base, lacc->offset,
2422 racc, NULL, false);
2423 lacc->grp_no_warning = true;
2426 return false;
2429 for (rchild = racc->first_child; rchild; rchild = rchild->next_sibling)
2431 struct access *new_acc = NULL;
2432 HOST_WIDE_INT norm_offset = rchild->offset + norm_delta;
2434 if (rchild->grp_unscalarizable_region)
2435 continue;
2437 if (child_would_conflict_in_lacc (lacc, norm_offset, rchild->size,
2438 &new_acc))
2440 if (new_acc)
2442 rchild->grp_hint = 1;
2443 new_acc->grp_hint |= new_acc->grp_read;
2444 if (rchild->first_child)
2445 ret |= propagate_subaccesses_across_link (new_acc, rchild);
2447 continue;
2450 rchild->grp_hint = 1;
2451 new_acc = create_artificial_child_access (lacc, rchild, norm_offset);
2452 if (new_acc)
2454 ret = true;
2455 if (racc->first_child)
2456 propagate_subaccesses_across_link (new_acc, rchild);
2460 return ret;
2463 /* Propagate all subaccesses across assignment links. */
2465 static void
2466 propagate_all_subaccesses (void)
2468 while (work_queue_head)
2470 struct access *racc = pop_access_from_work_queue ();
2471 struct assign_link *link;
2473 gcc_assert (racc->first_link);
2475 for (link = racc->first_link; link; link = link->next)
2477 struct access *lacc = link->lacc;
2479 if (!bitmap_bit_p (candidate_bitmap, DECL_UID (lacc->base)))
2480 continue;
2481 lacc = lacc->group_representative;
2482 if (propagate_subaccesses_across_link (lacc, racc)
2483 && lacc->first_link)
2484 add_access_to_work_queue (lacc);
2489 /* Go through all accesses collected throughout the (intraprocedural) analysis
2490 stage, exclude overlapping ones, identify representatives and build trees
2491 out of them, making decisions about scalarization on the way. Return true
2492 iff there are any to-be-scalarized variables after this stage. */
2494 static bool
2495 analyze_all_variable_accesses (void)
2497 int res = 0;
2498 bitmap tmp = BITMAP_ALLOC (NULL);
2499 bitmap_iterator bi;
2500 unsigned i, max_total_scalarization_size;
2502 max_total_scalarization_size = UNITS_PER_WORD * BITS_PER_UNIT
2503 * MOVE_RATIO (optimize_function_for_speed_p (cfun));
2505 EXECUTE_IF_SET_IN_BITMAP (candidate_bitmap, 0, i, bi)
2506 if (bitmap_bit_p (should_scalarize_away_bitmap, i)
2507 && !bitmap_bit_p (cannot_scalarize_away_bitmap, i))
2509 tree var = candidate (i);
2511 if (TREE_CODE (var) == VAR_DECL
2512 && type_consists_of_records_p (TREE_TYPE (var)))
2514 if (tree_to_uhwi (TYPE_SIZE (TREE_TYPE (var)))
2515 <= max_total_scalarization_size)
2517 completely_scalarize_var (var);
2518 if (dump_file && (dump_flags & TDF_DETAILS))
2520 fprintf (dump_file, "Will attempt to totally scalarize ");
2521 print_generic_expr (dump_file, var, 0);
2522 fprintf (dump_file, " (UID: %u): \n", DECL_UID (var));
2525 else if (dump_file && (dump_flags & TDF_DETAILS))
2527 fprintf (dump_file, "Too big to totally scalarize: ");
2528 print_generic_expr (dump_file, var, 0);
2529 fprintf (dump_file, " (UID: %u)\n", DECL_UID (var));
2534 bitmap_copy (tmp, candidate_bitmap);
2535 EXECUTE_IF_SET_IN_BITMAP (tmp, 0, i, bi)
2537 tree var = candidate (i);
2538 struct access *access;
2540 access = sort_and_splice_var_accesses (var);
2541 if (!access || !build_access_trees (access))
2542 disqualify_candidate (var,
2543 "No or inhibitingly overlapping accesses.");
2546 propagate_all_subaccesses ();
2548 bitmap_copy (tmp, candidate_bitmap);
2549 EXECUTE_IF_SET_IN_BITMAP (tmp, 0, i, bi)
2551 tree var = candidate (i);
2552 struct access *access = get_first_repr_for_decl (var);
2554 if (analyze_access_trees (access))
2556 res++;
2557 if (dump_file && (dump_flags & TDF_DETAILS))
2559 fprintf (dump_file, "\nAccess trees for ");
2560 print_generic_expr (dump_file, var, 0);
2561 fprintf (dump_file, " (UID: %u): \n", DECL_UID (var));
2562 dump_access_tree (dump_file, access);
2563 fprintf (dump_file, "\n");
2566 else
2567 disqualify_candidate (var, "No scalar replacements to be created.");
2570 BITMAP_FREE (tmp);
2572 if (res)
2574 statistics_counter_event (cfun, "Scalarized aggregates", res);
2575 return true;
2577 else
2578 return false;
2581 /* Generate statements copying scalar replacements of accesses within a subtree
2582 into or out of AGG. ACCESS, all its children, siblings and their children
2583 are to be processed. AGG is an aggregate type expression (can be a
2584 declaration but does not have to be, it can for example also be a mem_ref or
2585 a series of handled components). TOP_OFFSET is the offset of the processed
2586 subtree which has to be subtracted from offsets of individual accesses to
2587 get corresponding offsets for AGG. If CHUNK_SIZE is non-null, copy only
2588 replacements in the interval <start_offset, start_offset + chunk_size>,
2589 otherwise copy all. GSI is a statement iterator used to place the new
2590 statements. WRITE should be true when the statements should write from AGG
2591 to the replacement and false if vice versa. if INSERT_AFTER is true, new
2592 statements will be added after the current statement in GSI, they will be
2593 added before the statement otherwise. */
2595 static void
2596 generate_subtree_copies (struct access *access, tree agg,
2597 HOST_WIDE_INT top_offset,
2598 HOST_WIDE_INT start_offset, HOST_WIDE_INT chunk_size,
2599 gimple_stmt_iterator *gsi, bool write,
2600 bool insert_after, location_t loc)
2604 if (chunk_size && access->offset >= start_offset + chunk_size)
2605 return;
2607 if (access->grp_to_be_replaced
2608 && (chunk_size == 0
2609 || access->offset + access->size > start_offset))
2611 tree expr, repl = get_access_replacement (access);
2612 gimple stmt;
2614 expr = build_ref_for_model (loc, agg, access->offset - top_offset,
2615 access, gsi, insert_after);
2617 if (write)
2619 if (access->grp_partial_lhs)
2620 expr = force_gimple_operand_gsi (gsi, expr, true, NULL_TREE,
2621 !insert_after,
2622 insert_after ? GSI_NEW_STMT
2623 : GSI_SAME_STMT);
2624 stmt = gimple_build_assign (repl, expr);
2626 else
2628 TREE_NO_WARNING (repl) = 1;
2629 if (access->grp_partial_lhs)
2630 repl = force_gimple_operand_gsi (gsi, repl, true, NULL_TREE,
2631 !insert_after,
2632 insert_after ? GSI_NEW_STMT
2633 : GSI_SAME_STMT);
2634 stmt = gimple_build_assign (expr, repl);
2636 gimple_set_location (stmt, loc);
2638 if (insert_after)
2639 gsi_insert_after (gsi, stmt, GSI_NEW_STMT);
2640 else
2641 gsi_insert_before (gsi, stmt, GSI_SAME_STMT);
2642 update_stmt (stmt);
2643 sra_stats.subtree_copies++;
2645 else if (write
2646 && access->grp_to_be_debug_replaced
2647 && (chunk_size == 0
2648 || access->offset + access->size > start_offset))
2650 gimple ds;
2651 tree drhs = build_debug_ref_for_model (loc, agg,
2652 access->offset - top_offset,
2653 access);
2654 ds = gimple_build_debug_bind (get_access_replacement (access),
2655 drhs, gsi_stmt (*gsi));
2656 if (insert_after)
2657 gsi_insert_after (gsi, ds, GSI_NEW_STMT);
2658 else
2659 gsi_insert_before (gsi, ds, GSI_SAME_STMT);
2662 if (access->first_child)
2663 generate_subtree_copies (access->first_child, agg, top_offset,
2664 start_offset, chunk_size, gsi,
2665 write, insert_after, loc);
2667 access = access->next_sibling;
2669 while (access);
2672 /* Assign zero to all scalar replacements in an access subtree. ACCESS is the
2673 the root of the subtree to be processed. GSI is the statement iterator used
2674 for inserting statements which are added after the current statement if
2675 INSERT_AFTER is true or before it otherwise. */
2677 static void
2678 init_subtree_with_zero (struct access *access, gimple_stmt_iterator *gsi,
2679 bool insert_after, location_t loc)
2682 struct access *child;
2684 if (access->grp_to_be_replaced)
2686 gimple stmt;
2688 stmt = gimple_build_assign (get_access_replacement (access),
2689 build_zero_cst (access->type));
2690 if (insert_after)
2691 gsi_insert_after (gsi, stmt, GSI_NEW_STMT);
2692 else
2693 gsi_insert_before (gsi, stmt, GSI_SAME_STMT);
2694 update_stmt (stmt);
2695 gimple_set_location (stmt, loc);
2697 else if (access->grp_to_be_debug_replaced)
2699 gimple ds = gimple_build_debug_bind (get_access_replacement (access),
2700 build_zero_cst (access->type),
2701 gsi_stmt (*gsi));
2702 if (insert_after)
2703 gsi_insert_after (gsi, ds, GSI_NEW_STMT);
2704 else
2705 gsi_insert_before (gsi, ds, GSI_SAME_STMT);
2708 for (child = access->first_child; child; child = child->next_sibling)
2709 init_subtree_with_zero (child, gsi, insert_after, loc);
2712 /* Search for an access representative for the given expression EXPR and
2713 return it or NULL if it cannot be found. */
2715 static struct access *
2716 get_access_for_expr (tree expr)
2718 HOST_WIDE_INT offset, size, max_size;
2719 tree base;
2721 /* FIXME: This should not be necessary but Ada produces V_C_Es with a type of
2722 a different size than the size of its argument and we need the latter
2723 one. */
2724 if (TREE_CODE (expr) == VIEW_CONVERT_EXPR)
2725 expr = TREE_OPERAND (expr, 0);
2727 base = get_ref_base_and_extent (expr, &offset, &size, &max_size);
2728 if (max_size == -1 || !DECL_P (base))
2729 return NULL;
2731 if (!bitmap_bit_p (candidate_bitmap, DECL_UID (base)))
2732 return NULL;
2734 return get_var_base_offset_size_access (base, offset, max_size);
2737 /* Replace the expression EXPR with a scalar replacement if there is one and
2738 generate other statements to do type conversion or subtree copying if
2739 necessary. GSI is used to place newly created statements, WRITE is true if
2740 the expression is being written to (it is on a LHS of a statement or output
2741 in an assembly statement). */
2743 static bool
2744 sra_modify_expr (tree *expr, gimple_stmt_iterator *gsi, bool write)
2746 location_t loc;
2747 struct access *access;
2748 tree type, bfr;
2750 if (TREE_CODE (*expr) == BIT_FIELD_REF)
2752 bfr = *expr;
2753 expr = &TREE_OPERAND (*expr, 0);
2755 else
2756 bfr = NULL_TREE;
2758 if (TREE_CODE (*expr) == REALPART_EXPR || TREE_CODE (*expr) == IMAGPART_EXPR)
2759 expr = &TREE_OPERAND (*expr, 0);
2760 access = get_access_for_expr (*expr);
2761 if (!access)
2762 return false;
2763 type = TREE_TYPE (*expr);
2765 loc = gimple_location (gsi_stmt (*gsi));
2766 if (access->grp_to_be_replaced)
2768 tree repl = get_access_replacement (access);
2769 /* If we replace a non-register typed access simply use the original
2770 access expression to extract the scalar component afterwards.
2771 This happens if scalarizing a function return value or parameter
2772 like in gcc.c-torture/execute/20041124-1.c, 20050316-1.c and
2773 gcc.c-torture/compile/20011217-1.c.
2775 We also want to use this when accessing a complex or vector which can
2776 be accessed as a different type too, potentially creating a need for
2777 type conversion (see PR42196) and when scalarized unions are involved
2778 in assembler statements (see PR42398). */
2779 if (!useless_type_conversion_p (type, access->type))
2781 tree ref;
2783 ref = build_ref_for_model (loc, access->base, access->offset, access,
2784 NULL, false);
2786 if (write)
2788 gimple stmt;
2790 if (access->grp_partial_lhs)
2791 ref = force_gimple_operand_gsi (gsi, ref, true, NULL_TREE,
2792 false, GSI_NEW_STMT);
2793 stmt = gimple_build_assign (repl, ref);
2794 gimple_set_location (stmt, loc);
2795 gsi_insert_after (gsi, stmt, GSI_NEW_STMT);
2797 else
2799 gimple stmt;
2801 if (access->grp_partial_lhs)
2802 repl = force_gimple_operand_gsi (gsi, repl, true, NULL_TREE,
2803 true, GSI_SAME_STMT);
2804 stmt = gimple_build_assign (ref, repl);
2805 gimple_set_location (stmt, loc);
2806 gsi_insert_before (gsi, stmt, GSI_SAME_STMT);
2809 else
2810 *expr = repl;
2811 sra_stats.exprs++;
2813 else if (write && access->grp_to_be_debug_replaced)
2815 gimple ds = gimple_build_debug_bind (get_access_replacement (access),
2816 NULL_TREE,
2817 gsi_stmt (*gsi));
2818 gsi_insert_after (gsi, ds, GSI_NEW_STMT);
2821 if (access->first_child)
2823 HOST_WIDE_INT start_offset, chunk_size;
2824 if (bfr
2825 && tree_fits_uhwi_p (TREE_OPERAND (bfr, 1))
2826 && tree_fits_uhwi_p (TREE_OPERAND (bfr, 2)))
2828 chunk_size = tree_to_uhwi (TREE_OPERAND (bfr, 1));
2829 start_offset = access->offset
2830 + tree_to_uhwi (TREE_OPERAND (bfr, 2));
2832 else
2833 start_offset = chunk_size = 0;
2835 generate_subtree_copies (access->first_child, access->base, 0,
2836 start_offset, chunk_size, gsi, write, write,
2837 loc);
2839 return true;
2842 /* Where scalar replacements of the RHS have been written to when a replacement
2843 of a LHS of an assigments cannot be direclty loaded from a replacement of
2844 the RHS. */
2845 enum unscalarized_data_handling { SRA_UDH_NONE, /* Nothing done so far. */
2846 SRA_UDH_RIGHT, /* Data flushed to the RHS. */
2847 SRA_UDH_LEFT }; /* Data flushed to the LHS. */
2849 /* Store all replacements in the access tree rooted in TOP_RACC either to their
2850 base aggregate if there are unscalarized data or directly to LHS of the
2851 statement that is pointed to by GSI otherwise. */
2853 static enum unscalarized_data_handling
2854 handle_unscalarized_data_in_subtree (struct access *top_racc,
2855 gimple_stmt_iterator *gsi)
2857 if (top_racc->grp_unscalarized_data)
2859 generate_subtree_copies (top_racc->first_child, top_racc->base, 0, 0, 0,
2860 gsi, false, false,
2861 gimple_location (gsi_stmt (*gsi)));
2862 return SRA_UDH_RIGHT;
2864 else
2866 tree lhs = gimple_assign_lhs (gsi_stmt (*gsi));
2867 generate_subtree_copies (top_racc->first_child, lhs, top_racc->offset,
2868 0, 0, gsi, false, false,
2869 gimple_location (gsi_stmt (*gsi)));
2870 return SRA_UDH_LEFT;
2875 /* Try to generate statements to load all sub-replacements in an access subtree
2876 formed by children of LACC from scalar replacements in the TOP_RACC subtree.
2877 If that is not possible, refresh the TOP_RACC base aggregate and load the
2878 accesses from it. LEFT_OFFSET is the offset of the left whole subtree being
2879 copied. NEW_GSI is stmt iterator used for statement insertions after the
2880 original assignment, OLD_GSI is used to insert statements before the
2881 assignment. *REFRESHED keeps the information whether we have needed to
2882 refresh replacements of the LHS and from which side of the assignments this
2883 takes place. */
2885 static void
2886 load_assign_lhs_subreplacements (struct access *lacc, struct access *top_racc,
2887 HOST_WIDE_INT left_offset,
2888 gimple_stmt_iterator *old_gsi,
2889 gimple_stmt_iterator *new_gsi,
2890 enum unscalarized_data_handling *refreshed)
2892 location_t loc = gimple_location (gsi_stmt (*old_gsi));
2893 for (lacc = lacc->first_child; lacc; lacc = lacc->next_sibling)
2895 HOST_WIDE_INT offset = lacc->offset - left_offset + top_racc->offset;
2897 if (lacc->grp_to_be_replaced)
2899 struct access *racc;
2900 gimple stmt;
2901 tree rhs;
2903 racc = find_access_in_subtree (top_racc, offset, lacc->size);
2904 if (racc && racc->grp_to_be_replaced)
2906 rhs = get_access_replacement (racc);
2907 if (!useless_type_conversion_p (lacc->type, racc->type))
2908 rhs = fold_build1_loc (loc, VIEW_CONVERT_EXPR, lacc->type, rhs);
2910 if (racc->grp_partial_lhs && lacc->grp_partial_lhs)
2911 rhs = force_gimple_operand_gsi (old_gsi, rhs, true, NULL_TREE,
2912 true, GSI_SAME_STMT);
2914 else
2916 /* No suitable access on the right hand side, need to load from
2917 the aggregate. See if we have to update it first... */
2918 if (*refreshed == SRA_UDH_NONE)
2919 *refreshed = handle_unscalarized_data_in_subtree (top_racc,
2920 old_gsi);
2922 if (*refreshed == SRA_UDH_LEFT)
2923 rhs = build_ref_for_model (loc, lacc->base, lacc->offset, lacc,
2924 new_gsi, true);
2925 else
2926 rhs = build_ref_for_model (loc, top_racc->base, offset, lacc,
2927 new_gsi, true);
2928 if (lacc->grp_partial_lhs)
2929 rhs = force_gimple_operand_gsi (new_gsi, rhs, true, NULL_TREE,
2930 false, GSI_NEW_STMT);
2933 stmt = gimple_build_assign (get_access_replacement (lacc), rhs);
2934 gsi_insert_after (new_gsi, stmt, GSI_NEW_STMT);
2935 gimple_set_location (stmt, loc);
2936 update_stmt (stmt);
2937 sra_stats.subreplacements++;
2939 else
2941 if (*refreshed == SRA_UDH_NONE
2942 && lacc->grp_read && !lacc->grp_covered)
2943 *refreshed = handle_unscalarized_data_in_subtree (top_racc,
2944 old_gsi);
2945 if (lacc && lacc->grp_to_be_debug_replaced)
2947 gimple ds;
2948 tree drhs;
2949 struct access *racc = find_access_in_subtree (top_racc, offset,
2950 lacc->size);
2952 if (racc && racc->grp_to_be_replaced)
2954 if (racc->grp_write)
2955 drhs = get_access_replacement (racc);
2956 else
2957 drhs = NULL;
2959 else if (*refreshed == SRA_UDH_LEFT)
2960 drhs = build_debug_ref_for_model (loc, lacc->base, lacc->offset,
2961 lacc);
2962 else if (*refreshed == SRA_UDH_RIGHT)
2963 drhs = build_debug_ref_for_model (loc, top_racc->base, offset,
2964 lacc);
2965 else
2966 drhs = NULL_TREE;
2967 if (drhs
2968 && !useless_type_conversion_p (lacc->type, TREE_TYPE (drhs)))
2969 drhs = fold_build1_loc (loc, VIEW_CONVERT_EXPR,
2970 lacc->type, drhs);
2971 ds = gimple_build_debug_bind (get_access_replacement (lacc),
2972 drhs, gsi_stmt (*old_gsi));
2973 gsi_insert_after (new_gsi, ds, GSI_NEW_STMT);
2977 if (lacc->first_child)
2978 load_assign_lhs_subreplacements (lacc, top_racc, left_offset,
2979 old_gsi, new_gsi, refreshed);
2983 /* Result code for SRA assignment modification. */
2984 enum assignment_mod_result { SRA_AM_NONE, /* nothing done for the stmt */
2985 SRA_AM_MODIFIED, /* stmt changed but not
2986 removed */
2987 SRA_AM_REMOVED }; /* stmt eliminated */
2989 /* Modify assignments with a CONSTRUCTOR on their RHS. STMT contains a pointer
2990 to the assignment and GSI is the statement iterator pointing at it. Returns
2991 the same values as sra_modify_assign. */
2993 static enum assignment_mod_result
2994 sra_modify_constructor_assign (gimple *stmt, gimple_stmt_iterator *gsi)
2996 tree lhs = gimple_assign_lhs (*stmt);
2997 struct access *acc;
2998 location_t loc;
3000 acc = get_access_for_expr (lhs);
3001 if (!acc)
3002 return SRA_AM_NONE;
3004 if (gimple_clobber_p (*stmt))
3006 /* Remove clobbers of fully scalarized variables, otherwise
3007 do nothing. */
3008 if (acc->grp_covered)
3010 unlink_stmt_vdef (*stmt);
3011 gsi_remove (gsi, true);
3012 release_defs (*stmt);
3013 return SRA_AM_REMOVED;
3015 else
3016 return SRA_AM_NONE;
3019 loc = gimple_location (*stmt);
3020 if (vec_safe_length (CONSTRUCTOR_ELTS (gimple_assign_rhs1 (*stmt))) > 0)
3022 /* I have never seen this code path trigger but if it can happen the
3023 following should handle it gracefully. */
3024 if (access_has_children_p (acc))
3025 generate_subtree_copies (acc->first_child, acc->base, 0, 0, 0, gsi,
3026 true, true, loc);
3027 return SRA_AM_MODIFIED;
3030 if (acc->grp_covered)
3032 init_subtree_with_zero (acc, gsi, false, loc);
3033 unlink_stmt_vdef (*stmt);
3034 gsi_remove (gsi, true);
3035 release_defs (*stmt);
3036 return SRA_AM_REMOVED;
3038 else
3040 init_subtree_with_zero (acc, gsi, true, loc);
3041 return SRA_AM_MODIFIED;
3045 /* Create and return a new suitable default definition SSA_NAME for RACC which
3046 is an access describing an uninitialized part of an aggregate that is being
3047 loaded. */
3049 static tree
3050 get_repl_default_def_ssa_name (struct access *racc)
3052 gcc_checking_assert (!racc->grp_to_be_replaced
3053 && !racc->grp_to_be_debug_replaced);
3054 if (!racc->replacement_decl)
3055 racc->replacement_decl = create_access_replacement (racc);
3056 return get_or_create_ssa_default_def (cfun, racc->replacement_decl);
3059 /* Return true if REF has an VIEW_CONVERT_EXPR or a COMPONENT_REF with a
3060 bit-field field declaration somewhere in it. */
3062 static inline bool
3063 contains_vce_or_bfcref_p (const_tree ref)
3065 while (handled_component_p (ref))
3067 if (TREE_CODE (ref) == VIEW_CONVERT_EXPR
3068 || (TREE_CODE (ref) == COMPONENT_REF
3069 && DECL_BIT_FIELD (TREE_OPERAND (ref, 1))))
3070 return true;
3071 ref = TREE_OPERAND (ref, 0);
3074 return false;
3077 /* Examine both sides of the assignment statement pointed to by STMT, replace
3078 them with a scalare replacement if there is one and generate copying of
3079 replacements if scalarized aggregates have been used in the assignment. GSI
3080 is used to hold generated statements for type conversions and subtree
3081 copying. */
3083 static enum assignment_mod_result
3084 sra_modify_assign (gimple *stmt, gimple_stmt_iterator *gsi)
3086 struct access *lacc, *racc;
3087 tree lhs, rhs;
3088 bool modify_this_stmt = false;
3089 bool force_gimple_rhs = false;
3090 location_t loc;
3091 gimple_stmt_iterator orig_gsi = *gsi;
3093 if (!gimple_assign_single_p (*stmt))
3094 return SRA_AM_NONE;
3095 lhs = gimple_assign_lhs (*stmt);
3096 rhs = gimple_assign_rhs1 (*stmt);
3098 if (TREE_CODE (rhs) == CONSTRUCTOR)
3099 return sra_modify_constructor_assign (stmt, gsi);
3101 if (TREE_CODE (rhs) == REALPART_EXPR || TREE_CODE (lhs) == REALPART_EXPR
3102 || TREE_CODE (rhs) == IMAGPART_EXPR || TREE_CODE (lhs) == IMAGPART_EXPR
3103 || TREE_CODE (rhs) == BIT_FIELD_REF || TREE_CODE (lhs) == BIT_FIELD_REF)
3105 modify_this_stmt = sra_modify_expr (gimple_assign_rhs1_ptr (*stmt),
3106 gsi, false);
3107 modify_this_stmt |= sra_modify_expr (gimple_assign_lhs_ptr (*stmt),
3108 gsi, true);
3109 return modify_this_stmt ? SRA_AM_MODIFIED : SRA_AM_NONE;
3112 lacc = get_access_for_expr (lhs);
3113 racc = get_access_for_expr (rhs);
3114 if (!lacc && !racc)
3115 return SRA_AM_NONE;
3117 loc = gimple_location (*stmt);
3118 if (lacc && lacc->grp_to_be_replaced)
3120 lhs = get_access_replacement (lacc);
3121 gimple_assign_set_lhs (*stmt, lhs);
3122 modify_this_stmt = true;
3123 if (lacc->grp_partial_lhs)
3124 force_gimple_rhs = true;
3125 sra_stats.exprs++;
3128 if (racc && racc->grp_to_be_replaced)
3130 rhs = get_access_replacement (racc);
3131 modify_this_stmt = true;
3132 if (racc->grp_partial_lhs)
3133 force_gimple_rhs = true;
3134 sra_stats.exprs++;
3136 else if (racc
3137 && !racc->grp_unscalarized_data
3138 && TREE_CODE (lhs) == SSA_NAME
3139 && !access_has_replacements_p (racc))
3141 rhs = get_repl_default_def_ssa_name (racc);
3142 modify_this_stmt = true;
3143 sra_stats.exprs++;
3146 if (modify_this_stmt)
3148 if (!useless_type_conversion_p (TREE_TYPE (lhs), TREE_TYPE (rhs)))
3150 /* If we can avoid creating a VIEW_CONVERT_EXPR do so.
3151 ??? This should move to fold_stmt which we simply should
3152 call after building a VIEW_CONVERT_EXPR here. */
3153 if (AGGREGATE_TYPE_P (TREE_TYPE (lhs))
3154 && !contains_bitfld_component_ref_p (lhs))
3156 lhs = build_ref_for_model (loc, lhs, 0, racc, gsi, false);
3157 gimple_assign_set_lhs (*stmt, lhs);
3159 else if (AGGREGATE_TYPE_P (TREE_TYPE (rhs))
3160 && !contains_vce_or_bfcref_p (rhs))
3161 rhs = build_ref_for_model (loc, rhs, 0, lacc, gsi, false);
3163 if (!useless_type_conversion_p (TREE_TYPE (lhs), TREE_TYPE (rhs)))
3165 rhs = fold_build1_loc (loc, VIEW_CONVERT_EXPR, TREE_TYPE (lhs),
3166 rhs);
3167 if (is_gimple_reg_type (TREE_TYPE (lhs))
3168 && TREE_CODE (lhs) != SSA_NAME)
3169 force_gimple_rhs = true;
3174 if (lacc && lacc->grp_to_be_debug_replaced)
3176 tree dlhs = get_access_replacement (lacc);
3177 tree drhs = unshare_expr (rhs);
3178 if (!useless_type_conversion_p (TREE_TYPE (dlhs), TREE_TYPE (drhs)))
3180 if (AGGREGATE_TYPE_P (TREE_TYPE (drhs))
3181 && !contains_vce_or_bfcref_p (drhs))
3182 drhs = build_debug_ref_for_model (loc, drhs, 0, lacc);
3183 if (drhs
3184 && !useless_type_conversion_p (TREE_TYPE (dlhs),
3185 TREE_TYPE (drhs)))
3186 drhs = fold_build1_loc (loc, VIEW_CONVERT_EXPR,
3187 TREE_TYPE (dlhs), drhs);
3189 gimple ds = gimple_build_debug_bind (dlhs, drhs, *stmt);
3190 gsi_insert_before (gsi, ds, GSI_SAME_STMT);
3193 /* From this point on, the function deals with assignments in between
3194 aggregates when at least one has scalar reductions of some of its
3195 components. There are three possible scenarios: Both the LHS and RHS have
3196 to-be-scalarized components, 2) only the RHS has or 3) only the LHS has.
3198 In the first case, we would like to load the LHS components from RHS
3199 components whenever possible. If that is not possible, we would like to
3200 read it directly from the RHS (after updating it by storing in it its own
3201 components). If there are some necessary unscalarized data in the LHS,
3202 those will be loaded by the original assignment too. If neither of these
3203 cases happen, the original statement can be removed. Most of this is done
3204 by load_assign_lhs_subreplacements.
3206 In the second case, we would like to store all RHS scalarized components
3207 directly into LHS and if they cover the aggregate completely, remove the
3208 statement too. In the third case, we want the LHS components to be loaded
3209 directly from the RHS (DSE will remove the original statement if it
3210 becomes redundant).
3212 This is a bit complex but manageable when types match and when unions do
3213 not cause confusion in a way that we cannot really load a component of LHS
3214 from the RHS or vice versa (the access representing this level can have
3215 subaccesses that are accessible only through a different union field at a
3216 higher level - different from the one used in the examined expression).
3217 Unions are fun.
3219 Therefore, I specially handle a fourth case, happening when there is a
3220 specific type cast or it is impossible to locate a scalarized subaccess on
3221 the other side of the expression. If that happens, I simply "refresh" the
3222 RHS by storing in it is scalarized components leave the original statement
3223 there to do the copying and then load the scalar replacements of the LHS.
3224 This is what the first branch does. */
3226 if (modify_this_stmt
3227 || gimple_has_volatile_ops (*stmt)
3228 || contains_vce_or_bfcref_p (rhs)
3229 || contains_vce_or_bfcref_p (lhs))
3231 if (access_has_children_p (racc))
3232 generate_subtree_copies (racc->first_child, racc->base, 0, 0, 0,
3233 gsi, false, false, loc);
3234 if (access_has_children_p (lacc))
3235 generate_subtree_copies (lacc->first_child, lacc->base, 0, 0, 0,
3236 gsi, true, true, loc);
3237 sra_stats.separate_lhs_rhs_handling++;
3239 /* This gimplification must be done after generate_subtree_copies,
3240 lest we insert the subtree copies in the middle of the gimplified
3241 sequence. */
3242 if (force_gimple_rhs)
3243 rhs = force_gimple_operand_gsi (&orig_gsi, rhs, true, NULL_TREE,
3244 true, GSI_SAME_STMT);
3245 if (gimple_assign_rhs1 (*stmt) != rhs)
3247 modify_this_stmt = true;
3248 gimple_assign_set_rhs_from_tree (&orig_gsi, rhs);
3249 gcc_assert (*stmt == gsi_stmt (orig_gsi));
3252 return modify_this_stmt ? SRA_AM_MODIFIED : SRA_AM_NONE;
3254 else
3256 if (access_has_children_p (lacc)
3257 && access_has_children_p (racc)
3258 /* When an access represents an unscalarizable region, it usually
3259 represents accesses with variable offset and thus must not be used
3260 to generate new memory accesses. */
3261 && !lacc->grp_unscalarizable_region
3262 && !racc->grp_unscalarizable_region)
3264 gimple_stmt_iterator orig_gsi = *gsi;
3265 enum unscalarized_data_handling refreshed;
3267 if (lacc->grp_read && !lacc->grp_covered)
3268 refreshed = handle_unscalarized_data_in_subtree (racc, gsi);
3269 else
3270 refreshed = SRA_UDH_NONE;
3272 load_assign_lhs_subreplacements (lacc, racc, lacc->offset,
3273 &orig_gsi, gsi, &refreshed);
3274 if (refreshed != SRA_UDH_RIGHT)
3276 gsi_next (gsi);
3277 unlink_stmt_vdef (*stmt);
3278 gsi_remove (&orig_gsi, true);
3279 release_defs (*stmt);
3280 sra_stats.deleted++;
3281 return SRA_AM_REMOVED;
3284 else
3286 if (access_has_children_p (racc)
3287 && !racc->grp_unscalarized_data)
3289 if (dump_file)
3291 fprintf (dump_file, "Removing load: ");
3292 print_gimple_stmt (dump_file, *stmt, 0, 0);
3294 generate_subtree_copies (racc->first_child, lhs,
3295 racc->offset, 0, 0, gsi,
3296 false, false, loc);
3297 gcc_assert (*stmt == gsi_stmt (*gsi));
3298 unlink_stmt_vdef (*stmt);
3299 gsi_remove (gsi, true);
3300 release_defs (*stmt);
3301 sra_stats.deleted++;
3302 return SRA_AM_REMOVED;
3304 /* Restore the aggregate RHS from its components so the
3305 prevailing aggregate copy does the right thing. */
3306 if (access_has_children_p (racc))
3307 generate_subtree_copies (racc->first_child, racc->base, 0, 0, 0,
3308 gsi, false, false, loc);
3309 /* Re-load the components of the aggregate copy destination.
3310 But use the RHS aggregate to load from to expose more
3311 optimization opportunities. */
3312 if (access_has_children_p (lacc))
3313 generate_subtree_copies (lacc->first_child, rhs, lacc->offset,
3314 0, 0, gsi, true, true, loc);
3317 return SRA_AM_NONE;
3321 /* Traverse the function body and all modifications as decided in
3322 analyze_all_variable_accesses. Return true iff the CFG has been
3323 changed. */
3325 static bool
3326 sra_modify_function_body (void)
3328 bool cfg_changed = false;
3329 basic_block bb;
3331 FOR_EACH_BB_FN (bb, cfun)
3333 gimple_stmt_iterator gsi = gsi_start_bb (bb);
3334 while (!gsi_end_p (gsi))
3336 gimple stmt = gsi_stmt (gsi);
3337 enum assignment_mod_result assign_result;
3338 bool modified = false, deleted = false;
3339 tree *t;
3340 unsigned i;
3342 switch (gimple_code (stmt))
3344 case GIMPLE_RETURN:
3345 t = gimple_return_retval_ptr (stmt);
3346 if (*t != NULL_TREE)
3347 modified |= sra_modify_expr (t, &gsi, false);
3348 break;
3350 case GIMPLE_ASSIGN:
3351 assign_result = sra_modify_assign (&stmt, &gsi);
3352 modified |= assign_result == SRA_AM_MODIFIED;
3353 deleted = assign_result == SRA_AM_REMOVED;
3354 break;
3356 case GIMPLE_CALL:
3357 /* Operands must be processed before the lhs. */
3358 for (i = 0; i < gimple_call_num_args (stmt); i++)
3360 t = gimple_call_arg_ptr (stmt, i);
3361 modified |= sra_modify_expr (t, &gsi, false);
3364 if (gimple_call_lhs (stmt))
3366 t = gimple_call_lhs_ptr (stmt);
3367 modified |= sra_modify_expr (t, &gsi, true);
3369 break;
3371 case GIMPLE_ASM:
3372 for (i = 0; i < gimple_asm_ninputs (stmt); i++)
3374 t = &TREE_VALUE (gimple_asm_input_op (stmt, i));
3375 modified |= sra_modify_expr (t, &gsi, false);
3377 for (i = 0; i < gimple_asm_noutputs (stmt); i++)
3379 t = &TREE_VALUE (gimple_asm_output_op (stmt, i));
3380 modified |= sra_modify_expr (t, &gsi, true);
3382 break;
3384 default:
3385 break;
3388 if (modified)
3390 update_stmt (stmt);
3391 if (maybe_clean_eh_stmt (stmt)
3392 && gimple_purge_dead_eh_edges (gimple_bb (stmt)))
3393 cfg_changed = true;
3395 if (!deleted)
3396 gsi_next (&gsi);
3400 return cfg_changed;
3403 /* Generate statements initializing scalar replacements of parts of function
3404 parameters. */
3406 static void
3407 initialize_parameter_reductions (void)
3409 gimple_stmt_iterator gsi;
3410 gimple_seq seq = NULL;
3411 tree parm;
3413 gsi = gsi_start (seq);
3414 for (parm = DECL_ARGUMENTS (current_function_decl);
3415 parm;
3416 parm = DECL_CHAIN (parm))
3418 vec<access_p> *access_vec;
3419 struct access *access;
3421 if (!bitmap_bit_p (candidate_bitmap, DECL_UID (parm)))
3422 continue;
3423 access_vec = get_base_access_vector (parm);
3424 if (!access_vec)
3425 continue;
3427 for (access = (*access_vec)[0];
3428 access;
3429 access = access->next_grp)
3430 generate_subtree_copies (access, parm, 0, 0, 0, &gsi, true, true,
3431 EXPR_LOCATION (parm));
3434 seq = gsi_seq (gsi);
3435 if (seq)
3436 gsi_insert_seq_on_edge_immediate (single_succ_edge (ENTRY_BLOCK_PTR_FOR_FN (cfun)), seq);
3439 /* The "main" function of intraprocedural SRA passes. Runs the analysis and if
3440 it reveals there are components of some aggregates to be scalarized, it runs
3441 the required transformations. */
3442 static unsigned int
3443 perform_intra_sra (void)
3445 int ret = 0;
3446 sra_initialize ();
3448 if (!find_var_candidates ())
3449 goto out;
3451 if (!scan_function ())
3452 goto out;
3454 if (!analyze_all_variable_accesses ())
3455 goto out;
3457 if (sra_modify_function_body ())
3458 ret = TODO_update_ssa | TODO_cleanup_cfg;
3459 else
3460 ret = TODO_update_ssa;
3461 initialize_parameter_reductions ();
3463 statistics_counter_event (cfun, "Scalar replacements created",
3464 sra_stats.replacements);
3465 statistics_counter_event (cfun, "Modified expressions", sra_stats.exprs);
3466 statistics_counter_event (cfun, "Subtree copy stmts",
3467 sra_stats.subtree_copies);
3468 statistics_counter_event (cfun, "Subreplacement stmts",
3469 sra_stats.subreplacements);
3470 statistics_counter_event (cfun, "Deleted stmts", sra_stats.deleted);
3471 statistics_counter_event (cfun, "Separate LHS and RHS handling",
3472 sra_stats.separate_lhs_rhs_handling);
3474 out:
3475 sra_deinitialize ();
3476 return ret;
3479 /* Perform early intraprocedural SRA. */
3480 static unsigned int
3481 early_intra_sra (void)
3483 sra_mode = SRA_MODE_EARLY_INTRA;
3484 return perform_intra_sra ();
3487 /* Perform "late" intraprocedural SRA. */
3488 static unsigned int
3489 late_intra_sra (void)
3491 sra_mode = SRA_MODE_INTRA;
3492 return perform_intra_sra ();
3496 static bool
3497 gate_intra_sra (void)
3499 return flag_tree_sra != 0 && dbg_cnt (tree_sra);
3503 namespace {
3505 const pass_data pass_data_sra_early =
3507 GIMPLE_PASS, /* type */
3508 "esra", /* name */
3509 OPTGROUP_NONE, /* optinfo_flags */
3510 true, /* has_gate */
3511 true, /* has_execute */
3512 TV_TREE_SRA, /* tv_id */
3513 ( PROP_cfg | PROP_ssa ), /* properties_required */
3514 0, /* properties_provided */
3515 0, /* properties_destroyed */
3516 0, /* todo_flags_start */
3517 ( TODO_update_ssa | TODO_verify_ssa ), /* todo_flags_finish */
3520 class pass_sra_early : public gimple_opt_pass
3522 public:
3523 pass_sra_early (gcc::context *ctxt)
3524 : gimple_opt_pass (pass_data_sra_early, ctxt)
3527 /* opt_pass methods: */
3528 bool gate () { return gate_intra_sra (); }
3529 unsigned int execute () { return early_intra_sra (); }
3531 }; // class pass_sra_early
3533 } // anon namespace
3535 gimple_opt_pass *
3536 make_pass_sra_early (gcc::context *ctxt)
3538 return new pass_sra_early (ctxt);
3541 namespace {
3543 const pass_data pass_data_sra =
3545 GIMPLE_PASS, /* type */
3546 "sra", /* name */
3547 OPTGROUP_NONE, /* optinfo_flags */
3548 true, /* has_gate */
3549 true, /* has_execute */
3550 TV_TREE_SRA, /* tv_id */
3551 ( PROP_cfg | PROP_ssa ), /* properties_required */
3552 0, /* properties_provided */
3553 0, /* properties_destroyed */
3554 TODO_update_address_taken, /* todo_flags_start */
3555 ( TODO_update_ssa | TODO_verify_ssa ), /* todo_flags_finish */
3558 class pass_sra : public gimple_opt_pass
3560 public:
3561 pass_sra (gcc::context *ctxt)
3562 : gimple_opt_pass (pass_data_sra, ctxt)
3565 /* opt_pass methods: */
3566 bool gate () { return gate_intra_sra (); }
3567 unsigned int execute () { return late_intra_sra (); }
3569 }; // class pass_sra
3571 } // anon namespace
3573 gimple_opt_pass *
3574 make_pass_sra (gcc::context *ctxt)
3576 return new pass_sra (ctxt);
3580 /* Return true iff PARM (which must be a parm_decl) is an unused scalar
3581 parameter. */
3583 static bool
3584 is_unused_scalar_param (tree parm)
3586 tree name;
3587 return (is_gimple_reg (parm)
3588 && (!(name = ssa_default_def (cfun, parm))
3589 || has_zero_uses (name)));
3592 /* Scan immediate uses of a default definition SSA name of a parameter PARM and
3593 examine whether there are any direct or otherwise infeasible ones. If so,
3594 return true, otherwise return false. PARM must be a gimple register with a
3595 non-NULL default definition. */
3597 static bool
3598 ptr_parm_has_direct_uses (tree parm)
3600 imm_use_iterator ui;
3601 gimple stmt;
3602 tree name = ssa_default_def (cfun, parm);
3603 bool ret = false;
3605 FOR_EACH_IMM_USE_STMT (stmt, ui, name)
3607 int uses_ok = 0;
3608 use_operand_p use_p;
3610 if (is_gimple_debug (stmt))
3611 continue;
3613 /* Valid uses include dereferences on the lhs and the rhs. */
3614 if (gimple_has_lhs (stmt))
3616 tree lhs = gimple_get_lhs (stmt);
3617 while (handled_component_p (lhs))
3618 lhs = TREE_OPERAND (lhs, 0);
3619 if (TREE_CODE (lhs) == MEM_REF
3620 && TREE_OPERAND (lhs, 0) == name
3621 && integer_zerop (TREE_OPERAND (lhs, 1))
3622 && types_compatible_p (TREE_TYPE (lhs),
3623 TREE_TYPE (TREE_TYPE (name)))
3624 && !TREE_THIS_VOLATILE (lhs))
3625 uses_ok++;
3627 if (gimple_assign_single_p (stmt))
3629 tree rhs = gimple_assign_rhs1 (stmt);
3630 while (handled_component_p (rhs))
3631 rhs = TREE_OPERAND (rhs, 0);
3632 if (TREE_CODE (rhs) == MEM_REF
3633 && TREE_OPERAND (rhs, 0) == name
3634 && integer_zerop (TREE_OPERAND (rhs, 1))
3635 && types_compatible_p (TREE_TYPE (rhs),
3636 TREE_TYPE (TREE_TYPE (name)))
3637 && !TREE_THIS_VOLATILE (rhs))
3638 uses_ok++;
3640 else if (is_gimple_call (stmt))
3642 unsigned i;
3643 for (i = 0; i < gimple_call_num_args (stmt); ++i)
3645 tree arg = gimple_call_arg (stmt, i);
3646 while (handled_component_p (arg))
3647 arg = TREE_OPERAND (arg, 0);
3648 if (TREE_CODE (arg) == MEM_REF
3649 && TREE_OPERAND (arg, 0) == name
3650 && integer_zerop (TREE_OPERAND (arg, 1))
3651 && types_compatible_p (TREE_TYPE (arg),
3652 TREE_TYPE (TREE_TYPE (name)))
3653 && !TREE_THIS_VOLATILE (arg))
3654 uses_ok++;
3658 /* If the number of valid uses does not match the number of
3659 uses in this stmt there is an unhandled use. */
3660 FOR_EACH_IMM_USE_ON_STMT (use_p, ui)
3661 --uses_ok;
3663 if (uses_ok != 0)
3664 ret = true;
3666 if (ret)
3667 BREAK_FROM_IMM_USE_STMT (ui);
3670 return ret;
3673 /* Identify candidates for reduction for IPA-SRA based on their type and mark
3674 them in candidate_bitmap. Note that these do not necessarily include
3675 parameter which are unused and thus can be removed. Return true iff any
3676 such candidate has been found. */
3678 static bool
3679 find_param_candidates (void)
3681 tree parm;
3682 int count = 0;
3683 bool ret = false;
3684 const char *msg;
3686 for (parm = DECL_ARGUMENTS (current_function_decl);
3687 parm;
3688 parm = DECL_CHAIN (parm))
3690 tree type = TREE_TYPE (parm);
3691 tree_node **slot;
3693 count++;
3695 if (TREE_THIS_VOLATILE (parm)
3696 || TREE_ADDRESSABLE (parm)
3697 || (!is_gimple_reg_type (type) && is_va_list_type (type)))
3698 continue;
3700 if (is_unused_scalar_param (parm))
3702 ret = true;
3703 continue;
3706 if (POINTER_TYPE_P (type))
3708 type = TREE_TYPE (type);
3710 if (TREE_CODE (type) == FUNCTION_TYPE
3711 || TYPE_VOLATILE (type)
3712 || (TREE_CODE (type) == ARRAY_TYPE
3713 && TYPE_NONALIASED_COMPONENT (type))
3714 || !is_gimple_reg (parm)
3715 || is_va_list_type (type)
3716 || ptr_parm_has_direct_uses (parm))
3717 continue;
3719 else if (!AGGREGATE_TYPE_P (type))
3720 continue;
3722 if (!COMPLETE_TYPE_P (type)
3723 || !tree_fits_uhwi_p (TYPE_SIZE (type))
3724 || tree_to_uhwi (TYPE_SIZE (type)) == 0
3725 || (AGGREGATE_TYPE_P (type)
3726 && type_internals_preclude_sra_p (type, &msg)))
3727 continue;
3729 bitmap_set_bit (candidate_bitmap, DECL_UID (parm));
3730 slot = candidates.find_slot_with_hash (parm, DECL_UID (parm), INSERT);
3731 *slot = parm;
3733 ret = true;
3734 if (dump_file && (dump_flags & TDF_DETAILS))
3736 fprintf (dump_file, "Candidate (%d): ", DECL_UID (parm));
3737 print_generic_expr (dump_file, parm, 0);
3738 fprintf (dump_file, "\n");
3742 func_param_count = count;
3743 return ret;
3746 /* Callback of walk_aliased_vdefs, marks the access passed as DATA as
3747 maybe_modified. */
3749 static bool
3750 mark_maybe_modified (ao_ref *ao ATTRIBUTE_UNUSED, tree vdef ATTRIBUTE_UNUSED,
3751 void *data)
3753 struct access *repr = (struct access *) data;
3755 repr->grp_maybe_modified = 1;
3756 return true;
3759 /* Analyze what representatives (in linked lists accessible from
3760 REPRESENTATIVES) can be modified by side effects of statements in the
3761 current function. */
3763 static void
3764 analyze_modified_params (vec<access_p> representatives)
3766 int i;
3768 for (i = 0; i < func_param_count; i++)
3770 struct access *repr;
3772 for (repr = representatives[i];
3773 repr;
3774 repr = repr->next_grp)
3776 struct access *access;
3777 bitmap visited;
3778 ao_ref ar;
3780 if (no_accesses_p (repr))
3781 continue;
3782 if (!POINTER_TYPE_P (TREE_TYPE (repr->base))
3783 || repr->grp_maybe_modified)
3784 continue;
3786 ao_ref_init (&ar, repr->expr);
3787 visited = BITMAP_ALLOC (NULL);
3788 for (access = repr; access; access = access->next_sibling)
3790 /* All accesses are read ones, otherwise grp_maybe_modified would
3791 be trivially set. */
3792 walk_aliased_vdefs (&ar, gimple_vuse (access->stmt),
3793 mark_maybe_modified, repr, &visited);
3794 if (repr->grp_maybe_modified)
3795 break;
3797 BITMAP_FREE (visited);
3802 /* Propagate distances in bb_dereferences in the opposite direction than the
3803 control flow edges, in each step storing the maximum of the current value
3804 and the minimum of all successors. These steps are repeated until the table
3805 stabilizes. Note that BBs which might terminate the functions (according to
3806 final_bbs bitmap) never updated in this way. */
3808 static void
3809 propagate_dereference_distances (void)
3811 basic_block bb;
3813 auto_vec<basic_block> queue (last_basic_block_for_fn (cfun));
3814 queue.quick_push (ENTRY_BLOCK_PTR_FOR_FN (cfun));
3815 FOR_EACH_BB_FN (bb, cfun)
3817 queue.quick_push (bb);
3818 bb->aux = bb;
3821 while (!queue.is_empty ())
3823 edge_iterator ei;
3824 edge e;
3825 bool change = false;
3826 int i;
3828 bb = queue.pop ();
3829 bb->aux = NULL;
3831 if (bitmap_bit_p (final_bbs, bb->index))
3832 continue;
3834 for (i = 0; i < func_param_count; i++)
3836 int idx = bb->index * func_param_count + i;
3837 bool first = true;
3838 HOST_WIDE_INT inh = 0;
3840 FOR_EACH_EDGE (e, ei, bb->succs)
3842 int succ_idx = e->dest->index * func_param_count + i;
3844 if (e->src == EXIT_BLOCK_PTR_FOR_FN (cfun))
3845 continue;
3847 if (first)
3849 first = false;
3850 inh = bb_dereferences [succ_idx];
3852 else if (bb_dereferences [succ_idx] < inh)
3853 inh = bb_dereferences [succ_idx];
3856 if (!first && bb_dereferences[idx] < inh)
3858 bb_dereferences[idx] = inh;
3859 change = true;
3863 if (change && !bitmap_bit_p (final_bbs, bb->index))
3864 FOR_EACH_EDGE (e, ei, bb->preds)
3866 if (e->src->aux)
3867 continue;
3869 e->src->aux = e->src;
3870 queue.quick_push (e->src);
3875 /* Dump a dereferences TABLE with heading STR to file F. */
3877 static void
3878 dump_dereferences_table (FILE *f, const char *str, HOST_WIDE_INT *table)
3880 basic_block bb;
3882 fprintf (dump_file, str);
3883 FOR_BB_BETWEEN (bb, ENTRY_BLOCK_PTR_FOR_FN (cfun),
3884 EXIT_BLOCK_PTR_FOR_FN (cfun), next_bb)
3886 fprintf (f, "%4i %i ", bb->index, bitmap_bit_p (final_bbs, bb->index));
3887 if (bb != EXIT_BLOCK_PTR_FOR_FN (cfun))
3889 int i;
3890 for (i = 0; i < func_param_count; i++)
3892 int idx = bb->index * func_param_count + i;
3893 fprintf (f, " %4" HOST_WIDE_INT_PRINT "d", table[idx]);
3896 fprintf (f, "\n");
3898 fprintf (dump_file, "\n");
3901 /* Determine what (parts of) parameters passed by reference that are not
3902 assigned to are not certainly dereferenced in this function and thus the
3903 dereferencing cannot be safely moved to the caller without potentially
3904 introducing a segfault. Mark such REPRESENTATIVES as
3905 grp_not_necessarilly_dereferenced.
3907 The dereferenced maximum "distance," i.e. the offset + size of the accessed
3908 part is calculated rather than simple booleans are calculated for each
3909 pointer parameter to handle cases when only a fraction of the whole
3910 aggregate is allocated (see testsuite/gcc.c-torture/execute/ipa-sra-2.c for
3911 an example).
3913 The maximum dereference distances for each pointer parameter and BB are
3914 already stored in bb_dereference. This routine simply propagates these
3915 values upwards by propagate_dereference_distances and then compares the
3916 distances of individual parameters in the ENTRY BB to the equivalent
3917 distances of each representative of a (fraction of a) parameter. */
3919 static void
3920 analyze_caller_dereference_legality (vec<access_p> representatives)
3922 int i;
3924 if (dump_file && (dump_flags & TDF_DETAILS))
3925 dump_dereferences_table (dump_file,
3926 "Dereference table before propagation:\n",
3927 bb_dereferences);
3929 propagate_dereference_distances ();
3931 if (dump_file && (dump_flags & TDF_DETAILS))
3932 dump_dereferences_table (dump_file,
3933 "Dereference table after propagation:\n",
3934 bb_dereferences);
3936 for (i = 0; i < func_param_count; i++)
3938 struct access *repr = representatives[i];
3939 int idx = ENTRY_BLOCK_PTR_FOR_FN (cfun)->index * func_param_count + i;
3941 if (!repr || no_accesses_p (repr))
3942 continue;
3946 if ((repr->offset + repr->size) > bb_dereferences[idx])
3947 repr->grp_not_necessarilly_dereferenced = 1;
3948 repr = repr->next_grp;
3950 while (repr);
3954 /* Return the representative access for the parameter declaration PARM if it is
3955 a scalar passed by reference which is not written to and the pointer value
3956 is not used directly. Thus, if it is legal to dereference it in the caller
3957 and we can rule out modifications through aliases, such parameter should be
3958 turned into one passed by value. Return NULL otherwise. */
3960 static struct access *
3961 unmodified_by_ref_scalar_representative (tree parm)
3963 int i, access_count;
3964 struct access *repr;
3965 vec<access_p> *access_vec;
3967 access_vec = get_base_access_vector (parm);
3968 gcc_assert (access_vec);
3969 repr = (*access_vec)[0];
3970 if (repr->write)
3971 return NULL;
3972 repr->group_representative = repr;
3974 access_count = access_vec->length ();
3975 for (i = 1; i < access_count; i++)
3977 struct access *access = (*access_vec)[i];
3978 if (access->write)
3979 return NULL;
3980 access->group_representative = repr;
3981 access->next_sibling = repr->next_sibling;
3982 repr->next_sibling = access;
3985 repr->grp_read = 1;
3986 repr->grp_scalar_ptr = 1;
3987 return repr;
3990 /* Return true iff this ACCESS precludes IPA-SRA of the parameter it is
3991 associated with. REQ_ALIGN is the minimum required alignment. */
3993 static bool
3994 access_precludes_ipa_sra_p (struct access *access, unsigned int req_align)
3996 unsigned int exp_align;
3997 /* Avoid issues such as the second simple testcase in PR 42025. The problem
3998 is incompatible assign in a call statement (and possibly even in asm
3999 statements). This can be relaxed by using a new temporary but only for
4000 non-TREE_ADDRESSABLE types and is probably not worth the complexity. (In
4001 intraprocedural SRA we deal with this by keeping the old aggregate around,
4002 something we cannot do in IPA-SRA.) */
4003 if (access->write
4004 && (is_gimple_call (access->stmt)
4005 || gimple_code (access->stmt) == GIMPLE_ASM))
4006 return true;
4008 exp_align = get_object_alignment (access->expr);
4009 if (exp_align < req_align)
4010 return true;
4012 return false;
4016 /* Sort collected accesses for parameter PARM, identify representatives for
4017 each accessed region and link them together. Return NULL if there are
4018 different but overlapping accesses, return the special ptr value meaning
4019 there are no accesses for this parameter if that is the case and return the
4020 first representative otherwise. Set *RO_GRP if there is a group of accesses
4021 with only read (i.e. no write) accesses. */
4023 static struct access *
4024 splice_param_accesses (tree parm, bool *ro_grp)
4026 int i, j, access_count, group_count;
4027 int agg_size, total_size = 0;
4028 struct access *access, *res, **prev_acc_ptr = &res;
4029 vec<access_p> *access_vec;
4031 access_vec = get_base_access_vector (parm);
4032 if (!access_vec)
4033 return &no_accesses_representant;
4034 access_count = access_vec->length ();
4036 access_vec->qsort (compare_access_positions);
4038 i = 0;
4039 total_size = 0;
4040 group_count = 0;
4041 while (i < access_count)
4043 bool modification;
4044 tree a1_alias_type;
4045 access = (*access_vec)[i];
4046 modification = access->write;
4047 if (access_precludes_ipa_sra_p (access, TYPE_ALIGN (access->type)))
4048 return NULL;
4049 a1_alias_type = reference_alias_ptr_type (access->expr);
4051 /* Access is about to become group representative unless we find some
4052 nasty overlap which would preclude us from breaking this parameter
4053 apart. */
4055 j = i + 1;
4056 while (j < access_count)
4058 struct access *ac2 = (*access_vec)[j];
4059 if (ac2->offset != access->offset)
4061 /* All or nothing law for parameters. */
4062 if (access->offset + access->size > ac2->offset)
4063 return NULL;
4064 else
4065 break;
4067 else if (ac2->size != access->size)
4068 return NULL;
4070 if (access_precludes_ipa_sra_p (ac2, TYPE_ALIGN (access->type))
4071 || (ac2->type != access->type
4072 && (TREE_ADDRESSABLE (ac2->type)
4073 || TREE_ADDRESSABLE (access->type)))
4074 || (reference_alias_ptr_type (ac2->expr) != a1_alias_type))
4075 return NULL;
4077 modification |= ac2->write;
4078 ac2->group_representative = access;
4079 ac2->next_sibling = access->next_sibling;
4080 access->next_sibling = ac2;
4081 j++;
4084 group_count++;
4085 access->grp_maybe_modified = modification;
4086 if (!modification)
4087 *ro_grp = true;
4088 *prev_acc_ptr = access;
4089 prev_acc_ptr = &access->next_grp;
4090 total_size += access->size;
4091 i = j;
4094 if (POINTER_TYPE_P (TREE_TYPE (parm)))
4095 agg_size = tree_to_uhwi (TYPE_SIZE (TREE_TYPE (TREE_TYPE (parm))));
4096 else
4097 agg_size = tree_to_uhwi (TYPE_SIZE (TREE_TYPE (parm)));
4098 if (total_size >= agg_size)
4099 return NULL;
4101 gcc_assert (group_count > 0);
4102 return res;
4105 /* Decide whether parameters with representative accesses given by REPR should
4106 be reduced into components. */
4108 static int
4109 decide_one_param_reduction (struct access *repr)
4111 int total_size, cur_parm_size, agg_size, new_param_count, parm_size_limit;
4112 bool by_ref;
4113 tree parm;
4115 parm = repr->base;
4116 cur_parm_size = tree_to_uhwi (TYPE_SIZE (TREE_TYPE (parm)));
4117 gcc_assert (cur_parm_size > 0);
4119 if (POINTER_TYPE_P (TREE_TYPE (parm)))
4121 by_ref = true;
4122 agg_size = tree_to_uhwi (TYPE_SIZE (TREE_TYPE (TREE_TYPE (parm))));
4124 else
4126 by_ref = false;
4127 agg_size = cur_parm_size;
4130 if (dump_file)
4132 struct access *acc;
4133 fprintf (dump_file, "Evaluating PARAM group sizes for ");
4134 print_generic_expr (dump_file, parm, 0);
4135 fprintf (dump_file, " (UID: %u): \n", DECL_UID (parm));
4136 for (acc = repr; acc; acc = acc->next_grp)
4137 dump_access (dump_file, acc, true);
4140 total_size = 0;
4141 new_param_count = 0;
4143 for (; repr; repr = repr->next_grp)
4145 gcc_assert (parm == repr->base);
4147 /* Taking the address of a non-addressable field is verboten. */
4148 if (by_ref && repr->non_addressable)
4149 return 0;
4151 /* Do not decompose a non-BLKmode param in a way that would
4152 create BLKmode params. Especially for by-reference passing
4153 (thus, pointer-type param) this is hardly worthwhile. */
4154 if (DECL_MODE (parm) != BLKmode
4155 && TYPE_MODE (repr->type) == BLKmode)
4156 return 0;
4158 if (!by_ref || (!repr->grp_maybe_modified
4159 && !repr->grp_not_necessarilly_dereferenced))
4160 total_size += repr->size;
4161 else
4162 total_size += cur_parm_size;
4164 new_param_count++;
4167 gcc_assert (new_param_count > 0);
4169 if (optimize_function_for_size_p (cfun))
4170 parm_size_limit = cur_parm_size;
4171 else
4172 parm_size_limit = (PARAM_VALUE (PARAM_IPA_SRA_PTR_GROWTH_FACTOR)
4173 * cur_parm_size);
4175 if (total_size < agg_size
4176 && total_size <= parm_size_limit)
4178 if (dump_file)
4179 fprintf (dump_file, " ....will be split into %i components\n",
4180 new_param_count);
4181 return new_param_count;
4183 else
4184 return 0;
4187 /* The order of the following enums is important, we need to do extra work for
4188 UNUSED_PARAMS, BY_VAL_ACCESSES and UNMODIF_BY_REF_ACCESSES. */
4189 enum ipa_splicing_result { NO_GOOD_ACCESS, UNUSED_PARAMS, BY_VAL_ACCESSES,
4190 MODIF_BY_REF_ACCESSES, UNMODIF_BY_REF_ACCESSES };
4192 /* Identify representatives of all accesses to all candidate parameters for
4193 IPA-SRA. Return result based on what representatives have been found. */
4195 static enum ipa_splicing_result
4196 splice_all_param_accesses (vec<access_p> &representatives)
4198 enum ipa_splicing_result result = NO_GOOD_ACCESS;
4199 tree parm;
4200 struct access *repr;
4202 representatives.create (func_param_count);
4204 for (parm = DECL_ARGUMENTS (current_function_decl);
4205 parm;
4206 parm = DECL_CHAIN (parm))
4208 if (is_unused_scalar_param (parm))
4210 representatives.quick_push (&no_accesses_representant);
4211 if (result == NO_GOOD_ACCESS)
4212 result = UNUSED_PARAMS;
4214 else if (POINTER_TYPE_P (TREE_TYPE (parm))
4215 && is_gimple_reg_type (TREE_TYPE (TREE_TYPE (parm)))
4216 && bitmap_bit_p (candidate_bitmap, DECL_UID (parm)))
4218 repr = unmodified_by_ref_scalar_representative (parm);
4219 representatives.quick_push (repr);
4220 if (repr)
4221 result = UNMODIF_BY_REF_ACCESSES;
4223 else if (bitmap_bit_p (candidate_bitmap, DECL_UID (parm)))
4225 bool ro_grp = false;
4226 repr = splice_param_accesses (parm, &ro_grp);
4227 representatives.quick_push (repr);
4229 if (repr && !no_accesses_p (repr))
4231 if (POINTER_TYPE_P (TREE_TYPE (parm)))
4233 if (ro_grp)
4234 result = UNMODIF_BY_REF_ACCESSES;
4235 else if (result < MODIF_BY_REF_ACCESSES)
4236 result = MODIF_BY_REF_ACCESSES;
4238 else if (result < BY_VAL_ACCESSES)
4239 result = BY_VAL_ACCESSES;
4241 else if (no_accesses_p (repr) && (result == NO_GOOD_ACCESS))
4242 result = UNUSED_PARAMS;
4244 else
4245 representatives.quick_push (NULL);
4248 if (result == NO_GOOD_ACCESS)
4250 representatives.release ();
4251 return NO_GOOD_ACCESS;
4254 return result;
4257 /* Return the index of BASE in PARMS. Abort if it is not found. */
4259 static inline int
4260 get_param_index (tree base, vec<tree> parms)
4262 int i, len;
4264 len = parms.length ();
4265 for (i = 0; i < len; i++)
4266 if (parms[i] == base)
4267 return i;
4268 gcc_unreachable ();
4271 /* Convert the decisions made at the representative level into compact
4272 parameter adjustments. REPRESENTATIVES are pointers to first
4273 representatives of each param accesses, ADJUSTMENTS_COUNT is the expected
4274 final number of adjustments. */
4276 static ipa_parm_adjustment_vec
4277 turn_representatives_into_adjustments (vec<access_p> representatives,
4278 int adjustments_count)
4280 vec<tree> parms;
4281 ipa_parm_adjustment_vec adjustments;
4282 tree parm;
4283 int i;
4285 gcc_assert (adjustments_count > 0);
4286 parms = ipa_get_vector_of_formal_parms (current_function_decl);
4287 adjustments.create (adjustments_count);
4288 parm = DECL_ARGUMENTS (current_function_decl);
4289 for (i = 0; i < func_param_count; i++, parm = DECL_CHAIN (parm))
4291 struct access *repr = representatives[i];
4293 if (!repr || no_accesses_p (repr))
4295 struct ipa_parm_adjustment adj;
4297 memset (&adj, 0, sizeof (adj));
4298 adj.base_index = get_param_index (parm, parms);
4299 adj.base = parm;
4300 if (!repr)
4301 adj.op = IPA_PARM_OP_COPY;
4302 else
4303 adj.op = IPA_PARM_OP_REMOVE;
4304 adj.arg_prefix = "ISRA";
4305 adjustments.quick_push (adj);
4307 else
4309 struct ipa_parm_adjustment adj;
4310 int index = get_param_index (parm, parms);
4312 for (; repr; repr = repr->next_grp)
4314 memset (&adj, 0, sizeof (adj));
4315 gcc_assert (repr->base == parm);
4316 adj.base_index = index;
4317 adj.base = repr->base;
4318 adj.type = repr->type;
4319 adj.alias_ptr_type = reference_alias_ptr_type (repr->expr);
4320 adj.offset = repr->offset;
4321 adj.by_ref = (POINTER_TYPE_P (TREE_TYPE (repr->base))
4322 && (repr->grp_maybe_modified
4323 || repr->grp_not_necessarilly_dereferenced));
4324 adj.arg_prefix = "ISRA";
4325 adjustments.quick_push (adj);
4329 parms.release ();
4330 return adjustments;
4333 /* Analyze the collected accesses and produce a plan what to do with the
4334 parameters in the form of adjustments, NULL meaning nothing. */
4336 static ipa_parm_adjustment_vec
4337 analyze_all_param_acesses (void)
4339 enum ipa_splicing_result repr_state;
4340 bool proceed = false;
4341 int i, adjustments_count = 0;
4342 vec<access_p> representatives;
4343 ipa_parm_adjustment_vec adjustments;
4345 repr_state = splice_all_param_accesses (representatives);
4346 if (repr_state == NO_GOOD_ACCESS)
4347 return ipa_parm_adjustment_vec ();
4349 /* If there are any parameters passed by reference which are not modified
4350 directly, we need to check whether they can be modified indirectly. */
4351 if (repr_state == UNMODIF_BY_REF_ACCESSES)
4353 analyze_caller_dereference_legality (representatives);
4354 analyze_modified_params (representatives);
4357 for (i = 0; i < func_param_count; i++)
4359 struct access *repr = representatives[i];
4361 if (repr && !no_accesses_p (repr))
4363 if (repr->grp_scalar_ptr)
4365 adjustments_count++;
4366 if (repr->grp_not_necessarilly_dereferenced
4367 || repr->grp_maybe_modified)
4368 representatives[i] = NULL;
4369 else
4371 proceed = true;
4372 sra_stats.scalar_by_ref_to_by_val++;
4375 else
4377 int new_components = decide_one_param_reduction (repr);
4379 if (new_components == 0)
4381 representatives[i] = NULL;
4382 adjustments_count++;
4384 else
4386 adjustments_count += new_components;
4387 sra_stats.aggregate_params_reduced++;
4388 sra_stats.param_reductions_created += new_components;
4389 proceed = true;
4393 else
4395 if (no_accesses_p (repr))
4397 proceed = true;
4398 sra_stats.deleted_unused_parameters++;
4400 adjustments_count++;
4404 if (!proceed && dump_file)
4405 fprintf (dump_file, "NOT proceeding to change params.\n");
4407 if (proceed)
4408 adjustments = turn_representatives_into_adjustments (representatives,
4409 adjustments_count);
4410 else
4411 adjustments = ipa_parm_adjustment_vec ();
4413 representatives.release ();
4414 return adjustments;
4417 /* If a parameter replacement identified by ADJ does not yet exist in the form
4418 of declaration, create it and record it, otherwise return the previously
4419 created one. */
4421 static tree
4422 get_replaced_param_substitute (struct ipa_parm_adjustment *adj)
4424 tree repl;
4425 if (!adj->new_ssa_base)
4427 char *pretty_name = make_fancy_name (adj->base);
4429 repl = create_tmp_reg (TREE_TYPE (adj->base), "ISR");
4430 DECL_NAME (repl) = get_identifier (pretty_name);
4431 obstack_free (&name_obstack, pretty_name);
4433 adj->new_ssa_base = repl;
4435 else
4436 repl = adj->new_ssa_base;
4437 return repl;
4440 /* Find the first adjustment for a particular parameter BASE in a vector of
4441 ADJUSTMENTS which is not a copy_param. Return NULL if there is no such
4442 adjustment. */
4444 static struct ipa_parm_adjustment *
4445 get_adjustment_for_base (ipa_parm_adjustment_vec adjustments, tree base)
4447 int i, len;
4449 len = adjustments.length ();
4450 for (i = 0; i < len; i++)
4452 struct ipa_parm_adjustment *adj;
4454 adj = &adjustments[i];
4455 if (adj->op != IPA_PARM_OP_COPY && adj->base == base)
4456 return adj;
4459 return NULL;
4462 /* If the statement STMT defines an SSA_NAME of a parameter which is to be
4463 removed because its value is not used, replace the SSA_NAME with a one
4464 relating to a created VAR_DECL together all of its uses and return true.
4465 ADJUSTMENTS is a pointer to an adjustments vector. */
4467 static bool
4468 replace_removed_params_ssa_names (gimple stmt,
4469 ipa_parm_adjustment_vec adjustments)
4471 struct ipa_parm_adjustment *adj;
4472 tree lhs, decl, repl, name;
4474 if (gimple_code (stmt) == GIMPLE_PHI)
4475 lhs = gimple_phi_result (stmt);
4476 else if (is_gimple_assign (stmt))
4477 lhs = gimple_assign_lhs (stmt);
4478 else if (is_gimple_call (stmt))
4479 lhs = gimple_call_lhs (stmt);
4480 else
4481 gcc_unreachable ();
4483 if (TREE_CODE (lhs) != SSA_NAME)
4484 return false;
4486 decl = SSA_NAME_VAR (lhs);
4487 if (decl == NULL_TREE
4488 || TREE_CODE (decl) != PARM_DECL)
4489 return false;
4491 adj = get_adjustment_for_base (adjustments, decl);
4492 if (!adj)
4493 return false;
4495 repl = get_replaced_param_substitute (adj);
4496 name = make_ssa_name (repl, stmt);
4498 if (dump_file)
4500 fprintf (dump_file, "replacing an SSA name of a removed param ");
4501 print_generic_expr (dump_file, lhs, 0);
4502 fprintf (dump_file, " with ");
4503 print_generic_expr (dump_file, name, 0);
4504 fprintf (dump_file, "\n");
4507 if (is_gimple_assign (stmt))
4508 gimple_assign_set_lhs (stmt, name);
4509 else if (is_gimple_call (stmt))
4510 gimple_call_set_lhs (stmt, name);
4511 else
4512 gimple_phi_set_result (stmt, name);
4514 replace_uses_by (lhs, name);
4515 release_ssa_name (lhs);
4516 return true;
4519 /* If the statement pointed to by STMT_PTR contains any expressions that need
4520 to replaced with a different one as noted by ADJUSTMENTS, do so. Handle any
4521 potential type incompatibilities (GSI is used to accommodate conversion
4522 statements and must point to the statement). Return true iff the statement
4523 was modified. */
4525 static bool
4526 sra_ipa_modify_assign (gimple *stmt_ptr, gimple_stmt_iterator *gsi,
4527 ipa_parm_adjustment_vec adjustments)
4529 gimple stmt = *stmt_ptr;
4530 tree *lhs_p, *rhs_p;
4531 bool any;
4533 if (!gimple_assign_single_p (stmt))
4534 return false;
4536 rhs_p = gimple_assign_rhs1_ptr (stmt);
4537 lhs_p = gimple_assign_lhs_ptr (stmt);
4539 any = ipa_modify_expr (rhs_p, false, adjustments);
4540 any |= ipa_modify_expr (lhs_p, false, adjustments);
4541 if (any)
4543 tree new_rhs = NULL_TREE;
4545 if (!useless_type_conversion_p (TREE_TYPE (*lhs_p), TREE_TYPE (*rhs_p)))
4547 if (TREE_CODE (*rhs_p) == CONSTRUCTOR)
4549 /* V_C_Es of constructors can cause trouble (PR 42714). */
4550 if (is_gimple_reg_type (TREE_TYPE (*lhs_p)))
4551 *rhs_p = build_zero_cst (TREE_TYPE (*lhs_p));
4552 else
4553 *rhs_p = build_constructor (TREE_TYPE (*lhs_p),
4554 NULL);
4556 else
4557 new_rhs = fold_build1_loc (gimple_location (stmt),
4558 VIEW_CONVERT_EXPR, TREE_TYPE (*lhs_p),
4559 *rhs_p);
4561 else if (REFERENCE_CLASS_P (*rhs_p)
4562 && is_gimple_reg_type (TREE_TYPE (*lhs_p))
4563 && !is_gimple_reg (*lhs_p))
4564 /* This can happen when an assignment in between two single field
4565 structures is turned into an assignment in between two pointers to
4566 scalars (PR 42237). */
4567 new_rhs = *rhs_p;
4569 if (new_rhs)
4571 tree tmp = force_gimple_operand_gsi (gsi, new_rhs, true, NULL_TREE,
4572 true, GSI_SAME_STMT);
4574 gimple_assign_set_rhs_from_tree (gsi, tmp);
4577 return true;
4580 return false;
4583 /* Traverse the function body and all modifications as described in
4584 ADJUSTMENTS. Return true iff the CFG has been changed. */
4586 bool
4587 ipa_sra_modify_function_body (ipa_parm_adjustment_vec adjustments)
4589 bool cfg_changed = false;
4590 basic_block bb;
4592 FOR_EACH_BB_FN (bb, cfun)
4594 gimple_stmt_iterator gsi;
4596 for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi))
4597 replace_removed_params_ssa_names (gsi_stmt (gsi), adjustments);
4599 gsi = gsi_start_bb (bb);
4600 while (!gsi_end_p (gsi))
4602 gimple stmt = gsi_stmt (gsi);
4603 bool modified = false;
4604 tree *t;
4605 unsigned i;
4607 switch (gimple_code (stmt))
4609 case GIMPLE_RETURN:
4610 t = gimple_return_retval_ptr (stmt);
4611 if (*t != NULL_TREE)
4612 modified |= ipa_modify_expr (t, true, adjustments);
4613 break;
4615 case GIMPLE_ASSIGN:
4616 modified |= sra_ipa_modify_assign (&stmt, &gsi, adjustments);
4617 modified |= replace_removed_params_ssa_names (stmt, adjustments);
4618 break;
4620 case GIMPLE_CALL:
4621 /* Operands must be processed before the lhs. */
4622 for (i = 0; i < gimple_call_num_args (stmt); i++)
4624 t = gimple_call_arg_ptr (stmt, i);
4625 modified |= ipa_modify_expr (t, true, adjustments);
4628 if (gimple_call_lhs (stmt))
4630 t = gimple_call_lhs_ptr (stmt);
4631 modified |= ipa_modify_expr (t, false, adjustments);
4632 modified |= replace_removed_params_ssa_names (stmt,
4633 adjustments);
4635 break;
4637 case GIMPLE_ASM:
4638 for (i = 0; i < gimple_asm_ninputs (stmt); i++)
4640 t = &TREE_VALUE (gimple_asm_input_op (stmt, i));
4641 modified |= ipa_modify_expr (t, true, adjustments);
4643 for (i = 0; i < gimple_asm_noutputs (stmt); i++)
4645 t = &TREE_VALUE (gimple_asm_output_op (stmt, i));
4646 modified |= ipa_modify_expr (t, false, adjustments);
4648 break;
4650 default:
4651 break;
4654 if (modified)
4656 update_stmt (stmt);
4657 if (maybe_clean_eh_stmt (stmt)
4658 && gimple_purge_dead_eh_edges (gimple_bb (stmt)))
4659 cfg_changed = true;
4661 gsi_next (&gsi);
4665 return cfg_changed;
4668 /* Call gimple_debug_bind_reset_value on all debug statements describing
4669 gimple register parameters that are being removed or replaced. */
4671 static void
4672 sra_ipa_reset_debug_stmts (ipa_parm_adjustment_vec adjustments)
4674 int i, len;
4675 gimple_stmt_iterator *gsip = NULL, gsi;
4677 if (MAY_HAVE_DEBUG_STMTS && single_succ_p (ENTRY_BLOCK_PTR_FOR_FN (cfun)))
4679 gsi = gsi_after_labels (single_succ (ENTRY_BLOCK_PTR_FOR_FN (cfun)));
4680 gsip = &gsi;
4682 len = adjustments.length ();
4683 for (i = 0; i < len; i++)
4685 struct ipa_parm_adjustment *adj;
4686 imm_use_iterator ui;
4687 gimple stmt, def_temp;
4688 tree name, vexpr, copy = NULL_TREE;
4689 use_operand_p use_p;
4691 adj = &adjustments[i];
4692 if (adj->op == IPA_PARM_OP_COPY || !is_gimple_reg (adj->base))
4693 continue;
4694 name = ssa_default_def (cfun, adj->base);
4695 vexpr = NULL;
4696 if (name)
4697 FOR_EACH_IMM_USE_STMT (stmt, ui, name)
4699 if (gimple_clobber_p (stmt))
4701 gimple_stmt_iterator cgsi = gsi_for_stmt (stmt);
4702 unlink_stmt_vdef (stmt);
4703 gsi_remove (&cgsi, true);
4704 release_defs (stmt);
4705 continue;
4707 /* All other users must have been removed by
4708 ipa_sra_modify_function_body. */
4709 gcc_assert (is_gimple_debug (stmt));
4710 if (vexpr == NULL && gsip != NULL)
4712 gcc_assert (TREE_CODE (adj->base) == PARM_DECL);
4713 vexpr = make_node (DEBUG_EXPR_DECL);
4714 def_temp = gimple_build_debug_source_bind (vexpr, adj->base,
4715 NULL);
4716 DECL_ARTIFICIAL (vexpr) = 1;
4717 TREE_TYPE (vexpr) = TREE_TYPE (name);
4718 DECL_MODE (vexpr) = DECL_MODE (adj->base);
4719 gsi_insert_before (gsip, def_temp, GSI_SAME_STMT);
4721 if (vexpr)
4723 FOR_EACH_IMM_USE_ON_STMT (use_p, ui)
4724 SET_USE (use_p, vexpr);
4726 else
4727 gimple_debug_bind_reset_value (stmt);
4728 update_stmt (stmt);
4730 /* Create a VAR_DECL for debug info purposes. */
4731 if (!DECL_IGNORED_P (adj->base))
4733 copy = build_decl (DECL_SOURCE_LOCATION (current_function_decl),
4734 VAR_DECL, DECL_NAME (adj->base),
4735 TREE_TYPE (adj->base));
4736 if (DECL_PT_UID_SET_P (adj->base))
4737 SET_DECL_PT_UID (copy, DECL_PT_UID (adj->base));
4738 TREE_ADDRESSABLE (copy) = TREE_ADDRESSABLE (adj->base);
4739 TREE_READONLY (copy) = TREE_READONLY (adj->base);
4740 TREE_THIS_VOLATILE (copy) = TREE_THIS_VOLATILE (adj->base);
4741 DECL_GIMPLE_REG_P (copy) = DECL_GIMPLE_REG_P (adj->base);
4742 DECL_ARTIFICIAL (copy) = DECL_ARTIFICIAL (adj->base);
4743 DECL_IGNORED_P (copy) = DECL_IGNORED_P (adj->base);
4744 DECL_ABSTRACT_ORIGIN (copy) = DECL_ORIGIN (adj->base);
4745 DECL_SEEN_IN_BIND_EXPR_P (copy) = 1;
4746 SET_DECL_RTL (copy, 0);
4747 TREE_USED (copy) = 1;
4748 DECL_CONTEXT (copy) = current_function_decl;
4749 add_local_decl (cfun, copy);
4750 DECL_CHAIN (copy) =
4751 BLOCK_VARS (DECL_INITIAL (current_function_decl));
4752 BLOCK_VARS (DECL_INITIAL (current_function_decl)) = copy;
4754 if (gsip != NULL && copy && target_for_debug_bind (adj->base))
4756 gcc_assert (TREE_CODE (adj->base) == PARM_DECL);
4757 if (vexpr)
4758 def_temp = gimple_build_debug_bind (copy, vexpr, NULL);
4759 else
4760 def_temp = gimple_build_debug_source_bind (copy, adj->base,
4761 NULL);
4762 gsi_insert_before (gsip, def_temp, GSI_SAME_STMT);
4767 /* Return false if all callers have at least as many actual arguments as there
4768 are formal parameters in the current function and that their types
4769 match. */
4771 static bool
4772 some_callers_have_mismatched_arguments_p (struct cgraph_node *node,
4773 void *data ATTRIBUTE_UNUSED)
4775 struct cgraph_edge *cs;
4776 for (cs = node->callers; cs; cs = cs->next_caller)
4777 if (!callsite_arguments_match_p (cs->call_stmt))
4778 return true;
4780 return false;
4783 /* Convert all callers of NODE. */
4785 static bool
4786 convert_callers_for_node (struct cgraph_node *node,
4787 void *data)
4789 ipa_parm_adjustment_vec *adjustments = (ipa_parm_adjustment_vec *) data;
4790 bitmap recomputed_callers = BITMAP_ALLOC (NULL);
4791 struct cgraph_edge *cs;
4793 for (cs = node->callers; cs; cs = cs->next_caller)
4795 push_cfun (DECL_STRUCT_FUNCTION (cs->caller->decl));
4797 if (dump_file)
4798 fprintf (dump_file, "Adjusting call %s/%i -> %s/%i\n",
4799 xstrdup (cs->caller->name ()),
4800 cs->caller->order,
4801 xstrdup (cs->callee->name ()),
4802 cs->callee->order);
4804 ipa_modify_call_arguments (cs, cs->call_stmt, *adjustments);
4806 pop_cfun ();
4809 for (cs = node->callers; cs; cs = cs->next_caller)
4810 if (bitmap_set_bit (recomputed_callers, cs->caller->uid)
4811 && gimple_in_ssa_p (DECL_STRUCT_FUNCTION (cs->caller->decl)))
4812 compute_inline_parameters (cs->caller, true);
4813 BITMAP_FREE (recomputed_callers);
4815 return true;
4818 /* Convert all callers of NODE to pass parameters as given in ADJUSTMENTS. */
4820 static void
4821 convert_callers (struct cgraph_node *node, tree old_decl,
4822 ipa_parm_adjustment_vec adjustments)
4824 basic_block this_block;
4826 cgraph_for_node_and_aliases (node, convert_callers_for_node,
4827 &adjustments, false);
4829 if (!encountered_recursive_call)
4830 return;
4832 FOR_EACH_BB_FN (this_block, cfun)
4834 gimple_stmt_iterator gsi;
4836 for (gsi = gsi_start_bb (this_block); !gsi_end_p (gsi); gsi_next (&gsi))
4838 gimple stmt = gsi_stmt (gsi);
4839 tree call_fndecl;
4840 if (gimple_code (stmt) != GIMPLE_CALL)
4841 continue;
4842 call_fndecl = gimple_call_fndecl (stmt);
4843 if (call_fndecl == old_decl)
4845 if (dump_file)
4846 fprintf (dump_file, "Adjusting recursive call");
4847 gimple_call_set_fndecl (stmt, node->decl);
4848 ipa_modify_call_arguments (NULL, stmt, adjustments);
4853 return;
4856 /* Perform all the modification required in IPA-SRA for NODE to have parameters
4857 as given in ADJUSTMENTS. Return true iff the CFG has been changed. */
4859 static bool
4860 modify_function (struct cgraph_node *node, ipa_parm_adjustment_vec adjustments)
4862 struct cgraph_node *new_node;
4863 bool cfg_changed;
4864 vec<cgraph_edge_p> redirect_callers = collect_callers_of_node (node);
4866 rebuild_cgraph_edges ();
4867 free_dominance_info (CDI_DOMINATORS);
4868 pop_cfun ();
4870 new_node = cgraph_function_versioning (node, redirect_callers,
4871 NULL,
4872 NULL, false, NULL, NULL, "isra");
4873 redirect_callers.release ();
4875 push_cfun (DECL_STRUCT_FUNCTION (new_node->decl));
4876 ipa_modify_formal_parameters (current_function_decl, adjustments);
4877 cfg_changed = ipa_sra_modify_function_body (adjustments);
4878 sra_ipa_reset_debug_stmts (adjustments);
4879 convert_callers (new_node, node->decl, adjustments);
4880 cgraph_make_node_local (new_node);
4881 return cfg_changed;
4884 /* If NODE has a caller, return true. */
4886 static bool
4887 has_caller_p (struct cgraph_node *node, void *data ATTRIBUTE_UNUSED)
4889 if (node->callers)
4890 return true;
4891 return false;
4894 /* Return false the function is apparently unsuitable for IPA-SRA based on it's
4895 attributes, return true otherwise. NODE is the cgraph node of the current
4896 function. */
4898 static bool
4899 ipa_sra_preliminary_function_checks (struct cgraph_node *node)
4901 if (!cgraph_node_can_be_local_p (node))
4903 if (dump_file)
4904 fprintf (dump_file, "Function not local to this compilation unit.\n");
4905 return false;
4908 if (!node->local.can_change_signature)
4910 if (dump_file)
4911 fprintf (dump_file, "Function can not change signature.\n");
4912 return false;
4915 if (!tree_versionable_function_p (node->decl))
4917 if (dump_file)
4918 fprintf (dump_file, "Function is not versionable.\n");
4919 return false;
4922 if (!opt_for_fn (node->decl, optimize)
4923 || !opt_for_fn (node->decl, flag_ipa_sra))
4925 if (dump_file)
4926 fprintf (dump_file, "Function not optimized.\n");
4927 return false;
4930 if (DECL_VIRTUAL_P (current_function_decl))
4932 if (dump_file)
4933 fprintf (dump_file, "Function is a virtual method.\n");
4934 return false;
4937 if ((DECL_COMDAT (node->decl) || DECL_EXTERNAL (node->decl))
4938 && inline_summary (node)->size >= MAX_INLINE_INSNS_AUTO)
4940 if (dump_file)
4941 fprintf (dump_file, "Function too big to be made truly local.\n");
4942 return false;
4945 if (!cgraph_for_node_and_aliases (node, has_caller_p, NULL, true))
4947 if (dump_file)
4948 fprintf (dump_file,
4949 "Function has no callers in this compilation unit.\n");
4950 return false;
4953 if (cfun->stdarg)
4955 if (dump_file)
4956 fprintf (dump_file, "Function uses stdarg. \n");
4957 return false;
4960 if (TYPE_ATTRIBUTES (TREE_TYPE (node->decl)))
4961 return false;
4963 if (DECL_DISREGARD_INLINE_LIMITS (node->decl))
4965 if (dump_file)
4966 fprintf (dump_file, "Always inline function will be inlined "
4967 "anyway. \n");
4968 return false;
4971 return true;
4974 /* Perform early interprocedural SRA. */
4976 static unsigned int
4977 ipa_early_sra (void)
4979 struct cgraph_node *node = cgraph_get_node (current_function_decl);
4980 ipa_parm_adjustment_vec adjustments;
4981 int ret = 0;
4983 if (!ipa_sra_preliminary_function_checks (node))
4984 return 0;
4986 sra_initialize ();
4987 sra_mode = SRA_MODE_EARLY_IPA;
4989 if (!find_param_candidates ())
4991 if (dump_file)
4992 fprintf (dump_file, "Function has no IPA-SRA candidates.\n");
4993 goto simple_out;
4996 if (cgraph_for_node_and_aliases (node,
4997 some_callers_have_mismatched_arguments_p,
4998 NULL, true))
5000 if (dump_file)
5001 fprintf (dump_file, "There are callers with insufficient number of "
5002 "arguments or arguments with type mismatches.\n");
5003 goto simple_out;
5006 bb_dereferences = XCNEWVEC (HOST_WIDE_INT,
5007 func_param_count
5008 * last_basic_block_for_fn (cfun));
5009 final_bbs = BITMAP_ALLOC (NULL);
5011 scan_function ();
5012 if (encountered_apply_args)
5014 if (dump_file)
5015 fprintf (dump_file, "Function calls __builtin_apply_args().\n");
5016 goto out;
5019 if (encountered_unchangable_recursive_call)
5021 if (dump_file)
5022 fprintf (dump_file, "Function calls itself with insufficient "
5023 "number of arguments.\n");
5024 goto out;
5027 adjustments = analyze_all_param_acesses ();
5028 if (!adjustments.exists ())
5029 goto out;
5030 if (dump_file)
5031 ipa_dump_param_adjustments (dump_file, adjustments, current_function_decl);
5033 if (modify_function (node, adjustments))
5034 ret = TODO_update_ssa | TODO_cleanup_cfg;
5035 else
5036 ret = TODO_update_ssa;
5037 adjustments.release ();
5039 statistics_counter_event (cfun, "Unused parameters deleted",
5040 sra_stats.deleted_unused_parameters);
5041 statistics_counter_event (cfun, "Scalar parameters converted to by-value",
5042 sra_stats.scalar_by_ref_to_by_val);
5043 statistics_counter_event (cfun, "Aggregate parameters broken up",
5044 sra_stats.aggregate_params_reduced);
5045 statistics_counter_event (cfun, "Aggregate parameter components created",
5046 sra_stats.param_reductions_created);
5048 out:
5049 BITMAP_FREE (final_bbs);
5050 free (bb_dereferences);
5051 simple_out:
5052 sra_deinitialize ();
5053 return ret;
5056 /* Return if early ipa sra shall be performed. */
5057 static bool
5058 ipa_early_sra_gate (void)
5060 return flag_ipa_sra && dbg_cnt (eipa_sra);
5063 namespace {
5065 const pass_data pass_data_early_ipa_sra =
5067 GIMPLE_PASS, /* type */
5068 "eipa_sra", /* name */
5069 OPTGROUP_NONE, /* optinfo_flags */
5070 true, /* has_gate */
5071 true, /* has_execute */
5072 TV_IPA_SRA, /* tv_id */
5073 0, /* properties_required */
5074 0, /* properties_provided */
5075 0, /* properties_destroyed */
5076 0, /* todo_flags_start */
5077 TODO_dump_symtab, /* todo_flags_finish */
5080 class pass_early_ipa_sra : public gimple_opt_pass
5082 public:
5083 pass_early_ipa_sra (gcc::context *ctxt)
5084 : gimple_opt_pass (pass_data_early_ipa_sra, ctxt)
5087 /* opt_pass methods: */
5088 bool gate () { return ipa_early_sra_gate (); }
5089 unsigned int execute () { return ipa_early_sra (); }
5091 }; // class pass_early_ipa_sra
5093 } // anon namespace
5095 gimple_opt_pass *
5096 make_pass_early_ipa_sra (gcc::context *ctxt)
5098 return new pass_early_ipa_sra (ctxt);