[ARM 4/5 big.LITTLE] Add support for -mcpu=cortex-a57
[official-gcc.git] / gcc / ipa-devirt.c
blobf5b5926504c73ec0fcf7cfc3b736ac598e641c36
1 /* Basic IPA utilities for type inheritance graph construction and
2 devirtualization.
3 Copyright (C) 2013 Free Software Foundation, Inc.
4 Contributed by Jan Hubicka
6 This file is part of GCC.
8 GCC is free software; you can redistribute it and/or modify it under
9 the terms of the GNU General Public License as published by the Free
10 Software Foundation; either version 3, or (at your option) any later
11 version.
13 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
14 WARRANTY; without even the implied warranty of MERCHANTABILITY or
15 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
16 for more details.
18 You should have received a copy of the GNU General Public License
19 along with GCC; see the file COPYING3. If not see
20 <http://www.gnu.org/licenses/>. */
22 /* Brief vocalburary:
23 ODR = One Definition Rule
24 In short, the ODR states that:
25 1 In any translation unit, a template, type, function, or object can
26 have no more than one definition. Some of these can have any number
27 of declarations. A definition provides an instance.
28 2 In the entire program, an object or non-inline function cannot have
29 more than one definition; if an object or function is used, it must
30 have exactly one definition. You can declare an object or function
31 that is never used, in which case you don't have to provide
32 a definition. In no event can there be more than one definition.
33 3 Some things, like types, templates, and extern inline functions, can
34 be defined in more than one translation unit. For a given entity,
35 each definition must be the same. Non-extern objects and functions
36 in different translation units are different entities, even if their
37 names and types are the same.
39 OTR = OBJ_TYPE_REF
40 This is the Gimple representation of type information of a polymorphic call.
41 It contains two parameters:
42 otr_type is a type of class whose method is called.
43 otr_token is the index into virtual table where address is taken.
45 BINFO
46 This is the type inheritance information attached to each tree
47 RECORD_TYPE by the C++ frotend. It provides information about base
48 types and virtual tables.
50 BINFO is linked to the RECORD_TYPE by TYPE_BINFO.
51 BINFO also links to its type by BINFO_TYPE and to the virtual table by
52 BINFO_VTABLE.
54 Base types of a given type are enumerated by BINFO_BASE_BINFO
55 vector. Members of this vectors are not BINFOs associated
56 with a base type. Rather they are new copies of BINFOs
57 (base BINFOs). Their virtual tables may differ from
58 virtual table of the base type. Also BINFO_OFFSET specifies
59 offset of the base within the type.
61 In the case of single inheritance, the virtual table is shared
62 and BINFO_VTABLE of base BINFO is NULL. In the case of multiple
63 inheritance the individual virtual tables are pointer to by
64 BINFO_VTABLE of base binfos (that differs of BINFO_VTABLE of
65 binfo associated to the base type).
67 BINFO lookup for a given base type and offset can be done by
68 get_binfo_at_offset. It returns proper BINFO whose virtual table
69 can be used for lookup of virtual methods associated with the
70 base type.
72 token
73 This is an index of virtual method in virtual table associated
74 to the type defining it. Token can be looked up from OBJ_TYPE_REF
75 or from DECL_VINDEX of a given virtual table.
77 polymorphic (indirect) call
78 This is callgraph represention of virtual method call. Every
79 polymorphic call contains otr_type and otr_token taken from
80 original OBJ_TYPE_REF at callgraph construction time.
82 What we do here:
84 build_type_inheritance_graph triggers a construction of the type inheritance
85 graph.
87 We reconstruct it based on types of methods we see in the unit.
88 This means that the graph is not complete. Types with no methods are not
89 inserted into the graph. Also types without virtual methods are not
90 represented at all, though it may be easy to add this.
92 The inheritance graph is represented as follows:
94 Vertices are structures odr_type. Every odr_type may correspond
95 to one or more tree type nodes that are equivalent by ODR rule.
96 (the multiple type nodes appear only with linktime optimization)
98 Edges are represented by odr_type->base and odr_type->derived_types.
99 At the moment we do not track offsets of types for multiple inheritance.
100 Adding this is easy.
102 possible_polymorphic_call_targets returns, given an parameters found in
103 indirect polymorphic edge all possible polymorphic call targets of the call.
105 pass_ipa_devirt performs simple speculative devirtualization.
108 #include "config.h"
109 #include "system.h"
110 #include "coretypes.h"
111 #include "tm.h"
112 #include "tree.h"
113 #include "print-tree.h"
114 #include "calls.h"
115 #include "cgraph.h"
116 #include "expr.h"
117 #include "tree-pass.h"
118 #include "pointer-set.h"
119 #include "target.h"
120 #include "hash-table.h"
121 #include "tree-pretty-print.h"
122 #include "ipa-utils.h"
123 #include "tree-ssa-alias.h"
124 #include "internal-fn.h"
125 #include "gimple-fold.h"
126 #include "gimple-expr.h"
127 #include "gimple.h"
128 #include "ipa-inline.h"
129 #include "diagnostic.h"
130 #include "tree-dfa.h"
132 /* Dummy polymorphic call context. */
134 const ipa_polymorphic_call_context ipa_dummy_polymorphic_call_context
135 = {0, NULL, false, true};
137 /* Pointer set of all call targets appearing in the cache. */
138 static pointer_set_t *cached_polymorphic_call_targets;
140 /* The node of type inheritance graph. For each type unique in
141 One Defintion Rule (ODR) sense, we produce one node linking all
142 main variants of types equivalent to it, bases and derived types. */
144 struct GTY(()) odr_type_d
146 /* leader type. */
147 tree type;
148 /* All bases. */
149 vec<odr_type> GTY((skip)) bases;
150 /* All derrived types with virtual methods seen in unit. */
151 vec<odr_type> GTY((skip)) derived_types;
153 /* All equivalent types, if more than one. */
154 vec<tree, va_gc> *types;
155 /* Set of all equivalent types, if NON-NULL. */
156 pointer_set_t * GTY((skip)) types_set;
158 /* Unique ID indexing the type in odr_types array. */
159 int id;
160 /* Is it in anonymous namespace? */
161 bool anonymous_namespace;
165 /* Return true if BINFO corresponds to a type with virtual methods.
167 Every type has several BINFOs. One is the BINFO associated by the type
168 while other represents bases of derived types. The BINFOs representing
169 bases do not have BINFO_VTABLE pointer set when this is the single
170 inheritance (because vtables are shared). Look up the BINFO of type
171 and check presence of its vtable. */
173 static inline bool
174 polymorphic_type_binfo_p (tree binfo)
176 /* See if BINFO's type has an virtual table associtated with it. */
177 return BINFO_VTABLE (TYPE_BINFO (BINFO_TYPE (binfo)));
180 /* One Definition Rule hashtable helpers. */
182 struct odr_hasher
184 typedef odr_type_d value_type;
185 typedef union tree_node compare_type;
186 static inline hashval_t hash (const value_type *);
187 static inline bool equal (const value_type *, const compare_type *);
188 static inline void remove (value_type *);
191 /* Produce hash based on type name. */
193 hashval_t
194 hash_type_name (tree t)
196 gcc_checking_assert (TYPE_MAIN_VARIANT (t) == t);
198 /* If not in LTO, all main variants are unique, so we can do
199 pointer hash. */
200 if (!in_lto_p)
201 return htab_hash_pointer (t);
203 /* Anonymous types are unique. */
204 if (type_in_anonymous_namespace_p (t))
205 return htab_hash_pointer (t);
207 /* For polymorphic types, we can simply hash the virtual table. */
208 if (TYPE_BINFO (t) && BINFO_VTABLE (TYPE_BINFO (t)))
210 tree v = BINFO_VTABLE (TYPE_BINFO (t));
211 hashval_t hash = 0;
213 if (TREE_CODE (v) == POINTER_PLUS_EXPR)
215 hash = TREE_INT_CST_LOW (TREE_OPERAND (v, 1));
216 v = TREE_OPERAND (TREE_OPERAND (v, 0), 0);
219 v = DECL_ASSEMBLER_NAME (v);
220 hash = iterative_hash_hashval_t (hash, htab_hash_pointer (v));
221 return hash;
224 /* Rest is not implemented yet. */
225 gcc_unreachable ();
228 /* Return the computed hashcode for ODR_TYPE. */
230 inline hashval_t
231 odr_hasher::hash (const value_type *odr_type)
233 return hash_type_name (odr_type->type);
236 /* Compare types T1 and T2 and return true if they are
237 equivalent. */
239 inline bool
240 odr_hasher::equal (const value_type *t1, const compare_type *ct2)
242 tree t2 = const_cast <tree> (ct2);
244 gcc_checking_assert (TYPE_MAIN_VARIANT (ct2) == ct2);
245 if (t1->type == t2)
246 return true;
247 if (!in_lto_p)
248 return false;
249 return types_same_for_odr (t1->type, t2);
252 /* Free ODR type V. */
254 inline void
255 odr_hasher::remove (value_type *v)
257 v->bases.release ();
258 v->derived_types.release ();
259 if (v->types_set)
260 pointer_set_destroy (v->types_set);
261 ggc_free (v);
264 /* ODR type hash used to lookup ODR type based on tree type node. */
266 typedef hash_table <odr_hasher> odr_hash_type;
267 static odr_hash_type odr_hash;
269 /* ODR types are also stored into ODR_TYPE vector to allow consistent
270 walking. Bases appear before derived types. Vector is garbage collected
271 so we won't end up visiting empty types. */
273 static GTY(()) vec <odr_type, va_gc> *odr_types_ptr;
274 #define odr_types (*odr_types_ptr)
276 /* TYPE is equivalent to VAL by ODR, but its tree representation differs
277 from VAL->type. This may happen in LTO where tree merging did not merge
278 all variants of the same type. It may or may not mean the ODR violation.
279 Add it to the list of duplicates and warn on some violations. */
281 static void
282 add_type_duplicate (odr_type val, tree type)
284 if (!val->types_set)
285 val->types_set = pointer_set_create ();
287 /* See if this duplicate is new. */
288 if (!pointer_set_insert (val->types_set, type))
290 bool merge = true;
291 bool base_mismatch = false;
292 gcc_assert (in_lto_p);
293 vec_safe_push (val->types, type);
294 unsigned int i,j;
296 /* First we compare memory layout. */
297 if (!types_compatible_p (val->type, type))
299 merge = false;
300 if (BINFO_VTABLE (TYPE_BINFO (val->type))
301 && warning_at (DECL_SOURCE_LOCATION (TYPE_NAME (type)), 0,
302 "type %qD violates one definition rule ",
303 type))
304 inform (DECL_SOURCE_LOCATION (TYPE_NAME (val->type)),
305 "a type with the same name but different layout is "
306 "defined in another translation unit");
307 if (cgraph_dump_file)
309 fprintf (cgraph_dump_file, "ODR violation or merging or ODR type bug?\n");
311 print_node (cgraph_dump_file, "", val->type, 0);
312 putc ('\n',cgraph_dump_file);
313 print_node (cgraph_dump_file, "", type, 0);
314 putc ('\n',cgraph_dump_file);
318 /* Next sanity check that bases are the same. If not, we will end
319 up producing wrong answers. */
320 for (j = 0, i = 0; i < BINFO_N_BASE_BINFOS (TYPE_BINFO (type)); i++)
321 if (polymorphic_type_binfo_p (BINFO_BASE_BINFO (TYPE_BINFO (type), i)))
323 odr_type base = get_odr_type
324 (BINFO_TYPE
325 (BINFO_BASE_BINFO (TYPE_BINFO (type),
326 i)),
327 true);
328 if (val->bases.length () <= j || val->bases[j] != base)
329 base_mismatch = true;
330 j++;
332 if (base_mismatch)
334 merge = false;
336 if (warning_at (DECL_SOURCE_LOCATION (TYPE_NAME (type)), 0,
337 "type %qD violates one definition rule ",
338 type))
339 inform (DECL_SOURCE_LOCATION (TYPE_NAME (val->type)),
340 "a type with the same name but different bases is "
341 "defined in another translation unit");
342 if (cgraph_dump_file)
344 fprintf (cgraph_dump_file, "ODR bse violation or merging bug?\n");
346 print_node (cgraph_dump_file, "", val->type, 0);
347 putc ('\n',cgraph_dump_file);
348 print_node (cgraph_dump_file, "", type, 0);
349 putc ('\n',cgraph_dump_file);
353 /* Regularize things a little. During LTO same types may come with
354 different BINFOs. Either because their virtual table was
355 not merged by tree merging and only later at decl merging or
356 because one type comes with external vtable, while other
357 with internal. We want to merge equivalent binfos to conserve
358 memory and streaming overhead.
360 The external vtables are more harmful: they contain references
361 to external declarations of methods that may be defined in the
362 merged LTO unit. For this reason we absolutely need to remove
363 them and replace by internal variants. Not doing so will lead
364 to incomplete answers from possible_polymorphic_call_targets. */
365 if (!flag_ltrans && merge)
367 tree master_binfo = TYPE_BINFO (val->type);
368 tree v1 = BINFO_VTABLE (master_binfo);
369 tree v2 = BINFO_VTABLE (TYPE_BINFO (type));
371 if (TREE_CODE (v1) == POINTER_PLUS_EXPR)
373 gcc_assert (TREE_CODE (v2) == POINTER_PLUS_EXPR
374 && operand_equal_p (TREE_OPERAND (v1, 1),
375 TREE_OPERAND (v2, 1), 0));
376 v1 = TREE_OPERAND (TREE_OPERAND (v1, 0), 0);
377 v2 = TREE_OPERAND (TREE_OPERAND (v2, 0), 0);
379 gcc_assert (DECL_ASSEMBLER_NAME (v1)
380 == DECL_ASSEMBLER_NAME (v2));
382 if (DECL_EXTERNAL (v1) && !DECL_EXTERNAL (v2))
384 unsigned int i;
386 TYPE_BINFO (val->type) = TYPE_BINFO (type);
387 for (i = 0; i < val->types->length (); i++)
389 if (TYPE_BINFO ((*val->types)[i])
390 == master_binfo)
391 TYPE_BINFO ((*val->types)[i]) = TYPE_BINFO (type);
394 else
395 TYPE_BINFO (type) = master_binfo;
400 /* Get ODR type hash entry for TYPE. If INSERT is true, create
401 possibly new entry. */
403 odr_type
404 get_odr_type (tree type, bool insert)
406 odr_type_d **slot;
407 odr_type val;
408 hashval_t hash;
410 type = TYPE_MAIN_VARIANT (type);
411 gcc_checking_assert (TYPE_MAIN_VARIANT (type) == type);
412 hash = hash_type_name (type);
413 slot = odr_hash.find_slot_with_hash (type, hash, insert ? INSERT : NO_INSERT);
414 if (!slot)
415 return NULL;
417 /* See if we already have entry for type. */
418 if (*slot)
420 val = *slot;
422 /* With LTO we need to support multiple tree representation of
423 the same ODR type. */
424 if (val->type != type)
425 add_type_duplicate (val, type);
427 else
429 tree binfo = TYPE_BINFO (type);
430 unsigned int i;
432 val = ggc_alloc_cleared_odr_type_d ();
433 val->type = type;
434 val->bases = vNULL;
435 val->derived_types = vNULL;
436 val->anonymous_namespace = type_in_anonymous_namespace_p (type);
437 *slot = val;
438 for (i = 0; i < BINFO_N_BASE_BINFOS (binfo); i++)
439 /* For now record only polymorphic types. other are
440 pointless for devirtualization and we can not precisely
441 determine ODR equivalency of these during LTO. */
442 if (polymorphic_type_binfo_p (BINFO_BASE_BINFO (binfo, i)))
444 odr_type base = get_odr_type (BINFO_TYPE (BINFO_BASE_BINFO (binfo,
445 i)),
446 true);
447 base->derived_types.safe_push (val);
448 val->bases.safe_push (base);
450 /* First record bases, then add into array so ids are increasing. */
451 if (odr_types_ptr)
452 val->id = odr_types.length ();
453 vec_safe_push (odr_types_ptr, val);
455 return val;
458 /* Dump ODR type T and all its derrived type. INDENT specify indentation for
459 recusive printing. */
461 static void
462 dump_odr_type (FILE *f, odr_type t, int indent=0)
464 unsigned int i;
465 fprintf (f, "%*s type %i: ", indent * 2, "", t->id);
466 print_generic_expr (f, t->type, TDF_SLIM);
467 fprintf (f, "%s\n", t->anonymous_namespace ? " (anonymous namespace)":"");
468 if (TYPE_NAME (t->type))
470 fprintf (f, "%*s defined at: %s:%i\n", indent * 2, "",
471 DECL_SOURCE_FILE (TYPE_NAME (t->type)),
472 DECL_SOURCE_LINE (TYPE_NAME (t->type)));
474 if (t->bases.length ())
476 fprintf (f, "%*s base odr type ids: ", indent * 2, "");
477 for (i = 0; i < t->bases.length (); i++)
478 fprintf (f, " %i", t->bases[i]->id);
479 fprintf (f, "\n");
481 if (t->derived_types.length ())
483 fprintf (f, "%*s derived types:\n", indent * 2, "");
484 for (i = 0; i < t->derived_types.length (); i++)
485 dump_odr_type (f, t->derived_types[i], indent + 1);
487 fprintf (f, "\n");
490 /* Dump the type inheritance graph. */
492 static void
493 dump_type_inheritance_graph (FILE *f)
495 unsigned int i;
496 if (!odr_types_ptr)
497 return;
498 fprintf (f, "\n\nType inheritance graph:\n");
499 for (i = 0; i < odr_types.length (); i++)
501 if (odr_types[i]->bases.length () == 0)
502 dump_odr_type (f, odr_types[i]);
504 for (i = 0; i < odr_types.length (); i++)
506 if (odr_types[i]->types && odr_types[i]->types->length ())
508 unsigned int j;
509 fprintf (f, "Duplicate tree types for odr type %i\n", i);
510 print_node (f, "", odr_types[i]->type, 0);
511 for (j = 0; j < odr_types[i]->types->length (); j++)
513 tree t;
514 fprintf (f, "duplicate #%i\n", j);
515 print_node (f, "", (*odr_types[i]->types)[j], 0);
516 t = (*odr_types[i]->types)[j];
517 while (TYPE_P (t) && TYPE_CONTEXT (t))
519 t = TYPE_CONTEXT (t);
520 print_node (f, "", t, 0);
522 putc ('\n',f);
528 /* Given method type T, return type of class it belongs to.
529 Lookup this pointer and get its type. */
531 tree
532 method_class_type (tree t)
534 tree first_parm_type = TREE_VALUE (TYPE_ARG_TYPES (t));
535 gcc_assert (TREE_CODE (t) == METHOD_TYPE);
537 return TREE_TYPE (first_parm_type);
540 /* Initialize IPA devirt and build inheritance tree graph. */
542 void
543 build_type_inheritance_graph (void)
545 struct cgraph_node *n;
546 FILE *inheritance_dump_file;
547 int flags;
549 if (odr_hash.is_created ())
550 return;
551 timevar_push (TV_IPA_INHERITANCE);
552 inheritance_dump_file = dump_begin (TDI_inheritance, &flags);
553 odr_hash.create (23);
555 /* We reconstruct the graph starting of types of all methods seen in the
556 the unit. */
557 FOR_EACH_FUNCTION (n)
558 if (DECL_VIRTUAL_P (n->decl)
559 && symtab_real_symbol_p (n))
560 get_odr_type (method_class_type (TREE_TYPE (n->decl)), true);
561 if (inheritance_dump_file)
563 dump_type_inheritance_graph (inheritance_dump_file);
564 dump_end (TDI_inheritance, inheritance_dump_file);
566 timevar_pop (TV_IPA_INHERITANCE);
569 /* If TARGET has associated node, record it in the NODES array.
570 if TARGET can not be inserted (for example because its body was
571 already removed and there is no way to refer to it), clear COMPLETEP. */
573 static void
574 maybe_record_node (vec <cgraph_node *> &nodes,
575 tree target, pointer_set_t *inserted,
576 bool *completep)
578 struct cgraph_node *target_node;
579 enum built_in_function fcode;
581 if (!target
582 /* Those are used to mark impossible scenarios. */
583 || (fcode = DECL_FUNCTION_CODE (target))
584 == BUILT_IN_UNREACHABLE
585 || fcode == BUILT_IN_TRAP)
586 return;
588 target_node = cgraph_get_node (target);
590 if (target_node != NULL
591 && (TREE_PUBLIC (target)
592 || target_node->definition)
593 && symtab_real_symbol_p (target_node))
595 gcc_assert (!target_node->global.inlined_to);
596 gcc_assert (symtab_real_symbol_p (target_node));
597 if (!pointer_set_insert (inserted, target))
599 pointer_set_insert (cached_polymorphic_call_targets,
600 target_node);
601 nodes.safe_push (target_node);
604 else if (completep
605 && !type_in_anonymous_namespace_p
606 (method_class_type (TREE_TYPE (target))))
607 *completep = true;
610 /* See if BINFO's type match OUTER_TYPE. If so, lookup
611 BINFO of subtype of OTR_TYPE at OFFSET and in that BINFO find
612 method in vtable and insert method to NODES array.
613 Otherwise recurse to base BINFOs.
614 This match what get_binfo_at_offset does, but with offset
615 being unknown.
617 TYPE_BINFO is binfo holding an virtual table matching
618 BINFO's type. In the case of single inheritance, this
619 is binfo of BINFO's type ancestor (vtable is shared),
620 otherwise it is binfo of BINFO's type.
622 MATCHED_VTABLES tracks virtual tables we already did lookup
623 for virtual function in. INSERTED tracks nodes we already
624 inserted.
626 ANONYMOUS is true if BINFO is part of anonymous namespace.
629 static void
630 record_target_from_binfo (vec <cgraph_node *> &nodes,
631 tree binfo,
632 tree otr_type,
633 tree type_binfo,
634 HOST_WIDE_INT otr_token,
635 tree outer_type,
636 HOST_WIDE_INT offset,
637 pointer_set_t *inserted,
638 pointer_set_t *matched_vtables,
639 bool anonymous)
641 tree type = BINFO_TYPE (binfo);
642 int i;
643 tree base_binfo;
645 gcc_checking_assert (BINFO_VTABLE (type_binfo));
647 if (types_same_for_odr (type, outer_type))
649 tree inner_binfo = get_binfo_at_offset (type_binfo,
650 offset, otr_type);
651 /* For types in anonymous namespace first check if the respective vtable
652 is alive. If not, we know the type can't be called. */
653 if (!flag_ltrans && anonymous)
655 tree vtable = BINFO_VTABLE (inner_binfo);
656 varpool_node *vnode;
658 if (TREE_CODE (vtable) == POINTER_PLUS_EXPR)
659 vtable = TREE_OPERAND (TREE_OPERAND (vtable, 0), 0);
660 vnode = varpool_get_node (vtable);
661 if (!vnode || !vnode->definition)
662 return;
664 gcc_assert (inner_binfo);
665 if (!pointer_set_insert (matched_vtables, BINFO_VTABLE (inner_binfo)))
667 tree target = gimple_get_virt_method_for_binfo (otr_token, inner_binfo);
668 if (target)
669 maybe_record_node (nodes, target, inserted, NULL);
671 return;
674 /* Walk bases. */
675 for (i = 0; BINFO_BASE_ITERATE (binfo, i, base_binfo); i++)
676 /* Walking bases that have no virtual method is pointless excercise. */
677 if (polymorphic_type_binfo_p (base_binfo))
678 record_target_from_binfo (nodes, base_binfo, otr_type,
679 /* In the case of single inheritance,
680 the virtual table is shared with
681 the outer type. */
682 BINFO_VTABLE (base_binfo) ? base_binfo : type_binfo,
683 otr_token, outer_type, offset, inserted,
684 matched_vtables, anonymous);
687 /* Lookup virtual methods matching OTR_TYPE (with OFFSET and OTR_TOKEN)
688 of TYPE, insert them to NODES, recurse into derived nodes.
689 INSERTED is used to avoid duplicate insertions of methods into NODES.
690 MATCHED_VTABLES are used to avoid duplicate walking vtables. */
692 static void
693 possible_polymorphic_call_targets_1 (vec <cgraph_node *> &nodes,
694 pointer_set_t *inserted,
695 pointer_set_t *matched_vtables,
696 tree otr_type,
697 odr_type type,
698 HOST_WIDE_INT otr_token,
699 tree outer_type,
700 HOST_WIDE_INT offset)
702 tree binfo = TYPE_BINFO (type->type);
703 unsigned int i;
705 record_target_from_binfo (nodes, binfo, otr_type, binfo, otr_token,
706 outer_type, offset,
707 inserted, matched_vtables,
708 type->anonymous_namespace);
709 for (i = 0; i < type->derived_types.length (); i++)
710 possible_polymorphic_call_targets_1 (nodes, inserted,
711 matched_vtables,
712 otr_type,
713 type->derived_types[i],
714 otr_token, outer_type, offset);
717 /* Cache of queries for polymorphic call targets.
719 Enumerating all call targets may get expensive when there are many
720 polymorphic calls in the program, so we memoize all the previous
721 queries and avoid duplicated work. */
723 struct polymorphic_call_target_d
725 HOST_WIDE_INT otr_token;
726 ipa_polymorphic_call_context context;
727 odr_type type;
728 vec <cgraph_node *> targets;
729 bool final;
732 /* Polymorphic call target cache helpers. */
734 struct polymorphic_call_target_hasher
736 typedef polymorphic_call_target_d value_type;
737 typedef polymorphic_call_target_d compare_type;
738 static inline hashval_t hash (const value_type *);
739 static inline bool equal (const value_type *, const compare_type *);
740 static inline void remove (value_type *);
743 /* Return the computed hashcode for ODR_QUERY. */
745 inline hashval_t
746 polymorphic_call_target_hasher::hash (const value_type *odr_query)
748 hashval_t hash;
750 hash = iterative_hash_host_wide_int
751 (odr_query->otr_token,
752 odr_query->type->id);
753 hash = iterative_hash_hashval_t (TYPE_UID (odr_query->context.outer_type),
754 hash);
755 hash = iterative_hash_host_wide_int (odr_query->context.offset, hash);
756 return iterative_hash_hashval_t
757 (((int)odr_query->context.maybe_in_construction << 1)
758 | (int)odr_query->context.maybe_derived_type, hash);
761 /* Compare cache entries T1 and T2. */
763 inline bool
764 polymorphic_call_target_hasher::equal (const value_type *t1,
765 const compare_type *t2)
767 return (t1->type == t2->type && t1->otr_token == t2->otr_token
768 && t1->context.offset == t2->context.offset
769 && t1->context.outer_type == t2->context.outer_type
770 && t1->context.maybe_in_construction
771 == t2->context.maybe_in_construction
772 && t1->context.maybe_derived_type == t2->context.maybe_derived_type);
775 /* Remove entry in polymorphic call target cache hash. */
777 inline void
778 polymorphic_call_target_hasher::remove (value_type *v)
780 v->targets.release ();
781 free (v);
784 /* Polymorphic call target query cache. */
786 typedef hash_table <polymorphic_call_target_hasher>
787 polymorphic_call_target_hash_type;
788 static polymorphic_call_target_hash_type polymorphic_call_target_hash;
790 /* Destroy polymorphic call target query cache. */
792 static void
793 free_polymorphic_call_targets_hash ()
795 if (cached_polymorphic_call_targets)
797 polymorphic_call_target_hash.dispose ();
798 pointer_set_destroy (cached_polymorphic_call_targets);
799 cached_polymorphic_call_targets = NULL;
803 /* When virtual function is removed, we may need to flush the cache. */
805 static void
806 devirt_node_removal_hook (struct cgraph_node *n, void *d ATTRIBUTE_UNUSED)
808 if (cached_polymorphic_call_targets
809 && pointer_set_contains (cached_polymorphic_call_targets, n))
810 free_polymorphic_call_targets_hash ();
813 /* CONTEXT->OUTER_TYPE is a type of memory object where object of EXPECTED_TYPE
814 is contained at CONTEXT->OFFSET. Walk the memory representation of
815 CONTEXT->OUTER_TYPE and find the outermost class type that match
816 EXPECTED_TYPE or contain EXPECTED_TYPE as a base. Update CONTEXT
817 to represent it.
819 For example when CONTEXT represents type
820 class A
822 int a;
823 class B b;
825 and we look for type at offset sizeof(int), we end up with B and offset 0.
826 If the same is produced by multiple inheritance, we end up with A and offset
827 sizeof(int).
829 If we can not find corresponding class, give up by setting
830 CONTEXT->OUTER_TYPE to EXPECTED_TYPE and CONTEXT->OFFSET to NULL.
831 Return true when lookup was sucesful. */
833 static bool
834 get_class_context (ipa_polymorphic_call_context *context,
835 tree expected_type)
837 tree type = context->outer_type;
838 HOST_WIDE_INT offset = context->offset;
840 /* Find the sub-object the constant actually refers to and mark whether it is
841 an artificial one (as opposed to a user-defined one). */
842 while (true)
844 HOST_WIDE_INT pos, size;
845 tree fld;
847 /* On a match, just return what we found. */
848 if (TREE_CODE (type) == TREE_CODE (expected_type)
849 && types_same_for_odr (type, expected_type))
851 /* Type can not contain itself on an non-zero offset. In that case
852 just give up. */
853 if (offset != 0)
854 goto give_up;
855 gcc_assert (offset == 0);
856 return true;
859 /* Walk fields and find corresponding on at OFFSET. */
860 if (TREE_CODE (type) == RECORD_TYPE)
862 for (fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
864 if (TREE_CODE (fld) != FIELD_DECL)
865 continue;
867 pos = int_bit_position (fld);
868 size = tree_to_uhwi (DECL_SIZE (fld));
869 if (pos <= offset && (pos + size) > offset)
870 break;
873 if (!fld)
874 goto give_up;
876 type = TREE_TYPE (fld);
877 offset -= pos;
878 /* DECL_ARTIFICIAL represents a basetype. */
879 if (!DECL_ARTIFICIAL (fld))
881 context->outer_type = type;
882 context->offset = offset;
883 /* As soon as we se an field containing the type,
884 we know we are not looking for derivations. */
885 context->maybe_derived_type = false;
888 else if (TREE_CODE (type) == ARRAY_TYPE)
890 tree subtype = TREE_TYPE (type);
892 /* Give up if we don't know array size. */
893 if (!tree_fits_shwi_p (TYPE_SIZE (subtype))
894 || !tree_to_shwi (TYPE_SIZE (subtype)) <= 0)
895 goto give_up;
896 offset = offset % tree_to_shwi (TYPE_SIZE (subtype));
897 type = subtype;
898 context->outer_type = type;
899 context->offset = offset;
900 context->maybe_derived_type = false;
902 /* Give up on anything else. */
903 else
904 goto give_up;
907 /* If we failed to find subtype we look for, give up and fall back to the
908 most generic query. */
909 give_up:
910 context->outer_type = expected_type;
911 context->offset = 0;
912 context->maybe_derived_type = true;
913 return false;
916 /* Return true if OUTER_TYPE contains OTR_TYPE at OFFSET. */
918 static bool
919 contains_type_p (tree outer_type, HOST_WIDE_INT offset,
920 tree otr_type)
922 ipa_polymorphic_call_context context = {offset, outer_type,
923 false, true};
924 return get_class_context (&context, otr_type);
927 /* Given REF call in FNDECL, determine class of the polymorphic
928 call (OTR_TYPE), its token (OTR_TOKEN) and CONTEXT.
929 Return pointer to object described by the context */
931 tree
932 get_polymorphic_call_info (tree fndecl,
933 tree ref,
934 tree *otr_type,
935 HOST_WIDE_INT *otr_token,
936 ipa_polymorphic_call_context *context)
938 tree base_pointer;
939 *otr_type = obj_type_ref_class (ref);
940 *otr_token = tree_to_uhwi (OBJ_TYPE_REF_TOKEN (ref));
942 /* Set up basic info in case we find nothing interesting in the analysis. */
943 context->outer_type = *otr_type;
944 context->offset = 0;
945 base_pointer = OBJ_TYPE_REF_OBJECT (ref);
946 context->maybe_derived_type = true;
947 context->maybe_in_construction = false;
949 /* Walk SSA for outer object. */
952 if (TREE_CODE (base_pointer) == SSA_NAME
953 && !SSA_NAME_IS_DEFAULT_DEF (base_pointer)
954 && SSA_NAME_DEF_STMT (base_pointer)
955 && gimple_assign_single_p (SSA_NAME_DEF_STMT (base_pointer)))
957 base_pointer = gimple_assign_rhs1 (SSA_NAME_DEF_STMT (base_pointer));
958 STRIP_NOPS (base_pointer);
960 else if (TREE_CODE (base_pointer) == ADDR_EXPR)
962 HOST_WIDE_INT size, max_size;
963 HOST_WIDE_INT offset2;
964 tree base = get_ref_base_and_extent (TREE_OPERAND (base_pointer, 0),
965 &offset2, &size, &max_size);
967 /* If this is a varying address, punt. */
968 if ((TREE_CODE (base) == MEM_REF || DECL_P (base))
969 && max_size != -1
970 && max_size == size)
972 /* We found dereference of a pointer. Type of the pointer
973 and MEM_REF is meaningless, but we can look futher. */
974 if (TREE_CODE (base) == MEM_REF)
976 base_pointer = TREE_OPERAND (base, 0);
977 context->offset
978 += offset2 + mem_ref_offset (base).low * BITS_PER_UNIT;
979 context->outer_type = NULL;
981 /* We found base object. In this case the outer_type
982 is known. */
983 else if (DECL_P (base))
985 context->outer_type = TREE_TYPE (base);
986 gcc_assert (!POINTER_TYPE_P (context->outer_type));
988 /* Only type inconsistent programs can have otr_type that is
989 not part of outer type. */
990 if (!contains_type_p (context->outer_type,
991 context->offset, *otr_type))
992 return base_pointer;
993 context->offset += offset2;
994 base_pointer = NULL;
995 /* Make very conservative assumption that all objects
996 may be in construction.
997 TODO: ipa-prop already contains code to tell better.
998 merge it later. */
999 context->maybe_in_construction = true;
1000 context->maybe_derived_type = false;
1001 return base_pointer;
1003 else
1004 break;
1006 else
1007 break;
1009 else if (TREE_CODE (base_pointer) == POINTER_PLUS_EXPR
1010 && tree_fits_uhwi_p (TREE_OPERAND (base_pointer, 1)))
1012 context->offset += tree_to_shwi (TREE_OPERAND (base_pointer, 1))
1013 * BITS_PER_UNIT;
1014 base_pointer = TREE_OPERAND (base_pointer, 0);
1016 else
1017 break;
1019 while (true);
1021 /* Try to determine type of the outer object. */
1022 if (TREE_CODE (base_pointer) == SSA_NAME
1023 && SSA_NAME_IS_DEFAULT_DEF (base_pointer)
1024 && TREE_CODE (SSA_NAME_VAR (base_pointer)) == PARM_DECL)
1026 /* See if parameter is THIS pointer of a method. */
1027 if (TREE_CODE (TREE_TYPE (fndecl)) == METHOD_TYPE
1028 && SSA_NAME_VAR (base_pointer) == DECL_ARGUMENTS (fndecl))
1030 context->outer_type = TREE_TYPE (TREE_TYPE (base_pointer));
1031 gcc_assert (TREE_CODE (context->outer_type) == RECORD_TYPE);
1033 /* Dynamic casting has possibly upcasted the type
1034 in the hiearchy. In this case outer type is less
1035 informative than inner type and we should forget
1036 about it. */
1037 if (!contains_type_p (context->outer_type, context->offset,
1038 *otr_type))
1040 context->outer_type = NULL;
1041 return base_pointer;
1044 /* If the function is constructor or destructor, then
1045 the type is possibly in consturction, but we know
1046 it is not derived type. */
1047 if (DECL_CXX_CONSTRUCTOR_P (fndecl)
1048 || DECL_CXX_DESTRUCTOR_P (fndecl))
1050 context->maybe_in_construction = true;
1051 context->maybe_derived_type = false;
1053 else
1055 context->maybe_derived_type = true;
1056 context->maybe_in_construction = false;
1058 return base_pointer;
1060 /* Non-PODs passed by value are really passed by invisible
1061 reference. In this case we also know the type of the
1062 object. */
1063 if (DECL_BY_REFERENCE (SSA_NAME_VAR (base_pointer)))
1065 context->outer_type = TREE_TYPE (TREE_TYPE (base_pointer));
1066 gcc_assert (!POINTER_TYPE_P (context->outer_type));
1067 /* Only type inconsistent programs can have otr_type that is
1068 not part of outer type. */
1069 if (!contains_type_p (context->outer_type, context->offset,
1070 *otr_type))
1072 context->outer_type = NULL;
1073 gcc_unreachable ();
1074 return base_pointer;
1076 context->maybe_derived_type = false;
1077 context->maybe_in_construction = false;
1078 return base_pointer;
1081 /* TODO: There are multiple ways to derive a type. For instance
1082 if BASE_POINTER is passed to an constructor call prior our refernece.
1083 We do not make this type of flow sensitive analysis yet. */
1084 return base_pointer;
1087 /* Walk bases of OUTER_TYPE that contain OTR_TYPE at OFFSET.
1088 Lookup their respecitve virtual methods for OTR_TOKEN and OTR_TYPE
1089 and insert them to NODES.
1091 MATCHED_VTABLES and INSERTED is used to avoid duplicated work. */
1093 static void
1094 record_targets_from_bases (tree otr_type,
1095 HOST_WIDE_INT otr_token,
1096 tree outer_type,
1097 HOST_WIDE_INT offset,
1098 vec <cgraph_node *> nodes,
1099 pointer_set_t *inserted,
1100 pointer_set_t *matched_vtables,
1101 bool *completep)
1103 while (true)
1105 HOST_WIDE_INT pos, size;
1106 tree base_binfo;
1107 tree fld;
1109 if (types_same_for_odr (outer_type, otr_type))
1110 return;
1112 for (fld = TYPE_FIELDS (outer_type); fld; fld = DECL_CHAIN (fld))
1114 if (TREE_CODE (fld) != FIELD_DECL)
1115 continue;
1117 pos = int_bit_position (fld);
1118 size = tree_to_shwi (DECL_SIZE (fld));
1119 if (pos <= offset && (pos + size) > offset)
1120 break;
1122 /* Within a class type we should always find correcponding fields. */
1123 gcc_assert (fld && TREE_CODE (TREE_TYPE (fld)) == RECORD_TYPE);
1125 /* Nonbasetypes should have been stripped by outer_class_type. */
1126 gcc_assert (DECL_ARTIFICIAL (fld));
1128 outer_type = TREE_TYPE (fld);
1129 offset -= pos;
1131 base_binfo = get_binfo_at_offset (TYPE_BINFO (outer_type),
1132 offset, otr_type);
1133 gcc_assert (base_binfo);
1134 if (!pointer_set_insert (matched_vtables, BINFO_VTABLE (base_binfo)))
1136 tree target = gimple_get_virt_method_for_binfo (otr_token, base_binfo);
1137 if (target)
1138 maybe_record_node (nodes, target, inserted, completep);
1139 /* The only way method in anonymous namespace can become unreferable
1140 is that it has been fully optimized out. */
1141 else if (flag_ltrans || !type_in_anonymous_namespace_p (outer_type))
1142 *completep = false;
1143 pointer_set_insert (matched_vtables, BINFO_VTABLE (base_binfo));
1148 /* When virtual table is removed, we may need to flush the cache. */
1150 static void
1151 devirt_variable_node_removal_hook (varpool_node *n,
1152 void *d ATTRIBUTE_UNUSED)
1154 if (cached_polymorphic_call_targets
1155 && DECL_VIRTUAL_P (n->decl)
1156 && type_in_anonymous_namespace_p (DECL_CONTEXT (n->decl)))
1157 free_polymorphic_call_targets_hash ();
1160 /* Return vector containing possible targets of polymorphic call of type
1161 OTR_TYPE caling method OTR_TOKEN within type of OTR_OUTER_TYPE and OFFSET.
1162 If INCLUDE_BASES is true, walk also base types of OUTER_TYPES containig
1163 OTR_TYPE and include their virtual method. This is useful for types
1164 possibly in construction or destruction where the virtual table may
1165 temporarily change to one of base types. INCLUDE_DERIVER_TYPES make
1166 us to walk the inheritance graph for all derivations.
1168 If COMPLETEP is non-NULL, store true if the list is complette.
1169 CACHE_TOKEN (if non-NULL) will get stored to an unique ID of entry
1170 in the target cache. If user needs to visit every target list
1171 just once, it can memoize them.
1173 Returned vector is placed into cache. It is NOT caller's responsibility
1174 to free it. The vector can be freed on cgraph_remove_node call if
1175 the particular node is a virtual function present in the cache. */
1177 vec <cgraph_node *>
1178 possible_polymorphic_call_targets (tree otr_type,
1179 HOST_WIDE_INT otr_token,
1180 ipa_polymorphic_call_context context,
1181 bool *completep,
1182 void **cache_token)
1184 static struct cgraph_node_hook_list *node_removal_hook_holder;
1185 pointer_set_t *inserted;
1186 pointer_set_t *matched_vtables;
1187 vec <cgraph_node *> nodes=vNULL;
1188 odr_type type, outer_type;
1189 polymorphic_call_target_d key;
1190 polymorphic_call_target_d **slot;
1191 unsigned int i;
1192 tree binfo, target;
1193 bool final;
1195 type = get_odr_type (otr_type, true);
1197 /* Lookup the outer class type we want to walk. */
1198 if (context.outer_type)
1199 get_class_context (&context, otr_type);
1201 /* We now canonicalize our query, so we do not need extra hashtable entries. */
1203 /* Without outer type, we have no use for offset. Just do the
1204 basic search from innter type */
1205 if (!context.outer_type)
1207 context.outer_type = otr_type;
1208 context.offset = 0;
1210 /* We need to update our hiearchy if the type does not exist. */
1211 outer_type = get_odr_type (context.outer_type, true);
1212 /* If outer and inner type match, there are no bases to see. */
1213 if (type == outer_type)
1214 context.maybe_in_construction = false;
1215 /* If the type is final, there are no derivations. */
1216 if (TYPE_FINAL_P (outer_type->type))
1217 context.maybe_derived_type = false;
1219 /* Initialize query cache. */
1220 if (!cached_polymorphic_call_targets)
1222 cached_polymorphic_call_targets = pointer_set_create ();
1223 polymorphic_call_target_hash.create (23);
1224 if (!node_removal_hook_holder)
1226 node_removal_hook_holder =
1227 cgraph_add_node_removal_hook (&devirt_node_removal_hook, NULL);
1228 varpool_add_node_removal_hook (&devirt_variable_node_removal_hook,
1229 NULL);
1233 /* Lookup cached answer. */
1234 key.type = type;
1235 key.otr_token = otr_token;
1236 key.context = context;
1237 slot = polymorphic_call_target_hash.find_slot (&key, INSERT);
1238 if (cache_token)
1239 *cache_token = (void *)*slot;
1240 if (*slot)
1242 if (completep)
1243 *completep = (*slot)->final;
1244 return (*slot)->targets;
1247 final = true;
1249 /* Do actual search. */
1250 timevar_push (TV_IPA_VIRTUAL_CALL);
1251 *slot = XCNEW (polymorphic_call_target_d);
1252 if (cache_token)
1253 *cache_token = (void *)*slot;
1254 (*slot)->type = type;
1255 (*slot)->otr_token = otr_token;
1256 (*slot)->context = context;
1258 inserted = pointer_set_create ();
1259 matched_vtables = pointer_set_create ();
1261 /* First see virtual method of type itself. */
1263 binfo = get_binfo_at_offset (TYPE_BINFO (outer_type->type),
1264 context.offset, otr_type);
1265 target = gimple_get_virt_method_for_binfo (otr_token, binfo);
1266 if (target)
1268 maybe_record_node (nodes, target, inserted, &final);
1270 /* In the case we get final method, we don't need
1271 to walk derivations. */
1272 if (DECL_FINAL_P (target))
1273 context.maybe_derived_type = false;
1275 /* The only way method in anonymous namespace can become unreferable
1276 is that it has been fully optimized out. */
1277 else if (flag_ltrans || !type->anonymous_namespace)
1278 final = false;
1279 pointer_set_insert (matched_vtables, BINFO_VTABLE (binfo));
1281 /* Next walk bases, if asked to. */
1282 if (context.maybe_in_construction)
1283 record_targets_from_bases (otr_type, otr_token, outer_type->type,
1284 context.offset, nodes, inserted,
1285 matched_vtables, &final);
1287 /* Finally walk recursively all derived types. */
1288 if (context.maybe_derived_type)
1290 /* For anonymous namespace types we can attempt to build full type.
1291 All derivations must be in this unit (unless we see partial unit). */
1292 if (!type->anonymous_namespace || flag_ltrans)
1293 final = false;
1294 for (i = 0; i < outer_type->derived_types.length(); i++)
1295 possible_polymorphic_call_targets_1 (nodes, inserted,
1296 matched_vtables,
1297 otr_type, outer_type->derived_types[i],
1298 otr_token, outer_type->type,
1299 context.offset);
1301 (*slot)->targets = nodes;
1302 (*slot)->final = final;
1303 if (completep)
1304 *completep = final;
1306 pointer_set_destroy (inserted);
1307 pointer_set_destroy (matched_vtables);
1308 timevar_pop (TV_IPA_VIRTUAL_CALL);
1309 return nodes;
1312 /* Dump all possible targets of a polymorphic call. */
1314 void
1315 dump_possible_polymorphic_call_targets (FILE *f,
1316 tree otr_type,
1317 HOST_WIDE_INT otr_token,
1318 const ipa_polymorphic_call_context &ctx)
1320 vec <cgraph_node *> targets;
1321 bool final;
1322 odr_type type = get_odr_type (otr_type, false);
1323 unsigned int i;
1325 if (!type)
1326 return;
1327 targets = possible_polymorphic_call_targets (otr_type, otr_token,
1328 ctx,
1329 &final);
1330 fprintf (f, " Targets of polymorphic call of type %i:", type->id);
1331 print_generic_expr (f, type->type, TDF_SLIM);
1332 fprintf (f, " token %i\n"
1333 " Contained in type:",
1334 (int)otr_token);
1335 print_generic_expr (f, ctx.outer_type, TDF_SLIM);
1336 fprintf (f, " at offset "HOST_WIDE_INT_PRINT_DEC"\n"
1337 " %s%s%s\n ",
1338 ctx.offset,
1339 final ? "This is full list." :
1340 "This is partial list; extra targets may be defined in other units.",
1341 ctx.maybe_in_construction ? " (base types included)" : "",
1342 ctx.maybe_derived_type ? " (derived types included)" : "");
1343 for (i = 0; i < targets.length (); i++)
1344 fprintf (f, " %s/%i", targets[i]->name (),
1345 targets[i]->order);
1346 fprintf (f, "\n\n");
1350 /* Return true if N can be possibly target of a polymorphic call of
1351 OTR_TYPE/OTR_TOKEN. */
1353 bool
1354 possible_polymorphic_call_target_p (tree otr_type,
1355 HOST_WIDE_INT otr_token,
1356 const ipa_polymorphic_call_context &ctx,
1357 struct cgraph_node *n)
1359 vec <cgraph_node *> targets;
1360 unsigned int i;
1361 enum built_in_function fcode;
1362 bool final;
1364 if (TREE_CODE (TREE_TYPE (n->decl)) == FUNCTION_TYPE
1365 && ((fcode = DECL_FUNCTION_CODE (n->decl))
1366 == BUILT_IN_UNREACHABLE
1367 || fcode == BUILT_IN_TRAP))
1368 return true;
1370 if (!odr_hash.is_created ())
1371 return true;
1372 targets = possible_polymorphic_call_targets (otr_type, otr_token, ctx, &final);
1373 for (i = 0; i < targets.length (); i++)
1374 if (symtab_semantically_equivalent_p (n, targets[i]))
1375 return true;
1377 /* At a moment we allow middle end to dig out new external declarations
1378 as a targets of polymorphic calls. */
1379 if (!final && !n->definition)
1380 return true;
1381 return false;
1385 /* After callgraph construction new external nodes may appear.
1386 Add them into the graph. */
1388 void
1389 update_type_inheritance_graph (void)
1391 struct cgraph_node *n;
1393 if (!odr_hash.is_created ())
1394 return;
1395 free_polymorphic_call_targets_hash ();
1396 timevar_push (TV_IPA_INHERITANCE);
1397 /* We reconstruct the graph starting from types of all methods seen in the
1398 the unit. */
1399 FOR_EACH_FUNCTION (n)
1400 if (DECL_VIRTUAL_P (n->decl)
1401 && !n->definition
1402 && symtab_real_symbol_p (n))
1403 get_odr_type (method_class_type (TREE_TYPE (n->decl)), true);
1404 timevar_pop (TV_IPA_INHERITANCE);
1408 /* Return true if N looks like likely target of a polymorphic call.
1409 Rule out cxa_pure_virtual, noreturns, function declared cold and
1410 other obvious cases. */
1412 bool
1413 likely_target_p (struct cgraph_node *n)
1415 int flags;
1416 /* cxa_pure_virtual and similar things are not likely. */
1417 if (TREE_CODE (TREE_TYPE (n->decl)) != METHOD_TYPE)
1418 return false;
1419 flags = flags_from_decl_or_type (n->decl);
1420 if (flags & ECF_NORETURN)
1421 return false;
1422 if (lookup_attribute ("cold",
1423 DECL_ATTRIBUTES (n->decl)))
1424 return false;
1425 if (n->frequency < NODE_FREQUENCY_NORMAL)
1426 return false;
1427 return true;
1430 /* The ipa-devirt pass.
1431 When polymorphic call has only one likely target in the unit,
1432 turn it into speculative call. */
1434 static unsigned int
1435 ipa_devirt (void)
1437 struct cgraph_node *n;
1438 struct pointer_set_t *bad_call_targets = pointer_set_create ();
1439 struct cgraph_edge *e;
1441 int npolymorphic = 0, nspeculated = 0, nconverted = 0, ncold = 0;
1442 int nmultiple = 0, noverwritable = 0, ndevirtualized = 0, nnotdefined = 0;
1443 int nwrong = 0, nok = 0, nexternal = 0;;
1445 FOR_EACH_DEFINED_FUNCTION (n)
1447 bool update = false;
1448 if (dump_file && n->indirect_calls)
1449 fprintf (dump_file, "\n\nProcesing function %s/%i\n",
1450 n->name (), n->order);
1451 for (e = n->indirect_calls; e; e = e->next_callee)
1452 if (e->indirect_info->polymorphic)
1454 struct cgraph_node *likely_target = NULL;
1455 void *cache_token;
1456 bool final;
1457 vec <cgraph_node *>targets
1458 = possible_polymorphic_call_targets
1459 (e, &final, &cache_token);
1460 unsigned int i;
1462 if (dump_file)
1463 dump_possible_polymorphic_call_targets
1464 (dump_file, e);
1466 npolymorphic++;
1468 if (!cgraph_maybe_hot_edge_p (e))
1470 if (dump_file)
1471 fprintf (dump_file, "Call is cold\n");
1472 ncold++;
1473 continue;
1475 if (e->speculative)
1477 if (dump_file)
1478 fprintf (dump_file, "Call is aready speculated\n");
1479 nspeculated++;
1481 /* When dumping see if we agree with speculation. */
1482 if (!dump_file)
1483 continue;
1485 if (pointer_set_contains (bad_call_targets,
1486 cache_token))
1488 if (dump_file)
1489 fprintf (dump_file, "Target list is known to be useless\n");
1490 nmultiple++;
1491 continue;
1493 for (i = 0; i < targets.length (); i++)
1494 if (likely_target_p (targets[i]))
1496 if (likely_target)
1498 likely_target = NULL;
1499 if (dump_file)
1500 fprintf (dump_file, "More than one likely target\n");
1501 nmultiple++;
1502 break;
1504 likely_target = targets[i];
1506 if (!likely_target)
1508 pointer_set_insert (bad_call_targets, cache_token);
1509 continue;
1511 /* This is reached only when dumping; check if we agree or disagree
1512 with the speculation. */
1513 if (e->speculative)
1515 struct cgraph_edge *e2;
1516 struct ipa_ref *ref;
1517 cgraph_speculative_call_info (e, e2, e, ref);
1518 if (cgraph_function_or_thunk_node (e2->callee, NULL)
1519 == cgraph_function_or_thunk_node (likely_target, NULL))
1521 fprintf (dump_file, "We agree with speculation\n");
1522 nok++;
1524 else
1526 fprintf (dump_file, "We disagree with speculation\n");
1527 nwrong++;
1529 continue;
1531 if (!likely_target->definition)
1533 if (dump_file)
1534 fprintf (dump_file, "Target is not an definition\n");
1535 nnotdefined++;
1536 continue;
1538 /* Do not introduce new references to external symbols. While we
1539 can handle these just well, it is common for programs to
1540 incorrectly with headers defining methods they are linked
1541 with. */
1542 if (DECL_EXTERNAL (likely_target->decl))
1544 if (dump_file)
1545 fprintf (dump_file, "Target is external\n");
1546 nexternal++;
1547 continue;
1549 if (cgraph_function_body_availability (likely_target)
1550 <= AVAIL_OVERWRITABLE
1551 && symtab_can_be_discarded (likely_target))
1553 if (dump_file)
1554 fprintf (dump_file, "Target is overwritable\n");
1555 noverwritable++;
1556 continue;
1558 else
1560 if (dump_file)
1561 fprintf (dump_file,
1562 "Speculatively devirtualizing call in %s/%i to %s/%i\n",
1563 n->name (), n->order,
1564 likely_target->name (),
1565 likely_target->order);
1566 if (!symtab_can_be_discarded (likely_target))
1568 cgraph_node *alias;
1569 alias = cgraph (symtab_nonoverwritable_alias
1570 (likely_target));
1571 if (alias)
1572 likely_target = alias;
1574 nconverted++;
1575 update = true;
1576 cgraph_turn_edge_to_speculative
1577 (e, likely_target, e->count * 8 / 10, e->frequency * 8 / 10);
1580 if (update)
1581 inline_update_overall_summary (n);
1583 pointer_set_destroy (bad_call_targets);
1585 if (dump_file)
1586 fprintf (dump_file,
1587 "%i polymorphic calls, %i devirtualized,"
1588 " %i speculatively devirtualized, %i cold\n"
1589 "%i have multiple targets, %i overwritable,"
1590 " %i already speculated (%i agree, %i disagree),"
1591 " %i external, %i not defined\n",
1592 npolymorphic, ndevirtualized, nconverted, ncold,
1593 nmultiple, noverwritable, nspeculated, nok, nwrong,
1594 nexternal, nnotdefined);
1595 return ndevirtualized ? TODO_remove_functions : 0;
1598 /* Gate for speculative IPA devirtualization optimization. */
1600 static bool
1601 gate_ipa_devirt (void)
1603 return (flag_devirtualize
1604 && flag_devirtualize_speculatively
1605 && optimize);
1608 namespace {
1610 const pass_data pass_data_ipa_devirt =
1612 IPA_PASS, /* type */
1613 "devirt", /* name */
1614 OPTGROUP_NONE, /* optinfo_flags */
1615 true, /* has_gate */
1616 true, /* has_execute */
1617 TV_IPA_DEVIRT, /* tv_id */
1618 0, /* properties_required */
1619 0, /* properties_provided */
1620 0, /* properties_destroyed */
1621 0, /* todo_flags_start */
1622 ( TODO_dump_symtab ), /* todo_flags_finish */
1625 class pass_ipa_devirt : public ipa_opt_pass_d
1627 public:
1628 pass_ipa_devirt (gcc::context *ctxt)
1629 : ipa_opt_pass_d (pass_data_ipa_devirt, ctxt,
1630 NULL, /* generate_summary */
1631 NULL, /* write_summary */
1632 NULL, /* read_summary */
1633 NULL, /* write_optimization_summary */
1634 NULL, /* read_optimization_summary */
1635 NULL, /* stmt_fixup */
1636 0, /* function_transform_todo_flags_start */
1637 NULL, /* function_transform */
1638 NULL) /* variable_transform */
1641 /* opt_pass methods: */
1642 bool gate () { return gate_ipa_devirt (); }
1643 unsigned int execute () { return ipa_devirt (); }
1645 }; // class pass_ipa_devirt
1647 } // anon namespace
1649 ipa_opt_pass_d *
1650 make_pass_ipa_devirt (gcc::context *ctxt)
1652 return new pass_ipa_devirt (ctxt);
1655 #include "gt-ipa-devirt.h"