[i386] Fold __builtin_ia32_shufpd to VEC_PERM_EXPR
[official-gcc.git] / gcc / ipa-polymorphic-call.c
blob75d8ebc1c446f87cead4b276aca13ad620b4f525
1 /* Analysis of polymorphic call context.
2 Copyright (C) 2013-2019 Free Software Foundation, Inc.
3 Contributed by Jan Hubicka
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify it under
8 the terms of the GNU General Public License as published by the Free
9 Software Foundation; either version 3, or (at your option) any later
10 version.
12 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
15 for more details.
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
21 #include "config.h"
22 #include "system.h"
23 #include "coretypes.h"
24 #include "backend.h"
25 #include "rtl.h"
26 #include "tree.h"
27 #include "gimple.h"
28 #include "tree-pass.h"
29 #include "tree-ssa-operands.h"
30 #include "streamer-hooks.h"
31 #include "cgraph.h"
32 #include "data-streamer.h"
33 #include "diagnostic.h"
34 #include "alias.h"
35 #include "fold-const.h"
36 #include "calls.h"
37 #include "ipa-utils.h"
38 #include "tree-dfa.h"
39 #include "gimple-pretty-print.h"
40 #include "tree-into-ssa.h"
41 #include "params.h"
43 /* Return true when TYPE contains an polymorphic type and thus is interesting
44 for devirtualization machinery. */
46 static bool contains_type_p (tree, HOST_WIDE_INT, tree,
47 bool consider_placement_new = true,
48 bool consider_bases = true);
50 bool
51 contains_polymorphic_type_p (const_tree type)
53 type = TYPE_MAIN_VARIANT (type);
55 if (RECORD_OR_UNION_TYPE_P (type))
57 if (TYPE_BINFO (type)
58 && polymorphic_type_binfo_p (TYPE_BINFO (type)))
59 return true;
60 for (tree fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
61 if (TREE_CODE (fld) == FIELD_DECL
62 && !DECL_ARTIFICIAL (fld)
63 && contains_polymorphic_type_p (TREE_TYPE (fld)))
64 return true;
65 return false;
67 if (TREE_CODE (type) == ARRAY_TYPE)
68 return contains_polymorphic_type_p (TREE_TYPE (type));
69 return false;
72 /* Return true if it seems valid to use placement new to build EXPECTED_TYPE
73 at possition CUR_OFFSET within TYPE.
75 POD can be changed to an instance of a polymorphic type by
76 placement new. Here we play safe and assume that any
77 non-polymorphic type is POD. */
78 bool
79 possible_placement_new (tree type, tree expected_type,
80 HOST_WIDE_INT cur_offset)
82 if (cur_offset < 0)
83 return true;
84 return ((TREE_CODE (type) != RECORD_TYPE
85 || !TYPE_BINFO (type)
86 || cur_offset >= POINTER_SIZE
87 || !polymorphic_type_binfo_p (TYPE_BINFO (type)))
88 && (!TYPE_SIZE (type)
89 || !tree_fits_shwi_p (TYPE_SIZE (type))
90 || (cur_offset
91 + (expected_type ? tree_to_uhwi (TYPE_SIZE (expected_type))
92 : POINTER_SIZE)
93 <= tree_to_uhwi (TYPE_SIZE (type)))));
96 /* THIS->OUTER_TYPE is a type of memory object where object of OTR_TYPE
97 is contained at THIS->OFFSET. Walk the memory representation of
98 THIS->OUTER_TYPE and find the outermost class type that match
99 OTR_TYPE or contain OTR_TYPE as a base. Update THIS
100 to represent it.
102 If OTR_TYPE is NULL, just find outermost polymorphic type with
103 virtual table present at possition OFFSET.
105 For example when THIS represents type
106 class A
108 int a;
109 class B b;
111 and we look for type at offset sizeof(int), we end up with B and offset 0.
112 If the same is produced by multiple inheritance, we end up with A and offset
113 sizeof(int).
115 If we cannot find corresponding class, give up by setting
116 THIS->OUTER_TYPE to OTR_TYPE and THIS->OFFSET to NULL.
117 Return true when lookup was sucesful.
119 When CONSIDER_PLACEMENT_NEW is false, reject contexts that may be made
120 valid only via allocation of new polymorphic type inside by means
121 of placement new.
123 When CONSIDER_BASES is false, only look for actual fields, not base types
124 of TYPE. */
126 bool
127 ipa_polymorphic_call_context::restrict_to_inner_class (tree otr_type,
128 bool consider_placement_new,
129 bool consider_bases)
131 tree type = outer_type;
132 HOST_WIDE_INT cur_offset = offset;
133 bool speculative = false;
134 bool size_unknown = false;
135 unsigned HOST_WIDE_INT otr_type_size = POINTER_SIZE;
137 /* Update OUTER_TYPE to match EXPECTED_TYPE if it is not set. */
138 if (!outer_type)
140 clear_outer_type (otr_type);
141 type = otr_type;
142 cur_offset = 0;
144 /* See if OFFSET points inside OUTER_TYPE. If it does not, we know
145 that the context is either invalid, or the instance type must be
146 derived from OUTER_TYPE.
148 Because the instance type may contain field whose type is of OUTER_TYPE,
149 we cannot derive any effective information about it.
151 TODO: In the case we know all derrived types, we can definitely do better
152 here. */
153 else if (TYPE_SIZE (outer_type)
154 && tree_fits_shwi_p (TYPE_SIZE (outer_type))
155 && tree_to_shwi (TYPE_SIZE (outer_type)) >= 0
156 && tree_to_shwi (TYPE_SIZE (outer_type)) <= offset)
158 bool der = maybe_derived_type; /* clear_outer_type will reset it. */
159 bool dyn = dynamic;
160 clear_outer_type (otr_type);
161 type = otr_type;
162 cur_offset = 0;
164 /* If derived type is not allowed, we know that the context is invalid.
165 For dynamic types, we really do not have information about
166 size of the memory location. It is possible that completely
167 different type is stored after outer_type. */
168 if (!der && !dyn)
170 clear_speculation ();
171 invalid = true;
172 return false;
176 if (otr_type && TYPE_SIZE (otr_type)
177 && tree_fits_shwi_p (TYPE_SIZE (otr_type)))
178 otr_type_size = tree_to_uhwi (TYPE_SIZE (otr_type));
180 if (!type || offset < 0)
181 goto no_useful_type_info;
183 /* Find the sub-object the constant actually refers to and mark whether it is
184 an artificial one (as opposed to a user-defined one).
186 This loop is performed twice; first time for outer_type and second time
187 for speculative_outer_type. The second run has SPECULATIVE set. */
188 while (true)
190 unsigned HOST_WIDE_INT pos, size;
191 tree fld;
193 /* If we do not know size of TYPE, we need to be more conservative
194 about accepting cases where we cannot find EXPECTED_TYPE.
195 Generally the types that do matter here are of constant size.
196 Size_unknown case should be very rare. */
197 if (TYPE_SIZE (type)
198 && tree_fits_shwi_p (TYPE_SIZE (type))
199 && tree_to_shwi (TYPE_SIZE (type)) >= 0)
200 size_unknown = false;
201 else
202 size_unknown = true;
204 /* On a match, just return what we found. */
205 if ((otr_type
206 && types_odr_comparable (type, otr_type)
207 && types_same_for_odr (type, otr_type))
208 || (!otr_type
209 && TREE_CODE (type) == RECORD_TYPE
210 && TYPE_BINFO (type)
211 && polymorphic_type_binfo_p (TYPE_BINFO (type))))
213 if (speculative)
215 /* If we did not match the offset, just give up on speculation. */
216 if (cur_offset != 0
217 /* Also check if speculation did not end up being same as
218 non-speculation. */
219 || (types_must_be_same_for_odr (speculative_outer_type,
220 outer_type)
221 && (maybe_derived_type
222 == speculative_maybe_derived_type)))
223 clear_speculation ();
224 return true;
226 else
228 /* If type is known to be final, do not worry about derived
229 types. Testing it here may help us to avoid speculation. */
230 if (otr_type && TREE_CODE (outer_type) == RECORD_TYPE
231 && (!in_lto_p || odr_type_p (outer_type))
232 && type_with_linkage_p (outer_type)
233 && type_known_to_have_no_derivations_p (outer_type))
234 maybe_derived_type = false;
236 /* Type cannot contain itself on an non-zero offset. In that case
237 just give up. Still accept the case where size is now known.
238 Either the second copy may appear past the end of type or within
239 the non-POD buffer located inside the variably sized type
240 itself. */
241 if (cur_offset != 0)
242 goto no_useful_type_info;
243 /* If we determined type precisely or we have no clue on
244 speuclation, we are done. */
245 if (!maybe_derived_type || !speculative_outer_type
246 || !speculation_consistent_p (speculative_outer_type,
247 speculative_offset,
248 speculative_maybe_derived_type,
249 otr_type))
251 clear_speculation ();
252 return true;
254 /* Otherwise look into speculation now. */
255 else
257 speculative = true;
258 type = speculative_outer_type;
259 cur_offset = speculative_offset;
260 continue;
265 /* Walk fields and find corresponding on at OFFSET. */
266 if (TREE_CODE (type) == RECORD_TYPE)
268 for (fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
270 if (TREE_CODE (fld) != FIELD_DECL
271 || TREE_TYPE (fld) == error_mark_node)
272 continue;
274 pos = int_bit_position (fld);
275 if (pos > (unsigned HOST_WIDE_INT)cur_offset)
276 continue;
278 /* Do not consider vptr itself. Not even for placement new. */
279 if (!pos && DECL_ARTIFICIAL (fld)
280 && POINTER_TYPE_P (TREE_TYPE (fld))
281 && TYPE_BINFO (type)
282 && polymorphic_type_binfo_p (TYPE_BINFO (type)))
283 continue;
285 if (!DECL_SIZE (fld) || !tree_fits_uhwi_p (DECL_SIZE (fld)))
286 goto no_useful_type_info;
287 size = tree_to_uhwi (DECL_SIZE (fld));
289 /* We can always skip types smaller than pointer size:
290 those cannot contain a virtual table pointer.
292 Disqualifying fields that are too small to fit OTR_TYPE
293 saves work needed to walk them for no benefit.
294 Because of the way the bases are packed into a class, the
295 field's size may be smaller than type size, so it needs
296 to be done with a care. */
298 if (pos <= (unsigned HOST_WIDE_INT)cur_offset
299 && (pos + size) >= (unsigned HOST_WIDE_INT)cur_offset
300 + POINTER_SIZE
301 && (!otr_type
302 || !TYPE_SIZE (TREE_TYPE (fld))
303 || !tree_fits_shwi_p (TYPE_SIZE (TREE_TYPE (fld)))
304 || (pos + tree_to_uhwi (TYPE_SIZE (TREE_TYPE (fld))))
305 >= cur_offset + otr_type_size))
306 break;
309 if (!fld)
310 goto no_useful_type_info;
312 type = TYPE_MAIN_VARIANT (TREE_TYPE (fld));
313 cur_offset -= pos;
314 /* DECL_ARTIFICIAL represents a basetype. */
315 if (!DECL_ARTIFICIAL (fld))
317 if (!speculative)
319 outer_type = type;
320 offset = cur_offset;
321 /* As soon as we se an field containing the type,
322 we know we are not looking for derivations. */
323 maybe_derived_type = false;
325 else
327 speculative_outer_type = type;
328 speculative_offset = cur_offset;
329 speculative_maybe_derived_type = false;
332 else if (!consider_bases)
333 goto no_useful_type_info;
335 else if (TREE_CODE (type) == ARRAY_TYPE)
337 tree subtype = TYPE_MAIN_VARIANT (TREE_TYPE (type));
339 /* Give up if we don't know array field size.
340 Also give up on non-polymorphic types as they are used
341 as buffers for placement new. */
342 if (!TYPE_SIZE (subtype)
343 || !tree_fits_shwi_p (TYPE_SIZE (subtype))
344 || tree_to_shwi (TYPE_SIZE (subtype)) <= 0
345 || !contains_polymorphic_type_p (subtype))
346 goto no_useful_type_info;
348 HOST_WIDE_INT new_offset = cur_offset % tree_to_shwi (TYPE_SIZE (subtype));
350 /* We may see buffer for placement new. In this case the expected type
351 can be bigger than the subtype. */
352 if (TYPE_SIZE (subtype)
353 && (cur_offset + otr_type_size
354 > tree_to_uhwi (TYPE_SIZE (subtype))))
355 goto no_useful_type_info;
357 cur_offset = new_offset;
358 type = TYPE_MAIN_VARIANT (subtype);
359 if (!speculative)
361 outer_type = type;
362 offset = cur_offset;
363 maybe_derived_type = false;
365 else
367 speculative_outer_type = type;
368 speculative_offset = cur_offset;
369 speculative_maybe_derived_type = false;
372 /* Give up on anything else. */
373 else
375 no_useful_type_info:
376 if (maybe_derived_type && !speculative
377 && TREE_CODE (outer_type) == RECORD_TYPE
378 && TREE_CODE (otr_type) == RECORD_TYPE
379 && TYPE_BINFO (otr_type)
380 && !offset
381 && get_binfo_at_offset (TYPE_BINFO (otr_type), 0, outer_type))
383 clear_outer_type (otr_type);
384 if (!speculative_outer_type
385 || !speculation_consistent_p (speculative_outer_type,
386 speculative_offset,
387 speculative_maybe_derived_type,
388 otr_type))
389 clear_speculation ();
390 if (speculative_outer_type)
392 speculative = true;
393 type = speculative_outer_type;
394 cur_offset = speculative_offset;
396 else
397 return true;
399 /* We found no way to embedd EXPECTED_TYPE in TYPE.
400 We still permit two special cases - placement new and
401 the case of variadic types containing themselves. */
402 if (!speculative
403 && consider_placement_new
404 && (size_unknown || !type || maybe_derived_type
405 || possible_placement_new (type, otr_type, cur_offset)))
407 /* In these weird cases we want to accept the context.
408 In non-speculative run we have no useful outer_type info
409 (TODO: we may eventually want to record upper bound on the
410 type size that can be used to prune the walk),
411 but we still want to consider speculation that may
412 give useful info. */
413 if (!speculative)
415 clear_outer_type (otr_type);
416 if (!speculative_outer_type
417 || !speculation_consistent_p (speculative_outer_type,
418 speculative_offset,
419 speculative_maybe_derived_type,
420 otr_type))
421 clear_speculation ();
422 if (speculative_outer_type)
424 speculative = true;
425 type = speculative_outer_type;
426 cur_offset = speculative_offset;
428 else
429 return true;
431 else
433 clear_speculation ();
434 return true;
437 else
439 clear_speculation ();
440 if (speculative)
441 return true;
442 clear_outer_type (otr_type);
443 invalid = true;
444 return false;
450 /* Return true if OUTER_TYPE contains OTR_TYPE at OFFSET.
451 CONSIDER_PLACEMENT_NEW makes function to accept cases where OTR_TYPE can
452 be built within OUTER_TYPE by means of placement new. CONSIDER_BASES makes
453 function to accept cases where OTR_TYPE appears as base of OUTER_TYPE or as
454 base of one of fields of OUTER_TYPE. */
456 static bool
457 contains_type_p (tree outer_type, HOST_WIDE_INT offset,
458 tree otr_type,
459 bool consider_placement_new,
460 bool consider_bases)
462 ipa_polymorphic_call_context context;
464 /* Check that type is within range. */
465 if (offset < 0)
466 return false;
468 /* PR ipa/71207
469 As OUTER_TYPE can be a type which has a diamond virtual inheritance,
470 it's not necessary that INNER_TYPE will fit within OUTER_TYPE with
471 a given offset. It can happen that INNER_TYPE also contains a base object,
472 however it would point to the same instance in the OUTER_TYPE. */
474 context.offset = offset;
475 context.outer_type = TYPE_MAIN_VARIANT (outer_type);
476 context.maybe_derived_type = false;
477 context.dynamic = false;
478 return context.restrict_to_inner_class (otr_type, consider_placement_new,
479 consider_bases);
483 /* Return a FUNCTION_DECL if FN represent a constructor or destructor.
484 If CHECK_CLONES is true, also check for clones of ctor/dtors. */
486 tree
487 polymorphic_ctor_dtor_p (tree fn, bool check_clones)
489 if (TREE_CODE (TREE_TYPE (fn)) != METHOD_TYPE
490 || (!DECL_CXX_CONSTRUCTOR_P (fn) && !DECL_CXX_DESTRUCTOR_P (fn)))
492 if (!check_clones)
493 return NULL_TREE;
495 /* Watch for clones where we constant propagated the first
496 argument (pointer to the instance). */
497 fn = DECL_ABSTRACT_ORIGIN (fn);
498 if (!fn
499 || TREE_CODE (TREE_TYPE (fn)) != METHOD_TYPE
500 || (!DECL_CXX_CONSTRUCTOR_P (fn) && !DECL_CXX_DESTRUCTOR_P (fn)))
501 return NULL_TREE;
504 if (flags_from_decl_or_type (fn) & (ECF_PURE | ECF_CONST))
505 return NULL_TREE;
507 return fn;
510 /* Return a FUNCTION_DECL if BLOCK represents a constructor or destructor.
511 If CHECK_CLONES is true, also check for clones of ctor/dtors. */
513 tree
514 inlined_polymorphic_ctor_dtor_block_p (tree block, bool check_clones)
516 tree fn = block_ultimate_origin (block);
517 if (fn == NULL || TREE_CODE (fn) != FUNCTION_DECL)
518 return NULL_TREE;
520 return polymorphic_ctor_dtor_p (fn, check_clones);
524 /* We know that the instance is stored in variable or parameter
525 (not dynamically allocated) and we want to disprove the fact
526 that it may be in construction at invocation of CALL.
528 BASE represents memory location where instance is stored.
529 If BASE is NULL, it is assumed to be global memory.
530 OUTER_TYPE is known type of the instance or NULL if not
531 known.
533 For the variable to be in construction we actually need to
534 be in constructor of corresponding global variable or
535 the inline stack of CALL must contain the constructor.
536 Check this condition. This check works safely only before
537 IPA passes, because inline stacks may become out of date
538 later. */
540 bool
541 decl_maybe_in_construction_p (tree base, tree outer_type,
542 gimple *call, tree function)
544 if (outer_type)
545 outer_type = TYPE_MAIN_VARIANT (outer_type);
546 gcc_assert (!base || DECL_P (base));
548 /* After inlining the code unification optimizations may invalidate
549 inline stacks. Also we need to give up on global variables after
550 IPA, because addresses of these may have been propagated to their
551 constructors. */
552 if (DECL_STRUCT_FUNCTION (function)->after_inlining)
553 return true;
555 /* Pure functions cannot do any changes on the dynamic type;
556 that require writting to memory. */
557 if ((!base || !auto_var_in_fn_p (base, function))
558 && flags_from_decl_or_type (function) & (ECF_PURE | ECF_CONST))
559 return false;
561 bool check_clones = !base || is_global_var (base);
562 for (tree block = gimple_block (call); block && TREE_CODE (block) == BLOCK;
563 block = BLOCK_SUPERCONTEXT (block))
564 if (tree fn = inlined_polymorphic_ctor_dtor_block_p (block, check_clones))
566 tree type = TYPE_METHOD_BASETYPE (TREE_TYPE (fn));
568 if (!outer_type || !types_odr_comparable (type, outer_type))
570 if (TREE_CODE (type) == RECORD_TYPE
571 && TYPE_BINFO (type)
572 && polymorphic_type_binfo_p (TYPE_BINFO (type)))
573 return true;
575 else if (types_same_for_odr (type, outer_type))
576 return true;
579 if (!base || (VAR_P (base) && is_global_var (base)))
581 if (TREE_CODE (TREE_TYPE (function)) != METHOD_TYPE
582 || (!DECL_CXX_CONSTRUCTOR_P (function)
583 && !DECL_CXX_DESTRUCTOR_P (function)))
585 if (!DECL_ABSTRACT_ORIGIN (function))
586 return false;
587 /* Watch for clones where we constant propagated the first
588 argument (pointer to the instance). */
589 function = DECL_ABSTRACT_ORIGIN (function);
590 if (!function
591 || TREE_CODE (TREE_TYPE (function)) != METHOD_TYPE
592 || (!DECL_CXX_CONSTRUCTOR_P (function)
593 && !DECL_CXX_DESTRUCTOR_P (function)))
594 return false;
596 tree type = TYPE_METHOD_BASETYPE (TREE_TYPE (function));
597 if (!outer_type || !types_odr_comparable (type, outer_type))
599 if (TREE_CODE (type) == RECORD_TYPE
600 && TYPE_BINFO (type)
601 && polymorphic_type_binfo_p (TYPE_BINFO (type)))
602 return true;
604 else if (types_same_for_odr (type, outer_type))
605 return true;
607 return false;
610 /* Dump human readable context to F. If NEWLINE is true, it will be terminated
611 by a newline. */
613 void
614 ipa_polymorphic_call_context::dump (FILE *f, bool newline) const
616 fprintf (f, " ");
617 if (invalid)
618 fprintf (f, "Call is known to be undefined");
619 else
621 if (useless_p ())
622 fprintf (f, "nothing known");
623 if (outer_type || offset)
625 fprintf (f, "Outer type%s:", dynamic ? " (dynamic)":"");
626 print_generic_expr (f, outer_type, TDF_SLIM);
627 if (maybe_derived_type)
628 fprintf (f, " (or a derived type)");
629 if (maybe_in_construction)
630 fprintf (f, " (maybe in construction)");
631 fprintf (f, " offset " HOST_WIDE_INT_PRINT_DEC,
632 offset);
634 if (speculative_outer_type)
636 if (outer_type || offset)
637 fprintf (f, " ");
638 fprintf (f, "Speculative outer type:");
639 print_generic_expr (f, speculative_outer_type, TDF_SLIM);
640 if (speculative_maybe_derived_type)
641 fprintf (f, " (or a derived type)");
642 fprintf (f, " at offset " HOST_WIDE_INT_PRINT_DEC,
643 speculative_offset);
646 if (newline)
647 fprintf(f, "\n");
650 /* Print context to stderr. */
652 void
653 ipa_polymorphic_call_context::debug () const
655 dump (stderr);
658 /* Stream out the context to OB. */
660 void
661 ipa_polymorphic_call_context::stream_out (struct output_block *ob) const
663 struct bitpack_d bp = bitpack_create (ob->main_stream);
665 bp_pack_value (&bp, invalid, 1);
666 bp_pack_value (&bp, maybe_in_construction, 1);
667 bp_pack_value (&bp, maybe_derived_type, 1);
668 bp_pack_value (&bp, speculative_maybe_derived_type, 1);
669 bp_pack_value (&bp, dynamic, 1);
670 bp_pack_value (&bp, outer_type != NULL, 1);
671 bp_pack_value (&bp, offset != 0, 1);
672 bp_pack_value (&bp, speculative_outer_type != NULL, 1);
673 streamer_write_bitpack (&bp);
675 if (outer_type != NULL)
676 stream_write_tree (ob, outer_type, true);
677 if (offset)
678 streamer_write_hwi (ob, offset);
679 if (speculative_outer_type != NULL)
681 stream_write_tree (ob, speculative_outer_type, true);
682 streamer_write_hwi (ob, speculative_offset);
684 else
685 gcc_assert (!speculative_offset);
688 /* Stream in the context from IB and DATA_IN. */
690 void
691 ipa_polymorphic_call_context::stream_in (struct lto_input_block *ib,
692 struct data_in *data_in)
694 struct bitpack_d bp = streamer_read_bitpack (ib);
696 invalid = bp_unpack_value (&bp, 1);
697 maybe_in_construction = bp_unpack_value (&bp, 1);
698 maybe_derived_type = bp_unpack_value (&bp, 1);
699 speculative_maybe_derived_type = bp_unpack_value (&bp, 1);
700 dynamic = bp_unpack_value (&bp, 1);
701 bool outer_type_p = bp_unpack_value (&bp, 1);
702 bool offset_p = bp_unpack_value (&bp, 1);
703 bool speculative_outer_type_p = bp_unpack_value (&bp, 1);
705 if (outer_type_p)
706 outer_type = stream_read_tree (ib, data_in);
707 else
708 outer_type = NULL;
709 if (offset_p)
710 offset = (HOST_WIDE_INT) streamer_read_hwi (ib);
711 else
712 offset = 0;
713 if (speculative_outer_type_p)
715 speculative_outer_type = stream_read_tree (ib, data_in);
716 speculative_offset = (HOST_WIDE_INT) streamer_read_hwi (ib);
718 else
720 speculative_outer_type = NULL;
721 speculative_offset = 0;
725 /* Proudce polymorphic call context for call method of instance
726 that is located within BASE (that is assumed to be a decl) at offset OFF. */
728 void
729 ipa_polymorphic_call_context::set_by_decl (tree base, HOST_WIDE_INT off)
731 gcc_assert (DECL_P (base));
732 clear_speculation ();
734 if (!contains_polymorphic_type_p (TREE_TYPE (base)))
736 clear_outer_type ();
737 offset = off;
738 return;
740 outer_type = TYPE_MAIN_VARIANT (TREE_TYPE (base));
741 offset = off;
742 /* Make very conservative assumption that all objects
743 may be in construction.
745 It is up to caller to revisit this via
746 get_dynamic_type or decl_maybe_in_construction_p. */
747 maybe_in_construction = true;
748 maybe_derived_type = false;
749 dynamic = false;
752 /* CST is an invariant (address of decl), try to get meaningful
753 polymorphic call context for polymorphic call of method
754 if instance of OTR_TYPE that is located at offset OFF of this invariant.
755 Return FALSE if nothing meaningful can be found. */
757 bool
758 ipa_polymorphic_call_context::set_by_invariant (tree cst,
759 tree otr_type,
760 HOST_WIDE_INT off)
762 poly_int64 offset2, size, max_size;
763 bool reverse;
764 tree base;
766 invalid = false;
767 off = 0;
768 clear_outer_type (otr_type);
770 if (TREE_CODE (cst) != ADDR_EXPR)
771 return false;
773 cst = TREE_OPERAND (cst, 0);
774 base = get_ref_base_and_extent (cst, &offset2, &size, &max_size, &reverse);
775 if (!DECL_P (base) || !known_size_p (max_size) || maybe_ne (max_size, size))
776 return false;
778 /* Only type inconsistent programs can have otr_type that is
779 not part of outer type. */
780 if (otr_type && !contains_type_p (TREE_TYPE (base), off, otr_type))
781 return false;
783 set_by_decl (base, off);
784 return true;
787 /* See if OP is SSA name initialized as a copy or by single assignment.
788 If so, walk the SSA graph up. Because simple PHI conditional is considered
789 copy, GLOBAL_VISITED may be used to avoid infinite loop walking the SSA
790 graph. */
792 static tree
793 walk_ssa_copies (tree op, hash_set<tree> **global_visited = NULL)
795 hash_set <tree> *visited = NULL;
796 STRIP_NOPS (op);
797 while (TREE_CODE (op) == SSA_NAME
798 && !SSA_NAME_IS_DEFAULT_DEF (op)
799 /* We might be called via fold_stmt during cfgcleanup where
800 SSA form need not be up-to-date. */
801 && !name_registered_for_update_p (op)
802 && (gimple_assign_single_p (SSA_NAME_DEF_STMT (op))
803 || gimple_code (SSA_NAME_DEF_STMT (op)) == GIMPLE_PHI))
805 if (global_visited)
807 if (!*global_visited)
808 *global_visited = new hash_set<tree>;
809 if ((*global_visited)->add (op))
810 goto done;
812 else
814 if (!visited)
815 visited = new hash_set<tree>;
816 if (visited->add (op))
817 goto done;
819 /* Special case
820 if (ptr == 0)
821 ptr = 0;
822 else
823 ptr = ptr.foo;
824 This pattern is implicitly produced for casts to non-primary
825 bases. When doing context analysis, we do not really care
826 about the case pointer is NULL, because the call will be
827 undefined anyway. */
828 if (gimple_code (SSA_NAME_DEF_STMT (op)) == GIMPLE_PHI)
830 gimple *phi = SSA_NAME_DEF_STMT (op);
832 if (gimple_phi_num_args (phi) > 2)
833 goto done;
834 if (gimple_phi_num_args (phi) == 1)
835 op = gimple_phi_arg_def (phi, 0);
836 else if (integer_zerop (gimple_phi_arg_def (phi, 0)))
837 op = gimple_phi_arg_def (phi, 1);
838 else if (integer_zerop (gimple_phi_arg_def (phi, 1)))
839 op = gimple_phi_arg_def (phi, 0);
840 else
841 goto done;
843 else
845 if (gimple_assign_load_p (SSA_NAME_DEF_STMT (op)))
846 goto done;
847 op = gimple_assign_rhs1 (SSA_NAME_DEF_STMT (op));
849 STRIP_NOPS (op);
851 done:
852 if (visited)
853 delete (visited);
854 return op;
857 /* Create polymorphic call context from IP invariant CST.
858 This is typically &global_var.
859 OTR_TYPE specify type of polymorphic call or NULL if unknown, OFF
860 is offset of call. */
862 ipa_polymorphic_call_context::ipa_polymorphic_call_context (tree cst,
863 tree otr_type,
864 HOST_WIDE_INT off)
866 clear_speculation ();
867 set_by_invariant (cst, otr_type, off);
870 /* Build context for pointer REF contained in FNDECL at statement STMT.
871 if INSTANCE is non-NULL, return pointer to the object described by
872 the context or DECL where context is contained in. */
874 ipa_polymorphic_call_context::ipa_polymorphic_call_context (tree fndecl,
875 tree ref,
876 gimple *stmt,
877 tree *instance)
879 tree otr_type = NULL;
880 tree base_pointer;
881 hash_set <tree> *visited = NULL;
883 if (TREE_CODE (ref) == OBJ_TYPE_REF)
885 otr_type = obj_type_ref_class (ref);
886 base_pointer = OBJ_TYPE_REF_OBJECT (ref);
888 else
889 base_pointer = ref;
891 /* Set up basic info in case we find nothing interesting in the analysis. */
892 clear_speculation ();
893 clear_outer_type (otr_type);
894 invalid = false;
896 /* Walk SSA for outer object. */
897 while (true)
899 base_pointer = walk_ssa_copies (base_pointer, &visited);
900 if (TREE_CODE (base_pointer) == ADDR_EXPR)
902 HOST_WIDE_INT offset2, size;
903 bool reverse;
904 tree base
905 = get_ref_base_and_extent_hwi (TREE_OPERAND (base_pointer, 0),
906 &offset2, &size, &reverse);
907 if (!base)
908 break;
910 combine_speculation_with (TYPE_MAIN_VARIANT (TREE_TYPE (base)),
911 offset + offset2,
912 true,
913 NULL /* Do not change outer type. */);
915 /* If this is a varying address, punt. */
916 if (TREE_CODE (base) == MEM_REF || DECL_P (base))
918 /* We found dereference of a pointer. Type of the pointer
919 and MEM_REF is meaningless, but we can look futher. */
920 offset_int mem_offset;
921 if (TREE_CODE (base) == MEM_REF
922 && mem_ref_offset (base).is_constant (&mem_offset))
924 offset_int o = mem_offset * BITS_PER_UNIT;
925 o += offset;
926 o += offset2;
927 if (!wi::fits_shwi_p (o))
928 break;
929 base_pointer = TREE_OPERAND (base, 0);
930 offset = o.to_shwi ();
931 outer_type = NULL;
933 /* We found base object. In this case the outer_type
934 is known. */
935 else if (DECL_P (base))
937 if (visited)
938 delete (visited);
939 /* Only type inconsistent programs can have otr_type that is
940 not part of outer type. */
941 if (otr_type
942 && !contains_type_p (TREE_TYPE (base),
943 offset + offset2, otr_type))
945 invalid = true;
946 if (instance)
947 *instance = base_pointer;
948 return;
950 set_by_decl (base, offset + offset2);
951 if (outer_type && maybe_in_construction && stmt)
952 maybe_in_construction
953 = decl_maybe_in_construction_p (base,
954 outer_type,
955 stmt,
956 fndecl);
957 if (instance)
958 *instance = base;
959 return;
961 else
962 break;
964 else
965 break;
967 else if (TREE_CODE (base_pointer) == POINTER_PLUS_EXPR
968 && TREE_CODE (TREE_OPERAND (base_pointer, 1)) == INTEGER_CST)
970 offset_int o
971 = offset_int::from (wi::to_wide (TREE_OPERAND (base_pointer, 1)),
972 SIGNED);
973 o *= BITS_PER_UNIT;
974 o += offset;
975 if (!wi::fits_shwi_p (o))
976 break;
977 offset = o.to_shwi ();
978 base_pointer = TREE_OPERAND (base_pointer, 0);
980 else
981 break;
984 if (visited)
985 delete (visited);
987 /* Try to determine type of the outer object. */
988 if (TREE_CODE (base_pointer) == SSA_NAME
989 && SSA_NAME_IS_DEFAULT_DEF (base_pointer)
990 && TREE_CODE (SSA_NAME_VAR (base_pointer)) == PARM_DECL)
992 /* See if parameter is THIS pointer of a method. */
993 if (TREE_CODE (TREE_TYPE (fndecl)) == METHOD_TYPE
994 && SSA_NAME_VAR (base_pointer) == DECL_ARGUMENTS (fndecl))
996 outer_type
997 = TYPE_MAIN_VARIANT (TREE_TYPE (TREE_TYPE (base_pointer)));
998 cgraph_node *node = cgraph_node::get (current_function_decl);
999 gcc_assert (TREE_CODE (outer_type) == RECORD_TYPE
1000 || TREE_CODE (outer_type) == UNION_TYPE);
1002 /* Handle the case we inlined into a thunk. In this case
1003 thunk has THIS pointer of type bar, but it really receives
1004 address to its base type foo which sits in bar at
1005 0-thunk.fixed_offset. It starts with code that adds
1006 think.fixed_offset to the pointer to compensate for this.
1008 Because we walked all the way to the begining of thunk, we now
1009 see pointer &bar-thunk.fixed_offset and need to compensate
1010 for it. */
1011 if (node->thunk.fixed_offset)
1012 offset -= node->thunk.fixed_offset * BITS_PER_UNIT;
1014 /* Dynamic casting has possibly upcasted the type
1015 in the hiearchy. In this case outer type is less
1016 informative than inner type and we should forget
1017 about it. */
1018 if ((otr_type
1019 && !contains_type_p (outer_type, offset,
1020 otr_type))
1021 || !contains_polymorphic_type_p (outer_type)
1022 /* If we compile thunk with virtual offset, the THIS pointer
1023 is adjusted by unknown value. We can't thus use outer info
1024 at all. */
1025 || node->thunk.virtual_offset_p)
1027 outer_type = NULL;
1028 if (instance)
1029 *instance = base_pointer;
1030 return;
1033 dynamic = true;
1035 /* If the function is constructor or destructor, then
1036 the type is possibly in construction, but we know
1037 it is not derived type. */
1038 if (DECL_CXX_CONSTRUCTOR_P (fndecl)
1039 || DECL_CXX_DESTRUCTOR_P (fndecl))
1041 maybe_in_construction = true;
1042 maybe_derived_type = false;
1044 else
1046 maybe_derived_type = true;
1047 maybe_in_construction = false;
1049 if (instance)
1051 /* If method is expanded thunk, we need to apply thunk offset
1052 to instance pointer. */
1053 if (node->thunk.virtual_offset_p
1054 || node->thunk.fixed_offset)
1055 *instance = NULL;
1056 else
1057 *instance = base_pointer;
1059 return;
1061 /* Non-PODs passed by value are really passed by invisible
1062 reference. In this case we also know the type of the
1063 object. */
1064 if (DECL_BY_REFERENCE (SSA_NAME_VAR (base_pointer)))
1066 outer_type
1067 = TYPE_MAIN_VARIANT (TREE_TYPE (TREE_TYPE (base_pointer)));
1068 /* Only type inconsistent programs can have otr_type that is
1069 not part of outer type. */
1070 if (otr_type && !contains_type_p (outer_type, offset,
1071 otr_type))
1073 invalid = true;
1074 if (instance)
1075 *instance = base_pointer;
1076 return;
1078 /* Non-polymorphic types have no interest for us. */
1079 else if (!otr_type && !contains_polymorphic_type_p (outer_type))
1081 outer_type = NULL;
1082 if (instance)
1083 *instance = base_pointer;
1084 return;
1086 maybe_derived_type = false;
1087 maybe_in_construction = false;
1088 if (instance)
1089 *instance = base_pointer;
1090 return;
1094 tree base_type = TREE_TYPE (base_pointer);
1096 if (TREE_CODE (base_pointer) == SSA_NAME
1097 && SSA_NAME_IS_DEFAULT_DEF (base_pointer)
1098 && !(TREE_CODE (SSA_NAME_VAR (base_pointer)) == PARM_DECL
1099 || TREE_CODE (SSA_NAME_VAR (base_pointer)) == RESULT_DECL))
1101 invalid = true;
1102 if (instance)
1103 *instance = base_pointer;
1104 return;
1106 if (TREE_CODE (base_pointer) == SSA_NAME
1107 && SSA_NAME_DEF_STMT (base_pointer)
1108 && gimple_assign_single_p (SSA_NAME_DEF_STMT (base_pointer)))
1109 base_type = TREE_TYPE (gimple_assign_rhs1
1110 (SSA_NAME_DEF_STMT (base_pointer)));
1112 if (base_type && POINTER_TYPE_P (base_type))
1113 combine_speculation_with (TYPE_MAIN_VARIANT (TREE_TYPE (base_type)),
1114 offset,
1115 true, NULL /* Do not change type here */);
1116 /* TODO: There are multiple ways to derive a type. For instance
1117 if BASE_POINTER is passed to an constructor call prior our refernece.
1118 We do not make this type of flow sensitive analysis yet. */
1119 if (instance)
1120 *instance = base_pointer;
1121 return;
1124 /* Structure to be passed in between detect_type_change and
1125 check_stmt_for_type_change. */
1127 struct type_change_info
1129 /* Offset into the object where there is the virtual method pointer we are
1130 looking for. */
1131 HOST_WIDE_INT offset;
1132 /* The declaration or SSA_NAME pointer of the base that we are checking for
1133 type change. */
1134 tree instance;
1135 /* The reference to virtual table pointer used. */
1136 tree vtbl_ptr_ref;
1137 tree otr_type;
1138 /* If we actually can tell the type that the object has changed to, it is
1139 stored in this field. Otherwise it remains NULL_TREE. */
1140 tree known_current_type;
1141 HOST_WIDE_INT known_current_offset;
1143 /* Set to nonzero if we possibly missed some dynamic type changes and we
1144 should consider the set to be speculative. */
1145 unsigned speculative;
1147 /* Set to true if dynamic type change has been detected. */
1148 bool type_maybe_changed;
1149 /* Set to true if multiple types have been encountered. known_current_type
1150 must be disregarded in that case. */
1151 bool multiple_types_encountered;
1152 bool seen_unanalyzed_store;
1155 /* Return true if STMT is not call and can modify a virtual method table pointer.
1156 We take advantage of fact that vtable stores must appear within constructor
1157 and destructor functions. */
1159 static bool
1160 noncall_stmt_may_be_vtbl_ptr_store (gimple *stmt)
1162 if (is_gimple_assign (stmt))
1164 tree lhs = gimple_assign_lhs (stmt);
1166 if (gimple_clobber_p (stmt))
1167 return false;
1168 if (!AGGREGATE_TYPE_P (TREE_TYPE (lhs)))
1170 if (flag_strict_aliasing
1171 && !POINTER_TYPE_P (TREE_TYPE (lhs)))
1172 return false;
1174 if (TREE_CODE (lhs) == COMPONENT_REF
1175 && !DECL_VIRTUAL_P (TREE_OPERAND (lhs, 1)))
1176 return false;
1177 /* In the future we might want to use get_ref_base_and_extent to find
1178 if there is a field corresponding to the offset and if so, proceed
1179 almost like if it was a component ref. */
1183 /* Code unification may mess with inline stacks. */
1184 if (cfun->after_inlining)
1185 return true;
1187 /* Walk the inline stack and watch out for ctors/dtors.
1188 TODO: Maybe we can require the store to appear in toplevel
1189 block of CTOR/DTOR. */
1190 for (tree block = gimple_block (stmt); block && TREE_CODE (block) == BLOCK;
1191 block = BLOCK_SUPERCONTEXT (block))
1192 if (BLOCK_ABSTRACT_ORIGIN (block)
1193 && TREE_CODE (block_ultimate_origin (block)) == FUNCTION_DECL)
1194 return inlined_polymorphic_ctor_dtor_block_p (block, false);
1195 return (TREE_CODE (TREE_TYPE (current_function_decl)) == METHOD_TYPE
1196 && (DECL_CXX_CONSTRUCTOR_P (current_function_decl)
1197 || DECL_CXX_DESTRUCTOR_P (current_function_decl)));
1200 /* If STMT can be proved to be an assignment to the virtual method table
1201 pointer of ANALYZED_OBJ and the type associated with the new table
1202 identified, return the type. Otherwise return NULL_TREE if type changes
1203 in unknown way or ERROR_MARK_NODE if type is unchanged. */
1205 static tree
1206 extr_type_from_vtbl_ptr_store (gimple *stmt, struct type_change_info *tci,
1207 HOST_WIDE_INT *type_offset)
1209 poly_int64 offset, size, max_size;
1210 tree lhs, rhs, base;
1211 bool reverse;
1213 if (!gimple_assign_single_p (stmt))
1214 return NULL_TREE;
1216 lhs = gimple_assign_lhs (stmt);
1217 rhs = gimple_assign_rhs1 (stmt);
1218 if (TREE_CODE (lhs) != COMPONENT_REF
1219 || !DECL_VIRTUAL_P (TREE_OPERAND (lhs, 1)))
1221 if (dump_file)
1222 fprintf (dump_file, " LHS is not virtual table.\n");
1223 return NULL_TREE;
1226 if (tci->vtbl_ptr_ref && operand_equal_p (lhs, tci->vtbl_ptr_ref, 0))
1228 else
1230 base = get_ref_base_and_extent (lhs, &offset, &size, &max_size, &reverse);
1231 if (DECL_P (tci->instance))
1233 if (base != tci->instance)
1235 if (dump_file)
1237 fprintf (dump_file, " base:");
1238 print_generic_expr (dump_file, base, TDF_SLIM);
1239 fprintf (dump_file, " does not match instance:");
1240 print_generic_expr (dump_file, tci->instance, TDF_SLIM);
1241 fprintf (dump_file, "\n");
1243 return NULL_TREE;
1246 else if (TREE_CODE (base) == MEM_REF)
1248 if (!operand_equal_p (tci->instance, TREE_OPERAND (base, 0), 0))
1250 if (dump_file)
1252 fprintf (dump_file, " base mem ref:");
1253 print_generic_expr (dump_file, base, TDF_SLIM);
1254 fprintf (dump_file, " does not match instance:");
1255 print_generic_expr (dump_file, tci->instance, TDF_SLIM);
1256 fprintf (dump_file, "\n");
1258 return NULL_TREE;
1260 if (!integer_zerop (TREE_OPERAND (base, 1)))
1262 if (!tree_fits_shwi_p (TREE_OPERAND (base, 1)))
1264 if (dump_file)
1266 fprintf (dump_file, " base mem ref:");
1267 print_generic_expr (dump_file, base, TDF_SLIM);
1268 fprintf (dump_file, " has non-representable offset:");
1269 print_generic_expr (dump_file, tci->instance, TDF_SLIM);
1270 fprintf (dump_file, "\n");
1272 return NULL_TREE;
1274 else
1275 offset += tree_to_shwi (TREE_OPERAND (base, 1)) * BITS_PER_UNIT;
1278 else if (!operand_equal_p (tci->instance, base, 0)
1279 || tci->offset)
1281 if (dump_file)
1283 fprintf (dump_file, " base:");
1284 print_generic_expr (dump_file, base, TDF_SLIM);
1285 fprintf (dump_file, " does not match instance:");
1286 print_generic_expr (dump_file, tci->instance, TDF_SLIM);
1287 fprintf (dump_file, " with offset %i\n", (int)tci->offset);
1289 return tci->offset > POINTER_SIZE ? error_mark_node : NULL_TREE;
1291 if (maybe_ne (offset, tci->offset)
1292 || maybe_ne (size, POINTER_SIZE)
1293 || maybe_ne (max_size, POINTER_SIZE))
1295 if (dump_file)
1297 fprintf (dump_file, " wrong offset ");
1298 print_dec (offset, dump_file);
1299 fprintf (dump_file, "!=%i or size ", (int) tci->offset);
1300 print_dec (size, dump_file);
1301 fprintf (dump_file, "\n");
1303 return (known_le (offset + POINTER_SIZE, tci->offset)
1304 || (known_size_p (max_size)
1305 && known_gt (tci->offset + POINTER_SIZE,
1306 offset + max_size))
1307 ? error_mark_node : NULL);
1311 tree vtable;
1312 unsigned HOST_WIDE_INT offset2;
1314 if (!vtable_pointer_value_to_vtable (rhs, &vtable, &offset2))
1316 if (dump_file)
1317 fprintf (dump_file, " Failed to lookup binfo\n");
1318 return NULL;
1321 tree binfo = subbinfo_with_vtable_at_offset (TYPE_BINFO (DECL_CONTEXT (vtable)),
1322 offset2, vtable);
1323 if (!binfo)
1325 if (dump_file)
1326 fprintf (dump_file, " Construction vtable used\n");
1327 /* FIXME: We should suport construction contexts. */
1328 return NULL;
1331 *type_offset = tree_to_shwi (BINFO_OFFSET (binfo)) * BITS_PER_UNIT;
1332 return DECL_CONTEXT (vtable);
1335 /* Record dynamic type change of TCI to TYPE. */
1337 static void
1338 record_known_type (struct type_change_info *tci, tree type, HOST_WIDE_INT offset)
1340 if (dump_file)
1342 if (type)
1344 fprintf (dump_file, " Recording type: ");
1345 print_generic_expr (dump_file, type, TDF_SLIM);
1346 fprintf (dump_file, " at offset %i\n", (int)offset);
1348 else
1349 fprintf (dump_file, " Recording unknown type\n");
1352 /* If we found a constructor of type that is not polymorphic or
1353 that may contain the type in question as a field (not as base),
1354 restrict to the inner class first to make type matching bellow
1355 happier. */
1356 if (type
1357 && (offset
1358 || (TREE_CODE (type) != RECORD_TYPE
1359 || !TYPE_BINFO (type)
1360 || !polymorphic_type_binfo_p (TYPE_BINFO (type)))))
1362 ipa_polymorphic_call_context context;
1364 context.offset = offset;
1365 context.outer_type = type;
1366 context.maybe_in_construction = false;
1367 context.maybe_derived_type = false;
1368 context.dynamic = true;
1369 /* If we failed to find the inner type, we know that the call
1370 would be undefined for type produced here. */
1371 if (!context.restrict_to_inner_class (tci->otr_type))
1373 if (dump_file)
1374 fprintf (dump_file, " Ignoring; does not contain otr_type\n");
1375 return;
1377 /* Watch for case we reached an POD type and anticipate placement
1378 new. */
1379 if (!context.maybe_derived_type)
1381 type = context.outer_type;
1382 offset = context.offset;
1385 if (tci->type_maybe_changed
1386 && (!types_same_for_odr (type, tci->known_current_type)
1387 || offset != tci->known_current_offset))
1388 tci->multiple_types_encountered = true;
1389 tci->known_current_type = TYPE_MAIN_VARIANT (type);
1390 tci->known_current_offset = offset;
1391 tci->type_maybe_changed = true;
1395 /* The maximum number of may-defs we visit when looking for a must-def
1396 that changes the dynamic type in check_stmt_for_type_change. Tuned
1397 after the PR12392 testcase which unlimited spends 40% time within
1398 these alias walks and 8% with the following limit. */
1400 static inline bool
1401 csftc_abort_walking_p (unsigned speculative)
1403 unsigned max = PARAM_VALUE (PARAM_MAX_SPECULATIVE_DEVIRT_MAYDEFS);
1404 return speculative > max ? true : false;
1407 /* Callback of walk_aliased_vdefs and a helper function for
1408 detect_type_change to check whether a particular statement may modify
1409 the virtual table pointer, and if possible also determine the new type of
1410 the (sub-)object. It stores its result into DATA, which points to a
1411 type_change_info structure. */
1413 static bool
1414 check_stmt_for_type_change (ao_ref *ao ATTRIBUTE_UNUSED, tree vdef, void *data)
1416 gimple *stmt = SSA_NAME_DEF_STMT (vdef);
1417 struct type_change_info *tci = (struct type_change_info *) data;
1418 tree fn;
1420 /* If we already gave up, just terminate the rest of walk. */
1421 if (tci->multiple_types_encountered)
1422 return true;
1424 if (is_gimple_call (stmt))
1426 if (gimple_call_flags (stmt) & (ECF_CONST | ECF_PURE))
1427 return false;
1429 /* Check for a constructor call. */
1430 if ((fn = gimple_call_fndecl (stmt)) != NULL_TREE
1431 && DECL_CXX_CONSTRUCTOR_P (fn)
1432 && TREE_CODE (TREE_TYPE (fn)) == METHOD_TYPE
1433 && gimple_call_num_args (stmt))
1435 tree op = walk_ssa_copies (gimple_call_arg (stmt, 0));
1436 tree type = TYPE_METHOD_BASETYPE (TREE_TYPE (fn));
1437 HOST_WIDE_INT offset = 0;
1438 bool reverse;
1440 if (dump_file)
1442 fprintf (dump_file, " Checking constructor call: ");
1443 print_gimple_stmt (dump_file, stmt, 0);
1446 /* See if THIS parameter seems like instance pointer. */
1447 if (TREE_CODE (op) == ADDR_EXPR)
1449 HOST_WIDE_INT size;
1450 op = get_ref_base_and_extent_hwi (TREE_OPERAND (op, 0),
1451 &offset, &size, &reverse);
1452 if (!op)
1454 tci->speculative++;
1455 return csftc_abort_walking_p (tci->speculative);
1457 if (TREE_CODE (op) == MEM_REF)
1459 if (!tree_fits_shwi_p (TREE_OPERAND (op, 1)))
1461 tci->speculative++;
1462 return csftc_abort_walking_p (tci->speculative);
1464 offset += tree_to_shwi (TREE_OPERAND (op, 1))
1465 * BITS_PER_UNIT;
1466 op = TREE_OPERAND (op, 0);
1468 else if (DECL_P (op))
1470 else
1472 tci->speculative++;
1473 return csftc_abort_walking_p (tci->speculative);
1475 op = walk_ssa_copies (op);
1477 if (operand_equal_p (op, tci->instance, 0)
1478 && TYPE_SIZE (type)
1479 && TREE_CODE (TYPE_SIZE (type)) == INTEGER_CST
1480 && tree_fits_shwi_p (TYPE_SIZE (type))
1481 && tree_to_shwi (TYPE_SIZE (type)) + offset > tci->offset
1482 /* Some inlined constructors may look as follows:
1483 _3 = operator new (16);
1484 MEM[(struct &)_3] ={v} {CLOBBER};
1485 MEM[(struct CompositeClass *)_3]._vptr.CompositeClass
1486 = &MEM[(void *)&_ZTV14CompositeClass + 16B];
1487 _7 = &MEM[(struct CompositeClass *)_3].object;
1488 EmptyClass::EmptyClass (_7);
1490 When determining dynamic type of _3 and because we stop at first
1491 dynamic type found, we would stop on EmptyClass::EmptyClass (_7).
1492 In this case the emptyclass is not even polymorphic and we miss
1493 it is contained in an outer type that is polymorphic. */
1495 && (tci->offset == offset || contains_polymorphic_type_p (type)))
1497 record_known_type (tci, type, tci->offset - offset);
1498 return true;
1501 /* Calls may possibly change dynamic type by placement new. Assume
1502 it will not happen, but make result speculative only. */
1503 if (dump_file)
1505 fprintf (dump_file, " Function call may change dynamic type:");
1506 print_gimple_stmt (dump_file, stmt, 0);
1508 tci->speculative++;
1509 return csftc_abort_walking_p (tci->speculative);
1511 /* Check for inlined virtual table store. */
1512 else if (noncall_stmt_may_be_vtbl_ptr_store (stmt))
1514 tree type;
1515 HOST_WIDE_INT offset = 0;
1516 if (dump_file)
1518 fprintf (dump_file, " Checking vtbl store: ");
1519 print_gimple_stmt (dump_file, stmt, 0);
1522 type = extr_type_from_vtbl_ptr_store (stmt, tci, &offset);
1523 if (type == error_mark_node)
1524 return false;
1525 gcc_assert (!type || TYPE_MAIN_VARIANT (type) == type);
1526 if (!type)
1528 if (dump_file)
1529 fprintf (dump_file, " Unanalyzed store may change type.\n");
1530 tci->seen_unanalyzed_store = true;
1531 tci->speculative++;
1533 else
1534 record_known_type (tci, type, offset);
1535 return true;
1537 else
1538 return false;
1541 /* THIS is polymorphic call context obtained from get_polymorphic_context.
1542 OTR_OBJECT is pointer to the instance returned by OBJ_TYPE_REF_OBJECT.
1543 INSTANCE is pointer to the outer instance as returned by
1544 get_polymorphic_context. To avoid creation of temporary expressions,
1545 INSTANCE may also be an declaration of get_polymorphic_context found the
1546 value to be in static storage.
1548 If the type of instance is not fully determined
1549 (either OUTER_TYPE is unknown or MAYBE_IN_CONSTRUCTION/INCLUDE_DERIVED_TYPES
1550 is set), try to walk memory writes and find the actual construction of the
1551 instance.
1553 Return true if memory is unchanged from function entry.
1555 We do not include this analysis in the context analysis itself, because
1556 it needs memory SSA to be fully built and the walk may be expensive.
1557 So it is not suitable for use withing fold_stmt and similar uses.
1559 AA_WALK_BUDGET_P, if not NULL, is how statements we should allow
1560 walk_aliased_vdefs to examine. The value should be decremented by the
1561 number of stetements we examined or set to zero if exhausted. */
1563 bool
1564 ipa_polymorphic_call_context::get_dynamic_type (tree instance,
1565 tree otr_object,
1566 tree otr_type,
1567 gimple *call,
1568 unsigned *aa_walk_budget_p)
1570 struct type_change_info tci;
1571 ao_ref ao;
1572 bool function_entry_reached = false;
1573 tree instance_ref = NULL;
1574 gimple *stmt = call;
1575 /* Remember OFFSET before it is modified by restrict_to_inner_class.
1576 This is because we do not update INSTANCE when walking inwards. */
1577 HOST_WIDE_INT instance_offset = offset;
1578 tree instance_outer_type = outer_type;
1580 if (!instance)
1581 return false;
1583 if (otr_type)
1584 otr_type = TYPE_MAIN_VARIANT (otr_type);
1586 /* Walk into inner type. This may clear maybe_derived_type and save us
1587 from useless work. It also makes later comparsions with static type
1588 easier. */
1589 if (outer_type && otr_type)
1591 if (!restrict_to_inner_class (otr_type))
1592 return false;
1595 if (!maybe_in_construction && !maybe_derived_type)
1596 return false;
1598 /* If we are in fact not looking at any object object or the instance is
1599 some placement new into a random load, give up straight away. */
1600 if (TREE_CODE (instance) == MEM_REF)
1601 return false;
1603 /* We need to obtain refernce to virtual table pointer. It is better
1604 to look it up in the code rather than build our own. This require bit
1605 of pattern matching, but we end up verifying that what we found is
1606 correct.
1608 What we pattern match is:
1610 tmp = instance->_vptr.A; // vtbl ptr load
1611 tmp2 = tmp[otr_token]; // vtable lookup
1612 OBJ_TYPE_REF(tmp2;instance->0) (instance);
1614 We want to start alias oracle walk from vtbl pointer load,
1615 but we may not be able to identify it, for example, when PRE moved the
1616 load around. */
1618 if (gimple_code (call) == GIMPLE_CALL)
1620 tree ref = gimple_call_fn (call);
1621 bool reverse;
1623 if (TREE_CODE (ref) == OBJ_TYPE_REF)
1625 ref = OBJ_TYPE_REF_EXPR (ref);
1626 ref = walk_ssa_copies (ref);
1628 /* If call target is already known, no need to do the expensive
1629 memory walk. */
1630 if (is_gimple_min_invariant (ref))
1631 return false;
1633 /* Check if definition looks like vtable lookup. */
1634 if (TREE_CODE (ref) == SSA_NAME
1635 && !SSA_NAME_IS_DEFAULT_DEF (ref)
1636 && gimple_assign_load_p (SSA_NAME_DEF_STMT (ref))
1637 && TREE_CODE (gimple_assign_rhs1
1638 (SSA_NAME_DEF_STMT (ref))) == MEM_REF)
1640 ref = get_base_address
1641 (TREE_OPERAND (gimple_assign_rhs1
1642 (SSA_NAME_DEF_STMT (ref)), 0));
1643 ref = walk_ssa_copies (ref);
1644 /* Find base address of the lookup and see if it looks like
1645 vptr load. */
1646 if (TREE_CODE (ref) == SSA_NAME
1647 && !SSA_NAME_IS_DEFAULT_DEF (ref)
1648 && gimple_assign_load_p (SSA_NAME_DEF_STMT (ref)))
1650 HOST_WIDE_INT offset2, size;
1651 tree ref_exp = gimple_assign_rhs1 (SSA_NAME_DEF_STMT (ref));
1652 tree base_ref
1653 = get_ref_base_and_extent_hwi (ref_exp, &offset2,
1654 &size, &reverse);
1656 /* Finally verify that what we found looks like read from
1657 OTR_OBJECT or from INSTANCE with offset OFFSET. */
1658 if (base_ref
1659 && ((TREE_CODE (base_ref) == MEM_REF
1660 && ((offset2 == instance_offset
1661 && TREE_OPERAND (base_ref, 0) == instance)
1662 || (!offset2
1663 && TREE_OPERAND (base_ref, 0)
1664 == otr_object)))
1665 || (DECL_P (instance) && base_ref == instance
1666 && offset2 == instance_offset)))
1668 stmt = SSA_NAME_DEF_STMT (ref);
1669 instance_ref = ref_exp;
1676 /* If we failed to look up the reference in code, build our own. */
1677 if (!instance_ref)
1679 /* If the statement in question does not use memory, we can't tell
1680 anything. */
1681 if (!gimple_vuse (stmt))
1682 return false;
1683 ao_ref_init_from_ptr_and_size (&ao, otr_object, NULL);
1685 else
1686 /* Otherwise use the real reference. */
1687 ao_ref_init (&ao, instance_ref);
1689 /* We look for vtbl pointer read. */
1690 ao.size = POINTER_SIZE;
1691 ao.max_size = ao.size;
1692 /* We are looking for stores to vptr pointer within the instance of
1693 outer type.
1694 TODO: The vptr pointer type is globally known, we probably should
1695 keep it and do that even when otr_type is unknown. */
1696 if (otr_type)
1698 ao.base_alias_set
1699 = get_alias_set (outer_type ? outer_type : otr_type);
1700 ao.ref_alias_set
1701 = get_alias_set (TREE_TYPE (BINFO_VTABLE (TYPE_BINFO (otr_type))));
1704 if (dump_file)
1706 fprintf (dump_file, "Determining dynamic type for call: ");
1707 print_gimple_stmt (dump_file, call, 0);
1708 fprintf (dump_file, " Starting walk at: ");
1709 print_gimple_stmt (dump_file, stmt, 0);
1710 fprintf (dump_file, " instance pointer: ");
1711 print_generic_expr (dump_file, otr_object, TDF_SLIM);
1712 fprintf (dump_file, " Outer instance pointer: ");
1713 print_generic_expr (dump_file, instance, TDF_SLIM);
1714 fprintf (dump_file, " offset: %i (bits)", (int)instance_offset);
1715 fprintf (dump_file, " vtbl reference: ");
1716 print_generic_expr (dump_file, instance_ref, TDF_SLIM);
1717 fprintf (dump_file, "\n");
1720 tci.offset = instance_offset;
1721 tci.instance = instance;
1722 tci.vtbl_ptr_ref = instance_ref;
1723 tci.known_current_type = NULL_TREE;
1724 tci.known_current_offset = 0;
1725 tci.otr_type = otr_type;
1726 tci.type_maybe_changed = false;
1727 tci.multiple_types_encountered = false;
1728 tci.speculative = 0;
1729 tci.seen_unanalyzed_store = false;
1731 unsigned aa_walk_budget = 0;
1732 if (aa_walk_budget_p)
1733 aa_walk_budget = *aa_walk_budget_p + 1;
1735 int walked
1736 = walk_aliased_vdefs (&ao, gimple_vuse (stmt), check_stmt_for_type_change,
1737 &tci, NULL, &function_entry_reached, aa_walk_budget);
1739 /* If we did not find any type changing statements, we may still drop
1740 maybe_in_construction flag if the context already have outer type.
1742 Here we make special assumptions about both constructors and
1743 destructors which are all the functions that are allowed to alter the
1744 VMT pointers. It assumes that destructors begin with assignment into
1745 all VMT pointers and that constructors essentially look in the
1746 following way:
1748 1) The very first thing they do is that they call constructors of
1749 ancestor sub-objects that have them.
1751 2) Then VMT pointers of this and all its ancestors is set to new
1752 values corresponding to the type corresponding to the constructor.
1754 3) Only afterwards, other stuff such as constructor of member
1755 sub-objects and the code written by the user is run. Only this may
1756 include calling virtual functions, directly or indirectly.
1758 4) placement new cannot be used to change type of non-POD statically
1759 allocated variables.
1761 There is no way to call a constructor of an ancestor sub-object in any
1762 other way.
1764 This means that we do not have to care whether constructors get the
1765 correct type information because they will always change it (in fact,
1766 if we define the type to be given by the VMT pointer, it is undefined).
1768 The most important fact to derive from the above is that if, for some
1769 statement in the section 3, we try to detect whether the dynamic type
1770 has changed, we can safely ignore all calls as we examine the function
1771 body backwards until we reach statements in section 2 because these
1772 calls cannot be ancestor constructors or destructors (if the input is
1773 not bogus) and so do not change the dynamic type (this holds true only
1774 for automatically allocated objects but at the moment we devirtualize
1775 only these). We then must detect that statements in section 2 change
1776 the dynamic type and can try to derive the new type. That is enough
1777 and we can stop, we will never see the calls into constructors of
1778 sub-objects in this code.
1780 Therefore if the static outer type was found (outer_type)
1781 we can safely ignore tci.speculative that is set on calls and give up
1782 only if there was dyanmic type store that may affect given variable
1783 (seen_unanalyzed_store) */
1785 if (walked < 0)
1787 if (dump_file)
1788 fprintf (dump_file, " AA walk budget exhausted.\n");
1789 *aa_walk_budget_p = 0;
1790 return false;
1792 else if (aa_walk_budget_p)
1793 *aa_walk_budget_p -= walked;
1795 if (!tci.type_maybe_changed
1796 || (outer_type
1797 && !dynamic
1798 && !tci.seen_unanalyzed_store
1799 && !tci.multiple_types_encountered
1800 && ((offset == tci.offset
1801 && types_same_for_odr (tci.known_current_type,
1802 outer_type))
1803 || (instance_offset == offset
1804 && types_same_for_odr (tci.known_current_type,
1805 instance_outer_type)))))
1807 if (!outer_type || tci.seen_unanalyzed_store)
1808 return false;
1809 if (maybe_in_construction)
1810 maybe_in_construction = false;
1811 if (dump_file)
1812 fprintf (dump_file, " No dynamic type change found.\n");
1813 return true;
1816 if (tci.known_current_type
1817 && !function_entry_reached
1818 && !tci.multiple_types_encountered)
1820 if (!tci.speculative)
1822 outer_type = TYPE_MAIN_VARIANT (tci.known_current_type);
1823 offset = tci.known_current_offset;
1824 dynamic = true;
1825 maybe_in_construction = false;
1826 maybe_derived_type = false;
1827 if (dump_file)
1828 fprintf (dump_file, " Determined dynamic type.\n");
1830 else if (!speculative_outer_type
1831 || speculative_maybe_derived_type)
1833 speculative_outer_type = TYPE_MAIN_VARIANT (tci.known_current_type);
1834 speculative_offset = tci.known_current_offset;
1835 speculative_maybe_derived_type = false;
1836 if (dump_file)
1837 fprintf (dump_file, " Determined speculative dynamic type.\n");
1840 else if (dump_file)
1842 fprintf (dump_file, " Found multiple types%s%s\n",
1843 function_entry_reached ? " (function entry reached)" : "",
1844 function_entry_reached ? " (multiple types encountered)" : "");
1847 return false;
1850 /* See if speculation given by SPEC_OUTER_TYPE, SPEC_OFFSET and SPEC_MAYBE_DERIVED_TYPE
1851 seems consistent (and useful) with what we already have in the non-speculative context. */
1853 bool
1854 ipa_polymorphic_call_context::speculation_consistent_p (tree spec_outer_type,
1855 HOST_WIDE_INT spec_offset,
1856 bool spec_maybe_derived_type,
1857 tree otr_type) const
1859 if (!flag_devirtualize_speculatively)
1860 return false;
1862 /* Non-polymorphic types are useless for deriving likely polymorphic
1863 call targets. */
1864 if (!spec_outer_type || !contains_polymorphic_type_p (spec_outer_type))
1865 return false;
1867 /* If we know nothing, speculation is always good. */
1868 if (!outer_type)
1869 return true;
1871 /* Speculation is only useful to avoid derived types.
1872 This is not 100% true for placement new, where the outer context may
1873 turn out to be useless, but ignore these for now. */
1874 if (!maybe_derived_type)
1875 return false;
1877 /* If types agrees, speculation is consistent, but it makes sense only
1878 when it says something new. */
1879 if (types_must_be_same_for_odr (spec_outer_type, outer_type))
1880 return maybe_derived_type && !spec_maybe_derived_type;
1882 /* If speculation does not contain the type in question, ignore it. */
1883 if (otr_type
1884 && !contains_type_p (spec_outer_type, spec_offset, otr_type, false, true))
1885 return false;
1887 /* If outer type already contains speculation as a filed,
1888 it is useless. We already know from OUTER_TYPE
1889 SPEC_TYPE and that it is not in the construction. */
1890 if (contains_type_p (outer_type, offset - spec_offset,
1891 spec_outer_type, false, false))
1892 return false;
1894 /* If speculative outer type is not more specified than outer
1895 type, just give up.
1896 We can only decide this safely if we can compare types with OUTER_TYPE.
1898 if ((!in_lto_p || odr_type_p (outer_type))
1899 && !contains_type_p (spec_outer_type,
1900 spec_offset - offset,
1901 outer_type, false))
1902 return false;
1903 return true;
1906 /* Improve THIS with speculation described by NEW_OUTER_TYPE, NEW_OFFSET,
1907 NEW_MAYBE_DERIVED_TYPE
1908 If OTR_TYPE is set, assume the context is used with OTR_TYPE. */
1910 bool
1911 ipa_polymorphic_call_context::combine_speculation_with
1912 (tree new_outer_type, HOST_WIDE_INT new_offset, bool new_maybe_derived_type,
1913 tree otr_type)
1915 if (!new_outer_type)
1916 return false;
1918 /* restrict_to_inner_class may eliminate wrong speculation making our job
1919 easeier. */
1920 if (otr_type)
1921 restrict_to_inner_class (otr_type);
1923 if (!speculation_consistent_p (new_outer_type, new_offset,
1924 new_maybe_derived_type, otr_type))
1925 return false;
1927 /* New speculation is a win in case we have no speculation or new
1928 speculation does not consider derivations. */
1929 if (!speculative_outer_type
1930 || (speculative_maybe_derived_type
1931 && !new_maybe_derived_type))
1933 speculative_outer_type = new_outer_type;
1934 speculative_offset = new_offset;
1935 speculative_maybe_derived_type = new_maybe_derived_type;
1936 return true;
1938 else if (types_must_be_same_for_odr (speculative_outer_type,
1939 new_outer_type))
1941 if (speculative_offset != new_offset)
1943 /* OK we have two contexts that seems valid but they disagree,
1944 just give up.
1946 This is not a lattice operation, so we may want to drop it later. */
1947 if (dump_file && (dump_flags & TDF_DETAILS))
1948 fprintf (dump_file,
1949 "Speculative outer types match, "
1950 "offset mismatch -> invalid speculation\n");
1951 clear_speculation ();
1952 return true;
1954 else
1956 if (speculative_maybe_derived_type && !new_maybe_derived_type)
1958 speculative_maybe_derived_type = false;
1959 return true;
1961 else
1962 return false;
1965 /* Choose type that contains the other. This one either contains the outer
1966 as a field (thus giving exactly one target) or is deeper in the type
1967 hiearchy. */
1968 else if (speculative_outer_type
1969 && speculative_maybe_derived_type
1970 && (new_offset > speculative_offset
1971 || (new_offset == speculative_offset
1972 && contains_type_p (new_outer_type,
1973 0, speculative_outer_type, false))))
1975 tree old_outer_type = speculative_outer_type;
1976 HOST_WIDE_INT old_offset = speculative_offset;
1977 bool old_maybe_derived_type = speculative_maybe_derived_type;
1979 speculative_outer_type = new_outer_type;
1980 speculative_offset = new_offset;
1981 speculative_maybe_derived_type = new_maybe_derived_type;
1983 if (otr_type)
1984 restrict_to_inner_class (otr_type);
1986 /* If the speculation turned out to make no sense, revert to sensible
1987 one. */
1988 if (!speculative_outer_type)
1990 speculative_outer_type = old_outer_type;
1991 speculative_offset = old_offset;
1992 speculative_maybe_derived_type = old_maybe_derived_type;
1993 return false;
1995 return (old_offset != speculative_offset
1996 || old_maybe_derived_type != speculative_maybe_derived_type
1997 || types_must_be_same_for_odr (speculative_outer_type,
1998 new_outer_type));
2000 return false;
2003 /* Make speculation less specific so
2004 NEW_OUTER_TYPE, NEW_OFFSET, NEW_MAYBE_DERIVED_TYPE is also included.
2005 If OTR_TYPE is set, assume the context is used with OTR_TYPE. */
2007 bool
2008 ipa_polymorphic_call_context::meet_speculation_with
2009 (tree new_outer_type, HOST_WIDE_INT new_offset, bool new_maybe_derived_type,
2010 tree otr_type)
2012 if (!new_outer_type && speculative_outer_type)
2014 clear_speculation ();
2015 return true;
2018 /* restrict_to_inner_class may eliminate wrong speculation making our job
2019 easeier. */
2020 if (otr_type)
2021 restrict_to_inner_class (otr_type);
2023 if (!speculative_outer_type
2024 || !speculation_consistent_p (speculative_outer_type,
2025 speculative_offset,
2026 speculative_maybe_derived_type,
2027 otr_type))
2028 return false;
2030 if (!speculation_consistent_p (new_outer_type, new_offset,
2031 new_maybe_derived_type, otr_type))
2033 clear_speculation ();
2034 return true;
2037 else if (types_must_be_same_for_odr (speculative_outer_type,
2038 new_outer_type))
2040 if (speculative_offset != new_offset)
2042 clear_speculation ();
2043 return true;
2045 else
2047 if (!speculative_maybe_derived_type && new_maybe_derived_type)
2049 speculative_maybe_derived_type = true;
2050 return true;
2052 else
2053 return false;
2056 /* See if one type contains the other as a field (not base). */
2057 else if (contains_type_p (new_outer_type, new_offset - speculative_offset,
2058 speculative_outer_type, false, false))
2059 return false;
2060 else if (contains_type_p (speculative_outer_type,
2061 speculative_offset - new_offset,
2062 new_outer_type, false, false))
2064 speculative_outer_type = new_outer_type;
2065 speculative_offset = new_offset;
2066 speculative_maybe_derived_type = new_maybe_derived_type;
2067 return true;
2069 /* See if OUTER_TYPE is base of CTX.OUTER_TYPE. */
2070 else if (contains_type_p (new_outer_type,
2071 new_offset - speculative_offset,
2072 speculative_outer_type, false, true))
2074 if (!speculative_maybe_derived_type)
2076 speculative_maybe_derived_type = true;
2077 return true;
2079 return false;
2081 /* See if CTX.OUTER_TYPE is base of OUTER_TYPE. */
2082 else if (contains_type_p (speculative_outer_type,
2083 speculative_offset - new_offset, new_outer_type, false, true))
2085 speculative_outer_type = new_outer_type;
2086 speculative_offset = new_offset;
2087 speculative_maybe_derived_type = true;
2088 return true;
2090 else
2092 if (dump_file && (dump_flags & TDF_DETAILS))
2093 fprintf (dump_file, "Giving up on speculative meet\n");
2094 clear_speculation ();
2095 return true;
2099 /* Assume that both THIS and a given context is valid and strenghten THIS
2100 if possible. Return true if any strenghtening was made.
2101 If actual type the context is being used in is known, OTR_TYPE should be
2102 set accordingly. This improves quality of combined result. */
2104 bool
2105 ipa_polymorphic_call_context::combine_with (ipa_polymorphic_call_context ctx,
2106 tree otr_type)
2108 bool updated = false;
2110 if (ctx.useless_p () || invalid)
2111 return false;
2113 /* Restricting context to inner type makes merging easier, however do not
2114 do that unless we know how the context is used (OTR_TYPE is non-NULL) */
2115 if (otr_type && !invalid && !ctx.invalid)
2117 restrict_to_inner_class (otr_type);
2118 ctx.restrict_to_inner_class (otr_type);
2119 if(invalid)
2120 return false;
2123 if (dump_file && (dump_flags & TDF_DETAILS))
2125 fprintf (dump_file, "Polymorphic call context combine:");
2126 dump (dump_file);
2127 fprintf (dump_file, "With context: ");
2128 ctx.dump (dump_file);
2129 if (otr_type)
2131 fprintf (dump_file, "To be used with type: ");
2132 print_generic_expr (dump_file, otr_type, TDF_SLIM);
2133 fprintf (dump_file, "\n");
2137 /* If call is known to be invalid, we are done. */
2138 if (ctx.invalid)
2140 if (dump_file && (dump_flags & TDF_DETAILS))
2141 fprintf (dump_file, "-> Invalid context\n");
2142 goto invalidate;
2145 if (!ctx.outer_type)
2147 else if (!outer_type)
2149 outer_type = ctx.outer_type;
2150 offset = ctx.offset;
2151 dynamic = ctx.dynamic;
2152 maybe_in_construction = ctx.maybe_in_construction;
2153 maybe_derived_type = ctx.maybe_derived_type;
2154 updated = true;
2156 /* If types are known to be same, merging is quite easy. */
2157 else if (types_must_be_same_for_odr (outer_type, ctx.outer_type))
2159 if (offset != ctx.offset
2160 && TYPE_SIZE (outer_type)
2161 && TREE_CODE (TYPE_SIZE (outer_type)) == INTEGER_CST)
2163 if (dump_file && (dump_flags & TDF_DETAILS))
2164 fprintf (dump_file, "Outer types match, offset mismatch -> invalid\n");
2165 clear_speculation ();
2166 clear_outer_type ();
2167 invalid = true;
2168 return true;
2170 if (dump_file && (dump_flags & TDF_DETAILS))
2171 fprintf (dump_file, "Outer types match, merging flags\n");
2172 if (maybe_in_construction && !ctx.maybe_in_construction)
2174 updated = true;
2175 maybe_in_construction = false;
2177 if (maybe_derived_type && !ctx.maybe_derived_type)
2179 updated = true;
2180 maybe_derived_type = false;
2182 if (dynamic && !ctx.dynamic)
2184 updated = true;
2185 dynamic = false;
2188 /* If we know the type precisely, there is not much to improve. */
2189 else if (!maybe_derived_type && !maybe_in_construction
2190 && !ctx.maybe_derived_type && !ctx.maybe_in_construction)
2192 /* It may be easy to check if second context permits the first
2193 and set INVALID otherwise. This is not easy to do in general;
2194 contains_type_p may return false negatives for non-comparable
2195 types.
2197 If OTR_TYPE is known, we however can expect that
2198 restrict_to_inner_class should have discovered the same base
2199 type. */
2200 if (otr_type && !ctx.maybe_in_construction && !ctx.maybe_derived_type)
2202 if (dump_file && (dump_flags & TDF_DETAILS))
2203 fprintf (dump_file, "Contextes disagree -> invalid\n");
2204 goto invalidate;
2207 /* See if one type contains the other as a field (not base).
2208 In this case we want to choose the wider type, because it contains
2209 more information. */
2210 else if (contains_type_p (ctx.outer_type, ctx.offset - offset,
2211 outer_type, false, false))
2213 if (dump_file && (dump_flags & TDF_DETAILS))
2214 fprintf (dump_file, "Second type contain the first as a field\n");
2216 if (maybe_derived_type)
2218 outer_type = ctx.outer_type;
2219 maybe_derived_type = ctx.maybe_derived_type;
2220 offset = ctx.offset;
2221 dynamic = ctx.dynamic;
2222 updated = true;
2225 /* If we do not know how the context is being used, we cannot
2226 clear MAYBE_IN_CONSTRUCTION because it may be offseted
2227 to other component of OUTER_TYPE later and we know nothing
2228 about it. */
2229 if (otr_type && maybe_in_construction
2230 && !ctx.maybe_in_construction)
2232 maybe_in_construction = false;
2233 updated = true;
2236 else if (contains_type_p (outer_type, offset - ctx.offset,
2237 ctx.outer_type, false, false))
2239 if (dump_file && (dump_flags & TDF_DETAILS))
2240 fprintf (dump_file, "First type contain the second as a field\n");
2242 if (otr_type && maybe_in_construction
2243 && !ctx.maybe_in_construction)
2245 maybe_in_construction = false;
2246 updated = true;
2249 /* See if OUTER_TYPE is base of CTX.OUTER_TYPE. */
2250 else if (contains_type_p (ctx.outer_type,
2251 ctx.offset - offset, outer_type, false, true))
2253 if (dump_file && (dump_flags & TDF_DETAILS))
2254 fprintf (dump_file, "First type is base of second\n");
2255 if (!maybe_derived_type)
2257 if (!ctx.maybe_in_construction
2258 && types_odr_comparable (outer_type, ctx.outer_type))
2260 if (dump_file && (dump_flags & TDF_DETAILS))
2261 fprintf (dump_file, "Second context does not permit base -> invalid\n");
2262 goto invalidate;
2265 /* Pick variant deeper in the hiearchy. */
2266 else
2268 outer_type = ctx.outer_type;
2269 maybe_in_construction = ctx.maybe_in_construction;
2270 maybe_derived_type = ctx.maybe_derived_type;
2271 offset = ctx.offset;
2272 dynamic = ctx.dynamic;
2273 updated = true;
2276 /* See if CTX.OUTER_TYPE is base of OUTER_TYPE. */
2277 else if (contains_type_p (outer_type,
2278 offset - ctx.offset, ctx.outer_type, false, true))
2280 if (dump_file && (dump_flags & TDF_DETAILS))
2281 fprintf (dump_file, "Second type is base of first\n");
2282 if (!ctx.maybe_derived_type)
2284 if (!maybe_in_construction
2285 && types_odr_comparable (outer_type, ctx.outer_type))
2287 if (dump_file && (dump_flags & TDF_DETAILS))
2288 fprintf (dump_file, "First context does not permit base -> invalid\n");
2289 goto invalidate;
2291 /* Pick the base type. */
2292 else if (maybe_in_construction)
2294 outer_type = ctx.outer_type;
2295 maybe_in_construction = ctx.maybe_in_construction;
2296 maybe_derived_type = ctx.maybe_derived_type;
2297 offset = ctx.offset;
2298 dynamic = ctx.dynamic;
2299 updated = true;
2303 /* TODO handle merging using hiearchy. */
2304 else if (dump_file && (dump_flags & TDF_DETAILS))
2305 fprintf (dump_file, "Giving up on merge\n");
2307 updated |= combine_speculation_with (ctx.speculative_outer_type,
2308 ctx.speculative_offset,
2309 ctx.speculative_maybe_derived_type,
2310 otr_type);
2312 if (updated && dump_file && (dump_flags & TDF_DETAILS))
2314 fprintf (dump_file, "Updated as: ");
2315 dump (dump_file);
2316 fprintf (dump_file, "\n");
2318 return updated;
2320 invalidate:
2321 invalid = true;
2322 clear_speculation ();
2323 clear_outer_type ();
2324 return true;
2327 /* Take non-speculative info, merge it with speculative and clear speculation.
2328 Used when we no longer manage to keep track of actual outer type, but we
2329 think it is still there.
2331 If OTR_TYPE is set, the transformation can be done more effectively assuming
2332 that context is going to be used only that way. */
2334 void
2335 ipa_polymorphic_call_context::make_speculative (tree otr_type)
2337 tree spec_outer_type = outer_type;
2338 HOST_WIDE_INT spec_offset = offset;
2339 bool spec_maybe_derived_type = maybe_derived_type;
2341 if (invalid)
2343 invalid = false;
2344 clear_outer_type ();
2345 clear_speculation ();
2346 return;
2348 if (!outer_type)
2349 return;
2350 clear_outer_type ();
2351 combine_speculation_with (spec_outer_type, spec_offset,
2352 spec_maybe_derived_type,
2353 otr_type);
2356 /* Use when we cannot track dynamic type change. This speculatively assume
2357 type change is not happening. */
2359 void
2360 ipa_polymorphic_call_context::possible_dynamic_type_change (bool in_poly_cdtor,
2361 tree otr_type)
2363 if (dynamic)
2364 make_speculative (otr_type);
2365 else if (in_poly_cdtor)
2366 maybe_in_construction = true;
2369 /* Return TRUE if this context conveys the same information as OTHER. */
2371 bool
2372 ipa_polymorphic_call_context::equal_to
2373 (const ipa_polymorphic_call_context &x) const
2375 if (useless_p ())
2376 return x.useless_p ();
2377 if (invalid)
2378 return x.invalid;
2379 if (x.useless_p () || x.invalid)
2380 return false;
2382 if (outer_type)
2384 if (!x.outer_type
2385 || !types_odr_comparable (outer_type, x.outer_type)
2386 || !types_same_for_odr (outer_type, x.outer_type)
2387 || offset != x.offset
2388 || maybe_in_construction != x.maybe_in_construction
2389 || maybe_derived_type != x.maybe_derived_type
2390 || dynamic != x.dynamic)
2391 return false;
2393 else if (x.outer_type)
2394 return false;
2397 if (speculative_outer_type
2398 && speculation_consistent_p (speculative_outer_type, speculative_offset,
2399 speculative_maybe_derived_type, NULL_TREE))
2401 if (!x.speculative_outer_type)
2402 return false;
2404 if (!types_odr_comparable (speculative_outer_type,
2405 x.speculative_outer_type)
2406 || !types_same_for_odr (speculative_outer_type,
2407 x.speculative_outer_type)
2408 || speculative_offset != x.speculative_offset
2409 || speculative_maybe_derived_type != x.speculative_maybe_derived_type)
2410 return false;
2412 else if (x.speculative_outer_type
2413 && x.speculation_consistent_p (x.speculative_outer_type,
2414 x.speculative_offset,
2415 x.speculative_maybe_derived_type,
2416 NULL))
2417 return false;
2419 return true;
2422 /* Modify context to be strictly less restrictive than CTX. */
2424 bool
2425 ipa_polymorphic_call_context::meet_with (ipa_polymorphic_call_context ctx,
2426 tree otr_type)
2428 bool updated = false;
2430 if (useless_p () || ctx.invalid)
2431 return false;
2433 /* Restricting context to inner type makes merging easier, however do not
2434 do that unless we know how the context is used (OTR_TYPE is non-NULL) */
2435 if (otr_type && !useless_p () && !ctx.useless_p ())
2437 restrict_to_inner_class (otr_type);
2438 ctx.restrict_to_inner_class (otr_type);
2439 if(invalid)
2440 return false;
2443 if (equal_to (ctx))
2444 return false;
2446 if (ctx.useless_p () || invalid)
2448 *this = ctx;
2449 return true;
2452 if (dump_file && (dump_flags & TDF_DETAILS))
2454 fprintf (dump_file, "Polymorphic call context meet:");
2455 dump (dump_file);
2456 fprintf (dump_file, "With context: ");
2457 ctx.dump (dump_file);
2458 if (otr_type)
2460 fprintf (dump_file, "To be used with type: ");
2461 print_generic_expr (dump_file, otr_type, TDF_SLIM);
2462 fprintf (dump_file, "\n");
2466 if (!dynamic && ctx.dynamic)
2468 dynamic = true;
2469 updated = true;
2472 /* If call is known to be invalid, we are done. */
2473 if (!outer_type)
2475 else if (!ctx.outer_type)
2477 clear_outer_type ();
2478 updated = true;
2480 /* If types are known to be same, merging is quite easy. */
2481 else if (types_must_be_same_for_odr (outer_type, ctx.outer_type))
2483 if (offset != ctx.offset
2484 && TYPE_SIZE (outer_type)
2485 && TREE_CODE (TYPE_SIZE (outer_type)) == INTEGER_CST)
2487 if (dump_file && (dump_flags & TDF_DETAILS))
2488 fprintf (dump_file, "Outer types match, offset mismatch -> clearing\n");
2489 clear_outer_type ();
2490 return true;
2492 if (dump_file && (dump_flags & TDF_DETAILS))
2493 fprintf (dump_file, "Outer types match, merging flags\n");
2494 if (!maybe_in_construction && ctx.maybe_in_construction)
2496 updated = true;
2497 maybe_in_construction = true;
2499 if (!maybe_derived_type && ctx.maybe_derived_type)
2501 updated = true;
2502 maybe_derived_type = true;
2504 if (!dynamic && ctx.dynamic)
2506 updated = true;
2507 dynamic = true;
2510 /* See if one type contains the other as a field (not base). */
2511 else if (contains_type_p (ctx.outer_type, ctx.offset - offset,
2512 outer_type, false, false))
2514 if (dump_file && (dump_flags & TDF_DETAILS))
2515 fprintf (dump_file, "Second type contain the first as a field\n");
2517 /* The second type is more specified, so we keep the first.
2518 We need to set DYNAMIC flag to avoid declaring context INVALID
2519 of OFFSET ends up being out of range. */
2520 if (!dynamic
2521 && (ctx.dynamic
2522 || (!otr_type
2523 && (!TYPE_SIZE (ctx.outer_type)
2524 || !TYPE_SIZE (outer_type)
2525 || !operand_equal_p (TYPE_SIZE (ctx.outer_type),
2526 TYPE_SIZE (outer_type), 0)))))
2528 dynamic = true;
2529 updated = true;
2532 else if (contains_type_p (outer_type, offset - ctx.offset,
2533 ctx.outer_type, false, false))
2535 if (dump_file && (dump_flags & TDF_DETAILS))
2536 fprintf (dump_file, "First type contain the second as a field\n");
2538 if (!dynamic
2539 && (ctx.dynamic
2540 || (!otr_type
2541 && (!TYPE_SIZE (ctx.outer_type)
2542 || !TYPE_SIZE (outer_type)
2543 || !operand_equal_p (TYPE_SIZE (ctx.outer_type),
2544 TYPE_SIZE (outer_type), 0)))))
2545 dynamic = true;
2546 outer_type = ctx.outer_type;
2547 offset = ctx.offset;
2548 dynamic = ctx.dynamic;
2549 maybe_in_construction = ctx.maybe_in_construction;
2550 maybe_derived_type = ctx.maybe_derived_type;
2551 updated = true;
2553 /* See if OUTER_TYPE is base of CTX.OUTER_TYPE. */
2554 else if (contains_type_p (ctx.outer_type,
2555 ctx.offset - offset, outer_type, false, true))
2557 if (dump_file && (dump_flags & TDF_DETAILS))
2558 fprintf (dump_file, "First type is base of second\n");
2559 if (!maybe_derived_type)
2561 maybe_derived_type = true;
2562 updated = true;
2564 if (!maybe_in_construction && ctx.maybe_in_construction)
2566 maybe_in_construction = true;
2567 updated = true;
2569 if (!dynamic && ctx.dynamic)
2571 dynamic = true;
2572 updated = true;
2575 /* See if CTX.OUTER_TYPE is base of OUTER_TYPE. */
2576 else if (contains_type_p (outer_type,
2577 offset - ctx.offset, ctx.outer_type, false, true))
2579 if (dump_file && (dump_flags & TDF_DETAILS))
2580 fprintf (dump_file, "Second type is base of first\n");
2581 outer_type = ctx.outer_type;
2582 offset = ctx.offset;
2583 updated = true;
2584 if (!maybe_derived_type)
2585 maybe_derived_type = true;
2586 if (!maybe_in_construction && ctx.maybe_in_construction)
2587 maybe_in_construction = true;
2588 if (!dynamic && ctx.dynamic)
2589 dynamic = true;
2591 /* TODO handle merging using hiearchy. */
2592 else
2594 if (dump_file && (dump_flags & TDF_DETAILS))
2595 fprintf (dump_file, "Giving up on meet\n");
2596 clear_outer_type ();
2597 updated = true;
2600 updated |= meet_speculation_with (ctx.speculative_outer_type,
2601 ctx.speculative_offset,
2602 ctx.speculative_maybe_derived_type,
2603 otr_type);
2605 if (updated && dump_file && (dump_flags & TDF_DETAILS))
2607 fprintf (dump_file, "Updated as: ");
2608 dump (dump_file);
2609 fprintf (dump_file, "\n");
2611 return updated;