1 /* Write and read the cgraph to the memory mapped representation of a
4 Copyright (C) 2009-2014 Free Software Foundation, Inc.
5 Contributed by Kenneth Zadeck <zadeck@naturalbridge.com>
7 This file is part of GCC.
9 GCC is free software; you can redistribute it and/or modify it under
10 the terms of the GNU General Public License as published by the Free
11 Software Foundation; either version 3, or (at your option) any later
14 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
15 WARRANTY; without even the implied warranty of MERCHANTABILITY or
16 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
19 You should have received a copy of the GNU General Public License
20 along with GCC; see the file COPYING3. If not see
21 <http://www.gnu.org/licenses/>. */
25 #include "coretypes.h"
28 #include "stringpool.h"
29 #include "basic-block.h"
30 #include "tree-ssa-alias.h"
31 #include "internal-fn.h"
32 #include "gimple-expr.h"
41 #include "langhooks.h"
45 #include "hard-reg-set.h"
47 #include "diagnostic-core.h"
50 #include "lto-streamer.h"
51 #include "data-streamer.h"
52 #include "tree-streamer.h"
54 #include "tree-pass.h"
57 #include "pass_manager.h"
58 #include "ipa-utils.h"
60 /* True when asm nodes has been output. */
61 bool asm_nodes_output
= false;
63 static void output_cgraph_opt_summary (void);
64 static void input_cgraph_opt_summary (vec
<symtab_node
*> nodes
);
66 /* Number of LDPR values known to GCC. */
67 #define LDPR_NUM_KNOWN (LDPR_PREVAILING_DEF_IRONLY_EXP + 1)
69 /* All node orders are ofsetted by ORDER_BASE. */
70 static int order_base
;
72 /* Cgraph streaming is organized as set of record whose type
73 is indicated by a tag. */
76 /* Must leave 0 for the stopper. */
78 /* Cgraph node without body available. */
79 LTO_symtab_unavail_node
= 1,
80 /* Cgraph node with function body. */
81 LTO_symtab_analyzed_node
,
84 LTO_symtab_indirect_edge
,
89 /* Create a new symtab encoder.
90 if FOR_INPUT, the encoder allocate only datastructures needed
91 to read the symtab. */
94 lto_symtab_encoder_new (bool for_input
)
96 lto_symtab_encoder_t encoder
= XCNEW (struct lto_symtab_encoder_d
);
99 encoder
->map
= new hash_map
<symtab_node
*, size_t>;
100 encoder
->nodes
.create (0);
105 /* Delete ENCODER and its components. */
108 lto_symtab_encoder_delete (lto_symtab_encoder_t encoder
)
110 encoder
->nodes
.release ();
117 /* Return the existing reference number of NODE in the symtab encoder in
118 output block OB. Assign a new reference if this is the first time
122 lto_symtab_encoder_encode (lto_symtab_encoder_t encoder
,
129 lto_encoder_entry entry
= {node
, false, false, false};
131 ref
= encoder
->nodes
.length ();
132 encoder
->nodes
.safe_push (entry
);
136 size_t *slot
= encoder
->map
->get (node
);
139 lto_encoder_entry entry
= {node
, false, false, false};
140 ref
= encoder
->nodes
.length ();
142 encoder
->map
->put (node
, ref
+ 1);
143 encoder
->nodes
.safe_push (entry
);
151 /* Remove NODE from encoder. */
154 lto_symtab_encoder_delete_node (lto_symtab_encoder_t encoder
,
158 lto_encoder_entry last_node
;
160 size_t *slot
= encoder
->map
->get (node
);
161 if (slot
== NULL
|| !*slot
)
165 gcc_checking_assert (encoder
->nodes
[index
].node
== node
);
167 /* Remove from vector. We do this by swapping node with the last element
169 last_node
= encoder
->nodes
.pop ();
170 if (last_node
.node
!= node
)
172 gcc_assert (encoder
->map
->put (last_node
.node
, index
+ 1));
174 /* Move the last element to the original spot of NODE. */
175 encoder
->nodes
[index
] = last_node
;
178 /* Remove element from hash table. */
179 encoder
->map
->remove (node
);
184 /* Return TRUE if we should encode initializer of NODE (if any). */
187 lto_symtab_encoder_encode_body_p (lto_symtab_encoder_t encoder
,
188 struct cgraph_node
*node
)
190 int index
= lto_symtab_encoder_lookup (encoder
, node
);
191 return encoder
->nodes
[index
].body
;
194 /* Return TRUE if we should encode body of NODE (if any). */
197 lto_set_symtab_encoder_encode_body (lto_symtab_encoder_t encoder
,
198 struct cgraph_node
*node
)
200 int index
= lto_symtab_encoder_encode (encoder
, node
);
201 gcc_checking_assert (encoder
->nodes
[index
].node
== node
);
202 encoder
->nodes
[index
].body
= true;
205 /* Return TRUE if we should encode initializer of NODE (if any). */
208 lto_symtab_encoder_encode_initializer_p (lto_symtab_encoder_t encoder
,
211 int index
= lto_symtab_encoder_lookup (encoder
, node
);
212 if (index
== LCC_NOT_FOUND
)
214 return encoder
->nodes
[index
].initializer
;
217 /* Return TRUE if we should encode initializer of NODE (if any). */
220 lto_set_symtab_encoder_encode_initializer (lto_symtab_encoder_t encoder
,
223 int index
= lto_symtab_encoder_lookup (encoder
, node
);
224 encoder
->nodes
[index
].initializer
= true;
227 /* Return TRUE if we should encode initializer of NODE (if any). */
230 lto_symtab_encoder_in_partition_p (lto_symtab_encoder_t encoder
,
233 int index
= lto_symtab_encoder_lookup (encoder
, node
);
234 if (index
== LCC_NOT_FOUND
)
236 return encoder
->nodes
[index
].in_partition
;
239 /* Return TRUE if we should encode body of NODE (if any). */
242 lto_set_symtab_encoder_in_partition (lto_symtab_encoder_t encoder
,
245 int index
= lto_symtab_encoder_encode (encoder
, node
);
246 encoder
->nodes
[index
].in_partition
= true;
249 /* Output the cgraph EDGE to OB using ENCODER. */
252 lto_output_edge (struct lto_simple_output_block
*ob
, struct cgraph_edge
*edge
,
253 lto_symtab_encoder_t encoder
)
259 if (edge
->indirect_unknown_callee
)
260 streamer_write_enum (ob
->main_stream
, LTO_symtab_tags
, LTO_symtab_last_tag
,
261 LTO_symtab_indirect_edge
);
263 streamer_write_enum (ob
->main_stream
, LTO_symtab_tags
, LTO_symtab_last_tag
,
266 ref
= lto_symtab_encoder_lookup (encoder
, edge
->caller
);
267 gcc_assert (ref
!= LCC_NOT_FOUND
);
268 streamer_write_hwi_stream (ob
->main_stream
, ref
);
270 if (!edge
->indirect_unknown_callee
)
272 ref
= lto_symtab_encoder_lookup (encoder
, edge
->callee
);
273 gcc_assert (ref
!= LCC_NOT_FOUND
);
274 streamer_write_hwi_stream (ob
->main_stream
, ref
);
277 streamer_write_gcov_count_stream (ob
->main_stream
, edge
->count
);
279 bp
= bitpack_create (ob
->main_stream
);
280 uid
= (!gimple_has_body_p (edge
->caller
->decl
)
281 ? edge
->lto_stmt_uid
: gimple_uid (edge
->call_stmt
) + 1);
282 bp_pack_enum (&bp
, cgraph_inline_failed_t
,
283 CIF_N_REASONS
, edge
->inline_failed
);
284 bp_pack_var_len_unsigned (&bp
, uid
);
285 bp_pack_var_len_unsigned (&bp
, edge
->frequency
);
286 bp_pack_value (&bp
, edge
->indirect_inlining_edge
, 1);
287 bp_pack_value (&bp
, edge
->speculative
, 1);
288 bp_pack_value (&bp
, edge
->call_stmt_cannot_inline_p
, 1);
289 bp_pack_value (&bp
, edge
->can_throw_external
, 1);
290 bp_pack_value (&bp
, edge
->in_polymorphic_cdtor
, 1);
291 if (edge
->indirect_unknown_callee
)
293 int flags
= edge
->indirect_info
->ecf_flags
;
294 bp_pack_value (&bp
, (flags
& ECF_CONST
) != 0, 1);
295 bp_pack_value (&bp
, (flags
& ECF_PURE
) != 0, 1);
296 bp_pack_value (&bp
, (flags
& ECF_NORETURN
) != 0, 1);
297 bp_pack_value (&bp
, (flags
& ECF_MALLOC
) != 0, 1);
298 bp_pack_value (&bp
, (flags
& ECF_NOTHROW
) != 0, 1);
299 bp_pack_value (&bp
, (flags
& ECF_RETURNS_TWICE
) != 0, 1);
300 /* Flags that should not appear on indirect calls. */
301 gcc_assert (!(flags
& (ECF_LOOPING_CONST_OR_PURE
307 streamer_write_bitpack (&bp
);
308 if (edge
->indirect_unknown_callee
)
310 streamer_write_hwi_stream (ob
->main_stream
,
311 edge
->indirect_info
->common_target_id
);
312 if (edge
->indirect_info
->common_target_id
)
313 streamer_write_hwi_stream
314 (ob
->main_stream
, edge
->indirect_info
->common_target_probability
);
318 /* Return if NODE contain references from other partitions. */
321 referenced_from_other_partition_p (symtab_node
*node
, lto_symtab_encoder_t encoder
)
324 struct ipa_ref
*ref
= NULL
;
326 for (i
= 0; node
->iterate_referring (i
, ref
); i
++)
328 if (ref
->referring
->in_other_partition
329 || !lto_symtab_encoder_in_partition_p (encoder
, ref
->referring
))
335 /* Return true when node is reachable from other partition. */
338 reachable_from_other_partition_p (struct cgraph_node
*node
, lto_symtab_encoder_t encoder
)
340 struct cgraph_edge
*e
;
341 if (!node
->definition
)
343 if (node
->global
.inlined_to
)
345 for (e
= node
->callers
; e
; e
= e
->next_caller
)
346 if (e
->caller
->in_other_partition
347 || !lto_symtab_encoder_in_partition_p (encoder
, e
->caller
))
352 /* Return if NODE contain references from other partitions. */
355 referenced_from_this_partition_p (symtab_node
*node
,
356 lto_symtab_encoder_t encoder
)
359 struct ipa_ref
*ref
= NULL
;
361 for (i
= 0; node
->iterate_referring (i
, ref
); i
++)
362 if (lto_symtab_encoder_in_partition_p (encoder
, ref
->referring
))
367 /* Return true when node is reachable from other partition. */
370 reachable_from_this_partition_p (struct cgraph_node
*node
, lto_symtab_encoder_t encoder
)
372 struct cgraph_edge
*e
;
373 for (e
= node
->callers
; e
; e
= e
->next_caller
)
374 if (lto_symtab_encoder_in_partition_p (encoder
, e
->caller
))
379 /* Output the cgraph NODE to OB. ENCODER is used to find the
380 reference number of NODE->inlined_to. SET is the set of nodes we
381 are writing to the current file. If NODE is not in SET, then NODE
382 is a boundary of a cgraph_node_set and we pretend NODE just has a
383 decl and no callees. WRITTEN_DECLS is the set of FUNCTION_DECLs
384 that have had their callgraph node written so far. This is used to
385 determine if NODE is a clone of a previously written node. */
388 lto_output_node (struct lto_simple_output_block
*ob
, struct cgraph_node
*node
,
389 lto_symtab_encoder_t encoder
)
395 bool in_other_partition
= false;
396 struct cgraph_node
*clone_of
, *ultimate_clone_of
;
397 ipa_opt_pass_d
*pass
;
404 boundary_p
= !lto_symtab_encoder_in_partition_p (encoder
, node
);
406 if (node
->analyzed
&& !boundary_p
)
407 tag
= LTO_symtab_analyzed_node
;
409 tag
= LTO_symtab_unavail_node
;
411 streamer_write_enum (ob
->main_stream
, LTO_symtab_tags
, LTO_symtab_last_tag
,
413 streamer_write_hwi_stream (ob
->main_stream
, node
->order
);
415 /* In WPA mode, we only output part of the call-graph. Also, we
416 fake cgraph node attributes. There are two cases that we care.
418 Boundary nodes: There are nodes that are not part of SET but are
419 called from within SET. We artificially make them look like
420 externally visible nodes with no function body.
422 Cherry-picked nodes: These are nodes we pulled from other
423 translation units into SET during IPA-inlining. We make them as
424 local static nodes to prevent clashes with other local statics. */
425 if (boundary_p
&& node
->analyzed
426 && node
->get_partitioning_class () == SYMBOL_PARTITION
)
428 /* Inline clones can not be part of boundary.
429 gcc_assert (!node->global.inlined_to);
431 FIXME: At the moment they can be, when partition contains an inline
432 clone that is clone of inline clone from outside partition. We can
433 reshape the clone tree and make other tree to be the root, but it
434 needs a bit extra work and will be promplty done by cgraph_remove_node
435 after reading back. */
436 in_other_partition
= 1;
439 clone_of
= node
->clone_of
;
441 && (ref
= lto_symtab_encoder_lookup (encoder
, clone_of
)) == LCC_NOT_FOUND
)
442 if (clone_of
->prev_sibling_clone
)
443 clone_of
= clone_of
->prev_sibling_clone
;
445 clone_of
= clone_of
->clone_of
;
447 /* See if body of the master function is output. If not, we are seeing only
448 an declaration and we do not need to pass down clone tree. */
449 ultimate_clone_of
= clone_of
;
450 while (ultimate_clone_of
&& ultimate_clone_of
->clone_of
)
451 ultimate_clone_of
= ultimate_clone_of
->clone_of
;
453 if (clone_of
&& !lto_symtab_encoder_encode_body_p (encoder
, ultimate_clone_of
))
456 if (tag
== LTO_symtab_analyzed_node
)
457 gcc_assert (clone_of
|| !node
->clone_of
);
459 streamer_write_hwi_stream (ob
->main_stream
, LCC_NOT_FOUND
);
461 streamer_write_hwi_stream (ob
->main_stream
, ref
);
464 lto_output_fn_decl_index (ob
->decl_state
, ob
->main_stream
, node
->decl
);
465 streamer_write_gcov_count_stream (ob
->main_stream
, node
->count
);
466 streamer_write_hwi_stream (ob
->main_stream
, node
->count_materialization_scale
);
468 streamer_write_hwi_stream (ob
->main_stream
,
469 node
->ipa_transforms_to_apply
.length ());
470 FOR_EACH_VEC_ELT (node
->ipa_transforms_to_apply
, i
, pass
)
471 streamer_write_hwi_stream (ob
->main_stream
, pass
->static_pass_number
);
473 if (tag
== LTO_symtab_analyzed_node
)
475 if (node
->global
.inlined_to
)
477 ref
= lto_symtab_encoder_lookup (encoder
, node
->global
.inlined_to
);
478 gcc_assert (ref
!= LCC_NOT_FOUND
);
483 streamer_write_hwi_stream (ob
->main_stream
, ref
);
486 group
= node
->get_comdat_group ();
488 comdat
= IDENTIFIER_POINTER (group
);
491 streamer_write_data_stream (ob
->main_stream
, comdat
, strlen (comdat
) + 1);
495 if (node
->same_comdat_group
&& !boundary_p
)
497 ref
= lto_symtab_encoder_lookup (encoder
,
498 node
->same_comdat_group
);
499 gcc_assert (ref
!= LCC_NOT_FOUND
);
503 streamer_write_hwi_stream (ob
->main_stream
, ref
);
506 section
= node
->get_section ();
510 streamer_write_hwi_stream (ob
->main_stream
, node
->tp_first_run
);
512 bp
= bitpack_create (ob
->main_stream
);
513 bp_pack_value (&bp
, node
->local
.local
, 1);
514 bp_pack_value (&bp
, node
->externally_visible
, 1);
515 bp_pack_value (&bp
, node
->no_reorder
, 1);
516 bp_pack_value (&bp
, node
->definition
, 1);
517 bp_pack_value (&bp
, node
->local
.versionable
, 1);
518 bp_pack_value (&bp
, node
->local
.can_change_signature
, 1);
519 bp_pack_value (&bp
, node
->local
.redefined_extern_inline
, 1);
520 bp_pack_value (&bp
, node
->force_output
, 1);
521 bp_pack_value (&bp
, node
->forced_by_abi
, 1);
522 bp_pack_value (&bp
, node
->unique_name
, 1);
523 bp_pack_value (&bp
, node
->body_removed
, 1);
524 bp_pack_value (&bp
, node
->implicit_section
, 1);
525 bp_pack_value (&bp
, node
->address_taken
, 1);
526 bp_pack_value (&bp
, tag
== LTO_symtab_analyzed_node
527 && node
->get_partitioning_class () == SYMBOL_PARTITION
528 && (reachable_from_other_partition_p (node
, encoder
)
529 || referenced_from_other_partition_p (node
, encoder
)), 1);
530 bp_pack_value (&bp
, node
->lowered
, 1);
531 bp_pack_value (&bp
, in_other_partition
, 1);
532 /* Real aliases in a boundary become non-aliases. However we still stream
533 alias info on weakrefs.
534 TODO: We lose a bit of information here - when we know that variable is
535 defined in other unit, we may use the info on aliases to resolve
536 symbol1 != symbol2 type tests that we can do only for locally defined objects
538 alias_p
= node
->alias
&& (!boundary_p
|| node
->weakref
);
539 bp_pack_value (&bp
, alias_p
, 1);
540 bp_pack_value (&bp
, node
->weakref
, 1);
541 bp_pack_value (&bp
, node
->frequency
, 2);
542 bp_pack_value (&bp
, node
->only_called_at_startup
, 1);
543 bp_pack_value (&bp
, node
->only_called_at_exit
, 1);
544 bp_pack_value (&bp
, node
->tm_clone
, 1);
545 bp_pack_value (&bp
, node
->calls_comdat_local
, 1);
546 bp_pack_value (&bp
, node
->icf_merged
, 1);
547 bp_pack_value (&bp
, node
->thunk
.thunk_p
&& !boundary_p
, 1);
548 bp_pack_enum (&bp
, ld_plugin_symbol_resolution
,
549 LDPR_NUM_KNOWN
, node
->resolution
);
550 streamer_write_bitpack (&bp
);
551 streamer_write_data_stream (ob
->main_stream
, section
, strlen (section
) + 1);
553 if (node
->thunk
.thunk_p
&& !boundary_p
)
555 streamer_write_uhwi_stream
557 1 + (node
->thunk
.this_adjusting
!= 0) * 2
558 + (node
->thunk
.virtual_offset_p
!= 0) * 4);
559 streamer_write_uhwi_stream (ob
->main_stream
, node
->thunk
.fixed_offset
);
560 streamer_write_uhwi_stream (ob
->main_stream
, node
->thunk
.virtual_value
);
562 streamer_write_hwi_stream (ob
->main_stream
, node
->profile_id
);
563 if (DECL_STATIC_CONSTRUCTOR (node
->decl
))
564 streamer_write_hwi_stream (ob
->main_stream
, node
->get_init_priority ());
565 if (DECL_STATIC_DESTRUCTOR (node
->decl
))
566 streamer_write_hwi_stream (ob
->main_stream
, node
->get_fini_priority ());
569 /* Output the varpool NODE to OB.
570 If NODE is not in SET, then NODE is a boundary. */
573 lto_output_varpool_node (struct lto_simple_output_block
*ob
, varpool_node
*node
,
574 lto_symtab_encoder_t encoder
)
576 bool boundary_p
= !lto_symtab_encoder_in_partition_p (encoder
, node
);
584 streamer_write_enum (ob
->main_stream
, LTO_symtab_tags
, LTO_symtab_last_tag
,
585 LTO_symtab_variable
);
586 streamer_write_hwi_stream (ob
->main_stream
, node
->order
);
587 lto_output_var_decl_index (ob
->decl_state
, ob
->main_stream
, node
->decl
);
588 bp
= bitpack_create (ob
->main_stream
);
589 bp_pack_value (&bp
, node
->externally_visible
, 1);
590 bp_pack_value (&bp
, node
->no_reorder
, 1);
591 bp_pack_value (&bp
, node
->force_output
, 1);
592 bp_pack_value (&bp
, node
->forced_by_abi
, 1);
593 bp_pack_value (&bp
, node
->unique_name
, 1);
594 bp_pack_value (&bp
, node
->body_removed
, 1);
595 bp_pack_value (&bp
, node
->implicit_section
, 1);
596 bp_pack_value (&bp
, node
->writeonly
, 1);
597 bp_pack_value (&bp
, node
->definition
, 1);
598 alias_p
= node
->alias
&& (!boundary_p
|| node
->weakref
);
599 bp_pack_value (&bp
, alias_p
, 1);
600 bp_pack_value (&bp
, node
->weakref
, 1);
601 bp_pack_value (&bp
, node
->analyzed
&& !boundary_p
, 1);
602 gcc_assert (node
->definition
|| !node
->analyzed
);
603 /* Constant pool initializers can be de-unified into individual ltrans units.
604 FIXME: Alternatively at -Os we may want to avoid generating for them the local
605 labels and share them across LTRANS partitions. */
606 if (node
->get_partitioning_class () != SYMBOL_PARTITION
)
608 bp_pack_value (&bp
, 0, 1); /* used_from_other_parition. */
609 bp_pack_value (&bp
, 0, 1); /* in_other_partition. */
613 bp_pack_value (&bp
, node
->definition
614 && referenced_from_other_partition_p (node
, encoder
), 1);
615 bp_pack_value (&bp
, node
->analyzed
616 && boundary_p
&& !DECL_EXTERNAL (node
->decl
), 1);
617 /* in_other_partition. */
619 bp_pack_value (&bp
, node
->tls_model
, 3);
620 bp_pack_value (&bp
, node
->used_by_single_function
, 1);
621 streamer_write_bitpack (&bp
);
623 group
= node
->get_comdat_group ();
625 comdat
= IDENTIFIER_POINTER (group
);
628 streamer_write_data_stream (ob
->main_stream
, comdat
, strlen (comdat
) + 1);
632 if (node
->same_comdat_group
&& !boundary_p
)
634 ref
= lto_symtab_encoder_lookup (encoder
,
635 node
->same_comdat_group
);
636 gcc_assert (ref
!= LCC_NOT_FOUND
);
640 streamer_write_hwi_stream (ob
->main_stream
, ref
);
643 section
= node
->get_section ();
646 streamer_write_data_stream (ob
->main_stream
, section
, strlen (section
) + 1);
648 streamer_write_enum (ob
->main_stream
, ld_plugin_symbol_resolution
,
649 LDPR_NUM_KNOWN
, node
->resolution
);
652 /* Output the varpool NODE to OB.
653 If NODE is not in SET, then NODE is a boundary. */
656 lto_output_ref (struct lto_simple_output_block
*ob
, struct ipa_ref
*ref
,
657 lto_symtab_encoder_t encoder
)
661 int uid
= ref
->lto_stmt_uid
;
662 struct cgraph_node
*node
;
664 bp
= bitpack_create (ob
->main_stream
);
665 bp_pack_value (&bp
, ref
->use
, 2);
666 bp_pack_value (&bp
, ref
->speculative
, 1);
667 streamer_write_bitpack (&bp
);
668 nref
= lto_symtab_encoder_lookup (encoder
, ref
->referred
);
669 gcc_assert (nref
!= LCC_NOT_FOUND
);
670 streamer_write_hwi_stream (ob
->main_stream
, nref
);
672 node
= dyn_cast
<cgraph_node
*> (ref
->referring
);
676 uid
= gimple_uid (ref
->stmt
) + 1;
677 streamer_write_hwi_stream (ob
->main_stream
, uid
);
681 /* Stream out profile_summary to OB. */
684 output_profile_summary (struct lto_simple_output_block
*ob
)
691 /* We do not output num and run_max, they are not used by
692 GCC profile feedback and they are difficult to merge from multiple
694 gcc_assert (profile_info
->runs
);
695 streamer_write_uhwi_stream (ob
->main_stream
, profile_info
->runs
);
696 streamer_write_gcov_count_stream (ob
->main_stream
, profile_info
->sum_max
);
698 /* sum_all is needed for computing the working set with the
700 streamer_write_gcov_count_stream (ob
->main_stream
, profile_info
->sum_all
);
702 /* Create and output a bitpack of non-zero histogram entries indices. */
703 bp
= bitpack_create (ob
->main_stream
);
704 for (h_ix
= 0; h_ix
< GCOV_HISTOGRAM_SIZE
; h_ix
++)
705 bp_pack_value (&bp
, profile_info
->histogram
[h_ix
].num_counters
> 0, 1);
706 streamer_write_bitpack (&bp
);
707 /* Now stream out only those non-zero entries. */
708 for (h_ix
= 0; h_ix
< GCOV_HISTOGRAM_SIZE
; h_ix
++)
710 if (!profile_info
->histogram
[h_ix
].num_counters
)
712 streamer_write_gcov_count_stream (ob
->main_stream
,
713 profile_info
->histogram
[h_ix
].num_counters
);
714 streamer_write_gcov_count_stream (ob
->main_stream
,
715 profile_info
->histogram
[h_ix
].min_value
);
716 streamer_write_gcov_count_stream (ob
->main_stream
,
717 profile_info
->histogram
[h_ix
].cum_value
);
719 /* IPA-profile computes hot bb threshold based on cumulated
720 whole program profile. We need to stream it down to ltrans. */
722 streamer_write_gcov_count_stream (ob
->main_stream
,
723 get_hot_bb_threshold ());
726 streamer_write_uhwi_stream (ob
->main_stream
, 0);
729 /* Output all callees or indirect outgoing edges. EDGE must be the first such
733 output_outgoing_cgraph_edges (struct cgraph_edge
*edge
,
734 struct lto_simple_output_block
*ob
,
735 lto_symtab_encoder_t encoder
)
740 /* Output edges in backward direction, so the reconstructed callgraph match
741 and it is easy to associate call sites in the IPA pass summaries. */
742 while (edge
->next_callee
)
743 edge
= edge
->next_callee
;
744 for (; edge
; edge
= edge
->prev_callee
)
745 lto_output_edge (ob
, edge
, encoder
);
748 /* Output the part of the cgraph in SET. */
751 output_refs (lto_symtab_encoder_t encoder
)
753 lto_symtab_encoder_iterator lsei
;
754 struct lto_simple_output_block
*ob
;
759 ob
= lto_create_simple_output_block (LTO_section_refs
);
761 for (lsei
= lsei_start_in_partition (encoder
); !lsei_end_p (lsei
);
762 lsei_next_in_partition (&lsei
))
764 symtab_node
*node
= lsei_node (lsei
);
766 count
= node
->ref_list
.nreferences ();
769 streamer_write_gcov_count_stream (ob
->main_stream
, count
);
770 streamer_write_uhwi_stream (ob
->main_stream
,
771 lto_symtab_encoder_lookup (encoder
, node
));
772 for (i
= 0; node
->iterate_reference (i
, ref
); i
++)
773 lto_output_ref (ob
, ref
, encoder
);
777 streamer_write_uhwi_stream (ob
->main_stream
, 0);
779 lto_destroy_simple_output_block (ob
);
782 /* Add NODE into encoder as well as nodes it is cloned from.
783 Do it in a way so clones appear first. */
786 add_node_to (lto_symtab_encoder_t encoder
, struct cgraph_node
*node
,
790 add_node_to (encoder
, node
->clone_of
, include_body
);
791 else if (include_body
)
792 lto_set_symtab_encoder_encode_body (encoder
, node
);
793 lto_symtab_encoder_encode (encoder
, node
);
796 /* Add all references in NODE to encoders. */
799 create_references (lto_symtab_encoder_t encoder
, symtab_node
*node
)
802 struct ipa_ref
*ref
= NULL
;
803 for (i
= 0; node
->iterate_reference (i
, ref
); i
++)
804 if (is_a
<cgraph_node
*> (ref
->referred
))
805 add_node_to (encoder
, dyn_cast
<cgraph_node
*> (ref
->referred
), false);
807 lto_symtab_encoder_encode (encoder
, ref
->referred
);
810 /* Find all symbols we want to stream into given partition and insert them
813 The function actually replaces IN_ENCODER by new one. The reason is that
814 streaming code needs clone's origin to be streamed before clone. This
815 means that we need to insert the nodes in specific order. This order is
816 ignored by the partitioning logic earlier. */
819 compute_ltrans_boundary (lto_symtab_encoder_t in_encoder
)
821 struct cgraph_edge
*edge
;
823 lto_symtab_encoder_t encoder
;
824 lto_symtab_encoder_iterator lsei
;
825 hash_set
<void *> reachable_call_targets
;
827 encoder
= lto_symtab_encoder_new (false);
829 /* Go over all entries in the IN_ENCODER and duplicate them to
830 ENCODER. At the same time insert masters of clones so
831 every master appears before clone. */
832 for (lsei
= lsei_start_function_in_partition (in_encoder
);
833 !lsei_end_p (lsei
); lsei_next_function_in_partition (&lsei
))
835 struct cgraph_node
*node
= lsei_cgraph_node (lsei
);
836 add_node_to (encoder
, node
, true);
837 lto_set_symtab_encoder_in_partition (encoder
, node
);
838 create_references (encoder
, node
);
839 /* For proper debug info, we need to ship the origins, too. */
840 if (DECL_ABSTRACT_ORIGIN (node
->decl
))
842 struct cgraph_node
*origin_node
843 = cgraph_node::get (DECL_ABSTRACT_ORIGIN (node
->decl
));
844 add_node_to (encoder
, origin_node
, true);
847 for (lsei
= lsei_start_variable_in_partition (in_encoder
);
848 !lsei_end_p (lsei
); lsei_next_variable_in_partition (&lsei
))
850 varpool_node
*vnode
= lsei_varpool_node (lsei
);
852 lto_set_symtab_encoder_in_partition (encoder
, vnode
);
853 lto_set_symtab_encoder_encode_initializer (encoder
, vnode
);
854 create_references (encoder
, vnode
);
855 /* For proper debug info, we need to ship the origins, too. */
856 if (DECL_ABSTRACT_ORIGIN (vnode
->decl
))
858 varpool_node
*origin_node
859 = varpool_node::get (DECL_ABSTRACT_ORIGIN (vnode
->decl
));
860 lto_set_symtab_encoder_in_partition (encoder
, origin_node
);
863 /* Pickle in also the initializer of all referenced readonly variables
864 to help folding. Constant pool variables are not shared, so we must
866 for (i
= 0; i
< lto_symtab_encoder_size (encoder
); i
++)
868 symtab_node
*node
= lto_symtab_encoder_deref (encoder
, i
);
869 if (varpool_node
*vnode
= dyn_cast
<varpool_node
*> (node
))
871 if (!lto_symtab_encoder_encode_initializer_p (encoder
,
873 && vnode
->ctor_useable_for_folding_p ())
875 lto_set_symtab_encoder_encode_initializer (encoder
, vnode
);
876 create_references (encoder
, vnode
);
881 /* Go over all the nodes again to include callees that are not in
883 for (lsei
= lsei_start_function_in_partition (encoder
);
884 !lsei_end_p (lsei
); lsei_next_function_in_partition (&lsei
))
886 struct cgraph_node
*node
= lsei_cgraph_node (lsei
);
887 for (edge
= node
->callees
; edge
; edge
= edge
->next_callee
)
889 struct cgraph_node
*callee
= edge
->callee
;
890 if (!lto_symtab_encoder_in_partition_p (encoder
, callee
))
892 /* We should have moved all the inlines. */
893 gcc_assert (!callee
->global
.inlined_to
);
894 add_node_to (encoder
, callee
, false);
897 /* Add all possible targets for late devirtualization. */
898 if (flag_devirtualize
)
899 for (edge
= node
->indirect_calls
; edge
; edge
= edge
->next_callee
)
900 if (edge
->indirect_info
->polymorphic
)
905 vec
<cgraph_node
*>targets
906 = possible_polymorphic_call_targets
907 (edge
, &final
, &cache_token
);
908 if (!reachable_call_targets
.add (cache_token
))
910 for (i
= 0; i
< targets
.length (); i
++)
912 struct cgraph_node
*callee
= targets
[i
];
914 /* Adding an external declarations into the unit serves
915 no purpose and just increases its boundary. */
916 if (callee
->definition
917 && !lto_symtab_encoder_in_partition_p
920 gcc_assert (!callee
->global
.inlined_to
);
921 add_node_to (encoder
, callee
, false);
927 lto_symtab_encoder_delete (in_encoder
);
931 /* Output the part of the symtab in SET and VSET. */
936 struct cgraph_node
*node
;
937 struct lto_simple_output_block
*ob
;
938 lto_symtab_encoder_iterator lsei
;
940 lto_symtab_encoder_t encoder
;
943 output_cgraph_opt_summary ();
945 ob
= lto_create_simple_output_block (LTO_section_symtab_nodes
);
947 output_profile_summary (ob
);
949 /* An encoder for cgraph nodes should have been created by
950 ipa_write_summaries_1. */
951 gcc_assert (ob
->decl_state
->symtab_node_encoder
);
952 encoder
= ob
->decl_state
->symtab_node_encoder
;
954 /* Write out the nodes. We must first output a node and then its clones,
955 otherwise at a time reading back the node there would be nothing to clone
957 n_nodes
= lto_symtab_encoder_size (encoder
);
958 for (i
= 0; i
< n_nodes
; i
++)
960 symtab_node
*node
= lto_symtab_encoder_deref (encoder
, i
);
961 if (cgraph_node
*cnode
= dyn_cast
<cgraph_node
*> (node
))
962 lto_output_node (ob
, cnode
, encoder
);
964 lto_output_varpool_node (ob
, dyn_cast
<varpool_node
*> (node
), encoder
);
967 /* Go over the nodes in SET again to write edges. */
968 for (lsei
= lsei_start_function_in_partition (encoder
); !lsei_end_p (lsei
);
969 lsei_next_function_in_partition (&lsei
))
971 node
= lsei_cgraph_node (lsei
);
972 output_outgoing_cgraph_edges (node
->callees
, ob
, encoder
);
973 output_outgoing_cgraph_edges (node
->indirect_calls
, ob
, encoder
);
976 streamer_write_uhwi_stream (ob
->main_stream
, 0);
978 lto_destroy_simple_output_block (ob
);
980 /* Emit toplevel asms.
981 When doing WPA we must output every asm just once. Since we do not partition asm
982 nodes at all, output them to first output. This is kind of hack, but should work
984 if (!asm_nodes_output
)
986 asm_nodes_output
= true;
987 lto_output_toplevel_asms ();
990 output_refs (encoder
);
993 /* Return identifier encoded in IB as a plain string. */
996 read_identifier (struct lto_input_block
*ib
)
998 unsigned int len
= strnlen (ib
->data
+ ib
->p
, ib
->len
- ib
->p
- 1);
1001 if (ib
->data
[ib
->p
+ len
])
1002 lto_section_overrun (ib
);
1008 id
= get_identifier (ib
->data
+ ib
->p
);
1013 /* Return string encoded in IB, NULL if string is empty. */
1016 read_string (struct lto_input_block
*ib
)
1018 unsigned int len
= strnlen (ib
->data
+ ib
->p
, ib
->len
- ib
->p
- 1);
1021 if (ib
->data
[ib
->p
+ len
])
1022 lto_section_overrun (ib
);
1028 str
= ib
->data
+ ib
->p
;
1033 /* Overwrite the information in NODE based on FILE_DATA, TAG, FLAGS,
1034 STACK_SIZE, SELF_TIME and SELF_SIZE. This is called either to initialize
1035 NODE or to replace the values in it, for instance because the first
1036 time we saw it, the function body was not available but now it
1037 is. BP is a bitpack with all the bitflags for NODE read from the
1041 input_overwrite_node (struct lto_file_decl_data
*file_data
,
1042 struct cgraph_node
*node
,
1043 enum LTO_symtab_tags tag
,
1044 struct bitpack_d
*bp
)
1046 node
->aux
= (void *) tag
;
1047 node
->lto_file_data
= file_data
;
1049 node
->local
.local
= bp_unpack_value (bp
, 1);
1050 node
->externally_visible
= bp_unpack_value (bp
, 1);
1051 node
->no_reorder
= bp_unpack_value (bp
, 1);
1052 node
->definition
= bp_unpack_value (bp
, 1);
1053 node
->local
.versionable
= bp_unpack_value (bp
, 1);
1054 node
->local
.can_change_signature
= bp_unpack_value (bp
, 1);
1055 node
->local
.redefined_extern_inline
= bp_unpack_value (bp
, 1);
1056 node
->force_output
= bp_unpack_value (bp
, 1);
1057 node
->forced_by_abi
= bp_unpack_value (bp
, 1);
1058 node
->unique_name
= bp_unpack_value (bp
, 1);
1059 node
->body_removed
= bp_unpack_value (bp
, 1);
1060 node
->implicit_section
= bp_unpack_value (bp
, 1);
1061 node
->address_taken
= bp_unpack_value (bp
, 1);
1062 node
->used_from_other_partition
= bp_unpack_value (bp
, 1);
1063 node
->lowered
= bp_unpack_value (bp
, 1);
1064 node
->analyzed
= tag
== LTO_symtab_analyzed_node
;
1065 node
->in_other_partition
= bp_unpack_value (bp
, 1);
1066 if (node
->in_other_partition
1067 /* Avoid updating decl when we are seeing just inline clone.
1068 When inlining function that has functions already inlined into it,
1069 we produce clones of inline clones.
1071 WPA partitioning might put each clone into different unit and
1072 we might end up streaming inline clone from other partition
1073 to support clone we are interested in. */
1075 || node
->clone_of
->decl
!= node
->decl
))
1077 DECL_EXTERNAL (node
->decl
) = 1;
1078 TREE_STATIC (node
->decl
) = 0;
1080 node
->alias
= bp_unpack_value (bp
, 1);
1081 node
->weakref
= bp_unpack_value (bp
, 1);
1082 node
->frequency
= (enum node_frequency
)bp_unpack_value (bp
, 2);
1083 node
->only_called_at_startup
= bp_unpack_value (bp
, 1);
1084 node
->only_called_at_exit
= bp_unpack_value (bp
, 1);
1085 node
->tm_clone
= bp_unpack_value (bp
, 1);
1086 node
->calls_comdat_local
= bp_unpack_value (bp
, 1);
1087 node
->icf_merged
= bp_unpack_value (bp
, 1);
1088 node
->thunk
.thunk_p
= bp_unpack_value (bp
, 1);
1089 node
->resolution
= bp_unpack_enum (bp
, ld_plugin_symbol_resolution
,
1091 gcc_assert (flag_ltrans
1092 || (!node
->in_other_partition
1093 && !node
->used_from_other_partition
));
1096 /* Return string alias is alias of. */
1099 get_alias_symbol (tree decl
)
1101 tree alias
= lookup_attribute ("alias", DECL_ATTRIBUTES (decl
));
1102 return get_identifier (TREE_STRING_POINTER
1103 (TREE_VALUE (TREE_VALUE (alias
))));
1106 /* Read a node from input_block IB. TAG is the node's tag just read.
1107 Return the node read or overwriten. */
1109 static struct cgraph_node
*
1110 input_node (struct lto_file_decl_data
*file_data
,
1111 struct lto_input_block
*ib
,
1112 enum LTO_symtab_tags tag
,
1113 vec
<symtab_node
*> nodes
)
1115 gcc::pass_manager
*passes
= g
->get_passes ();
1117 struct cgraph_node
*node
;
1118 struct bitpack_d bp
;
1119 unsigned decl_index
;
1120 int ref
= LCC_NOT_FOUND
, ref2
= LCC_NOT_FOUND
;
1125 const char *section
;
1126 order
= streamer_read_hwi (ib
) + order_base
;
1127 clone_ref
= streamer_read_hwi (ib
);
1129 decl_index
= streamer_read_uhwi (ib
);
1130 fn_decl
= lto_file_decl_data_get_fn_decl (file_data
, decl_index
);
1132 if (clone_ref
!= LCC_NOT_FOUND
)
1134 node
= dyn_cast
<cgraph_node
*> (nodes
[clone_ref
])->create_clone (fn_decl
,
1135 0, CGRAPH_FREQ_BASE
, false,
1136 vNULL
, false, NULL
, NULL
);
1140 /* Declaration of functions can be already merged with a declaration
1141 from other input file. We keep cgraph unmerged until after streaming
1142 of ipa passes is done. Alays forcingly create a fresh node. */
1143 node
= symtab
->create_empty ();
1144 node
->decl
= fn_decl
;
1145 node
->register_symbol ();
1148 node
->order
= order
;
1149 if (order
>= symtab
->order
)
1150 symtab
->order
= order
+ 1;
1152 node
->count
= streamer_read_gcov_count (ib
);
1153 node
->count_materialization_scale
= streamer_read_hwi (ib
);
1155 count
= streamer_read_hwi (ib
);
1156 node
->ipa_transforms_to_apply
= vNULL
;
1157 for (i
= 0; i
< count
; i
++)
1160 int pid
= streamer_read_hwi (ib
);
1162 gcc_assert (pid
< passes
->passes_by_id_size
);
1163 pass
= passes
->passes_by_id
[pid
];
1164 node
->ipa_transforms_to_apply
.safe_push ((ipa_opt_pass_d
*) pass
);
1167 if (tag
== LTO_symtab_analyzed_node
)
1168 ref
= streamer_read_hwi (ib
);
1170 group
= read_identifier (ib
);
1172 ref2
= streamer_read_hwi (ib
);
1174 /* Make sure that we have not read this node before. Nodes that
1175 have already been read will have their tag stored in the 'aux'
1176 field. Since built-in functions can be referenced in multiple
1177 functions, they are expected to be read more than once. */
1178 if (node
->aux
&& !DECL_BUILT_IN (node
->decl
))
1179 internal_error ("bytecode stream: found multiple instances of cgraph "
1180 "node with uid %d", node
->uid
);
1182 node
->tp_first_run
= streamer_read_uhwi (ib
);
1184 bp
= streamer_read_bitpack (ib
);
1186 input_overwrite_node (file_data
, node
, tag
, &bp
);
1188 /* Store a reference for now, and fix up later to be a pointer. */
1189 node
->global
.inlined_to
= (cgraph_node
*) (intptr_t) ref
;
1193 node
->set_comdat_group (group
);
1194 /* Store a reference for now, and fix up later to be a pointer. */
1195 node
->same_comdat_group
= (symtab_node
*) (intptr_t) ref2
;
1198 node
->same_comdat_group
= (symtab_node
*) (intptr_t) LCC_NOT_FOUND
;
1199 section
= read_string (ib
);
1201 node
->set_section_for_node (section
);
1203 if (node
->thunk
.thunk_p
)
1205 int type
= streamer_read_uhwi (ib
);
1206 HOST_WIDE_INT fixed_offset
= streamer_read_uhwi (ib
);
1207 HOST_WIDE_INT virtual_value
= streamer_read_uhwi (ib
);
1209 node
->thunk
.fixed_offset
= fixed_offset
;
1210 node
->thunk
.this_adjusting
= (type
& 2);
1211 node
->thunk
.virtual_value
= virtual_value
;
1212 node
->thunk
.virtual_offset_p
= (type
& 4);
1214 if (node
->alias
&& !node
->analyzed
&& node
->weakref
)
1215 node
->alias_target
= get_alias_symbol (node
->decl
);
1216 node
->profile_id
= streamer_read_hwi (ib
);
1217 if (DECL_STATIC_CONSTRUCTOR (node
->decl
))
1218 node
->set_init_priority (streamer_read_hwi (ib
));
1219 if (DECL_STATIC_DESTRUCTOR (node
->decl
))
1220 node
->set_fini_priority (streamer_read_hwi (ib
));
1224 /* Read a node from input_block IB. TAG is the node's tag just read.
1225 Return the node read or overwriten. */
1227 static varpool_node
*
1228 input_varpool_node (struct lto_file_decl_data
*file_data
,
1229 struct lto_input_block
*ib
)
1234 struct bitpack_d bp
;
1235 int ref
= LCC_NOT_FOUND
;
1238 const char *section
;
1240 order
= streamer_read_hwi (ib
) + order_base
;
1241 decl_index
= streamer_read_uhwi (ib
);
1242 var_decl
= lto_file_decl_data_get_var_decl (file_data
, decl_index
);
1244 /* Declaration of functions can be already merged with a declaration
1245 from other input file. We keep cgraph unmerged until after streaming
1246 of ipa passes is done. Alays forcingly create a fresh node. */
1247 node
= varpool_node::create_empty ();
1248 node
->decl
= var_decl
;
1249 node
->register_symbol ();
1251 node
->order
= order
;
1252 if (order
>= symtab
->order
)
1253 symtab
->order
= order
+ 1;
1254 node
->lto_file_data
= file_data
;
1256 bp
= streamer_read_bitpack (ib
);
1257 node
->externally_visible
= bp_unpack_value (&bp
, 1);
1258 node
->no_reorder
= bp_unpack_value (&bp
, 1);
1259 node
->force_output
= bp_unpack_value (&bp
, 1);
1260 node
->forced_by_abi
= bp_unpack_value (&bp
, 1);
1261 node
->unique_name
= bp_unpack_value (&bp
, 1);
1262 node
->body_removed
= bp_unpack_value (&bp
, 1);
1263 node
->implicit_section
= bp_unpack_value (&bp
, 1);
1264 node
->writeonly
= bp_unpack_value (&bp
, 1);
1265 node
->definition
= bp_unpack_value (&bp
, 1);
1266 node
->alias
= bp_unpack_value (&bp
, 1);
1267 node
->weakref
= bp_unpack_value (&bp
, 1);
1268 node
->analyzed
= bp_unpack_value (&bp
, 1);
1269 node
->used_from_other_partition
= bp_unpack_value (&bp
, 1);
1270 node
->in_other_partition
= bp_unpack_value (&bp
, 1);
1271 if (node
->in_other_partition
)
1273 DECL_EXTERNAL (node
->decl
) = 1;
1274 TREE_STATIC (node
->decl
) = 0;
1276 if (node
->alias
&& !node
->analyzed
&& node
->weakref
)
1277 node
->alias_target
= get_alias_symbol (node
->decl
);
1278 node
->tls_model
= (enum tls_model
)bp_unpack_value (&bp
, 3);
1279 node
->used_by_single_function
= (enum tls_model
)bp_unpack_value (&bp
, 1);
1280 group
= read_identifier (ib
);
1283 node
->set_comdat_group (group
);
1284 ref
= streamer_read_hwi (ib
);
1285 /* Store a reference for now, and fix up later to be a pointer. */
1286 node
->same_comdat_group
= (symtab_node
*) (intptr_t) ref
;
1289 node
->same_comdat_group
= (symtab_node
*) (intptr_t) LCC_NOT_FOUND
;
1290 section
= read_string (ib
);
1292 node
->set_section_for_node (section
);
1293 node
->resolution
= streamer_read_enum (ib
, ld_plugin_symbol_resolution
,
1295 gcc_assert (flag_ltrans
1296 || (!node
->in_other_partition
1297 && !node
->used_from_other_partition
));
1302 /* Read a node from input_block IB. TAG is the node's tag just read.
1303 Return the node read or overwriten. */
1306 input_ref (struct lto_input_block
*ib
,
1307 symtab_node
*referring_node
,
1308 vec
<symtab_node
*> nodes
)
1310 symtab_node
*node
= NULL
;
1311 struct bitpack_d bp
;
1312 enum ipa_ref_use use
;
1314 struct ipa_ref
*ref
;
1316 bp
= streamer_read_bitpack (ib
);
1317 use
= (enum ipa_ref_use
) bp_unpack_value (&bp
, 2);
1318 speculative
= (enum ipa_ref_use
) bp_unpack_value (&bp
, 1);
1319 node
= nodes
[streamer_read_hwi (ib
)];
1320 ref
= referring_node
->create_reference (node
, use
);
1321 ref
->speculative
= speculative
;
1322 if (is_a
<cgraph_node
*> (referring_node
))
1323 ref
->lto_stmt_uid
= streamer_read_hwi (ib
);
1326 /* Read an edge from IB. NODES points to a vector of previously read nodes for
1327 decoding caller and callee of the edge to be read. If INDIRECT is true, the
1328 edge being read is indirect (in the sense that it has
1329 indirect_unknown_callee set). */
1332 input_edge (struct lto_input_block
*ib
, vec
<symtab_node
*> nodes
,
1335 struct cgraph_node
*caller
, *callee
;
1336 struct cgraph_edge
*edge
;
1337 unsigned int stmt_id
;
1340 cgraph_inline_failed_t inline_failed
;
1341 struct bitpack_d bp
;
1344 caller
= dyn_cast
<cgraph_node
*> (nodes
[streamer_read_hwi (ib
)]);
1345 if (caller
== NULL
|| caller
->decl
== NULL_TREE
)
1346 internal_error ("bytecode stream: no caller found while reading edge");
1350 callee
= dyn_cast
<cgraph_node
*> (nodes
[streamer_read_hwi (ib
)]);
1351 if (callee
== NULL
|| callee
->decl
== NULL_TREE
)
1352 internal_error ("bytecode stream: no callee found while reading edge");
1357 count
= streamer_read_gcov_count (ib
);
1359 bp
= streamer_read_bitpack (ib
);
1360 inline_failed
= bp_unpack_enum (&bp
, cgraph_inline_failed_t
, CIF_N_REASONS
);
1361 stmt_id
= bp_unpack_var_len_unsigned (&bp
);
1362 freq
= (int) bp_unpack_var_len_unsigned (&bp
);
1365 edge
= caller
->create_indirect_edge (NULL
, 0, count
, freq
);
1367 edge
= caller
->create_edge (callee
, NULL
, count
, freq
);
1369 edge
->indirect_inlining_edge
= bp_unpack_value (&bp
, 1);
1370 edge
->speculative
= bp_unpack_value (&bp
, 1);
1371 edge
->lto_stmt_uid
= stmt_id
;
1372 edge
->inline_failed
= inline_failed
;
1373 edge
->call_stmt_cannot_inline_p
= bp_unpack_value (&bp
, 1);
1374 edge
->can_throw_external
= bp_unpack_value (&bp
, 1);
1375 edge
->in_polymorphic_cdtor
= bp_unpack_value (&bp
, 1);
1378 if (bp_unpack_value (&bp
, 1))
1379 ecf_flags
|= ECF_CONST
;
1380 if (bp_unpack_value (&bp
, 1))
1381 ecf_flags
|= ECF_PURE
;
1382 if (bp_unpack_value (&bp
, 1))
1383 ecf_flags
|= ECF_NORETURN
;
1384 if (bp_unpack_value (&bp
, 1))
1385 ecf_flags
|= ECF_MALLOC
;
1386 if (bp_unpack_value (&bp
, 1))
1387 ecf_flags
|= ECF_NOTHROW
;
1388 if (bp_unpack_value (&bp
, 1))
1389 ecf_flags
|= ECF_RETURNS_TWICE
;
1390 edge
->indirect_info
->ecf_flags
= ecf_flags
;
1391 edge
->indirect_info
->common_target_id
= streamer_read_hwi (ib
);
1392 if (edge
->indirect_info
->common_target_id
)
1393 edge
->indirect_info
->common_target_probability
= streamer_read_hwi (ib
);
1398 /* Read a cgraph from IB using the info in FILE_DATA. */
1400 static vec
<symtab_node
*>
1401 input_cgraph_1 (struct lto_file_decl_data
*file_data
,
1402 struct lto_input_block
*ib
)
1404 enum LTO_symtab_tags tag
;
1405 vec
<symtab_node
*> nodes
= vNULL
;
1409 tag
= streamer_read_enum (ib
, LTO_symtab_tags
, LTO_symtab_last_tag
);
1410 order_base
= symtab
->order
;
1413 if (tag
== LTO_symtab_edge
)
1414 input_edge (ib
, nodes
, false);
1415 else if (tag
== LTO_symtab_indirect_edge
)
1416 input_edge (ib
, nodes
, true);
1417 else if (tag
== LTO_symtab_variable
)
1419 node
= input_varpool_node (file_data
, ib
);
1420 nodes
.safe_push (node
);
1421 lto_symtab_encoder_encode (file_data
->symtab_node_encoder
, node
);
1425 node
= input_node (file_data
, ib
, tag
, nodes
);
1426 if (node
== NULL
|| node
->decl
== NULL_TREE
)
1427 internal_error ("bytecode stream: found empty cgraph node");
1428 nodes
.safe_push (node
);
1429 lto_symtab_encoder_encode (file_data
->symtab_node_encoder
, node
);
1432 tag
= streamer_read_enum (ib
, LTO_symtab_tags
, LTO_symtab_last_tag
);
1435 lto_input_toplevel_asms (file_data
, order_base
);
1437 /* AUX pointers should be all non-zero for function nodes read from the stream. */
1438 #ifdef ENABLE_CHECKING
1439 FOR_EACH_VEC_ELT (nodes
, i
, node
)
1440 gcc_assert (node
->aux
|| !is_a
<cgraph_node
*> (node
));
1442 FOR_EACH_VEC_ELT (nodes
, i
, node
)
1445 if (cgraph_node
*cnode
= dyn_cast
<cgraph_node
*> (node
))
1447 ref
= (int) (intptr_t) cnode
->global
.inlined_to
;
1449 /* We share declaration of builtins, so we may read same node twice. */
1454 /* Fixup inlined_to from reference to pointer. */
1455 if (ref
!= LCC_NOT_FOUND
)
1456 dyn_cast
<cgraph_node
*> (node
)->global
.inlined_to
1457 = dyn_cast
<cgraph_node
*> (nodes
[ref
]);
1459 cnode
->global
.inlined_to
= NULL
;
1462 ref
= (int) (intptr_t) node
->same_comdat_group
;
1464 /* Fixup same_comdat_group from reference to pointer. */
1465 if (ref
!= LCC_NOT_FOUND
)
1466 node
->same_comdat_group
= nodes
[ref
];
1468 node
->same_comdat_group
= NULL
;
1470 FOR_EACH_VEC_ELT (nodes
, i
, node
)
1471 node
->aux
= is_a
<cgraph_node
*> (node
) ? (void *)1 : NULL
;
1475 /* Input ipa_refs. */
1478 input_refs (struct lto_input_block
*ib
,
1479 vec
<symtab_node
*> nodes
)
1486 count
= streamer_read_uhwi (ib
);
1489 idx
= streamer_read_uhwi (ib
);
1493 input_ref (ib
, node
, nodes
);
1500 static struct gcov_ctr_summary lto_gcov_summary
;
1502 /* Input profile_info from IB. */
1504 input_profile_summary (struct lto_input_block
*ib
,
1505 struct lto_file_decl_data
*file_data
)
1508 struct bitpack_d bp
;
1509 unsigned int runs
= streamer_read_uhwi (ib
);
1512 file_data
->profile_info
.runs
= runs
;
1513 file_data
->profile_info
.sum_max
= streamer_read_gcov_count (ib
);
1514 file_data
->profile_info
.sum_all
= streamer_read_gcov_count (ib
);
1516 memset (file_data
->profile_info
.histogram
, 0,
1517 sizeof (gcov_bucket_type
) * GCOV_HISTOGRAM_SIZE
);
1518 /* Input the bitpack of non-zero histogram indices. */
1519 bp
= streamer_read_bitpack (ib
);
1520 /* Read in and unpack the full bitpack, flagging non-zero
1521 histogram entries by setting the num_counters non-zero. */
1522 for (h_ix
= 0; h_ix
< GCOV_HISTOGRAM_SIZE
; h_ix
++)
1524 file_data
->profile_info
.histogram
[h_ix
].num_counters
1525 = bp_unpack_value (&bp
, 1);
1527 for (h_ix
= 0; h_ix
< GCOV_HISTOGRAM_SIZE
; h_ix
++)
1529 if (!file_data
->profile_info
.histogram
[h_ix
].num_counters
)
1532 file_data
->profile_info
.histogram
[h_ix
].num_counters
1533 = streamer_read_gcov_count (ib
);
1534 file_data
->profile_info
.histogram
[h_ix
].min_value
1535 = streamer_read_gcov_count (ib
);
1536 file_data
->profile_info
.histogram
[h_ix
].cum_value
1537 = streamer_read_gcov_count (ib
);
1539 /* IPA-profile computes hot bb threshold based on cumulated
1540 whole program profile. We need to stream it down to ltrans. */
1542 set_hot_bb_threshold (streamer_read_gcov_count (ib
));
1547 /* Rescale profile summaries to the same number of runs in the whole unit. */
1550 merge_profile_summaries (struct lto_file_decl_data
**file_data_vec
)
1552 struct lto_file_decl_data
*file_data
;
1553 unsigned int j
, h_ix
;
1554 gcov_unsigned_t max_runs
= 0;
1555 struct cgraph_node
*node
;
1556 struct cgraph_edge
*edge
;
1557 gcov_type saved_sum_all
= 0;
1558 gcov_ctr_summary
*saved_profile_info
= 0;
1559 int saved_scale
= 0;
1561 /* Find unit with maximal number of runs. If we ever get serious about
1562 roundoff errors, we might also consider computing smallest common
1564 for (j
= 0; (file_data
= file_data_vec
[j
]) != NULL
; j
++)
1565 if (max_runs
< file_data
->profile_info
.runs
)
1566 max_runs
= file_data
->profile_info
.runs
;
1571 /* Simple overflow check. We probably don't need to support that many train
1572 runs. Such a large value probably imply data corruption anyway. */
1573 if (max_runs
> INT_MAX
/ REG_BR_PROB_BASE
)
1575 sorry ("At most %i profile runs is supported. Perhaps corrupted profile?",
1576 INT_MAX
/ REG_BR_PROB_BASE
);
1580 profile_info
= <o_gcov_summary
;
1581 lto_gcov_summary
.runs
= max_runs
;
1582 lto_gcov_summary
.sum_max
= 0;
1583 memset (lto_gcov_summary
.histogram
, 0,
1584 sizeof (gcov_bucket_type
) * GCOV_HISTOGRAM_SIZE
);
1586 /* Rescale all units to the maximal number of runs.
1587 sum_max can not be easily merged, as we have no idea what files come from
1588 the same run. We do not use the info anyway, so leave it 0. */
1589 for (j
= 0; (file_data
= file_data_vec
[j
]) != NULL
; j
++)
1590 if (file_data
->profile_info
.runs
)
1592 int scale
= GCOV_COMPUTE_SCALE (max_runs
,
1593 file_data
->profile_info
.runs
);
1594 lto_gcov_summary
.sum_max
1595 = MAX (lto_gcov_summary
.sum_max
,
1596 apply_scale (file_data
->profile_info
.sum_max
, scale
));
1597 lto_gcov_summary
.sum_all
1598 = MAX (lto_gcov_summary
.sum_all
,
1599 apply_scale (file_data
->profile_info
.sum_all
, scale
));
1600 /* Save a pointer to the profile_info with the largest
1601 scaled sum_all and the scale for use in merging the
1603 if (!saved_profile_info
1604 || lto_gcov_summary
.sum_all
> saved_sum_all
)
1606 saved_profile_info
= &file_data
->profile_info
;
1607 saved_sum_all
= lto_gcov_summary
.sum_all
;
1608 saved_scale
= scale
;
1612 gcc_assert (saved_profile_info
);
1614 /* Scale up the histogram from the profile that had the largest
1615 scaled sum_all above. */
1616 for (h_ix
= 0; h_ix
< GCOV_HISTOGRAM_SIZE
; h_ix
++)
1618 /* Scale up the min value as we did the corresponding sum_all
1619 above. Use that to find the new histogram index. */
1620 gcov_type scaled_min
1621 = apply_scale (saved_profile_info
->histogram
[h_ix
].min_value
,
1623 /* The new index may be shared with another scaled histogram entry,
1624 so we need to account for a non-zero histogram entry at new_ix. */
1625 unsigned new_ix
= gcov_histo_index (scaled_min
);
1626 lto_gcov_summary
.histogram
[new_ix
].min_value
1627 = (lto_gcov_summary
.histogram
[new_ix
].num_counters
1628 ? MIN (lto_gcov_summary
.histogram
[new_ix
].min_value
, scaled_min
)
1630 /* Some of the scaled counter values would ostensibly need to be placed
1631 into different (larger) histogram buckets, but we keep things simple
1632 here and place the scaled cumulative counter value in the bucket
1633 corresponding to the scaled minimum counter value. */
1634 lto_gcov_summary
.histogram
[new_ix
].cum_value
1635 += apply_scale (saved_profile_info
->histogram
[h_ix
].cum_value
,
1637 lto_gcov_summary
.histogram
[new_ix
].num_counters
1638 += saved_profile_info
->histogram
[h_ix
].num_counters
;
1641 /* Watch roundoff errors. */
1642 if (lto_gcov_summary
.sum_max
< max_runs
)
1643 lto_gcov_summary
.sum_max
= max_runs
;
1645 /* If merging already happent at WPA time, we are done. */
1649 /* Now compute count_materialization_scale of each node.
1650 During LTRANS we already have values of count_materialization_scale
1651 computed, so just update them. */
1652 FOR_EACH_FUNCTION (node
)
1653 if (node
->lto_file_data
1654 && node
->lto_file_data
->profile_info
.runs
)
1658 scale
= RDIV (node
->count_materialization_scale
* max_runs
,
1659 node
->lto_file_data
->profile_info
.runs
);
1660 node
->count_materialization_scale
= scale
;
1662 fatal_error ("Profile information in %s corrupted",
1663 file_data
->file_name
);
1665 if (scale
== REG_BR_PROB_BASE
)
1667 for (edge
= node
->callees
; edge
; edge
= edge
->next_callee
)
1668 edge
->count
= apply_scale (edge
->count
, scale
);
1669 node
->count
= apply_scale (node
->count
, scale
);
1673 /* Input and merge the symtab from each of the .o files passed to
1679 struct lto_file_decl_data
**file_data_vec
= lto_get_file_decl_data ();
1680 struct lto_file_decl_data
*file_data
;
1682 struct cgraph_node
*node
;
1684 while ((file_data
= file_data_vec
[j
++]))
1688 struct lto_input_block
*ib
;
1689 vec
<symtab_node
*> nodes
;
1691 ib
= lto_create_simple_input_block (file_data
, LTO_section_symtab_nodes
,
1694 fatal_error ("cannot find LTO cgraph in %s", file_data
->file_name
);
1695 input_profile_summary (ib
, file_data
);
1696 file_data
->symtab_node_encoder
= lto_symtab_encoder_new (true);
1697 nodes
= input_cgraph_1 (file_data
, ib
);
1698 lto_destroy_simple_input_block (file_data
, LTO_section_symtab_nodes
,
1701 ib
= lto_create_simple_input_block (file_data
, LTO_section_refs
,
1704 fatal_error ("cannot find LTO section refs in %s",
1705 file_data
->file_name
);
1706 input_refs (ib
, nodes
);
1707 lto_destroy_simple_input_block (file_data
, LTO_section_refs
,
1710 input_cgraph_opt_summary (nodes
);
1714 merge_profile_summaries (file_data_vec
);
1715 get_working_sets ();
1718 /* Clear out the aux field that was used to store enough state to
1719 tell which nodes should be overwritten. */
1720 FOR_EACH_FUNCTION (node
)
1722 /* Some nodes may have been created by cgraph_node. This
1723 happens when the callgraph contains nested functions. If the
1724 node for the parent function was never emitted to the gimple
1725 file, cgraph_node will create a node for it when setting the
1726 context of the nested function. */
1727 if (node
->lto_file_data
)
1732 /* True when we need optimization summary for NODE. */
1735 output_cgraph_opt_summary_p (struct cgraph_node
*node
)
1737 return (node
->clone_of
1738 && (node
->clone
.tree_map
1739 || node
->clone
.args_to_skip
1740 || node
->clone
.combined_args_to_skip
));
1743 /* Output optimization summary for EDGE to OB. */
1745 output_edge_opt_summary (struct output_block
*ob ATTRIBUTE_UNUSED
,
1746 struct cgraph_edge
*edge ATTRIBUTE_UNUSED
)
1750 /* Output optimization summary for NODE to OB. */
1753 output_node_opt_summary (struct output_block
*ob
,
1754 struct cgraph_node
*node
,
1755 lto_symtab_encoder_t encoder
)
1759 struct ipa_replace_map
*map
;
1760 struct bitpack_d bp
;
1762 struct cgraph_edge
*e
;
1764 if (node
->clone
.args_to_skip
)
1766 streamer_write_uhwi (ob
, bitmap_count_bits (node
->clone
.args_to_skip
));
1767 EXECUTE_IF_SET_IN_BITMAP (node
->clone
.args_to_skip
, 0, index
, bi
)
1768 streamer_write_uhwi (ob
, index
);
1771 streamer_write_uhwi (ob
, 0);
1772 if (node
->clone
.combined_args_to_skip
)
1774 streamer_write_uhwi (ob
, bitmap_count_bits (node
->clone
.combined_args_to_skip
));
1775 EXECUTE_IF_SET_IN_BITMAP (node
->clone
.combined_args_to_skip
, 0, index
, bi
)
1776 streamer_write_uhwi (ob
, index
);
1779 streamer_write_uhwi (ob
, 0);
1780 streamer_write_uhwi (ob
, vec_safe_length (node
->clone
.tree_map
));
1781 FOR_EACH_VEC_SAFE_ELT (node
->clone
.tree_map
, i
, map
)
1783 /* At the moment we assume all old trees to be PARM_DECLs, because we have no
1784 mechanism to store function local declarations into summaries. */
1785 gcc_assert (!map
->old_tree
);
1786 streamer_write_uhwi (ob
, map
->parm_num
);
1787 gcc_assert (EXPR_LOCATION (map
->new_tree
) == UNKNOWN_LOCATION
);
1788 stream_write_tree (ob
, map
->new_tree
, true);
1789 bp
= bitpack_create (ob
->main_stream
);
1790 bp_pack_value (&bp
, map
->replace_p
, 1);
1791 bp_pack_value (&bp
, map
->ref_p
, 1);
1792 streamer_write_bitpack (&bp
);
1795 if (lto_symtab_encoder_in_partition_p (encoder
, node
))
1797 for (e
= node
->callees
; e
; e
= e
->next_callee
)
1798 output_edge_opt_summary (ob
, e
);
1799 for (e
= node
->indirect_calls
; e
; e
= e
->next_callee
)
1800 output_edge_opt_summary (ob
, e
);
1804 /* Output optimization summaries stored in callgraph.
1805 At the moment it is the clone info structure. */
1808 output_cgraph_opt_summary (void)
1811 lto_symtab_encoder_t encoder
;
1812 struct output_block
*ob
= create_output_block (LTO_section_cgraph_opt_sum
);
1816 encoder
= ob
->decl_state
->symtab_node_encoder
;
1817 n_nodes
= lto_symtab_encoder_size (encoder
);
1818 for (i
= 0; i
< n_nodes
; i
++)
1820 symtab_node
*node
= lto_symtab_encoder_deref (encoder
, i
);
1821 cgraph_node
*cnode
= dyn_cast
<cgraph_node
*> (node
);
1822 if (cnode
&& output_cgraph_opt_summary_p (cnode
))
1825 streamer_write_uhwi (ob
, count
);
1826 for (i
= 0; i
< n_nodes
; i
++)
1828 symtab_node
*node
= lto_symtab_encoder_deref (encoder
, i
);
1829 cgraph_node
*cnode
= dyn_cast
<cgraph_node
*> (node
);
1830 if (cnode
&& output_cgraph_opt_summary_p (cnode
))
1832 streamer_write_uhwi (ob
, i
);
1833 output_node_opt_summary (ob
, cnode
, encoder
);
1836 produce_asm (ob
, NULL
);
1837 destroy_output_block (ob
);
1840 /* Input optimisation summary of EDGE. */
1843 input_edge_opt_summary (struct cgraph_edge
*edge ATTRIBUTE_UNUSED
,
1844 struct lto_input_block
*ib_main ATTRIBUTE_UNUSED
)
1848 /* Input optimisation summary of NODE. */
1851 input_node_opt_summary (struct cgraph_node
*node
,
1852 struct lto_input_block
*ib_main
,
1853 struct data_in
*data_in
)
1858 struct bitpack_d bp
;
1859 struct cgraph_edge
*e
;
1861 count
= streamer_read_uhwi (ib_main
);
1863 node
->clone
.args_to_skip
= BITMAP_GGC_ALLOC ();
1864 for (i
= 0; i
< count
; i
++)
1866 bit
= streamer_read_uhwi (ib_main
);
1867 bitmap_set_bit (node
->clone
.args_to_skip
, bit
);
1869 count
= streamer_read_uhwi (ib_main
);
1871 node
->clone
.combined_args_to_skip
= BITMAP_GGC_ALLOC ();
1872 for (i
= 0; i
< count
; i
++)
1874 bit
= streamer_read_uhwi (ib_main
);
1875 bitmap_set_bit (node
->clone
.combined_args_to_skip
, bit
);
1877 count
= streamer_read_uhwi (ib_main
);
1878 for (i
= 0; i
< count
; i
++)
1880 struct ipa_replace_map
*map
= ggc_alloc
<ipa_replace_map
> ();
1882 vec_safe_push (node
->clone
.tree_map
, map
);
1883 map
->parm_num
= streamer_read_uhwi (ib_main
);
1884 map
->old_tree
= NULL
;
1885 map
->new_tree
= stream_read_tree (ib_main
, data_in
);
1886 bp
= streamer_read_bitpack (ib_main
);
1887 map
->replace_p
= bp_unpack_value (&bp
, 1);
1888 map
->ref_p
= bp_unpack_value (&bp
, 1);
1890 for (e
= node
->callees
; e
; e
= e
->next_callee
)
1891 input_edge_opt_summary (e
, ib_main
);
1892 for (e
= node
->indirect_calls
; e
; e
= e
->next_callee
)
1893 input_edge_opt_summary (e
, ib_main
);
1896 /* Read section in file FILE_DATA of length LEN with data DATA. */
1899 input_cgraph_opt_section (struct lto_file_decl_data
*file_data
,
1900 const char *data
, size_t len
,
1901 vec
<symtab_node
*> nodes
)
1903 const struct lto_function_header
*header
=
1904 (const struct lto_function_header
*) data
;
1905 const int cfg_offset
= sizeof (struct lto_function_header
);
1906 const int main_offset
= cfg_offset
+ header
->cfg_size
;
1907 const int string_offset
= main_offset
+ header
->main_size
;
1908 struct data_in
*data_in
;
1912 lto_input_block
ib_main ((const char *) data
+ main_offset
,
1916 lto_data_in_create (file_data
, (const char *) data
+ string_offset
,
1917 header
->string_size
, vNULL
);
1918 count
= streamer_read_uhwi (&ib_main
);
1920 for (i
= 0; i
< count
; i
++)
1922 int ref
= streamer_read_uhwi (&ib_main
);
1923 input_node_opt_summary (dyn_cast
<cgraph_node
*> (nodes
[ref
]),
1926 lto_free_section_data (file_data
, LTO_section_cgraph_opt_sum
, NULL
, data
,
1928 lto_data_in_delete (data_in
);
1931 /* Input optimization summary of cgraph. */
1934 input_cgraph_opt_summary (vec
<symtab_node
*> nodes
)
1936 struct lto_file_decl_data
**file_data_vec
= lto_get_file_decl_data ();
1937 struct lto_file_decl_data
*file_data
;
1940 while ((file_data
= file_data_vec
[j
++]))
1944 lto_get_section_data (file_data
, LTO_section_cgraph_opt_sum
, NULL
,
1948 input_cgraph_opt_section (file_data
, data
, len
, nodes
);