* doc/extend.texi (Size of an asm): Really move node to its position.
[official-gcc.git] / gcc / lto-cgraph.c
blob2329d1a6be52b1aa9a54119f7cbe7a8648549b46
1 /* Write and read the cgraph to the memory mapped representation of a
2 .o file.
4 Copyright (C) 2009-2014 Free Software Foundation, Inc.
5 Contributed by Kenneth Zadeck <zadeck@naturalbridge.com>
7 This file is part of GCC.
9 GCC is free software; you can redistribute it and/or modify it under
10 the terms of the GNU General Public License as published by the Free
11 Software Foundation; either version 3, or (at your option) any later
12 version.
14 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
15 WARRANTY; without even the implied warranty of MERCHANTABILITY or
16 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 for more details.
19 You should have received a copy of the GNU General Public License
20 along with GCC; see the file COPYING3. If not see
21 <http://www.gnu.org/licenses/>. */
23 #include "config.h"
24 #include "system.h"
25 #include "coretypes.h"
26 #include "tm.h"
27 #include "tree.h"
28 #include "stringpool.h"
29 #include "basic-block.h"
30 #include "tree-ssa-alias.h"
31 #include "internal-fn.h"
32 #include "gimple-expr.h"
33 #include "is-a.h"
34 #include "gimple.h"
35 #include "expr.h"
36 #include "flags.h"
37 #include "params.h"
38 #include "input.h"
39 #include "hashtab.h"
40 #include "langhooks.h"
41 #include "bitmap.h"
42 #include "function.h"
43 #include "diagnostic-core.h"
44 #include "except.h"
45 #include "timevar.h"
46 #include "lto-streamer.h"
47 #include "data-streamer.h"
48 #include "tree-streamer.h"
49 #include "gcov-io.h"
50 #include "tree-pass.h"
51 #include "profile.h"
52 #include "context.h"
53 #include "pass_manager.h"
54 #include "ipa-utils.h"
56 /* True when asm nodes has been output. */
57 bool asm_nodes_output = false;
59 static void output_cgraph_opt_summary (void);
60 static void input_cgraph_opt_summary (vec<symtab_node *> nodes);
62 /* Number of LDPR values known to GCC. */
63 #define LDPR_NUM_KNOWN (LDPR_PREVAILING_DEF_IRONLY_EXP + 1)
65 /* All node orders are ofsetted by ORDER_BASE. */
66 static int order_base;
68 /* Cgraph streaming is organized as set of record whose type
69 is indicated by a tag. */
70 enum LTO_symtab_tags
72 /* Must leave 0 for the stopper. */
74 /* Cgraph node without body available. */
75 LTO_symtab_unavail_node = 1,
76 /* Cgraph node with function body. */
77 LTO_symtab_analyzed_node,
78 /* Cgraph edges. */
79 LTO_symtab_edge,
80 LTO_symtab_indirect_edge,
81 LTO_symtab_variable,
82 LTO_symtab_last_tag
85 /* Create a new symtab encoder.
86 if FOR_INPUT, the encoder allocate only datastructures needed
87 to read the symtab. */
89 lto_symtab_encoder_t
90 lto_symtab_encoder_new (bool for_input)
92 lto_symtab_encoder_t encoder = XCNEW (struct lto_symtab_encoder_d);
94 if (!for_input)
95 encoder->map = pointer_map_create ();
96 encoder->nodes.create (0);
97 return encoder;
101 /* Delete ENCODER and its components. */
103 void
104 lto_symtab_encoder_delete (lto_symtab_encoder_t encoder)
106 encoder->nodes.release ();
107 if (encoder->map)
108 pointer_map_destroy (encoder->map);
109 free (encoder);
113 /* Return the existing reference number of NODE in the symtab encoder in
114 output block OB. Assign a new reference if this is the first time
115 NODE is encoded. */
118 lto_symtab_encoder_encode (lto_symtab_encoder_t encoder,
119 symtab_node *node)
121 int ref;
122 void **slot;
124 if (!encoder->map)
126 lto_encoder_entry entry = {node, false, false, false};
128 ref = encoder->nodes.length ();
129 encoder->nodes.safe_push (entry);
130 return ref;
133 slot = pointer_map_contains (encoder->map, node);
134 if (!slot || !*slot)
136 lto_encoder_entry entry = {node, false, false, false};
137 ref = encoder->nodes.length ();
138 if (!slot)
139 slot = pointer_map_insert (encoder->map, node);
140 *slot = (void *) (intptr_t) (ref + 1);
141 encoder->nodes.safe_push (entry);
143 else
144 ref = (size_t) *slot - 1;
146 return ref;
149 /* Remove NODE from encoder. */
151 bool
152 lto_symtab_encoder_delete_node (lto_symtab_encoder_t encoder,
153 symtab_node *node)
155 void **slot, **last_slot;
156 int index;
157 lto_encoder_entry last_node;
159 slot = pointer_map_contains (encoder->map, node);
160 if (slot == NULL || !*slot)
161 return false;
163 index = (size_t) *slot - 1;
164 gcc_checking_assert (encoder->nodes[index].node == node);
166 /* Remove from vector. We do this by swapping node with the last element
167 of the vector. */
168 last_node = encoder->nodes.pop ();
169 if (last_node.node != node)
171 last_slot = pointer_map_contains (encoder->map, last_node.node);
172 gcc_checking_assert (last_slot && *last_slot);
173 *last_slot = (void *)(size_t) (index + 1);
175 /* Move the last element to the original spot of NODE. */
176 encoder->nodes[index] = last_node;
179 /* Remove element from hash table. */
180 *slot = NULL;
181 return true;
185 /* Return TRUE if we should encode initializer of NODE (if any). */
187 bool
188 lto_symtab_encoder_encode_body_p (lto_symtab_encoder_t encoder,
189 struct cgraph_node *node)
191 int index = lto_symtab_encoder_lookup (encoder, node);
192 return encoder->nodes[index].body;
195 /* Return TRUE if we should encode body of NODE (if any). */
197 static void
198 lto_set_symtab_encoder_encode_body (lto_symtab_encoder_t encoder,
199 struct cgraph_node *node)
201 int index = lto_symtab_encoder_encode (encoder, node);
202 gcc_checking_assert (encoder->nodes[index].node == node);
203 encoder->nodes[index].body = true;
206 /* Return TRUE if we should encode initializer of NODE (if any). */
208 bool
209 lto_symtab_encoder_encode_initializer_p (lto_symtab_encoder_t encoder,
210 varpool_node *node)
212 int index = lto_symtab_encoder_lookup (encoder, node);
213 if (index == LCC_NOT_FOUND)
214 return false;
215 return encoder->nodes[index].initializer;
218 /* Return TRUE if we should encode initializer of NODE (if any). */
220 static void
221 lto_set_symtab_encoder_encode_initializer (lto_symtab_encoder_t encoder,
222 varpool_node *node)
224 int index = lto_symtab_encoder_lookup (encoder, node);
225 encoder->nodes[index].initializer = true;
228 /* Return TRUE if we should encode initializer of NODE (if any). */
230 bool
231 lto_symtab_encoder_in_partition_p (lto_symtab_encoder_t encoder,
232 symtab_node *node)
234 int index = lto_symtab_encoder_lookup (encoder, node);
235 if (index == LCC_NOT_FOUND)
236 return false;
237 return encoder->nodes[index].in_partition;
240 /* Return TRUE if we should encode body of NODE (if any). */
242 void
243 lto_set_symtab_encoder_in_partition (lto_symtab_encoder_t encoder,
244 symtab_node *node)
246 int index = lto_symtab_encoder_encode (encoder, node);
247 encoder->nodes[index].in_partition = true;
250 /* Output the cgraph EDGE to OB using ENCODER. */
252 static void
253 lto_output_edge (struct lto_simple_output_block *ob, struct cgraph_edge *edge,
254 lto_symtab_encoder_t encoder)
256 unsigned int uid;
257 intptr_t ref;
258 struct bitpack_d bp;
260 if (edge->indirect_unknown_callee)
261 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
262 LTO_symtab_indirect_edge);
263 else
264 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
265 LTO_symtab_edge);
267 ref = lto_symtab_encoder_lookup (encoder, edge->caller);
268 gcc_assert (ref != LCC_NOT_FOUND);
269 streamer_write_hwi_stream (ob->main_stream, ref);
271 if (!edge->indirect_unknown_callee)
273 ref = lto_symtab_encoder_lookup (encoder, edge->callee);
274 gcc_assert (ref != LCC_NOT_FOUND);
275 streamer_write_hwi_stream (ob->main_stream, ref);
278 streamer_write_gcov_count_stream (ob->main_stream, edge->count);
280 bp = bitpack_create (ob->main_stream);
281 uid = (!gimple_has_body_p (edge->caller->decl)
282 ? edge->lto_stmt_uid : gimple_uid (edge->call_stmt) + 1);
283 bp_pack_enum (&bp, cgraph_inline_failed_t,
284 CIF_N_REASONS, edge->inline_failed);
285 bp_pack_var_len_unsigned (&bp, uid);
286 bp_pack_var_len_unsigned (&bp, edge->frequency);
287 bp_pack_value (&bp, edge->indirect_inlining_edge, 1);
288 bp_pack_value (&bp, edge->speculative, 1);
289 bp_pack_value (&bp, edge->call_stmt_cannot_inline_p, 1);
290 bp_pack_value (&bp, edge->can_throw_external, 1);
291 if (edge->indirect_unknown_callee)
293 int flags = edge->indirect_info->ecf_flags;
294 bp_pack_value (&bp, (flags & ECF_CONST) != 0, 1);
295 bp_pack_value (&bp, (flags & ECF_PURE) != 0, 1);
296 bp_pack_value (&bp, (flags & ECF_NORETURN) != 0, 1);
297 bp_pack_value (&bp, (flags & ECF_MALLOC) != 0, 1);
298 bp_pack_value (&bp, (flags & ECF_NOTHROW) != 0, 1);
299 bp_pack_value (&bp, (flags & ECF_RETURNS_TWICE) != 0, 1);
300 /* Flags that should not appear on indirect calls. */
301 gcc_assert (!(flags & (ECF_LOOPING_CONST_OR_PURE
302 | ECF_MAY_BE_ALLOCA
303 | ECF_SIBCALL
304 | ECF_LEAF
305 | ECF_NOVOPS)));
307 streamer_write_bitpack (&bp);
308 if (edge->indirect_unknown_callee)
310 streamer_write_hwi_stream (ob->main_stream,
311 edge->indirect_info->common_target_id);
312 if (edge->indirect_info->common_target_id)
313 streamer_write_hwi_stream
314 (ob->main_stream, edge->indirect_info->common_target_probability);
318 /* Return if LIST contain references from other partitions. */
320 bool
321 referenced_from_other_partition_p (struct ipa_ref_list *list, lto_symtab_encoder_t encoder)
323 int i;
324 struct ipa_ref *ref;
325 for (i = 0; ipa_ref_list_referring_iterate (list, i, ref); i++)
327 if (ref->referring->in_other_partition
328 || !lto_symtab_encoder_in_partition_p (encoder, ref->referring))
329 return true;
331 return false;
334 /* Return true when node is reachable from other partition. */
336 bool
337 reachable_from_other_partition_p (struct cgraph_node *node, lto_symtab_encoder_t encoder)
339 struct cgraph_edge *e;
340 if (!node->definition)
341 return false;
342 if (node->global.inlined_to)
343 return false;
344 for (e = node->callers; e; e = e->next_caller)
345 if (e->caller->in_other_partition
346 || !lto_symtab_encoder_in_partition_p (encoder, e->caller))
347 return true;
348 return false;
351 /* Return if LIST contain references from other partitions. */
353 bool
354 referenced_from_this_partition_p (struct ipa_ref_list *list,
355 lto_symtab_encoder_t encoder)
357 int i;
358 struct ipa_ref *ref;
359 for (i = 0; ipa_ref_list_referring_iterate (list, i, ref); i++)
360 if (lto_symtab_encoder_in_partition_p (encoder, ref->referring))
361 return true;
362 return false;
365 /* Return true when node is reachable from other partition. */
367 bool
368 reachable_from_this_partition_p (struct cgraph_node *node, lto_symtab_encoder_t encoder)
370 struct cgraph_edge *e;
371 for (e = node->callers; e; e = e->next_caller)
372 if (lto_symtab_encoder_in_partition_p (encoder, e->caller))
373 return true;
374 return false;
377 /* Output the cgraph NODE to OB. ENCODER is used to find the
378 reference number of NODE->inlined_to. SET is the set of nodes we
379 are writing to the current file. If NODE is not in SET, then NODE
380 is a boundary of a cgraph_node_set and we pretend NODE just has a
381 decl and no callees. WRITTEN_DECLS is the set of FUNCTION_DECLs
382 that have had their callgraph node written so far. This is used to
383 determine if NODE is a clone of a previously written node. */
385 static void
386 lto_output_node (struct lto_simple_output_block *ob, struct cgraph_node *node,
387 lto_symtab_encoder_t encoder)
389 unsigned int tag;
390 struct bitpack_d bp;
391 bool boundary_p;
392 intptr_t ref;
393 bool in_other_partition = false;
394 struct cgraph_node *clone_of, *ultimate_clone_of;
395 ipa_opt_pass_d *pass;
396 int i;
397 bool alias_p;
399 boundary_p = !lto_symtab_encoder_in_partition_p (encoder, node);
401 if (node->analyzed && !boundary_p)
402 tag = LTO_symtab_analyzed_node;
403 else
404 tag = LTO_symtab_unavail_node;
406 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
407 tag);
408 streamer_write_hwi_stream (ob->main_stream, node->order);
410 /* In WPA mode, we only output part of the call-graph. Also, we
411 fake cgraph node attributes. There are two cases that we care.
413 Boundary nodes: There are nodes that are not part of SET but are
414 called from within SET. We artificially make them look like
415 externally visible nodes with no function body.
417 Cherry-picked nodes: These are nodes we pulled from other
418 translation units into SET during IPA-inlining. We make them as
419 local static nodes to prevent clashes with other local statics. */
420 if (boundary_p && node->analyzed
421 && symtab_get_symbol_partitioning_class (node) == SYMBOL_PARTITION)
423 /* Inline clones can not be part of boundary.
424 gcc_assert (!node->global.inlined_to);
426 FIXME: At the moment they can be, when partition contains an inline
427 clone that is clone of inline clone from outside partition. We can
428 reshape the clone tree and make other tree to be the root, but it
429 needs a bit extra work and will be promplty done by cgraph_remove_node
430 after reading back. */
431 in_other_partition = 1;
434 clone_of = node->clone_of;
435 while (clone_of
436 && (ref = lto_symtab_encoder_lookup (encoder, clone_of)) == LCC_NOT_FOUND)
437 if (clone_of->prev_sibling_clone)
438 clone_of = clone_of->prev_sibling_clone;
439 else
440 clone_of = clone_of->clone_of;
442 /* See if body of the master function is output. If not, we are seeing only
443 an declaration and we do not need to pass down clone tree. */
444 ultimate_clone_of = clone_of;
445 while (ultimate_clone_of && ultimate_clone_of->clone_of)
446 ultimate_clone_of = ultimate_clone_of->clone_of;
448 if (clone_of && !lto_symtab_encoder_encode_body_p (encoder, ultimate_clone_of))
449 clone_of = NULL;
451 if (tag == LTO_symtab_analyzed_node)
452 gcc_assert (clone_of || !node->clone_of);
453 if (!clone_of)
454 streamer_write_hwi_stream (ob->main_stream, LCC_NOT_FOUND);
455 else
456 streamer_write_hwi_stream (ob->main_stream, ref);
459 lto_output_fn_decl_index (ob->decl_state, ob->main_stream, node->decl);
460 streamer_write_gcov_count_stream (ob->main_stream, node->count);
461 streamer_write_hwi_stream (ob->main_stream, node->count_materialization_scale);
463 streamer_write_hwi_stream (ob->main_stream,
464 node->ipa_transforms_to_apply.length ());
465 FOR_EACH_VEC_ELT (node->ipa_transforms_to_apply, i, pass)
466 streamer_write_hwi_stream (ob->main_stream, pass->static_pass_number);
468 if (tag == LTO_symtab_analyzed_node)
470 if (node->global.inlined_to)
472 ref = lto_symtab_encoder_lookup (encoder, node->global.inlined_to);
473 gcc_assert (ref != LCC_NOT_FOUND);
475 else
476 ref = LCC_NOT_FOUND;
478 streamer_write_hwi_stream (ob->main_stream, ref);
481 if (node->same_comdat_group && !boundary_p)
483 ref = lto_symtab_encoder_lookup (encoder,
484 node->same_comdat_group);
485 gcc_assert (ref != LCC_NOT_FOUND);
487 else
488 ref = LCC_NOT_FOUND;
489 streamer_write_hwi_stream (ob->main_stream, ref);
491 streamer_write_hwi_stream (ob->main_stream, node->tp_first_run);
493 bp = bitpack_create (ob->main_stream);
494 bp_pack_value (&bp, node->local.local, 1);
495 bp_pack_value (&bp, node->externally_visible, 1);
496 bp_pack_value (&bp, node->definition, 1);
497 bp_pack_value (&bp, node->local.versionable, 1);
498 bp_pack_value (&bp, node->local.can_change_signature, 1);
499 bp_pack_value (&bp, node->local.redefined_extern_inline, 1);
500 bp_pack_value (&bp, node->force_output, 1);
501 bp_pack_value (&bp, node->forced_by_abi, 1);
502 bp_pack_value (&bp, node->unique_name, 1);
503 bp_pack_value (&bp, node->body_removed, 1);
504 bp_pack_value (&bp, node->address_taken, 1);
505 bp_pack_value (&bp, tag == LTO_symtab_analyzed_node
506 && symtab_get_symbol_partitioning_class (node) == SYMBOL_PARTITION
507 && (reachable_from_other_partition_p (node, encoder)
508 || referenced_from_other_partition_p (&node->ref_list,
509 encoder)), 1);
510 bp_pack_value (&bp, node->lowered, 1);
511 bp_pack_value (&bp, in_other_partition, 1);
512 /* Real aliases in a boundary become non-aliases. However we still stream
513 alias info on weakrefs.
514 TODO: We lose a bit of information here - when we know that variable is
515 defined in other unit, we may use the info on aliases to resolve
516 symbol1 != symbol2 type tests that we can do only for locally defined objects
517 otherwise. */
518 alias_p = node->alias && (!boundary_p || node->weakref);
519 bp_pack_value (&bp, alias_p, 1);
520 bp_pack_value (&bp, node->weakref, 1);
521 bp_pack_value (&bp, node->frequency, 2);
522 bp_pack_value (&bp, node->only_called_at_startup, 1);
523 bp_pack_value (&bp, node->only_called_at_exit, 1);
524 bp_pack_value (&bp, node->tm_clone, 1);
525 bp_pack_value (&bp, node->calls_comdat_local, 1);
526 bp_pack_value (&bp, node->thunk.thunk_p && !boundary_p, 1);
527 bp_pack_enum (&bp, ld_plugin_symbol_resolution,
528 LDPR_NUM_KNOWN, node->resolution);
529 streamer_write_bitpack (&bp);
531 if (node->thunk.thunk_p && !boundary_p)
533 streamer_write_uhwi_stream
534 (ob->main_stream,
535 1 + (node->thunk.this_adjusting != 0) * 2
536 + (node->thunk.virtual_offset_p != 0) * 4);
537 streamer_write_uhwi_stream (ob->main_stream, node->thunk.fixed_offset);
538 streamer_write_uhwi_stream (ob->main_stream, node->thunk.virtual_value);
540 streamer_write_hwi_stream (ob->main_stream, node->profile_id);
543 /* Output the varpool NODE to OB.
544 If NODE is not in SET, then NODE is a boundary. */
546 static void
547 lto_output_varpool_node (struct lto_simple_output_block *ob, varpool_node *node,
548 lto_symtab_encoder_t encoder)
550 bool boundary_p = !lto_symtab_encoder_in_partition_p (encoder, node);
551 struct bitpack_d bp;
552 int ref;
553 bool alias_p;
555 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
556 LTO_symtab_variable);
557 streamer_write_hwi_stream (ob->main_stream, node->order);
558 lto_output_var_decl_index (ob->decl_state, ob->main_stream, node->decl);
559 bp = bitpack_create (ob->main_stream);
560 bp_pack_value (&bp, node->externally_visible, 1);
561 bp_pack_value (&bp, node->force_output, 1);
562 bp_pack_value (&bp, node->forced_by_abi, 1);
563 bp_pack_value (&bp, node->unique_name, 1);
564 bp_pack_value (&bp, node->body_removed, 1);
565 bp_pack_value (&bp, node->writeonly, 1);
566 bp_pack_value (&bp, node->definition, 1);
567 alias_p = node->alias && (!boundary_p || node->weakref);
568 bp_pack_value (&bp, alias_p, 1);
569 bp_pack_value (&bp, node->weakref, 1);
570 bp_pack_value (&bp, node->analyzed && !boundary_p, 1);
571 gcc_assert (node->definition || !node->analyzed);
572 /* Constant pool initializers can be de-unified into individual ltrans units.
573 FIXME: Alternatively at -Os we may want to avoid generating for them the local
574 labels and share them across LTRANS partitions. */
575 if (symtab_get_symbol_partitioning_class (node) != SYMBOL_PARTITION)
577 bp_pack_value (&bp, 0, 1); /* used_from_other_parition. */
578 bp_pack_value (&bp, 0, 1); /* in_other_partition. */
580 else
582 bp_pack_value (&bp, node->definition
583 && referenced_from_other_partition_p (&node->ref_list,
584 encoder), 1);
585 bp_pack_value (&bp, node->analyzed
586 && boundary_p && !DECL_EXTERNAL (node->decl), 1);
587 /* in_other_partition. */
589 streamer_write_bitpack (&bp);
590 if (node->same_comdat_group && !boundary_p)
592 ref = lto_symtab_encoder_lookup (encoder,
593 node->same_comdat_group);
594 gcc_assert (ref != LCC_NOT_FOUND);
596 else
597 ref = LCC_NOT_FOUND;
598 streamer_write_hwi_stream (ob->main_stream, ref);
599 streamer_write_enum (ob->main_stream, ld_plugin_symbol_resolution,
600 LDPR_NUM_KNOWN, node->resolution);
603 /* Output the varpool NODE to OB.
604 If NODE is not in SET, then NODE is a boundary. */
606 static void
607 lto_output_ref (struct lto_simple_output_block *ob, struct ipa_ref *ref,
608 lto_symtab_encoder_t encoder)
610 struct bitpack_d bp;
611 int nref;
612 int uid = ref->lto_stmt_uid;
613 struct cgraph_node *node;
615 bp = bitpack_create (ob->main_stream);
616 bp_pack_value (&bp, ref->use, 2);
617 bp_pack_value (&bp, ref->speculative, 1);
618 streamer_write_bitpack (&bp);
619 nref = lto_symtab_encoder_lookup (encoder, ref->referred);
620 gcc_assert (nref != LCC_NOT_FOUND);
621 streamer_write_hwi_stream (ob->main_stream, nref);
623 node = dyn_cast <cgraph_node *> (ref->referring);
624 if (node)
626 if (ref->stmt)
627 uid = gimple_uid (ref->stmt) + 1;
628 streamer_write_hwi_stream (ob->main_stream, uid);
632 /* Stream out profile_summary to OB. */
634 static void
635 output_profile_summary (struct lto_simple_output_block *ob)
637 unsigned h_ix;
638 struct bitpack_d bp;
640 if (profile_info)
642 /* We do not output num and run_max, they are not used by
643 GCC profile feedback and they are difficult to merge from multiple
644 units. */
645 gcc_assert (profile_info->runs);
646 streamer_write_uhwi_stream (ob->main_stream, profile_info->runs);
647 streamer_write_gcov_count_stream (ob->main_stream, profile_info->sum_max);
649 /* sum_all is needed for computing the working set with the
650 histogram. */
651 streamer_write_gcov_count_stream (ob->main_stream, profile_info->sum_all);
653 /* Create and output a bitpack of non-zero histogram entries indices. */
654 bp = bitpack_create (ob->main_stream);
655 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
656 bp_pack_value (&bp, profile_info->histogram[h_ix].num_counters > 0, 1);
657 streamer_write_bitpack (&bp);
658 /* Now stream out only those non-zero entries. */
659 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
661 if (!profile_info->histogram[h_ix].num_counters)
662 continue;
663 streamer_write_gcov_count_stream (ob->main_stream,
664 profile_info->histogram[h_ix].num_counters);
665 streamer_write_gcov_count_stream (ob->main_stream,
666 profile_info->histogram[h_ix].min_value);
667 streamer_write_gcov_count_stream (ob->main_stream,
668 profile_info->histogram[h_ix].cum_value);
670 /* IPA-profile computes hot bb threshold based on cumulated
671 whole program profile. We need to stream it down to ltrans. */
672 if (flag_wpa)
673 streamer_write_gcov_count_stream (ob->main_stream,
674 get_hot_bb_threshold ());
676 else
677 streamer_write_uhwi_stream (ob->main_stream, 0);
680 /* Output all callees or indirect outgoing edges. EDGE must be the first such
681 edge. */
683 static void
684 output_outgoing_cgraph_edges (struct cgraph_edge *edge,
685 struct lto_simple_output_block *ob,
686 lto_symtab_encoder_t encoder)
688 if (!edge)
689 return;
691 /* Output edges in backward direction, so the reconstructed callgraph match
692 and it is easy to associate call sites in the IPA pass summaries. */
693 while (edge->next_callee)
694 edge = edge->next_callee;
695 for (; edge; edge = edge->prev_callee)
696 lto_output_edge (ob, edge, encoder);
699 /* Output the part of the cgraph in SET. */
701 static void
702 output_refs (lto_symtab_encoder_t encoder)
704 lto_symtab_encoder_iterator lsei;
705 struct lto_simple_output_block *ob;
706 int count;
707 struct ipa_ref *ref;
708 int i;
710 ob = lto_create_simple_output_block (LTO_section_refs);
712 for (lsei = lsei_start_in_partition (encoder); !lsei_end_p (lsei);
713 lsei_next_in_partition (&lsei))
715 symtab_node *node = lsei_node (lsei);
717 count = ipa_ref_list_nreferences (&node->ref_list);
718 if (count)
720 streamer_write_gcov_count_stream (ob->main_stream, count);
721 streamer_write_uhwi_stream (ob->main_stream,
722 lto_symtab_encoder_lookup (encoder, node));
723 for (i = 0; ipa_ref_list_reference_iterate (&node->ref_list,
724 i, ref); i++)
725 lto_output_ref (ob, ref, encoder);
729 streamer_write_uhwi_stream (ob->main_stream, 0);
731 lto_destroy_simple_output_block (ob);
734 /* Add NODE into encoder as well as nodes it is cloned from.
735 Do it in a way so clones appear first. */
737 static void
738 add_node_to (lto_symtab_encoder_t encoder, struct cgraph_node *node,
739 bool include_body)
741 if (node->clone_of)
742 add_node_to (encoder, node->clone_of, include_body);
743 else if (include_body)
744 lto_set_symtab_encoder_encode_body (encoder, node);
745 lto_symtab_encoder_encode (encoder, node);
748 /* Add all references in LIST to encoders. */
750 static void
751 add_references (lto_symtab_encoder_t encoder,
752 struct ipa_ref_list *list)
754 int i;
755 struct ipa_ref *ref;
756 for (i = 0; ipa_ref_list_reference_iterate (list, i, ref); i++)
757 if (is_a <cgraph_node *> (ref->referred))
758 add_node_to (encoder, ipa_ref_node (ref), false);
759 else
760 lto_symtab_encoder_encode (encoder, ref->referred);
763 /* Find all symbols we want to stream into given partition and insert them
764 to encoders.
766 The function actually replaces IN_ENCODER by new one. The reason is that
767 streaming code needs clone's origin to be streamed before clone. This
768 means that we need to insert the nodes in specific order. This order is
769 ignored by the partitioning logic earlier. */
771 lto_symtab_encoder_t
772 compute_ltrans_boundary (lto_symtab_encoder_t in_encoder)
774 struct cgraph_edge *edge;
775 int i;
776 lto_symtab_encoder_t encoder;
777 lto_symtab_encoder_iterator lsei;
778 struct pointer_set_t *reachable_call_targets = pointer_set_create ();
780 encoder = lto_symtab_encoder_new (false);
782 /* Go over all entries in the IN_ENCODER and duplicate them to
783 ENCODER. At the same time insert masters of clones so
784 every master appears before clone. */
785 for (lsei = lsei_start_function_in_partition (in_encoder);
786 !lsei_end_p (lsei); lsei_next_function_in_partition (&lsei))
788 struct cgraph_node *node = lsei_cgraph_node (lsei);
789 add_node_to (encoder, node, true);
790 lto_set_symtab_encoder_in_partition (encoder, node);
791 add_references (encoder, &node->ref_list);
792 /* For proper debug info, we need to ship the origins, too. */
793 if (DECL_ABSTRACT_ORIGIN (node->decl))
795 struct cgraph_node *origin_node
796 = cgraph_get_node (DECL_ABSTRACT_ORIGIN (node->decl));
797 add_node_to (encoder, origin_node, true);
800 for (lsei = lsei_start_variable_in_partition (in_encoder);
801 !lsei_end_p (lsei); lsei_next_variable_in_partition (&lsei))
803 varpool_node *vnode = lsei_varpool_node (lsei);
805 lto_set_symtab_encoder_in_partition (encoder, vnode);
806 lto_set_symtab_encoder_encode_initializer (encoder, vnode);
807 add_references (encoder, &vnode->ref_list);
808 /* For proper debug info, we need to ship the origins, too. */
809 if (DECL_ABSTRACT_ORIGIN (vnode->decl))
811 varpool_node *origin_node
812 = varpool_get_node (DECL_ABSTRACT_ORIGIN (vnode->decl));
813 lto_set_symtab_encoder_in_partition (encoder, origin_node);
816 /* Pickle in also the initializer of all referenced readonly variables
817 to help folding. Constant pool variables are not shared, so we must
818 pickle those too. */
819 for (i = 0; i < lto_symtab_encoder_size (encoder); i++)
821 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
822 if (varpool_node *vnode = dyn_cast <varpool_node *> (node))
824 if (!lto_symtab_encoder_encode_initializer_p (encoder,
825 vnode)
826 && ctor_for_folding (vnode->decl) != error_mark_node)
828 lto_set_symtab_encoder_encode_initializer (encoder, vnode);
829 add_references (encoder, &vnode->ref_list);
834 /* Go over all the nodes again to include callees that are not in
835 SET. */
836 for (lsei = lsei_start_function_in_partition (encoder);
837 !lsei_end_p (lsei); lsei_next_function_in_partition (&lsei))
839 struct cgraph_node *node = lsei_cgraph_node (lsei);
840 for (edge = node->callees; edge; edge = edge->next_callee)
842 struct cgraph_node *callee = edge->callee;
843 if (!lto_symtab_encoder_in_partition_p (encoder, callee))
845 /* We should have moved all the inlines. */
846 gcc_assert (!callee->global.inlined_to);
847 add_node_to (encoder, callee, false);
850 /* Add all possible targets for late devirtualization. */
851 if (flag_devirtualize)
852 for (edge = node->indirect_calls; edge; edge = edge->next_callee)
853 if (edge->indirect_info->polymorphic)
855 unsigned int i;
856 void *cache_token;
857 bool final;
858 vec <cgraph_node *>targets
859 = possible_polymorphic_call_targets
860 (edge, &final, &cache_token);
861 if (!pointer_set_insert (reachable_call_targets,
862 cache_token))
864 for (i = 0; i < targets.length (); i++)
866 struct cgraph_node *callee = targets[i];
868 /* Adding an external declarations into the unit serves
869 no purpose and just increases its boundary. */
870 if (callee->definition
871 && !lto_symtab_encoder_in_partition_p
872 (encoder, callee))
874 gcc_assert (!callee->global.inlined_to);
875 add_node_to (encoder, callee, false);
881 lto_symtab_encoder_delete (in_encoder);
882 pointer_set_destroy (reachable_call_targets);
883 return encoder;
886 /* Output the part of the symtab in SET and VSET. */
888 void
889 output_symtab (void)
891 struct cgraph_node *node;
892 struct lto_simple_output_block *ob;
893 lto_symtab_encoder_iterator lsei;
894 int i, n_nodes;
895 lto_symtab_encoder_t encoder;
897 if (flag_wpa)
898 output_cgraph_opt_summary ();
900 ob = lto_create_simple_output_block (LTO_section_symtab_nodes);
902 output_profile_summary (ob);
904 /* An encoder for cgraph nodes should have been created by
905 ipa_write_summaries_1. */
906 gcc_assert (ob->decl_state->symtab_node_encoder);
907 encoder = ob->decl_state->symtab_node_encoder;
909 /* Write out the nodes. We must first output a node and then its clones,
910 otherwise at a time reading back the node there would be nothing to clone
911 from. */
912 n_nodes = lto_symtab_encoder_size (encoder);
913 for (i = 0; i < n_nodes; i++)
915 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
916 if (cgraph_node *cnode = dyn_cast <cgraph_node *> (node))
917 lto_output_node (ob, cnode, encoder);
918 else
919 lto_output_varpool_node (ob, varpool (node), encoder);
923 /* Go over the nodes in SET again to write edges. */
924 for (lsei = lsei_start_function_in_partition (encoder); !lsei_end_p (lsei);
925 lsei_next_function_in_partition (&lsei))
927 node = lsei_cgraph_node (lsei);
928 output_outgoing_cgraph_edges (node->callees, ob, encoder);
929 output_outgoing_cgraph_edges (node->indirect_calls, ob, encoder);
932 streamer_write_uhwi_stream (ob->main_stream, 0);
934 lto_destroy_simple_output_block (ob);
936 /* Emit toplevel asms.
937 When doing WPA we must output every asm just once. Since we do not partition asm
938 nodes at all, output them to first output. This is kind of hack, but should work
939 well. */
940 if (!asm_nodes_output)
942 asm_nodes_output = true;
943 lto_output_toplevel_asms ();
946 output_refs (encoder);
949 /* Overwrite the information in NODE based on FILE_DATA, TAG, FLAGS,
950 STACK_SIZE, SELF_TIME and SELF_SIZE. This is called either to initialize
951 NODE or to replace the values in it, for instance because the first
952 time we saw it, the function body was not available but now it
953 is. BP is a bitpack with all the bitflags for NODE read from the
954 stream. */
956 static void
957 input_overwrite_node (struct lto_file_decl_data *file_data,
958 struct cgraph_node *node,
959 enum LTO_symtab_tags tag,
960 struct bitpack_d *bp)
962 node->aux = (void *) tag;
963 node->lto_file_data = file_data;
965 node->local.local = bp_unpack_value (bp, 1);
966 node->externally_visible = bp_unpack_value (bp, 1);
967 node->definition = bp_unpack_value (bp, 1);
968 node->local.versionable = bp_unpack_value (bp, 1);
969 node->local.can_change_signature = bp_unpack_value (bp, 1);
970 node->local.redefined_extern_inline = bp_unpack_value (bp, 1);
971 node->force_output = bp_unpack_value (bp, 1);
972 node->forced_by_abi = bp_unpack_value (bp, 1);
973 node->unique_name = bp_unpack_value (bp, 1);
974 node->body_removed = bp_unpack_value (bp, 1);
975 node->address_taken = bp_unpack_value (bp, 1);
976 node->used_from_other_partition = bp_unpack_value (bp, 1);
977 node->lowered = bp_unpack_value (bp, 1);
978 node->analyzed = tag == LTO_symtab_analyzed_node;
979 node->in_other_partition = bp_unpack_value (bp, 1);
980 if (node->in_other_partition
981 /* Avoid updating decl when we are seeing just inline clone.
982 When inlining function that has functions already inlined into it,
983 we produce clones of inline clones.
985 WPA partitioning might put each clone into different unit and
986 we might end up streaming inline clone from other partition
987 to support clone we are interested in. */
988 && (!node->clone_of
989 || node->clone_of->decl != node->decl))
991 DECL_EXTERNAL (node->decl) = 1;
992 TREE_STATIC (node->decl) = 0;
994 node->alias = bp_unpack_value (bp, 1);
995 node->weakref = bp_unpack_value (bp, 1);
996 node->frequency = (enum node_frequency)bp_unpack_value (bp, 2);
997 node->only_called_at_startup = bp_unpack_value (bp, 1);
998 node->only_called_at_exit = bp_unpack_value (bp, 1);
999 node->tm_clone = bp_unpack_value (bp, 1);
1000 node->calls_comdat_local = bp_unpack_value (bp, 1);
1001 node->thunk.thunk_p = bp_unpack_value (bp, 1);
1002 node->resolution = bp_unpack_enum (bp, ld_plugin_symbol_resolution,
1003 LDPR_NUM_KNOWN);
1004 gcc_assert (flag_ltrans
1005 || (!node->in_other_partition
1006 && !node->used_from_other_partition));
1009 /* Return string alias is alias of. */
1011 static tree
1012 get_alias_symbol (tree decl)
1014 tree alias = lookup_attribute ("alias", DECL_ATTRIBUTES (decl));
1015 return get_identifier (TREE_STRING_POINTER
1016 (TREE_VALUE (TREE_VALUE (alias))));
1019 /* Read a node from input_block IB. TAG is the node's tag just read.
1020 Return the node read or overwriten. */
1022 static struct cgraph_node *
1023 input_node (struct lto_file_decl_data *file_data,
1024 struct lto_input_block *ib,
1025 enum LTO_symtab_tags tag,
1026 vec<symtab_node *> nodes)
1028 gcc::pass_manager *passes = g->get_passes ();
1029 tree fn_decl;
1030 struct cgraph_node *node;
1031 struct bitpack_d bp;
1032 unsigned decl_index;
1033 int ref = LCC_NOT_FOUND, ref2 = LCC_NOT_FOUND;
1034 int clone_ref;
1035 int order;
1036 int i, count;
1038 order = streamer_read_hwi (ib) + order_base;
1039 clone_ref = streamer_read_hwi (ib);
1041 decl_index = streamer_read_uhwi (ib);
1042 fn_decl = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1044 if (clone_ref != LCC_NOT_FOUND)
1046 node = cgraph_clone_node (cgraph (nodes[clone_ref]), fn_decl,
1047 0, CGRAPH_FREQ_BASE, false,
1048 vNULL, false, NULL, NULL);
1050 else
1052 /* Declaration of functions can be already merged with a declaration
1053 from other input file. We keep cgraph unmerged until after streaming
1054 of ipa passes is done. Alays forcingly create a fresh node. */
1055 node = cgraph_create_empty_node ();
1056 node->decl = fn_decl;
1057 symtab_register_node (node);
1060 node->order = order;
1061 if (order >= symtab_order)
1062 symtab_order = order + 1;
1064 node->count = streamer_read_gcov_count (ib);
1065 node->count_materialization_scale = streamer_read_hwi (ib);
1067 count = streamer_read_hwi (ib);
1068 node->ipa_transforms_to_apply = vNULL;
1069 for (i = 0; i < count; i++)
1071 opt_pass *pass;
1072 int pid = streamer_read_hwi (ib);
1074 gcc_assert (pid < passes->passes_by_id_size);
1075 pass = passes->passes_by_id[pid];
1076 node->ipa_transforms_to_apply.safe_push ((ipa_opt_pass_d *) pass);
1079 if (tag == LTO_symtab_analyzed_node)
1080 ref = streamer_read_hwi (ib);
1082 ref2 = streamer_read_hwi (ib);
1084 /* Make sure that we have not read this node before. Nodes that
1085 have already been read will have their tag stored in the 'aux'
1086 field. Since built-in functions can be referenced in multiple
1087 functions, they are expected to be read more than once. */
1088 if (node->aux && !DECL_BUILT_IN (node->decl))
1089 internal_error ("bytecode stream: found multiple instances of cgraph "
1090 "node with uid %d", node->uid);
1092 node->tp_first_run = streamer_read_uhwi (ib);
1094 bp = streamer_read_bitpack (ib);
1096 input_overwrite_node (file_data, node, tag, &bp);
1098 /* Store a reference for now, and fix up later to be a pointer. */
1099 node->global.inlined_to = (cgraph_node_ptr) (intptr_t) ref;
1101 /* Store a reference for now, and fix up later to be a pointer. */
1102 node->same_comdat_group = (symtab_node *) (intptr_t) ref2;
1104 if (node->thunk.thunk_p)
1106 int type = streamer_read_uhwi (ib);
1107 HOST_WIDE_INT fixed_offset = streamer_read_uhwi (ib);
1108 HOST_WIDE_INT virtual_value = streamer_read_uhwi (ib);
1110 node->thunk.fixed_offset = fixed_offset;
1111 node->thunk.this_adjusting = (type & 2);
1112 node->thunk.virtual_value = virtual_value;
1113 node->thunk.virtual_offset_p = (type & 4);
1115 if (node->alias && !node->analyzed && node->weakref)
1116 node->alias_target = get_alias_symbol (node->decl);
1117 node->profile_id = streamer_read_hwi (ib);
1118 return node;
1121 /* Read a node from input_block IB. TAG is the node's tag just read.
1122 Return the node read or overwriten. */
1124 static varpool_node *
1125 input_varpool_node (struct lto_file_decl_data *file_data,
1126 struct lto_input_block *ib)
1128 int decl_index;
1129 tree var_decl;
1130 varpool_node *node;
1131 struct bitpack_d bp;
1132 int ref = LCC_NOT_FOUND;
1133 int order;
1135 order = streamer_read_hwi (ib) + order_base;
1136 decl_index = streamer_read_uhwi (ib);
1137 var_decl = lto_file_decl_data_get_var_decl (file_data, decl_index);
1139 /* Declaration of functions can be already merged with a declaration
1140 from other input file. We keep cgraph unmerged until after streaming
1141 of ipa passes is done. Alays forcingly create a fresh node. */
1142 node = varpool_create_empty_node ();
1143 node->decl = var_decl;
1144 symtab_register_node (node);
1146 node->order = order;
1147 if (order >= symtab_order)
1148 symtab_order = order + 1;
1149 node->lto_file_data = file_data;
1151 bp = streamer_read_bitpack (ib);
1152 node->externally_visible = bp_unpack_value (&bp, 1);
1153 node->force_output = bp_unpack_value (&bp, 1);
1154 node->forced_by_abi = bp_unpack_value (&bp, 1);
1155 node->unique_name = bp_unpack_value (&bp, 1);
1156 node->body_removed = bp_unpack_value (&bp, 1);
1157 node->writeonly = bp_unpack_value (&bp, 1);
1158 node->definition = bp_unpack_value (&bp, 1);
1159 node->alias = bp_unpack_value (&bp, 1);
1160 node->weakref = bp_unpack_value (&bp, 1);
1161 node->analyzed = bp_unpack_value (&bp, 1);
1162 node->used_from_other_partition = bp_unpack_value (&bp, 1);
1163 node->in_other_partition = bp_unpack_value (&bp, 1);
1164 if (node->in_other_partition)
1166 DECL_EXTERNAL (node->decl) = 1;
1167 TREE_STATIC (node->decl) = 0;
1169 if (node->alias && !node->analyzed && node->weakref)
1170 node->alias_target = get_alias_symbol (node->decl);
1171 ref = streamer_read_hwi (ib);
1172 /* Store a reference for now, and fix up later to be a pointer. */
1173 node->same_comdat_group = (symtab_node *) (intptr_t) ref;
1174 node->resolution = streamer_read_enum (ib, ld_plugin_symbol_resolution,
1175 LDPR_NUM_KNOWN);
1176 gcc_assert (flag_ltrans
1177 || (!node->in_other_partition
1178 && !node->used_from_other_partition));
1180 return node;
1183 /* Read a node from input_block IB. TAG is the node's tag just read.
1184 Return the node read or overwriten. */
1186 static void
1187 input_ref (struct lto_input_block *ib,
1188 symtab_node *referring_node,
1189 vec<symtab_node *> nodes)
1191 symtab_node *node = NULL;
1192 struct bitpack_d bp;
1193 enum ipa_ref_use use;
1194 bool speculative;
1195 struct ipa_ref *ref;
1197 bp = streamer_read_bitpack (ib);
1198 use = (enum ipa_ref_use) bp_unpack_value (&bp, 2);
1199 speculative = (enum ipa_ref_use) bp_unpack_value (&bp, 1);
1200 node = nodes[streamer_read_hwi (ib)];
1201 ref = ipa_record_reference (referring_node, node, use, NULL);
1202 ref->speculative = speculative;
1203 if (is_a <cgraph_node *> (referring_node))
1204 ref->lto_stmt_uid = streamer_read_hwi (ib);
1207 /* Read an edge from IB. NODES points to a vector of previously read nodes for
1208 decoding caller and callee of the edge to be read. If INDIRECT is true, the
1209 edge being read is indirect (in the sense that it has
1210 indirect_unknown_callee set). */
1212 static void
1213 input_edge (struct lto_input_block *ib, vec<symtab_node *> nodes,
1214 bool indirect)
1216 struct cgraph_node *caller, *callee;
1217 struct cgraph_edge *edge;
1218 unsigned int stmt_id;
1219 gcov_type count;
1220 int freq;
1221 cgraph_inline_failed_t inline_failed;
1222 struct bitpack_d bp;
1223 int ecf_flags = 0;
1225 caller = cgraph (nodes[streamer_read_hwi (ib)]);
1226 if (caller == NULL || caller->decl == NULL_TREE)
1227 internal_error ("bytecode stream: no caller found while reading edge");
1229 if (!indirect)
1231 callee = cgraph (nodes[streamer_read_hwi (ib)]);
1232 if (callee == NULL || callee->decl == NULL_TREE)
1233 internal_error ("bytecode stream: no callee found while reading edge");
1235 else
1236 callee = NULL;
1238 count = streamer_read_gcov_count (ib);
1240 bp = streamer_read_bitpack (ib);
1241 inline_failed = bp_unpack_enum (&bp, cgraph_inline_failed_t, CIF_N_REASONS);
1242 stmt_id = bp_unpack_var_len_unsigned (&bp);
1243 freq = (int) bp_unpack_var_len_unsigned (&bp);
1245 if (indirect)
1246 edge = cgraph_create_indirect_edge (caller, NULL, 0, count, freq);
1247 else
1248 edge = cgraph_create_edge (caller, callee, NULL, count, freq);
1250 edge->indirect_inlining_edge = bp_unpack_value (&bp, 1);
1251 edge->speculative = bp_unpack_value (&bp, 1);
1252 edge->lto_stmt_uid = stmt_id;
1253 edge->inline_failed = inline_failed;
1254 edge->call_stmt_cannot_inline_p = bp_unpack_value (&bp, 1);
1255 edge->can_throw_external = bp_unpack_value (&bp, 1);
1256 if (indirect)
1258 if (bp_unpack_value (&bp, 1))
1259 ecf_flags |= ECF_CONST;
1260 if (bp_unpack_value (&bp, 1))
1261 ecf_flags |= ECF_PURE;
1262 if (bp_unpack_value (&bp, 1))
1263 ecf_flags |= ECF_NORETURN;
1264 if (bp_unpack_value (&bp, 1))
1265 ecf_flags |= ECF_MALLOC;
1266 if (bp_unpack_value (&bp, 1))
1267 ecf_flags |= ECF_NOTHROW;
1268 if (bp_unpack_value (&bp, 1))
1269 ecf_flags |= ECF_RETURNS_TWICE;
1270 edge->indirect_info->ecf_flags = ecf_flags;
1271 edge->indirect_info->common_target_id = streamer_read_hwi (ib);
1272 if (edge->indirect_info->common_target_id)
1273 edge->indirect_info->common_target_probability = streamer_read_hwi (ib);
1278 /* Read a cgraph from IB using the info in FILE_DATA. */
1280 static vec<symtab_node *>
1281 input_cgraph_1 (struct lto_file_decl_data *file_data,
1282 struct lto_input_block *ib)
1284 enum LTO_symtab_tags tag;
1285 vec<symtab_node *> nodes = vNULL;
1286 symtab_node *node;
1287 unsigned i;
1289 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1290 order_base = symtab_order;
1291 while (tag)
1293 if (tag == LTO_symtab_edge)
1294 input_edge (ib, nodes, false);
1295 else if (tag == LTO_symtab_indirect_edge)
1296 input_edge (ib, nodes, true);
1297 else if (tag == LTO_symtab_variable)
1299 node = input_varpool_node (file_data, ib);
1300 nodes.safe_push (node);
1301 lto_symtab_encoder_encode (file_data->symtab_node_encoder, node);
1303 else
1305 node = input_node (file_data, ib, tag, nodes);
1306 if (node == NULL || node->decl == NULL_TREE)
1307 internal_error ("bytecode stream: found empty cgraph node");
1308 nodes.safe_push (node);
1309 lto_symtab_encoder_encode (file_data->symtab_node_encoder, node);
1312 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1315 lto_input_toplevel_asms (file_data, order_base);
1317 /* AUX pointers should be all non-zero for function nodes read from the stream. */
1318 #ifdef ENABLE_CHECKING
1319 FOR_EACH_VEC_ELT (nodes, i, node)
1320 gcc_assert (node->aux || !is_a <cgraph_node *> (node));
1321 #endif
1322 FOR_EACH_VEC_ELT (nodes, i, node)
1324 int ref;
1325 if (cgraph_node *cnode = dyn_cast <cgraph_node *> (node))
1327 ref = (int) (intptr_t) cnode->global.inlined_to;
1329 /* We share declaration of builtins, so we may read same node twice. */
1330 if (!node->aux)
1331 continue;
1332 node->aux = NULL;
1334 /* Fixup inlined_to from reference to pointer. */
1335 if (ref != LCC_NOT_FOUND)
1336 cgraph (node)->global.inlined_to = cgraph (nodes[ref]);
1337 else
1338 cnode->global.inlined_to = NULL;
1341 ref = (int) (intptr_t) node->same_comdat_group;
1343 /* Fixup same_comdat_group from reference to pointer. */
1344 if (ref != LCC_NOT_FOUND)
1345 node->same_comdat_group = nodes[ref];
1346 else
1347 node->same_comdat_group = NULL;
1349 FOR_EACH_VEC_ELT (nodes, i, node)
1350 node->aux = is_a <cgraph_node *> (node) ? (void *)1 : NULL;
1351 return nodes;
1354 /* Input ipa_refs. */
1356 static void
1357 input_refs (struct lto_input_block *ib,
1358 vec<symtab_node *> nodes)
1360 int count;
1361 int idx;
1362 while (true)
1364 symtab_node *node;
1365 count = streamer_read_uhwi (ib);
1366 if (!count)
1367 break;
1368 idx = streamer_read_uhwi (ib);
1369 node = nodes[idx];
1370 while (count)
1372 input_ref (ib, node, nodes);
1373 count--;
1379 static struct gcov_ctr_summary lto_gcov_summary;
1381 /* Input profile_info from IB. */
1382 static void
1383 input_profile_summary (struct lto_input_block *ib,
1384 struct lto_file_decl_data *file_data)
1386 unsigned h_ix;
1387 struct bitpack_d bp;
1388 unsigned int runs = streamer_read_uhwi (ib);
1389 if (runs)
1391 file_data->profile_info.runs = runs;
1392 file_data->profile_info.sum_max = streamer_read_gcov_count (ib);
1393 file_data->profile_info.sum_all = streamer_read_gcov_count (ib);
1395 memset (file_data->profile_info.histogram, 0,
1396 sizeof (gcov_bucket_type) * GCOV_HISTOGRAM_SIZE);
1397 /* Input the bitpack of non-zero histogram indices. */
1398 bp = streamer_read_bitpack (ib);
1399 /* Read in and unpack the full bitpack, flagging non-zero
1400 histogram entries by setting the num_counters non-zero. */
1401 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1403 file_data->profile_info.histogram[h_ix].num_counters
1404 = bp_unpack_value (&bp, 1);
1406 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1408 if (!file_data->profile_info.histogram[h_ix].num_counters)
1409 continue;
1411 file_data->profile_info.histogram[h_ix].num_counters
1412 = streamer_read_gcov_count (ib);
1413 file_data->profile_info.histogram[h_ix].min_value
1414 = streamer_read_gcov_count (ib);
1415 file_data->profile_info.histogram[h_ix].cum_value
1416 = streamer_read_gcov_count (ib);
1418 /* IPA-profile computes hot bb threshold based on cumulated
1419 whole program profile. We need to stream it down to ltrans. */
1420 if (flag_ltrans)
1421 set_hot_bb_threshold (streamer_read_gcov_count (ib));
1426 /* Rescale profile summaries to the same number of runs in the whole unit. */
1428 static void
1429 merge_profile_summaries (struct lto_file_decl_data **file_data_vec)
1431 struct lto_file_decl_data *file_data;
1432 unsigned int j, h_ix;
1433 gcov_unsigned_t max_runs = 0;
1434 struct cgraph_node *node;
1435 struct cgraph_edge *edge;
1436 gcov_type saved_sum_all = 0;
1437 gcov_ctr_summary *saved_profile_info = 0;
1438 int saved_scale = 0;
1440 /* Find unit with maximal number of runs. If we ever get serious about
1441 roundoff errors, we might also consider computing smallest common
1442 multiply. */
1443 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1444 if (max_runs < file_data->profile_info.runs)
1445 max_runs = file_data->profile_info.runs;
1447 if (!max_runs)
1448 return;
1450 /* Simple overflow check. We probably don't need to support that many train
1451 runs. Such a large value probably imply data corruption anyway. */
1452 if (max_runs > INT_MAX / REG_BR_PROB_BASE)
1454 sorry ("At most %i profile runs is supported. Perhaps corrupted profile?",
1455 INT_MAX / REG_BR_PROB_BASE);
1456 return;
1459 profile_info = &lto_gcov_summary;
1460 lto_gcov_summary.runs = max_runs;
1461 lto_gcov_summary.sum_max = 0;
1462 memset (lto_gcov_summary.histogram, 0,
1463 sizeof (gcov_bucket_type) * GCOV_HISTOGRAM_SIZE);
1465 /* Rescale all units to the maximal number of runs.
1466 sum_max can not be easily merged, as we have no idea what files come from
1467 the same run. We do not use the info anyway, so leave it 0. */
1468 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1469 if (file_data->profile_info.runs)
1471 int scale = GCOV_COMPUTE_SCALE (max_runs,
1472 file_data->profile_info.runs);
1473 lto_gcov_summary.sum_max
1474 = MAX (lto_gcov_summary.sum_max,
1475 apply_scale (file_data->profile_info.sum_max, scale));
1476 lto_gcov_summary.sum_all
1477 = MAX (lto_gcov_summary.sum_all,
1478 apply_scale (file_data->profile_info.sum_all, scale));
1479 /* Save a pointer to the profile_info with the largest
1480 scaled sum_all and the scale for use in merging the
1481 histogram. */
1482 if (!saved_profile_info
1483 || lto_gcov_summary.sum_all > saved_sum_all)
1485 saved_profile_info = &file_data->profile_info;
1486 saved_sum_all = lto_gcov_summary.sum_all;
1487 saved_scale = scale;
1491 gcc_assert (saved_profile_info);
1493 /* Scale up the histogram from the profile that had the largest
1494 scaled sum_all above. */
1495 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1497 /* Scale up the min value as we did the corresponding sum_all
1498 above. Use that to find the new histogram index. */
1499 gcov_type scaled_min
1500 = apply_scale (saved_profile_info->histogram[h_ix].min_value,
1501 saved_scale);
1502 /* The new index may be shared with another scaled histogram entry,
1503 so we need to account for a non-zero histogram entry at new_ix. */
1504 unsigned new_ix = gcov_histo_index (scaled_min);
1505 lto_gcov_summary.histogram[new_ix].min_value
1506 = (lto_gcov_summary.histogram[new_ix].num_counters
1507 ? MIN (lto_gcov_summary.histogram[new_ix].min_value, scaled_min)
1508 : scaled_min);
1509 /* Some of the scaled counter values would ostensibly need to be placed
1510 into different (larger) histogram buckets, but we keep things simple
1511 here and place the scaled cumulative counter value in the bucket
1512 corresponding to the scaled minimum counter value. */
1513 lto_gcov_summary.histogram[new_ix].cum_value
1514 += apply_scale (saved_profile_info->histogram[h_ix].cum_value,
1515 saved_scale);
1516 lto_gcov_summary.histogram[new_ix].num_counters
1517 += saved_profile_info->histogram[h_ix].num_counters;
1520 /* Watch roundoff errors. */
1521 if (lto_gcov_summary.sum_max < max_runs)
1522 lto_gcov_summary.sum_max = max_runs;
1524 /* If merging already happent at WPA time, we are done. */
1525 if (flag_ltrans)
1526 return;
1528 /* Now compute count_materialization_scale of each node.
1529 During LTRANS we already have values of count_materialization_scale
1530 computed, so just update them. */
1531 FOR_EACH_FUNCTION (node)
1532 if (node->lto_file_data
1533 && node->lto_file_data->profile_info.runs)
1535 int scale;
1537 scale = RDIV (node->count_materialization_scale * max_runs,
1538 node->lto_file_data->profile_info.runs);
1539 node->count_materialization_scale = scale;
1540 if (scale < 0)
1541 fatal_error ("Profile information in %s corrupted",
1542 file_data->file_name);
1544 if (scale == REG_BR_PROB_BASE)
1545 continue;
1546 for (edge = node->callees; edge; edge = edge->next_callee)
1547 edge->count = apply_scale (edge->count, scale);
1548 node->count = apply_scale (node->count, scale);
1552 /* Input and merge the symtab from each of the .o files passed to
1553 lto1. */
1555 void
1556 input_symtab (void)
1558 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1559 struct lto_file_decl_data *file_data;
1560 unsigned int j = 0;
1561 struct cgraph_node *node;
1563 while ((file_data = file_data_vec[j++]))
1565 const char *data;
1566 size_t len;
1567 struct lto_input_block *ib;
1568 vec<symtab_node *> nodes;
1570 ib = lto_create_simple_input_block (file_data, LTO_section_symtab_nodes,
1571 &data, &len);
1572 if (!ib)
1573 fatal_error ("cannot find LTO cgraph in %s", file_data->file_name);
1574 input_profile_summary (ib, file_data);
1575 file_data->symtab_node_encoder = lto_symtab_encoder_new (true);
1576 nodes = input_cgraph_1 (file_data, ib);
1577 lto_destroy_simple_input_block (file_data, LTO_section_symtab_nodes,
1578 ib, data, len);
1580 ib = lto_create_simple_input_block (file_data, LTO_section_refs,
1581 &data, &len);
1582 if (!ib)
1583 fatal_error ("cannot find LTO section refs in %s",
1584 file_data->file_name);
1585 input_refs (ib, nodes);
1586 lto_destroy_simple_input_block (file_data, LTO_section_refs,
1587 ib, data, len);
1588 if (flag_ltrans)
1589 input_cgraph_opt_summary (nodes);
1590 nodes.release ();
1593 merge_profile_summaries (file_data_vec);
1594 get_working_sets ();
1597 /* Clear out the aux field that was used to store enough state to
1598 tell which nodes should be overwritten. */
1599 FOR_EACH_FUNCTION (node)
1601 /* Some nodes may have been created by cgraph_node. This
1602 happens when the callgraph contains nested functions. If the
1603 node for the parent function was never emitted to the gimple
1604 file, cgraph_node will create a node for it when setting the
1605 context of the nested function. */
1606 if (node->lto_file_data)
1607 node->aux = NULL;
1611 /* True when we need optimization summary for NODE. */
1613 static int
1614 output_cgraph_opt_summary_p (struct cgraph_node *node)
1616 return (node->clone_of
1617 && (node->clone.tree_map
1618 || node->clone.args_to_skip
1619 || node->clone.combined_args_to_skip));
1622 /* Output optimization summary for EDGE to OB. */
1623 static void
1624 output_edge_opt_summary (struct output_block *ob ATTRIBUTE_UNUSED,
1625 struct cgraph_edge *edge ATTRIBUTE_UNUSED)
1629 /* Output optimization summary for NODE to OB. */
1631 static void
1632 output_node_opt_summary (struct output_block *ob,
1633 struct cgraph_node *node,
1634 lto_symtab_encoder_t encoder)
1636 unsigned int index;
1637 bitmap_iterator bi;
1638 struct ipa_replace_map *map;
1639 struct bitpack_d bp;
1640 int i;
1641 struct cgraph_edge *e;
1643 if (node->clone.args_to_skip)
1645 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.args_to_skip));
1646 EXECUTE_IF_SET_IN_BITMAP (node->clone.args_to_skip, 0, index, bi)
1647 streamer_write_uhwi (ob, index);
1649 else
1650 streamer_write_uhwi (ob, 0);
1651 if (node->clone.combined_args_to_skip)
1653 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.combined_args_to_skip));
1654 EXECUTE_IF_SET_IN_BITMAP (node->clone.combined_args_to_skip, 0, index, bi)
1655 streamer_write_uhwi (ob, index);
1657 else
1658 streamer_write_uhwi (ob, 0);
1659 streamer_write_uhwi (ob, vec_safe_length (node->clone.tree_map));
1660 FOR_EACH_VEC_SAFE_ELT (node->clone.tree_map, i, map)
1662 /* At the moment we assume all old trees to be PARM_DECLs, because we have no
1663 mechanism to store function local declarations into summaries. */
1664 gcc_assert (!map->old_tree);
1665 streamer_write_uhwi (ob, map->parm_num);
1666 gcc_assert (EXPR_LOCATION (map->new_tree) == UNKNOWN_LOCATION);
1667 stream_write_tree (ob, map->new_tree, true);
1668 bp = bitpack_create (ob->main_stream);
1669 bp_pack_value (&bp, map->replace_p, 1);
1670 bp_pack_value (&bp, map->ref_p, 1);
1671 streamer_write_bitpack (&bp);
1674 if (lto_symtab_encoder_in_partition_p (encoder, node))
1676 for (e = node->callees; e; e = e->next_callee)
1677 output_edge_opt_summary (ob, e);
1678 for (e = node->indirect_calls; e; e = e->next_callee)
1679 output_edge_opt_summary (ob, e);
1683 /* Output optimization summaries stored in callgraph.
1684 At the moment it is the clone info structure. */
1686 static void
1687 output_cgraph_opt_summary (void)
1689 int i, n_nodes;
1690 lto_symtab_encoder_t encoder;
1691 struct output_block *ob = create_output_block (LTO_section_cgraph_opt_sum);
1692 unsigned count = 0;
1694 ob->cgraph_node = NULL;
1695 encoder = ob->decl_state->symtab_node_encoder;
1696 n_nodes = lto_symtab_encoder_size (encoder);
1697 for (i = 0; i < n_nodes; i++)
1699 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
1700 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
1701 if (cnode && output_cgraph_opt_summary_p (cnode))
1702 count++;
1704 streamer_write_uhwi (ob, count);
1705 for (i = 0; i < n_nodes; i++)
1707 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
1708 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
1709 if (cnode && output_cgraph_opt_summary_p (cnode))
1711 streamer_write_uhwi (ob, i);
1712 output_node_opt_summary (ob, cnode, encoder);
1715 produce_asm (ob, NULL);
1716 destroy_output_block (ob);
1719 /* Input optimisation summary of EDGE. */
1721 static void
1722 input_edge_opt_summary (struct cgraph_edge *edge ATTRIBUTE_UNUSED,
1723 struct lto_input_block *ib_main ATTRIBUTE_UNUSED)
1727 /* Input optimisation summary of NODE. */
1729 static void
1730 input_node_opt_summary (struct cgraph_node *node,
1731 struct lto_input_block *ib_main,
1732 struct data_in *data_in)
1734 int i;
1735 int count;
1736 int bit;
1737 struct bitpack_d bp;
1738 struct cgraph_edge *e;
1740 count = streamer_read_uhwi (ib_main);
1741 if (count)
1742 node->clone.args_to_skip = BITMAP_GGC_ALLOC ();
1743 for (i = 0; i < count; i++)
1745 bit = streamer_read_uhwi (ib_main);
1746 bitmap_set_bit (node->clone.args_to_skip, bit);
1748 count = streamer_read_uhwi (ib_main);
1749 if (count)
1750 node->clone.combined_args_to_skip = BITMAP_GGC_ALLOC ();
1751 for (i = 0; i < count; i++)
1753 bit = streamer_read_uhwi (ib_main);
1754 bitmap_set_bit (node->clone.combined_args_to_skip, bit);
1756 count = streamer_read_uhwi (ib_main);
1757 for (i = 0; i < count; i++)
1759 struct ipa_replace_map *map = ggc_alloc<ipa_replace_map> ();
1761 vec_safe_push (node->clone.tree_map, map);
1762 map->parm_num = streamer_read_uhwi (ib_main);
1763 map->old_tree = NULL;
1764 map->new_tree = stream_read_tree (ib_main, data_in);
1765 bp = streamer_read_bitpack (ib_main);
1766 map->replace_p = bp_unpack_value (&bp, 1);
1767 map->ref_p = bp_unpack_value (&bp, 1);
1769 for (e = node->callees; e; e = e->next_callee)
1770 input_edge_opt_summary (e, ib_main);
1771 for (e = node->indirect_calls; e; e = e->next_callee)
1772 input_edge_opt_summary (e, ib_main);
1775 /* Read section in file FILE_DATA of length LEN with data DATA. */
1777 static void
1778 input_cgraph_opt_section (struct lto_file_decl_data *file_data,
1779 const char *data, size_t len,
1780 vec<symtab_node *> nodes)
1782 const struct lto_function_header *header =
1783 (const struct lto_function_header *) data;
1784 const int cfg_offset = sizeof (struct lto_function_header);
1785 const int main_offset = cfg_offset + header->cfg_size;
1786 const int string_offset = main_offset + header->main_size;
1787 struct data_in *data_in;
1788 struct lto_input_block ib_main;
1789 unsigned int i;
1790 unsigned int count;
1792 LTO_INIT_INPUT_BLOCK (ib_main, (const char *) data + main_offset, 0,
1793 header->main_size);
1795 data_in =
1796 lto_data_in_create (file_data, (const char *) data + string_offset,
1797 header->string_size, vNULL);
1798 count = streamer_read_uhwi (&ib_main);
1800 for (i = 0; i < count; i++)
1802 int ref = streamer_read_uhwi (&ib_main);
1803 input_node_opt_summary (cgraph (nodes[ref]),
1804 &ib_main, data_in);
1806 lto_free_section_data (file_data, LTO_section_cgraph_opt_sum, NULL, data,
1807 len);
1808 lto_data_in_delete (data_in);
1811 /* Input optimization summary of cgraph. */
1813 static void
1814 input_cgraph_opt_summary (vec<symtab_node *> nodes)
1816 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1817 struct lto_file_decl_data *file_data;
1818 unsigned int j = 0;
1820 while ((file_data = file_data_vec[j++]))
1822 size_t len;
1823 const char *data =
1824 lto_get_section_data (file_data, LTO_section_cgraph_opt_sum, NULL,
1825 &len);
1827 if (data)
1828 input_cgraph_opt_section (file_data, data, len, nodes);