Mark omp thread functions as parallelized
[official-gcc.git] / gcc / lto-cgraph.c
blob088de8606464c35d4f50d0452fecdc5d77a94ee1
1 /* Write and read the cgraph to the memory mapped representation of a
2 .o file.
4 Copyright (C) 2009-2015 Free Software Foundation, Inc.
5 Contributed by Kenneth Zadeck <zadeck@naturalbridge.com>
7 This file is part of GCC.
9 GCC is free software; you can redistribute it and/or modify it under
10 the terms of the GNU General Public License as published by the Free
11 Software Foundation; either version 3, or (at your option) any later
12 version.
14 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
15 WARRANTY; without even the implied warranty of MERCHANTABILITY or
16 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 for more details.
19 You should have received a copy of the GNU General Public License
20 along with GCC; see the file COPYING3. If not see
21 <http://www.gnu.org/licenses/>. */
23 #include "config.h"
24 #include "system.h"
25 #include "coretypes.h"
26 #include "tm.h"
27 #include "hash-set.h"
28 #include "machmode.h"
29 #include "vec.h"
30 #include "double-int.h"
31 #include "input.h"
32 #include "alias.h"
33 #include "symtab.h"
34 #include "wide-int.h"
35 #include "inchash.h"
36 #include "tree.h"
37 #include "fold-const.h"
38 #include "stringpool.h"
39 #include "predict.h"
40 #include "hard-reg-set.h"
41 #include "function.h"
42 #include "basic-block.h"
43 #include "tree-ssa-alias.h"
44 #include "internal-fn.h"
45 #include "gimple-expr.h"
46 #include "is-a.h"
47 #include "gimple.h"
48 #include "hashtab.h"
49 #include "rtl.h"
50 #include "flags.h"
51 #include "statistics.h"
52 #include "real.h"
53 #include "fixed-value.h"
54 #include "insn-config.h"
55 #include "expmed.h"
56 #include "dojump.h"
57 #include "explow.h"
58 #include "calls.h"
59 #include "emit-rtl.h"
60 #include "varasm.h"
61 #include "stmt.h"
62 #include "expr.h"
63 #include "params.h"
64 #include "langhooks.h"
65 #include "bitmap.h"
66 #include "diagnostic-core.h"
67 #include "except.h"
68 #include "timevar.h"
69 #include "hash-map.h"
70 #include "plugin-api.h"
71 #include "ipa-ref.h"
72 #include "cgraph.h"
73 #include "lto-streamer.h"
74 #include "data-streamer.h"
75 #include "tree-streamer.h"
76 #include "gcov-io.h"
77 #include "tree-pass.h"
78 #include "profile.h"
79 #include "context.h"
80 #include "pass_manager.h"
81 #include "ipa-utils.h"
82 #include "omp-low.h"
84 /* True when asm nodes has been output. */
85 bool asm_nodes_output = false;
87 static void output_cgraph_opt_summary (void);
88 static void input_cgraph_opt_summary (vec<symtab_node *> nodes);
90 /* Number of LDPR values known to GCC. */
91 #define LDPR_NUM_KNOWN (LDPR_PREVAILING_DEF_IRONLY_EXP + 1)
93 /* All node orders are ofsetted by ORDER_BASE. */
94 static int order_base;
96 /* Cgraph streaming is organized as set of record whose type
97 is indicated by a tag. */
98 enum LTO_symtab_tags
100 /* Must leave 0 for the stopper. */
102 /* Cgraph node without body available. */
103 LTO_symtab_unavail_node = 1,
104 /* Cgraph node with function body. */
105 LTO_symtab_analyzed_node,
106 /* Cgraph edges. */
107 LTO_symtab_edge,
108 LTO_symtab_indirect_edge,
109 LTO_symtab_variable,
110 LTO_symtab_last_tag
113 /* Create a new symtab encoder.
114 if FOR_INPUT, the encoder allocate only datastructures needed
115 to read the symtab. */
117 lto_symtab_encoder_t
118 lto_symtab_encoder_new (bool for_input)
120 lto_symtab_encoder_t encoder = XCNEW (struct lto_symtab_encoder_d);
122 if (!for_input)
123 encoder->map = new hash_map<symtab_node *, size_t>;
124 encoder->nodes.create (0);
125 return encoder;
129 /* Delete ENCODER and its components. */
131 void
132 lto_symtab_encoder_delete (lto_symtab_encoder_t encoder)
134 encoder->nodes.release ();
135 if (encoder->map)
136 delete encoder->map;
137 free (encoder);
141 /* Return the existing reference number of NODE in the symtab encoder in
142 output block OB. Assign a new reference if this is the first time
143 NODE is encoded. */
146 lto_symtab_encoder_encode (lto_symtab_encoder_t encoder,
147 symtab_node *node)
149 int ref;
151 if (!encoder->map)
153 lto_encoder_entry entry = {node, false, false, false};
155 ref = encoder->nodes.length ();
156 encoder->nodes.safe_push (entry);
157 return ref;
160 size_t *slot = encoder->map->get (node);
161 if (!slot || !*slot)
163 lto_encoder_entry entry = {node, false, false, false};
164 ref = encoder->nodes.length ();
165 if (!slot)
166 encoder->map->put (node, ref + 1);
167 encoder->nodes.safe_push (entry);
169 else
170 ref = *slot - 1;
172 return ref;
175 /* Remove NODE from encoder. */
177 bool
178 lto_symtab_encoder_delete_node (lto_symtab_encoder_t encoder,
179 symtab_node *node)
181 int index;
182 lto_encoder_entry last_node;
184 size_t *slot = encoder->map->get (node);
185 if (slot == NULL || !*slot)
186 return false;
188 index = *slot - 1;
189 gcc_checking_assert (encoder->nodes[index].node == node);
191 /* Remove from vector. We do this by swapping node with the last element
192 of the vector. */
193 last_node = encoder->nodes.pop ();
194 if (last_node.node != node)
196 gcc_assert (encoder->map->put (last_node.node, index + 1));
198 /* Move the last element to the original spot of NODE. */
199 encoder->nodes[index] = last_node;
202 /* Remove element from hash table. */
203 encoder->map->remove (node);
204 return true;
208 /* Return TRUE if we should encode the body of NODE (if any). */
210 bool
211 lto_symtab_encoder_encode_body_p (lto_symtab_encoder_t encoder,
212 struct cgraph_node *node)
214 int index = lto_symtab_encoder_lookup (encoder, node);
215 return encoder->nodes[index].body;
218 /* Specify that we encode the body of NODE in this partition. */
220 static void
221 lto_set_symtab_encoder_encode_body (lto_symtab_encoder_t encoder,
222 struct cgraph_node *node)
224 int index = lto_symtab_encoder_encode (encoder, node);
225 gcc_checking_assert (encoder->nodes[index].node == node);
226 encoder->nodes[index].body = true;
229 /* Return TRUE if we should encode initializer of NODE (if any). */
231 bool
232 lto_symtab_encoder_encode_initializer_p (lto_symtab_encoder_t encoder,
233 varpool_node *node)
235 int index = lto_symtab_encoder_lookup (encoder, node);
236 if (index == LCC_NOT_FOUND)
237 return false;
238 return encoder->nodes[index].initializer;
241 /* Specify that we should encode initializer of NODE (if any). */
243 static void
244 lto_set_symtab_encoder_encode_initializer (lto_symtab_encoder_t encoder,
245 varpool_node *node)
247 int index = lto_symtab_encoder_lookup (encoder, node);
248 encoder->nodes[index].initializer = true;
251 /* Return TRUE if NODE is in this partition. */
253 bool
254 lto_symtab_encoder_in_partition_p (lto_symtab_encoder_t encoder,
255 symtab_node *node)
257 int index = lto_symtab_encoder_lookup (encoder, node);
258 if (index == LCC_NOT_FOUND)
259 return false;
260 return encoder->nodes[index].in_partition;
263 /* Specify that NODE is in this partition. */
265 void
266 lto_set_symtab_encoder_in_partition (lto_symtab_encoder_t encoder,
267 symtab_node *node)
269 int index = lto_symtab_encoder_encode (encoder, node);
270 encoder->nodes[index].in_partition = true;
273 /* Output the cgraph EDGE to OB using ENCODER. */
275 static void
276 lto_output_edge (struct lto_simple_output_block *ob, struct cgraph_edge *edge,
277 lto_symtab_encoder_t encoder)
279 unsigned int uid;
280 intptr_t ref;
281 struct bitpack_d bp;
283 if (edge->indirect_unknown_callee)
284 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
285 LTO_symtab_indirect_edge);
286 else
287 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
288 LTO_symtab_edge);
290 ref = lto_symtab_encoder_lookup (encoder, edge->caller);
291 gcc_assert (ref != LCC_NOT_FOUND);
292 streamer_write_hwi_stream (ob->main_stream, ref);
294 if (!edge->indirect_unknown_callee)
296 ref = lto_symtab_encoder_lookup (encoder, edge->callee);
297 gcc_assert (ref != LCC_NOT_FOUND);
298 streamer_write_hwi_stream (ob->main_stream, ref);
301 streamer_write_gcov_count_stream (ob->main_stream, edge->count);
303 bp = bitpack_create (ob->main_stream);
304 uid = (!gimple_has_body_p (edge->caller->decl)
305 ? edge->lto_stmt_uid : gimple_uid (edge->call_stmt) + 1);
306 bp_pack_enum (&bp, cgraph_inline_failed_t,
307 CIF_N_REASONS, edge->inline_failed);
308 bp_pack_var_len_unsigned (&bp, uid);
309 bp_pack_var_len_unsigned (&bp, edge->frequency);
310 bp_pack_value (&bp, edge->indirect_inlining_edge, 1);
311 bp_pack_value (&bp, edge->speculative, 1);
312 bp_pack_value (&bp, edge->call_stmt_cannot_inline_p, 1);
313 bp_pack_value (&bp, edge->can_throw_external, 1);
314 bp_pack_value (&bp, edge->in_polymorphic_cdtor, 1);
315 if (edge->indirect_unknown_callee)
317 int flags = edge->indirect_info->ecf_flags;
318 bp_pack_value (&bp, (flags & ECF_CONST) != 0, 1);
319 bp_pack_value (&bp, (flags & ECF_PURE) != 0, 1);
320 bp_pack_value (&bp, (flags & ECF_NORETURN) != 0, 1);
321 bp_pack_value (&bp, (flags & ECF_MALLOC) != 0, 1);
322 bp_pack_value (&bp, (flags & ECF_NOTHROW) != 0, 1);
323 bp_pack_value (&bp, (flags & ECF_RETURNS_TWICE) != 0, 1);
324 /* Flags that should not appear on indirect calls. */
325 gcc_assert (!(flags & (ECF_LOOPING_CONST_OR_PURE
326 | ECF_MAY_BE_ALLOCA
327 | ECF_SIBCALL
328 | ECF_LEAF
329 | ECF_NOVOPS)));
331 streamer_write_bitpack (&bp);
332 if (edge->indirect_unknown_callee)
334 streamer_write_hwi_stream (ob->main_stream,
335 edge->indirect_info->common_target_id);
336 if (edge->indirect_info->common_target_id)
337 streamer_write_hwi_stream
338 (ob->main_stream, edge->indirect_info->common_target_probability);
342 /* Return if NODE contain references from other partitions. */
344 bool
345 referenced_from_other_partition_p (symtab_node *node, lto_symtab_encoder_t encoder)
347 int i;
348 struct ipa_ref *ref = NULL;
350 for (i = 0; node->iterate_referring (i, ref); i++)
352 /* Ignore references from non-offloadable nodes while streaming NODE into
353 offload LTO section. */
354 if (!ref->referring->need_lto_streaming)
355 continue;
357 if (ref->referring->in_other_partition
358 || !lto_symtab_encoder_in_partition_p (encoder, ref->referring))
359 return true;
361 return false;
364 /* Return true when node is reachable from other partition. */
366 bool
367 reachable_from_other_partition_p (struct cgraph_node *node, lto_symtab_encoder_t encoder)
369 struct cgraph_edge *e;
370 if (!node->definition)
371 return false;
372 if (node->global.inlined_to)
373 return false;
374 for (e = node->callers; e; e = e->next_caller)
376 /* Ignore references from non-offloadable nodes while streaming NODE into
377 offload LTO section. */
378 if (!e->caller->need_lto_streaming)
379 continue;
381 if (e->caller->in_other_partition
382 || !lto_symtab_encoder_in_partition_p (encoder, e->caller))
383 return true;
385 return false;
388 /* Return if NODE contain references from other partitions. */
390 bool
391 referenced_from_this_partition_p (symtab_node *node,
392 lto_symtab_encoder_t encoder)
394 int i;
395 struct ipa_ref *ref = NULL;
397 for (i = 0; node->iterate_referring (i, ref); i++)
398 if (lto_symtab_encoder_in_partition_p (encoder, ref->referring))
399 return true;
400 return false;
403 /* Return true when node is reachable from other partition. */
405 bool
406 reachable_from_this_partition_p (struct cgraph_node *node, lto_symtab_encoder_t encoder)
408 struct cgraph_edge *e;
409 for (e = node->callers; e; e = e->next_caller)
410 if (lto_symtab_encoder_in_partition_p (encoder, e->caller))
411 return true;
412 return false;
415 /* Output the cgraph NODE to OB. ENCODER is used to find the
416 reference number of NODE->inlined_to. SET is the set of nodes we
417 are writing to the current file. If NODE is not in SET, then NODE
418 is a boundary of a cgraph_node_set and we pretend NODE just has a
419 decl and no callees. WRITTEN_DECLS is the set of FUNCTION_DECLs
420 that have had their callgraph node written so far. This is used to
421 determine if NODE is a clone of a previously written node. */
423 static void
424 lto_output_node (struct lto_simple_output_block *ob, struct cgraph_node *node,
425 lto_symtab_encoder_t encoder)
427 unsigned int tag;
428 struct bitpack_d bp;
429 bool boundary_p;
430 intptr_t ref;
431 bool in_other_partition = false;
432 struct cgraph_node *clone_of, *ultimate_clone_of;
433 ipa_opt_pass_d *pass;
434 int i;
435 const char *comdat;
436 const char *section;
437 tree group;
439 boundary_p = !lto_symtab_encoder_in_partition_p (encoder, node);
441 if (node->analyzed && (!boundary_p || node->alias || node->thunk.thunk_p))
442 tag = LTO_symtab_analyzed_node;
443 else
444 tag = LTO_symtab_unavail_node;
446 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
447 tag);
448 streamer_write_hwi_stream (ob->main_stream, node->order);
450 /* In WPA mode, we only output part of the call-graph. Also, we
451 fake cgraph node attributes. There are two cases that we care.
453 Boundary nodes: There are nodes that are not part of SET but are
454 called from within SET. We artificially make them look like
455 externally visible nodes with no function body.
457 Cherry-picked nodes: These are nodes we pulled from other
458 translation units into SET during IPA-inlining. We make them as
459 local static nodes to prevent clashes with other local statics. */
460 if (boundary_p && node->analyzed
461 && node->get_partitioning_class () == SYMBOL_PARTITION)
463 /* Inline clones can not be part of boundary.
464 gcc_assert (!node->global.inlined_to);
466 FIXME: At the moment they can be, when partition contains an inline
467 clone that is clone of inline clone from outside partition. We can
468 reshape the clone tree and make other tree to be the root, but it
469 needs a bit extra work and will be promplty done by cgraph_remove_node
470 after reading back. */
471 in_other_partition = 1;
474 clone_of = node->clone_of;
475 while (clone_of
476 && (ref = lto_symtab_encoder_lookup (encoder, clone_of)) == LCC_NOT_FOUND)
477 if (clone_of->prev_sibling_clone)
478 clone_of = clone_of->prev_sibling_clone;
479 else
480 clone_of = clone_of->clone_of;
482 /* See if body of the master function is output. If not, we are seeing only
483 an declaration and we do not need to pass down clone tree. */
484 ultimate_clone_of = clone_of;
485 while (ultimate_clone_of && ultimate_clone_of->clone_of)
486 ultimate_clone_of = ultimate_clone_of->clone_of;
488 if (clone_of && !lto_symtab_encoder_encode_body_p (encoder, ultimate_clone_of))
489 clone_of = NULL;
491 if (tag == LTO_symtab_analyzed_node)
492 gcc_assert (clone_of || !node->clone_of);
493 if (!clone_of)
494 streamer_write_hwi_stream (ob->main_stream, LCC_NOT_FOUND);
495 else
496 streamer_write_hwi_stream (ob->main_stream, ref);
499 lto_output_fn_decl_index (ob->decl_state, ob->main_stream, node->decl);
500 streamer_write_gcov_count_stream (ob->main_stream, node->count);
501 streamer_write_hwi_stream (ob->main_stream, node->count_materialization_scale);
503 streamer_write_hwi_stream (ob->main_stream,
504 node->ipa_transforms_to_apply.length ());
505 FOR_EACH_VEC_ELT (node->ipa_transforms_to_apply, i, pass)
506 streamer_write_hwi_stream (ob->main_stream, pass->static_pass_number);
508 if (tag == LTO_symtab_analyzed_node)
510 if (node->global.inlined_to)
512 ref = lto_symtab_encoder_lookup (encoder, node->global.inlined_to);
513 gcc_assert (ref != LCC_NOT_FOUND);
515 else
516 ref = LCC_NOT_FOUND;
518 streamer_write_hwi_stream (ob->main_stream, ref);
521 group = node->get_comdat_group ();
522 if (group)
523 comdat = IDENTIFIER_POINTER (group);
524 else
525 comdat = "";
526 streamer_write_data_stream (ob->main_stream, comdat, strlen (comdat) + 1);
528 if (group)
530 if (node->same_comdat_group && !boundary_p)
532 ref = lto_symtab_encoder_lookup (encoder,
533 node->same_comdat_group);
534 gcc_assert (ref != LCC_NOT_FOUND);
536 else
537 ref = LCC_NOT_FOUND;
538 streamer_write_hwi_stream (ob->main_stream, ref);
541 section = node->get_section ();
542 if (!section)
543 section = "";
545 streamer_write_hwi_stream (ob->main_stream, node->tp_first_run);
547 bp = bitpack_create (ob->main_stream);
548 bp_pack_value (&bp, node->local.local, 1);
549 bp_pack_value (&bp, node->externally_visible, 1);
550 bp_pack_value (&bp, node->no_reorder, 1);
551 bp_pack_value (&bp, node->definition, 1);
552 bp_pack_value (&bp, node->local.versionable, 1);
553 bp_pack_value (&bp, node->local.can_change_signature, 1);
554 bp_pack_value (&bp, node->local.redefined_extern_inline, 1);
555 bp_pack_value (&bp, node->force_output, 1);
556 bp_pack_value (&bp, node->forced_by_abi, 1);
557 bp_pack_value (&bp, node->unique_name, 1);
558 bp_pack_value (&bp, node->body_removed, 1);
559 bp_pack_value (&bp, node->implicit_section, 1);
560 bp_pack_value (&bp, node->address_taken, 1);
561 bp_pack_value (&bp, tag == LTO_symtab_analyzed_node
562 && node->get_partitioning_class () == SYMBOL_PARTITION
563 && (reachable_from_other_partition_p (node, encoder)
564 || referenced_from_other_partition_p (node, encoder)), 1);
565 bp_pack_value (&bp, node->lowered, 1);
566 bp_pack_value (&bp, in_other_partition, 1);
567 bp_pack_value (&bp, node->alias, 1);
568 bp_pack_value (&bp, node->weakref, 1);
569 bp_pack_value (&bp, node->frequency, 2);
570 bp_pack_value (&bp, node->only_called_at_startup, 1);
571 bp_pack_value (&bp, node->only_called_at_exit, 1);
572 bp_pack_value (&bp, node->tm_clone, 1);
573 bp_pack_value (&bp, node->calls_comdat_local, 1);
574 bp_pack_value (&bp, node->icf_merged, 1);
575 bp_pack_value (&bp, node->nonfreeing_fn, 1);
576 bp_pack_value (&bp, node->thunk.thunk_p, 1);
577 bp_pack_value (&bp, node->parallelized_function, 1);
578 bp_pack_enum (&bp, ld_plugin_symbol_resolution,
579 LDPR_NUM_KNOWN, node->resolution);
580 bp_pack_value (&bp, node->instrumentation_clone, 1);
581 streamer_write_bitpack (&bp);
582 streamer_write_data_stream (ob->main_stream, section, strlen (section) + 1);
584 if (node->thunk.thunk_p)
586 streamer_write_uhwi_stream
587 (ob->main_stream,
588 1 + (node->thunk.this_adjusting != 0) * 2
589 + (node->thunk.virtual_offset_p != 0) * 4
590 + (node->thunk.add_pointer_bounds_args != 0) * 8);
591 streamer_write_uhwi_stream (ob->main_stream, node->thunk.fixed_offset);
592 streamer_write_uhwi_stream (ob->main_stream, node->thunk.virtual_value);
594 streamer_write_hwi_stream (ob->main_stream, node->profile_id);
595 if (DECL_STATIC_CONSTRUCTOR (node->decl))
596 streamer_write_hwi_stream (ob->main_stream, node->get_init_priority ());
597 if (DECL_STATIC_DESTRUCTOR (node->decl))
598 streamer_write_hwi_stream (ob->main_stream, node->get_fini_priority ());
600 if (node->instrumentation_clone)
601 lto_output_fn_decl_index (ob->decl_state, ob->main_stream, node->orig_decl);
604 /* Output the varpool NODE to OB.
605 If NODE is not in SET, then NODE is a boundary. */
607 static void
608 lto_output_varpool_node (struct lto_simple_output_block *ob, varpool_node *node,
609 lto_symtab_encoder_t encoder)
611 bool boundary_p = !lto_symtab_encoder_in_partition_p (encoder, node);
612 bool encode_initializer_p
613 = (node->definition
614 && lto_symtab_encoder_encode_initializer_p (encoder, node));
615 struct bitpack_d bp;
616 int ref;
617 const char *comdat;
618 const char *section;
619 tree group;
621 gcc_assert (!encode_initializer_p || node->definition);
622 gcc_assert (boundary_p || encode_initializer_p);
624 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
625 LTO_symtab_variable);
626 streamer_write_hwi_stream (ob->main_stream, node->order);
627 lto_output_var_decl_index (ob->decl_state, ob->main_stream, node->decl);
628 bp = bitpack_create (ob->main_stream);
629 bp_pack_value (&bp, node->externally_visible, 1);
630 bp_pack_value (&bp, node->no_reorder, 1);
631 bp_pack_value (&bp, node->force_output, 1);
632 bp_pack_value (&bp, node->forced_by_abi, 1);
633 bp_pack_value (&bp, node->unique_name, 1);
634 bp_pack_value (&bp,
635 node->body_removed
636 || (!encode_initializer_p && !node->alias && node->definition),
638 bp_pack_value (&bp, node->implicit_section, 1);
639 bp_pack_value (&bp, node->writeonly, 1);
640 bp_pack_value (&bp, node->definition && (encode_initializer_p || node->alias),
642 bp_pack_value (&bp, node->alias, 1);
643 bp_pack_value (&bp, node->weakref, 1);
644 bp_pack_value (&bp, node->analyzed && !boundary_p, 1);
645 gcc_assert (node->definition || !node->analyzed);
646 /* Constant pool initializers can be de-unified into individual ltrans units.
647 FIXME: Alternatively at -Os we may want to avoid generating for them the local
648 labels and share them across LTRANS partitions. */
649 if (node->get_partitioning_class () != SYMBOL_PARTITION)
651 bp_pack_value (&bp, 0, 1); /* used_from_other_parition. */
652 bp_pack_value (&bp, 0, 1); /* in_other_partition. */
654 else
656 bp_pack_value (&bp, node->definition
657 && referenced_from_other_partition_p (node, encoder), 1);
658 bp_pack_value (&bp, node->analyzed
659 && boundary_p && !DECL_EXTERNAL (node->decl), 1);
660 /* in_other_partition. */
662 bp_pack_value (&bp, node->tls_model, 3);
663 bp_pack_value (&bp, node->used_by_single_function, 1);
664 bp_pack_value (&bp, node->need_bounds_init, 1);
665 streamer_write_bitpack (&bp);
667 group = node->get_comdat_group ();
668 if (group)
669 comdat = IDENTIFIER_POINTER (group);
670 else
671 comdat = "";
672 streamer_write_data_stream (ob->main_stream, comdat, strlen (comdat) + 1);
674 if (group)
676 if (node->same_comdat_group && !boundary_p)
678 ref = lto_symtab_encoder_lookup (encoder,
679 node->same_comdat_group);
680 gcc_assert (ref != LCC_NOT_FOUND);
682 else
683 ref = LCC_NOT_FOUND;
684 streamer_write_hwi_stream (ob->main_stream, ref);
687 section = node->get_section ();
688 if (!section)
689 section = "";
690 streamer_write_data_stream (ob->main_stream, section, strlen (section) + 1);
692 streamer_write_enum (ob->main_stream, ld_plugin_symbol_resolution,
693 LDPR_NUM_KNOWN, node->resolution);
696 /* Output the varpool NODE to OB.
697 If NODE is not in SET, then NODE is a boundary. */
699 static void
700 lto_output_ref (struct lto_simple_output_block *ob, struct ipa_ref *ref,
701 lto_symtab_encoder_t encoder)
703 struct bitpack_d bp;
704 int nref;
705 int uid = ref->lto_stmt_uid;
706 struct cgraph_node *node;
708 bp = bitpack_create (ob->main_stream);
709 bp_pack_value (&bp, ref->use, 3);
710 bp_pack_value (&bp, ref->speculative, 1);
711 streamer_write_bitpack (&bp);
712 nref = lto_symtab_encoder_lookup (encoder, ref->referred);
713 gcc_assert (nref != LCC_NOT_FOUND);
714 streamer_write_hwi_stream (ob->main_stream, nref);
716 node = dyn_cast <cgraph_node *> (ref->referring);
717 if (node)
719 if (ref->stmt)
720 uid = gimple_uid (ref->stmt) + 1;
721 streamer_write_hwi_stream (ob->main_stream, uid);
725 /* Stream out profile_summary to OB. */
727 static void
728 output_profile_summary (struct lto_simple_output_block *ob)
730 unsigned h_ix;
731 struct bitpack_d bp;
733 if (profile_info)
735 /* We do not output num and run_max, they are not used by
736 GCC profile feedback and they are difficult to merge from multiple
737 units. */
738 gcc_assert (profile_info->runs);
739 streamer_write_uhwi_stream (ob->main_stream, profile_info->runs);
740 streamer_write_gcov_count_stream (ob->main_stream, profile_info->sum_max);
742 /* sum_all is needed for computing the working set with the
743 histogram. */
744 streamer_write_gcov_count_stream (ob->main_stream, profile_info->sum_all);
746 /* Create and output a bitpack of non-zero histogram entries indices. */
747 bp = bitpack_create (ob->main_stream);
748 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
749 bp_pack_value (&bp, profile_info->histogram[h_ix].num_counters > 0, 1);
750 streamer_write_bitpack (&bp);
751 /* Now stream out only those non-zero entries. */
752 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
754 if (!profile_info->histogram[h_ix].num_counters)
755 continue;
756 streamer_write_gcov_count_stream (ob->main_stream,
757 profile_info->histogram[h_ix].num_counters);
758 streamer_write_gcov_count_stream (ob->main_stream,
759 profile_info->histogram[h_ix].min_value);
760 streamer_write_gcov_count_stream (ob->main_stream,
761 profile_info->histogram[h_ix].cum_value);
763 /* IPA-profile computes hot bb threshold based on cumulated
764 whole program profile. We need to stream it down to ltrans. */
765 if (flag_wpa)
766 streamer_write_gcov_count_stream (ob->main_stream,
767 get_hot_bb_threshold ());
769 else
770 streamer_write_uhwi_stream (ob->main_stream, 0);
773 /* Output all callees or indirect outgoing edges. EDGE must be the first such
774 edge. */
776 static void
777 output_outgoing_cgraph_edges (struct cgraph_edge *edge,
778 struct lto_simple_output_block *ob,
779 lto_symtab_encoder_t encoder)
781 if (!edge)
782 return;
784 /* Output edges in backward direction, so the reconstructed callgraph match
785 and it is easy to associate call sites in the IPA pass summaries. */
786 while (edge->next_callee)
787 edge = edge->next_callee;
788 for (; edge; edge = edge->prev_callee)
789 lto_output_edge (ob, edge, encoder);
792 /* Output the part of the cgraph in SET. */
794 static void
795 output_refs (lto_symtab_encoder_t encoder)
797 struct lto_simple_output_block *ob;
798 int count;
799 struct ipa_ref *ref;
801 ob = lto_create_simple_output_block (LTO_section_refs);
803 for (int i = 0; i < lto_symtab_encoder_size (encoder); i++)
805 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
807 if (!node->alias && !lto_symtab_encoder_in_partition_p (encoder, node))
808 continue;
810 count = node->ref_list.nreferences ();
811 if (count)
813 streamer_write_gcov_count_stream (ob->main_stream, count);
814 streamer_write_uhwi_stream (ob->main_stream,
815 lto_symtab_encoder_lookup (encoder, node));
816 for (int i = 0; node->iterate_reference (i, ref); i++)
817 lto_output_ref (ob, ref, encoder);
821 streamer_write_uhwi_stream (ob->main_stream, 0);
823 lto_destroy_simple_output_block (ob);
826 /* Add NODE into encoder as well as nodes it is cloned from.
827 Do it in a way so clones appear first. */
829 static void
830 add_node_to (lto_symtab_encoder_t encoder, struct cgraph_node *node,
831 bool include_body)
833 if (node->clone_of)
834 add_node_to (encoder, node->clone_of, include_body);
835 else if (include_body)
836 lto_set_symtab_encoder_encode_body (encoder, node);
837 lto_symtab_encoder_encode (encoder, node);
840 /* Add all references in NODE to encoders. */
842 static void
843 create_references (lto_symtab_encoder_t encoder, symtab_node *node)
845 int i;
846 struct ipa_ref *ref = NULL;
847 for (i = 0; node->iterate_reference (i, ref); i++)
848 if (is_a <cgraph_node *> (ref->referred))
849 add_node_to (encoder, dyn_cast <cgraph_node *> (ref->referred), false);
850 else
851 lto_symtab_encoder_encode (encoder, ref->referred);
854 /* Select what needs to be streamed out. In regular lto mode stream everything.
855 In offload lto mode stream only nodes marked as offloadable. */
856 void
857 select_what_to_stream (void)
859 struct symtab_node *snode;
860 FOR_EACH_SYMBOL (snode)
861 snode->need_lto_streaming = !lto_stream_offload_p || snode->offloadable;
864 /* Find all symbols we want to stream into given partition and insert them
865 to encoders.
867 The function actually replaces IN_ENCODER by new one. The reason is that
868 streaming code needs clone's origin to be streamed before clone. This
869 means that we need to insert the nodes in specific order. This order is
870 ignored by the partitioning logic earlier. */
872 lto_symtab_encoder_t
873 compute_ltrans_boundary (lto_symtab_encoder_t in_encoder)
875 struct cgraph_edge *edge;
876 int i;
877 lto_symtab_encoder_t encoder;
878 lto_symtab_encoder_iterator lsei;
879 hash_set<void *> reachable_call_targets;
881 encoder = lto_symtab_encoder_new (false);
883 /* Go over all entries in the IN_ENCODER and duplicate them to
884 ENCODER. At the same time insert masters of clones so
885 every master appears before clone. */
886 for (lsei = lsei_start_function_in_partition (in_encoder);
887 !lsei_end_p (lsei); lsei_next_function_in_partition (&lsei))
889 struct cgraph_node *node = lsei_cgraph_node (lsei);
890 if (!node->need_lto_streaming)
891 continue;
892 add_node_to (encoder, node, true);
893 lto_set_symtab_encoder_in_partition (encoder, node);
894 create_references (encoder, node);
895 /* For proper debug info, we need to ship the origins, too. */
896 if (DECL_ABSTRACT_ORIGIN (node->decl))
898 struct cgraph_node *origin_node
899 = cgraph_node::get_create (DECL_ABSTRACT_ORIGIN (node->decl));
900 origin_node->used_as_abstract_origin = true;
901 add_node_to (encoder, origin_node, true);
904 for (lsei = lsei_start_variable_in_partition (in_encoder);
905 !lsei_end_p (lsei); lsei_next_variable_in_partition (&lsei))
907 varpool_node *vnode = lsei_varpool_node (lsei);
909 if (!vnode->need_lto_streaming)
910 continue;
911 lto_set_symtab_encoder_in_partition (encoder, vnode);
912 lto_set_symtab_encoder_encode_initializer (encoder, vnode);
913 create_references (encoder, vnode);
914 /* For proper debug info, we need to ship the origins, too. */
915 if (DECL_ABSTRACT_ORIGIN (vnode->decl))
917 varpool_node *origin_node
918 = varpool_node::get (DECL_ABSTRACT_ORIGIN (vnode->decl));
919 lto_set_symtab_encoder_in_partition (encoder, origin_node);
922 /* Pickle in also the initializer of all referenced readonly variables
923 to help folding. Constant pool variables are not shared, so we must
924 pickle those too. */
925 for (i = 0; i < lto_symtab_encoder_size (encoder); i++)
927 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
928 if (varpool_node *vnode = dyn_cast <varpool_node *> (node))
930 if (!lto_symtab_encoder_encode_initializer_p (encoder,
931 vnode)
932 && (((vnode->ctor_useable_for_folding_p ()
933 && (!DECL_VIRTUAL_P (vnode->decl)
934 || !flag_wpa
935 || flag_ltrans_devirtualize))
936 || POINTER_BOUNDS_P (vnode->decl))))
938 lto_set_symtab_encoder_encode_initializer (encoder, vnode);
939 create_references (encoder, vnode);
944 /* Go over all the nodes again to include callees that are not in
945 SET. */
946 for (lsei = lsei_start_function_in_partition (encoder);
947 !lsei_end_p (lsei); lsei_next_function_in_partition (&lsei))
949 struct cgraph_node *node = lsei_cgraph_node (lsei);
950 for (edge = node->callees; edge; edge = edge->next_callee)
952 struct cgraph_node *callee = edge->callee;
953 if (!lto_symtab_encoder_in_partition_p (encoder, callee))
955 /* We should have moved all the inlines. */
956 gcc_assert (!callee->global.inlined_to);
957 add_node_to (encoder, callee, false);
960 /* Add all possible targets for late devirtualization. */
961 if (flag_ltrans_devirtualize || !flag_wpa)
962 for (edge = node->indirect_calls; edge; edge = edge->next_callee)
963 if (edge->indirect_info->polymorphic)
965 unsigned int i;
966 void *cache_token;
967 bool final;
968 vec <cgraph_node *>targets
969 = possible_polymorphic_call_targets
970 (edge, &final, &cache_token);
971 if (!reachable_call_targets.add (cache_token))
973 for (i = 0; i < targets.length (); i++)
975 struct cgraph_node *callee = targets[i];
977 /* Adding an external declarations into the unit serves
978 no purpose and just increases its boundary. */
979 if (callee->definition
980 && !lto_symtab_encoder_in_partition_p
981 (encoder, callee))
983 gcc_assert (!callee->global.inlined_to);
984 add_node_to (encoder, callee, false);
990 /* Be sure to also insert alias targert and thunk callees. These needs
991 to stay to aid local calling conventions. */
992 for (i = 0; i < lto_symtab_encoder_size (encoder); i++)
994 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
995 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
997 if (node->alias && node->analyzed)
998 create_references (encoder, node);
999 if (cnode
1000 && cnode->thunk.thunk_p)
1001 add_node_to (encoder, cnode->callees->callee, false);
1003 lto_symtab_encoder_delete (in_encoder);
1004 return encoder;
1007 /* Output the part of the symtab in SET and VSET. */
1009 void
1010 output_symtab (void)
1012 struct cgraph_node *node;
1013 struct lto_simple_output_block *ob;
1014 int i, n_nodes;
1015 lto_symtab_encoder_t encoder;
1017 if (flag_wpa)
1018 output_cgraph_opt_summary ();
1020 ob = lto_create_simple_output_block (LTO_section_symtab_nodes);
1022 output_profile_summary (ob);
1024 /* An encoder for cgraph nodes should have been created by
1025 ipa_write_summaries_1. */
1026 gcc_assert (ob->decl_state->symtab_node_encoder);
1027 encoder = ob->decl_state->symtab_node_encoder;
1029 /* Write out the nodes. We must first output a node and then its clones,
1030 otherwise at a time reading back the node there would be nothing to clone
1031 from. */
1032 n_nodes = lto_symtab_encoder_size (encoder);
1033 for (i = 0; i < n_nodes; i++)
1035 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
1036 if (cgraph_node *cnode = dyn_cast <cgraph_node *> (node))
1037 lto_output_node (ob, cnode, encoder);
1038 else
1039 lto_output_varpool_node (ob, dyn_cast<varpool_node *> (node), encoder);
1042 /* Go over the nodes in SET again to write edges. */
1043 for (int i = 0; i < lto_symtab_encoder_size (encoder); i++)
1045 node = dyn_cast <cgraph_node *> (lto_symtab_encoder_deref (encoder, i));
1046 if (node
1047 && (node->thunk.thunk_p
1048 || lto_symtab_encoder_in_partition_p (encoder, node)))
1050 output_outgoing_cgraph_edges (node->callees, ob, encoder);
1051 output_outgoing_cgraph_edges (node->indirect_calls, ob, encoder);
1055 streamer_write_uhwi_stream (ob->main_stream, 0);
1057 lto_destroy_simple_output_block (ob);
1059 /* Emit toplevel asms.
1060 When doing WPA we must output every asm just once. Since we do not partition asm
1061 nodes at all, output them to first output. This is kind of hack, but should work
1062 well. */
1063 if (!asm_nodes_output)
1065 asm_nodes_output = true;
1066 lto_output_toplevel_asms ();
1069 output_refs (encoder);
1072 /* Return identifier encoded in IB as a plain string. */
1074 static tree
1075 read_identifier (struct lto_input_block *ib)
1077 unsigned int len = strnlen (ib->data + ib->p, ib->len - ib->p - 1);
1078 tree id;
1080 if (ib->data[ib->p + len])
1081 lto_section_overrun (ib);
1082 if (!len)
1084 ib->p++;
1085 return NULL;
1087 id = get_identifier (ib->data + ib->p);
1088 ib->p += len + 1;
1089 return id;
1092 /* Return string encoded in IB, NULL if string is empty. */
1094 static const char *
1095 read_string (struct lto_input_block *ib)
1097 unsigned int len = strnlen (ib->data + ib->p, ib->len - ib->p - 1);
1098 const char *str;
1100 if (ib->data[ib->p + len])
1101 lto_section_overrun (ib);
1102 if (!len)
1104 ib->p++;
1105 return NULL;
1107 str = ib->data + ib->p;
1108 ib->p += len + 1;
1109 return str;
1112 /* Output function/variable tables that will allow libgomp to look up offload
1113 target code.
1114 OFFLOAD_FUNCS is filled in expand_omp_target, OFFLOAD_VARS is filled in
1115 varpool_node::get_create. In WHOPR (partitioned) mode during the WPA stage
1116 both OFFLOAD_FUNCS and OFFLOAD_VARS are filled by input_offload_tables. */
1118 void
1119 output_offload_tables (void)
1121 if (vec_safe_is_empty (offload_funcs) && vec_safe_is_empty (offload_vars))
1122 return;
1124 struct lto_simple_output_block *ob
1125 = lto_create_simple_output_block (LTO_section_offload_table);
1127 for (unsigned i = 0; i < vec_safe_length (offload_funcs); i++)
1129 streamer_write_enum (ob->main_stream, LTO_symtab_tags,
1130 LTO_symtab_last_tag, LTO_symtab_unavail_node);
1131 lto_output_fn_decl_index (ob->decl_state, ob->main_stream,
1132 (*offload_funcs)[i]);
1135 for (unsigned i = 0; i < vec_safe_length (offload_vars); i++)
1137 streamer_write_enum (ob->main_stream, LTO_symtab_tags,
1138 LTO_symtab_last_tag, LTO_symtab_variable);
1139 lto_output_var_decl_index (ob->decl_state, ob->main_stream,
1140 (*offload_vars)[i]);
1143 streamer_write_uhwi_stream (ob->main_stream, 0);
1144 lto_destroy_simple_output_block (ob);
1146 /* In WHOPR mode during the WPA stage the joint offload tables need to be
1147 streamed to one partition only. That's why we free offload_funcs and
1148 offload_vars after the first call of output_offload_tables. */
1149 if (flag_wpa)
1151 vec_free (offload_funcs);
1152 vec_free (offload_vars);
1156 /* Overwrite the information in NODE based on FILE_DATA, TAG, FLAGS,
1157 STACK_SIZE, SELF_TIME and SELF_SIZE. This is called either to initialize
1158 NODE or to replace the values in it, for instance because the first
1159 time we saw it, the function body was not available but now it
1160 is. BP is a bitpack with all the bitflags for NODE read from the
1161 stream. */
1163 static void
1164 input_overwrite_node (struct lto_file_decl_data *file_data,
1165 struct cgraph_node *node,
1166 enum LTO_symtab_tags tag,
1167 struct bitpack_d *bp)
1169 node->aux = (void *) tag;
1170 node->lto_file_data = file_data;
1172 node->local.local = bp_unpack_value (bp, 1);
1173 node->externally_visible = bp_unpack_value (bp, 1);
1174 node->no_reorder = bp_unpack_value (bp, 1);
1175 node->definition = bp_unpack_value (bp, 1);
1176 node->local.versionable = bp_unpack_value (bp, 1);
1177 node->local.can_change_signature = bp_unpack_value (bp, 1);
1178 node->local.redefined_extern_inline = bp_unpack_value (bp, 1);
1179 node->force_output = bp_unpack_value (bp, 1);
1180 node->forced_by_abi = bp_unpack_value (bp, 1);
1181 node->unique_name = bp_unpack_value (bp, 1);
1182 node->body_removed = bp_unpack_value (bp, 1);
1183 node->implicit_section = bp_unpack_value (bp, 1);
1184 node->address_taken = bp_unpack_value (bp, 1);
1185 node->used_from_other_partition = bp_unpack_value (bp, 1);
1186 node->lowered = bp_unpack_value (bp, 1);
1187 node->analyzed = tag == LTO_symtab_analyzed_node;
1188 node->in_other_partition = bp_unpack_value (bp, 1);
1189 if (node->in_other_partition
1190 /* Avoid updating decl when we are seeing just inline clone.
1191 When inlining function that has functions already inlined into it,
1192 we produce clones of inline clones.
1194 WPA partitioning might put each clone into different unit and
1195 we might end up streaming inline clone from other partition
1196 to support clone we are interested in. */
1197 && (!node->clone_of
1198 || node->clone_of->decl != node->decl))
1200 DECL_EXTERNAL (node->decl) = 1;
1201 TREE_STATIC (node->decl) = 0;
1203 node->alias = bp_unpack_value (bp, 1);
1204 node->weakref = bp_unpack_value (bp, 1);
1205 node->frequency = (enum node_frequency)bp_unpack_value (bp, 2);
1206 node->only_called_at_startup = bp_unpack_value (bp, 1);
1207 node->only_called_at_exit = bp_unpack_value (bp, 1);
1208 node->tm_clone = bp_unpack_value (bp, 1);
1209 node->calls_comdat_local = bp_unpack_value (bp, 1);
1210 node->icf_merged = bp_unpack_value (bp, 1);
1211 node->nonfreeing_fn = bp_unpack_value (bp, 1);
1212 node->thunk.thunk_p = bp_unpack_value (bp, 1);
1213 node->parallelized_function = bp_unpack_value (bp, 1);
1214 node->resolution = bp_unpack_enum (bp, ld_plugin_symbol_resolution,
1215 LDPR_NUM_KNOWN);
1216 node->instrumentation_clone = bp_unpack_value (bp, 1);
1217 gcc_assert (flag_ltrans
1218 || (!node->in_other_partition
1219 && !node->used_from_other_partition));
1222 /* Return string alias is alias of. */
1224 static tree
1225 get_alias_symbol (tree decl)
1227 tree alias = lookup_attribute ("alias", DECL_ATTRIBUTES (decl));
1228 return get_identifier (TREE_STRING_POINTER
1229 (TREE_VALUE (TREE_VALUE (alias))));
1232 /* Read a node from input_block IB. TAG is the node's tag just read.
1233 Return the node read or overwriten. */
1235 static struct cgraph_node *
1236 input_node (struct lto_file_decl_data *file_data,
1237 struct lto_input_block *ib,
1238 enum LTO_symtab_tags tag,
1239 vec<symtab_node *> nodes)
1241 gcc::pass_manager *passes = g->get_passes ();
1242 tree fn_decl;
1243 struct cgraph_node *node;
1244 struct bitpack_d bp;
1245 unsigned decl_index;
1246 int ref = LCC_NOT_FOUND, ref2 = LCC_NOT_FOUND;
1247 int clone_ref;
1248 int order;
1249 int i, count;
1250 tree group;
1251 const char *section;
1252 order = streamer_read_hwi (ib) + order_base;
1253 clone_ref = streamer_read_hwi (ib);
1255 decl_index = streamer_read_uhwi (ib);
1256 fn_decl = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1258 if (clone_ref != LCC_NOT_FOUND)
1260 node = dyn_cast<cgraph_node *> (nodes[clone_ref])->create_clone (fn_decl,
1261 0, CGRAPH_FREQ_BASE, false,
1262 vNULL, false, NULL, NULL);
1264 else
1266 /* Declaration of functions can be already merged with a declaration
1267 from other input file. We keep cgraph unmerged until after streaming
1268 of ipa passes is done. Alays forcingly create a fresh node. */
1269 node = symtab->create_empty ();
1270 node->decl = fn_decl;
1271 node->register_symbol ();
1274 node->order = order;
1275 if (order >= symtab->order)
1276 symtab->order = order + 1;
1278 node->count = streamer_read_gcov_count (ib);
1279 node->count_materialization_scale = streamer_read_hwi (ib);
1281 count = streamer_read_hwi (ib);
1282 node->ipa_transforms_to_apply = vNULL;
1283 for (i = 0; i < count; i++)
1285 opt_pass *pass;
1286 int pid = streamer_read_hwi (ib);
1288 gcc_assert (pid < passes->passes_by_id_size);
1289 pass = passes->passes_by_id[pid];
1290 node->ipa_transforms_to_apply.safe_push ((ipa_opt_pass_d *) pass);
1293 if (tag == LTO_symtab_analyzed_node)
1294 ref = streamer_read_hwi (ib);
1296 group = read_identifier (ib);
1297 if (group)
1298 ref2 = streamer_read_hwi (ib);
1300 /* Make sure that we have not read this node before. Nodes that
1301 have already been read will have their tag stored in the 'aux'
1302 field. Since built-in functions can be referenced in multiple
1303 functions, they are expected to be read more than once. */
1304 if (node->aux && !DECL_BUILT_IN (node->decl))
1305 internal_error ("bytecode stream: found multiple instances of cgraph "
1306 "node with uid %d", node->uid);
1308 node->tp_first_run = streamer_read_uhwi (ib);
1310 bp = streamer_read_bitpack (ib);
1312 input_overwrite_node (file_data, node, tag, &bp);
1314 /* Store a reference for now, and fix up later to be a pointer. */
1315 node->global.inlined_to = (cgraph_node *) (intptr_t) ref;
1317 if (group)
1319 node->set_comdat_group (group);
1320 /* Store a reference for now, and fix up later to be a pointer. */
1321 node->same_comdat_group = (symtab_node *) (intptr_t) ref2;
1323 else
1324 node->same_comdat_group = (symtab_node *) (intptr_t) LCC_NOT_FOUND;
1325 section = read_string (ib);
1326 if (section)
1327 node->set_section_for_node (section);
1329 if (node->thunk.thunk_p)
1331 int type = streamer_read_uhwi (ib);
1332 HOST_WIDE_INT fixed_offset = streamer_read_uhwi (ib);
1333 HOST_WIDE_INT virtual_value = streamer_read_uhwi (ib);
1335 node->thunk.fixed_offset = fixed_offset;
1336 node->thunk.this_adjusting = (type & 2);
1337 node->thunk.virtual_value = virtual_value;
1338 node->thunk.virtual_offset_p = (type & 4);
1339 node->thunk.add_pointer_bounds_args = (type & 8);
1341 if (node->alias && !node->analyzed && node->weakref)
1342 node->alias_target = get_alias_symbol (node->decl);
1343 node->profile_id = streamer_read_hwi (ib);
1344 if (DECL_STATIC_CONSTRUCTOR (node->decl))
1345 node->set_init_priority (streamer_read_hwi (ib));
1346 if (DECL_STATIC_DESTRUCTOR (node->decl))
1347 node->set_fini_priority (streamer_read_hwi (ib));
1349 if (node->instrumentation_clone)
1351 decl_index = streamer_read_uhwi (ib);
1352 fn_decl = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1353 node->orig_decl = fn_decl;
1356 return node;
1359 /* Read a node from input_block IB. TAG is the node's tag just read.
1360 Return the node read or overwriten. */
1362 static varpool_node *
1363 input_varpool_node (struct lto_file_decl_data *file_data,
1364 struct lto_input_block *ib)
1366 int decl_index;
1367 tree var_decl;
1368 varpool_node *node;
1369 struct bitpack_d bp;
1370 int ref = LCC_NOT_FOUND;
1371 int order;
1372 tree group;
1373 const char *section;
1375 order = streamer_read_hwi (ib) + order_base;
1376 decl_index = streamer_read_uhwi (ib);
1377 var_decl = lto_file_decl_data_get_var_decl (file_data, decl_index);
1379 /* Declaration of functions can be already merged with a declaration
1380 from other input file. We keep cgraph unmerged until after streaming
1381 of ipa passes is done. Alays forcingly create a fresh node. */
1382 node = varpool_node::create_empty ();
1383 node->decl = var_decl;
1384 node->register_symbol ();
1386 node->order = order;
1387 if (order >= symtab->order)
1388 symtab->order = order + 1;
1389 node->lto_file_data = file_data;
1391 bp = streamer_read_bitpack (ib);
1392 node->externally_visible = bp_unpack_value (&bp, 1);
1393 node->no_reorder = bp_unpack_value (&bp, 1);
1394 node->force_output = bp_unpack_value (&bp, 1);
1395 node->forced_by_abi = bp_unpack_value (&bp, 1);
1396 node->unique_name = bp_unpack_value (&bp, 1);
1397 node->body_removed = bp_unpack_value (&bp, 1);
1398 node->implicit_section = bp_unpack_value (&bp, 1);
1399 node->writeonly = bp_unpack_value (&bp, 1);
1400 node->definition = bp_unpack_value (&bp, 1);
1401 node->alias = bp_unpack_value (&bp, 1);
1402 node->weakref = bp_unpack_value (&bp, 1);
1403 node->analyzed = bp_unpack_value (&bp, 1);
1404 node->used_from_other_partition = bp_unpack_value (&bp, 1);
1405 node->in_other_partition = bp_unpack_value (&bp, 1);
1406 if (node->in_other_partition)
1408 DECL_EXTERNAL (node->decl) = 1;
1409 TREE_STATIC (node->decl) = 0;
1411 if (node->alias && !node->analyzed && node->weakref)
1412 node->alias_target = get_alias_symbol (node->decl);
1413 node->tls_model = (enum tls_model)bp_unpack_value (&bp, 3);
1414 node->used_by_single_function = (enum tls_model)bp_unpack_value (&bp, 1);
1415 node->need_bounds_init = bp_unpack_value (&bp, 1);
1416 group = read_identifier (ib);
1417 if (group)
1419 node->set_comdat_group (group);
1420 ref = streamer_read_hwi (ib);
1421 /* Store a reference for now, and fix up later to be a pointer. */
1422 node->same_comdat_group = (symtab_node *) (intptr_t) ref;
1424 else
1425 node->same_comdat_group = (symtab_node *) (intptr_t) LCC_NOT_FOUND;
1426 section = read_string (ib);
1427 if (section)
1428 node->set_section_for_node (section);
1429 node->resolution = streamer_read_enum (ib, ld_plugin_symbol_resolution,
1430 LDPR_NUM_KNOWN);
1431 gcc_assert (flag_ltrans
1432 || (!node->in_other_partition
1433 && !node->used_from_other_partition));
1435 return node;
1438 /* Read a node from input_block IB. TAG is the node's tag just read.
1439 Return the node read or overwriten. */
1441 static void
1442 input_ref (struct lto_input_block *ib,
1443 symtab_node *referring_node,
1444 vec<symtab_node *> nodes)
1446 symtab_node *node = NULL;
1447 struct bitpack_d bp;
1448 enum ipa_ref_use use;
1449 bool speculative;
1450 struct ipa_ref *ref;
1452 bp = streamer_read_bitpack (ib);
1453 use = (enum ipa_ref_use) bp_unpack_value (&bp, 3);
1454 speculative = (enum ipa_ref_use) bp_unpack_value (&bp, 1);
1455 node = nodes[streamer_read_hwi (ib)];
1456 ref = referring_node->create_reference (node, use);
1457 ref->speculative = speculative;
1458 if (is_a <cgraph_node *> (referring_node))
1459 ref->lto_stmt_uid = streamer_read_hwi (ib);
1462 /* Read an edge from IB. NODES points to a vector of previously read nodes for
1463 decoding caller and callee of the edge to be read. If INDIRECT is true, the
1464 edge being read is indirect (in the sense that it has
1465 indirect_unknown_callee set). */
1467 static void
1468 input_edge (struct lto_input_block *ib, vec<symtab_node *> nodes,
1469 bool indirect)
1471 struct cgraph_node *caller, *callee;
1472 struct cgraph_edge *edge;
1473 unsigned int stmt_id;
1474 gcov_type count;
1475 int freq;
1476 cgraph_inline_failed_t inline_failed;
1477 struct bitpack_d bp;
1478 int ecf_flags = 0;
1480 caller = dyn_cast<cgraph_node *> (nodes[streamer_read_hwi (ib)]);
1481 if (caller == NULL || caller->decl == NULL_TREE)
1482 internal_error ("bytecode stream: no caller found while reading edge");
1484 if (!indirect)
1486 callee = dyn_cast<cgraph_node *> (nodes[streamer_read_hwi (ib)]);
1487 if (callee == NULL || callee->decl == NULL_TREE)
1488 internal_error ("bytecode stream: no callee found while reading edge");
1490 else
1491 callee = NULL;
1493 count = streamer_read_gcov_count (ib);
1495 bp = streamer_read_bitpack (ib);
1496 inline_failed = bp_unpack_enum (&bp, cgraph_inline_failed_t, CIF_N_REASONS);
1497 stmt_id = bp_unpack_var_len_unsigned (&bp);
1498 freq = (int) bp_unpack_var_len_unsigned (&bp);
1500 if (indirect)
1501 edge = caller->create_indirect_edge (NULL, 0, count, freq);
1502 else
1503 edge = caller->create_edge (callee, NULL, count, freq);
1505 edge->indirect_inlining_edge = bp_unpack_value (&bp, 1);
1506 edge->speculative = bp_unpack_value (&bp, 1);
1507 edge->lto_stmt_uid = stmt_id;
1508 edge->inline_failed = inline_failed;
1509 edge->call_stmt_cannot_inline_p = bp_unpack_value (&bp, 1);
1510 edge->can_throw_external = bp_unpack_value (&bp, 1);
1511 edge->in_polymorphic_cdtor = bp_unpack_value (&bp, 1);
1512 if (indirect)
1514 if (bp_unpack_value (&bp, 1))
1515 ecf_flags |= ECF_CONST;
1516 if (bp_unpack_value (&bp, 1))
1517 ecf_flags |= ECF_PURE;
1518 if (bp_unpack_value (&bp, 1))
1519 ecf_flags |= ECF_NORETURN;
1520 if (bp_unpack_value (&bp, 1))
1521 ecf_flags |= ECF_MALLOC;
1522 if (bp_unpack_value (&bp, 1))
1523 ecf_flags |= ECF_NOTHROW;
1524 if (bp_unpack_value (&bp, 1))
1525 ecf_flags |= ECF_RETURNS_TWICE;
1526 edge->indirect_info->ecf_flags = ecf_flags;
1527 edge->indirect_info->common_target_id = streamer_read_hwi (ib);
1528 if (edge->indirect_info->common_target_id)
1529 edge->indirect_info->common_target_probability = streamer_read_hwi (ib);
1534 /* Read a cgraph from IB using the info in FILE_DATA. */
1536 static vec<symtab_node *>
1537 input_cgraph_1 (struct lto_file_decl_data *file_data,
1538 struct lto_input_block *ib)
1540 enum LTO_symtab_tags tag;
1541 vec<symtab_node *> nodes = vNULL;
1542 symtab_node *node;
1543 unsigned i;
1545 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1546 order_base = symtab->order;
1547 while (tag)
1549 if (tag == LTO_symtab_edge)
1550 input_edge (ib, nodes, false);
1551 else if (tag == LTO_symtab_indirect_edge)
1552 input_edge (ib, nodes, true);
1553 else if (tag == LTO_symtab_variable)
1555 node = input_varpool_node (file_data, ib);
1556 nodes.safe_push (node);
1557 lto_symtab_encoder_encode (file_data->symtab_node_encoder, node);
1559 else
1561 node = input_node (file_data, ib, tag, nodes);
1562 if (node == NULL || node->decl == NULL_TREE)
1563 internal_error ("bytecode stream: found empty cgraph node");
1564 nodes.safe_push (node);
1565 lto_symtab_encoder_encode (file_data->symtab_node_encoder, node);
1568 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1571 lto_input_toplevel_asms (file_data, order_base);
1573 /* AUX pointers should be all non-zero for function nodes read from the stream. */
1574 #ifdef ENABLE_CHECKING
1575 FOR_EACH_VEC_ELT (nodes, i, node)
1576 gcc_assert (node->aux || !is_a <cgraph_node *> (node));
1577 #endif
1578 FOR_EACH_VEC_ELT (nodes, i, node)
1580 int ref;
1581 if (cgraph_node *cnode = dyn_cast <cgraph_node *> (node))
1583 ref = (int) (intptr_t) cnode->global.inlined_to;
1585 /* We share declaration of builtins, so we may read same node twice. */
1586 if (!node->aux)
1587 continue;
1588 node->aux = NULL;
1590 /* Fixup inlined_to from reference to pointer. */
1591 if (ref != LCC_NOT_FOUND)
1592 dyn_cast<cgraph_node *> (node)->global.inlined_to
1593 = dyn_cast<cgraph_node *> (nodes[ref]);
1594 else
1595 cnode->global.inlined_to = NULL;
1597 /* Compute instrumented_version. */
1598 if (cnode->instrumentation_clone)
1600 gcc_assert (cnode->orig_decl);
1602 cnode->instrumented_version = cgraph_node::get (cnode->orig_decl);
1603 if (cnode->instrumented_version)
1605 /* We may have multiple nodes for a single function which
1606 will be merged later. To have a proper merge we need
1607 to keep instrumentation_version reference between nodes
1608 consistent: each instrumented_version reference should
1609 have proper reverse reference. Thus don't break existing
1610 instrumented_version reference if it already exists. */
1611 if (cnode->instrumented_version->instrumented_version)
1612 cnode->instrumented_version = NULL;
1613 else
1614 cnode->instrumented_version->instrumented_version = cnode;
1617 /* Restore decl names reference. */
1618 if (IDENTIFIER_TRANSPARENT_ALIAS (DECL_ASSEMBLER_NAME (cnode->decl))
1619 && !TREE_CHAIN (DECL_ASSEMBLER_NAME (cnode->decl)))
1620 TREE_CHAIN (DECL_ASSEMBLER_NAME (cnode->decl))
1621 = DECL_ASSEMBLER_NAME (cnode->orig_decl);
1625 ref = (int) (intptr_t) node->same_comdat_group;
1627 /* Fixup same_comdat_group from reference to pointer. */
1628 if (ref != LCC_NOT_FOUND)
1629 node->same_comdat_group = nodes[ref];
1630 else
1631 node->same_comdat_group = NULL;
1633 FOR_EACH_VEC_ELT (nodes, i, node)
1634 node->aux = is_a <cgraph_node *> (node) ? (void *)1 : NULL;
1635 return nodes;
1638 /* Input ipa_refs. */
1640 static void
1641 input_refs (struct lto_input_block *ib,
1642 vec<symtab_node *> nodes)
1644 int count;
1645 int idx;
1646 while (true)
1648 symtab_node *node;
1649 count = streamer_read_uhwi (ib);
1650 if (!count)
1651 break;
1652 idx = streamer_read_uhwi (ib);
1653 node = nodes[idx];
1654 while (count)
1656 input_ref (ib, node, nodes);
1657 count--;
1663 static struct gcov_ctr_summary lto_gcov_summary;
1665 /* Input profile_info from IB. */
1666 static void
1667 input_profile_summary (struct lto_input_block *ib,
1668 struct lto_file_decl_data *file_data)
1670 unsigned h_ix;
1671 struct bitpack_d bp;
1672 unsigned int runs = streamer_read_uhwi (ib);
1673 if (runs)
1675 file_data->profile_info.runs = runs;
1676 file_data->profile_info.sum_max = streamer_read_gcov_count (ib);
1677 file_data->profile_info.sum_all = streamer_read_gcov_count (ib);
1679 memset (file_data->profile_info.histogram, 0,
1680 sizeof (gcov_bucket_type) * GCOV_HISTOGRAM_SIZE);
1681 /* Input the bitpack of non-zero histogram indices. */
1682 bp = streamer_read_bitpack (ib);
1683 /* Read in and unpack the full bitpack, flagging non-zero
1684 histogram entries by setting the num_counters non-zero. */
1685 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1687 file_data->profile_info.histogram[h_ix].num_counters
1688 = bp_unpack_value (&bp, 1);
1690 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1692 if (!file_data->profile_info.histogram[h_ix].num_counters)
1693 continue;
1695 file_data->profile_info.histogram[h_ix].num_counters
1696 = streamer_read_gcov_count (ib);
1697 file_data->profile_info.histogram[h_ix].min_value
1698 = streamer_read_gcov_count (ib);
1699 file_data->profile_info.histogram[h_ix].cum_value
1700 = streamer_read_gcov_count (ib);
1702 /* IPA-profile computes hot bb threshold based on cumulated
1703 whole program profile. We need to stream it down to ltrans. */
1704 if (flag_ltrans)
1705 set_hot_bb_threshold (streamer_read_gcov_count (ib));
1710 /* Rescale profile summaries to the same number of runs in the whole unit. */
1712 static void
1713 merge_profile_summaries (struct lto_file_decl_data **file_data_vec)
1715 struct lto_file_decl_data *file_data;
1716 unsigned int j, h_ix;
1717 gcov_unsigned_t max_runs = 0;
1718 struct cgraph_node *node;
1719 struct cgraph_edge *edge;
1720 gcov_type saved_sum_all = 0;
1721 gcov_ctr_summary *saved_profile_info = 0;
1722 int saved_scale = 0;
1724 /* Find unit with maximal number of runs. If we ever get serious about
1725 roundoff errors, we might also consider computing smallest common
1726 multiply. */
1727 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1728 if (max_runs < file_data->profile_info.runs)
1729 max_runs = file_data->profile_info.runs;
1731 if (!max_runs)
1732 return;
1734 /* Simple overflow check. We probably don't need to support that many train
1735 runs. Such a large value probably imply data corruption anyway. */
1736 if (max_runs > INT_MAX / REG_BR_PROB_BASE)
1738 sorry ("At most %i profile runs is supported. Perhaps corrupted profile?",
1739 INT_MAX / REG_BR_PROB_BASE);
1740 return;
1743 profile_info = &lto_gcov_summary;
1744 lto_gcov_summary.runs = max_runs;
1745 lto_gcov_summary.sum_max = 0;
1746 memset (lto_gcov_summary.histogram, 0,
1747 sizeof (gcov_bucket_type) * GCOV_HISTOGRAM_SIZE);
1749 /* Rescale all units to the maximal number of runs.
1750 sum_max can not be easily merged, as we have no idea what files come from
1751 the same run. We do not use the info anyway, so leave it 0. */
1752 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1753 if (file_data->profile_info.runs)
1755 int scale = GCOV_COMPUTE_SCALE (max_runs,
1756 file_data->profile_info.runs);
1757 lto_gcov_summary.sum_max
1758 = MAX (lto_gcov_summary.sum_max,
1759 apply_scale (file_data->profile_info.sum_max, scale));
1760 lto_gcov_summary.sum_all
1761 = MAX (lto_gcov_summary.sum_all,
1762 apply_scale (file_data->profile_info.sum_all, scale));
1763 /* Save a pointer to the profile_info with the largest
1764 scaled sum_all and the scale for use in merging the
1765 histogram. */
1766 if (!saved_profile_info
1767 || lto_gcov_summary.sum_all > saved_sum_all)
1769 saved_profile_info = &file_data->profile_info;
1770 saved_sum_all = lto_gcov_summary.sum_all;
1771 saved_scale = scale;
1775 gcc_assert (saved_profile_info);
1777 /* Scale up the histogram from the profile that had the largest
1778 scaled sum_all above. */
1779 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1781 /* Scale up the min value as we did the corresponding sum_all
1782 above. Use that to find the new histogram index. */
1783 gcov_type scaled_min
1784 = apply_scale (saved_profile_info->histogram[h_ix].min_value,
1785 saved_scale);
1786 /* The new index may be shared with another scaled histogram entry,
1787 so we need to account for a non-zero histogram entry at new_ix. */
1788 unsigned new_ix = gcov_histo_index (scaled_min);
1789 lto_gcov_summary.histogram[new_ix].min_value
1790 = (lto_gcov_summary.histogram[new_ix].num_counters
1791 ? MIN (lto_gcov_summary.histogram[new_ix].min_value, scaled_min)
1792 : scaled_min);
1793 /* Some of the scaled counter values would ostensibly need to be placed
1794 into different (larger) histogram buckets, but we keep things simple
1795 here and place the scaled cumulative counter value in the bucket
1796 corresponding to the scaled minimum counter value. */
1797 lto_gcov_summary.histogram[new_ix].cum_value
1798 += apply_scale (saved_profile_info->histogram[h_ix].cum_value,
1799 saved_scale);
1800 lto_gcov_summary.histogram[new_ix].num_counters
1801 += saved_profile_info->histogram[h_ix].num_counters;
1804 /* Watch roundoff errors. */
1805 if (lto_gcov_summary.sum_max < max_runs)
1806 lto_gcov_summary.sum_max = max_runs;
1808 /* If merging already happent at WPA time, we are done. */
1809 if (flag_ltrans)
1810 return;
1812 /* Now compute count_materialization_scale of each node.
1813 During LTRANS we already have values of count_materialization_scale
1814 computed, so just update them. */
1815 FOR_EACH_FUNCTION (node)
1816 if (node->lto_file_data
1817 && node->lto_file_data->profile_info.runs)
1819 int scale;
1821 scale = RDIV (node->count_materialization_scale * max_runs,
1822 node->lto_file_data->profile_info.runs);
1823 node->count_materialization_scale = scale;
1824 if (scale < 0)
1825 fatal_error (input_location, "Profile information in %s corrupted",
1826 file_data->file_name);
1828 if (scale == REG_BR_PROB_BASE)
1829 continue;
1830 for (edge = node->callees; edge; edge = edge->next_callee)
1831 edge->count = apply_scale (edge->count, scale);
1832 node->count = apply_scale (node->count, scale);
1836 /* Input and merge the symtab from each of the .o files passed to
1837 lto1. */
1839 void
1840 input_symtab (void)
1842 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1843 struct lto_file_decl_data *file_data;
1844 unsigned int j = 0;
1845 struct cgraph_node *node;
1847 while ((file_data = file_data_vec[j++]))
1849 const char *data;
1850 size_t len;
1851 struct lto_input_block *ib;
1852 vec<symtab_node *> nodes;
1854 ib = lto_create_simple_input_block (file_data, LTO_section_symtab_nodes,
1855 &data, &len);
1856 if (!ib)
1857 fatal_error (input_location,
1858 "cannot find LTO cgraph in %s", file_data->file_name);
1859 input_profile_summary (ib, file_data);
1860 file_data->symtab_node_encoder = lto_symtab_encoder_new (true);
1861 nodes = input_cgraph_1 (file_data, ib);
1862 lto_destroy_simple_input_block (file_data, LTO_section_symtab_nodes,
1863 ib, data, len);
1865 ib = lto_create_simple_input_block (file_data, LTO_section_refs,
1866 &data, &len);
1867 if (!ib)
1868 fatal_error (input_location, "cannot find LTO section refs in %s",
1869 file_data->file_name);
1870 input_refs (ib, nodes);
1871 lto_destroy_simple_input_block (file_data, LTO_section_refs,
1872 ib, data, len);
1873 if (flag_ltrans)
1874 input_cgraph_opt_summary (nodes);
1875 nodes.release ();
1878 merge_profile_summaries (file_data_vec);
1879 get_working_sets ();
1882 /* Clear out the aux field that was used to store enough state to
1883 tell which nodes should be overwritten. */
1884 FOR_EACH_FUNCTION (node)
1886 /* Some nodes may have been created by cgraph_node. This
1887 happens when the callgraph contains nested functions. If the
1888 node for the parent function was never emitted to the gimple
1889 file, cgraph_node will create a node for it when setting the
1890 context of the nested function. */
1891 if (node->lto_file_data)
1892 node->aux = NULL;
1896 /* Input function/variable tables that will allow libgomp to look up offload
1897 target code, and store them into OFFLOAD_FUNCS and OFFLOAD_VARS. */
1899 void
1900 input_offload_tables (void)
1902 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1903 struct lto_file_decl_data *file_data;
1904 unsigned int j = 0;
1906 while ((file_data = file_data_vec[j++]))
1908 const char *data;
1909 size_t len;
1910 struct lto_input_block *ib
1911 = lto_create_simple_input_block (file_data, LTO_section_offload_table,
1912 &data, &len);
1913 if (!ib)
1914 continue;
1916 enum LTO_symtab_tags tag
1917 = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1918 while (tag)
1920 if (tag == LTO_symtab_unavail_node)
1922 int decl_index = streamer_read_uhwi (ib);
1923 tree fn_decl
1924 = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1925 vec_safe_push (offload_funcs, fn_decl);
1927 else if (tag == LTO_symtab_variable)
1929 int decl_index = streamer_read_uhwi (ib);
1930 tree var_decl
1931 = lto_file_decl_data_get_var_decl (file_data, decl_index);
1932 vec_safe_push (offload_vars, var_decl);
1934 else
1935 fatal_error (input_location,
1936 "invalid offload table in %s", file_data->file_name);
1938 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1941 lto_destroy_simple_input_block (file_data, LTO_section_offload_table,
1942 ib, data, len);
1946 /* True when we need optimization summary for NODE. */
1948 static int
1949 output_cgraph_opt_summary_p (struct cgraph_node *node)
1951 return (node->clone_of
1952 && (node->clone.tree_map
1953 || node->clone.args_to_skip
1954 || node->clone.combined_args_to_skip));
1957 /* Output optimization summary for EDGE to OB. */
1958 static void
1959 output_edge_opt_summary (struct output_block *ob ATTRIBUTE_UNUSED,
1960 struct cgraph_edge *edge ATTRIBUTE_UNUSED)
1964 /* Output optimization summary for NODE to OB. */
1966 static void
1967 output_node_opt_summary (struct output_block *ob,
1968 struct cgraph_node *node,
1969 lto_symtab_encoder_t encoder)
1971 unsigned int index;
1972 bitmap_iterator bi;
1973 struct ipa_replace_map *map;
1974 struct bitpack_d bp;
1975 int i;
1976 struct cgraph_edge *e;
1978 if (node->clone.args_to_skip)
1980 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.args_to_skip));
1981 EXECUTE_IF_SET_IN_BITMAP (node->clone.args_to_skip, 0, index, bi)
1982 streamer_write_uhwi (ob, index);
1984 else
1985 streamer_write_uhwi (ob, 0);
1986 if (node->clone.combined_args_to_skip)
1988 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.combined_args_to_skip));
1989 EXECUTE_IF_SET_IN_BITMAP (node->clone.combined_args_to_skip, 0, index, bi)
1990 streamer_write_uhwi (ob, index);
1992 else
1993 streamer_write_uhwi (ob, 0);
1994 streamer_write_uhwi (ob, vec_safe_length (node->clone.tree_map));
1995 FOR_EACH_VEC_SAFE_ELT (node->clone.tree_map, i, map)
1997 /* At the moment we assume all old trees to be PARM_DECLs, because we have no
1998 mechanism to store function local declarations into summaries. */
1999 gcc_assert (!map->old_tree);
2000 streamer_write_uhwi (ob, map->parm_num);
2001 gcc_assert (EXPR_LOCATION (map->new_tree) == UNKNOWN_LOCATION);
2002 stream_write_tree (ob, map->new_tree, true);
2003 bp = bitpack_create (ob->main_stream);
2004 bp_pack_value (&bp, map->replace_p, 1);
2005 bp_pack_value (&bp, map->ref_p, 1);
2006 streamer_write_bitpack (&bp);
2009 if (lto_symtab_encoder_in_partition_p (encoder, node))
2011 for (e = node->callees; e; e = e->next_callee)
2012 output_edge_opt_summary (ob, e);
2013 for (e = node->indirect_calls; e; e = e->next_callee)
2014 output_edge_opt_summary (ob, e);
2018 /* Output optimization summaries stored in callgraph.
2019 At the moment it is the clone info structure. */
2021 static void
2022 output_cgraph_opt_summary (void)
2024 int i, n_nodes;
2025 lto_symtab_encoder_t encoder;
2026 struct output_block *ob = create_output_block (LTO_section_cgraph_opt_sum);
2027 unsigned count = 0;
2029 ob->symbol = NULL;
2030 encoder = ob->decl_state->symtab_node_encoder;
2031 n_nodes = lto_symtab_encoder_size (encoder);
2032 for (i = 0; i < n_nodes; i++)
2034 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
2035 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
2036 if (cnode && output_cgraph_opt_summary_p (cnode))
2037 count++;
2039 streamer_write_uhwi (ob, count);
2040 for (i = 0; i < n_nodes; i++)
2042 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
2043 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
2044 if (cnode && output_cgraph_opt_summary_p (cnode))
2046 streamer_write_uhwi (ob, i);
2047 output_node_opt_summary (ob, cnode, encoder);
2050 produce_asm (ob, NULL);
2051 destroy_output_block (ob);
2054 /* Input optimisation summary of EDGE. */
2056 static void
2057 input_edge_opt_summary (struct cgraph_edge *edge ATTRIBUTE_UNUSED,
2058 struct lto_input_block *ib_main ATTRIBUTE_UNUSED)
2062 /* Input optimisation summary of NODE. */
2064 static void
2065 input_node_opt_summary (struct cgraph_node *node,
2066 struct lto_input_block *ib_main,
2067 struct data_in *data_in)
2069 int i;
2070 int count;
2071 int bit;
2072 struct bitpack_d bp;
2073 struct cgraph_edge *e;
2075 count = streamer_read_uhwi (ib_main);
2076 if (count)
2077 node->clone.args_to_skip = BITMAP_GGC_ALLOC ();
2078 for (i = 0; i < count; i++)
2080 bit = streamer_read_uhwi (ib_main);
2081 bitmap_set_bit (node->clone.args_to_skip, bit);
2083 count = streamer_read_uhwi (ib_main);
2084 if (count)
2085 node->clone.combined_args_to_skip = BITMAP_GGC_ALLOC ();
2086 for (i = 0; i < count; i++)
2088 bit = streamer_read_uhwi (ib_main);
2089 bitmap_set_bit (node->clone.combined_args_to_skip, bit);
2091 count = streamer_read_uhwi (ib_main);
2092 for (i = 0; i < count; i++)
2094 struct ipa_replace_map *map = ggc_alloc<ipa_replace_map> ();
2096 vec_safe_push (node->clone.tree_map, map);
2097 map->parm_num = streamer_read_uhwi (ib_main);
2098 map->old_tree = NULL;
2099 map->new_tree = stream_read_tree (ib_main, data_in);
2100 bp = streamer_read_bitpack (ib_main);
2101 map->replace_p = bp_unpack_value (&bp, 1);
2102 map->ref_p = bp_unpack_value (&bp, 1);
2104 for (e = node->callees; e; e = e->next_callee)
2105 input_edge_opt_summary (e, ib_main);
2106 for (e = node->indirect_calls; e; e = e->next_callee)
2107 input_edge_opt_summary (e, ib_main);
2110 /* Read section in file FILE_DATA of length LEN with data DATA. */
2112 static void
2113 input_cgraph_opt_section (struct lto_file_decl_data *file_data,
2114 const char *data, size_t len,
2115 vec<symtab_node *> nodes)
2117 const struct lto_function_header *header =
2118 (const struct lto_function_header *) data;
2119 const int cfg_offset = sizeof (struct lto_function_header);
2120 const int main_offset = cfg_offset + header->cfg_size;
2121 const int string_offset = main_offset + header->main_size;
2122 struct data_in *data_in;
2123 unsigned int i;
2124 unsigned int count;
2126 lto_input_block ib_main ((const char *) data + main_offset,
2127 header->main_size, file_data->mode_table);
2129 data_in =
2130 lto_data_in_create (file_data, (const char *) data + string_offset,
2131 header->string_size, vNULL);
2132 count = streamer_read_uhwi (&ib_main);
2134 for (i = 0; i < count; i++)
2136 int ref = streamer_read_uhwi (&ib_main);
2137 input_node_opt_summary (dyn_cast<cgraph_node *> (nodes[ref]),
2138 &ib_main, data_in);
2140 lto_free_section_data (file_data, LTO_section_cgraph_opt_sum, NULL, data,
2141 len);
2142 lto_data_in_delete (data_in);
2145 /* Input optimization summary of cgraph. */
2147 static void
2148 input_cgraph_opt_summary (vec<symtab_node *> nodes)
2150 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
2151 struct lto_file_decl_data *file_data;
2152 unsigned int j = 0;
2154 while ((file_data = file_data_vec[j++]))
2156 size_t len;
2157 const char *data =
2158 lto_get_section_data (file_data, LTO_section_cgraph_opt_sum, NULL,
2159 &len);
2161 if (data)
2162 input_cgraph_opt_section (file_data, data, len, nodes);