Daily bump.
[official-gcc.git] / gcc / lto-cgraph.c
blob1e6a7adeaa20a7b1594169e1f1f7be043f81d685
1 /* Write and read the cgraph to the memory mapped representation of a
2 .o file.
4 Copyright (C) 2009-2018 Free Software Foundation, Inc.
5 Contributed by Kenneth Zadeck <zadeck@naturalbridge.com>
7 This file is part of GCC.
9 GCC is free software; you can redistribute it and/or modify it under
10 the terms of the GNU General Public License as published by the Free
11 Software Foundation; either version 3, or (at your option) any later
12 version.
14 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
15 WARRANTY; without even the implied warranty of MERCHANTABILITY or
16 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 for more details.
19 You should have received a copy of the GNU General Public License
20 along with GCC; see the file COPYING3. If not see
21 <http://www.gnu.org/licenses/>. */
23 #include "config.h"
24 #include "system.h"
25 #include "coretypes.h"
26 #include "backend.h"
27 #include "rtl.h"
28 #include "tree.h"
29 #include "gimple.h"
30 #include "predict.h"
31 #include "stringpool.h"
32 #include "tree-streamer.h"
33 #include "cgraph.h"
34 #include "tree-pass.h"
35 #include "profile.h"
36 #include "context.h"
37 #include "pass_manager.h"
38 #include "ipa-utils.h"
39 #include "omp-offload.h"
40 #include "stringpool.h"
41 #include "attribs.h"
43 /* True when asm nodes has been output. */
44 bool asm_nodes_output = false;
46 static void output_cgraph_opt_summary (void);
47 static void input_cgraph_opt_summary (vec<symtab_node *> nodes);
49 /* Number of LDPR values known to GCC. */
50 #define LDPR_NUM_KNOWN (LDPR_PREVAILING_DEF_IRONLY_EXP + 1)
52 /* All node orders are ofsetted by ORDER_BASE. */
53 static int order_base;
55 /* Cgraph streaming is organized as set of record whose type
56 is indicated by a tag. */
57 enum LTO_symtab_tags
59 /* Must leave 0 for the stopper. */
61 /* Cgraph node without body available. */
62 LTO_symtab_unavail_node = 1,
63 /* Cgraph node with function body. */
64 LTO_symtab_analyzed_node,
65 /* Cgraph edges. */
66 LTO_symtab_edge,
67 LTO_symtab_indirect_edge,
68 LTO_symtab_variable,
69 LTO_symtab_last_tag
72 /* Create a new symtab encoder.
73 if FOR_INPUT, the encoder allocate only datastructures needed
74 to read the symtab. */
76 lto_symtab_encoder_t
77 lto_symtab_encoder_new (bool for_input)
79 lto_symtab_encoder_t encoder = XCNEW (struct lto_symtab_encoder_d);
81 if (!for_input)
82 encoder->map = new hash_map<symtab_node *, size_t>;
83 encoder->nodes.create (0);
84 return encoder;
88 /* Delete ENCODER and its components. */
90 void
91 lto_symtab_encoder_delete (lto_symtab_encoder_t encoder)
93 encoder->nodes.release ();
94 if (encoder->map)
95 delete encoder->map;
96 free (encoder);
100 /* Return the existing reference number of NODE in the symtab encoder in
101 output block OB. Assign a new reference if this is the first time
102 NODE is encoded. */
105 lto_symtab_encoder_encode (lto_symtab_encoder_t encoder,
106 symtab_node *node)
108 int ref;
110 if (!encoder->map)
112 lto_encoder_entry entry = {node, false, false, false};
114 ref = encoder->nodes.length ();
115 encoder->nodes.safe_push (entry);
116 return ref;
119 size_t *slot = encoder->map->get (node);
120 if (!slot || !*slot)
122 lto_encoder_entry entry = {node, false, false, false};
123 ref = encoder->nodes.length ();
124 if (!slot)
125 encoder->map->put (node, ref + 1);
126 encoder->nodes.safe_push (entry);
128 else
129 ref = *slot - 1;
131 return ref;
134 /* Remove NODE from encoder. */
136 bool
137 lto_symtab_encoder_delete_node (lto_symtab_encoder_t encoder,
138 symtab_node *node)
140 int index;
141 lto_encoder_entry last_node;
143 size_t *slot = encoder->map->get (node);
144 if (slot == NULL || !*slot)
145 return false;
147 index = *slot - 1;
148 gcc_checking_assert (encoder->nodes[index].node == node);
150 /* Remove from vector. We do this by swapping node with the last element
151 of the vector. */
152 last_node = encoder->nodes.pop ();
153 if (last_node.node != node)
155 gcc_assert (encoder->map->put (last_node.node, index + 1));
157 /* Move the last element to the original spot of NODE. */
158 encoder->nodes[index] = last_node;
161 /* Remove element from hash table. */
162 encoder->map->remove (node);
163 return true;
167 /* Return TRUE if we should encode the body of NODE (if any). */
169 bool
170 lto_symtab_encoder_encode_body_p (lto_symtab_encoder_t encoder,
171 struct cgraph_node *node)
173 int index = lto_symtab_encoder_lookup (encoder, node);
174 return encoder->nodes[index].body;
177 /* Specify that we encode the body of NODE in this partition. */
179 static void
180 lto_set_symtab_encoder_encode_body (lto_symtab_encoder_t encoder,
181 struct cgraph_node *node)
183 int index = lto_symtab_encoder_encode (encoder, node);
184 gcc_checking_assert (encoder->nodes[index].node == node);
185 encoder->nodes[index].body = true;
188 /* Return TRUE if we should encode initializer of NODE (if any). */
190 bool
191 lto_symtab_encoder_encode_initializer_p (lto_symtab_encoder_t encoder,
192 varpool_node *node)
194 int index = lto_symtab_encoder_lookup (encoder, node);
195 if (index == LCC_NOT_FOUND)
196 return false;
197 return encoder->nodes[index].initializer;
200 /* Specify that we should encode initializer of NODE (if any). */
202 static void
203 lto_set_symtab_encoder_encode_initializer (lto_symtab_encoder_t encoder,
204 varpool_node *node)
206 int index = lto_symtab_encoder_lookup (encoder, node);
207 encoder->nodes[index].initializer = true;
210 /* Return TRUE if NODE is in this partition. */
212 bool
213 lto_symtab_encoder_in_partition_p (lto_symtab_encoder_t encoder,
214 symtab_node *node)
216 int index = lto_symtab_encoder_lookup (encoder, node);
217 if (index == LCC_NOT_FOUND)
218 return false;
219 return encoder->nodes[index].in_partition;
222 /* Specify that NODE is in this partition. */
224 void
225 lto_set_symtab_encoder_in_partition (lto_symtab_encoder_t encoder,
226 symtab_node *node)
228 int index = lto_symtab_encoder_encode (encoder, node);
229 encoder->nodes[index].in_partition = true;
232 /* Output the cgraph EDGE to OB using ENCODER. */
234 static void
235 lto_output_edge (struct lto_simple_output_block *ob, struct cgraph_edge *edge,
236 lto_symtab_encoder_t encoder)
238 unsigned int uid;
239 intptr_t ref;
240 struct bitpack_d bp;
242 if (edge->indirect_unknown_callee)
243 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
244 LTO_symtab_indirect_edge);
245 else
246 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
247 LTO_symtab_edge);
249 ref = lto_symtab_encoder_lookup (encoder, edge->caller);
250 gcc_assert (ref != LCC_NOT_FOUND);
251 streamer_write_hwi_stream (ob->main_stream, ref);
253 if (!edge->indirect_unknown_callee)
255 ref = lto_symtab_encoder_lookup (encoder, edge->callee);
256 gcc_assert (ref != LCC_NOT_FOUND);
257 streamer_write_hwi_stream (ob->main_stream, ref);
260 edge->count.stream_out (ob->main_stream);
262 bp = bitpack_create (ob->main_stream);
263 uid = (!gimple_has_body_p (edge->caller->decl) || edge->caller->thunk.thunk_p
264 ? edge->lto_stmt_uid : gimple_uid (edge->call_stmt) + 1);
265 bp_pack_enum (&bp, cgraph_inline_failed_t,
266 CIF_N_REASONS, edge->inline_failed);
267 bp_pack_var_len_unsigned (&bp, uid);
268 bp_pack_value (&bp, edge->indirect_inlining_edge, 1);
269 bp_pack_value (&bp, edge->speculative, 1);
270 bp_pack_value (&bp, edge->call_stmt_cannot_inline_p, 1);
271 gcc_assert (!edge->call_stmt_cannot_inline_p
272 || edge->inline_failed != CIF_BODY_NOT_AVAILABLE);
273 bp_pack_value (&bp, edge->can_throw_external, 1);
274 bp_pack_value (&bp, edge->in_polymorphic_cdtor, 1);
275 if (edge->indirect_unknown_callee)
277 int flags = edge->indirect_info->ecf_flags;
278 bp_pack_value (&bp, (flags & ECF_CONST) != 0, 1);
279 bp_pack_value (&bp, (flags & ECF_PURE) != 0, 1);
280 bp_pack_value (&bp, (flags & ECF_NORETURN) != 0, 1);
281 bp_pack_value (&bp, (flags & ECF_MALLOC) != 0, 1);
282 bp_pack_value (&bp, (flags & ECF_NOTHROW) != 0, 1);
283 bp_pack_value (&bp, (flags & ECF_RETURNS_TWICE) != 0, 1);
284 /* Flags that should not appear on indirect calls. */
285 gcc_assert (!(flags & (ECF_LOOPING_CONST_OR_PURE
286 | ECF_MAY_BE_ALLOCA
287 | ECF_SIBCALL
288 | ECF_LEAF
289 | ECF_NOVOPS)));
291 streamer_write_bitpack (&bp);
292 if (edge->indirect_unknown_callee)
294 streamer_write_hwi_stream (ob->main_stream,
295 edge->indirect_info->common_target_id);
296 if (edge->indirect_info->common_target_id)
297 streamer_write_hwi_stream
298 (ob->main_stream, edge->indirect_info->common_target_probability);
302 /* Return if NODE contain references from other partitions. */
304 bool
305 referenced_from_other_partition_p (symtab_node *node, lto_symtab_encoder_t encoder)
307 int i;
308 struct ipa_ref *ref = NULL;
310 for (i = 0; node->iterate_referring (i, ref); i++)
312 /* Ignore references from non-offloadable nodes while streaming NODE into
313 offload LTO section. */
314 if (!ref->referring->need_lto_streaming)
315 continue;
317 if (ref->referring->in_other_partition
318 || !lto_symtab_encoder_in_partition_p (encoder, ref->referring))
319 return true;
321 return false;
324 /* Return true when node is reachable from other partition. */
326 bool
327 reachable_from_other_partition_p (struct cgraph_node *node, lto_symtab_encoder_t encoder)
329 struct cgraph_edge *e;
330 if (!node->definition)
331 return false;
332 if (node->global.inlined_to)
333 return false;
334 for (e = node->callers; e; e = e->next_caller)
336 /* Ignore references from non-offloadable nodes while streaming NODE into
337 offload LTO section. */
338 if (!e->caller->need_lto_streaming)
339 continue;
341 if (e->caller->in_other_partition
342 || !lto_symtab_encoder_in_partition_p (encoder, e->caller))
343 return true;
345 return false;
348 /* Return if NODE contain references from other partitions. */
350 bool
351 referenced_from_this_partition_p (symtab_node *node,
352 lto_symtab_encoder_t encoder)
354 int i;
355 struct ipa_ref *ref = NULL;
357 for (i = 0; node->iterate_referring (i, ref); i++)
358 if (lto_symtab_encoder_in_partition_p (encoder, ref->referring))
359 return true;
360 return false;
363 /* Return true when node is reachable from other partition. */
365 bool
366 reachable_from_this_partition_p (struct cgraph_node *node, lto_symtab_encoder_t encoder)
368 struct cgraph_edge *e;
369 for (e = node->callers; e; e = e->next_caller)
370 if (lto_symtab_encoder_in_partition_p (encoder, e->caller))
371 return true;
372 return false;
375 /* Output the cgraph NODE to OB. ENCODER is used to find the
376 reference number of NODE->inlined_to. SET is the set of nodes we
377 are writing to the current file. If NODE is not in SET, then NODE
378 is a boundary of a cgraph_node_set and we pretend NODE just has a
379 decl and no callees. WRITTEN_DECLS is the set of FUNCTION_DECLs
380 that have had their callgraph node written so far. This is used to
381 determine if NODE is a clone of a previously written node. */
383 static void
384 lto_output_node (struct lto_simple_output_block *ob, struct cgraph_node *node,
385 lto_symtab_encoder_t encoder)
387 unsigned int tag;
388 struct bitpack_d bp;
389 bool boundary_p;
390 intptr_t ref;
391 bool in_other_partition = false;
392 struct cgraph_node *clone_of, *ultimate_clone_of;
393 ipa_opt_pass_d *pass;
394 int i;
395 const char *comdat;
396 const char *section;
397 tree group;
399 boundary_p = !lto_symtab_encoder_in_partition_p (encoder, node);
401 if (node->analyzed && (!boundary_p || node->alias
402 || (node->thunk.thunk_p && !node->global.inlined_to)))
403 tag = LTO_symtab_analyzed_node;
404 else
405 tag = LTO_symtab_unavail_node;
407 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
408 tag);
409 streamer_write_hwi_stream (ob->main_stream, node->order);
411 /* In WPA mode, we only output part of the call-graph. Also, we
412 fake cgraph node attributes. There are two cases that we care.
414 Boundary nodes: There are nodes that are not part of SET but are
415 called from within SET. We artificially make them look like
416 externally visible nodes with no function body.
418 Cherry-picked nodes: These are nodes we pulled from other
419 translation units into SET during IPA-inlining. We make them as
420 local static nodes to prevent clashes with other local statics. */
421 if (boundary_p && node->analyzed
422 && node->get_partitioning_class () == SYMBOL_PARTITION)
424 /* Inline clones can not be part of boundary.
425 gcc_assert (!node->global.inlined_to);
427 FIXME: At the moment they can be, when partition contains an inline
428 clone that is clone of inline clone from outside partition. We can
429 reshape the clone tree and make other tree to be the root, but it
430 needs a bit extra work and will be promplty done by cgraph_remove_node
431 after reading back. */
432 in_other_partition = 1;
435 clone_of = node->clone_of;
436 while (clone_of
437 && (ref = lto_symtab_encoder_lookup (encoder, clone_of)) == LCC_NOT_FOUND)
438 if (clone_of->prev_sibling_clone)
439 clone_of = clone_of->prev_sibling_clone;
440 else
441 clone_of = clone_of->clone_of;
443 /* See if body of the master function is output. If not, we are seeing only
444 an declaration and we do not need to pass down clone tree. */
445 ultimate_clone_of = clone_of;
446 while (ultimate_clone_of && ultimate_clone_of->clone_of)
447 ultimate_clone_of = ultimate_clone_of->clone_of;
449 if (clone_of && !lto_symtab_encoder_encode_body_p (encoder, ultimate_clone_of))
450 clone_of = NULL;
452 if (tag == LTO_symtab_analyzed_node)
453 gcc_assert (clone_of || !node->clone_of);
454 if (!clone_of)
455 streamer_write_hwi_stream (ob->main_stream, LCC_NOT_FOUND);
456 else
457 streamer_write_hwi_stream (ob->main_stream, ref);
460 lto_output_fn_decl_index (ob->decl_state, ob->main_stream, node->decl);
461 node->count.stream_out (ob->main_stream);
462 streamer_write_hwi_stream (ob->main_stream, node->count_materialization_scale);
464 streamer_write_hwi_stream (ob->main_stream,
465 node->ipa_transforms_to_apply.length ());
466 FOR_EACH_VEC_ELT (node->ipa_transforms_to_apply, i, pass)
467 streamer_write_hwi_stream (ob->main_stream, pass->static_pass_number);
469 if (tag == LTO_symtab_analyzed_node)
471 if (node->global.inlined_to)
473 ref = lto_symtab_encoder_lookup (encoder, node->global.inlined_to);
474 gcc_assert (ref != LCC_NOT_FOUND);
476 else
477 ref = LCC_NOT_FOUND;
479 streamer_write_hwi_stream (ob->main_stream, ref);
482 group = node->get_comdat_group ();
483 if (group)
484 comdat = IDENTIFIER_POINTER (group);
485 else
486 comdat = "";
487 streamer_write_data_stream (ob->main_stream, comdat, strlen (comdat) + 1);
489 if (group)
491 if (node->same_comdat_group)
493 ref = LCC_NOT_FOUND;
494 for (struct symtab_node *n = node->same_comdat_group;
495 ref == LCC_NOT_FOUND && n != node; n = n->same_comdat_group)
496 ref = lto_symtab_encoder_lookup (encoder, n);
498 else
499 ref = LCC_NOT_FOUND;
500 streamer_write_hwi_stream (ob->main_stream, ref);
503 section = node->get_section ();
504 if (!section)
505 section = "";
507 streamer_write_hwi_stream (ob->main_stream, node->tp_first_run);
509 bp = bitpack_create (ob->main_stream);
510 bp_pack_value (&bp, node->local.local, 1);
511 bp_pack_value (&bp, node->externally_visible, 1);
512 bp_pack_value (&bp, node->no_reorder, 1);
513 bp_pack_value (&bp, node->definition, 1);
514 bp_pack_value (&bp, node->local.versionable, 1);
515 bp_pack_value (&bp, node->local.can_change_signature, 1);
516 bp_pack_value (&bp, node->local.redefined_extern_inline, 1);
517 bp_pack_value (&bp, node->force_output, 1);
518 bp_pack_value (&bp, node->forced_by_abi, 1);
519 bp_pack_value (&bp, node->unique_name, 1);
520 bp_pack_value (&bp, node->body_removed, 1);
521 bp_pack_value (&bp, node->implicit_section, 1);
522 bp_pack_value (&bp, node->address_taken, 1);
523 bp_pack_value (&bp, tag == LTO_symtab_analyzed_node
524 && node->get_partitioning_class () == SYMBOL_PARTITION
525 && (reachable_from_other_partition_p (node, encoder)
526 || referenced_from_other_partition_p (node, encoder)), 1);
527 bp_pack_value (&bp, node->lowered, 1);
528 bp_pack_value (&bp, in_other_partition, 1);
529 bp_pack_value (&bp, node->alias, 1);
530 bp_pack_value (&bp, node->transparent_alias, 1);
531 bp_pack_value (&bp, node->weakref, 1);
532 bp_pack_value (&bp, node->frequency, 2);
533 bp_pack_value (&bp, node->only_called_at_startup, 1);
534 bp_pack_value (&bp, node->only_called_at_exit, 1);
535 bp_pack_value (&bp, node->tm_clone, 1);
536 bp_pack_value (&bp, node->calls_comdat_local, 1);
537 bp_pack_value (&bp, node->icf_merged, 1);
538 bp_pack_value (&bp, node->nonfreeing_fn, 1);
539 bp_pack_value (&bp, node->thunk.thunk_p, 1);
540 bp_pack_value (&bp, node->parallelized_function, 1);
541 bp_pack_enum (&bp, ld_plugin_symbol_resolution,
542 LDPR_NUM_KNOWN,
543 /* When doing incremental link, we will get new resolution
544 info next time we process the file. */
545 flag_incremental_link ? LDPR_UNKNOWN : node->resolution);
546 bp_pack_value (&bp, node->split_part, 1);
547 streamer_write_bitpack (&bp);
548 streamer_write_data_stream (ob->main_stream, section, strlen (section) + 1);
550 if (node->thunk.thunk_p)
552 streamer_write_uhwi_stream
553 (ob->main_stream,
554 1 + (node->thunk.this_adjusting != 0) * 2
555 + (node->thunk.virtual_offset_p != 0) * 4
556 + (node->thunk.add_pointer_bounds_args != 0) * 8);
557 streamer_write_uhwi_stream (ob->main_stream, node->thunk.fixed_offset);
558 streamer_write_uhwi_stream (ob->main_stream, node->thunk.virtual_value);
560 streamer_write_hwi_stream (ob->main_stream, node->profile_id);
561 if (DECL_STATIC_CONSTRUCTOR (node->decl))
562 streamer_write_hwi_stream (ob->main_stream, node->get_init_priority ());
563 if (DECL_STATIC_DESTRUCTOR (node->decl))
564 streamer_write_hwi_stream (ob->main_stream, node->get_fini_priority ());
567 /* Output the varpool NODE to OB.
568 If NODE is not in SET, then NODE is a boundary. */
570 static void
571 lto_output_varpool_node (struct lto_simple_output_block *ob, varpool_node *node,
572 lto_symtab_encoder_t encoder)
574 bool boundary_p = !lto_symtab_encoder_in_partition_p (encoder, node);
575 bool encode_initializer_p
576 = (node->definition
577 && lto_symtab_encoder_encode_initializer_p (encoder, node));
578 struct bitpack_d bp;
579 int ref;
580 const char *comdat;
581 const char *section;
582 tree group;
584 gcc_assert (!encode_initializer_p || node->definition);
585 gcc_assert (boundary_p || encode_initializer_p);
587 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
588 LTO_symtab_variable);
589 streamer_write_hwi_stream (ob->main_stream, node->order);
590 lto_output_var_decl_index (ob->decl_state, ob->main_stream, node->decl);
591 bp = bitpack_create (ob->main_stream);
592 bp_pack_value (&bp, node->externally_visible, 1);
593 bp_pack_value (&bp, node->no_reorder, 1);
594 bp_pack_value (&bp, node->force_output, 1);
595 bp_pack_value (&bp, node->forced_by_abi, 1);
596 bp_pack_value (&bp, node->unique_name, 1);
597 bp_pack_value (&bp,
598 node->body_removed
599 || (!encode_initializer_p && !node->alias && node->definition),
601 bp_pack_value (&bp, node->implicit_section, 1);
602 bp_pack_value (&bp, node->writeonly, 1);
603 bp_pack_value (&bp, node->definition && (encode_initializer_p || node->alias),
605 bp_pack_value (&bp, node->alias, 1);
606 bp_pack_value (&bp, node->transparent_alias, 1);
607 bp_pack_value (&bp, node->weakref, 1);
608 bp_pack_value (&bp, node->analyzed && (!boundary_p || node->alias), 1);
609 gcc_assert (node->definition || !node->analyzed);
610 /* Constant pool initializers can be de-unified into individual ltrans units.
611 FIXME: Alternatively at -Os we may want to avoid generating for them the local
612 labels and share them across LTRANS partitions. */
613 if (node->get_partitioning_class () != SYMBOL_PARTITION)
615 bp_pack_value (&bp, 0, 1); /* used_from_other_parition. */
616 bp_pack_value (&bp, 0, 1); /* in_other_partition. */
618 else
620 bp_pack_value (&bp, node->definition
621 && referenced_from_other_partition_p (node, encoder), 1);
622 bp_pack_value (&bp, node->analyzed
623 && boundary_p && !DECL_EXTERNAL (node->decl), 1);
624 /* in_other_partition. */
626 bp_pack_value (&bp, node->tls_model, 3);
627 bp_pack_value (&bp, node->used_by_single_function, 1);
628 bp_pack_value (&bp, node->dynamically_initialized, 1);
629 bp_pack_value (&bp, node->need_bounds_init, 1);
630 streamer_write_bitpack (&bp);
632 group = node->get_comdat_group ();
633 if (group)
634 comdat = IDENTIFIER_POINTER (group);
635 else
636 comdat = "";
637 streamer_write_data_stream (ob->main_stream, comdat, strlen (comdat) + 1);
639 if (group)
641 if (node->same_comdat_group)
643 ref = LCC_NOT_FOUND;
644 for (struct symtab_node *n = node->same_comdat_group;
645 ref == LCC_NOT_FOUND && n != node; n = n->same_comdat_group)
646 ref = lto_symtab_encoder_lookup (encoder, n);
648 else
649 ref = LCC_NOT_FOUND;
650 streamer_write_hwi_stream (ob->main_stream, ref);
653 section = node->get_section ();
654 if (!section)
655 section = "";
656 streamer_write_data_stream (ob->main_stream, section, strlen (section) + 1);
658 streamer_write_enum (ob->main_stream, ld_plugin_symbol_resolution,
659 LDPR_NUM_KNOWN, node->resolution);
662 /* Output the varpool NODE to OB.
663 If NODE is not in SET, then NODE is a boundary. */
665 static void
666 lto_output_ref (struct lto_simple_output_block *ob, struct ipa_ref *ref,
667 lto_symtab_encoder_t encoder)
669 struct bitpack_d bp;
670 int nref;
671 int uid = ref->lto_stmt_uid;
672 struct cgraph_node *node;
674 bp = bitpack_create (ob->main_stream);
675 bp_pack_value (&bp, ref->use, 3);
676 bp_pack_value (&bp, ref->speculative, 1);
677 streamer_write_bitpack (&bp);
678 nref = lto_symtab_encoder_lookup (encoder, ref->referred);
679 gcc_assert (nref != LCC_NOT_FOUND);
680 streamer_write_hwi_stream (ob->main_stream, nref);
682 node = dyn_cast <cgraph_node *> (ref->referring);
683 if (node)
685 if (ref->stmt)
686 uid = gimple_uid (ref->stmt) + 1;
687 streamer_write_hwi_stream (ob->main_stream, uid);
691 /* Stream out profile_summary to OB. */
693 static void
694 output_profile_summary (struct lto_simple_output_block *ob)
696 unsigned h_ix;
697 struct bitpack_d bp;
699 if (profile_info)
701 /* We do not output num and run_max, they are not used by
702 GCC profile feedback and they are difficult to merge from multiple
703 units. */
704 gcc_assert (profile_info->runs);
705 streamer_write_uhwi_stream (ob->main_stream, profile_info->runs);
706 streamer_write_gcov_count_stream (ob->main_stream, profile_info->sum_max);
708 /* sum_all is needed for computing the working set with the
709 histogram. */
710 streamer_write_gcov_count_stream (ob->main_stream, profile_info->sum_all);
712 /* Create and output a bitpack of non-zero histogram entries indices. */
713 bp = bitpack_create (ob->main_stream);
714 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
715 bp_pack_value (&bp, profile_info->histogram[h_ix].num_counters > 0, 1);
716 streamer_write_bitpack (&bp);
717 /* Now stream out only those non-zero entries. */
718 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
720 if (!profile_info->histogram[h_ix].num_counters)
721 continue;
722 streamer_write_gcov_count_stream (ob->main_stream,
723 profile_info->histogram[h_ix].num_counters);
724 streamer_write_gcov_count_stream (ob->main_stream,
725 profile_info->histogram[h_ix].min_value);
726 streamer_write_gcov_count_stream (ob->main_stream,
727 profile_info->histogram[h_ix].cum_value);
729 /* IPA-profile computes hot bb threshold based on cumulated
730 whole program profile. We need to stream it down to ltrans. */
731 if (flag_wpa)
732 streamer_write_gcov_count_stream (ob->main_stream,
733 get_hot_bb_threshold ());
735 else
736 streamer_write_uhwi_stream (ob->main_stream, 0);
739 /* Output all callees or indirect outgoing edges. EDGE must be the first such
740 edge. */
742 static void
743 output_outgoing_cgraph_edges (struct cgraph_edge *edge,
744 struct lto_simple_output_block *ob,
745 lto_symtab_encoder_t encoder)
747 if (!edge)
748 return;
750 /* Output edges in backward direction, so the reconstructed callgraph match
751 and it is easy to associate call sites in the IPA pass summaries. */
752 while (edge->next_callee)
753 edge = edge->next_callee;
754 for (; edge; edge = edge->prev_callee)
755 lto_output_edge (ob, edge, encoder);
758 /* Output the part of the cgraph in SET. */
760 static void
761 output_refs (lto_symtab_encoder_t encoder)
763 struct lto_simple_output_block *ob;
764 int count;
765 struct ipa_ref *ref;
767 ob = lto_create_simple_output_block (LTO_section_refs);
769 for (int i = 0; i < lto_symtab_encoder_size (encoder); i++)
771 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
773 /* IPA_REF_ALIAS references are always preserved
774 in the boundary. Alias node can't have other references and
775 can be always handled as if it's not in the boundary. */
776 if (!node->alias && !lto_symtab_encoder_in_partition_p (encoder, node))
777 continue;
779 count = node->ref_list.nreferences ();
780 if (count)
782 streamer_write_gcov_count_stream (ob->main_stream, count);
783 streamer_write_uhwi_stream (ob->main_stream,
784 lto_symtab_encoder_lookup (encoder, node));
785 for (int i = 0; node->iterate_reference (i, ref); i++)
786 lto_output_ref (ob, ref, encoder);
790 streamer_write_uhwi_stream (ob->main_stream, 0);
792 lto_destroy_simple_output_block (ob);
795 /* Add NODE into encoder as well as nodes it is cloned from.
796 Do it in a way so clones appear first. */
798 static void
799 add_node_to (lto_symtab_encoder_t encoder, struct cgraph_node *node,
800 bool include_body)
802 if (node->clone_of)
803 add_node_to (encoder, node->clone_of, include_body);
804 else if (include_body)
805 lto_set_symtab_encoder_encode_body (encoder, node);
806 lto_symtab_encoder_encode (encoder, node);
809 /* Add all references in NODE to encoders. */
811 static void
812 create_references (lto_symtab_encoder_t encoder, symtab_node *node)
814 int i;
815 struct ipa_ref *ref = NULL;
816 for (i = 0; node->iterate_reference (i, ref); i++)
817 if (is_a <cgraph_node *> (ref->referred))
818 add_node_to (encoder, dyn_cast <cgraph_node *> (ref->referred), false);
819 else
820 lto_symtab_encoder_encode (encoder, ref->referred);
823 /* Select what needs to be streamed out. In regular lto mode stream everything.
824 In offload lto mode stream only nodes marked as offloadable. */
825 void
826 select_what_to_stream (void)
828 struct symtab_node *snode;
829 FOR_EACH_SYMBOL (snode)
830 snode->need_lto_streaming = !lto_stream_offload_p || snode->offloadable;
833 /* Find all symbols we want to stream into given partition and insert them
834 to encoders.
836 The function actually replaces IN_ENCODER by new one. The reason is that
837 streaming code needs clone's origin to be streamed before clone. This
838 means that we need to insert the nodes in specific order. This order is
839 ignored by the partitioning logic earlier. */
841 lto_symtab_encoder_t
842 compute_ltrans_boundary (lto_symtab_encoder_t in_encoder)
844 struct cgraph_edge *edge;
845 int i;
846 lto_symtab_encoder_t encoder;
847 lto_symtab_encoder_iterator lsei;
848 hash_set<void *> reachable_call_targets;
850 encoder = lto_symtab_encoder_new (false);
852 /* Go over all entries in the IN_ENCODER and duplicate them to
853 ENCODER. At the same time insert masters of clones so
854 every master appears before clone. */
855 for (lsei = lsei_start_function_in_partition (in_encoder);
856 !lsei_end_p (lsei); lsei_next_function_in_partition (&lsei))
858 struct cgraph_node *node = lsei_cgraph_node (lsei);
859 if (!node->need_lto_streaming)
860 continue;
861 add_node_to (encoder, node, true);
862 lto_set_symtab_encoder_in_partition (encoder, node);
863 create_references (encoder, node);
865 for (lsei = lsei_start_variable_in_partition (in_encoder);
866 !lsei_end_p (lsei); lsei_next_variable_in_partition (&lsei))
868 varpool_node *vnode = lsei_varpool_node (lsei);
870 if (!vnode->need_lto_streaming)
871 continue;
872 lto_set_symtab_encoder_in_partition (encoder, vnode);
873 lto_set_symtab_encoder_encode_initializer (encoder, vnode);
874 create_references (encoder, vnode);
876 /* Pickle in also the initializer of all referenced readonly variables
877 to help folding. Constant pool variables are not shared, so we must
878 pickle those too. */
879 for (i = 0; i < lto_symtab_encoder_size (encoder); i++)
881 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
882 if (varpool_node *vnode = dyn_cast <varpool_node *> (node))
884 if (!lto_symtab_encoder_encode_initializer_p (encoder,
885 vnode)
886 && (((vnode->ctor_useable_for_folding_p ()
887 && (!DECL_VIRTUAL_P (vnode->decl)
888 || !flag_wpa
889 || flag_ltrans_devirtualize)))))
891 lto_set_symtab_encoder_encode_initializer (encoder, vnode);
892 create_references (encoder, vnode);
897 /* Go over all the nodes again to include callees that are not in
898 SET. */
899 for (lsei = lsei_start_function_in_partition (encoder);
900 !lsei_end_p (lsei); lsei_next_function_in_partition (&lsei))
902 struct cgraph_node *node = lsei_cgraph_node (lsei);
903 for (edge = node->callees; edge; edge = edge->next_callee)
905 struct cgraph_node *callee = edge->callee;
906 if (!lto_symtab_encoder_in_partition_p (encoder, callee))
908 /* We should have moved all the inlines. */
909 gcc_assert (!callee->global.inlined_to);
910 add_node_to (encoder, callee, false);
913 /* Add all possible targets for late devirtualization. */
914 if (flag_ltrans_devirtualize || !flag_wpa)
915 for (edge = node->indirect_calls; edge; edge = edge->next_callee)
916 if (edge->indirect_info->polymorphic)
918 unsigned int i;
919 void *cache_token;
920 bool final;
921 vec <cgraph_node *>targets
922 = possible_polymorphic_call_targets
923 (edge, &final, &cache_token);
924 if (!reachable_call_targets.add (cache_token))
926 for (i = 0; i < targets.length (); i++)
928 struct cgraph_node *callee = targets[i];
930 /* Adding an external declarations into the unit serves
931 no purpose and just increases its boundary. */
932 if (callee->definition
933 && !lto_symtab_encoder_in_partition_p
934 (encoder, callee))
936 gcc_assert (!callee->global.inlined_to);
937 add_node_to (encoder, callee, false);
943 /* Be sure to also insert alias targert and thunk callees. These needs
944 to stay to aid local calling conventions. */
945 for (i = 0; i < lto_symtab_encoder_size (encoder); i++)
947 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
948 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
950 if (node->alias && node->analyzed)
951 create_references (encoder, node);
952 if (cnode
953 && cnode->thunk.thunk_p && !cnode->global.inlined_to)
954 add_node_to (encoder, cnode->callees->callee, false);
955 while (node->transparent_alias && node->analyzed)
957 node = node->get_alias_target ();
958 if (is_a <cgraph_node *> (node))
959 add_node_to (encoder, dyn_cast <cgraph_node *> (node),
960 false);
961 else
962 lto_symtab_encoder_encode (encoder, node);
965 lto_symtab_encoder_delete (in_encoder);
966 return encoder;
969 /* Output the part of the symtab in SET and VSET. */
971 void
972 output_symtab (void)
974 struct cgraph_node *node;
975 struct lto_simple_output_block *ob;
976 int i, n_nodes;
977 lto_symtab_encoder_t encoder;
979 if (flag_wpa)
980 output_cgraph_opt_summary ();
982 ob = lto_create_simple_output_block (LTO_section_symtab_nodes);
984 output_profile_summary (ob);
986 /* An encoder for cgraph nodes should have been created by
987 ipa_write_summaries_1. */
988 gcc_assert (ob->decl_state->symtab_node_encoder);
989 encoder = ob->decl_state->symtab_node_encoder;
991 /* Write out the nodes. We must first output a node and then its clones,
992 otherwise at a time reading back the node there would be nothing to clone
993 from. */
994 n_nodes = lto_symtab_encoder_size (encoder);
995 for (i = 0; i < n_nodes; i++)
997 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
998 if (cgraph_node *cnode = dyn_cast <cgraph_node *> (node))
999 lto_output_node (ob, cnode, encoder);
1000 else
1001 lto_output_varpool_node (ob, dyn_cast<varpool_node *> (node), encoder);
1004 /* Go over the nodes in SET again to write edges. */
1005 for (int i = 0; i < lto_symtab_encoder_size (encoder); i++)
1007 node = dyn_cast <cgraph_node *> (lto_symtab_encoder_deref (encoder, i));
1008 if (node
1009 && ((node->thunk.thunk_p && !node->global.inlined_to)
1010 || lto_symtab_encoder_in_partition_p (encoder, node)))
1012 output_outgoing_cgraph_edges (node->callees, ob, encoder);
1013 output_outgoing_cgraph_edges (node->indirect_calls, ob, encoder);
1017 streamer_write_uhwi_stream (ob->main_stream, 0);
1019 lto_destroy_simple_output_block (ob);
1021 /* Emit toplevel asms.
1022 When doing WPA we must output every asm just once. Since we do not partition asm
1023 nodes at all, output them to first output. This is kind of hack, but should work
1024 well. */
1025 if (!asm_nodes_output)
1027 asm_nodes_output = true;
1028 lto_output_toplevel_asms ();
1031 output_refs (encoder);
1034 /* Return identifier encoded in IB as a plain string. */
1036 static tree
1037 read_identifier (struct lto_input_block *ib)
1039 unsigned int len = strnlen (ib->data + ib->p, ib->len - ib->p - 1);
1040 tree id;
1042 if (ib->data[ib->p + len])
1043 lto_section_overrun (ib);
1044 if (!len)
1046 ib->p++;
1047 return NULL;
1049 id = get_identifier (ib->data + ib->p);
1050 ib->p += len + 1;
1051 return id;
1054 /* Return string encoded in IB, NULL if string is empty. */
1056 static const char *
1057 read_string (struct lto_input_block *ib)
1059 unsigned int len = strnlen (ib->data + ib->p, ib->len - ib->p - 1);
1060 const char *str;
1062 if (ib->data[ib->p + len])
1063 lto_section_overrun (ib);
1064 if (!len)
1066 ib->p++;
1067 return NULL;
1069 str = ib->data + ib->p;
1070 ib->p += len + 1;
1071 return str;
1074 /* Output function/variable tables that will allow libgomp to look up offload
1075 target code.
1076 OFFLOAD_FUNCS is filled in expand_omp_target, OFFLOAD_VARS is filled in
1077 varpool_node::get_create. In WHOPR (partitioned) mode during the WPA stage
1078 both OFFLOAD_FUNCS and OFFLOAD_VARS are filled by input_offload_tables. */
1080 void
1081 output_offload_tables (void)
1083 if (vec_safe_is_empty (offload_funcs) && vec_safe_is_empty (offload_vars))
1084 return;
1086 struct lto_simple_output_block *ob
1087 = lto_create_simple_output_block (LTO_section_offload_table);
1089 for (unsigned i = 0; i < vec_safe_length (offload_funcs); i++)
1091 streamer_write_enum (ob->main_stream, LTO_symtab_tags,
1092 LTO_symtab_last_tag, LTO_symtab_unavail_node);
1093 lto_output_fn_decl_index (ob->decl_state, ob->main_stream,
1094 (*offload_funcs)[i]);
1097 for (unsigned i = 0; i < vec_safe_length (offload_vars); i++)
1099 streamer_write_enum (ob->main_stream, LTO_symtab_tags,
1100 LTO_symtab_last_tag, LTO_symtab_variable);
1101 lto_output_var_decl_index (ob->decl_state, ob->main_stream,
1102 (*offload_vars)[i]);
1105 streamer_write_uhwi_stream (ob->main_stream, 0);
1106 lto_destroy_simple_output_block (ob);
1108 /* In WHOPR mode during the WPA stage the joint offload tables need to be
1109 streamed to one partition only. That's why we free offload_funcs and
1110 offload_vars after the first call of output_offload_tables. */
1111 if (flag_wpa)
1113 vec_free (offload_funcs);
1114 vec_free (offload_vars);
1118 /* Overwrite the information in NODE based on FILE_DATA, TAG, FLAGS,
1119 STACK_SIZE, SELF_TIME and SELF_SIZE. This is called either to initialize
1120 NODE or to replace the values in it, for instance because the first
1121 time we saw it, the function body was not available but now it
1122 is. BP is a bitpack with all the bitflags for NODE read from the
1123 stream. */
1125 static void
1126 input_overwrite_node (struct lto_file_decl_data *file_data,
1127 struct cgraph_node *node,
1128 enum LTO_symtab_tags tag,
1129 struct bitpack_d *bp)
1131 node->aux = (void *) tag;
1132 node->lto_file_data = file_data;
1134 node->local.local = bp_unpack_value (bp, 1);
1135 node->externally_visible = bp_unpack_value (bp, 1);
1136 node->no_reorder = bp_unpack_value (bp, 1);
1137 node->definition = bp_unpack_value (bp, 1);
1138 node->local.versionable = bp_unpack_value (bp, 1);
1139 node->local.can_change_signature = bp_unpack_value (bp, 1);
1140 node->local.redefined_extern_inline = bp_unpack_value (bp, 1);
1141 node->force_output = bp_unpack_value (bp, 1);
1142 node->forced_by_abi = bp_unpack_value (bp, 1);
1143 node->unique_name = bp_unpack_value (bp, 1);
1144 node->body_removed = bp_unpack_value (bp, 1);
1145 node->implicit_section = bp_unpack_value (bp, 1);
1146 node->address_taken = bp_unpack_value (bp, 1);
1147 node->used_from_other_partition = bp_unpack_value (bp, 1);
1148 node->lowered = bp_unpack_value (bp, 1);
1149 node->analyzed = tag == LTO_symtab_analyzed_node;
1150 node->in_other_partition = bp_unpack_value (bp, 1);
1151 if (node->in_other_partition
1152 /* Avoid updating decl when we are seeing just inline clone.
1153 When inlining function that has functions already inlined into it,
1154 we produce clones of inline clones.
1156 WPA partitioning might put each clone into different unit and
1157 we might end up streaming inline clone from other partition
1158 to support clone we are interested in. */
1159 && (!node->clone_of
1160 || node->clone_of->decl != node->decl))
1162 DECL_EXTERNAL (node->decl) = 1;
1163 TREE_STATIC (node->decl) = 0;
1165 node->alias = bp_unpack_value (bp, 1);
1166 node->transparent_alias = bp_unpack_value (bp, 1);
1167 node->weakref = bp_unpack_value (bp, 1);
1168 node->frequency = (enum node_frequency)bp_unpack_value (bp, 2);
1169 node->only_called_at_startup = bp_unpack_value (bp, 1);
1170 node->only_called_at_exit = bp_unpack_value (bp, 1);
1171 node->tm_clone = bp_unpack_value (bp, 1);
1172 node->calls_comdat_local = bp_unpack_value (bp, 1);
1173 node->icf_merged = bp_unpack_value (bp, 1);
1174 node->nonfreeing_fn = bp_unpack_value (bp, 1);
1175 node->thunk.thunk_p = bp_unpack_value (bp, 1);
1176 node->parallelized_function = bp_unpack_value (bp, 1);
1177 node->resolution = bp_unpack_enum (bp, ld_plugin_symbol_resolution,
1178 LDPR_NUM_KNOWN);
1179 node->split_part = bp_unpack_value (bp, 1);
1180 gcc_assert (flag_ltrans
1181 || (!node->in_other_partition
1182 && !node->used_from_other_partition));
1185 /* Return string alias is alias of. */
1187 static tree
1188 get_alias_symbol (tree decl)
1190 tree alias = lookup_attribute ("alias", DECL_ATTRIBUTES (decl));
1191 return get_identifier (TREE_STRING_POINTER
1192 (TREE_VALUE (TREE_VALUE (alias))));
1195 /* Read a node from input_block IB. TAG is the node's tag just read.
1196 Return the node read or overwriten. */
1198 static struct cgraph_node *
1199 input_node (struct lto_file_decl_data *file_data,
1200 struct lto_input_block *ib,
1201 enum LTO_symtab_tags tag,
1202 vec<symtab_node *> nodes)
1204 gcc::pass_manager *passes = g->get_passes ();
1205 tree fn_decl;
1206 struct cgraph_node *node;
1207 struct bitpack_d bp;
1208 unsigned decl_index;
1209 int ref = LCC_NOT_FOUND, ref2 = LCC_NOT_FOUND;
1210 int clone_ref;
1211 int order;
1212 int i, count;
1213 tree group;
1214 const char *section;
1215 order = streamer_read_hwi (ib) + order_base;
1216 clone_ref = streamer_read_hwi (ib);
1218 decl_index = streamer_read_uhwi (ib);
1219 fn_decl = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1221 if (clone_ref != LCC_NOT_FOUND)
1223 node = dyn_cast<cgraph_node *> (nodes[clone_ref])->create_clone (fn_decl,
1224 profile_count::uninitialized (), false,
1225 vNULL, false, NULL, NULL);
1227 else
1229 /* Declaration of functions can be already merged with a declaration
1230 from other input file. We keep cgraph unmerged until after streaming
1231 of ipa passes is done. Alays forcingly create a fresh node. */
1232 node = symtab->create_empty ();
1233 node->decl = fn_decl;
1234 if (lookup_attribute ("ifunc", DECL_ATTRIBUTES (fn_decl)))
1235 node->ifunc_resolver = 1;
1236 node->register_symbol ();
1239 node->order = order;
1240 if (order >= symtab->order)
1241 symtab->order = order + 1;
1243 node->count = profile_count::stream_in (ib);
1244 node->count_materialization_scale = streamer_read_hwi (ib);
1246 count = streamer_read_hwi (ib);
1247 node->ipa_transforms_to_apply = vNULL;
1248 for (i = 0; i < count; i++)
1250 opt_pass *pass;
1251 int pid = streamer_read_hwi (ib);
1253 gcc_assert (pid < passes->passes_by_id_size);
1254 pass = passes->passes_by_id[pid];
1255 node->ipa_transforms_to_apply.safe_push ((ipa_opt_pass_d *) pass);
1258 if (tag == LTO_symtab_analyzed_node)
1259 ref = streamer_read_hwi (ib);
1261 group = read_identifier (ib);
1262 if (group)
1263 ref2 = streamer_read_hwi (ib);
1265 /* Make sure that we have not read this node before. Nodes that
1266 have already been read will have their tag stored in the 'aux'
1267 field. Since built-in functions can be referenced in multiple
1268 functions, they are expected to be read more than once. */
1269 if (node->aux && !fndecl_built_in_p (node->decl))
1270 internal_error ("bytecode stream: found multiple instances of cgraph "
1271 "node with uid %d", node->get_uid ());
1273 node->tp_first_run = streamer_read_uhwi (ib);
1275 bp = streamer_read_bitpack (ib);
1277 input_overwrite_node (file_data, node, tag, &bp);
1279 /* Store a reference for now, and fix up later to be a pointer. */
1280 node->global.inlined_to = (cgraph_node *) (intptr_t) ref;
1282 if (group)
1284 node->set_comdat_group (group);
1285 /* Store a reference for now, and fix up later to be a pointer. */
1286 node->same_comdat_group = (symtab_node *) (intptr_t) ref2;
1288 else
1289 node->same_comdat_group = (symtab_node *) (intptr_t) LCC_NOT_FOUND;
1290 section = read_string (ib);
1291 if (section)
1292 node->set_section_for_node (section);
1294 if (node->thunk.thunk_p)
1296 int type = streamer_read_uhwi (ib);
1297 HOST_WIDE_INT fixed_offset = streamer_read_uhwi (ib);
1298 HOST_WIDE_INT virtual_value = streamer_read_uhwi (ib);
1300 node->thunk.fixed_offset = fixed_offset;
1301 node->thunk.this_adjusting = (type & 2);
1302 node->thunk.virtual_value = virtual_value;
1303 node->thunk.virtual_offset_p = (type & 4);
1304 node->thunk.add_pointer_bounds_args = (type & 8);
1306 if (node->alias && !node->analyzed && node->weakref)
1307 node->alias_target = get_alias_symbol (node->decl);
1308 node->profile_id = streamer_read_hwi (ib);
1309 if (DECL_STATIC_CONSTRUCTOR (node->decl))
1310 node->set_init_priority (streamer_read_hwi (ib));
1311 if (DECL_STATIC_DESTRUCTOR (node->decl))
1312 node->set_fini_priority (streamer_read_hwi (ib));
1314 return node;
1317 /* Read a node from input_block IB. TAG is the node's tag just read.
1318 Return the node read or overwriten. */
1320 static varpool_node *
1321 input_varpool_node (struct lto_file_decl_data *file_data,
1322 struct lto_input_block *ib)
1324 int decl_index;
1325 tree var_decl;
1326 varpool_node *node;
1327 struct bitpack_d bp;
1328 int ref = LCC_NOT_FOUND;
1329 int order;
1330 tree group;
1331 const char *section;
1333 order = streamer_read_hwi (ib) + order_base;
1334 decl_index = streamer_read_uhwi (ib);
1335 var_decl = lto_file_decl_data_get_var_decl (file_data, decl_index);
1337 /* Declaration of functions can be already merged with a declaration
1338 from other input file. We keep cgraph unmerged until after streaming
1339 of ipa passes is done. Alays forcingly create a fresh node. */
1340 node = varpool_node::create_empty ();
1341 node->decl = var_decl;
1342 node->register_symbol ();
1344 node->order = order;
1345 if (order >= symtab->order)
1346 symtab->order = order + 1;
1347 node->lto_file_data = file_data;
1349 bp = streamer_read_bitpack (ib);
1350 node->externally_visible = bp_unpack_value (&bp, 1);
1351 node->no_reorder = bp_unpack_value (&bp, 1);
1352 node->force_output = bp_unpack_value (&bp, 1);
1353 node->forced_by_abi = bp_unpack_value (&bp, 1);
1354 node->unique_name = bp_unpack_value (&bp, 1);
1355 node->body_removed = bp_unpack_value (&bp, 1);
1356 node->implicit_section = bp_unpack_value (&bp, 1);
1357 node->writeonly = bp_unpack_value (&bp, 1);
1358 node->definition = bp_unpack_value (&bp, 1);
1359 node->alias = bp_unpack_value (&bp, 1);
1360 node->transparent_alias = bp_unpack_value (&bp, 1);
1361 node->weakref = bp_unpack_value (&bp, 1);
1362 node->analyzed = bp_unpack_value (&bp, 1);
1363 node->used_from_other_partition = bp_unpack_value (&bp, 1);
1364 node->in_other_partition = bp_unpack_value (&bp, 1);
1365 if (node->in_other_partition)
1367 DECL_EXTERNAL (node->decl) = 1;
1368 TREE_STATIC (node->decl) = 0;
1370 if (node->alias && !node->analyzed && node->weakref)
1371 node->alias_target = get_alias_symbol (node->decl);
1372 node->tls_model = (enum tls_model)bp_unpack_value (&bp, 3);
1373 node->used_by_single_function = (enum tls_model)bp_unpack_value (&bp, 1);
1374 node->dynamically_initialized = bp_unpack_value (&bp, 1);
1375 node->need_bounds_init = bp_unpack_value (&bp, 1);
1376 group = read_identifier (ib);
1377 if (group)
1379 node->set_comdat_group (group);
1380 ref = streamer_read_hwi (ib);
1381 /* Store a reference for now, and fix up later to be a pointer. */
1382 node->same_comdat_group = (symtab_node *) (intptr_t) ref;
1384 else
1385 node->same_comdat_group = (symtab_node *) (intptr_t) LCC_NOT_FOUND;
1386 section = read_string (ib);
1387 if (section)
1388 node->set_section_for_node (section);
1389 node->resolution = streamer_read_enum (ib, ld_plugin_symbol_resolution,
1390 LDPR_NUM_KNOWN);
1391 gcc_assert (flag_ltrans
1392 || (!node->in_other_partition
1393 && !node->used_from_other_partition));
1395 return node;
1398 /* Read a node from input_block IB. TAG is the node's tag just read.
1399 Return the node read or overwriten. */
1401 static void
1402 input_ref (struct lto_input_block *ib,
1403 symtab_node *referring_node,
1404 vec<symtab_node *> nodes)
1406 symtab_node *node = NULL;
1407 struct bitpack_d bp;
1408 enum ipa_ref_use use;
1409 bool speculative;
1410 struct ipa_ref *ref;
1412 bp = streamer_read_bitpack (ib);
1413 use = (enum ipa_ref_use) bp_unpack_value (&bp, 3);
1414 speculative = (enum ipa_ref_use) bp_unpack_value (&bp, 1);
1415 node = nodes[streamer_read_hwi (ib)];
1416 ref = referring_node->create_reference (node, use);
1417 ref->speculative = speculative;
1418 if (is_a <cgraph_node *> (referring_node))
1419 ref->lto_stmt_uid = streamer_read_hwi (ib);
1422 /* Read an edge from IB. NODES points to a vector of previously read nodes for
1423 decoding caller and callee of the edge to be read. If INDIRECT is true, the
1424 edge being read is indirect (in the sense that it has
1425 indirect_unknown_callee set). */
1427 static void
1428 input_edge (struct lto_input_block *ib, vec<symtab_node *> nodes,
1429 bool indirect)
1431 struct cgraph_node *caller, *callee;
1432 struct cgraph_edge *edge;
1433 unsigned int stmt_id;
1434 profile_count count;
1435 cgraph_inline_failed_t inline_failed;
1436 struct bitpack_d bp;
1437 int ecf_flags = 0;
1439 caller = dyn_cast<cgraph_node *> (nodes[streamer_read_hwi (ib)]);
1440 if (caller == NULL || caller->decl == NULL_TREE)
1441 internal_error ("bytecode stream: no caller found while reading edge");
1443 if (!indirect)
1445 callee = dyn_cast<cgraph_node *> (nodes[streamer_read_hwi (ib)]);
1446 if (callee == NULL || callee->decl == NULL_TREE)
1447 internal_error ("bytecode stream: no callee found while reading edge");
1449 else
1450 callee = NULL;
1452 count = profile_count::stream_in (ib);
1454 bp = streamer_read_bitpack (ib);
1455 inline_failed = bp_unpack_enum (&bp, cgraph_inline_failed_t, CIF_N_REASONS);
1456 stmt_id = bp_unpack_var_len_unsigned (&bp);
1458 if (indirect)
1459 edge = caller->create_indirect_edge (NULL, 0, count);
1460 else
1461 edge = caller->create_edge (callee, NULL, count);
1463 edge->indirect_inlining_edge = bp_unpack_value (&bp, 1);
1464 edge->speculative = bp_unpack_value (&bp, 1);
1465 edge->lto_stmt_uid = stmt_id;
1466 edge->inline_failed = inline_failed;
1467 edge->call_stmt_cannot_inline_p = bp_unpack_value (&bp, 1);
1468 edge->can_throw_external = bp_unpack_value (&bp, 1);
1469 edge->in_polymorphic_cdtor = bp_unpack_value (&bp, 1);
1470 if (indirect)
1472 if (bp_unpack_value (&bp, 1))
1473 ecf_flags |= ECF_CONST;
1474 if (bp_unpack_value (&bp, 1))
1475 ecf_flags |= ECF_PURE;
1476 if (bp_unpack_value (&bp, 1))
1477 ecf_flags |= ECF_NORETURN;
1478 if (bp_unpack_value (&bp, 1))
1479 ecf_flags |= ECF_MALLOC;
1480 if (bp_unpack_value (&bp, 1))
1481 ecf_flags |= ECF_NOTHROW;
1482 if (bp_unpack_value (&bp, 1))
1483 ecf_flags |= ECF_RETURNS_TWICE;
1484 edge->indirect_info->ecf_flags = ecf_flags;
1485 edge->indirect_info->common_target_id = streamer_read_hwi (ib);
1486 if (edge->indirect_info->common_target_id)
1487 edge->indirect_info->common_target_probability = streamer_read_hwi (ib);
1492 /* Read a cgraph from IB using the info in FILE_DATA. */
1494 static vec<symtab_node *>
1495 input_cgraph_1 (struct lto_file_decl_data *file_data,
1496 struct lto_input_block *ib)
1498 enum LTO_symtab_tags tag;
1499 vec<symtab_node *> nodes = vNULL;
1500 symtab_node *node;
1501 unsigned i;
1503 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1504 order_base = symtab->order;
1505 while (tag)
1507 if (tag == LTO_symtab_edge)
1508 input_edge (ib, nodes, false);
1509 else if (tag == LTO_symtab_indirect_edge)
1510 input_edge (ib, nodes, true);
1511 else if (tag == LTO_symtab_variable)
1513 node = input_varpool_node (file_data, ib);
1514 nodes.safe_push (node);
1515 lto_symtab_encoder_encode (file_data->symtab_node_encoder, node);
1517 else
1519 node = input_node (file_data, ib, tag, nodes);
1520 if (node == NULL || node->decl == NULL_TREE)
1521 internal_error ("bytecode stream: found empty cgraph node");
1522 nodes.safe_push (node);
1523 lto_symtab_encoder_encode (file_data->symtab_node_encoder, node);
1526 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1529 lto_input_toplevel_asms (file_data, order_base);
1531 /* AUX pointers should be all non-zero for function nodes read from the stream. */
1532 if (flag_checking)
1534 FOR_EACH_VEC_ELT (nodes, i, node)
1535 gcc_assert (node->aux || !is_a <cgraph_node *> (node));
1537 FOR_EACH_VEC_ELT (nodes, i, node)
1539 int ref;
1540 if (cgraph_node *cnode = dyn_cast <cgraph_node *> (node))
1542 ref = (int) (intptr_t) cnode->global.inlined_to;
1544 /* We share declaration of builtins, so we may read same node twice. */
1545 if (!node->aux)
1546 continue;
1547 node->aux = NULL;
1549 /* Fixup inlined_to from reference to pointer. */
1550 if (ref != LCC_NOT_FOUND)
1551 dyn_cast<cgraph_node *> (node)->global.inlined_to
1552 = dyn_cast<cgraph_node *> (nodes[ref]);
1553 else
1554 cnode->global.inlined_to = NULL;
1557 ref = (int) (intptr_t) node->same_comdat_group;
1559 /* Fixup same_comdat_group from reference to pointer. */
1560 if (ref != LCC_NOT_FOUND)
1561 node->same_comdat_group = nodes[ref];
1562 else
1563 node->same_comdat_group = NULL;
1565 FOR_EACH_VEC_ELT (nodes, i, node)
1566 node->aux = is_a <cgraph_node *> (node) ? (void *)1 : NULL;
1567 return nodes;
1570 /* Input ipa_refs. */
1572 static void
1573 input_refs (struct lto_input_block *ib,
1574 vec<symtab_node *> nodes)
1576 int count;
1577 int idx;
1578 while (true)
1580 symtab_node *node;
1581 count = streamer_read_uhwi (ib);
1582 if (!count)
1583 break;
1584 idx = streamer_read_uhwi (ib);
1585 node = nodes[idx];
1586 while (count)
1588 input_ref (ib, node, nodes);
1589 count--;
1595 static gcov_summary lto_gcov_summary;
1597 /* Input profile_info from IB. */
1598 static void
1599 input_profile_summary (struct lto_input_block *ib,
1600 struct lto_file_decl_data *file_data)
1602 unsigned h_ix;
1603 struct bitpack_d bp;
1604 unsigned int runs = streamer_read_uhwi (ib);
1605 if (runs)
1607 file_data->profile_info.runs = runs;
1608 file_data->profile_info.sum_max = streamer_read_gcov_count (ib);
1609 file_data->profile_info.sum_all = streamer_read_gcov_count (ib);
1611 memset (file_data->profile_info.histogram, 0,
1612 sizeof (gcov_bucket_type) * GCOV_HISTOGRAM_SIZE);
1613 /* Input the bitpack of non-zero histogram indices. */
1614 bp = streamer_read_bitpack (ib);
1615 /* Read in and unpack the full bitpack, flagging non-zero
1616 histogram entries by setting the num_counters non-zero. */
1617 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1619 file_data->profile_info.histogram[h_ix].num_counters
1620 = bp_unpack_value (&bp, 1);
1622 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1624 if (!file_data->profile_info.histogram[h_ix].num_counters)
1625 continue;
1627 file_data->profile_info.histogram[h_ix].num_counters
1628 = streamer_read_gcov_count (ib);
1629 file_data->profile_info.histogram[h_ix].min_value
1630 = streamer_read_gcov_count (ib);
1631 file_data->profile_info.histogram[h_ix].cum_value
1632 = streamer_read_gcov_count (ib);
1634 /* IPA-profile computes hot bb threshold based on cumulated
1635 whole program profile. We need to stream it down to ltrans. */
1636 if (flag_ltrans)
1637 set_hot_bb_threshold (streamer_read_gcov_count (ib));
1642 /* Rescale profile summaries to the same number of runs in the whole unit. */
1644 static void
1645 merge_profile_summaries (struct lto_file_decl_data **file_data_vec)
1647 struct lto_file_decl_data *file_data;
1648 unsigned int j, h_ix;
1649 gcov_unsigned_t max_runs = 0;
1650 struct cgraph_node *node;
1651 struct cgraph_edge *edge;
1652 gcov_type saved_sum_all = 0;
1653 gcov_summary *saved_profile_info = 0;
1654 int saved_scale = 0;
1656 /* Find unit with maximal number of runs. If we ever get serious about
1657 roundoff errors, we might also consider computing smallest common
1658 multiply. */
1659 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1660 if (max_runs < file_data->profile_info.runs)
1661 max_runs = file_data->profile_info.runs;
1663 if (!max_runs)
1664 return;
1666 /* Simple overflow check. We probably don't need to support that many train
1667 runs. Such a large value probably imply data corruption anyway. */
1668 if (max_runs > INT_MAX / REG_BR_PROB_BASE)
1670 sorry ("At most %i profile runs is supported. Perhaps corrupted profile?",
1671 INT_MAX / REG_BR_PROB_BASE);
1672 return;
1675 profile_info = &lto_gcov_summary;
1676 lto_gcov_summary.runs = max_runs;
1677 lto_gcov_summary.sum_max = 0;
1678 memset (lto_gcov_summary.histogram, 0,
1679 sizeof (gcov_bucket_type) * GCOV_HISTOGRAM_SIZE);
1681 /* Rescale all units to the maximal number of runs.
1682 sum_max can not be easily merged, as we have no idea what files come from
1683 the same run. We do not use the info anyway, so leave it 0. */
1684 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1685 if (file_data->profile_info.runs)
1687 int scale = GCOV_COMPUTE_SCALE (max_runs,
1688 file_data->profile_info.runs);
1689 lto_gcov_summary.sum_max
1690 = MAX (lto_gcov_summary.sum_max,
1691 apply_scale (file_data->profile_info.sum_max, scale));
1692 lto_gcov_summary.sum_all
1693 = MAX (lto_gcov_summary.sum_all,
1694 apply_scale (file_data->profile_info.sum_all, scale));
1695 /* Save a pointer to the profile_info with the largest
1696 scaled sum_all and the scale for use in merging the
1697 histogram. */
1698 if (!saved_profile_info
1699 || lto_gcov_summary.sum_all > saved_sum_all)
1701 saved_profile_info = &file_data->profile_info;
1702 saved_sum_all = lto_gcov_summary.sum_all;
1703 saved_scale = scale;
1707 gcc_assert (saved_profile_info);
1709 /* Scale up the histogram from the profile that had the largest
1710 scaled sum_all above. */
1711 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1713 /* Scale up the min value as we did the corresponding sum_all
1714 above. Use that to find the new histogram index. */
1715 gcov_type scaled_min
1716 = apply_scale (saved_profile_info->histogram[h_ix].min_value,
1717 saved_scale);
1718 /* The new index may be shared with another scaled histogram entry,
1719 so we need to account for a non-zero histogram entry at new_ix. */
1720 unsigned new_ix = gcov_histo_index (scaled_min);
1721 lto_gcov_summary.histogram[new_ix].min_value
1722 = (lto_gcov_summary.histogram[new_ix].num_counters
1723 ? MIN (lto_gcov_summary.histogram[new_ix].min_value, scaled_min)
1724 : scaled_min);
1725 /* Some of the scaled counter values would ostensibly need to be placed
1726 into different (larger) histogram buckets, but we keep things simple
1727 here and place the scaled cumulative counter value in the bucket
1728 corresponding to the scaled minimum counter value. */
1729 lto_gcov_summary.histogram[new_ix].cum_value
1730 += apply_scale (saved_profile_info->histogram[h_ix].cum_value,
1731 saved_scale);
1732 lto_gcov_summary.histogram[new_ix].num_counters
1733 += saved_profile_info->histogram[h_ix].num_counters;
1736 /* Watch roundoff errors. */
1737 if (lto_gcov_summary.sum_max < max_runs)
1738 lto_gcov_summary.sum_max = max_runs;
1740 /* If merging already happent at WPA time, we are done. */
1741 if (flag_ltrans)
1742 return;
1744 /* Now compute count_materialization_scale of each node.
1745 During LTRANS we already have values of count_materialization_scale
1746 computed, so just update them. */
1747 FOR_EACH_FUNCTION (node)
1748 if (node->lto_file_data
1749 && node->lto_file_data->profile_info.runs)
1751 int scale;
1753 scale = RDIV (node->count_materialization_scale * max_runs,
1754 node->lto_file_data->profile_info.runs);
1755 node->count_materialization_scale = scale;
1756 if (scale < 0)
1757 fatal_error (input_location, "Profile information in %s corrupted",
1758 file_data->file_name);
1760 if (scale == REG_BR_PROB_BASE)
1761 continue;
1762 for (edge = node->callees; edge; edge = edge->next_callee)
1763 if (edge->count.ipa ().nonzero_p ())
1764 edge->count = edge->count.apply_scale (scale, REG_BR_PROB_BASE);
1765 for (edge = node->indirect_calls; edge; edge = edge->next_callee)
1766 if (edge->count.ipa ().nonzero_p ())
1767 edge->count = edge->count.apply_scale (scale, REG_BR_PROB_BASE);
1768 if (node->count.ipa ().nonzero_p ())
1769 node->count = node->count.apply_scale (scale, REG_BR_PROB_BASE);
1773 /* Input and merge the symtab from each of the .o files passed to
1774 lto1. */
1776 void
1777 input_symtab (void)
1779 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1780 struct lto_file_decl_data *file_data;
1781 unsigned int j = 0;
1782 struct cgraph_node *node;
1784 while ((file_data = file_data_vec[j++]))
1786 const char *data;
1787 size_t len;
1788 struct lto_input_block *ib;
1789 vec<symtab_node *> nodes;
1791 ib = lto_create_simple_input_block (file_data, LTO_section_symtab_nodes,
1792 &data, &len);
1793 if (!ib)
1794 fatal_error (input_location,
1795 "cannot find LTO cgraph in %s", file_data->file_name);
1796 input_profile_summary (ib, file_data);
1797 file_data->symtab_node_encoder = lto_symtab_encoder_new (true);
1798 nodes = input_cgraph_1 (file_data, ib);
1799 lto_destroy_simple_input_block (file_data, LTO_section_symtab_nodes,
1800 ib, data, len);
1802 ib = lto_create_simple_input_block (file_data, LTO_section_refs,
1803 &data, &len);
1804 if (!ib)
1805 fatal_error (input_location, "cannot find LTO section refs in %s",
1806 file_data->file_name);
1807 input_refs (ib, nodes);
1808 lto_destroy_simple_input_block (file_data, LTO_section_refs,
1809 ib, data, len);
1810 if (flag_ltrans)
1811 input_cgraph_opt_summary (nodes);
1812 nodes.release ();
1815 merge_profile_summaries (file_data_vec);
1817 if (!flag_auto_profile)
1818 get_working_sets ();
1821 /* Clear out the aux field that was used to store enough state to
1822 tell which nodes should be overwritten. */
1823 FOR_EACH_FUNCTION (node)
1825 /* Some nodes may have been created by cgraph_node. This
1826 happens when the callgraph contains nested functions. If the
1827 node for the parent function was never emitted to the gimple
1828 file, cgraph_node will create a node for it when setting the
1829 context of the nested function. */
1830 if (node->lto_file_data)
1831 node->aux = NULL;
1835 /* Input function/variable tables that will allow libgomp to look up offload
1836 target code, and store them into OFFLOAD_FUNCS and OFFLOAD_VARS. */
1838 void
1839 input_offload_tables (bool do_force_output)
1841 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1842 struct lto_file_decl_data *file_data;
1843 unsigned int j = 0;
1845 while ((file_data = file_data_vec[j++]))
1847 const char *data;
1848 size_t len;
1849 struct lto_input_block *ib
1850 = lto_create_simple_input_block (file_data, LTO_section_offload_table,
1851 &data, &len);
1852 if (!ib)
1853 continue;
1855 enum LTO_symtab_tags tag
1856 = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1857 while (tag)
1859 if (tag == LTO_symtab_unavail_node)
1861 int decl_index = streamer_read_uhwi (ib);
1862 tree fn_decl
1863 = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1864 vec_safe_push (offload_funcs, fn_decl);
1866 /* Prevent IPA from removing fn_decl as unreachable, since there
1867 may be no refs from the parent function to child_fn in offload
1868 LTO mode. */
1869 if (do_force_output)
1870 cgraph_node::get (fn_decl)->mark_force_output ();
1872 else if (tag == LTO_symtab_variable)
1874 int decl_index = streamer_read_uhwi (ib);
1875 tree var_decl
1876 = lto_file_decl_data_get_var_decl (file_data, decl_index);
1877 vec_safe_push (offload_vars, var_decl);
1879 /* Prevent IPA from removing var_decl as unused, since there
1880 may be no refs to var_decl in offload LTO mode. */
1881 if (do_force_output)
1882 varpool_node::get (var_decl)->force_output = 1;
1884 else
1885 fatal_error (input_location,
1886 "invalid offload table in %s", file_data->file_name);
1888 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1891 lto_destroy_simple_input_block (file_data, LTO_section_offload_table,
1892 ib, data, len);
1896 /* True when we need optimization summary for NODE. */
1898 static int
1899 output_cgraph_opt_summary_p (struct cgraph_node *node)
1901 return ((node->clone_of || node->former_clone_of)
1902 && (node->clone.tree_map
1903 || node->clone.args_to_skip
1904 || node->clone.combined_args_to_skip));
1907 /* Output optimization summary for EDGE to OB. */
1908 static void
1909 output_edge_opt_summary (struct output_block *ob ATTRIBUTE_UNUSED,
1910 struct cgraph_edge *edge ATTRIBUTE_UNUSED)
1914 /* Output optimization summary for NODE to OB. */
1916 static void
1917 output_node_opt_summary (struct output_block *ob,
1918 struct cgraph_node *node,
1919 lto_symtab_encoder_t encoder)
1921 unsigned int index;
1922 bitmap_iterator bi;
1923 struct ipa_replace_map *map;
1924 struct bitpack_d bp;
1925 int i;
1926 struct cgraph_edge *e;
1928 if (node->clone.args_to_skip)
1930 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.args_to_skip));
1931 EXECUTE_IF_SET_IN_BITMAP (node->clone.args_to_skip, 0, index, bi)
1932 streamer_write_uhwi (ob, index);
1934 else
1935 streamer_write_uhwi (ob, 0);
1936 if (node->clone.combined_args_to_skip)
1938 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.combined_args_to_skip));
1939 EXECUTE_IF_SET_IN_BITMAP (node->clone.combined_args_to_skip, 0, index, bi)
1940 streamer_write_uhwi (ob, index);
1942 else
1943 streamer_write_uhwi (ob, 0);
1944 streamer_write_uhwi (ob, vec_safe_length (node->clone.tree_map));
1945 FOR_EACH_VEC_SAFE_ELT (node->clone.tree_map, i, map)
1947 /* At the moment we assume all old trees to be PARM_DECLs, because we have no
1948 mechanism to store function local declarations into summaries. */
1949 gcc_assert (!map->old_tree);
1950 streamer_write_uhwi (ob, map->parm_num);
1951 gcc_assert (EXPR_LOCATION (map->new_tree) == UNKNOWN_LOCATION);
1952 stream_write_tree (ob, map->new_tree, true);
1953 bp = bitpack_create (ob->main_stream);
1954 bp_pack_value (&bp, map->replace_p, 1);
1955 bp_pack_value (&bp, map->ref_p, 1);
1956 streamer_write_bitpack (&bp);
1959 if (lto_symtab_encoder_in_partition_p (encoder, node))
1961 for (e = node->callees; e; e = e->next_callee)
1962 output_edge_opt_summary (ob, e);
1963 for (e = node->indirect_calls; e; e = e->next_callee)
1964 output_edge_opt_summary (ob, e);
1968 /* Output optimization summaries stored in callgraph.
1969 At the moment it is the clone info structure. */
1971 static void
1972 output_cgraph_opt_summary (void)
1974 int i, n_nodes;
1975 lto_symtab_encoder_t encoder;
1976 struct output_block *ob = create_output_block (LTO_section_cgraph_opt_sum);
1977 unsigned count = 0;
1979 ob->symbol = NULL;
1980 encoder = ob->decl_state->symtab_node_encoder;
1981 n_nodes = lto_symtab_encoder_size (encoder);
1982 for (i = 0; i < n_nodes; i++)
1984 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
1985 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
1986 if (cnode && output_cgraph_opt_summary_p (cnode))
1987 count++;
1989 streamer_write_uhwi (ob, count);
1990 for (i = 0; i < n_nodes; i++)
1992 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
1993 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
1994 if (cnode && output_cgraph_opt_summary_p (cnode))
1996 streamer_write_uhwi (ob, i);
1997 output_node_opt_summary (ob, cnode, encoder);
2000 produce_asm (ob, NULL);
2001 destroy_output_block (ob);
2004 /* Input optimisation summary of EDGE. */
2006 static void
2007 input_edge_opt_summary (struct cgraph_edge *edge ATTRIBUTE_UNUSED,
2008 struct lto_input_block *ib_main ATTRIBUTE_UNUSED)
2012 /* Input optimisation summary of NODE. */
2014 static void
2015 input_node_opt_summary (struct cgraph_node *node,
2016 struct lto_input_block *ib_main,
2017 struct data_in *data_in)
2019 int i;
2020 int count;
2021 int bit;
2022 struct bitpack_d bp;
2023 struct cgraph_edge *e;
2025 count = streamer_read_uhwi (ib_main);
2026 if (count)
2027 node->clone.args_to_skip = BITMAP_GGC_ALLOC ();
2028 for (i = 0; i < count; i++)
2030 bit = streamer_read_uhwi (ib_main);
2031 bitmap_set_bit (node->clone.args_to_skip, bit);
2033 count = streamer_read_uhwi (ib_main);
2034 if (count)
2035 node->clone.combined_args_to_skip = BITMAP_GGC_ALLOC ();
2036 for (i = 0; i < count; i++)
2038 bit = streamer_read_uhwi (ib_main);
2039 bitmap_set_bit (node->clone.combined_args_to_skip, bit);
2041 count = streamer_read_uhwi (ib_main);
2042 for (i = 0; i < count; i++)
2044 struct ipa_replace_map *map = ggc_alloc<ipa_replace_map> ();
2046 vec_safe_push (node->clone.tree_map, map);
2047 map->parm_num = streamer_read_uhwi (ib_main);
2048 map->old_tree = NULL;
2049 map->new_tree = stream_read_tree (ib_main, data_in);
2050 bp = streamer_read_bitpack (ib_main);
2051 map->replace_p = bp_unpack_value (&bp, 1);
2052 map->ref_p = bp_unpack_value (&bp, 1);
2054 for (e = node->callees; e; e = e->next_callee)
2055 input_edge_opt_summary (e, ib_main);
2056 for (e = node->indirect_calls; e; e = e->next_callee)
2057 input_edge_opt_summary (e, ib_main);
2060 /* Read section in file FILE_DATA of length LEN with data DATA. */
2062 static void
2063 input_cgraph_opt_section (struct lto_file_decl_data *file_data,
2064 const char *data, size_t len,
2065 vec<symtab_node *> nodes)
2067 const struct lto_function_header *header =
2068 (const struct lto_function_header *) data;
2069 const int cfg_offset = sizeof (struct lto_function_header);
2070 const int main_offset = cfg_offset + header->cfg_size;
2071 const int string_offset = main_offset + header->main_size;
2072 struct data_in *data_in;
2073 unsigned int i;
2074 unsigned int count;
2076 lto_input_block ib_main ((const char *) data + main_offset,
2077 header->main_size, file_data->mode_table);
2079 data_in =
2080 lto_data_in_create (file_data, (const char *) data + string_offset,
2081 header->string_size, vNULL);
2082 count = streamer_read_uhwi (&ib_main);
2084 for (i = 0; i < count; i++)
2086 int ref = streamer_read_uhwi (&ib_main);
2087 input_node_opt_summary (dyn_cast<cgraph_node *> (nodes[ref]),
2088 &ib_main, data_in);
2090 lto_free_section_data (file_data, LTO_section_cgraph_opt_sum, NULL, data,
2091 len);
2092 lto_data_in_delete (data_in);
2095 /* Input optimization summary of cgraph. */
2097 static void
2098 input_cgraph_opt_summary (vec<symtab_node *> nodes)
2100 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
2101 struct lto_file_decl_data *file_data;
2102 unsigned int j = 0;
2104 while ((file_data = file_data_vec[j++]))
2106 size_t len;
2107 const char *data =
2108 lto_get_section_data (file_data, LTO_section_cgraph_opt_sum, NULL,
2109 &len);
2111 if (data)
2112 input_cgraph_opt_section (file_data, data, len, nodes);