2016-05-04 Thomas Preud'homme <thomas.preudhomme@arm.com>
[official-gcc.git] / gcc / lto-cgraph.c
blob14302138efa050a69bb9d1250715cd6947fb9ffe
1 /* Write and read the cgraph to the memory mapped representation of a
2 .o file.
4 Copyright (C) 2009-2016 Free Software Foundation, Inc.
5 Contributed by Kenneth Zadeck <zadeck@naturalbridge.com>
7 This file is part of GCC.
9 GCC is free software; you can redistribute it and/or modify it under
10 the terms of the GNU General Public License as published by the Free
11 Software Foundation; either version 3, or (at your option) any later
12 version.
14 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
15 WARRANTY; without even the implied warranty of MERCHANTABILITY or
16 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 for more details.
19 You should have received a copy of the GNU General Public License
20 along with GCC; see the file COPYING3. If not see
21 <http://www.gnu.org/licenses/>. */
23 #include "config.h"
24 #include "system.h"
25 #include "coretypes.h"
26 #include "backend.h"
27 #include "rtl.h"
28 #include "tree.h"
29 #include "gimple.h"
30 #include "predict.h"
31 #include "stringpool.h"
32 #include "tree-streamer.h"
33 #include "cgraph.h"
34 #include "tree-pass.h"
35 #include "profile.h"
36 #include "context.h"
37 #include "pass_manager.h"
38 #include "ipa-utils.h"
39 #include "omp-low.h"
40 #include "ipa-chkp.h"
42 /* True when asm nodes has been output. */
43 bool asm_nodes_output = false;
45 static void output_cgraph_opt_summary (void);
46 static void input_cgraph_opt_summary (vec<symtab_node *> nodes);
48 /* Number of LDPR values known to GCC. */
49 #define LDPR_NUM_KNOWN (LDPR_PREVAILING_DEF_IRONLY_EXP + 1)
51 /* All node orders are ofsetted by ORDER_BASE. */
52 static int order_base;
54 /* Cgraph streaming is organized as set of record whose type
55 is indicated by a tag. */
56 enum LTO_symtab_tags
58 /* Must leave 0 for the stopper. */
60 /* Cgraph node without body available. */
61 LTO_symtab_unavail_node = 1,
62 /* Cgraph node with function body. */
63 LTO_symtab_analyzed_node,
64 /* Cgraph edges. */
65 LTO_symtab_edge,
66 LTO_symtab_indirect_edge,
67 LTO_symtab_variable,
68 LTO_symtab_last_tag
71 /* Create a new symtab encoder.
72 if FOR_INPUT, the encoder allocate only datastructures needed
73 to read the symtab. */
75 lto_symtab_encoder_t
76 lto_symtab_encoder_new (bool for_input)
78 lto_symtab_encoder_t encoder = XCNEW (struct lto_symtab_encoder_d);
80 if (!for_input)
81 encoder->map = new hash_map<symtab_node *, size_t>;
82 encoder->nodes.create (0);
83 return encoder;
87 /* Delete ENCODER and its components. */
89 void
90 lto_symtab_encoder_delete (lto_symtab_encoder_t encoder)
92 encoder->nodes.release ();
93 if (encoder->map)
94 delete encoder->map;
95 free (encoder);
99 /* Return the existing reference number of NODE in the symtab encoder in
100 output block OB. Assign a new reference if this is the first time
101 NODE is encoded. */
104 lto_symtab_encoder_encode (lto_symtab_encoder_t encoder,
105 symtab_node *node)
107 int ref;
109 if (!encoder->map)
111 lto_encoder_entry entry = {node, false, false, false};
113 ref = encoder->nodes.length ();
114 encoder->nodes.safe_push (entry);
115 return ref;
118 size_t *slot = encoder->map->get (node);
119 if (!slot || !*slot)
121 lto_encoder_entry entry = {node, false, false, false};
122 ref = encoder->nodes.length ();
123 if (!slot)
124 encoder->map->put (node, ref + 1);
125 encoder->nodes.safe_push (entry);
127 else
128 ref = *slot - 1;
130 return ref;
133 /* Remove NODE from encoder. */
135 bool
136 lto_symtab_encoder_delete_node (lto_symtab_encoder_t encoder,
137 symtab_node *node)
139 int index;
140 lto_encoder_entry last_node;
142 size_t *slot = encoder->map->get (node);
143 if (slot == NULL || !*slot)
144 return false;
146 index = *slot - 1;
147 gcc_checking_assert (encoder->nodes[index].node == node);
149 /* Remove from vector. We do this by swapping node with the last element
150 of the vector. */
151 last_node = encoder->nodes.pop ();
152 if (last_node.node != node)
154 gcc_assert (encoder->map->put (last_node.node, index + 1));
156 /* Move the last element to the original spot of NODE. */
157 encoder->nodes[index] = last_node;
160 /* Remove element from hash table. */
161 encoder->map->remove (node);
162 return true;
166 /* Return TRUE if we should encode the body of NODE (if any). */
168 bool
169 lto_symtab_encoder_encode_body_p (lto_symtab_encoder_t encoder,
170 struct cgraph_node *node)
172 int index = lto_symtab_encoder_lookup (encoder, node);
173 return encoder->nodes[index].body;
176 /* Specify that we encode the body of NODE in this partition. */
178 static void
179 lto_set_symtab_encoder_encode_body (lto_symtab_encoder_t encoder,
180 struct cgraph_node *node)
182 int index = lto_symtab_encoder_encode (encoder, node);
183 gcc_checking_assert (encoder->nodes[index].node == node);
184 encoder->nodes[index].body = true;
187 /* Return TRUE if we should encode initializer of NODE (if any). */
189 bool
190 lto_symtab_encoder_encode_initializer_p (lto_symtab_encoder_t encoder,
191 varpool_node *node)
193 int index = lto_symtab_encoder_lookup (encoder, node);
194 if (index == LCC_NOT_FOUND)
195 return false;
196 return encoder->nodes[index].initializer;
199 /* Specify that we should encode initializer of NODE (if any). */
201 static void
202 lto_set_symtab_encoder_encode_initializer (lto_symtab_encoder_t encoder,
203 varpool_node *node)
205 int index = lto_symtab_encoder_lookup (encoder, node);
206 encoder->nodes[index].initializer = true;
209 /* Return TRUE if NODE is in this partition. */
211 bool
212 lto_symtab_encoder_in_partition_p (lto_symtab_encoder_t encoder,
213 symtab_node *node)
215 int index = lto_symtab_encoder_lookup (encoder, node);
216 if (index == LCC_NOT_FOUND)
217 return false;
218 return encoder->nodes[index].in_partition;
221 /* Specify that NODE is in this partition. */
223 void
224 lto_set_symtab_encoder_in_partition (lto_symtab_encoder_t encoder,
225 symtab_node *node)
227 int index = lto_symtab_encoder_encode (encoder, node);
228 encoder->nodes[index].in_partition = true;
231 /* Output the cgraph EDGE to OB using ENCODER. */
233 static void
234 lto_output_edge (struct lto_simple_output_block *ob, struct cgraph_edge *edge,
235 lto_symtab_encoder_t encoder)
237 unsigned int uid;
238 intptr_t ref;
239 struct bitpack_d bp;
241 if (edge->indirect_unknown_callee)
242 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
243 LTO_symtab_indirect_edge);
244 else
245 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
246 LTO_symtab_edge);
248 ref = lto_symtab_encoder_lookup (encoder, edge->caller);
249 gcc_assert (ref != LCC_NOT_FOUND);
250 streamer_write_hwi_stream (ob->main_stream, ref);
252 if (!edge->indirect_unknown_callee)
254 ref = lto_symtab_encoder_lookup (encoder, edge->callee);
255 gcc_assert (ref != LCC_NOT_FOUND);
256 streamer_write_hwi_stream (ob->main_stream, ref);
259 streamer_write_gcov_count_stream (ob->main_stream, edge->count);
261 bp = bitpack_create (ob->main_stream);
262 uid = (!gimple_has_body_p (edge->caller->decl)
263 ? edge->lto_stmt_uid : gimple_uid (edge->call_stmt) + 1);
264 bp_pack_enum (&bp, cgraph_inline_failed_t,
265 CIF_N_REASONS, edge->inline_failed);
266 bp_pack_var_len_unsigned (&bp, uid);
267 bp_pack_var_len_unsigned (&bp, edge->frequency);
268 bp_pack_value (&bp, edge->indirect_inlining_edge, 1);
269 bp_pack_value (&bp, edge->speculative, 1);
270 bp_pack_value (&bp, edge->call_stmt_cannot_inline_p, 1);
271 gcc_assert (!edge->call_stmt_cannot_inline_p
272 || edge->inline_failed != CIF_BODY_NOT_AVAILABLE);
273 bp_pack_value (&bp, edge->can_throw_external, 1);
274 bp_pack_value (&bp, edge->in_polymorphic_cdtor, 1);
275 if (edge->indirect_unknown_callee)
277 int flags = edge->indirect_info->ecf_flags;
278 bp_pack_value (&bp, (flags & ECF_CONST) != 0, 1);
279 bp_pack_value (&bp, (flags & ECF_PURE) != 0, 1);
280 bp_pack_value (&bp, (flags & ECF_NORETURN) != 0, 1);
281 bp_pack_value (&bp, (flags & ECF_MALLOC) != 0, 1);
282 bp_pack_value (&bp, (flags & ECF_NOTHROW) != 0, 1);
283 bp_pack_value (&bp, (flags & ECF_RETURNS_TWICE) != 0, 1);
284 /* Flags that should not appear on indirect calls. */
285 gcc_assert (!(flags & (ECF_LOOPING_CONST_OR_PURE
286 | ECF_MAY_BE_ALLOCA
287 | ECF_SIBCALL
288 | ECF_LEAF
289 | ECF_NOVOPS)));
291 streamer_write_bitpack (&bp);
292 if (edge->indirect_unknown_callee)
294 streamer_write_hwi_stream (ob->main_stream,
295 edge->indirect_info->common_target_id);
296 if (edge->indirect_info->common_target_id)
297 streamer_write_hwi_stream
298 (ob->main_stream, edge->indirect_info->common_target_probability);
302 /* Return if NODE contain references from other partitions. */
304 bool
305 referenced_from_other_partition_p (symtab_node *node, lto_symtab_encoder_t encoder)
307 int i;
308 struct ipa_ref *ref = NULL;
310 for (i = 0; node->iterate_referring (i, ref); i++)
312 /* Ignore references from non-offloadable nodes while streaming NODE into
313 offload LTO section. */
314 if (!ref->referring->need_lto_streaming)
315 continue;
317 if (ref->referring->in_other_partition
318 || !lto_symtab_encoder_in_partition_p (encoder, ref->referring))
319 return true;
321 return false;
324 /* Return true when node is reachable from other partition. */
326 bool
327 reachable_from_other_partition_p (struct cgraph_node *node, lto_symtab_encoder_t encoder)
329 struct cgraph_edge *e;
330 if (!node->definition)
331 return false;
332 if (node->global.inlined_to)
333 return false;
334 for (e = node->callers; e; e = e->next_caller)
336 /* Ignore references from non-offloadable nodes while streaming NODE into
337 offload LTO section. */
338 if (!e->caller->need_lto_streaming)
339 continue;
341 if (e->caller->in_other_partition
342 || !lto_symtab_encoder_in_partition_p (encoder, e->caller))
343 return true;
345 return false;
348 /* Return if NODE contain references from other partitions. */
350 bool
351 referenced_from_this_partition_p (symtab_node *node,
352 lto_symtab_encoder_t encoder)
354 int i;
355 struct ipa_ref *ref = NULL;
357 for (i = 0; node->iterate_referring (i, ref); i++)
358 if (lto_symtab_encoder_in_partition_p (encoder, ref->referring))
359 return true;
360 return false;
363 /* Return true when node is reachable from other partition. */
365 bool
366 reachable_from_this_partition_p (struct cgraph_node *node, lto_symtab_encoder_t encoder)
368 struct cgraph_edge *e;
369 for (e = node->callers; e; e = e->next_caller)
370 if (lto_symtab_encoder_in_partition_p (encoder, e->caller))
371 return true;
372 return false;
375 /* Output the cgraph NODE to OB. ENCODER is used to find the
376 reference number of NODE->inlined_to. SET is the set of nodes we
377 are writing to the current file. If NODE is not in SET, then NODE
378 is a boundary of a cgraph_node_set and we pretend NODE just has a
379 decl and no callees. WRITTEN_DECLS is the set of FUNCTION_DECLs
380 that have had their callgraph node written so far. This is used to
381 determine if NODE is a clone of a previously written node. */
383 static void
384 lto_output_node (struct lto_simple_output_block *ob, struct cgraph_node *node,
385 lto_symtab_encoder_t encoder)
387 unsigned int tag;
388 struct bitpack_d bp;
389 bool boundary_p;
390 intptr_t ref;
391 bool in_other_partition = false;
392 struct cgraph_node *clone_of, *ultimate_clone_of;
393 ipa_opt_pass_d *pass;
394 int i;
395 const char *comdat;
396 const char *section;
397 tree group;
399 boundary_p = !lto_symtab_encoder_in_partition_p (encoder, node);
401 if (node->analyzed && (!boundary_p || node->alias || node->thunk.thunk_p))
402 tag = LTO_symtab_analyzed_node;
403 else
404 tag = LTO_symtab_unavail_node;
406 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
407 tag);
408 streamer_write_hwi_stream (ob->main_stream, node->order);
410 /* In WPA mode, we only output part of the call-graph. Also, we
411 fake cgraph node attributes. There are two cases that we care.
413 Boundary nodes: There are nodes that are not part of SET but are
414 called from within SET. We artificially make them look like
415 externally visible nodes with no function body.
417 Cherry-picked nodes: These are nodes we pulled from other
418 translation units into SET during IPA-inlining. We make them as
419 local static nodes to prevent clashes with other local statics. */
420 if (boundary_p && node->analyzed
421 && node->get_partitioning_class () == SYMBOL_PARTITION)
423 /* Inline clones can not be part of boundary.
424 gcc_assert (!node->global.inlined_to);
426 FIXME: At the moment they can be, when partition contains an inline
427 clone that is clone of inline clone from outside partition. We can
428 reshape the clone tree and make other tree to be the root, but it
429 needs a bit extra work and will be promplty done by cgraph_remove_node
430 after reading back. */
431 in_other_partition = 1;
434 clone_of = node->clone_of;
435 while (clone_of
436 && (ref = lto_symtab_encoder_lookup (encoder, clone_of)) == LCC_NOT_FOUND)
437 if (clone_of->prev_sibling_clone)
438 clone_of = clone_of->prev_sibling_clone;
439 else
440 clone_of = clone_of->clone_of;
442 /* See if body of the master function is output. If not, we are seeing only
443 an declaration and we do not need to pass down clone tree. */
444 ultimate_clone_of = clone_of;
445 while (ultimate_clone_of && ultimate_clone_of->clone_of)
446 ultimate_clone_of = ultimate_clone_of->clone_of;
448 if (clone_of && !lto_symtab_encoder_encode_body_p (encoder, ultimate_clone_of))
449 clone_of = NULL;
451 if (tag == LTO_symtab_analyzed_node)
452 gcc_assert (clone_of || !node->clone_of);
453 if (!clone_of)
454 streamer_write_hwi_stream (ob->main_stream, LCC_NOT_FOUND);
455 else
456 streamer_write_hwi_stream (ob->main_stream, ref);
459 lto_output_fn_decl_index (ob->decl_state, ob->main_stream, node->decl);
460 streamer_write_gcov_count_stream (ob->main_stream, node->count);
461 streamer_write_hwi_stream (ob->main_stream, node->count_materialization_scale);
463 streamer_write_hwi_stream (ob->main_stream,
464 node->ipa_transforms_to_apply.length ());
465 FOR_EACH_VEC_ELT (node->ipa_transforms_to_apply, i, pass)
466 streamer_write_hwi_stream (ob->main_stream, pass->static_pass_number);
468 if (tag == LTO_symtab_analyzed_node)
470 if (node->global.inlined_to)
472 ref = lto_symtab_encoder_lookup (encoder, node->global.inlined_to);
473 gcc_assert (ref != LCC_NOT_FOUND);
475 else
476 ref = LCC_NOT_FOUND;
478 streamer_write_hwi_stream (ob->main_stream, ref);
481 group = node->get_comdat_group ();
482 if (group)
483 comdat = IDENTIFIER_POINTER (group);
484 else
485 comdat = "";
486 streamer_write_data_stream (ob->main_stream, comdat, strlen (comdat) + 1);
488 if (group)
490 if (node->same_comdat_group)
492 ref = LCC_NOT_FOUND;
493 for (struct symtab_node *n = node->same_comdat_group;
494 ref == LCC_NOT_FOUND && n != node; n = n->same_comdat_group)
495 ref = lto_symtab_encoder_lookup (encoder, n);
497 else
498 ref = LCC_NOT_FOUND;
499 streamer_write_hwi_stream (ob->main_stream, ref);
502 section = node->get_section ();
503 if (!section)
504 section = "";
506 streamer_write_hwi_stream (ob->main_stream, node->tp_first_run);
508 bp = bitpack_create (ob->main_stream);
509 bp_pack_value (&bp, node->local.local, 1);
510 bp_pack_value (&bp, node->externally_visible, 1);
511 bp_pack_value (&bp, node->no_reorder, 1);
512 bp_pack_value (&bp, node->definition, 1);
513 bp_pack_value (&bp, node->local.versionable, 1);
514 bp_pack_value (&bp, node->local.can_change_signature, 1);
515 bp_pack_value (&bp, node->local.redefined_extern_inline, 1);
516 bp_pack_value (&bp, node->force_output, 1);
517 bp_pack_value (&bp, node->forced_by_abi, 1);
518 bp_pack_value (&bp, node->unique_name, 1);
519 bp_pack_value (&bp, node->body_removed, 1);
520 bp_pack_value (&bp, node->implicit_section, 1);
521 bp_pack_value (&bp, node->address_taken, 1);
522 bp_pack_value (&bp, tag == LTO_symtab_analyzed_node
523 && node->get_partitioning_class () == SYMBOL_PARTITION
524 && (reachable_from_other_partition_p (node, encoder)
525 || referenced_from_other_partition_p (node, encoder)), 1);
526 bp_pack_value (&bp, node->lowered, 1);
527 bp_pack_value (&bp, in_other_partition, 1);
528 bp_pack_value (&bp, node->alias, 1);
529 bp_pack_value (&bp, node->transparent_alias, 1);
530 bp_pack_value (&bp, node->weakref, 1);
531 bp_pack_value (&bp, node->frequency, 2);
532 bp_pack_value (&bp, node->only_called_at_startup, 1);
533 bp_pack_value (&bp, node->only_called_at_exit, 1);
534 bp_pack_value (&bp, node->tm_clone, 1);
535 bp_pack_value (&bp, node->calls_comdat_local, 1);
536 bp_pack_value (&bp, node->icf_merged, 1);
537 bp_pack_value (&bp, node->nonfreeing_fn, 1);
538 bp_pack_value (&bp, node->thunk.thunk_p, 1);
539 bp_pack_value (&bp, node->parallelized_function, 1);
540 bp_pack_enum (&bp, ld_plugin_symbol_resolution,
541 LDPR_NUM_KNOWN, node->resolution);
542 bp_pack_value (&bp, node->instrumentation_clone, 1);
543 bp_pack_value (&bp, node->split_part, 1);
544 streamer_write_bitpack (&bp);
545 streamer_write_data_stream (ob->main_stream, section, strlen (section) + 1);
547 if (node->thunk.thunk_p)
549 streamer_write_uhwi_stream
550 (ob->main_stream,
551 1 + (node->thunk.this_adjusting != 0) * 2
552 + (node->thunk.virtual_offset_p != 0) * 4
553 + (node->thunk.add_pointer_bounds_args != 0) * 8);
554 streamer_write_uhwi_stream (ob->main_stream, node->thunk.fixed_offset);
555 streamer_write_uhwi_stream (ob->main_stream, node->thunk.virtual_value);
557 streamer_write_hwi_stream (ob->main_stream, node->profile_id);
558 if (DECL_STATIC_CONSTRUCTOR (node->decl))
559 streamer_write_hwi_stream (ob->main_stream, node->get_init_priority ());
560 if (DECL_STATIC_DESTRUCTOR (node->decl))
561 streamer_write_hwi_stream (ob->main_stream, node->get_fini_priority ());
563 if (node->instrumentation_clone)
564 lto_output_fn_decl_index (ob->decl_state, ob->main_stream, node->orig_decl);
567 /* Output the varpool NODE to OB.
568 If NODE is not in SET, then NODE is a boundary. */
570 static void
571 lto_output_varpool_node (struct lto_simple_output_block *ob, varpool_node *node,
572 lto_symtab_encoder_t encoder)
574 bool boundary_p = !lto_symtab_encoder_in_partition_p (encoder, node);
575 bool encode_initializer_p
576 = (node->definition
577 && lto_symtab_encoder_encode_initializer_p (encoder, node));
578 struct bitpack_d bp;
579 int ref;
580 const char *comdat;
581 const char *section;
582 tree group;
584 gcc_assert (!encode_initializer_p || node->definition);
585 gcc_assert (boundary_p || encode_initializer_p);
587 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
588 LTO_symtab_variable);
589 streamer_write_hwi_stream (ob->main_stream, node->order);
590 lto_output_var_decl_index (ob->decl_state, ob->main_stream, node->decl);
591 bp = bitpack_create (ob->main_stream);
592 bp_pack_value (&bp, node->externally_visible, 1);
593 bp_pack_value (&bp, node->no_reorder, 1);
594 bp_pack_value (&bp, node->force_output, 1);
595 bp_pack_value (&bp, node->forced_by_abi, 1);
596 bp_pack_value (&bp, node->unique_name, 1);
597 bp_pack_value (&bp,
598 node->body_removed
599 || (!encode_initializer_p && !node->alias && node->definition),
601 bp_pack_value (&bp, node->implicit_section, 1);
602 bp_pack_value (&bp, node->writeonly, 1);
603 bp_pack_value (&bp, node->definition && (encode_initializer_p || node->alias),
605 bp_pack_value (&bp, node->alias, 1);
606 bp_pack_value (&bp, node->transparent_alias, 1);
607 bp_pack_value (&bp, node->weakref, 1);
608 bp_pack_value (&bp, node->analyzed && (!boundary_p || node->alias), 1);
609 gcc_assert (node->definition || !node->analyzed);
610 /* Constant pool initializers can be de-unified into individual ltrans units.
611 FIXME: Alternatively at -Os we may want to avoid generating for them the local
612 labels and share them across LTRANS partitions. */
613 if (node->get_partitioning_class () != SYMBOL_PARTITION)
615 bp_pack_value (&bp, 0, 1); /* used_from_other_parition. */
616 bp_pack_value (&bp, 0, 1); /* in_other_partition. */
618 else
620 bp_pack_value (&bp, node->definition
621 && referenced_from_other_partition_p (node, encoder), 1);
622 bp_pack_value (&bp, node->analyzed
623 && boundary_p && !DECL_EXTERNAL (node->decl), 1);
624 /* in_other_partition. */
626 bp_pack_value (&bp, node->tls_model, 3);
627 bp_pack_value (&bp, node->used_by_single_function, 1);
628 bp_pack_value (&bp, node->need_bounds_init, 1);
629 streamer_write_bitpack (&bp);
631 group = node->get_comdat_group ();
632 if (group)
633 comdat = IDENTIFIER_POINTER (group);
634 else
635 comdat = "";
636 streamer_write_data_stream (ob->main_stream, comdat, strlen (comdat) + 1);
638 if (group)
640 if (node->same_comdat_group)
642 ref = LCC_NOT_FOUND;
643 for (struct symtab_node *n = node->same_comdat_group;
644 ref == LCC_NOT_FOUND && n != node; n = n->same_comdat_group)
645 ref = lto_symtab_encoder_lookup (encoder, n);
647 else
648 ref = LCC_NOT_FOUND;
649 streamer_write_hwi_stream (ob->main_stream, ref);
652 section = node->get_section ();
653 if (!section)
654 section = "";
655 streamer_write_data_stream (ob->main_stream, section, strlen (section) + 1);
657 streamer_write_enum (ob->main_stream, ld_plugin_symbol_resolution,
658 LDPR_NUM_KNOWN, node->resolution);
661 /* Output the varpool NODE to OB.
662 If NODE is not in SET, then NODE is a boundary. */
664 static void
665 lto_output_ref (struct lto_simple_output_block *ob, struct ipa_ref *ref,
666 lto_symtab_encoder_t encoder)
668 struct bitpack_d bp;
669 int nref;
670 int uid = ref->lto_stmt_uid;
671 struct cgraph_node *node;
673 bp = bitpack_create (ob->main_stream);
674 bp_pack_value (&bp, ref->use, 3);
675 bp_pack_value (&bp, ref->speculative, 1);
676 streamer_write_bitpack (&bp);
677 nref = lto_symtab_encoder_lookup (encoder, ref->referred);
678 gcc_assert (nref != LCC_NOT_FOUND);
679 streamer_write_hwi_stream (ob->main_stream, nref);
681 node = dyn_cast <cgraph_node *> (ref->referring);
682 if (node)
684 if (ref->stmt)
685 uid = gimple_uid (ref->stmt) + 1;
686 streamer_write_hwi_stream (ob->main_stream, uid);
690 /* Stream out profile_summary to OB. */
692 static void
693 output_profile_summary (struct lto_simple_output_block *ob)
695 unsigned h_ix;
696 struct bitpack_d bp;
698 if (profile_info)
700 /* We do not output num and run_max, they are not used by
701 GCC profile feedback and they are difficult to merge from multiple
702 units. */
703 gcc_assert (profile_info->runs);
704 streamer_write_uhwi_stream (ob->main_stream, profile_info->runs);
705 streamer_write_gcov_count_stream (ob->main_stream, profile_info->sum_max);
707 /* sum_all is needed for computing the working set with the
708 histogram. */
709 streamer_write_gcov_count_stream (ob->main_stream, profile_info->sum_all);
711 /* Create and output a bitpack of non-zero histogram entries indices. */
712 bp = bitpack_create (ob->main_stream);
713 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
714 bp_pack_value (&bp, profile_info->histogram[h_ix].num_counters > 0, 1);
715 streamer_write_bitpack (&bp);
716 /* Now stream out only those non-zero entries. */
717 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
719 if (!profile_info->histogram[h_ix].num_counters)
720 continue;
721 streamer_write_gcov_count_stream (ob->main_stream,
722 profile_info->histogram[h_ix].num_counters);
723 streamer_write_gcov_count_stream (ob->main_stream,
724 profile_info->histogram[h_ix].min_value);
725 streamer_write_gcov_count_stream (ob->main_stream,
726 profile_info->histogram[h_ix].cum_value);
728 /* IPA-profile computes hot bb threshold based on cumulated
729 whole program profile. We need to stream it down to ltrans. */
730 if (flag_wpa)
731 streamer_write_gcov_count_stream (ob->main_stream,
732 get_hot_bb_threshold ());
734 else
735 streamer_write_uhwi_stream (ob->main_stream, 0);
738 /* Output all callees or indirect outgoing edges. EDGE must be the first such
739 edge. */
741 static void
742 output_outgoing_cgraph_edges (struct cgraph_edge *edge,
743 struct lto_simple_output_block *ob,
744 lto_symtab_encoder_t encoder)
746 if (!edge)
747 return;
749 /* Output edges in backward direction, so the reconstructed callgraph match
750 and it is easy to associate call sites in the IPA pass summaries. */
751 while (edge->next_callee)
752 edge = edge->next_callee;
753 for (; edge; edge = edge->prev_callee)
754 lto_output_edge (ob, edge, encoder);
757 /* Output the part of the cgraph in SET. */
759 static void
760 output_refs (lto_symtab_encoder_t encoder)
762 struct lto_simple_output_block *ob;
763 int count;
764 struct ipa_ref *ref;
766 ob = lto_create_simple_output_block (LTO_section_refs);
768 for (int i = 0; i < lto_symtab_encoder_size (encoder); i++)
770 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
772 /* IPA_REF_ALIAS and IPA_REF_CHKP references are always preserved
773 in the boundary. Alias node can't have other references and
774 can be always handled as if it's not in the boundary. */
775 if (!node->alias && !lto_symtab_encoder_in_partition_p (encoder, node))
777 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
778 /* Output IPA_REF_CHKP reference. */
779 if (cnode
780 && cnode->instrumented_version
781 && !cnode->instrumentation_clone)
783 for (int i = 0; node->iterate_reference (i, ref); i++)
784 if (ref->use == IPA_REF_CHKP)
786 if (lto_symtab_encoder_lookup (encoder, ref->referred)
787 != LCC_NOT_FOUND)
789 int nref = lto_symtab_encoder_lookup (encoder, node);
790 streamer_write_gcov_count_stream (ob->main_stream, 1);
791 streamer_write_uhwi_stream (ob->main_stream, nref);
792 lto_output_ref (ob, ref, encoder);
794 break;
797 continue;
800 count = node->ref_list.nreferences ();
801 if (count)
803 streamer_write_gcov_count_stream (ob->main_stream, count);
804 streamer_write_uhwi_stream (ob->main_stream,
805 lto_symtab_encoder_lookup (encoder, node));
806 for (int i = 0; node->iterate_reference (i, ref); i++)
807 lto_output_ref (ob, ref, encoder);
811 streamer_write_uhwi_stream (ob->main_stream, 0);
813 lto_destroy_simple_output_block (ob);
816 /* Add NODE into encoder as well as nodes it is cloned from.
817 Do it in a way so clones appear first. */
819 static void
820 add_node_to (lto_symtab_encoder_t encoder, struct cgraph_node *node,
821 bool include_body)
823 if (node->clone_of)
824 add_node_to (encoder, node->clone_of, include_body);
825 else if (include_body)
826 lto_set_symtab_encoder_encode_body (encoder, node);
827 lto_symtab_encoder_encode (encoder, node);
830 /* Add all references in NODE to encoders. */
832 static void
833 create_references (lto_symtab_encoder_t encoder, symtab_node *node)
835 int i;
836 struct ipa_ref *ref = NULL;
837 for (i = 0; node->iterate_reference (i, ref); i++)
838 if (is_a <cgraph_node *> (ref->referred))
839 add_node_to (encoder, dyn_cast <cgraph_node *> (ref->referred), false);
840 else
841 lto_symtab_encoder_encode (encoder, ref->referred);
844 /* Select what needs to be streamed out. In regular lto mode stream everything.
845 In offload lto mode stream only nodes marked as offloadable. */
846 void
847 select_what_to_stream (void)
849 struct symtab_node *snode;
850 FOR_EACH_SYMBOL (snode)
851 snode->need_lto_streaming = !lto_stream_offload_p || snode->offloadable;
854 /* Find all symbols we want to stream into given partition and insert them
855 to encoders.
857 The function actually replaces IN_ENCODER by new one. The reason is that
858 streaming code needs clone's origin to be streamed before clone. This
859 means that we need to insert the nodes in specific order. This order is
860 ignored by the partitioning logic earlier. */
862 lto_symtab_encoder_t
863 compute_ltrans_boundary (lto_symtab_encoder_t in_encoder)
865 struct cgraph_edge *edge;
866 int i;
867 lto_symtab_encoder_t encoder;
868 lto_symtab_encoder_iterator lsei;
869 hash_set<void *> reachable_call_targets;
871 encoder = lto_symtab_encoder_new (false);
873 /* Go over all entries in the IN_ENCODER and duplicate them to
874 ENCODER. At the same time insert masters of clones so
875 every master appears before clone. */
876 for (lsei = lsei_start_function_in_partition (in_encoder);
877 !lsei_end_p (lsei); lsei_next_function_in_partition (&lsei))
879 struct cgraph_node *node = lsei_cgraph_node (lsei);
880 if (!node->need_lto_streaming)
881 continue;
882 add_node_to (encoder, node, true);
883 lto_set_symtab_encoder_in_partition (encoder, node);
884 create_references (encoder, node);
886 for (lsei = lsei_start_variable_in_partition (in_encoder);
887 !lsei_end_p (lsei); lsei_next_variable_in_partition (&lsei))
889 varpool_node *vnode = lsei_varpool_node (lsei);
891 if (!vnode->need_lto_streaming)
892 continue;
893 lto_set_symtab_encoder_in_partition (encoder, vnode);
894 lto_set_symtab_encoder_encode_initializer (encoder, vnode);
895 create_references (encoder, vnode);
897 /* Pickle in also the initializer of all referenced readonly variables
898 to help folding. Constant pool variables are not shared, so we must
899 pickle those too. */
900 for (i = 0; i < lto_symtab_encoder_size (encoder); i++)
902 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
903 if (varpool_node *vnode = dyn_cast <varpool_node *> (node))
905 if (!lto_symtab_encoder_encode_initializer_p (encoder,
906 vnode)
907 && (((vnode->ctor_useable_for_folding_p ()
908 && (!DECL_VIRTUAL_P (vnode->decl)
909 || !flag_wpa
910 || flag_ltrans_devirtualize))
911 || POINTER_BOUNDS_P (vnode->decl))))
913 lto_set_symtab_encoder_encode_initializer (encoder, vnode);
914 create_references (encoder, vnode);
919 /* Go over all the nodes again to include callees that are not in
920 SET. */
921 for (lsei = lsei_start_function_in_partition (encoder);
922 !lsei_end_p (lsei); lsei_next_function_in_partition (&lsei))
924 struct cgraph_node *node = lsei_cgraph_node (lsei);
925 for (edge = node->callees; edge; edge = edge->next_callee)
927 struct cgraph_node *callee = edge->callee;
928 if (!lto_symtab_encoder_in_partition_p (encoder, callee))
930 /* We should have moved all the inlines. */
931 gcc_assert (!callee->global.inlined_to);
932 add_node_to (encoder, callee, false);
935 /* Add all possible targets for late devirtualization. */
936 if (flag_ltrans_devirtualize || !flag_wpa)
937 for (edge = node->indirect_calls; edge; edge = edge->next_callee)
938 if (edge->indirect_info->polymorphic)
940 unsigned int i;
941 void *cache_token;
942 bool final;
943 vec <cgraph_node *>targets
944 = possible_polymorphic_call_targets
945 (edge, &final, &cache_token);
946 if (!reachable_call_targets.add (cache_token))
948 for (i = 0; i < targets.length (); i++)
950 struct cgraph_node *callee = targets[i];
952 /* Adding an external declarations into the unit serves
953 no purpose and just increases its boundary. */
954 if (callee->definition
955 && !lto_symtab_encoder_in_partition_p
956 (encoder, callee))
958 gcc_assert (!callee->global.inlined_to);
959 add_node_to (encoder, callee, false);
965 /* Be sure to also insert alias targert and thunk callees. These needs
966 to stay to aid local calling conventions. */
967 for (i = 0; i < lto_symtab_encoder_size (encoder); i++)
969 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
970 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
972 if (node->alias && node->analyzed)
973 create_references (encoder, node);
974 if (cnode
975 && cnode->thunk.thunk_p)
976 add_node_to (encoder, cnode->callees->callee, false);
977 while (node->transparent_alias && node->analyzed)
979 node = node->get_alias_target ();
980 if (is_a <cgraph_node *> (node))
981 add_node_to (encoder, dyn_cast <cgraph_node *> (node),
982 false);
983 else
984 lto_symtab_encoder_encode (encoder, node);
987 lto_symtab_encoder_delete (in_encoder);
988 return encoder;
991 /* Output the part of the symtab in SET and VSET. */
993 void
994 output_symtab (void)
996 struct cgraph_node *node;
997 struct lto_simple_output_block *ob;
998 int i, n_nodes;
999 lto_symtab_encoder_t encoder;
1001 if (flag_wpa)
1002 output_cgraph_opt_summary ();
1004 ob = lto_create_simple_output_block (LTO_section_symtab_nodes);
1006 output_profile_summary (ob);
1008 /* An encoder for cgraph nodes should have been created by
1009 ipa_write_summaries_1. */
1010 gcc_assert (ob->decl_state->symtab_node_encoder);
1011 encoder = ob->decl_state->symtab_node_encoder;
1013 /* Write out the nodes. We must first output a node and then its clones,
1014 otherwise at a time reading back the node there would be nothing to clone
1015 from. */
1016 n_nodes = lto_symtab_encoder_size (encoder);
1017 for (i = 0; i < n_nodes; i++)
1019 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
1020 if (cgraph_node *cnode = dyn_cast <cgraph_node *> (node))
1021 lto_output_node (ob, cnode, encoder);
1022 else
1023 lto_output_varpool_node (ob, dyn_cast<varpool_node *> (node), encoder);
1026 /* Go over the nodes in SET again to write edges. */
1027 for (int i = 0; i < lto_symtab_encoder_size (encoder); i++)
1029 node = dyn_cast <cgraph_node *> (lto_symtab_encoder_deref (encoder, i));
1030 if (node
1031 && (node->thunk.thunk_p
1032 || lto_symtab_encoder_in_partition_p (encoder, node)))
1034 output_outgoing_cgraph_edges (node->callees, ob, encoder);
1035 output_outgoing_cgraph_edges (node->indirect_calls, ob, encoder);
1039 streamer_write_uhwi_stream (ob->main_stream, 0);
1041 lto_destroy_simple_output_block (ob);
1043 /* Emit toplevel asms.
1044 When doing WPA we must output every asm just once. Since we do not partition asm
1045 nodes at all, output them to first output. This is kind of hack, but should work
1046 well. */
1047 if (!asm_nodes_output)
1049 asm_nodes_output = true;
1050 lto_output_toplevel_asms ();
1053 output_refs (encoder);
1056 /* Return identifier encoded in IB as a plain string. */
1058 static tree
1059 read_identifier (struct lto_input_block *ib)
1061 unsigned int len = strnlen (ib->data + ib->p, ib->len - ib->p - 1);
1062 tree id;
1064 if (ib->data[ib->p + len])
1065 lto_section_overrun (ib);
1066 if (!len)
1068 ib->p++;
1069 return NULL;
1071 id = get_identifier (ib->data + ib->p);
1072 ib->p += len + 1;
1073 return id;
1076 /* Return string encoded in IB, NULL if string is empty. */
1078 static const char *
1079 read_string (struct lto_input_block *ib)
1081 unsigned int len = strnlen (ib->data + ib->p, ib->len - ib->p - 1);
1082 const char *str;
1084 if (ib->data[ib->p + len])
1085 lto_section_overrun (ib);
1086 if (!len)
1088 ib->p++;
1089 return NULL;
1091 str = ib->data + ib->p;
1092 ib->p += len + 1;
1093 return str;
1096 /* Output function/variable tables that will allow libgomp to look up offload
1097 target code.
1098 OFFLOAD_FUNCS is filled in expand_omp_target, OFFLOAD_VARS is filled in
1099 varpool_node::get_create. In WHOPR (partitioned) mode during the WPA stage
1100 both OFFLOAD_FUNCS and OFFLOAD_VARS are filled by input_offload_tables. */
1102 void
1103 output_offload_tables (void)
1105 if (vec_safe_is_empty (offload_funcs) && vec_safe_is_empty (offload_vars))
1106 return;
1108 struct lto_simple_output_block *ob
1109 = lto_create_simple_output_block (LTO_section_offload_table);
1111 for (unsigned i = 0; i < vec_safe_length (offload_funcs); i++)
1113 streamer_write_enum (ob->main_stream, LTO_symtab_tags,
1114 LTO_symtab_last_tag, LTO_symtab_unavail_node);
1115 lto_output_fn_decl_index (ob->decl_state, ob->main_stream,
1116 (*offload_funcs)[i]);
1119 for (unsigned i = 0; i < vec_safe_length (offload_vars); i++)
1121 streamer_write_enum (ob->main_stream, LTO_symtab_tags,
1122 LTO_symtab_last_tag, LTO_symtab_variable);
1123 lto_output_var_decl_index (ob->decl_state, ob->main_stream,
1124 (*offload_vars)[i]);
1127 streamer_write_uhwi_stream (ob->main_stream, 0);
1128 lto_destroy_simple_output_block (ob);
1130 /* In WHOPR mode during the WPA stage the joint offload tables need to be
1131 streamed to one partition only. That's why we free offload_funcs and
1132 offload_vars after the first call of output_offload_tables. */
1133 if (flag_wpa)
1135 vec_free (offload_funcs);
1136 vec_free (offload_vars);
1140 /* Overwrite the information in NODE based on FILE_DATA, TAG, FLAGS,
1141 STACK_SIZE, SELF_TIME and SELF_SIZE. This is called either to initialize
1142 NODE or to replace the values in it, for instance because the first
1143 time we saw it, the function body was not available but now it
1144 is. BP is a bitpack with all the bitflags for NODE read from the
1145 stream. */
1147 static void
1148 input_overwrite_node (struct lto_file_decl_data *file_data,
1149 struct cgraph_node *node,
1150 enum LTO_symtab_tags tag,
1151 struct bitpack_d *bp)
1153 node->aux = (void *) tag;
1154 node->lto_file_data = file_data;
1156 node->local.local = bp_unpack_value (bp, 1);
1157 node->externally_visible = bp_unpack_value (bp, 1);
1158 node->no_reorder = bp_unpack_value (bp, 1);
1159 node->definition = bp_unpack_value (bp, 1);
1160 node->local.versionable = bp_unpack_value (bp, 1);
1161 node->local.can_change_signature = bp_unpack_value (bp, 1);
1162 node->local.redefined_extern_inline = bp_unpack_value (bp, 1);
1163 node->force_output = bp_unpack_value (bp, 1);
1164 node->forced_by_abi = bp_unpack_value (bp, 1);
1165 node->unique_name = bp_unpack_value (bp, 1);
1166 node->body_removed = bp_unpack_value (bp, 1);
1167 node->implicit_section = bp_unpack_value (bp, 1);
1168 node->address_taken = bp_unpack_value (bp, 1);
1169 node->used_from_other_partition = bp_unpack_value (bp, 1);
1170 node->lowered = bp_unpack_value (bp, 1);
1171 node->analyzed = tag == LTO_symtab_analyzed_node;
1172 node->in_other_partition = bp_unpack_value (bp, 1);
1173 if (node->in_other_partition
1174 /* Avoid updating decl when we are seeing just inline clone.
1175 When inlining function that has functions already inlined into it,
1176 we produce clones of inline clones.
1178 WPA partitioning might put each clone into different unit and
1179 we might end up streaming inline clone from other partition
1180 to support clone we are interested in. */
1181 && (!node->clone_of
1182 || node->clone_of->decl != node->decl))
1184 DECL_EXTERNAL (node->decl) = 1;
1185 TREE_STATIC (node->decl) = 0;
1187 node->alias = bp_unpack_value (bp, 1);
1188 node->transparent_alias = bp_unpack_value (bp, 1);
1189 node->weakref = bp_unpack_value (bp, 1);
1190 node->frequency = (enum node_frequency)bp_unpack_value (bp, 2);
1191 node->only_called_at_startup = bp_unpack_value (bp, 1);
1192 node->only_called_at_exit = bp_unpack_value (bp, 1);
1193 node->tm_clone = bp_unpack_value (bp, 1);
1194 node->calls_comdat_local = bp_unpack_value (bp, 1);
1195 node->icf_merged = bp_unpack_value (bp, 1);
1196 node->nonfreeing_fn = bp_unpack_value (bp, 1);
1197 node->thunk.thunk_p = bp_unpack_value (bp, 1);
1198 node->parallelized_function = bp_unpack_value (bp, 1);
1199 node->resolution = bp_unpack_enum (bp, ld_plugin_symbol_resolution,
1200 LDPR_NUM_KNOWN);
1201 node->instrumentation_clone = bp_unpack_value (bp, 1);
1202 node->split_part = bp_unpack_value (bp, 1);
1203 gcc_assert (flag_ltrans
1204 || (!node->in_other_partition
1205 && !node->used_from_other_partition));
1208 /* Return string alias is alias of. */
1210 static tree
1211 get_alias_symbol (tree decl)
1213 tree alias = lookup_attribute ("alias", DECL_ATTRIBUTES (decl));
1214 return get_identifier (TREE_STRING_POINTER
1215 (TREE_VALUE (TREE_VALUE (alias))));
1218 /* Read a node from input_block IB. TAG is the node's tag just read.
1219 Return the node read or overwriten. */
1221 static struct cgraph_node *
1222 input_node (struct lto_file_decl_data *file_data,
1223 struct lto_input_block *ib,
1224 enum LTO_symtab_tags tag,
1225 vec<symtab_node *> nodes)
1227 gcc::pass_manager *passes = g->get_passes ();
1228 tree fn_decl;
1229 struct cgraph_node *node;
1230 struct bitpack_d bp;
1231 unsigned decl_index;
1232 int ref = LCC_NOT_FOUND, ref2 = LCC_NOT_FOUND;
1233 int clone_ref;
1234 int order;
1235 int i, count;
1236 tree group;
1237 const char *section;
1238 order = streamer_read_hwi (ib) + order_base;
1239 clone_ref = streamer_read_hwi (ib);
1241 decl_index = streamer_read_uhwi (ib);
1242 fn_decl = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1244 if (clone_ref != LCC_NOT_FOUND)
1246 node = dyn_cast<cgraph_node *> (nodes[clone_ref])->create_clone (fn_decl,
1247 0, CGRAPH_FREQ_BASE, false,
1248 vNULL, false, NULL, NULL);
1250 else
1252 /* Declaration of functions can be already merged with a declaration
1253 from other input file. We keep cgraph unmerged until after streaming
1254 of ipa passes is done. Alays forcingly create a fresh node. */
1255 node = symtab->create_empty ();
1256 node->decl = fn_decl;
1257 node->register_symbol ();
1260 node->order = order;
1261 if (order >= symtab->order)
1262 symtab->order = order + 1;
1264 node->count = streamer_read_gcov_count (ib);
1265 node->count_materialization_scale = streamer_read_hwi (ib);
1267 count = streamer_read_hwi (ib);
1268 node->ipa_transforms_to_apply = vNULL;
1269 for (i = 0; i < count; i++)
1271 opt_pass *pass;
1272 int pid = streamer_read_hwi (ib);
1274 gcc_assert (pid < passes->passes_by_id_size);
1275 pass = passes->passes_by_id[pid];
1276 node->ipa_transforms_to_apply.safe_push ((ipa_opt_pass_d *) pass);
1279 if (tag == LTO_symtab_analyzed_node)
1280 ref = streamer_read_hwi (ib);
1282 group = read_identifier (ib);
1283 if (group)
1284 ref2 = streamer_read_hwi (ib);
1286 /* Make sure that we have not read this node before. Nodes that
1287 have already been read will have their tag stored in the 'aux'
1288 field. Since built-in functions can be referenced in multiple
1289 functions, they are expected to be read more than once. */
1290 if (node->aux && !DECL_BUILT_IN (node->decl))
1291 internal_error ("bytecode stream: found multiple instances of cgraph "
1292 "node with uid %d", node->uid);
1294 node->tp_first_run = streamer_read_uhwi (ib);
1296 bp = streamer_read_bitpack (ib);
1298 input_overwrite_node (file_data, node, tag, &bp);
1300 /* Store a reference for now, and fix up later to be a pointer. */
1301 node->global.inlined_to = (cgraph_node *) (intptr_t) ref;
1303 if (group)
1305 node->set_comdat_group (group);
1306 /* Store a reference for now, and fix up later to be a pointer. */
1307 node->same_comdat_group = (symtab_node *) (intptr_t) ref2;
1309 else
1310 node->same_comdat_group = (symtab_node *) (intptr_t) LCC_NOT_FOUND;
1311 section = read_string (ib);
1312 if (section)
1313 node->set_section_for_node (section);
1315 if (node->thunk.thunk_p)
1317 int type = streamer_read_uhwi (ib);
1318 HOST_WIDE_INT fixed_offset = streamer_read_uhwi (ib);
1319 HOST_WIDE_INT virtual_value = streamer_read_uhwi (ib);
1321 node->thunk.fixed_offset = fixed_offset;
1322 node->thunk.this_adjusting = (type & 2);
1323 node->thunk.virtual_value = virtual_value;
1324 node->thunk.virtual_offset_p = (type & 4);
1325 node->thunk.add_pointer_bounds_args = (type & 8);
1327 if (node->alias && !node->analyzed && node->weakref)
1328 node->alias_target = get_alias_symbol (node->decl);
1329 node->profile_id = streamer_read_hwi (ib);
1330 if (DECL_STATIC_CONSTRUCTOR (node->decl))
1331 node->set_init_priority (streamer_read_hwi (ib));
1332 if (DECL_STATIC_DESTRUCTOR (node->decl))
1333 node->set_fini_priority (streamer_read_hwi (ib));
1335 if (node->instrumentation_clone)
1337 decl_index = streamer_read_uhwi (ib);
1338 fn_decl = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1339 node->orig_decl = fn_decl;
1342 return node;
1345 /* Read a node from input_block IB. TAG is the node's tag just read.
1346 Return the node read or overwriten. */
1348 static varpool_node *
1349 input_varpool_node (struct lto_file_decl_data *file_data,
1350 struct lto_input_block *ib)
1352 int decl_index;
1353 tree var_decl;
1354 varpool_node *node;
1355 struct bitpack_d bp;
1356 int ref = LCC_NOT_FOUND;
1357 int order;
1358 tree group;
1359 const char *section;
1361 order = streamer_read_hwi (ib) + order_base;
1362 decl_index = streamer_read_uhwi (ib);
1363 var_decl = lto_file_decl_data_get_var_decl (file_data, decl_index);
1365 /* Declaration of functions can be already merged with a declaration
1366 from other input file. We keep cgraph unmerged until after streaming
1367 of ipa passes is done. Alays forcingly create a fresh node. */
1368 node = varpool_node::create_empty ();
1369 node->decl = var_decl;
1370 node->register_symbol ();
1372 node->order = order;
1373 if (order >= symtab->order)
1374 symtab->order = order + 1;
1375 node->lto_file_data = file_data;
1377 bp = streamer_read_bitpack (ib);
1378 node->externally_visible = bp_unpack_value (&bp, 1);
1379 node->no_reorder = bp_unpack_value (&bp, 1);
1380 node->force_output = bp_unpack_value (&bp, 1);
1381 node->forced_by_abi = bp_unpack_value (&bp, 1);
1382 node->unique_name = bp_unpack_value (&bp, 1);
1383 node->body_removed = bp_unpack_value (&bp, 1);
1384 node->implicit_section = bp_unpack_value (&bp, 1);
1385 node->writeonly = bp_unpack_value (&bp, 1);
1386 node->definition = bp_unpack_value (&bp, 1);
1387 node->alias = bp_unpack_value (&bp, 1);
1388 node->transparent_alias = bp_unpack_value (&bp, 1);
1389 node->weakref = bp_unpack_value (&bp, 1);
1390 node->analyzed = bp_unpack_value (&bp, 1);
1391 node->used_from_other_partition = bp_unpack_value (&bp, 1);
1392 node->in_other_partition = bp_unpack_value (&bp, 1);
1393 if (node->in_other_partition)
1395 DECL_EXTERNAL (node->decl) = 1;
1396 TREE_STATIC (node->decl) = 0;
1398 if (node->alias && !node->analyzed && node->weakref)
1399 node->alias_target = get_alias_symbol (node->decl);
1400 node->tls_model = (enum tls_model)bp_unpack_value (&bp, 3);
1401 node->used_by_single_function = (enum tls_model)bp_unpack_value (&bp, 1);
1402 node->need_bounds_init = bp_unpack_value (&bp, 1);
1403 group = read_identifier (ib);
1404 if (group)
1406 node->set_comdat_group (group);
1407 ref = streamer_read_hwi (ib);
1408 /* Store a reference for now, and fix up later to be a pointer. */
1409 node->same_comdat_group = (symtab_node *) (intptr_t) ref;
1411 else
1412 node->same_comdat_group = (symtab_node *) (intptr_t) LCC_NOT_FOUND;
1413 section = read_string (ib);
1414 if (section)
1415 node->set_section_for_node (section);
1416 node->resolution = streamer_read_enum (ib, ld_plugin_symbol_resolution,
1417 LDPR_NUM_KNOWN);
1418 gcc_assert (flag_ltrans
1419 || (!node->in_other_partition
1420 && !node->used_from_other_partition));
1422 return node;
1425 /* Read a node from input_block IB. TAG is the node's tag just read.
1426 Return the node read or overwriten. */
1428 static void
1429 input_ref (struct lto_input_block *ib,
1430 symtab_node *referring_node,
1431 vec<symtab_node *> nodes)
1433 symtab_node *node = NULL;
1434 struct bitpack_d bp;
1435 enum ipa_ref_use use;
1436 bool speculative;
1437 struct ipa_ref *ref;
1439 bp = streamer_read_bitpack (ib);
1440 use = (enum ipa_ref_use) bp_unpack_value (&bp, 3);
1441 speculative = (enum ipa_ref_use) bp_unpack_value (&bp, 1);
1442 node = nodes[streamer_read_hwi (ib)];
1443 ref = referring_node->create_reference (node, use);
1444 ref->speculative = speculative;
1445 if (is_a <cgraph_node *> (referring_node))
1446 ref->lto_stmt_uid = streamer_read_hwi (ib);
1449 /* Read an edge from IB. NODES points to a vector of previously read nodes for
1450 decoding caller and callee of the edge to be read. If INDIRECT is true, the
1451 edge being read is indirect (in the sense that it has
1452 indirect_unknown_callee set). */
1454 static void
1455 input_edge (struct lto_input_block *ib, vec<symtab_node *> nodes,
1456 bool indirect)
1458 struct cgraph_node *caller, *callee;
1459 struct cgraph_edge *edge;
1460 unsigned int stmt_id;
1461 gcov_type count;
1462 int freq;
1463 cgraph_inline_failed_t inline_failed;
1464 struct bitpack_d bp;
1465 int ecf_flags = 0;
1467 caller = dyn_cast<cgraph_node *> (nodes[streamer_read_hwi (ib)]);
1468 if (caller == NULL || caller->decl == NULL_TREE)
1469 internal_error ("bytecode stream: no caller found while reading edge");
1471 if (!indirect)
1473 callee = dyn_cast<cgraph_node *> (nodes[streamer_read_hwi (ib)]);
1474 if (callee == NULL || callee->decl == NULL_TREE)
1475 internal_error ("bytecode stream: no callee found while reading edge");
1477 else
1478 callee = NULL;
1480 count = streamer_read_gcov_count (ib);
1482 bp = streamer_read_bitpack (ib);
1483 inline_failed = bp_unpack_enum (&bp, cgraph_inline_failed_t, CIF_N_REASONS);
1484 stmt_id = bp_unpack_var_len_unsigned (&bp);
1485 freq = (int) bp_unpack_var_len_unsigned (&bp);
1487 if (indirect)
1488 edge = caller->create_indirect_edge (NULL, 0, count, freq);
1489 else
1490 edge = caller->create_edge (callee, NULL, count, freq);
1492 edge->indirect_inlining_edge = bp_unpack_value (&bp, 1);
1493 edge->speculative = bp_unpack_value (&bp, 1);
1494 edge->lto_stmt_uid = stmt_id;
1495 edge->inline_failed = inline_failed;
1496 edge->call_stmt_cannot_inline_p = bp_unpack_value (&bp, 1);
1497 edge->can_throw_external = bp_unpack_value (&bp, 1);
1498 edge->in_polymorphic_cdtor = bp_unpack_value (&bp, 1);
1499 if (indirect)
1501 if (bp_unpack_value (&bp, 1))
1502 ecf_flags |= ECF_CONST;
1503 if (bp_unpack_value (&bp, 1))
1504 ecf_flags |= ECF_PURE;
1505 if (bp_unpack_value (&bp, 1))
1506 ecf_flags |= ECF_NORETURN;
1507 if (bp_unpack_value (&bp, 1))
1508 ecf_flags |= ECF_MALLOC;
1509 if (bp_unpack_value (&bp, 1))
1510 ecf_flags |= ECF_NOTHROW;
1511 if (bp_unpack_value (&bp, 1))
1512 ecf_flags |= ECF_RETURNS_TWICE;
1513 edge->indirect_info->ecf_flags = ecf_flags;
1514 edge->indirect_info->common_target_id = streamer_read_hwi (ib);
1515 if (edge->indirect_info->common_target_id)
1516 edge->indirect_info->common_target_probability = streamer_read_hwi (ib);
1521 /* Read a cgraph from IB using the info in FILE_DATA. */
1523 static vec<symtab_node *>
1524 input_cgraph_1 (struct lto_file_decl_data *file_data,
1525 struct lto_input_block *ib)
1527 enum LTO_symtab_tags tag;
1528 vec<symtab_node *> nodes = vNULL;
1529 symtab_node *node;
1530 unsigned i;
1532 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1533 order_base = symtab->order;
1534 while (tag)
1536 if (tag == LTO_symtab_edge)
1537 input_edge (ib, nodes, false);
1538 else if (tag == LTO_symtab_indirect_edge)
1539 input_edge (ib, nodes, true);
1540 else if (tag == LTO_symtab_variable)
1542 node = input_varpool_node (file_data, ib);
1543 nodes.safe_push (node);
1544 lto_symtab_encoder_encode (file_data->symtab_node_encoder, node);
1546 else
1548 node = input_node (file_data, ib, tag, nodes);
1549 if (node == NULL || node->decl == NULL_TREE)
1550 internal_error ("bytecode stream: found empty cgraph node");
1551 nodes.safe_push (node);
1552 lto_symtab_encoder_encode (file_data->symtab_node_encoder, node);
1555 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1558 lto_input_toplevel_asms (file_data, order_base);
1560 /* AUX pointers should be all non-zero for function nodes read from the stream. */
1561 if (flag_checking)
1563 FOR_EACH_VEC_ELT (nodes, i, node)
1564 gcc_assert (node->aux || !is_a <cgraph_node *> (node));
1566 FOR_EACH_VEC_ELT (nodes, i, node)
1568 int ref;
1569 if (cgraph_node *cnode = dyn_cast <cgraph_node *> (node))
1571 ref = (int) (intptr_t) cnode->global.inlined_to;
1573 /* We share declaration of builtins, so we may read same node twice. */
1574 if (!node->aux)
1575 continue;
1576 node->aux = NULL;
1578 /* Fixup inlined_to from reference to pointer. */
1579 if (ref != LCC_NOT_FOUND)
1580 dyn_cast<cgraph_node *> (node)->global.inlined_to
1581 = dyn_cast<cgraph_node *> (nodes[ref]);
1582 else
1583 cnode->global.inlined_to = NULL;
1585 /* Compute instrumented_version. */
1586 if (cnode->instrumentation_clone)
1588 gcc_assert (cnode->orig_decl);
1590 cnode->instrumented_version = cgraph_node::get (cnode->orig_decl);
1591 if (cnode->instrumented_version)
1593 /* We may have multiple nodes for a single function which
1594 will be merged later. To have a proper merge we need
1595 to keep instrumentation_version reference between nodes
1596 consistent: each instrumented_version reference should
1597 have proper reverse reference. Thus don't break existing
1598 instrumented_version reference if it already exists. */
1599 if (cnode->instrumented_version->instrumented_version)
1600 cnode->instrumented_version = NULL;
1601 else
1602 cnode->instrumented_version->instrumented_version = cnode;
1605 /* Restore decl names reference except for wrapper functions. */
1606 if (!chkp_wrap_function (cnode->orig_decl))
1608 tree name = DECL_ASSEMBLER_NAME (cnode->decl);
1609 IDENTIFIER_TRANSPARENT_ALIAS (name) = 1;
1610 TREE_CHAIN (name) = DECL_ASSEMBLER_NAME (cnode->orig_decl);
1615 ref = (int) (intptr_t) node->same_comdat_group;
1617 /* Fixup same_comdat_group from reference to pointer. */
1618 if (ref != LCC_NOT_FOUND)
1619 node->same_comdat_group = nodes[ref];
1620 else
1621 node->same_comdat_group = NULL;
1623 FOR_EACH_VEC_ELT (nodes, i, node)
1624 node->aux = is_a <cgraph_node *> (node) ? (void *)1 : NULL;
1625 return nodes;
1628 /* Input ipa_refs. */
1630 static void
1631 input_refs (struct lto_input_block *ib,
1632 vec<symtab_node *> nodes)
1634 int count;
1635 int idx;
1636 while (true)
1638 symtab_node *node;
1639 count = streamer_read_uhwi (ib);
1640 if (!count)
1641 break;
1642 idx = streamer_read_uhwi (ib);
1643 node = nodes[idx];
1644 while (count)
1646 input_ref (ib, node, nodes);
1647 count--;
1653 static struct gcov_ctr_summary lto_gcov_summary;
1655 /* Input profile_info from IB. */
1656 static void
1657 input_profile_summary (struct lto_input_block *ib,
1658 struct lto_file_decl_data *file_data)
1660 unsigned h_ix;
1661 struct bitpack_d bp;
1662 unsigned int runs = streamer_read_uhwi (ib);
1663 if (runs)
1665 file_data->profile_info.runs = runs;
1666 file_data->profile_info.sum_max = streamer_read_gcov_count (ib);
1667 file_data->profile_info.sum_all = streamer_read_gcov_count (ib);
1669 memset (file_data->profile_info.histogram, 0,
1670 sizeof (gcov_bucket_type) * GCOV_HISTOGRAM_SIZE);
1671 /* Input the bitpack of non-zero histogram indices. */
1672 bp = streamer_read_bitpack (ib);
1673 /* Read in and unpack the full bitpack, flagging non-zero
1674 histogram entries by setting the num_counters non-zero. */
1675 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1677 file_data->profile_info.histogram[h_ix].num_counters
1678 = bp_unpack_value (&bp, 1);
1680 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1682 if (!file_data->profile_info.histogram[h_ix].num_counters)
1683 continue;
1685 file_data->profile_info.histogram[h_ix].num_counters
1686 = streamer_read_gcov_count (ib);
1687 file_data->profile_info.histogram[h_ix].min_value
1688 = streamer_read_gcov_count (ib);
1689 file_data->profile_info.histogram[h_ix].cum_value
1690 = streamer_read_gcov_count (ib);
1692 /* IPA-profile computes hot bb threshold based on cumulated
1693 whole program profile. We need to stream it down to ltrans. */
1694 if (flag_ltrans)
1695 set_hot_bb_threshold (streamer_read_gcov_count (ib));
1700 /* Rescale profile summaries to the same number of runs in the whole unit. */
1702 static void
1703 merge_profile_summaries (struct lto_file_decl_data **file_data_vec)
1705 struct lto_file_decl_data *file_data;
1706 unsigned int j, h_ix;
1707 gcov_unsigned_t max_runs = 0;
1708 struct cgraph_node *node;
1709 struct cgraph_edge *edge;
1710 gcov_type saved_sum_all = 0;
1711 gcov_ctr_summary *saved_profile_info = 0;
1712 int saved_scale = 0;
1714 /* Find unit with maximal number of runs. If we ever get serious about
1715 roundoff errors, we might also consider computing smallest common
1716 multiply. */
1717 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1718 if (max_runs < file_data->profile_info.runs)
1719 max_runs = file_data->profile_info.runs;
1721 if (!max_runs)
1722 return;
1724 /* Simple overflow check. We probably don't need to support that many train
1725 runs. Such a large value probably imply data corruption anyway. */
1726 if (max_runs > INT_MAX / REG_BR_PROB_BASE)
1728 sorry ("At most %i profile runs is supported. Perhaps corrupted profile?",
1729 INT_MAX / REG_BR_PROB_BASE);
1730 return;
1733 profile_info = &lto_gcov_summary;
1734 lto_gcov_summary.runs = max_runs;
1735 lto_gcov_summary.sum_max = 0;
1736 memset (lto_gcov_summary.histogram, 0,
1737 sizeof (gcov_bucket_type) * GCOV_HISTOGRAM_SIZE);
1739 /* Rescale all units to the maximal number of runs.
1740 sum_max can not be easily merged, as we have no idea what files come from
1741 the same run. We do not use the info anyway, so leave it 0. */
1742 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1743 if (file_data->profile_info.runs)
1745 int scale = GCOV_COMPUTE_SCALE (max_runs,
1746 file_data->profile_info.runs);
1747 lto_gcov_summary.sum_max
1748 = MAX (lto_gcov_summary.sum_max,
1749 apply_scale (file_data->profile_info.sum_max, scale));
1750 lto_gcov_summary.sum_all
1751 = MAX (lto_gcov_summary.sum_all,
1752 apply_scale (file_data->profile_info.sum_all, scale));
1753 /* Save a pointer to the profile_info with the largest
1754 scaled sum_all and the scale for use in merging the
1755 histogram. */
1756 if (!saved_profile_info
1757 || lto_gcov_summary.sum_all > saved_sum_all)
1759 saved_profile_info = &file_data->profile_info;
1760 saved_sum_all = lto_gcov_summary.sum_all;
1761 saved_scale = scale;
1765 gcc_assert (saved_profile_info);
1767 /* Scale up the histogram from the profile that had the largest
1768 scaled sum_all above. */
1769 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1771 /* Scale up the min value as we did the corresponding sum_all
1772 above. Use that to find the new histogram index. */
1773 gcov_type scaled_min
1774 = apply_scale (saved_profile_info->histogram[h_ix].min_value,
1775 saved_scale);
1776 /* The new index may be shared with another scaled histogram entry,
1777 so we need to account for a non-zero histogram entry at new_ix. */
1778 unsigned new_ix = gcov_histo_index (scaled_min);
1779 lto_gcov_summary.histogram[new_ix].min_value
1780 = (lto_gcov_summary.histogram[new_ix].num_counters
1781 ? MIN (lto_gcov_summary.histogram[new_ix].min_value, scaled_min)
1782 : scaled_min);
1783 /* Some of the scaled counter values would ostensibly need to be placed
1784 into different (larger) histogram buckets, but we keep things simple
1785 here and place the scaled cumulative counter value in the bucket
1786 corresponding to the scaled minimum counter value. */
1787 lto_gcov_summary.histogram[new_ix].cum_value
1788 += apply_scale (saved_profile_info->histogram[h_ix].cum_value,
1789 saved_scale);
1790 lto_gcov_summary.histogram[new_ix].num_counters
1791 += saved_profile_info->histogram[h_ix].num_counters;
1794 /* Watch roundoff errors. */
1795 if (lto_gcov_summary.sum_max < max_runs)
1796 lto_gcov_summary.sum_max = max_runs;
1798 /* If merging already happent at WPA time, we are done. */
1799 if (flag_ltrans)
1800 return;
1802 /* Now compute count_materialization_scale of each node.
1803 During LTRANS we already have values of count_materialization_scale
1804 computed, so just update them. */
1805 FOR_EACH_FUNCTION (node)
1806 if (node->lto_file_data
1807 && node->lto_file_data->profile_info.runs)
1809 int scale;
1811 scale = RDIV (node->count_materialization_scale * max_runs,
1812 node->lto_file_data->profile_info.runs);
1813 node->count_materialization_scale = scale;
1814 if (scale < 0)
1815 fatal_error (input_location, "Profile information in %s corrupted",
1816 file_data->file_name);
1818 if (scale == REG_BR_PROB_BASE)
1819 continue;
1820 for (edge = node->callees; edge; edge = edge->next_callee)
1821 edge->count = apply_scale (edge->count, scale);
1822 node->count = apply_scale (node->count, scale);
1826 /* Input and merge the symtab from each of the .o files passed to
1827 lto1. */
1829 void
1830 input_symtab (void)
1832 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1833 struct lto_file_decl_data *file_data;
1834 unsigned int j = 0;
1835 struct cgraph_node *node;
1837 while ((file_data = file_data_vec[j++]))
1839 const char *data;
1840 size_t len;
1841 struct lto_input_block *ib;
1842 vec<symtab_node *> nodes;
1844 ib = lto_create_simple_input_block (file_data, LTO_section_symtab_nodes,
1845 &data, &len);
1846 if (!ib)
1847 fatal_error (input_location,
1848 "cannot find LTO cgraph in %s", file_data->file_name);
1849 input_profile_summary (ib, file_data);
1850 file_data->symtab_node_encoder = lto_symtab_encoder_new (true);
1851 nodes = input_cgraph_1 (file_data, ib);
1852 lto_destroy_simple_input_block (file_data, LTO_section_symtab_nodes,
1853 ib, data, len);
1855 ib = lto_create_simple_input_block (file_data, LTO_section_refs,
1856 &data, &len);
1857 if (!ib)
1858 fatal_error (input_location, "cannot find LTO section refs in %s",
1859 file_data->file_name);
1860 input_refs (ib, nodes);
1861 lto_destroy_simple_input_block (file_data, LTO_section_refs,
1862 ib, data, len);
1863 if (flag_ltrans)
1864 input_cgraph_opt_summary (nodes);
1865 nodes.release ();
1868 merge_profile_summaries (file_data_vec);
1869 get_working_sets ();
1872 /* Clear out the aux field that was used to store enough state to
1873 tell which nodes should be overwritten. */
1874 FOR_EACH_FUNCTION (node)
1876 /* Some nodes may have been created by cgraph_node. This
1877 happens when the callgraph contains nested functions. If the
1878 node for the parent function was never emitted to the gimple
1879 file, cgraph_node will create a node for it when setting the
1880 context of the nested function. */
1881 if (node->lto_file_data)
1882 node->aux = NULL;
1886 /* Input function/variable tables that will allow libgomp to look up offload
1887 target code, and store them into OFFLOAD_FUNCS and OFFLOAD_VARS. */
1889 void
1890 input_offload_tables (bool do_force_output)
1892 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1893 struct lto_file_decl_data *file_data;
1894 unsigned int j = 0;
1896 while ((file_data = file_data_vec[j++]))
1898 const char *data;
1899 size_t len;
1900 struct lto_input_block *ib
1901 = lto_create_simple_input_block (file_data, LTO_section_offload_table,
1902 &data, &len);
1903 if (!ib)
1904 continue;
1906 enum LTO_symtab_tags tag
1907 = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1908 while (tag)
1910 if (tag == LTO_symtab_unavail_node)
1912 int decl_index = streamer_read_uhwi (ib);
1913 tree fn_decl
1914 = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1915 vec_safe_push (offload_funcs, fn_decl);
1917 /* Prevent IPA from removing fn_decl as unreachable, since there
1918 may be no refs from the parent function to child_fn in offload
1919 LTO mode. */
1920 if (do_force_output)
1921 cgraph_node::get (fn_decl)->mark_force_output ();
1923 else if (tag == LTO_symtab_variable)
1925 int decl_index = streamer_read_uhwi (ib);
1926 tree var_decl
1927 = lto_file_decl_data_get_var_decl (file_data, decl_index);
1928 vec_safe_push (offload_vars, var_decl);
1930 /* Prevent IPA from removing var_decl as unused, since there
1931 may be no refs to var_decl in offload LTO mode. */
1932 if (do_force_output)
1933 varpool_node::get (var_decl)->force_output = 1;
1935 else
1936 fatal_error (input_location,
1937 "invalid offload table in %s", file_data->file_name);
1939 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1942 lto_destroy_simple_input_block (file_data, LTO_section_offload_table,
1943 ib, data, len);
1947 /* True when we need optimization summary for NODE. */
1949 static int
1950 output_cgraph_opt_summary_p (struct cgraph_node *node)
1952 return (node->clone_of
1953 && (node->clone.tree_map
1954 || node->clone.args_to_skip
1955 || node->clone.combined_args_to_skip));
1958 /* Output optimization summary for EDGE to OB. */
1959 static void
1960 output_edge_opt_summary (struct output_block *ob ATTRIBUTE_UNUSED,
1961 struct cgraph_edge *edge ATTRIBUTE_UNUSED)
1965 /* Output optimization summary for NODE to OB. */
1967 static void
1968 output_node_opt_summary (struct output_block *ob,
1969 struct cgraph_node *node,
1970 lto_symtab_encoder_t encoder)
1972 unsigned int index;
1973 bitmap_iterator bi;
1974 struct ipa_replace_map *map;
1975 struct bitpack_d bp;
1976 int i;
1977 struct cgraph_edge *e;
1979 if (node->clone.args_to_skip)
1981 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.args_to_skip));
1982 EXECUTE_IF_SET_IN_BITMAP (node->clone.args_to_skip, 0, index, bi)
1983 streamer_write_uhwi (ob, index);
1985 else
1986 streamer_write_uhwi (ob, 0);
1987 if (node->clone.combined_args_to_skip)
1989 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.combined_args_to_skip));
1990 EXECUTE_IF_SET_IN_BITMAP (node->clone.combined_args_to_skip, 0, index, bi)
1991 streamer_write_uhwi (ob, index);
1993 else
1994 streamer_write_uhwi (ob, 0);
1995 streamer_write_uhwi (ob, vec_safe_length (node->clone.tree_map));
1996 FOR_EACH_VEC_SAFE_ELT (node->clone.tree_map, i, map)
1998 /* At the moment we assume all old trees to be PARM_DECLs, because we have no
1999 mechanism to store function local declarations into summaries. */
2000 gcc_assert (!map->old_tree);
2001 streamer_write_uhwi (ob, map->parm_num);
2002 gcc_assert (EXPR_LOCATION (map->new_tree) == UNKNOWN_LOCATION);
2003 stream_write_tree (ob, map->new_tree, true);
2004 bp = bitpack_create (ob->main_stream);
2005 bp_pack_value (&bp, map->replace_p, 1);
2006 bp_pack_value (&bp, map->ref_p, 1);
2007 streamer_write_bitpack (&bp);
2010 if (lto_symtab_encoder_in_partition_p (encoder, node))
2012 for (e = node->callees; e; e = e->next_callee)
2013 output_edge_opt_summary (ob, e);
2014 for (e = node->indirect_calls; e; e = e->next_callee)
2015 output_edge_opt_summary (ob, e);
2019 /* Output optimization summaries stored in callgraph.
2020 At the moment it is the clone info structure. */
2022 static void
2023 output_cgraph_opt_summary (void)
2025 int i, n_nodes;
2026 lto_symtab_encoder_t encoder;
2027 struct output_block *ob = create_output_block (LTO_section_cgraph_opt_sum);
2028 unsigned count = 0;
2030 ob->symbol = NULL;
2031 encoder = ob->decl_state->symtab_node_encoder;
2032 n_nodes = lto_symtab_encoder_size (encoder);
2033 for (i = 0; i < n_nodes; i++)
2035 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
2036 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
2037 if (cnode && output_cgraph_opt_summary_p (cnode))
2038 count++;
2040 streamer_write_uhwi (ob, count);
2041 for (i = 0; i < n_nodes; i++)
2043 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
2044 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
2045 if (cnode && output_cgraph_opt_summary_p (cnode))
2047 streamer_write_uhwi (ob, i);
2048 output_node_opt_summary (ob, cnode, encoder);
2051 produce_asm (ob, NULL);
2052 destroy_output_block (ob);
2055 /* Input optimisation summary of EDGE. */
2057 static void
2058 input_edge_opt_summary (struct cgraph_edge *edge ATTRIBUTE_UNUSED,
2059 struct lto_input_block *ib_main ATTRIBUTE_UNUSED)
2063 /* Input optimisation summary of NODE. */
2065 static void
2066 input_node_opt_summary (struct cgraph_node *node,
2067 struct lto_input_block *ib_main,
2068 struct data_in *data_in)
2070 int i;
2071 int count;
2072 int bit;
2073 struct bitpack_d bp;
2074 struct cgraph_edge *e;
2076 count = streamer_read_uhwi (ib_main);
2077 if (count)
2078 node->clone.args_to_skip = BITMAP_GGC_ALLOC ();
2079 for (i = 0; i < count; i++)
2081 bit = streamer_read_uhwi (ib_main);
2082 bitmap_set_bit (node->clone.args_to_skip, bit);
2084 count = streamer_read_uhwi (ib_main);
2085 if (count)
2086 node->clone.combined_args_to_skip = BITMAP_GGC_ALLOC ();
2087 for (i = 0; i < count; i++)
2089 bit = streamer_read_uhwi (ib_main);
2090 bitmap_set_bit (node->clone.combined_args_to_skip, bit);
2092 count = streamer_read_uhwi (ib_main);
2093 for (i = 0; i < count; i++)
2095 struct ipa_replace_map *map = ggc_alloc<ipa_replace_map> ();
2097 vec_safe_push (node->clone.tree_map, map);
2098 map->parm_num = streamer_read_uhwi (ib_main);
2099 map->old_tree = NULL;
2100 map->new_tree = stream_read_tree (ib_main, data_in);
2101 bp = streamer_read_bitpack (ib_main);
2102 map->replace_p = bp_unpack_value (&bp, 1);
2103 map->ref_p = bp_unpack_value (&bp, 1);
2105 for (e = node->callees; e; e = e->next_callee)
2106 input_edge_opt_summary (e, ib_main);
2107 for (e = node->indirect_calls; e; e = e->next_callee)
2108 input_edge_opt_summary (e, ib_main);
2111 /* Read section in file FILE_DATA of length LEN with data DATA. */
2113 static void
2114 input_cgraph_opt_section (struct lto_file_decl_data *file_data,
2115 const char *data, size_t len,
2116 vec<symtab_node *> nodes)
2118 const struct lto_function_header *header =
2119 (const struct lto_function_header *) data;
2120 const int cfg_offset = sizeof (struct lto_function_header);
2121 const int main_offset = cfg_offset + header->cfg_size;
2122 const int string_offset = main_offset + header->main_size;
2123 struct data_in *data_in;
2124 unsigned int i;
2125 unsigned int count;
2127 lto_input_block ib_main ((const char *) data + main_offset,
2128 header->main_size, file_data->mode_table);
2130 data_in =
2131 lto_data_in_create (file_data, (const char *) data + string_offset,
2132 header->string_size, vNULL);
2133 count = streamer_read_uhwi (&ib_main);
2135 for (i = 0; i < count; i++)
2137 int ref = streamer_read_uhwi (&ib_main);
2138 input_node_opt_summary (dyn_cast<cgraph_node *> (nodes[ref]),
2139 &ib_main, data_in);
2141 lto_free_section_data (file_data, LTO_section_cgraph_opt_sum, NULL, data,
2142 len);
2143 lto_data_in_delete (data_in);
2146 /* Input optimization summary of cgraph. */
2148 static void
2149 input_cgraph_opt_summary (vec<symtab_node *> nodes)
2151 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
2152 struct lto_file_decl_data *file_data;
2153 unsigned int j = 0;
2155 while ((file_data = file_data_vec[j++]))
2157 size_t len;
2158 const char *data =
2159 lto_get_section_data (file_data, LTO_section_cgraph_opt_sum, NULL,
2160 &len);
2162 if (data)
2163 input_cgraph_opt_section (file_data, data, len, nodes);