* testsuite/26_numerics/headers/cmath/hypot.cc: XFAIL on AIX.
[official-gcc.git] / gcc / lto-cgraph.c
blob26420411b0d416b4bcbc0b22819ff6137a290140
1 /* Write and read the cgraph to the memory mapped representation of a
2 .o file.
4 Copyright (C) 2009-2016 Free Software Foundation, Inc.
5 Contributed by Kenneth Zadeck <zadeck@naturalbridge.com>
7 This file is part of GCC.
9 GCC is free software; you can redistribute it and/or modify it under
10 the terms of the GNU General Public License as published by the Free
11 Software Foundation; either version 3, or (at your option) any later
12 version.
14 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
15 WARRANTY; without even the implied warranty of MERCHANTABILITY or
16 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 for more details.
19 You should have received a copy of the GNU General Public License
20 along with GCC; see the file COPYING3. If not see
21 <http://www.gnu.org/licenses/>. */
23 #include "config.h"
24 #include "system.h"
25 #include "coretypes.h"
26 #include "backend.h"
27 #include "rtl.h"
28 #include "tree.h"
29 #include "gimple.h"
30 #include "predict.h"
31 #include "stringpool.h"
32 #include "tree-streamer.h"
33 #include "cgraph.h"
34 #include "tree-pass.h"
35 #include "profile.h"
36 #include "context.h"
37 #include "pass_manager.h"
38 #include "ipa-utils.h"
39 #include "omp-low.h"
40 #include "ipa-chkp.h"
42 /* True when asm nodes has been output. */
43 bool asm_nodes_output = false;
45 static void output_cgraph_opt_summary (void);
46 static void input_cgraph_opt_summary (vec<symtab_node *> nodes);
48 /* Number of LDPR values known to GCC. */
49 #define LDPR_NUM_KNOWN (LDPR_PREVAILING_DEF_IRONLY_EXP + 1)
51 /* All node orders are ofsetted by ORDER_BASE. */
52 static int order_base;
54 /* Cgraph streaming is organized as set of record whose type
55 is indicated by a tag. */
56 enum LTO_symtab_tags
58 /* Must leave 0 for the stopper. */
60 /* Cgraph node without body available. */
61 LTO_symtab_unavail_node = 1,
62 /* Cgraph node with function body. */
63 LTO_symtab_analyzed_node,
64 /* Cgraph edges. */
65 LTO_symtab_edge,
66 LTO_symtab_indirect_edge,
67 LTO_symtab_variable,
68 LTO_symtab_last_tag
71 /* Create a new symtab encoder.
72 if FOR_INPUT, the encoder allocate only datastructures needed
73 to read the symtab. */
75 lto_symtab_encoder_t
76 lto_symtab_encoder_new (bool for_input)
78 lto_symtab_encoder_t encoder = XCNEW (struct lto_symtab_encoder_d);
80 if (!for_input)
81 encoder->map = new hash_map<symtab_node *, size_t>;
82 encoder->nodes.create (0);
83 return encoder;
87 /* Delete ENCODER and its components. */
89 void
90 lto_symtab_encoder_delete (lto_symtab_encoder_t encoder)
92 encoder->nodes.release ();
93 if (encoder->map)
94 delete encoder->map;
95 free (encoder);
99 /* Return the existing reference number of NODE in the symtab encoder in
100 output block OB. Assign a new reference if this is the first time
101 NODE is encoded. */
104 lto_symtab_encoder_encode (lto_symtab_encoder_t encoder,
105 symtab_node *node)
107 int ref;
109 if (!encoder->map)
111 lto_encoder_entry entry = {node, false, false, false};
113 ref = encoder->nodes.length ();
114 encoder->nodes.safe_push (entry);
115 return ref;
118 size_t *slot = encoder->map->get (node);
119 if (!slot || !*slot)
121 lto_encoder_entry entry = {node, false, false, false};
122 ref = encoder->nodes.length ();
123 if (!slot)
124 encoder->map->put (node, ref + 1);
125 encoder->nodes.safe_push (entry);
127 else
128 ref = *slot - 1;
130 return ref;
133 /* Remove NODE from encoder. */
135 bool
136 lto_symtab_encoder_delete_node (lto_symtab_encoder_t encoder,
137 symtab_node *node)
139 int index;
140 lto_encoder_entry last_node;
142 size_t *slot = encoder->map->get (node);
143 if (slot == NULL || !*slot)
144 return false;
146 index = *slot - 1;
147 gcc_checking_assert (encoder->nodes[index].node == node);
149 /* Remove from vector. We do this by swapping node with the last element
150 of the vector. */
151 last_node = encoder->nodes.pop ();
152 if (last_node.node != node)
154 gcc_assert (encoder->map->put (last_node.node, index + 1));
156 /* Move the last element to the original spot of NODE. */
157 encoder->nodes[index] = last_node;
160 /* Remove element from hash table. */
161 encoder->map->remove (node);
162 return true;
166 /* Return TRUE if we should encode the body of NODE (if any). */
168 bool
169 lto_symtab_encoder_encode_body_p (lto_symtab_encoder_t encoder,
170 struct cgraph_node *node)
172 int index = lto_symtab_encoder_lookup (encoder, node);
173 return encoder->nodes[index].body;
176 /* Specify that we encode the body of NODE in this partition. */
178 static void
179 lto_set_symtab_encoder_encode_body (lto_symtab_encoder_t encoder,
180 struct cgraph_node *node)
182 int index = lto_symtab_encoder_encode (encoder, node);
183 gcc_checking_assert (encoder->nodes[index].node == node);
184 encoder->nodes[index].body = true;
187 /* Return TRUE if we should encode initializer of NODE (if any). */
189 bool
190 lto_symtab_encoder_encode_initializer_p (lto_symtab_encoder_t encoder,
191 varpool_node *node)
193 int index = lto_symtab_encoder_lookup (encoder, node);
194 if (index == LCC_NOT_FOUND)
195 return false;
196 return encoder->nodes[index].initializer;
199 /* Specify that we should encode initializer of NODE (if any). */
201 static void
202 lto_set_symtab_encoder_encode_initializer (lto_symtab_encoder_t encoder,
203 varpool_node *node)
205 int index = lto_symtab_encoder_lookup (encoder, node);
206 encoder->nodes[index].initializer = true;
209 /* Return TRUE if NODE is in this partition. */
211 bool
212 lto_symtab_encoder_in_partition_p (lto_symtab_encoder_t encoder,
213 symtab_node *node)
215 int index = lto_symtab_encoder_lookup (encoder, node);
216 if (index == LCC_NOT_FOUND)
217 return false;
218 return encoder->nodes[index].in_partition;
221 /* Specify that NODE is in this partition. */
223 void
224 lto_set_symtab_encoder_in_partition (lto_symtab_encoder_t encoder,
225 symtab_node *node)
227 int index = lto_symtab_encoder_encode (encoder, node);
228 encoder->nodes[index].in_partition = true;
231 /* Output the cgraph EDGE to OB using ENCODER. */
233 static void
234 lto_output_edge (struct lto_simple_output_block *ob, struct cgraph_edge *edge,
235 lto_symtab_encoder_t encoder)
237 unsigned int uid;
238 intptr_t ref;
239 struct bitpack_d bp;
241 if (edge->indirect_unknown_callee)
242 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
243 LTO_symtab_indirect_edge);
244 else
245 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
246 LTO_symtab_edge);
248 ref = lto_symtab_encoder_lookup (encoder, edge->caller);
249 gcc_assert (ref != LCC_NOT_FOUND);
250 streamer_write_hwi_stream (ob->main_stream, ref);
252 if (!edge->indirect_unknown_callee)
254 ref = lto_symtab_encoder_lookup (encoder, edge->callee);
255 gcc_assert (ref != LCC_NOT_FOUND);
256 streamer_write_hwi_stream (ob->main_stream, ref);
259 streamer_write_gcov_count_stream (ob->main_stream, edge->count);
261 bp = bitpack_create (ob->main_stream);
262 uid = (!gimple_has_body_p (edge->caller->decl) || edge->caller->thunk.thunk_p
263 ? edge->lto_stmt_uid : gimple_uid (edge->call_stmt) + 1);
264 bp_pack_enum (&bp, cgraph_inline_failed_t,
265 CIF_N_REASONS, edge->inline_failed);
266 bp_pack_var_len_unsigned (&bp, uid);
267 bp_pack_var_len_unsigned (&bp, edge->frequency);
268 bp_pack_value (&bp, edge->indirect_inlining_edge, 1);
269 bp_pack_value (&bp, edge->speculative, 1);
270 bp_pack_value (&bp, edge->call_stmt_cannot_inline_p, 1);
271 gcc_assert (!edge->call_stmt_cannot_inline_p
272 || edge->inline_failed != CIF_BODY_NOT_AVAILABLE);
273 bp_pack_value (&bp, edge->can_throw_external, 1);
274 bp_pack_value (&bp, edge->in_polymorphic_cdtor, 1);
275 if (edge->indirect_unknown_callee)
277 int flags = edge->indirect_info->ecf_flags;
278 bp_pack_value (&bp, (flags & ECF_CONST) != 0, 1);
279 bp_pack_value (&bp, (flags & ECF_PURE) != 0, 1);
280 bp_pack_value (&bp, (flags & ECF_NORETURN) != 0, 1);
281 bp_pack_value (&bp, (flags & ECF_MALLOC) != 0, 1);
282 bp_pack_value (&bp, (flags & ECF_NOTHROW) != 0, 1);
283 bp_pack_value (&bp, (flags & ECF_RETURNS_TWICE) != 0, 1);
284 /* Flags that should not appear on indirect calls. */
285 gcc_assert (!(flags & (ECF_LOOPING_CONST_OR_PURE
286 | ECF_MAY_BE_ALLOCA
287 | ECF_SIBCALL
288 | ECF_LEAF
289 | ECF_NOVOPS)));
291 streamer_write_bitpack (&bp);
292 if (edge->indirect_unknown_callee)
294 streamer_write_hwi_stream (ob->main_stream,
295 edge->indirect_info->common_target_id);
296 if (edge->indirect_info->common_target_id)
297 streamer_write_hwi_stream
298 (ob->main_stream, edge->indirect_info->common_target_probability);
302 /* Return if NODE contain references from other partitions. */
304 bool
305 referenced_from_other_partition_p (symtab_node *node, lto_symtab_encoder_t encoder)
307 int i;
308 struct ipa_ref *ref = NULL;
310 for (i = 0; node->iterate_referring (i, ref); i++)
312 /* Ignore references from non-offloadable nodes while streaming NODE into
313 offload LTO section. */
314 if (!ref->referring->need_lto_streaming)
315 continue;
317 if (ref->referring->in_other_partition
318 || !lto_symtab_encoder_in_partition_p (encoder, ref->referring))
319 return true;
321 return false;
324 /* Return true when node is reachable from other partition. */
326 bool
327 reachable_from_other_partition_p (struct cgraph_node *node, lto_symtab_encoder_t encoder)
329 struct cgraph_edge *e;
330 if (!node->definition)
331 return false;
332 if (node->global.inlined_to)
333 return false;
334 for (e = node->callers; e; e = e->next_caller)
336 /* Ignore references from non-offloadable nodes while streaming NODE into
337 offload LTO section. */
338 if (!e->caller->need_lto_streaming)
339 continue;
341 if (e->caller->in_other_partition
342 || !lto_symtab_encoder_in_partition_p (encoder, e->caller))
343 return true;
345 return false;
348 /* Return if NODE contain references from other partitions. */
350 bool
351 referenced_from_this_partition_p (symtab_node *node,
352 lto_symtab_encoder_t encoder)
354 int i;
355 struct ipa_ref *ref = NULL;
357 for (i = 0; node->iterate_referring (i, ref); i++)
358 if (lto_symtab_encoder_in_partition_p (encoder, ref->referring))
359 return true;
360 return false;
363 /* Return true when node is reachable from other partition. */
365 bool
366 reachable_from_this_partition_p (struct cgraph_node *node, lto_symtab_encoder_t encoder)
368 struct cgraph_edge *e;
369 for (e = node->callers; e; e = e->next_caller)
370 if (lto_symtab_encoder_in_partition_p (encoder, e->caller))
371 return true;
372 return false;
375 /* Output the cgraph NODE to OB. ENCODER is used to find the
376 reference number of NODE->inlined_to. SET is the set of nodes we
377 are writing to the current file. If NODE is not in SET, then NODE
378 is a boundary of a cgraph_node_set and we pretend NODE just has a
379 decl and no callees. WRITTEN_DECLS is the set of FUNCTION_DECLs
380 that have had their callgraph node written so far. This is used to
381 determine if NODE is a clone of a previously written node. */
383 static void
384 lto_output_node (struct lto_simple_output_block *ob, struct cgraph_node *node,
385 lto_symtab_encoder_t encoder)
387 unsigned int tag;
388 struct bitpack_d bp;
389 bool boundary_p;
390 intptr_t ref;
391 bool in_other_partition = false;
392 struct cgraph_node *clone_of, *ultimate_clone_of;
393 ipa_opt_pass_d *pass;
394 int i;
395 const char *comdat;
396 const char *section;
397 tree group;
399 boundary_p = !lto_symtab_encoder_in_partition_p (encoder, node);
401 if (node->analyzed && (!boundary_p || node->alias
402 || (node->thunk.thunk_p && !node->global.inlined_to)))
403 tag = LTO_symtab_analyzed_node;
404 else
405 tag = LTO_symtab_unavail_node;
407 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
408 tag);
409 streamer_write_hwi_stream (ob->main_stream, node->order);
411 /* In WPA mode, we only output part of the call-graph. Also, we
412 fake cgraph node attributes. There are two cases that we care.
414 Boundary nodes: There are nodes that are not part of SET but are
415 called from within SET. We artificially make them look like
416 externally visible nodes with no function body.
418 Cherry-picked nodes: These are nodes we pulled from other
419 translation units into SET during IPA-inlining. We make them as
420 local static nodes to prevent clashes with other local statics. */
421 if (boundary_p && node->analyzed
422 && node->get_partitioning_class () == SYMBOL_PARTITION)
424 /* Inline clones can not be part of boundary.
425 gcc_assert (!node->global.inlined_to);
427 FIXME: At the moment they can be, when partition contains an inline
428 clone that is clone of inline clone from outside partition. We can
429 reshape the clone tree and make other tree to be the root, but it
430 needs a bit extra work and will be promplty done by cgraph_remove_node
431 after reading back. */
432 in_other_partition = 1;
435 clone_of = node->clone_of;
436 while (clone_of
437 && (ref = lto_symtab_encoder_lookup (encoder, clone_of)) == LCC_NOT_FOUND)
438 if (clone_of->prev_sibling_clone)
439 clone_of = clone_of->prev_sibling_clone;
440 else
441 clone_of = clone_of->clone_of;
443 /* See if body of the master function is output. If not, we are seeing only
444 an declaration and we do not need to pass down clone tree. */
445 ultimate_clone_of = clone_of;
446 while (ultimate_clone_of && ultimate_clone_of->clone_of)
447 ultimate_clone_of = ultimate_clone_of->clone_of;
449 if (clone_of && !lto_symtab_encoder_encode_body_p (encoder, ultimate_clone_of))
450 clone_of = NULL;
452 if (tag == LTO_symtab_analyzed_node)
453 gcc_assert (clone_of || !node->clone_of);
454 if (!clone_of)
455 streamer_write_hwi_stream (ob->main_stream, LCC_NOT_FOUND);
456 else
457 streamer_write_hwi_stream (ob->main_stream, ref);
460 lto_output_fn_decl_index (ob->decl_state, ob->main_stream, node->decl);
461 streamer_write_gcov_count_stream (ob->main_stream, node->count);
462 streamer_write_hwi_stream (ob->main_stream, node->count_materialization_scale);
464 streamer_write_hwi_stream (ob->main_stream,
465 node->ipa_transforms_to_apply.length ());
466 FOR_EACH_VEC_ELT (node->ipa_transforms_to_apply, i, pass)
467 streamer_write_hwi_stream (ob->main_stream, pass->static_pass_number);
469 if (tag == LTO_symtab_analyzed_node)
471 if (node->global.inlined_to)
473 ref = lto_symtab_encoder_lookup (encoder, node->global.inlined_to);
474 gcc_assert (ref != LCC_NOT_FOUND);
476 else
477 ref = LCC_NOT_FOUND;
479 streamer_write_hwi_stream (ob->main_stream, ref);
482 group = node->get_comdat_group ();
483 if (group)
484 comdat = IDENTIFIER_POINTER (group);
485 else
486 comdat = "";
487 streamer_write_data_stream (ob->main_stream, comdat, strlen (comdat) + 1);
489 if (group)
491 if (node->same_comdat_group)
493 ref = LCC_NOT_FOUND;
494 for (struct symtab_node *n = node->same_comdat_group;
495 ref == LCC_NOT_FOUND && n != node; n = n->same_comdat_group)
496 ref = lto_symtab_encoder_lookup (encoder, n);
498 else
499 ref = LCC_NOT_FOUND;
500 streamer_write_hwi_stream (ob->main_stream, ref);
503 section = node->get_section ();
504 if (!section)
505 section = "";
507 streamer_write_hwi_stream (ob->main_stream, node->tp_first_run);
509 bp = bitpack_create (ob->main_stream);
510 bp_pack_value (&bp, node->local.local, 1);
511 bp_pack_value (&bp, node->externally_visible, 1);
512 bp_pack_value (&bp, node->no_reorder, 1);
513 bp_pack_value (&bp, node->definition, 1);
514 bp_pack_value (&bp, node->local.versionable, 1);
515 bp_pack_value (&bp, node->local.can_change_signature, 1);
516 bp_pack_value (&bp, node->local.redefined_extern_inline, 1);
517 bp_pack_value (&bp, node->force_output, 1);
518 bp_pack_value (&bp, node->forced_by_abi, 1);
519 bp_pack_value (&bp, node->unique_name, 1);
520 bp_pack_value (&bp, node->body_removed, 1);
521 bp_pack_value (&bp, node->implicit_section, 1);
522 bp_pack_value (&bp, node->address_taken, 1);
523 bp_pack_value (&bp, tag == LTO_symtab_analyzed_node
524 && node->get_partitioning_class () == SYMBOL_PARTITION
525 && (reachable_from_other_partition_p (node, encoder)
526 || referenced_from_other_partition_p (node, encoder)), 1);
527 bp_pack_value (&bp, node->lowered, 1);
528 bp_pack_value (&bp, in_other_partition, 1);
529 bp_pack_value (&bp, node->alias, 1);
530 bp_pack_value (&bp, node->transparent_alias, 1);
531 bp_pack_value (&bp, node->weakref, 1);
532 bp_pack_value (&bp, node->frequency, 2);
533 bp_pack_value (&bp, node->only_called_at_startup, 1);
534 bp_pack_value (&bp, node->only_called_at_exit, 1);
535 bp_pack_value (&bp, node->tm_clone, 1);
536 bp_pack_value (&bp, node->calls_comdat_local, 1);
537 bp_pack_value (&bp, node->icf_merged, 1);
538 bp_pack_value (&bp, node->nonfreeing_fn, 1);
539 bp_pack_value (&bp, node->thunk.thunk_p, 1);
540 bp_pack_value (&bp, node->parallelized_function, 1);
541 bp_pack_enum (&bp, ld_plugin_symbol_resolution,
542 LDPR_NUM_KNOWN, node->resolution);
543 bp_pack_value (&bp, node->instrumentation_clone, 1);
544 bp_pack_value (&bp, node->split_part, 1);
545 streamer_write_bitpack (&bp);
546 streamer_write_data_stream (ob->main_stream, section, strlen (section) + 1);
548 if (node->thunk.thunk_p)
550 streamer_write_uhwi_stream
551 (ob->main_stream,
552 1 + (node->thunk.this_adjusting != 0) * 2
553 + (node->thunk.virtual_offset_p != 0) * 4
554 + (node->thunk.add_pointer_bounds_args != 0) * 8);
555 streamer_write_uhwi_stream (ob->main_stream, node->thunk.fixed_offset);
556 streamer_write_uhwi_stream (ob->main_stream, node->thunk.virtual_value);
558 streamer_write_hwi_stream (ob->main_stream, node->profile_id);
559 if (DECL_STATIC_CONSTRUCTOR (node->decl))
560 streamer_write_hwi_stream (ob->main_stream, node->get_init_priority ());
561 if (DECL_STATIC_DESTRUCTOR (node->decl))
562 streamer_write_hwi_stream (ob->main_stream, node->get_fini_priority ());
564 if (node->instrumentation_clone)
565 lto_output_fn_decl_index (ob->decl_state, ob->main_stream, node->orig_decl);
568 /* Output the varpool NODE to OB.
569 If NODE is not in SET, then NODE is a boundary. */
571 static void
572 lto_output_varpool_node (struct lto_simple_output_block *ob, varpool_node *node,
573 lto_symtab_encoder_t encoder)
575 bool boundary_p = !lto_symtab_encoder_in_partition_p (encoder, node);
576 bool encode_initializer_p
577 = (node->definition
578 && lto_symtab_encoder_encode_initializer_p (encoder, node));
579 struct bitpack_d bp;
580 int ref;
581 const char *comdat;
582 const char *section;
583 tree group;
585 gcc_assert (!encode_initializer_p || node->definition);
586 gcc_assert (boundary_p || encode_initializer_p);
588 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
589 LTO_symtab_variable);
590 streamer_write_hwi_stream (ob->main_stream, node->order);
591 lto_output_var_decl_index (ob->decl_state, ob->main_stream, node->decl);
592 bp = bitpack_create (ob->main_stream);
593 bp_pack_value (&bp, node->externally_visible, 1);
594 bp_pack_value (&bp, node->no_reorder, 1);
595 bp_pack_value (&bp, node->force_output, 1);
596 bp_pack_value (&bp, node->forced_by_abi, 1);
597 bp_pack_value (&bp, node->unique_name, 1);
598 bp_pack_value (&bp,
599 node->body_removed
600 || (!encode_initializer_p && !node->alias && node->definition),
602 bp_pack_value (&bp, node->implicit_section, 1);
603 bp_pack_value (&bp, node->writeonly, 1);
604 bp_pack_value (&bp, node->definition && (encode_initializer_p || node->alias),
606 bp_pack_value (&bp, node->alias, 1);
607 bp_pack_value (&bp, node->transparent_alias, 1);
608 bp_pack_value (&bp, node->weakref, 1);
609 bp_pack_value (&bp, node->analyzed && (!boundary_p || node->alias), 1);
610 gcc_assert (node->definition || !node->analyzed);
611 /* Constant pool initializers can be de-unified into individual ltrans units.
612 FIXME: Alternatively at -Os we may want to avoid generating for them the local
613 labels and share them across LTRANS partitions. */
614 if (node->get_partitioning_class () != SYMBOL_PARTITION)
616 bp_pack_value (&bp, 0, 1); /* used_from_other_parition. */
617 bp_pack_value (&bp, 0, 1); /* in_other_partition. */
619 else
621 bp_pack_value (&bp, node->definition
622 && referenced_from_other_partition_p (node, encoder), 1);
623 bp_pack_value (&bp, node->analyzed
624 && boundary_p && !DECL_EXTERNAL (node->decl), 1);
625 /* in_other_partition. */
627 bp_pack_value (&bp, node->tls_model, 3);
628 bp_pack_value (&bp, node->used_by_single_function, 1);
629 bp_pack_value (&bp, node->need_bounds_init, 1);
630 streamer_write_bitpack (&bp);
632 group = node->get_comdat_group ();
633 if (group)
634 comdat = IDENTIFIER_POINTER (group);
635 else
636 comdat = "";
637 streamer_write_data_stream (ob->main_stream, comdat, strlen (comdat) + 1);
639 if (group)
641 if (node->same_comdat_group)
643 ref = LCC_NOT_FOUND;
644 for (struct symtab_node *n = node->same_comdat_group;
645 ref == LCC_NOT_FOUND && n != node; n = n->same_comdat_group)
646 ref = lto_symtab_encoder_lookup (encoder, n);
648 else
649 ref = LCC_NOT_FOUND;
650 streamer_write_hwi_stream (ob->main_stream, ref);
653 section = node->get_section ();
654 if (!section)
655 section = "";
656 streamer_write_data_stream (ob->main_stream, section, strlen (section) + 1);
658 streamer_write_enum (ob->main_stream, ld_plugin_symbol_resolution,
659 LDPR_NUM_KNOWN, node->resolution);
662 /* Output the varpool NODE to OB.
663 If NODE is not in SET, then NODE is a boundary. */
665 static void
666 lto_output_ref (struct lto_simple_output_block *ob, struct ipa_ref *ref,
667 lto_symtab_encoder_t encoder)
669 struct bitpack_d bp;
670 int nref;
671 int uid = ref->lto_stmt_uid;
672 struct cgraph_node *node;
674 bp = bitpack_create (ob->main_stream);
675 bp_pack_value (&bp, ref->use, 3);
676 bp_pack_value (&bp, ref->speculative, 1);
677 streamer_write_bitpack (&bp);
678 nref = lto_symtab_encoder_lookup (encoder, ref->referred);
679 gcc_assert (nref != LCC_NOT_FOUND);
680 streamer_write_hwi_stream (ob->main_stream, nref);
682 node = dyn_cast <cgraph_node *> (ref->referring);
683 if (node)
685 if (ref->stmt)
686 uid = gimple_uid (ref->stmt) + 1;
687 streamer_write_hwi_stream (ob->main_stream, uid);
691 /* Stream out profile_summary to OB. */
693 static void
694 output_profile_summary (struct lto_simple_output_block *ob)
696 unsigned h_ix;
697 struct bitpack_d bp;
699 if (profile_info)
701 /* We do not output num and run_max, they are not used by
702 GCC profile feedback and they are difficult to merge from multiple
703 units. */
704 gcc_assert (profile_info->runs);
705 streamer_write_uhwi_stream (ob->main_stream, profile_info->runs);
706 streamer_write_gcov_count_stream (ob->main_stream, profile_info->sum_max);
708 /* sum_all is needed for computing the working set with the
709 histogram. */
710 streamer_write_gcov_count_stream (ob->main_stream, profile_info->sum_all);
712 /* Create and output a bitpack of non-zero histogram entries indices. */
713 bp = bitpack_create (ob->main_stream);
714 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
715 bp_pack_value (&bp, profile_info->histogram[h_ix].num_counters > 0, 1);
716 streamer_write_bitpack (&bp);
717 /* Now stream out only those non-zero entries. */
718 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
720 if (!profile_info->histogram[h_ix].num_counters)
721 continue;
722 streamer_write_gcov_count_stream (ob->main_stream,
723 profile_info->histogram[h_ix].num_counters);
724 streamer_write_gcov_count_stream (ob->main_stream,
725 profile_info->histogram[h_ix].min_value);
726 streamer_write_gcov_count_stream (ob->main_stream,
727 profile_info->histogram[h_ix].cum_value);
729 /* IPA-profile computes hot bb threshold based on cumulated
730 whole program profile. We need to stream it down to ltrans. */
731 if (flag_wpa)
732 streamer_write_gcov_count_stream (ob->main_stream,
733 get_hot_bb_threshold ());
735 else
736 streamer_write_uhwi_stream (ob->main_stream, 0);
739 /* Output all callees or indirect outgoing edges. EDGE must be the first such
740 edge. */
742 static void
743 output_outgoing_cgraph_edges (struct cgraph_edge *edge,
744 struct lto_simple_output_block *ob,
745 lto_symtab_encoder_t encoder)
747 if (!edge)
748 return;
750 /* Output edges in backward direction, so the reconstructed callgraph match
751 and it is easy to associate call sites in the IPA pass summaries. */
752 while (edge->next_callee)
753 edge = edge->next_callee;
754 for (; edge; edge = edge->prev_callee)
755 lto_output_edge (ob, edge, encoder);
758 /* Output the part of the cgraph in SET. */
760 static void
761 output_refs (lto_symtab_encoder_t encoder)
763 struct lto_simple_output_block *ob;
764 int count;
765 struct ipa_ref *ref;
767 ob = lto_create_simple_output_block (LTO_section_refs);
769 for (int i = 0; i < lto_symtab_encoder_size (encoder); i++)
771 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
773 /* IPA_REF_ALIAS and IPA_REF_CHKP references are always preserved
774 in the boundary. Alias node can't have other references and
775 can be always handled as if it's not in the boundary. */
776 if (!node->alias && !lto_symtab_encoder_in_partition_p (encoder, node))
778 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
779 /* Output IPA_REF_CHKP reference. */
780 if (cnode
781 && cnode->instrumented_version
782 && !cnode->instrumentation_clone)
784 for (int i = 0; node->iterate_reference (i, ref); i++)
785 if (ref->use == IPA_REF_CHKP)
787 if (lto_symtab_encoder_lookup (encoder, ref->referred)
788 != LCC_NOT_FOUND)
790 int nref = lto_symtab_encoder_lookup (encoder, node);
791 streamer_write_gcov_count_stream (ob->main_stream, 1);
792 streamer_write_uhwi_stream (ob->main_stream, nref);
793 lto_output_ref (ob, ref, encoder);
795 break;
798 continue;
801 count = node->ref_list.nreferences ();
802 if (count)
804 streamer_write_gcov_count_stream (ob->main_stream, count);
805 streamer_write_uhwi_stream (ob->main_stream,
806 lto_symtab_encoder_lookup (encoder, node));
807 for (int i = 0; node->iterate_reference (i, ref); i++)
808 lto_output_ref (ob, ref, encoder);
812 streamer_write_uhwi_stream (ob->main_stream, 0);
814 lto_destroy_simple_output_block (ob);
817 /* Add NODE into encoder as well as nodes it is cloned from.
818 Do it in a way so clones appear first. */
820 static void
821 add_node_to (lto_symtab_encoder_t encoder, struct cgraph_node *node,
822 bool include_body)
824 if (node->clone_of)
825 add_node_to (encoder, node->clone_of, include_body);
826 else if (include_body)
827 lto_set_symtab_encoder_encode_body (encoder, node);
828 lto_symtab_encoder_encode (encoder, node);
831 /* Add all references in NODE to encoders. */
833 static void
834 create_references (lto_symtab_encoder_t encoder, symtab_node *node)
836 int i;
837 struct ipa_ref *ref = NULL;
838 for (i = 0; node->iterate_reference (i, ref); i++)
839 if (is_a <cgraph_node *> (ref->referred))
840 add_node_to (encoder, dyn_cast <cgraph_node *> (ref->referred), false);
841 else
842 lto_symtab_encoder_encode (encoder, ref->referred);
845 /* Select what needs to be streamed out. In regular lto mode stream everything.
846 In offload lto mode stream only nodes marked as offloadable. */
847 void
848 select_what_to_stream (void)
850 struct symtab_node *snode;
851 FOR_EACH_SYMBOL (snode)
852 snode->need_lto_streaming = !lto_stream_offload_p || snode->offloadable;
855 /* Find all symbols we want to stream into given partition and insert them
856 to encoders.
858 The function actually replaces IN_ENCODER by new one. The reason is that
859 streaming code needs clone's origin to be streamed before clone. This
860 means that we need to insert the nodes in specific order. This order is
861 ignored by the partitioning logic earlier. */
863 lto_symtab_encoder_t
864 compute_ltrans_boundary (lto_symtab_encoder_t in_encoder)
866 struct cgraph_edge *edge;
867 int i;
868 lto_symtab_encoder_t encoder;
869 lto_symtab_encoder_iterator lsei;
870 hash_set<void *> reachable_call_targets;
872 encoder = lto_symtab_encoder_new (false);
874 /* Go over all entries in the IN_ENCODER and duplicate them to
875 ENCODER. At the same time insert masters of clones so
876 every master appears before clone. */
877 for (lsei = lsei_start_function_in_partition (in_encoder);
878 !lsei_end_p (lsei); lsei_next_function_in_partition (&lsei))
880 struct cgraph_node *node = lsei_cgraph_node (lsei);
881 if (!node->need_lto_streaming)
882 continue;
883 add_node_to (encoder, node, true);
884 lto_set_symtab_encoder_in_partition (encoder, node);
885 create_references (encoder, node);
887 for (lsei = lsei_start_variable_in_partition (in_encoder);
888 !lsei_end_p (lsei); lsei_next_variable_in_partition (&lsei))
890 varpool_node *vnode = lsei_varpool_node (lsei);
892 if (!vnode->need_lto_streaming)
893 continue;
894 lto_set_symtab_encoder_in_partition (encoder, vnode);
895 lto_set_symtab_encoder_encode_initializer (encoder, vnode);
896 create_references (encoder, vnode);
898 /* Pickle in also the initializer of all referenced readonly variables
899 to help folding. Constant pool variables are not shared, so we must
900 pickle those too. */
901 for (i = 0; i < lto_symtab_encoder_size (encoder); i++)
903 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
904 if (varpool_node *vnode = dyn_cast <varpool_node *> (node))
906 if (!lto_symtab_encoder_encode_initializer_p (encoder,
907 vnode)
908 && (((vnode->ctor_useable_for_folding_p ()
909 && (!DECL_VIRTUAL_P (vnode->decl)
910 || !flag_wpa
911 || flag_ltrans_devirtualize))
912 || POINTER_BOUNDS_P (vnode->decl))))
914 lto_set_symtab_encoder_encode_initializer (encoder, vnode);
915 create_references (encoder, vnode);
920 /* Go over all the nodes again to include callees that are not in
921 SET. */
922 for (lsei = lsei_start_function_in_partition (encoder);
923 !lsei_end_p (lsei); lsei_next_function_in_partition (&lsei))
925 struct cgraph_node *node = lsei_cgraph_node (lsei);
926 for (edge = node->callees; edge; edge = edge->next_callee)
928 struct cgraph_node *callee = edge->callee;
929 if (!lto_symtab_encoder_in_partition_p (encoder, callee))
931 /* We should have moved all the inlines. */
932 gcc_assert (!callee->global.inlined_to);
933 add_node_to (encoder, callee, false);
936 /* Add all possible targets for late devirtualization. */
937 if (flag_ltrans_devirtualize || !flag_wpa)
938 for (edge = node->indirect_calls; edge; edge = edge->next_callee)
939 if (edge->indirect_info->polymorphic)
941 unsigned int i;
942 void *cache_token;
943 bool final;
944 vec <cgraph_node *>targets
945 = possible_polymorphic_call_targets
946 (edge, &final, &cache_token);
947 if (!reachable_call_targets.add (cache_token))
949 for (i = 0; i < targets.length (); i++)
951 struct cgraph_node *callee = targets[i];
953 /* Adding an external declarations into the unit serves
954 no purpose and just increases its boundary. */
955 if (callee->definition
956 && !lto_symtab_encoder_in_partition_p
957 (encoder, callee))
959 gcc_assert (!callee->global.inlined_to);
960 add_node_to (encoder, callee, false);
966 /* Be sure to also insert alias targert and thunk callees. These needs
967 to stay to aid local calling conventions. */
968 for (i = 0; i < lto_symtab_encoder_size (encoder); i++)
970 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
971 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
973 if (node->alias && node->analyzed)
974 create_references (encoder, node);
975 if (cnode
976 && cnode->thunk.thunk_p && !cnode->global.inlined_to)
977 add_node_to (encoder, cnode->callees->callee, false);
978 while (node->transparent_alias && node->analyzed)
980 node = node->get_alias_target ();
981 if (is_a <cgraph_node *> (node))
982 add_node_to (encoder, dyn_cast <cgraph_node *> (node),
983 false);
984 else
985 lto_symtab_encoder_encode (encoder, node);
988 lto_symtab_encoder_delete (in_encoder);
989 return encoder;
992 /* Output the part of the symtab in SET and VSET. */
994 void
995 output_symtab (void)
997 struct cgraph_node *node;
998 struct lto_simple_output_block *ob;
999 int i, n_nodes;
1000 lto_symtab_encoder_t encoder;
1002 if (flag_wpa)
1003 output_cgraph_opt_summary ();
1005 ob = lto_create_simple_output_block (LTO_section_symtab_nodes);
1007 output_profile_summary (ob);
1009 /* An encoder for cgraph nodes should have been created by
1010 ipa_write_summaries_1. */
1011 gcc_assert (ob->decl_state->symtab_node_encoder);
1012 encoder = ob->decl_state->symtab_node_encoder;
1014 /* Write out the nodes. We must first output a node and then its clones,
1015 otherwise at a time reading back the node there would be nothing to clone
1016 from. */
1017 n_nodes = lto_symtab_encoder_size (encoder);
1018 for (i = 0; i < n_nodes; i++)
1020 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
1021 if (cgraph_node *cnode = dyn_cast <cgraph_node *> (node))
1022 lto_output_node (ob, cnode, encoder);
1023 else
1024 lto_output_varpool_node (ob, dyn_cast<varpool_node *> (node), encoder);
1027 /* Go over the nodes in SET again to write edges. */
1028 for (int i = 0; i < lto_symtab_encoder_size (encoder); i++)
1030 node = dyn_cast <cgraph_node *> (lto_symtab_encoder_deref (encoder, i));
1031 if (node
1032 && ((node->thunk.thunk_p && !node->global.inlined_to)
1033 || lto_symtab_encoder_in_partition_p (encoder, node)))
1035 output_outgoing_cgraph_edges (node->callees, ob, encoder);
1036 output_outgoing_cgraph_edges (node->indirect_calls, ob, encoder);
1040 streamer_write_uhwi_stream (ob->main_stream, 0);
1042 lto_destroy_simple_output_block (ob);
1044 /* Emit toplevel asms.
1045 When doing WPA we must output every asm just once. Since we do not partition asm
1046 nodes at all, output them to first output. This is kind of hack, but should work
1047 well. */
1048 if (!asm_nodes_output)
1050 asm_nodes_output = true;
1051 lto_output_toplevel_asms ();
1054 output_refs (encoder);
1057 /* Return identifier encoded in IB as a plain string. */
1059 static tree
1060 read_identifier (struct lto_input_block *ib)
1062 unsigned int len = strnlen (ib->data + ib->p, ib->len - ib->p - 1);
1063 tree id;
1065 if (ib->data[ib->p + len])
1066 lto_section_overrun (ib);
1067 if (!len)
1069 ib->p++;
1070 return NULL;
1072 id = get_identifier (ib->data + ib->p);
1073 ib->p += len + 1;
1074 return id;
1077 /* Return string encoded in IB, NULL if string is empty. */
1079 static const char *
1080 read_string (struct lto_input_block *ib)
1082 unsigned int len = strnlen (ib->data + ib->p, ib->len - ib->p - 1);
1083 const char *str;
1085 if (ib->data[ib->p + len])
1086 lto_section_overrun (ib);
1087 if (!len)
1089 ib->p++;
1090 return NULL;
1092 str = ib->data + ib->p;
1093 ib->p += len + 1;
1094 return str;
1097 /* Output function/variable tables that will allow libgomp to look up offload
1098 target code.
1099 OFFLOAD_FUNCS is filled in expand_omp_target, OFFLOAD_VARS is filled in
1100 varpool_node::get_create. In WHOPR (partitioned) mode during the WPA stage
1101 both OFFLOAD_FUNCS and OFFLOAD_VARS are filled by input_offload_tables. */
1103 void
1104 output_offload_tables (void)
1106 if (vec_safe_is_empty (offload_funcs) && vec_safe_is_empty (offload_vars))
1107 return;
1109 struct lto_simple_output_block *ob
1110 = lto_create_simple_output_block (LTO_section_offload_table);
1112 for (unsigned i = 0; i < vec_safe_length (offload_funcs); i++)
1114 streamer_write_enum (ob->main_stream, LTO_symtab_tags,
1115 LTO_symtab_last_tag, LTO_symtab_unavail_node);
1116 lto_output_fn_decl_index (ob->decl_state, ob->main_stream,
1117 (*offload_funcs)[i]);
1120 for (unsigned i = 0; i < vec_safe_length (offload_vars); i++)
1122 streamer_write_enum (ob->main_stream, LTO_symtab_tags,
1123 LTO_symtab_last_tag, LTO_symtab_variable);
1124 lto_output_var_decl_index (ob->decl_state, ob->main_stream,
1125 (*offload_vars)[i]);
1128 streamer_write_uhwi_stream (ob->main_stream, 0);
1129 lto_destroy_simple_output_block (ob);
1131 /* In WHOPR mode during the WPA stage the joint offload tables need to be
1132 streamed to one partition only. That's why we free offload_funcs and
1133 offload_vars after the first call of output_offload_tables. */
1134 if (flag_wpa)
1136 vec_free (offload_funcs);
1137 vec_free (offload_vars);
1141 /* Overwrite the information in NODE based on FILE_DATA, TAG, FLAGS,
1142 STACK_SIZE, SELF_TIME and SELF_SIZE. This is called either to initialize
1143 NODE or to replace the values in it, for instance because the first
1144 time we saw it, the function body was not available but now it
1145 is. BP is a bitpack with all the bitflags for NODE read from the
1146 stream. */
1148 static void
1149 input_overwrite_node (struct lto_file_decl_data *file_data,
1150 struct cgraph_node *node,
1151 enum LTO_symtab_tags tag,
1152 struct bitpack_d *bp)
1154 node->aux = (void *) tag;
1155 node->lto_file_data = file_data;
1157 node->local.local = bp_unpack_value (bp, 1);
1158 node->externally_visible = bp_unpack_value (bp, 1);
1159 node->no_reorder = bp_unpack_value (bp, 1);
1160 node->definition = bp_unpack_value (bp, 1);
1161 node->local.versionable = bp_unpack_value (bp, 1);
1162 node->local.can_change_signature = bp_unpack_value (bp, 1);
1163 node->local.redefined_extern_inline = bp_unpack_value (bp, 1);
1164 node->force_output = bp_unpack_value (bp, 1);
1165 node->forced_by_abi = bp_unpack_value (bp, 1);
1166 node->unique_name = bp_unpack_value (bp, 1);
1167 node->body_removed = bp_unpack_value (bp, 1);
1168 node->implicit_section = bp_unpack_value (bp, 1);
1169 node->address_taken = bp_unpack_value (bp, 1);
1170 node->used_from_other_partition = bp_unpack_value (bp, 1);
1171 node->lowered = bp_unpack_value (bp, 1);
1172 node->analyzed = tag == LTO_symtab_analyzed_node;
1173 node->in_other_partition = bp_unpack_value (bp, 1);
1174 if (node->in_other_partition
1175 /* Avoid updating decl when we are seeing just inline clone.
1176 When inlining function that has functions already inlined into it,
1177 we produce clones of inline clones.
1179 WPA partitioning might put each clone into different unit and
1180 we might end up streaming inline clone from other partition
1181 to support clone we are interested in. */
1182 && (!node->clone_of
1183 || node->clone_of->decl != node->decl))
1185 DECL_EXTERNAL (node->decl) = 1;
1186 TREE_STATIC (node->decl) = 0;
1188 node->alias = bp_unpack_value (bp, 1);
1189 node->transparent_alias = bp_unpack_value (bp, 1);
1190 node->weakref = bp_unpack_value (bp, 1);
1191 node->frequency = (enum node_frequency)bp_unpack_value (bp, 2);
1192 node->only_called_at_startup = bp_unpack_value (bp, 1);
1193 node->only_called_at_exit = bp_unpack_value (bp, 1);
1194 node->tm_clone = bp_unpack_value (bp, 1);
1195 node->calls_comdat_local = bp_unpack_value (bp, 1);
1196 node->icf_merged = bp_unpack_value (bp, 1);
1197 node->nonfreeing_fn = bp_unpack_value (bp, 1);
1198 node->thunk.thunk_p = bp_unpack_value (bp, 1);
1199 node->parallelized_function = bp_unpack_value (bp, 1);
1200 node->resolution = bp_unpack_enum (bp, ld_plugin_symbol_resolution,
1201 LDPR_NUM_KNOWN);
1202 node->instrumentation_clone = bp_unpack_value (bp, 1);
1203 node->split_part = bp_unpack_value (bp, 1);
1204 gcc_assert (flag_ltrans
1205 || (!node->in_other_partition
1206 && !node->used_from_other_partition));
1209 /* Return string alias is alias of. */
1211 static tree
1212 get_alias_symbol (tree decl)
1214 tree alias = lookup_attribute ("alias", DECL_ATTRIBUTES (decl));
1215 return get_identifier (TREE_STRING_POINTER
1216 (TREE_VALUE (TREE_VALUE (alias))));
1219 /* Read a node from input_block IB. TAG is the node's tag just read.
1220 Return the node read or overwriten. */
1222 static struct cgraph_node *
1223 input_node (struct lto_file_decl_data *file_data,
1224 struct lto_input_block *ib,
1225 enum LTO_symtab_tags tag,
1226 vec<symtab_node *> nodes)
1228 gcc::pass_manager *passes = g->get_passes ();
1229 tree fn_decl;
1230 struct cgraph_node *node;
1231 struct bitpack_d bp;
1232 unsigned decl_index;
1233 int ref = LCC_NOT_FOUND, ref2 = LCC_NOT_FOUND;
1234 int clone_ref;
1235 int order;
1236 int i, count;
1237 tree group;
1238 const char *section;
1239 order = streamer_read_hwi (ib) + order_base;
1240 clone_ref = streamer_read_hwi (ib);
1242 decl_index = streamer_read_uhwi (ib);
1243 fn_decl = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1245 if (clone_ref != LCC_NOT_FOUND)
1247 node = dyn_cast<cgraph_node *> (nodes[clone_ref])->create_clone (fn_decl,
1248 0, CGRAPH_FREQ_BASE, false,
1249 vNULL, false, NULL, NULL);
1251 else
1253 /* Declaration of functions can be already merged with a declaration
1254 from other input file. We keep cgraph unmerged until after streaming
1255 of ipa passes is done. Alays forcingly create a fresh node. */
1256 node = symtab->create_empty ();
1257 node->decl = fn_decl;
1258 node->register_symbol ();
1261 node->order = order;
1262 if (order >= symtab->order)
1263 symtab->order = order + 1;
1265 node->count = streamer_read_gcov_count (ib);
1266 node->count_materialization_scale = streamer_read_hwi (ib);
1268 count = streamer_read_hwi (ib);
1269 node->ipa_transforms_to_apply = vNULL;
1270 for (i = 0; i < count; i++)
1272 opt_pass *pass;
1273 int pid = streamer_read_hwi (ib);
1275 gcc_assert (pid < passes->passes_by_id_size);
1276 pass = passes->passes_by_id[pid];
1277 node->ipa_transforms_to_apply.safe_push ((ipa_opt_pass_d *) pass);
1280 if (tag == LTO_symtab_analyzed_node)
1281 ref = streamer_read_hwi (ib);
1283 group = read_identifier (ib);
1284 if (group)
1285 ref2 = streamer_read_hwi (ib);
1287 /* Make sure that we have not read this node before. Nodes that
1288 have already been read will have their tag stored in the 'aux'
1289 field. Since built-in functions can be referenced in multiple
1290 functions, they are expected to be read more than once. */
1291 if (node->aux && !DECL_BUILT_IN (node->decl))
1292 internal_error ("bytecode stream: found multiple instances of cgraph "
1293 "node with uid %d", node->uid);
1295 node->tp_first_run = streamer_read_uhwi (ib);
1297 bp = streamer_read_bitpack (ib);
1299 input_overwrite_node (file_data, node, tag, &bp);
1301 /* Store a reference for now, and fix up later to be a pointer. */
1302 node->global.inlined_to = (cgraph_node *) (intptr_t) ref;
1304 if (group)
1306 node->set_comdat_group (group);
1307 /* Store a reference for now, and fix up later to be a pointer. */
1308 node->same_comdat_group = (symtab_node *) (intptr_t) ref2;
1310 else
1311 node->same_comdat_group = (symtab_node *) (intptr_t) LCC_NOT_FOUND;
1312 section = read_string (ib);
1313 if (section)
1314 node->set_section_for_node (section);
1316 if (node->thunk.thunk_p)
1318 int type = streamer_read_uhwi (ib);
1319 HOST_WIDE_INT fixed_offset = streamer_read_uhwi (ib);
1320 HOST_WIDE_INT virtual_value = streamer_read_uhwi (ib);
1322 node->thunk.fixed_offset = fixed_offset;
1323 node->thunk.this_adjusting = (type & 2);
1324 node->thunk.virtual_value = virtual_value;
1325 node->thunk.virtual_offset_p = (type & 4);
1326 node->thunk.add_pointer_bounds_args = (type & 8);
1328 if (node->alias && !node->analyzed && node->weakref)
1329 node->alias_target = get_alias_symbol (node->decl);
1330 node->profile_id = streamer_read_hwi (ib);
1331 if (DECL_STATIC_CONSTRUCTOR (node->decl))
1332 node->set_init_priority (streamer_read_hwi (ib));
1333 if (DECL_STATIC_DESTRUCTOR (node->decl))
1334 node->set_fini_priority (streamer_read_hwi (ib));
1336 if (node->instrumentation_clone)
1338 decl_index = streamer_read_uhwi (ib);
1339 fn_decl = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1340 node->orig_decl = fn_decl;
1343 return node;
1346 /* Read a node from input_block IB. TAG is the node's tag just read.
1347 Return the node read or overwriten. */
1349 static varpool_node *
1350 input_varpool_node (struct lto_file_decl_data *file_data,
1351 struct lto_input_block *ib)
1353 int decl_index;
1354 tree var_decl;
1355 varpool_node *node;
1356 struct bitpack_d bp;
1357 int ref = LCC_NOT_FOUND;
1358 int order;
1359 tree group;
1360 const char *section;
1362 order = streamer_read_hwi (ib) + order_base;
1363 decl_index = streamer_read_uhwi (ib);
1364 var_decl = lto_file_decl_data_get_var_decl (file_data, decl_index);
1366 /* Declaration of functions can be already merged with a declaration
1367 from other input file. We keep cgraph unmerged until after streaming
1368 of ipa passes is done. Alays forcingly create a fresh node. */
1369 node = varpool_node::create_empty ();
1370 node->decl = var_decl;
1371 node->register_symbol ();
1373 node->order = order;
1374 if (order >= symtab->order)
1375 symtab->order = order + 1;
1376 node->lto_file_data = file_data;
1378 bp = streamer_read_bitpack (ib);
1379 node->externally_visible = bp_unpack_value (&bp, 1);
1380 node->no_reorder = bp_unpack_value (&bp, 1);
1381 node->force_output = bp_unpack_value (&bp, 1);
1382 node->forced_by_abi = bp_unpack_value (&bp, 1);
1383 node->unique_name = bp_unpack_value (&bp, 1);
1384 node->body_removed = bp_unpack_value (&bp, 1);
1385 node->implicit_section = bp_unpack_value (&bp, 1);
1386 node->writeonly = bp_unpack_value (&bp, 1);
1387 node->definition = bp_unpack_value (&bp, 1);
1388 node->alias = bp_unpack_value (&bp, 1);
1389 node->transparent_alias = bp_unpack_value (&bp, 1);
1390 node->weakref = bp_unpack_value (&bp, 1);
1391 node->analyzed = bp_unpack_value (&bp, 1);
1392 node->used_from_other_partition = bp_unpack_value (&bp, 1);
1393 node->in_other_partition = bp_unpack_value (&bp, 1);
1394 if (node->in_other_partition)
1396 DECL_EXTERNAL (node->decl) = 1;
1397 TREE_STATIC (node->decl) = 0;
1399 if (node->alias && !node->analyzed && node->weakref)
1400 node->alias_target = get_alias_symbol (node->decl);
1401 node->tls_model = (enum tls_model)bp_unpack_value (&bp, 3);
1402 node->used_by_single_function = (enum tls_model)bp_unpack_value (&bp, 1);
1403 node->need_bounds_init = bp_unpack_value (&bp, 1);
1404 group = read_identifier (ib);
1405 if (group)
1407 node->set_comdat_group (group);
1408 ref = streamer_read_hwi (ib);
1409 /* Store a reference for now, and fix up later to be a pointer. */
1410 node->same_comdat_group = (symtab_node *) (intptr_t) ref;
1412 else
1413 node->same_comdat_group = (symtab_node *) (intptr_t) LCC_NOT_FOUND;
1414 section = read_string (ib);
1415 if (section)
1416 node->set_section_for_node (section);
1417 node->resolution = streamer_read_enum (ib, ld_plugin_symbol_resolution,
1418 LDPR_NUM_KNOWN);
1419 gcc_assert (flag_ltrans
1420 || (!node->in_other_partition
1421 && !node->used_from_other_partition));
1423 return node;
1426 /* Read a node from input_block IB. TAG is the node's tag just read.
1427 Return the node read or overwriten. */
1429 static void
1430 input_ref (struct lto_input_block *ib,
1431 symtab_node *referring_node,
1432 vec<symtab_node *> nodes)
1434 symtab_node *node = NULL;
1435 struct bitpack_d bp;
1436 enum ipa_ref_use use;
1437 bool speculative;
1438 struct ipa_ref *ref;
1440 bp = streamer_read_bitpack (ib);
1441 use = (enum ipa_ref_use) bp_unpack_value (&bp, 3);
1442 speculative = (enum ipa_ref_use) bp_unpack_value (&bp, 1);
1443 node = nodes[streamer_read_hwi (ib)];
1444 ref = referring_node->create_reference (node, use);
1445 ref->speculative = speculative;
1446 if (is_a <cgraph_node *> (referring_node))
1447 ref->lto_stmt_uid = streamer_read_hwi (ib);
1450 /* Read an edge from IB. NODES points to a vector of previously read nodes for
1451 decoding caller and callee of the edge to be read. If INDIRECT is true, the
1452 edge being read is indirect (in the sense that it has
1453 indirect_unknown_callee set). */
1455 static void
1456 input_edge (struct lto_input_block *ib, vec<symtab_node *> nodes,
1457 bool indirect)
1459 struct cgraph_node *caller, *callee;
1460 struct cgraph_edge *edge;
1461 unsigned int stmt_id;
1462 gcov_type count;
1463 int freq;
1464 cgraph_inline_failed_t inline_failed;
1465 struct bitpack_d bp;
1466 int ecf_flags = 0;
1468 caller = dyn_cast<cgraph_node *> (nodes[streamer_read_hwi (ib)]);
1469 if (caller == NULL || caller->decl == NULL_TREE)
1470 internal_error ("bytecode stream: no caller found while reading edge");
1472 if (!indirect)
1474 callee = dyn_cast<cgraph_node *> (nodes[streamer_read_hwi (ib)]);
1475 if (callee == NULL || callee->decl == NULL_TREE)
1476 internal_error ("bytecode stream: no callee found while reading edge");
1478 else
1479 callee = NULL;
1481 count = streamer_read_gcov_count (ib);
1483 bp = streamer_read_bitpack (ib);
1484 inline_failed = bp_unpack_enum (&bp, cgraph_inline_failed_t, CIF_N_REASONS);
1485 stmt_id = bp_unpack_var_len_unsigned (&bp);
1486 freq = (int) bp_unpack_var_len_unsigned (&bp);
1488 if (indirect)
1489 edge = caller->create_indirect_edge (NULL, 0, count, freq);
1490 else
1491 edge = caller->create_edge (callee, NULL, count, freq);
1493 edge->indirect_inlining_edge = bp_unpack_value (&bp, 1);
1494 edge->speculative = bp_unpack_value (&bp, 1);
1495 edge->lto_stmt_uid = stmt_id;
1496 edge->inline_failed = inline_failed;
1497 edge->call_stmt_cannot_inline_p = bp_unpack_value (&bp, 1);
1498 edge->can_throw_external = bp_unpack_value (&bp, 1);
1499 edge->in_polymorphic_cdtor = bp_unpack_value (&bp, 1);
1500 if (indirect)
1502 if (bp_unpack_value (&bp, 1))
1503 ecf_flags |= ECF_CONST;
1504 if (bp_unpack_value (&bp, 1))
1505 ecf_flags |= ECF_PURE;
1506 if (bp_unpack_value (&bp, 1))
1507 ecf_flags |= ECF_NORETURN;
1508 if (bp_unpack_value (&bp, 1))
1509 ecf_flags |= ECF_MALLOC;
1510 if (bp_unpack_value (&bp, 1))
1511 ecf_flags |= ECF_NOTHROW;
1512 if (bp_unpack_value (&bp, 1))
1513 ecf_flags |= ECF_RETURNS_TWICE;
1514 edge->indirect_info->ecf_flags = ecf_flags;
1515 edge->indirect_info->common_target_id = streamer_read_hwi (ib);
1516 if (edge->indirect_info->common_target_id)
1517 edge->indirect_info->common_target_probability = streamer_read_hwi (ib);
1522 /* Read a cgraph from IB using the info in FILE_DATA. */
1524 static vec<symtab_node *>
1525 input_cgraph_1 (struct lto_file_decl_data *file_data,
1526 struct lto_input_block *ib)
1528 enum LTO_symtab_tags tag;
1529 vec<symtab_node *> nodes = vNULL;
1530 symtab_node *node;
1531 unsigned i;
1533 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1534 order_base = symtab->order;
1535 while (tag)
1537 if (tag == LTO_symtab_edge)
1538 input_edge (ib, nodes, false);
1539 else if (tag == LTO_symtab_indirect_edge)
1540 input_edge (ib, nodes, true);
1541 else if (tag == LTO_symtab_variable)
1543 node = input_varpool_node (file_data, ib);
1544 nodes.safe_push (node);
1545 lto_symtab_encoder_encode (file_data->symtab_node_encoder, node);
1547 else
1549 node = input_node (file_data, ib, tag, nodes);
1550 if (node == NULL || node->decl == NULL_TREE)
1551 internal_error ("bytecode stream: found empty cgraph node");
1552 nodes.safe_push (node);
1553 lto_symtab_encoder_encode (file_data->symtab_node_encoder, node);
1556 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1559 lto_input_toplevel_asms (file_data, order_base);
1561 /* AUX pointers should be all non-zero for function nodes read from the stream. */
1562 if (flag_checking)
1564 FOR_EACH_VEC_ELT (nodes, i, node)
1565 gcc_assert (node->aux || !is_a <cgraph_node *> (node));
1567 FOR_EACH_VEC_ELT (nodes, i, node)
1569 int ref;
1570 if (cgraph_node *cnode = dyn_cast <cgraph_node *> (node))
1572 ref = (int) (intptr_t) cnode->global.inlined_to;
1574 /* We share declaration of builtins, so we may read same node twice. */
1575 if (!node->aux)
1576 continue;
1577 node->aux = NULL;
1579 /* Fixup inlined_to from reference to pointer. */
1580 if (ref != LCC_NOT_FOUND)
1581 dyn_cast<cgraph_node *> (node)->global.inlined_to
1582 = dyn_cast<cgraph_node *> (nodes[ref]);
1583 else
1584 cnode->global.inlined_to = NULL;
1586 /* Compute instrumented_version. */
1587 if (cnode->instrumentation_clone)
1589 gcc_assert (cnode->orig_decl);
1591 cnode->instrumented_version = cgraph_node::get (cnode->orig_decl);
1592 if (cnode->instrumented_version)
1594 /* We may have multiple nodes for a single function which
1595 will be merged later. To have a proper merge we need
1596 to keep instrumentation_version reference between nodes
1597 consistent: each instrumented_version reference should
1598 have proper reverse reference. Thus don't break existing
1599 instrumented_version reference if it already exists. */
1600 if (cnode->instrumented_version->instrumented_version)
1601 cnode->instrumented_version = NULL;
1602 else
1603 cnode->instrumented_version->instrumented_version = cnode;
1606 /* Restore decl names reference except for wrapper functions. */
1607 if (!chkp_wrap_function (cnode->orig_decl))
1609 tree name = DECL_ASSEMBLER_NAME (cnode->decl);
1610 IDENTIFIER_TRANSPARENT_ALIAS (name) = 1;
1611 TREE_CHAIN (name) = DECL_ASSEMBLER_NAME (cnode->orig_decl);
1616 ref = (int) (intptr_t) node->same_comdat_group;
1618 /* Fixup same_comdat_group from reference to pointer. */
1619 if (ref != LCC_NOT_FOUND)
1620 node->same_comdat_group = nodes[ref];
1621 else
1622 node->same_comdat_group = NULL;
1624 FOR_EACH_VEC_ELT (nodes, i, node)
1625 node->aux = is_a <cgraph_node *> (node) ? (void *)1 : NULL;
1626 return nodes;
1629 /* Input ipa_refs. */
1631 static void
1632 input_refs (struct lto_input_block *ib,
1633 vec<symtab_node *> nodes)
1635 int count;
1636 int idx;
1637 while (true)
1639 symtab_node *node;
1640 count = streamer_read_uhwi (ib);
1641 if (!count)
1642 break;
1643 idx = streamer_read_uhwi (ib);
1644 node = nodes[idx];
1645 while (count)
1647 input_ref (ib, node, nodes);
1648 count--;
1654 static struct gcov_ctr_summary lto_gcov_summary;
1656 /* Input profile_info from IB. */
1657 static void
1658 input_profile_summary (struct lto_input_block *ib,
1659 struct lto_file_decl_data *file_data)
1661 unsigned h_ix;
1662 struct bitpack_d bp;
1663 unsigned int runs = streamer_read_uhwi (ib);
1664 if (runs)
1666 file_data->profile_info.runs = runs;
1667 file_data->profile_info.sum_max = streamer_read_gcov_count (ib);
1668 file_data->profile_info.sum_all = streamer_read_gcov_count (ib);
1670 memset (file_data->profile_info.histogram, 0,
1671 sizeof (gcov_bucket_type) * GCOV_HISTOGRAM_SIZE);
1672 /* Input the bitpack of non-zero histogram indices. */
1673 bp = streamer_read_bitpack (ib);
1674 /* Read in and unpack the full bitpack, flagging non-zero
1675 histogram entries by setting the num_counters non-zero. */
1676 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1678 file_data->profile_info.histogram[h_ix].num_counters
1679 = bp_unpack_value (&bp, 1);
1681 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1683 if (!file_data->profile_info.histogram[h_ix].num_counters)
1684 continue;
1686 file_data->profile_info.histogram[h_ix].num_counters
1687 = streamer_read_gcov_count (ib);
1688 file_data->profile_info.histogram[h_ix].min_value
1689 = streamer_read_gcov_count (ib);
1690 file_data->profile_info.histogram[h_ix].cum_value
1691 = streamer_read_gcov_count (ib);
1693 /* IPA-profile computes hot bb threshold based on cumulated
1694 whole program profile. We need to stream it down to ltrans. */
1695 if (flag_ltrans)
1696 set_hot_bb_threshold (streamer_read_gcov_count (ib));
1701 /* Rescale profile summaries to the same number of runs in the whole unit. */
1703 static void
1704 merge_profile_summaries (struct lto_file_decl_data **file_data_vec)
1706 struct lto_file_decl_data *file_data;
1707 unsigned int j, h_ix;
1708 gcov_unsigned_t max_runs = 0;
1709 struct cgraph_node *node;
1710 struct cgraph_edge *edge;
1711 gcov_type saved_sum_all = 0;
1712 gcov_ctr_summary *saved_profile_info = 0;
1713 int saved_scale = 0;
1715 /* Find unit with maximal number of runs. If we ever get serious about
1716 roundoff errors, we might also consider computing smallest common
1717 multiply. */
1718 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1719 if (max_runs < file_data->profile_info.runs)
1720 max_runs = file_data->profile_info.runs;
1722 if (!max_runs)
1723 return;
1725 /* Simple overflow check. We probably don't need to support that many train
1726 runs. Such a large value probably imply data corruption anyway. */
1727 if (max_runs > INT_MAX / REG_BR_PROB_BASE)
1729 sorry ("At most %i profile runs is supported. Perhaps corrupted profile?",
1730 INT_MAX / REG_BR_PROB_BASE);
1731 return;
1734 profile_info = &lto_gcov_summary;
1735 lto_gcov_summary.runs = max_runs;
1736 lto_gcov_summary.sum_max = 0;
1737 memset (lto_gcov_summary.histogram, 0,
1738 sizeof (gcov_bucket_type) * GCOV_HISTOGRAM_SIZE);
1740 /* Rescale all units to the maximal number of runs.
1741 sum_max can not be easily merged, as we have no idea what files come from
1742 the same run. We do not use the info anyway, so leave it 0. */
1743 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1744 if (file_data->profile_info.runs)
1746 int scale = GCOV_COMPUTE_SCALE (max_runs,
1747 file_data->profile_info.runs);
1748 lto_gcov_summary.sum_max
1749 = MAX (lto_gcov_summary.sum_max,
1750 apply_scale (file_data->profile_info.sum_max, scale));
1751 lto_gcov_summary.sum_all
1752 = MAX (lto_gcov_summary.sum_all,
1753 apply_scale (file_data->profile_info.sum_all, scale));
1754 /* Save a pointer to the profile_info with the largest
1755 scaled sum_all and the scale for use in merging the
1756 histogram. */
1757 if (!saved_profile_info
1758 || lto_gcov_summary.sum_all > saved_sum_all)
1760 saved_profile_info = &file_data->profile_info;
1761 saved_sum_all = lto_gcov_summary.sum_all;
1762 saved_scale = scale;
1766 gcc_assert (saved_profile_info);
1768 /* Scale up the histogram from the profile that had the largest
1769 scaled sum_all above. */
1770 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1772 /* Scale up the min value as we did the corresponding sum_all
1773 above. Use that to find the new histogram index. */
1774 gcov_type scaled_min
1775 = apply_scale (saved_profile_info->histogram[h_ix].min_value,
1776 saved_scale);
1777 /* The new index may be shared with another scaled histogram entry,
1778 so we need to account for a non-zero histogram entry at new_ix. */
1779 unsigned new_ix = gcov_histo_index (scaled_min);
1780 lto_gcov_summary.histogram[new_ix].min_value
1781 = (lto_gcov_summary.histogram[new_ix].num_counters
1782 ? MIN (lto_gcov_summary.histogram[new_ix].min_value, scaled_min)
1783 : scaled_min);
1784 /* Some of the scaled counter values would ostensibly need to be placed
1785 into different (larger) histogram buckets, but we keep things simple
1786 here and place the scaled cumulative counter value in the bucket
1787 corresponding to the scaled minimum counter value. */
1788 lto_gcov_summary.histogram[new_ix].cum_value
1789 += apply_scale (saved_profile_info->histogram[h_ix].cum_value,
1790 saved_scale);
1791 lto_gcov_summary.histogram[new_ix].num_counters
1792 += saved_profile_info->histogram[h_ix].num_counters;
1795 /* Watch roundoff errors. */
1796 if (lto_gcov_summary.sum_max < max_runs)
1797 lto_gcov_summary.sum_max = max_runs;
1799 /* If merging already happent at WPA time, we are done. */
1800 if (flag_ltrans)
1801 return;
1803 /* Now compute count_materialization_scale of each node.
1804 During LTRANS we already have values of count_materialization_scale
1805 computed, so just update them. */
1806 FOR_EACH_FUNCTION (node)
1807 if (node->lto_file_data
1808 && node->lto_file_data->profile_info.runs)
1810 int scale;
1812 scale = RDIV (node->count_materialization_scale * max_runs,
1813 node->lto_file_data->profile_info.runs);
1814 node->count_materialization_scale = scale;
1815 if (scale < 0)
1816 fatal_error (input_location, "Profile information in %s corrupted",
1817 file_data->file_name);
1819 if (scale == REG_BR_PROB_BASE)
1820 continue;
1821 for (edge = node->callees; edge; edge = edge->next_callee)
1822 edge->count = apply_scale (edge->count, scale);
1823 node->count = apply_scale (node->count, scale);
1827 /* Input and merge the symtab from each of the .o files passed to
1828 lto1. */
1830 void
1831 input_symtab (void)
1833 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1834 struct lto_file_decl_data *file_data;
1835 unsigned int j = 0;
1836 struct cgraph_node *node;
1838 while ((file_data = file_data_vec[j++]))
1840 const char *data;
1841 size_t len;
1842 struct lto_input_block *ib;
1843 vec<symtab_node *> nodes;
1845 ib = lto_create_simple_input_block (file_data, LTO_section_symtab_nodes,
1846 &data, &len);
1847 if (!ib)
1848 fatal_error (input_location,
1849 "cannot find LTO cgraph in %s", file_data->file_name);
1850 input_profile_summary (ib, file_data);
1851 file_data->symtab_node_encoder = lto_symtab_encoder_new (true);
1852 nodes = input_cgraph_1 (file_data, ib);
1853 lto_destroy_simple_input_block (file_data, LTO_section_symtab_nodes,
1854 ib, data, len);
1856 ib = lto_create_simple_input_block (file_data, LTO_section_refs,
1857 &data, &len);
1858 if (!ib)
1859 fatal_error (input_location, "cannot find LTO section refs in %s",
1860 file_data->file_name);
1861 input_refs (ib, nodes);
1862 lto_destroy_simple_input_block (file_data, LTO_section_refs,
1863 ib, data, len);
1864 if (flag_ltrans)
1865 input_cgraph_opt_summary (nodes);
1866 nodes.release ();
1869 merge_profile_summaries (file_data_vec);
1871 if (!flag_auto_profile)
1872 get_working_sets ();
1875 /* Clear out the aux field that was used to store enough state to
1876 tell which nodes should be overwritten. */
1877 FOR_EACH_FUNCTION (node)
1879 /* Some nodes may have been created by cgraph_node. This
1880 happens when the callgraph contains nested functions. If the
1881 node for the parent function was never emitted to the gimple
1882 file, cgraph_node will create a node for it when setting the
1883 context of the nested function. */
1884 if (node->lto_file_data)
1885 node->aux = NULL;
1889 /* Input function/variable tables that will allow libgomp to look up offload
1890 target code, and store them into OFFLOAD_FUNCS and OFFLOAD_VARS. */
1892 void
1893 input_offload_tables (bool do_force_output)
1895 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1896 struct lto_file_decl_data *file_data;
1897 unsigned int j = 0;
1899 while ((file_data = file_data_vec[j++]))
1901 const char *data;
1902 size_t len;
1903 struct lto_input_block *ib
1904 = lto_create_simple_input_block (file_data, LTO_section_offload_table,
1905 &data, &len);
1906 if (!ib)
1907 continue;
1909 enum LTO_symtab_tags tag
1910 = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1911 while (tag)
1913 if (tag == LTO_symtab_unavail_node)
1915 int decl_index = streamer_read_uhwi (ib);
1916 tree fn_decl
1917 = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1918 vec_safe_push (offload_funcs, fn_decl);
1920 /* Prevent IPA from removing fn_decl as unreachable, since there
1921 may be no refs from the parent function to child_fn in offload
1922 LTO mode. */
1923 if (do_force_output)
1924 cgraph_node::get (fn_decl)->mark_force_output ();
1926 else if (tag == LTO_symtab_variable)
1928 int decl_index = streamer_read_uhwi (ib);
1929 tree var_decl
1930 = lto_file_decl_data_get_var_decl (file_data, decl_index);
1931 vec_safe_push (offload_vars, var_decl);
1933 /* Prevent IPA from removing var_decl as unused, since there
1934 may be no refs to var_decl in offload LTO mode. */
1935 if (do_force_output)
1936 varpool_node::get (var_decl)->force_output = 1;
1938 else
1939 fatal_error (input_location,
1940 "invalid offload table in %s", file_data->file_name);
1942 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1945 lto_destroy_simple_input_block (file_data, LTO_section_offload_table,
1946 ib, data, len);
1950 /* True when we need optimization summary for NODE. */
1952 static int
1953 output_cgraph_opt_summary_p (struct cgraph_node *node)
1955 return (node->clone_of
1956 && (node->clone.tree_map
1957 || node->clone.args_to_skip
1958 || node->clone.combined_args_to_skip));
1961 /* Output optimization summary for EDGE to OB. */
1962 static void
1963 output_edge_opt_summary (struct output_block *ob ATTRIBUTE_UNUSED,
1964 struct cgraph_edge *edge ATTRIBUTE_UNUSED)
1968 /* Output optimization summary for NODE to OB. */
1970 static void
1971 output_node_opt_summary (struct output_block *ob,
1972 struct cgraph_node *node,
1973 lto_symtab_encoder_t encoder)
1975 unsigned int index;
1976 bitmap_iterator bi;
1977 struct ipa_replace_map *map;
1978 struct bitpack_d bp;
1979 int i;
1980 struct cgraph_edge *e;
1982 if (node->clone.args_to_skip)
1984 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.args_to_skip));
1985 EXECUTE_IF_SET_IN_BITMAP (node->clone.args_to_skip, 0, index, bi)
1986 streamer_write_uhwi (ob, index);
1988 else
1989 streamer_write_uhwi (ob, 0);
1990 if (node->clone.combined_args_to_skip)
1992 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.combined_args_to_skip));
1993 EXECUTE_IF_SET_IN_BITMAP (node->clone.combined_args_to_skip, 0, index, bi)
1994 streamer_write_uhwi (ob, index);
1996 else
1997 streamer_write_uhwi (ob, 0);
1998 streamer_write_uhwi (ob, vec_safe_length (node->clone.tree_map));
1999 FOR_EACH_VEC_SAFE_ELT (node->clone.tree_map, i, map)
2001 /* At the moment we assume all old trees to be PARM_DECLs, because we have no
2002 mechanism to store function local declarations into summaries. */
2003 gcc_assert (!map->old_tree);
2004 streamer_write_uhwi (ob, map->parm_num);
2005 gcc_assert (EXPR_LOCATION (map->new_tree) == UNKNOWN_LOCATION);
2006 stream_write_tree (ob, map->new_tree, true);
2007 bp = bitpack_create (ob->main_stream);
2008 bp_pack_value (&bp, map->replace_p, 1);
2009 bp_pack_value (&bp, map->ref_p, 1);
2010 streamer_write_bitpack (&bp);
2013 if (lto_symtab_encoder_in_partition_p (encoder, node))
2015 for (e = node->callees; e; e = e->next_callee)
2016 output_edge_opt_summary (ob, e);
2017 for (e = node->indirect_calls; e; e = e->next_callee)
2018 output_edge_opt_summary (ob, e);
2022 /* Output optimization summaries stored in callgraph.
2023 At the moment it is the clone info structure. */
2025 static void
2026 output_cgraph_opt_summary (void)
2028 int i, n_nodes;
2029 lto_symtab_encoder_t encoder;
2030 struct output_block *ob = create_output_block (LTO_section_cgraph_opt_sum);
2031 unsigned count = 0;
2033 ob->symbol = NULL;
2034 encoder = ob->decl_state->symtab_node_encoder;
2035 n_nodes = lto_symtab_encoder_size (encoder);
2036 for (i = 0; i < n_nodes; i++)
2038 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
2039 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
2040 if (cnode && output_cgraph_opt_summary_p (cnode))
2041 count++;
2043 streamer_write_uhwi (ob, count);
2044 for (i = 0; i < n_nodes; i++)
2046 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
2047 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
2048 if (cnode && output_cgraph_opt_summary_p (cnode))
2050 streamer_write_uhwi (ob, i);
2051 output_node_opt_summary (ob, cnode, encoder);
2054 produce_asm (ob, NULL);
2055 destroy_output_block (ob);
2058 /* Input optimisation summary of EDGE. */
2060 static void
2061 input_edge_opt_summary (struct cgraph_edge *edge ATTRIBUTE_UNUSED,
2062 struct lto_input_block *ib_main ATTRIBUTE_UNUSED)
2066 /* Input optimisation summary of NODE. */
2068 static void
2069 input_node_opt_summary (struct cgraph_node *node,
2070 struct lto_input_block *ib_main,
2071 struct data_in *data_in)
2073 int i;
2074 int count;
2075 int bit;
2076 struct bitpack_d bp;
2077 struct cgraph_edge *e;
2079 count = streamer_read_uhwi (ib_main);
2080 if (count)
2081 node->clone.args_to_skip = BITMAP_GGC_ALLOC ();
2082 for (i = 0; i < count; i++)
2084 bit = streamer_read_uhwi (ib_main);
2085 bitmap_set_bit (node->clone.args_to_skip, bit);
2087 count = streamer_read_uhwi (ib_main);
2088 if (count)
2089 node->clone.combined_args_to_skip = BITMAP_GGC_ALLOC ();
2090 for (i = 0; i < count; i++)
2092 bit = streamer_read_uhwi (ib_main);
2093 bitmap_set_bit (node->clone.combined_args_to_skip, bit);
2095 count = streamer_read_uhwi (ib_main);
2096 for (i = 0; i < count; i++)
2098 struct ipa_replace_map *map = ggc_alloc<ipa_replace_map> ();
2100 vec_safe_push (node->clone.tree_map, map);
2101 map->parm_num = streamer_read_uhwi (ib_main);
2102 map->old_tree = NULL;
2103 map->new_tree = stream_read_tree (ib_main, data_in);
2104 bp = streamer_read_bitpack (ib_main);
2105 map->replace_p = bp_unpack_value (&bp, 1);
2106 map->ref_p = bp_unpack_value (&bp, 1);
2108 for (e = node->callees; e; e = e->next_callee)
2109 input_edge_opt_summary (e, ib_main);
2110 for (e = node->indirect_calls; e; e = e->next_callee)
2111 input_edge_opt_summary (e, ib_main);
2114 /* Read section in file FILE_DATA of length LEN with data DATA. */
2116 static void
2117 input_cgraph_opt_section (struct lto_file_decl_data *file_data,
2118 const char *data, size_t len,
2119 vec<symtab_node *> nodes)
2121 const struct lto_function_header *header =
2122 (const struct lto_function_header *) data;
2123 const int cfg_offset = sizeof (struct lto_function_header);
2124 const int main_offset = cfg_offset + header->cfg_size;
2125 const int string_offset = main_offset + header->main_size;
2126 struct data_in *data_in;
2127 unsigned int i;
2128 unsigned int count;
2130 lto_input_block ib_main ((const char *) data + main_offset,
2131 header->main_size, file_data->mode_table);
2133 data_in =
2134 lto_data_in_create (file_data, (const char *) data + string_offset,
2135 header->string_size, vNULL);
2136 count = streamer_read_uhwi (&ib_main);
2138 for (i = 0; i < count; i++)
2140 int ref = streamer_read_uhwi (&ib_main);
2141 input_node_opt_summary (dyn_cast<cgraph_node *> (nodes[ref]),
2142 &ib_main, data_in);
2144 lto_free_section_data (file_data, LTO_section_cgraph_opt_sum, NULL, data,
2145 len);
2146 lto_data_in_delete (data_in);
2149 /* Input optimization summary of cgraph. */
2151 static void
2152 input_cgraph_opt_summary (vec<symtab_node *> nodes)
2154 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
2155 struct lto_file_decl_data *file_data;
2156 unsigned int j = 0;
2158 while ((file_data = file_data_vec[j++]))
2160 size_t len;
2161 const char *data =
2162 lto_get_section_data (file_data, LTO_section_cgraph_opt_sum, NULL,
2163 &len);
2165 if (data)
2166 input_cgraph_opt_section (file_data, data, len, nodes);