[NDS32] Implement fp-as-gp optimization.
[official-gcc.git] / gcc / lto-cgraph.c
blobb23d1890fc0cea8d844df7e20ceb728e167a6842
1 /* Write and read the cgraph to the memory mapped representation of a
2 .o file.
4 Copyright (C) 2009-2018 Free Software Foundation, Inc.
5 Contributed by Kenneth Zadeck <zadeck@naturalbridge.com>
7 This file is part of GCC.
9 GCC is free software; you can redistribute it and/or modify it under
10 the terms of the GNU General Public License as published by the Free
11 Software Foundation; either version 3, or (at your option) any later
12 version.
14 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
15 WARRANTY; without even the implied warranty of MERCHANTABILITY or
16 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 for more details.
19 You should have received a copy of the GNU General Public License
20 along with GCC; see the file COPYING3. If not see
21 <http://www.gnu.org/licenses/>. */
23 #include "config.h"
24 #include "system.h"
25 #include "coretypes.h"
26 #include "backend.h"
27 #include "rtl.h"
28 #include "tree.h"
29 #include "gimple.h"
30 #include "predict.h"
31 #include "stringpool.h"
32 #include "tree-streamer.h"
33 #include "cgraph.h"
34 #include "tree-pass.h"
35 #include "profile.h"
36 #include "context.h"
37 #include "pass_manager.h"
38 #include "ipa-utils.h"
39 #include "omp-offload.h"
40 #include "ipa-chkp.h"
41 #include "stringpool.h"
42 #include "attribs.h"
44 /* True when asm nodes has been output. */
45 bool asm_nodes_output = false;
47 static void output_cgraph_opt_summary (void);
48 static void input_cgraph_opt_summary (vec<symtab_node *> nodes);
50 /* Number of LDPR values known to GCC. */
51 #define LDPR_NUM_KNOWN (LDPR_PREVAILING_DEF_IRONLY_EXP + 1)
53 /* All node orders are ofsetted by ORDER_BASE. */
54 static int order_base;
56 /* Cgraph streaming is organized as set of record whose type
57 is indicated by a tag. */
58 enum LTO_symtab_tags
60 /* Must leave 0 for the stopper. */
62 /* Cgraph node without body available. */
63 LTO_symtab_unavail_node = 1,
64 /* Cgraph node with function body. */
65 LTO_symtab_analyzed_node,
66 /* Cgraph edges. */
67 LTO_symtab_edge,
68 LTO_symtab_indirect_edge,
69 LTO_symtab_variable,
70 LTO_symtab_last_tag
73 /* Create a new symtab encoder.
74 if FOR_INPUT, the encoder allocate only datastructures needed
75 to read the symtab. */
77 lto_symtab_encoder_t
78 lto_symtab_encoder_new (bool for_input)
80 lto_symtab_encoder_t encoder = XCNEW (struct lto_symtab_encoder_d);
82 if (!for_input)
83 encoder->map = new hash_map<symtab_node *, size_t>;
84 encoder->nodes.create (0);
85 return encoder;
89 /* Delete ENCODER and its components. */
91 void
92 lto_symtab_encoder_delete (lto_symtab_encoder_t encoder)
94 encoder->nodes.release ();
95 if (encoder->map)
96 delete encoder->map;
97 free (encoder);
101 /* Return the existing reference number of NODE in the symtab encoder in
102 output block OB. Assign a new reference if this is the first time
103 NODE is encoded. */
106 lto_symtab_encoder_encode (lto_symtab_encoder_t encoder,
107 symtab_node *node)
109 int ref;
111 if (!encoder->map)
113 lto_encoder_entry entry = {node, false, false, false};
115 ref = encoder->nodes.length ();
116 encoder->nodes.safe_push (entry);
117 return ref;
120 size_t *slot = encoder->map->get (node);
121 if (!slot || !*slot)
123 lto_encoder_entry entry = {node, false, false, false};
124 ref = encoder->nodes.length ();
125 if (!slot)
126 encoder->map->put (node, ref + 1);
127 encoder->nodes.safe_push (entry);
129 else
130 ref = *slot - 1;
132 return ref;
135 /* Remove NODE from encoder. */
137 bool
138 lto_symtab_encoder_delete_node (lto_symtab_encoder_t encoder,
139 symtab_node *node)
141 int index;
142 lto_encoder_entry last_node;
144 size_t *slot = encoder->map->get (node);
145 if (slot == NULL || !*slot)
146 return false;
148 index = *slot - 1;
149 gcc_checking_assert (encoder->nodes[index].node == node);
151 /* Remove from vector. We do this by swapping node with the last element
152 of the vector. */
153 last_node = encoder->nodes.pop ();
154 if (last_node.node != node)
156 gcc_assert (encoder->map->put (last_node.node, index + 1));
158 /* Move the last element to the original spot of NODE. */
159 encoder->nodes[index] = last_node;
162 /* Remove element from hash table. */
163 encoder->map->remove (node);
164 return true;
168 /* Return TRUE if we should encode the body of NODE (if any). */
170 bool
171 lto_symtab_encoder_encode_body_p (lto_symtab_encoder_t encoder,
172 struct cgraph_node *node)
174 int index = lto_symtab_encoder_lookup (encoder, node);
175 return encoder->nodes[index].body;
178 /* Specify that we encode the body of NODE in this partition. */
180 static void
181 lto_set_symtab_encoder_encode_body (lto_symtab_encoder_t encoder,
182 struct cgraph_node *node)
184 int index = lto_symtab_encoder_encode (encoder, node);
185 gcc_checking_assert (encoder->nodes[index].node == node);
186 encoder->nodes[index].body = true;
189 /* Return TRUE if we should encode initializer of NODE (if any). */
191 bool
192 lto_symtab_encoder_encode_initializer_p (lto_symtab_encoder_t encoder,
193 varpool_node *node)
195 int index = lto_symtab_encoder_lookup (encoder, node);
196 if (index == LCC_NOT_FOUND)
197 return false;
198 return encoder->nodes[index].initializer;
201 /* Specify that we should encode initializer of NODE (if any). */
203 static void
204 lto_set_symtab_encoder_encode_initializer (lto_symtab_encoder_t encoder,
205 varpool_node *node)
207 int index = lto_symtab_encoder_lookup (encoder, node);
208 encoder->nodes[index].initializer = true;
211 /* Return TRUE if NODE is in this partition. */
213 bool
214 lto_symtab_encoder_in_partition_p (lto_symtab_encoder_t encoder,
215 symtab_node *node)
217 int index = lto_symtab_encoder_lookup (encoder, node);
218 if (index == LCC_NOT_FOUND)
219 return false;
220 return encoder->nodes[index].in_partition;
223 /* Specify that NODE is in this partition. */
225 void
226 lto_set_symtab_encoder_in_partition (lto_symtab_encoder_t encoder,
227 symtab_node *node)
229 int index = lto_symtab_encoder_encode (encoder, node);
230 encoder->nodes[index].in_partition = true;
233 /* Output the cgraph EDGE to OB using ENCODER. */
235 static void
236 lto_output_edge (struct lto_simple_output_block *ob, struct cgraph_edge *edge,
237 lto_symtab_encoder_t encoder)
239 unsigned int uid;
240 intptr_t ref;
241 struct bitpack_d bp;
243 if (edge->indirect_unknown_callee)
244 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
245 LTO_symtab_indirect_edge);
246 else
247 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
248 LTO_symtab_edge);
250 ref = lto_symtab_encoder_lookup (encoder, edge->caller);
251 gcc_assert (ref != LCC_NOT_FOUND);
252 streamer_write_hwi_stream (ob->main_stream, ref);
254 if (!edge->indirect_unknown_callee)
256 ref = lto_symtab_encoder_lookup (encoder, edge->callee);
257 gcc_assert (ref != LCC_NOT_FOUND);
258 streamer_write_hwi_stream (ob->main_stream, ref);
261 edge->count.stream_out (ob->main_stream);
263 bp = bitpack_create (ob->main_stream);
264 uid = (!gimple_has_body_p (edge->caller->decl) || edge->caller->thunk.thunk_p
265 ? edge->lto_stmt_uid : gimple_uid (edge->call_stmt) + 1);
266 bp_pack_enum (&bp, cgraph_inline_failed_t,
267 CIF_N_REASONS, edge->inline_failed);
268 bp_pack_var_len_unsigned (&bp, uid);
269 bp_pack_value (&bp, edge->indirect_inlining_edge, 1);
270 bp_pack_value (&bp, edge->speculative, 1);
271 bp_pack_value (&bp, edge->call_stmt_cannot_inline_p, 1);
272 gcc_assert (!edge->call_stmt_cannot_inline_p
273 || edge->inline_failed != CIF_BODY_NOT_AVAILABLE);
274 bp_pack_value (&bp, edge->can_throw_external, 1);
275 bp_pack_value (&bp, edge->in_polymorphic_cdtor, 1);
276 if (edge->indirect_unknown_callee)
278 int flags = edge->indirect_info->ecf_flags;
279 bp_pack_value (&bp, (flags & ECF_CONST) != 0, 1);
280 bp_pack_value (&bp, (flags & ECF_PURE) != 0, 1);
281 bp_pack_value (&bp, (flags & ECF_NORETURN) != 0, 1);
282 bp_pack_value (&bp, (flags & ECF_MALLOC) != 0, 1);
283 bp_pack_value (&bp, (flags & ECF_NOTHROW) != 0, 1);
284 bp_pack_value (&bp, (flags & ECF_RETURNS_TWICE) != 0, 1);
285 /* Flags that should not appear on indirect calls. */
286 gcc_assert (!(flags & (ECF_LOOPING_CONST_OR_PURE
287 | ECF_MAY_BE_ALLOCA
288 | ECF_SIBCALL
289 | ECF_LEAF
290 | ECF_NOVOPS)));
292 streamer_write_bitpack (&bp);
293 if (edge->indirect_unknown_callee)
295 streamer_write_hwi_stream (ob->main_stream,
296 edge->indirect_info->common_target_id);
297 if (edge->indirect_info->common_target_id)
298 streamer_write_hwi_stream
299 (ob->main_stream, edge->indirect_info->common_target_probability);
303 /* Return if NODE contain references from other partitions. */
305 bool
306 referenced_from_other_partition_p (symtab_node *node, lto_symtab_encoder_t encoder)
308 int i;
309 struct ipa_ref *ref = NULL;
311 for (i = 0; node->iterate_referring (i, ref); i++)
313 /* Ignore references from non-offloadable nodes while streaming NODE into
314 offload LTO section. */
315 if (!ref->referring->need_lto_streaming)
316 continue;
318 if (ref->referring->in_other_partition
319 || !lto_symtab_encoder_in_partition_p (encoder, ref->referring))
320 return true;
322 return false;
325 /* Return true when node is reachable from other partition. */
327 bool
328 reachable_from_other_partition_p (struct cgraph_node *node, lto_symtab_encoder_t encoder)
330 struct cgraph_edge *e;
331 if (!node->definition)
332 return false;
333 if (node->global.inlined_to)
334 return false;
335 for (e = node->callers; e; e = e->next_caller)
337 /* Ignore references from non-offloadable nodes while streaming NODE into
338 offload LTO section. */
339 if (!e->caller->need_lto_streaming)
340 continue;
342 if (e->caller->in_other_partition
343 || !lto_symtab_encoder_in_partition_p (encoder, e->caller))
344 return true;
346 return false;
349 /* Return if NODE contain references from other partitions. */
351 bool
352 referenced_from_this_partition_p (symtab_node *node,
353 lto_symtab_encoder_t encoder)
355 int i;
356 struct ipa_ref *ref = NULL;
358 for (i = 0; node->iterate_referring (i, ref); i++)
359 if (lto_symtab_encoder_in_partition_p (encoder, ref->referring))
360 return true;
361 return false;
364 /* Return true when node is reachable from other partition. */
366 bool
367 reachable_from_this_partition_p (struct cgraph_node *node, lto_symtab_encoder_t encoder)
369 struct cgraph_edge *e;
370 for (e = node->callers; e; e = e->next_caller)
371 if (lto_symtab_encoder_in_partition_p (encoder, e->caller))
372 return true;
373 return false;
376 /* Output the cgraph NODE to OB. ENCODER is used to find the
377 reference number of NODE->inlined_to. SET is the set of nodes we
378 are writing to the current file. If NODE is not in SET, then NODE
379 is a boundary of a cgraph_node_set and we pretend NODE just has a
380 decl and no callees. WRITTEN_DECLS is the set of FUNCTION_DECLs
381 that have had their callgraph node written so far. This is used to
382 determine if NODE is a clone of a previously written node. */
384 static void
385 lto_output_node (struct lto_simple_output_block *ob, struct cgraph_node *node,
386 lto_symtab_encoder_t encoder)
388 unsigned int tag;
389 struct bitpack_d bp;
390 bool boundary_p;
391 intptr_t ref;
392 bool in_other_partition = false;
393 struct cgraph_node *clone_of, *ultimate_clone_of;
394 ipa_opt_pass_d *pass;
395 int i;
396 const char *comdat;
397 const char *section;
398 tree group;
400 boundary_p = !lto_symtab_encoder_in_partition_p (encoder, node);
402 if (node->analyzed && (!boundary_p || node->alias
403 || (node->thunk.thunk_p && !node->global.inlined_to)))
404 tag = LTO_symtab_analyzed_node;
405 else
406 tag = LTO_symtab_unavail_node;
408 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
409 tag);
410 streamer_write_hwi_stream (ob->main_stream, node->order);
412 /* In WPA mode, we only output part of the call-graph. Also, we
413 fake cgraph node attributes. There are two cases that we care.
415 Boundary nodes: There are nodes that are not part of SET but are
416 called from within SET. We artificially make them look like
417 externally visible nodes with no function body.
419 Cherry-picked nodes: These are nodes we pulled from other
420 translation units into SET during IPA-inlining. We make them as
421 local static nodes to prevent clashes with other local statics. */
422 if (boundary_p && node->analyzed
423 && node->get_partitioning_class () == SYMBOL_PARTITION)
425 /* Inline clones can not be part of boundary.
426 gcc_assert (!node->global.inlined_to);
428 FIXME: At the moment they can be, when partition contains an inline
429 clone that is clone of inline clone from outside partition. We can
430 reshape the clone tree and make other tree to be the root, but it
431 needs a bit extra work and will be promplty done by cgraph_remove_node
432 after reading back. */
433 in_other_partition = 1;
436 clone_of = node->clone_of;
437 while (clone_of
438 && (ref = lto_symtab_encoder_lookup (encoder, clone_of)) == LCC_NOT_FOUND)
439 if (clone_of->prev_sibling_clone)
440 clone_of = clone_of->prev_sibling_clone;
441 else
442 clone_of = clone_of->clone_of;
444 /* See if body of the master function is output. If not, we are seeing only
445 an declaration and we do not need to pass down clone tree. */
446 ultimate_clone_of = clone_of;
447 while (ultimate_clone_of && ultimate_clone_of->clone_of)
448 ultimate_clone_of = ultimate_clone_of->clone_of;
450 if (clone_of && !lto_symtab_encoder_encode_body_p (encoder, ultimate_clone_of))
451 clone_of = NULL;
453 if (tag == LTO_symtab_analyzed_node)
454 gcc_assert (clone_of || !node->clone_of);
455 if (!clone_of)
456 streamer_write_hwi_stream (ob->main_stream, LCC_NOT_FOUND);
457 else
458 streamer_write_hwi_stream (ob->main_stream, ref);
461 lto_output_fn_decl_index (ob->decl_state, ob->main_stream, node->decl);
462 node->count.stream_out (ob->main_stream);
463 streamer_write_hwi_stream (ob->main_stream, node->count_materialization_scale);
465 streamer_write_hwi_stream (ob->main_stream,
466 node->ipa_transforms_to_apply.length ());
467 FOR_EACH_VEC_ELT (node->ipa_transforms_to_apply, i, pass)
468 streamer_write_hwi_stream (ob->main_stream, pass->static_pass_number);
470 if (tag == LTO_symtab_analyzed_node)
472 if (node->global.inlined_to)
474 ref = lto_symtab_encoder_lookup (encoder, node->global.inlined_to);
475 gcc_assert (ref != LCC_NOT_FOUND);
477 else
478 ref = LCC_NOT_FOUND;
480 streamer_write_hwi_stream (ob->main_stream, ref);
483 group = node->get_comdat_group ();
484 if (group)
485 comdat = IDENTIFIER_POINTER (group);
486 else
487 comdat = "";
488 streamer_write_data_stream (ob->main_stream, comdat, strlen (comdat) + 1);
490 if (group)
492 if (node->same_comdat_group)
494 ref = LCC_NOT_FOUND;
495 for (struct symtab_node *n = node->same_comdat_group;
496 ref == LCC_NOT_FOUND && n != node; n = n->same_comdat_group)
497 ref = lto_symtab_encoder_lookup (encoder, n);
499 else
500 ref = LCC_NOT_FOUND;
501 streamer_write_hwi_stream (ob->main_stream, ref);
504 section = node->get_section ();
505 if (!section)
506 section = "";
508 streamer_write_hwi_stream (ob->main_stream, node->tp_first_run);
510 bp = bitpack_create (ob->main_stream);
511 bp_pack_value (&bp, node->local.local, 1);
512 bp_pack_value (&bp, node->externally_visible, 1);
513 bp_pack_value (&bp, node->no_reorder, 1);
514 bp_pack_value (&bp, node->definition, 1);
515 bp_pack_value (&bp, node->local.versionable, 1);
516 bp_pack_value (&bp, node->local.can_change_signature, 1);
517 bp_pack_value (&bp, node->local.redefined_extern_inline, 1);
518 bp_pack_value (&bp, node->force_output, 1);
519 bp_pack_value (&bp, node->forced_by_abi, 1);
520 bp_pack_value (&bp, node->unique_name, 1);
521 bp_pack_value (&bp, node->body_removed, 1);
522 bp_pack_value (&bp, node->implicit_section, 1);
523 bp_pack_value (&bp, node->address_taken, 1);
524 bp_pack_value (&bp, tag == LTO_symtab_analyzed_node
525 && node->get_partitioning_class () == SYMBOL_PARTITION
526 && (reachable_from_other_partition_p (node, encoder)
527 || referenced_from_other_partition_p (node, encoder)), 1);
528 bp_pack_value (&bp, node->lowered, 1);
529 bp_pack_value (&bp, in_other_partition, 1);
530 bp_pack_value (&bp, node->alias, 1);
531 bp_pack_value (&bp, node->transparent_alias, 1);
532 bp_pack_value (&bp, node->weakref, 1);
533 bp_pack_value (&bp, node->frequency, 2);
534 bp_pack_value (&bp, node->only_called_at_startup, 1);
535 bp_pack_value (&bp, node->only_called_at_exit, 1);
536 bp_pack_value (&bp, node->tm_clone, 1);
537 bp_pack_value (&bp, node->calls_comdat_local, 1);
538 bp_pack_value (&bp, node->icf_merged, 1);
539 bp_pack_value (&bp, node->nonfreeing_fn, 1);
540 bp_pack_value (&bp, node->thunk.thunk_p, 1);
541 bp_pack_value (&bp, node->parallelized_function, 1);
542 bp_pack_enum (&bp, ld_plugin_symbol_resolution,
543 LDPR_NUM_KNOWN,
544 /* When doing incremental link, we will get new resolution
545 info next time we process the file. */
546 flag_incremental_link ? LDPR_UNKNOWN : node->resolution);
547 bp_pack_value (&bp, node->instrumentation_clone, 1);
548 bp_pack_value (&bp, node->split_part, 1);
549 streamer_write_bitpack (&bp);
550 streamer_write_data_stream (ob->main_stream, section, strlen (section) + 1);
552 if (node->thunk.thunk_p)
554 streamer_write_uhwi_stream
555 (ob->main_stream,
556 1 + (node->thunk.this_adjusting != 0) * 2
557 + (node->thunk.virtual_offset_p != 0) * 4
558 + (node->thunk.add_pointer_bounds_args != 0) * 8);
559 streamer_write_uhwi_stream (ob->main_stream, node->thunk.fixed_offset);
560 streamer_write_uhwi_stream (ob->main_stream, node->thunk.virtual_value);
562 streamer_write_hwi_stream (ob->main_stream, node->profile_id);
563 if (DECL_STATIC_CONSTRUCTOR (node->decl))
564 streamer_write_hwi_stream (ob->main_stream, node->get_init_priority ());
565 if (DECL_STATIC_DESTRUCTOR (node->decl))
566 streamer_write_hwi_stream (ob->main_stream, node->get_fini_priority ());
568 if (node->instrumentation_clone)
569 lto_output_fn_decl_index (ob->decl_state, ob->main_stream, node->orig_decl);
572 /* Output the varpool NODE to OB.
573 If NODE is not in SET, then NODE is a boundary. */
575 static void
576 lto_output_varpool_node (struct lto_simple_output_block *ob, varpool_node *node,
577 lto_symtab_encoder_t encoder)
579 bool boundary_p = !lto_symtab_encoder_in_partition_p (encoder, node);
580 bool encode_initializer_p
581 = (node->definition
582 && lto_symtab_encoder_encode_initializer_p (encoder, node));
583 struct bitpack_d bp;
584 int ref;
585 const char *comdat;
586 const char *section;
587 tree group;
589 gcc_assert (!encode_initializer_p || node->definition);
590 gcc_assert (boundary_p || encode_initializer_p);
592 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
593 LTO_symtab_variable);
594 streamer_write_hwi_stream (ob->main_stream, node->order);
595 lto_output_var_decl_index (ob->decl_state, ob->main_stream, node->decl);
596 bp = bitpack_create (ob->main_stream);
597 bp_pack_value (&bp, node->externally_visible, 1);
598 bp_pack_value (&bp, node->no_reorder, 1);
599 bp_pack_value (&bp, node->force_output, 1);
600 bp_pack_value (&bp, node->forced_by_abi, 1);
601 bp_pack_value (&bp, node->unique_name, 1);
602 bp_pack_value (&bp,
603 node->body_removed
604 || (!encode_initializer_p && !node->alias && node->definition),
606 bp_pack_value (&bp, node->implicit_section, 1);
607 bp_pack_value (&bp, node->writeonly, 1);
608 bp_pack_value (&bp, node->definition && (encode_initializer_p || node->alias),
610 bp_pack_value (&bp, node->alias, 1);
611 bp_pack_value (&bp, node->transparent_alias, 1);
612 bp_pack_value (&bp, node->weakref, 1);
613 bp_pack_value (&bp, node->analyzed && (!boundary_p || node->alias), 1);
614 gcc_assert (node->definition || !node->analyzed);
615 /* Constant pool initializers can be de-unified into individual ltrans units.
616 FIXME: Alternatively at -Os we may want to avoid generating for them the local
617 labels and share them across LTRANS partitions. */
618 if (node->get_partitioning_class () != SYMBOL_PARTITION)
620 bp_pack_value (&bp, 0, 1); /* used_from_other_parition. */
621 bp_pack_value (&bp, 0, 1); /* in_other_partition. */
623 else
625 bp_pack_value (&bp, node->definition
626 && referenced_from_other_partition_p (node, encoder), 1);
627 bp_pack_value (&bp, node->analyzed
628 && boundary_p && !DECL_EXTERNAL (node->decl), 1);
629 /* in_other_partition. */
631 bp_pack_value (&bp, node->tls_model, 3);
632 bp_pack_value (&bp, node->used_by_single_function, 1);
633 bp_pack_value (&bp, node->dynamically_initialized, 1);
634 bp_pack_value (&bp, node->need_bounds_init, 1);
635 streamer_write_bitpack (&bp);
637 group = node->get_comdat_group ();
638 if (group)
639 comdat = IDENTIFIER_POINTER (group);
640 else
641 comdat = "";
642 streamer_write_data_stream (ob->main_stream, comdat, strlen (comdat) + 1);
644 if (group)
646 if (node->same_comdat_group)
648 ref = LCC_NOT_FOUND;
649 for (struct symtab_node *n = node->same_comdat_group;
650 ref == LCC_NOT_FOUND && n != node; n = n->same_comdat_group)
651 ref = lto_symtab_encoder_lookup (encoder, n);
653 else
654 ref = LCC_NOT_FOUND;
655 streamer_write_hwi_stream (ob->main_stream, ref);
658 section = node->get_section ();
659 if (!section)
660 section = "";
661 streamer_write_data_stream (ob->main_stream, section, strlen (section) + 1);
663 streamer_write_enum (ob->main_stream, ld_plugin_symbol_resolution,
664 LDPR_NUM_KNOWN, node->resolution);
667 /* Output the varpool NODE to OB.
668 If NODE is not in SET, then NODE is a boundary. */
670 static void
671 lto_output_ref (struct lto_simple_output_block *ob, struct ipa_ref *ref,
672 lto_symtab_encoder_t encoder)
674 struct bitpack_d bp;
675 int nref;
676 int uid = ref->lto_stmt_uid;
677 struct cgraph_node *node;
679 bp = bitpack_create (ob->main_stream);
680 bp_pack_value (&bp, ref->use, 3);
681 bp_pack_value (&bp, ref->speculative, 1);
682 streamer_write_bitpack (&bp);
683 nref = lto_symtab_encoder_lookup (encoder, ref->referred);
684 gcc_assert (nref != LCC_NOT_FOUND);
685 streamer_write_hwi_stream (ob->main_stream, nref);
687 node = dyn_cast <cgraph_node *> (ref->referring);
688 if (node)
690 if (ref->stmt)
691 uid = gimple_uid (ref->stmt) + 1;
692 streamer_write_hwi_stream (ob->main_stream, uid);
696 /* Stream out profile_summary to OB. */
698 static void
699 output_profile_summary (struct lto_simple_output_block *ob)
701 unsigned h_ix;
702 struct bitpack_d bp;
704 if (profile_info)
706 /* We do not output num and run_max, they are not used by
707 GCC profile feedback and they are difficult to merge from multiple
708 units. */
709 gcc_assert (profile_info->runs);
710 streamer_write_uhwi_stream (ob->main_stream, profile_info->runs);
711 streamer_write_gcov_count_stream (ob->main_stream, profile_info->sum_max);
713 /* sum_all is needed for computing the working set with the
714 histogram. */
715 streamer_write_gcov_count_stream (ob->main_stream, profile_info->sum_all);
717 /* Create and output a bitpack of non-zero histogram entries indices. */
718 bp = bitpack_create (ob->main_stream);
719 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
720 bp_pack_value (&bp, profile_info->histogram[h_ix].num_counters > 0, 1);
721 streamer_write_bitpack (&bp);
722 /* Now stream out only those non-zero entries. */
723 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
725 if (!profile_info->histogram[h_ix].num_counters)
726 continue;
727 streamer_write_gcov_count_stream (ob->main_stream,
728 profile_info->histogram[h_ix].num_counters);
729 streamer_write_gcov_count_stream (ob->main_stream,
730 profile_info->histogram[h_ix].min_value);
731 streamer_write_gcov_count_stream (ob->main_stream,
732 profile_info->histogram[h_ix].cum_value);
734 /* IPA-profile computes hot bb threshold based on cumulated
735 whole program profile. We need to stream it down to ltrans. */
736 if (flag_wpa)
737 streamer_write_gcov_count_stream (ob->main_stream,
738 get_hot_bb_threshold ());
740 else
741 streamer_write_uhwi_stream (ob->main_stream, 0);
744 /* Output all callees or indirect outgoing edges. EDGE must be the first such
745 edge. */
747 static void
748 output_outgoing_cgraph_edges (struct cgraph_edge *edge,
749 struct lto_simple_output_block *ob,
750 lto_symtab_encoder_t encoder)
752 if (!edge)
753 return;
755 /* Output edges in backward direction, so the reconstructed callgraph match
756 and it is easy to associate call sites in the IPA pass summaries. */
757 while (edge->next_callee)
758 edge = edge->next_callee;
759 for (; edge; edge = edge->prev_callee)
760 lto_output_edge (ob, edge, encoder);
763 /* Output the part of the cgraph in SET. */
765 static void
766 output_refs (lto_symtab_encoder_t encoder)
768 struct lto_simple_output_block *ob;
769 int count;
770 struct ipa_ref *ref;
772 ob = lto_create_simple_output_block (LTO_section_refs);
774 for (int i = 0; i < lto_symtab_encoder_size (encoder); i++)
776 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
778 /* IPA_REF_ALIAS and IPA_REF_CHKP references are always preserved
779 in the boundary. Alias node can't have other references and
780 can be always handled as if it's not in the boundary. */
781 if (!node->alias && !lto_symtab_encoder_in_partition_p (encoder, node))
783 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
784 /* Output IPA_REF_CHKP reference. */
785 if (cnode
786 && cnode->instrumented_version
787 && !cnode->instrumentation_clone)
789 for (int i = 0; node->iterate_reference (i, ref); i++)
790 if (ref->use == IPA_REF_CHKP)
792 if (lto_symtab_encoder_lookup (encoder, ref->referred)
793 != LCC_NOT_FOUND)
795 int nref = lto_symtab_encoder_lookup (encoder, node);
796 streamer_write_gcov_count_stream (ob->main_stream, 1);
797 streamer_write_uhwi_stream (ob->main_stream, nref);
798 lto_output_ref (ob, ref, encoder);
800 break;
803 continue;
806 count = node->ref_list.nreferences ();
807 if (count)
809 streamer_write_gcov_count_stream (ob->main_stream, count);
810 streamer_write_uhwi_stream (ob->main_stream,
811 lto_symtab_encoder_lookup (encoder, node));
812 for (int i = 0; node->iterate_reference (i, ref); i++)
813 lto_output_ref (ob, ref, encoder);
817 streamer_write_uhwi_stream (ob->main_stream, 0);
819 lto_destroy_simple_output_block (ob);
822 /* Add NODE into encoder as well as nodes it is cloned from.
823 Do it in a way so clones appear first. */
825 static void
826 add_node_to (lto_symtab_encoder_t encoder, struct cgraph_node *node,
827 bool include_body)
829 if (node->clone_of)
830 add_node_to (encoder, node->clone_of, include_body);
831 else if (include_body)
832 lto_set_symtab_encoder_encode_body (encoder, node);
833 lto_symtab_encoder_encode (encoder, node);
836 /* Add all references in NODE to encoders. */
838 static void
839 create_references (lto_symtab_encoder_t encoder, symtab_node *node)
841 int i;
842 struct ipa_ref *ref = NULL;
843 for (i = 0; node->iterate_reference (i, ref); i++)
844 if (is_a <cgraph_node *> (ref->referred))
845 add_node_to (encoder, dyn_cast <cgraph_node *> (ref->referred), false);
846 else
847 lto_symtab_encoder_encode (encoder, ref->referred);
850 /* Select what needs to be streamed out. In regular lto mode stream everything.
851 In offload lto mode stream only nodes marked as offloadable. */
852 void
853 select_what_to_stream (void)
855 struct symtab_node *snode;
856 FOR_EACH_SYMBOL (snode)
857 snode->need_lto_streaming = !lto_stream_offload_p || snode->offloadable;
860 /* Find all symbols we want to stream into given partition and insert them
861 to encoders.
863 The function actually replaces IN_ENCODER by new one. The reason is that
864 streaming code needs clone's origin to be streamed before clone. This
865 means that we need to insert the nodes in specific order. This order is
866 ignored by the partitioning logic earlier. */
868 lto_symtab_encoder_t
869 compute_ltrans_boundary (lto_symtab_encoder_t in_encoder)
871 struct cgraph_edge *edge;
872 int i;
873 lto_symtab_encoder_t encoder;
874 lto_symtab_encoder_iterator lsei;
875 hash_set<void *> reachable_call_targets;
877 encoder = lto_symtab_encoder_new (false);
879 /* Go over all entries in the IN_ENCODER and duplicate them to
880 ENCODER. At the same time insert masters of clones so
881 every master appears before clone. */
882 for (lsei = lsei_start_function_in_partition (in_encoder);
883 !lsei_end_p (lsei); lsei_next_function_in_partition (&lsei))
885 struct cgraph_node *node = lsei_cgraph_node (lsei);
886 if (!node->need_lto_streaming)
887 continue;
888 add_node_to (encoder, node, true);
889 lto_set_symtab_encoder_in_partition (encoder, node);
890 create_references (encoder, node);
892 for (lsei = lsei_start_variable_in_partition (in_encoder);
893 !lsei_end_p (lsei); lsei_next_variable_in_partition (&lsei))
895 varpool_node *vnode = lsei_varpool_node (lsei);
897 if (!vnode->need_lto_streaming)
898 continue;
899 lto_set_symtab_encoder_in_partition (encoder, vnode);
900 lto_set_symtab_encoder_encode_initializer (encoder, vnode);
901 create_references (encoder, vnode);
903 /* Pickle in also the initializer of all referenced readonly variables
904 to help folding. Constant pool variables are not shared, so we must
905 pickle those too. */
906 for (i = 0; i < lto_symtab_encoder_size (encoder); i++)
908 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
909 if (varpool_node *vnode = dyn_cast <varpool_node *> (node))
911 if (!lto_symtab_encoder_encode_initializer_p (encoder,
912 vnode)
913 && (((vnode->ctor_useable_for_folding_p ()
914 && (!DECL_VIRTUAL_P (vnode->decl)
915 || !flag_wpa
916 || flag_ltrans_devirtualize))
917 || POINTER_BOUNDS_P (vnode->decl))))
919 lto_set_symtab_encoder_encode_initializer (encoder, vnode);
920 create_references (encoder, vnode);
925 /* Go over all the nodes again to include callees that are not in
926 SET. */
927 for (lsei = lsei_start_function_in_partition (encoder);
928 !lsei_end_p (lsei); lsei_next_function_in_partition (&lsei))
930 struct cgraph_node *node = lsei_cgraph_node (lsei);
931 for (edge = node->callees; edge; edge = edge->next_callee)
933 struct cgraph_node *callee = edge->callee;
934 if (!lto_symtab_encoder_in_partition_p (encoder, callee))
936 /* We should have moved all the inlines. */
937 gcc_assert (!callee->global.inlined_to);
938 add_node_to (encoder, callee, false);
941 /* Add all possible targets for late devirtualization. */
942 if (flag_ltrans_devirtualize || !flag_wpa)
943 for (edge = node->indirect_calls; edge; edge = edge->next_callee)
944 if (edge->indirect_info->polymorphic)
946 unsigned int i;
947 void *cache_token;
948 bool final;
949 vec <cgraph_node *>targets
950 = possible_polymorphic_call_targets
951 (edge, &final, &cache_token);
952 if (!reachable_call_targets.add (cache_token))
954 for (i = 0; i < targets.length (); i++)
956 struct cgraph_node *callee = targets[i];
958 /* Adding an external declarations into the unit serves
959 no purpose and just increases its boundary. */
960 if (callee->definition
961 && !lto_symtab_encoder_in_partition_p
962 (encoder, callee))
964 gcc_assert (!callee->global.inlined_to);
965 add_node_to (encoder, callee, false);
971 /* Be sure to also insert alias targert and thunk callees. These needs
972 to stay to aid local calling conventions. */
973 for (i = 0; i < lto_symtab_encoder_size (encoder); i++)
975 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
976 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
978 if (node->alias && node->analyzed)
979 create_references (encoder, node);
980 if (cnode
981 && cnode->thunk.thunk_p && !cnode->global.inlined_to)
982 add_node_to (encoder, cnode->callees->callee, false);
983 while (node->transparent_alias && node->analyzed)
985 node = node->get_alias_target ();
986 if (is_a <cgraph_node *> (node))
987 add_node_to (encoder, dyn_cast <cgraph_node *> (node),
988 false);
989 else
990 lto_symtab_encoder_encode (encoder, node);
993 lto_symtab_encoder_delete (in_encoder);
994 return encoder;
997 /* Output the part of the symtab in SET and VSET. */
999 void
1000 output_symtab (void)
1002 struct cgraph_node *node;
1003 struct lto_simple_output_block *ob;
1004 int i, n_nodes;
1005 lto_symtab_encoder_t encoder;
1007 if (flag_wpa)
1008 output_cgraph_opt_summary ();
1010 ob = lto_create_simple_output_block (LTO_section_symtab_nodes);
1012 output_profile_summary (ob);
1014 /* An encoder for cgraph nodes should have been created by
1015 ipa_write_summaries_1. */
1016 gcc_assert (ob->decl_state->symtab_node_encoder);
1017 encoder = ob->decl_state->symtab_node_encoder;
1019 /* Write out the nodes. We must first output a node and then its clones,
1020 otherwise at a time reading back the node there would be nothing to clone
1021 from. */
1022 n_nodes = lto_symtab_encoder_size (encoder);
1023 for (i = 0; i < n_nodes; i++)
1025 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
1026 if (cgraph_node *cnode = dyn_cast <cgraph_node *> (node))
1027 lto_output_node (ob, cnode, encoder);
1028 else
1029 lto_output_varpool_node (ob, dyn_cast<varpool_node *> (node), encoder);
1032 /* Go over the nodes in SET again to write edges. */
1033 for (int i = 0; i < lto_symtab_encoder_size (encoder); i++)
1035 node = dyn_cast <cgraph_node *> (lto_symtab_encoder_deref (encoder, i));
1036 if (node
1037 && ((node->thunk.thunk_p && !node->global.inlined_to)
1038 || lto_symtab_encoder_in_partition_p (encoder, node)))
1040 output_outgoing_cgraph_edges (node->callees, ob, encoder);
1041 output_outgoing_cgraph_edges (node->indirect_calls, ob, encoder);
1045 streamer_write_uhwi_stream (ob->main_stream, 0);
1047 lto_destroy_simple_output_block (ob);
1049 /* Emit toplevel asms.
1050 When doing WPA we must output every asm just once. Since we do not partition asm
1051 nodes at all, output them to first output. This is kind of hack, but should work
1052 well. */
1053 if (!asm_nodes_output)
1055 asm_nodes_output = true;
1056 lto_output_toplevel_asms ();
1059 output_refs (encoder);
1062 /* Return identifier encoded in IB as a plain string. */
1064 static tree
1065 read_identifier (struct lto_input_block *ib)
1067 unsigned int len = strnlen (ib->data + ib->p, ib->len - ib->p - 1);
1068 tree id;
1070 if (ib->data[ib->p + len])
1071 lto_section_overrun (ib);
1072 if (!len)
1074 ib->p++;
1075 return NULL;
1077 id = get_identifier (ib->data + ib->p);
1078 ib->p += len + 1;
1079 return id;
1082 /* Return string encoded in IB, NULL if string is empty. */
1084 static const char *
1085 read_string (struct lto_input_block *ib)
1087 unsigned int len = strnlen (ib->data + ib->p, ib->len - ib->p - 1);
1088 const char *str;
1090 if (ib->data[ib->p + len])
1091 lto_section_overrun (ib);
1092 if (!len)
1094 ib->p++;
1095 return NULL;
1097 str = ib->data + ib->p;
1098 ib->p += len + 1;
1099 return str;
1102 /* Output function/variable tables that will allow libgomp to look up offload
1103 target code.
1104 OFFLOAD_FUNCS is filled in expand_omp_target, OFFLOAD_VARS is filled in
1105 varpool_node::get_create. In WHOPR (partitioned) mode during the WPA stage
1106 both OFFLOAD_FUNCS and OFFLOAD_VARS are filled by input_offload_tables. */
1108 void
1109 output_offload_tables (void)
1111 if (vec_safe_is_empty (offload_funcs) && vec_safe_is_empty (offload_vars))
1112 return;
1114 struct lto_simple_output_block *ob
1115 = lto_create_simple_output_block (LTO_section_offload_table);
1117 for (unsigned i = 0; i < vec_safe_length (offload_funcs); i++)
1119 streamer_write_enum (ob->main_stream, LTO_symtab_tags,
1120 LTO_symtab_last_tag, LTO_symtab_unavail_node);
1121 lto_output_fn_decl_index (ob->decl_state, ob->main_stream,
1122 (*offload_funcs)[i]);
1125 for (unsigned i = 0; i < vec_safe_length (offload_vars); i++)
1127 streamer_write_enum (ob->main_stream, LTO_symtab_tags,
1128 LTO_symtab_last_tag, LTO_symtab_variable);
1129 lto_output_var_decl_index (ob->decl_state, ob->main_stream,
1130 (*offload_vars)[i]);
1133 streamer_write_uhwi_stream (ob->main_stream, 0);
1134 lto_destroy_simple_output_block (ob);
1136 /* In WHOPR mode during the WPA stage the joint offload tables need to be
1137 streamed to one partition only. That's why we free offload_funcs and
1138 offload_vars after the first call of output_offload_tables. */
1139 if (flag_wpa)
1141 vec_free (offload_funcs);
1142 vec_free (offload_vars);
1146 /* Overwrite the information in NODE based on FILE_DATA, TAG, FLAGS,
1147 STACK_SIZE, SELF_TIME and SELF_SIZE. This is called either to initialize
1148 NODE or to replace the values in it, for instance because the first
1149 time we saw it, the function body was not available but now it
1150 is. BP is a bitpack with all the bitflags for NODE read from the
1151 stream. */
1153 static void
1154 input_overwrite_node (struct lto_file_decl_data *file_data,
1155 struct cgraph_node *node,
1156 enum LTO_symtab_tags tag,
1157 struct bitpack_d *bp)
1159 node->aux = (void *) tag;
1160 node->lto_file_data = file_data;
1162 node->local.local = bp_unpack_value (bp, 1);
1163 node->externally_visible = bp_unpack_value (bp, 1);
1164 node->no_reorder = bp_unpack_value (bp, 1);
1165 node->definition = bp_unpack_value (bp, 1);
1166 node->local.versionable = bp_unpack_value (bp, 1);
1167 node->local.can_change_signature = bp_unpack_value (bp, 1);
1168 node->local.redefined_extern_inline = bp_unpack_value (bp, 1);
1169 node->force_output = bp_unpack_value (bp, 1);
1170 node->forced_by_abi = bp_unpack_value (bp, 1);
1171 node->unique_name = bp_unpack_value (bp, 1);
1172 node->body_removed = bp_unpack_value (bp, 1);
1173 node->implicit_section = bp_unpack_value (bp, 1);
1174 node->address_taken = bp_unpack_value (bp, 1);
1175 node->used_from_other_partition = bp_unpack_value (bp, 1);
1176 node->lowered = bp_unpack_value (bp, 1);
1177 node->analyzed = tag == LTO_symtab_analyzed_node;
1178 node->in_other_partition = bp_unpack_value (bp, 1);
1179 if (node->in_other_partition
1180 /* Avoid updating decl when we are seeing just inline clone.
1181 When inlining function that has functions already inlined into it,
1182 we produce clones of inline clones.
1184 WPA partitioning might put each clone into different unit and
1185 we might end up streaming inline clone from other partition
1186 to support clone we are interested in. */
1187 && (!node->clone_of
1188 || node->clone_of->decl != node->decl))
1190 DECL_EXTERNAL (node->decl) = 1;
1191 TREE_STATIC (node->decl) = 0;
1193 node->alias = bp_unpack_value (bp, 1);
1194 node->transparent_alias = bp_unpack_value (bp, 1);
1195 node->weakref = bp_unpack_value (bp, 1);
1196 node->frequency = (enum node_frequency)bp_unpack_value (bp, 2);
1197 node->only_called_at_startup = bp_unpack_value (bp, 1);
1198 node->only_called_at_exit = bp_unpack_value (bp, 1);
1199 node->tm_clone = bp_unpack_value (bp, 1);
1200 node->calls_comdat_local = bp_unpack_value (bp, 1);
1201 node->icf_merged = bp_unpack_value (bp, 1);
1202 node->nonfreeing_fn = bp_unpack_value (bp, 1);
1203 node->thunk.thunk_p = bp_unpack_value (bp, 1);
1204 node->parallelized_function = bp_unpack_value (bp, 1);
1205 node->resolution = bp_unpack_enum (bp, ld_plugin_symbol_resolution,
1206 LDPR_NUM_KNOWN);
1207 node->instrumentation_clone = bp_unpack_value (bp, 1);
1208 node->split_part = bp_unpack_value (bp, 1);
1209 gcc_assert (flag_ltrans
1210 || (!node->in_other_partition
1211 && !node->used_from_other_partition));
1214 /* Return string alias is alias of. */
1216 static tree
1217 get_alias_symbol (tree decl)
1219 tree alias = lookup_attribute ("alias", DECL_ATTRIBUTES (decl));
1220 return get_identifier (TREE_STRING_POINTER
1221 (TREE_VALUE (TREE_VALUE (alias))));
1224 /* Read a node from input_block IB. TAG is the node's tag just read.
1225 Return the node read or overwriten. */
1227 static struct cgraph_node *
1228 input_node (struct lto_file_decl_data *file_data,
1229 struct lto_input_block *ib,
1230 enum LTO_symtab_tags tag,
1231 vec<symtab_node *> nodes)
1233 gcc::pass_manager *passes = g->get_passes ();
1234 tree fn_decl;
1235 struct cgraph_node *node;
1236 struct bitpack_d bp;
1237 unsigned decl_index;
1238 int ref = LCC_NOT_FOUND, ref2 = LCC_NOT_FOUND;
1239 int clone_ref;
1240 int order;
1241 int i, count;
1242 tree group;
1243 const char *section;
1244 order = streamer_read_hwi (ib) + order_base;
1245 clone_ref = streamer_read_hwi (ib);
1247 decl_index = streamer_read_uhwi (ib);
1248 fn_decl = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1250 if (clone_ref != LCC_NOT_FOUND)
1252 node = dyn_cast<cgraph_node *> (nodes[clone_ref])->create_clone (fn_decl,
1253 profile_count::uninitialized (), false,
1254 vNULL, false, NULL, NULL);
1256 else
1258 /* Declaration of functions can be already merged with a declaration
1259 from other input file. We keep cgraph unmerged until after streaming
1260 of ipa passes is done. Alays forcingly create a fresh node. */
1261 node = symtab->create_empty ();
1262 node->decl = fn_decl;
1263 if (lookup_attribute ("ifunc", DECL_ATTRIBUTES (fn_decl)))
1264 node->ifunc_resolver = 1;
1265 node->register_symbol ();
1268 node->order = order;
1269 if (order >= symtab->order)
1270 symtab->order = order + 1;
1272 node->count = profile_count::stream_in (ib);
1273 node->count_materialization_scale = streamer_read_hwi (ib);
1275 count = streamer_read_hwi (ib);
1276 node->ipa_transforms_to_apply = vNULL;
1277 for (i = 0; i < count; i++)
1279 opt_pass *pass;
1280 int pid = streamer_read_hwi (ib);
1282 gcc_assert (pid < passes->passes_by_id_size);
1283 pass = passes->passes_by_id[pid];
1284 node->ipa_transforms_to_apply.safe_push ((ipa_opt_pass_d *) pass);
1287 if (tag == LTO_symtab_analyzed_node)
1288 ref = streamer_read_hwi (ib);
1290 group = read_identifier (ib);
1291 if (group)
1292 ref2 = streamer_read_hwi (ib);
1294 /* Make sure that we have not read this node before. Nodes that
1295 have already been read will have their tag stored in the 'aux'
1296 field. Since built-in functions can be referenced in multiple
1297 functions, they are expected to be read more than once. */
1298 if (node->aux && !DECL_BUILT_IN (node->decl))
1299 internal_error ("bytecode stream: found multiple instances of cgraph "
1300 "node with uid %d", node->uid);
1302 node->tp_first_run = streamer_read_uhwi (ib);
1304 bp = streamer_read_bitpack (ib);
1306 input_overwrite_node (file_data, node, tag, &bp);
1308 /* Store a reference for now, and fix up later to be a pointer. */
1309 node->global.inlined_to = (cgraph_node *) (intptr_t) ref;
1311 if (group)
1313 node->set_comdat_group (group);
1314 /* Store a reference for now, and fix up later to be a pointer. */
1315 node->same_comdat_group = (symtab_node *) (intptr_t) ref2;
1317 else
1318 node->same_comdat_group = (symtab_node *) (intptr_t) LCC_NOT_FOUND;
1319 section = read_string (ib);
1320 if (section)
1321 node->set_section_for_node (section);
1323 if (node->thunk.thunk_p)
1325 int type = streamer_read_uhwi (ib);
1326 HOST_WIDE_INT fixed_offset = streamer_read_uhwi (ib);
1327 HOST_WIDE_INT virtual_value = streamer_read_uhwi (ib);
1329 node->thunk.fixed_offset = fixed_offset;
1330 node->thunk.this_adjusting = (type & 2);
1331 node->thunk.virtual_value = virtual_value;
1332 node->thunk.virtual_offset_p = (type & 4);
1333 node->thunk.add_pointer_bounds_args = (type & 8);
1335 if (node->alias && !node->analyzed && node->weakref)
1336 node->alias_target = get_alias_symbol (node->decl);
1337 node->profile_id = streamer_read_hwi (ib);
1338 if (DECL_STATIC_CONSTRUCTOR (node->decl))
1339 node->set_init_priority (streamer_read_hwi (ib));
1340 if (DECL_STATIC_DESTRUCTOR (node->decl))
1341 node->set_fini_priority (streamer_read_hwi (ib));
1343 if (node->instrumentation_clone)
1345 decl_index = streamer_read_uhwi (ib);
1346 fn_decl = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1347 node->orig_decl = fn_decl;
1350 return node;
1353 /* Read a node from input_block IB. TAG is the node's tag just read.
1354 Return the node read or overwriten. */
1356 static varpool_node *
1357 input_varpool_node (struct lto_file_decl_data *file_data,
1358 struct lto_input_block *ib)
1360 int decl_index;
1361 tree var_decl;
1362 varpool_node *node;
1363 struct bitpack_d bp;
1364 int ref = LCC_NOT_FOUND;
1365 int order;
1366 tree group;
1367 const char *section;
1369 order = streamer_read_hwi (ib) + order_base;
1370 decl_index = streamer_read_uhwi (ib);
1371 var_decl = lto_file_decl_data_get_var_decl (file_data, decl_index);
1373 /* Declaration of functions can be already merged with a declaration
1374 from other input file. We keep cgraph unmerged until after streaming
1375 of ipa passes is done. Alays forcingly create a fresh node. */
1376 node = varpool_node::create_empty ();
1377 node->decl = var_decl;
1378 node->register_symbol ();
1380 node->order = order;
1381 if (order >= symtab->order)
1382 symtab->order = order + 1;
1383 node->lto_file_data = file_data;
1385 bp = streamer_read_bitpack (ib);
1386 node->externally_visible = bp_unpack_value (&bp, 1);
1387 node->no_reorder = bp_unpack_value (&bp, 1);
1388 node->force_output = bp_unpack_value (&bp, 1);
1389 node->forced_by_abi = bp_unpack_value (&bp, 1);
1390 node->unique_name = bp_unpack_value (&bp, 1);
1391 node->body_removed = bp_unpack_value (&bp, 1);
1392 node->implicit_section = bp_unpack_value (&bp, 1);
1393 node->writeonly = bp_unpack_value (&bp, 1);
1394 node->definition = bp_unpack_value (&bp, 1);
1395 node->alias = bp_unpack_value (&bp, 1);
1396 node->transparent_alias = bp_unpack_value (&bp, 1);
1397 node->weakref = bp_unpack_value (&bp, 1);
1398 node->analyzed = bp_unpack_value (&bp, 1);
1399 node->used_from_other_partition = bp_unpack_value (&bp, 1);
1400 node->in_other_partition = bp_unpack_value (&bp, 1);
1401 if (node->in_other_partition)
1403 DECL_EXTERNAL (node->decl) = 1;
1404 TREE_STATIC (node->decl) = 0;
1406 if (node->alias && !node->analyzed && node->weakref)
1407 node->alias_target = get_alias_symbol (node->decl);
1408 node->tls_model = (enum tls_model)bp_unpack_value (&bp, 3);
1409 node->used_by_single_function = (enum tls_model)bp_unpack_value (&bp, 1);
1410 node->dynamically_initialized = bp_unpack_value (&bp, 1);
1411 node->need_bounds_init = bp_unpack_value (&bp, 1);
1412 group = read_identifier (ib);
1413 if (group)
1415 node->set_comdat_group (group);
1416 ref = streamer_read_hwi (ib);
1417 /* Store a reference for now, and fix up later to be a pointer. */
1418 node->same_comdat_group = (symtab_node *) (intptr_t) ref;
1420 else
1421 node->same_comdat_group = (symtab_node *) (intptr_t) LCC_NOT_FOUND;
1422 section = read_string (ib);
1423 if (section)
1424 node->set_section_for_node (section);
1425 node->resolution = streamer_read_enum (ib, ld_plugin_symbol_resolution,
1426 LDPR_NUM_KNOWN);
1427 gcc_assert (flag_ltrans
1428 || (!node->in_other_partition
1429 && !node->used_from_other_partition));
1431 return node;
1434 /* Read a node from input_block IB. TAG is the node's tag just read.
1435 Return the node read or overwriten. */
1437 static void
1438 input_ref (struct lto_input_block *ib,
1439 symtab_node *referring_node,
1440 vec<symtab_node *> nodes)
1442 symtab_node *node = NULL;
1443 struct bitpack_d bp;
1444 enum ipa_ref_use use;
1445 bool speculative;
1446 struct ipa_ref *ref;
1448 bp = streamer_read_bitpack (ib);
1449 use = (enum ipa_ref_use) bp_unpack_value (&bp, 3);
1450 speculative = (enum ipa_ref_use) bp_unpack_value (&bp, 1);
1451 node = nodes[streamer_read_hwi (ib)];
1452 ref = referring_node->create_reference (node, use);
1453 ref->speculative = speculative;
1454 if (is_a <cgraph_node *> (referring_node))
1455 ref->lto_stmt_uid = streamer_read_hwi (ib);
1458 /* Read an edge from IB. NODES points to a vector of previously read nodes for
1459 decoding caller and callee of the edge to be read. If INDIRECT is true, the
1460 edge being read is indirect (in the sense that it has
1461 indirect_unknown_callee set). */
1463 static void
1464 input_edge (struct lto_input_block *ib, vec<symtab_node *> nodes,
1465 bool indirect)
1467 struct cgraph_node *caller, *callee;
1468 struct cgraph_edge *edge;
1469 unsigned int stmt_id;
1470 profile_count count;
1471 cgraph_inline_failed_t inline_failed;
1472 struct bitpack_d bp;
1473 int ecf_flags = 0;
1475 caller = dyn_cast<cgraph_node *> (nodes[streamer_read_hwi (ib)]);
1476 if (caller == NULL || caller->decl == NULL_TREE)
1477 internal_error ("bytecode stream: no caller found while reading edge");
1479 if (!indirect)
1481 callee = dyn_cast<cgraph_node *> (nodes[streamer_read_hwi (ib)]);
1482 if (callee == NULL || callee->decl == NULL_TREE)
1483 internal_error ("bytecode stream: no callee found while reading edge");
1485 else
1486 callee = NULL;
1488 count = profile_count::stream_in (ib);
1490 bp = streamer_read_bitpack (ib);
1491 inline_failed = bp_unpack_enum (&bp, cgraph_inline_failed_t, CIF_N_REASONS);
1492 stmt_id = bp_unpack_var_len_unsigned (&bp);
1494 if (indirect)
1495 edge = caller->create_indirect_edge (NULL, 0, count);
1496 else
1497 edge = caller->create_edge (callee, NULL, count);
1499 edge->indirect_inlining_edge = bp_unpack_value (&bp, 1);
1500 edge->speculative = bp_unpack_value (&bp, 1);
1501 edge->lto_stmt_uid = stmt_id;
1502 edge->inline_failed = inline_failed;
1503 edge->call_stmt_cannot_inline_p = bp_unpack_value (&bp, 1);
1504 edge->can_throw_external = bp_unpack_value (&bp, 1);
1505 edge->in_polymorphic_cdtor = bp_unpack_value (&bp, 1);
1506 if (indirect)
1508 if (bp_unpack_value (&bp, 1))
1509 ecf_flags |= ECF_CONST;
1510 if (bp_unpack_value (&bp, 1))
1511 ecf_flags |= ECF_PURE;
1512 if (bp_unpack_value (&bp, 1))
1513 ecf_flags |= ECF_NORETURN;
1514 if (bp_unpack_value (&bp, 1))
1515 ecf_flags |= ECF_MALLOC;
1516 if (bp_unpack_value (&bp, 1))
1517 ecf_flags |= ECF_NOTHROW;
1518 if (bp_unpack_value (&bp, 1))
1519 ecf_flags |= ECF_RETURNS_TWICE;
1520 edge->indirect_info->ecf_flags = ecf_flags;
1521 edge->indirect_info->common_target_id = streamer_read_hwi (ib);
1522 if (edge->indirect_info->common_target_id)
1523 edge->indirect_info->common_target_probability = streamer_read_hwi (ib);
1528 /* Read a cgraph from IB using the info in FILE_DATA. */
1530 static vec<symtab_node *>
1531 input_cgraph_1 (struct lto_file_decl_data *file_data,
1532 struct lto_input_block *ib)
1534 enum LTO_symtab_tags tag;
1535 vec<symtab_node *> nodes = vNULL;
1536 symtab_node *node;
1537 unsigned i;
1539 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1540 order_base = symtab->order;
1541 while (tag)
1543 if (tag == LTO_symtab_edge)
1544 input_edge (ib, nodes, false);
1545 else if (tag == LTO_symtab_indirect_edge)
1546 input_edge (ib, nodes, true);
1547 else if (tag == LTO_symtab_variable)
1549 node = input_varpool_node (file_data, ib);
1550 nodes.safe_push (node);
1551 lto_symtab_encoder_encode (file_data->symtab_node_encoder, node);
1553 else
1555 node = input_node (file_data, ib, tag, nodes);
1556 if (node == NULL || node->decl == NULL_TREE)
1557 internal_error ("bytecode stream: found empty cgraph node");
1558 nodes.safe_push (node);
1559 lto_symtab_encoder_encode (file_data->symtab_node_encoder, node);
1562 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1565 lto_input_toplevel_asms (file_data, order_base);
1567 /* AUX pointers should be all non-zero for function nodes read from the stream. */
1568 if (flag_checking)
1570 FOR_EACH_VEC_ELT (nodes, i, node)
1571 gcc_assert (node->aux || !is_a <cgraph_node *> (node));
1573 FOR_EACH_VEC_ELT (nodes, i, node)
1575 int ref;
1576 if (cgraph_node *cnode = dyn_cast <cgraph_node *> (node))
1578 ref = (int) (intptr_t) cnode->global.inlined_to;
1580 /* We share declaration of builtins, so we may read same node twice. */
1581 if (!node->aux)
1582 continue;
1583 node->aux = NULL;
1585 /* Fixup inlined_to from reference to pointer. */
1586 if (ref != LCC_NOT_FOUND)
1587 dyn_cast<cgraph_node *> (node)->global.inlined_to
1588 = dyn_cast<cgraph_node *> (nodes[ref]);
1589 else
1590 cnode->global.inlined_to = NULL;
1592 /* Compute instrumented_version. */
1593 if (cnode->instrumentation_clone)
1595 gcc_assert (cnode->orig_decl);
1597 cnode->instrumented_version = cgraph_node::get (cnode->orig_decl);
1598 if (cnode->instrumented_version)
1600 /* We may have multiple nodes for a single function which
1601 will be merged later. To have a proper merge we need
1602 to keep instrumentation_version reference between nodes
1603 consistent: each instrumented_version reference should
1604 have proper reverse reference. Thus don't break existing
1605 instrumented_version reference if it already exists. */
1606 if (cnode->instrumented_version->instrumented_version)
1607 cnode->instrumented_version = NULL;
1608 else
1609 cnode->instrumented_version->instrumented_version = cnode;
1612 /* Restore decl names reference except for wrapper functions. */
1613 if (!chkp_wrap_function (cnode->orig_decl))
1615 tree name = DECL_ASSEMBLER_NAME (cnode->decl);
1616 IDENTIFIER_TRANSPARENT_ALIAS (name) = 1;
1617 TREE_CHAIN (name) = DECL_ASSEMBLER_NAME (cnode->orig_decl);
1622 ref = (int) (intptr_t) node->same_comdat_group;
1624 /* Fixup same_comdat_group from reference to pointer. */
1625 if (ref != LCC_NOT_FOUND)
1626 node->same_comdat_group = nodes[ref];
1627 else
1628 node->same_comdat_group = NULL;
1630 FOR_EACH_VEC_ELT (nodes, i, node)
1631 node->aux = is_a <cgraph_node *> (node) ? (void *)1 : NULL;
1632 return nodes;
1635 /* Input ipa_refs. */
1637 static void
1638 input_refs (struct lto_input_block *ib,
1639 vec<symtab_node *> nodes)
1641 int count;
1642 int idx;
1643 while (true)
1645 symtab_node *node;
1646 count = streamer_read_uhwi (ib);
1647 if (!count)
1648 break;
1649 idx = streamer_read_uhwi (ib);
1650 node = nodes[idx];
1651 while (count)
1653 input_ref (ib, node, nodes);
1654 count--;
1660 static struct gcov_ctr_summary lto_gcov_summary;
1662 /* Input profile_info from IB. */
1663 static void
1664 input_profile_summary (struct lto_input_block *ib,
1665 struct lto_file_decl_data *file_data)
1667 unsigned h_ix;
1668 struct bitpack_d bp;
1669 unsigned int runs = streamer_read_uhwi (ib);
1670 if (runs)
1672 file_data->profile_info.runs = runs;
1673 file_data->profile_info.sum_max = streamer_read_gcov_count (ib);
1674 file_data->profile_info.sum_all = streamer_read_gcov_count (ib);
1676 memset (file_data->profile_info.histogram, 0,
1677 sizeof (gcov_bucket_type) * GCOV_HISTOGRAM_SIZE);
1678 /* Input the bitpack of non-zero histogram indices. */
1679 bp = streamer_read_bitpack (ib);
1680 /* Read in and unpack the full bitpack, flagging non-zero
1681 histogram entries by setting the num_counters non-zero. */
1682 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1684 file_data->profile_info.histogram[h_ix].num_counters
1685 = bp_unpack_value (&bp, 1);
1687 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1689 if (!file_data->profile_info.histogram[h_ix].num_counters)
1690 continue;
1692 file_data->profile_info.histogram[h_ix].num_counters
1693 = streamer_read_gcov_count (ib);
1694 file_data->profile_info.histogram[h_ix].min_value
1695 = streamer_read_gcov_count (ib);
1696 file_data->profile_info.histogram[h_ix].cum_value
1697 = streamer_read_gcov_count (ib);
1699 /* IPA-profile computes hot bb threshold based on cumulated
1700 whole program profile. We need to stream it down to ltrans. */
1701 if (flag_ltrans)
1702 set_hot_bb_threshold (streamer_read_gcov_count (ib));
1707 /* Rescale profile summaries to the same number of runs in the whole unit. */
1709 static void
1710 merge_profile_summaries (struct lto_file_decl_data **file_data_vec)
1712 struct lto_file_decl_data *file_data;
1713 unsigned int j, h_ix;
1714 gcov_unsigned_t max_runs = 0;
1715 struct cgraph_node *node;
1716 struct cgraph_edge *edge;
1717 gcov_type saved_sum_all = 0;
1718 gcov_ctr_summary *saved_profile_info = 0;
1719 int saved_scale = 0;
1721 /* Find unit with maximal number of runs. If we ever get serious about
1722 roundoff errors, we might also consider computing smallest common
1723 multiply. */
1724 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1725 if (max_runs < file_data->profile_info.runs)
1726 max_runs = file_data->profile_info.runs;
1728 if (!max_runs)
1729 return;
1731 /* Simple overflow check. We probably don't need to support that many train
1732 runs. Such a large value probably imply data corruption anyway. */
1733 if (max_runs > INT_MAX / REG_BR_PROB_BASE)
1735 sorry ("At most %i profile runs is supported. Perhaps corrupted profile?",
1736 INT_MAX / REG_BR_PROB_BASE);
1737 return;
1740 profile_info = &lto_gcov_summary;
1741 lto_gcov_summary.runs = max_runs;
1742 lto_gcov_summary.sum_max = 0;
1743 memset (lto_gcov_summary.histogram, 0,
1744 sizeof (gcov_bucket_type) * GCOV_HISTOGRAM_SIZE);
1746 /* Rescale all units to the maximal number of runs.
1747 sum_max can not be easily merged, as we have no idea what files come from
1748 the same run. We do not use the info anyway, so leave it 0. */
1749 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1750 if (file_data->profile_info.runs)
1752 int scale = GCOV_COMPUTE_SCALE (max_runs,
1753 file_data->profile_info.runs);
1754 lto_gcov_summary.sum_max
1755 = MAX (lto_gcov_summary.sum_max,
1756 apply_scale (file_data->profile_info.sum_max, scale));
1757 lto_gcov_summary.sum_all
1758 = MAX (lto_gcov_summary.sum_all,
1759 apply_scale (file_data->profile_info.sum_all, scale));
1760 /* Save a pointer to the profile_info with the largest
1761 scaled sum_all and the scale for use in merging the
1762 histogram. */
1763 if (!saved_profile_info
1764 || lto_gcov_summary.sum_all > saved_sum_all)
1766 saved_profile_info = &file_data->profile_info;
1767 saved_sum_all = lto_gcov_summary.sum_all;
1768 saved_scale = scale;
1772 gcc_assert (saved_profile_info);
1774 /* Scale up the histogram from the profile that had the largest
1775 scaled sum_all above. */
1776 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1778 /* Scale up the min value as we did the corresponding sum_all
1779 above. Use that to find the new histogram index. */
1780 gcov_type scaled_min
1781 = apply_scale (saved_profile_info->histogram[h_ix].min_value,
1782 saved_scale);
1783 /* The new index may be shared with another scaled histogram entry,
1784 so we need to account for a non-zero histogram entry at new_ix. */
1785 unsigned new_ix = gcov_histo_index (scaled_min);
1786 lto_gcov_summary.histogram[new_ix].min_value
1787 = (lto_gcov_summary.histogram[new_ix].num_counters
1788 ? MIN (lto_gcov_summary.histogram[new_ix].min_value, scaled_min)
1789 : scaled_min);
1790 /* Some of the scaled counter values would ostensibly need to be placed
1791 into different (larger) histogram buckets, but we keep things simple
1792 here and place the scaled cumulative counter value in the bucket
1793 corresponding to the scaled minimum counter value. */
1794 lto_gcov_summary.histogram[new_ix].cum_value
1795 += apply_scale (saved_profile_info->histogram[h_ix].cum_value,
1796 saved_scale);
1797 lto_gcov_summary.histogram[new_ix].num_counters
1798 += saved_profile_info->histogram[h_ix].num_counters;
1801 /* Watch roundoff errors. */
1802 if (lto_gcov_summary.sum_max < max_runs)
1803 lto_gcov_summary.sum_max = max_runs;
1805 /* If merging already happent at WPA time, we are done. */
1806 if (flag_ltrans)
1807 return;
1809 /* Now compute count_materialization_scale of each node.
1810 During LTRANS we already have values of count_materialization_scale
1811 computed, so just update them. */
1812 FOR_EACH_FUNCTION (node)
1813 if (node->lto_file_data
1814 && node->lto_file_data->profile_info.runs)
1816 int scale;
1818 scale = RDIV (node->count_materialization_scale * max_runs,
1819 node->lto_file_data->profile_info.runs);
1820 node->count_materialization_scale = scale;
1821 if (scale < 0)
1822 fatal_error (input_location, "Profile information in %s corrupted",
1823 file_data->file_name);
1825 if (scale == REG_BR_PROB_BASE)
1826 continue;
1827 for (edge = node->callees; edge; edge = edge->next_callee)
1828 if (edge->count.ipa ().nonzero_p ())
1829 edge->count = edge->count.apply_scale (scale, REG_BR_PROB_BASE);
1830 for (edge = node->indirect_calls; edge; edge = edge->next_callee)
1831 if (edge->count.ipa ().nonzero_p ())
1832 edge->count = edge->count.apply_scale (scale, REG_BR_PROB_BASE);
1833 if (node->count.ipa ().nonzero_p ())
1834 node->count = node->count.apply_scale (scale, REG_BR_PROB_BASE);
1838 /* Input and merge the symtab from each of the .o files passed to
1839 lto1. */
1841 void
1842 input_symtab (void)
1844 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1845 struct lto_file_decl_data *file_data;
1846 unsigned int j = 0;
1847 struct cgraph_node *node;
1849 while ((file_data = file_data_vec[j++]))
1851 const char *data;
1852 size_t len;
1853 struct lto_input_block *ib;
1854 vec<symtab_node *> nodes;
1856 ib = lto_create_simple_input_block (file_data, LTO_section_symtab_nodes,
1857 &data, &len);
1858 if (!ib)
1859 fatal_error (input_location,
1860 "cannot find LTO cgraph in %s", file_data->file_name);
1861 input_profile_summary (ib, file_data);
1862 file_data->symtab_node_encoder = lto_symtab_encoder_new (true);
1863 nodes = input_cgraph_1 (file_data, ib);
1864 lto_destroy_simple_input_block (file_data, LTO_section_symtab_nodes,
1865 ib, data, len);
1867 ib = lto_create_simple_input_block (file_data, LTO_section_refs,
1868 &data, &len);
1869 if (!ib)
1870 fatal_error (input_location, "cannot find LTO section refs in %s",
1871 file_data->file_name);
1872 input_refs (ib, nodes);
1873 lto_destroy_simple_input_block (file_data, LTO_section_refs,
1874 ib, data, len);
1875 if (flag_ltrans)
1876 input_cgraph_opt_summary (nodes);
1877 nodes.release ();
1880 merge_profile_summaries (file_data_vec);
1882 if (!flag_auto_profile)
1883 get_working_sets ();
1886 /* Clear out the aux field that was used to store enough state to
1887 tell which nodes should be overwritten. */
1888 FOR_EACH_FUNCTION (node)
1890 /* Some nodes may have been created by cgraph_node. This
1891 happens when the callgraph contains nested functions. If the
1892 node for the parent function was never emitted to the gimple
1893 file, cgraph_node will create a node for it when setting the
1894 context of the nested function. */
1895 if (node->lto_file_data)
1896 node->aux = NULL;
1900 /* Input function/variable tables that will allow libgomp to look up offload
1901 target code, and store them into OFFLOAD_FUNCS and OFFLOAD_VARS. */
1903 void
1904 input_offload_tables (bool do_force_output)
1906 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1907 struct lto_file_decl_data *file_data;
1908 unsigned int j = 0;
1910 while ((file_data = file_data_vec[j++]))
1912 const char *data;
1913 size_t len;
1914 struct lto_input_block *ib
1915 = lto_create_simple_input_block (file_data, LTO_section_offload_table,
1916 &data, &len);
1917 if (!ib)
1918 continue;
1920 enum LTO_symtab_tags tag
1921 = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1922 while (tag)
1924 if (tag == LTO_symtab_unavail_node)
1926 int decl_index = streamer_read_uhwi (ib);
1927 tree fn_decl
1928 = lto_file_decl_data_get_fn_decl (file_data, decl_index);
1929 vec_safe_push (offload_funcs, fn_decl);
1931 /* Prevent IPA from removing fn_decl as unreachable, since there
1932 may be no refs from the parent function to child_fn in offload
1933 LTO mode. */
1934 if (do_force_output)
1935 cgraph_node::get (fn_decl)->mark_force_output ();
1937 else if (tag == LTO_symtab_variable)
1939 int decl_index = streamer_read_uhwi (ib);
1940 tree var_decl
1941 = lto_file_decl_data_get_var_decl (file_data, decl_index);
1942 vec_safe_push (offload_vars, var_decl);
1944 /* Prevent IPA from removing var_decl as unused, since there
1945 may be no refs to var_decl in offload LTO mode. */
1946 if (do_force_output)
1947 varpool_node::get (var_decl)->force_output = 1;
1949 else
1950 fatal_error (input_location,
1951 "invalid offload table in %s", file_data->file_name);
1953 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1956 lto_destroy_simple_input_block (file_data, LTO_section_offload_table,
1957 ib, data, len);
1961 /* True when we need optimization summary for NODE. */
1963 static int
1964 output_cgraph_opt_summary_p (struct cgraph_node *node)
1966 return ((node->clone_of || node->former_clone_of)
1967 && (node->clone.tree_map
1968 || node->clone.args_to_skip
1969 || node->clone.combined_args_to_skip));
1972 /* Output optimization summary for EDGE to OB. */
1973 static void
1974 output_edge_opt_summary (struct output_block *ob ATTRIBUTE_UNUSED,
1975 struct cgraph_edge *edge ATTRIBUTE_UNUSED)
1979 /* Output optimization summary for NODE to OB. */
1981 static void
1982 output_node_opt_summary (struct output_block *ob,
1983 struct cgraph_node *node,
1984 lto_symtab_encoder_t encoder)
1986 unsigned int index;
1987 bitmap_iterator bi;
1988 struct ipa_replace_map *map;
1989 struct bitpack_d bp;
1990 int i;
1991 struct cgraph_edge *e;
1993 if (node->clone.args_to_skip)
1995 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.args_to_skip));
1996 EXECUTE_IF_SET_IN_BITMAP (node->clone.args_to_skip, 0, index, bi)
1997 streamer_write_uhwi (ob, index);
1999 else
2000 streamer_write_uhwi (ob, 0);
2001 if (node->clone.combined_args_to_skip)
2003 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.combined_args_to_skip));
2004 EXECUTE_IF_SET_IN_BITMAP (node->clone.combined_args_to_skip, 0, index, bi)
2005 streamer_write_uhwi (ob, index);
2007 else
2008 streamer_write_uhwi (ob, 0);
2009 streamer_write_uhwi (ob, vec_safe_length (node->clone.tree_map));
2010 FOR_EACH_VEC_SAFE_ELT (node->clone.tree_map, i, map)
2012 /* At the moment we assume all old trees to be PARM_DECLs, because we have no
2013 mechanism to store function local declarations into summaries. */
2014 gcc_assert (!map->old_tree);
2015 streamer_write_uhwi (ob, map->parm_num);
2016 gcc_assert (EXPR_LOCATION (map->new_tree) == UNKNOWN_LOCATION);
2017 stream_write_tree (ob, map->new_tree, true);
2018 bp = bitpack_create (ob->main_stream);
2019 bp_pack_value (&bp, map->replace_p, 1);
2020 bp_pack_value (&bp, map->ref_p, 1);
2021 streamer_write_bitpack (&bp);
2024 if (lto_symtab_encoder_in_partition_p (encoder, node))
2026 for (e = node->callees; e; e = e->next_callee)
2027 output_edge_opt_summary (ob, e);
2028 for (e = node->indirect_calls; e; e = e->next_callee)
2029 output_edge_opt_summary (ob, e);
2033 /* Output optimization summaries stored in callgraph.
2034 At the moment it is the clone info structure. */
2036 static void
2037 output_cgraph_opt_summary (void)
2039 int i, n_nodes;
2040 lto_symtab_encoder_t encoder;
2041 struct output_block *ob = create_output_block (LTO_section_cgraph_opt_sum);
2042 unsigned count = 0;
2044 ob->symbol = NULL;
2045 encoder = ob->decl_state->symtab_node_encoder;
2046 n_nodes = lto_symtab_encoder_size (encoder);
2047 for (i = 0; i < n_nodes; i++)
2049 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
2050 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
2051 if (cnode && output_cgraph_opt_summary_p (cnode))
2052 count++;
2054 streamer_write_uhwi (ob, count);
2055 for (i = 0; i < n_nodes; i++)
2057 symtab_node *node = lto_symtab_encoder_deref (encoder, i);
2058 cgraph_node *cnode = dyn_cast <cgraph_node *> (node);
2059 if (cnode && output_cgraph_opt_summary_p (cnode))
2061 streamer_write_uhwi (ob, i);
2062 output_node_opt_summary (ob, cnode, encoder);
2065 produce_asm (ob, NULL);
2066 destroy_output_block (ob);
2069 /* Input optimisation summary of EDGE. */
2071 static void
2072 input_edge_opt_summary (struct cgraph_edge *edge ATTRIBUTE_UNUSED,
2073 struct lto_input_block *ib_main ATTRIBUTE_UNUSED)
2077 /* Input optimisation summary of NODE. */
2079 static void
2080 input_node_opt_summary (struct cgraph_node *node,
2081 struct lto_input_block *ib_main,
2082 struct data_in *data_in)
2084 int i;
2085 int count;
2086 int bit;
2087 struct bitpack_d bp;
2088 struct cgraph_edge *e;
2090 count = streamer_read_uhwi (ib_main);
2091 if (count)
2092 node->clone.args_to_skip = BITMAP_GGC_ALLOC ();
2093 for (i = 0; i < count; i++)
2095 bit = streamer_read_uhwi (ib_main);
2096 bitmap_set_bit (node->clone.args_to_skip, bit);
2098 count = streamer_read_uhwi (ib_main);
2099 if (count)
2100 node->clone.combined_args_to_skip = BITMAP_GGC_ALLOC ();
2101 for (i = 0; i < count; i++)
2103 bit = streamer_read_uhwi (ib_main);
2104 bitmap_set_bit (node->clone.combined_args_to_skip, bit);
2106 count = streamer_read_uhwi (ib_main);
2107 for (i = 0; i < count; i++)
2109 struct ipa_replace_map *map = ggc_alloc<ipa_replace_map> ();
2111 vec_safe_push (node->clone.tree_map, map);
2112 map->parm_num = streamer_read_uhwi (ib_main);
2113 map->old_tree = NULL;
2114 map->new_tree = stream_read_tree (ib_main, data_in);
2115 bp = streamer_read_bitpack (ib_main);
2116 map->replace_p = bp_unpack_value (&bp, 1);
2117 map->ref_p = bp_unpack_value (&bp, 1);
2119 for (e = node->callees; e; e = e->next_callee)
2120 input_edge_opt_summary (e, ib_main);
2121 for (e = node->indirect_calls; e; e = e->next_callee)
2122 input_edge_opt_summary (e, ib_main);
2125 /* Read section in file FILE_DATA of length LEN with data DATA. */
2127 static void
2128 input_cgraph_opt_section (struct lto_file_decl_data *file_data,
2129 const char *data, size_t len,
2130 vec<symtab_node *> nodes)
2132 const struct lto_function_header *header =
2133 (const struct lto_function_header *) data;
2134 const int cfg_offset = sizeof (struct lto_function_header);
2135 const int main_offset = cfg_offset + header->cfg_size;
2136 const int string_offset = main_offset + header->main_size;
2137 struct data_in *data_in;
2138 unsigned int i;
2139 unsigned int count;
2141 lto_input_block ib_main ((const char *) data + main_offset,
2142 header->main_size, file_data->mode_table);
2144 data_in =
2145 lto_data_in_create (file_data, (const char *) data + string_offset,
2146 header->string_size, vNULL);
2147 count = streamer_read_uhwi (&ib_main);
2149 for (i = 0; i < count; i++)
2151 int ref = streamer_read_uhwi (&ib_main);
2152 input_node_opt_summary (dyn_cast<cgraph_node *> (nodes[ref]),
2153 &ib_main, data_in);
2155 lto_free_section_data (file_data, LTO_section_cgraph_opt_sum, NULL, data,
2156 len);
2157 lto_data_in_delete (data_in);
2160 /* Input optimization summary of cgraph. */
2162 static void
2163 input_cgraph_opt_summary (vec<symtab_node *> nodes)
2165 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
2166 struct lto_file_decl_data *file_data;
2167 unsigned int j = 0;
2169 while ((file_data = file_data_vec[j++]))
2171 size_t len;
2172 const char *data =
2173 lto_get_section_data (file_data, LTO_section_cgraph_opt_sum, NULL,
2174 &len);
2176 if (data)
2177 input_cgraph_opt_section (file_data, data, len, nodes);