1 /* Write and read the cgraph to the memory mapped representation of a
4 Copyright (C) 2009-2015 Free Software Foundation, Inc.
5 Contributed by Kenneth Zadeck <zadeck@naturalbridge.com>
7 This file is part of GCC.
9 GCC is free software; you can redistribute it and/or modify it under
10 the terms of the GNU General Public License as published by the Free
11 Software Foundation; either version 3, or (at your option) any later
14 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
15 WARRANTY; without even the implied warranty of MERCHANTABILITY or
16 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
19 You should have received a copy of the GNU General Public License
20 along with GCC; see the file COPYING3. If not see
21 <http://www.gnu.org/licenses/>. */
25 #include "coretypes.h"
32 #include "fold-const.h"
33 #include "stringpool.h"
34 #include "internal-fn.h"
36 #include "insn-config.h"
46 #include "langhooks.h"
47 #include "diagnostic-core.h"
51 #include "tree-streamer.h"
53 #include "tree-pass.h"
56 #include "pass_manager.h"
57 #include "ipa-utils.h"
61 /* True when asm nodes has been output. */
62 bool asm_nodes_output
= false;
64 static void output_cgraph_opt_summary (void);
65 static void input_cgraph_opt_summary (vec
<symtab_node
*> nodes
);
67 /* Number of LDPR values known to GCC. */
68 #define LDPR_NUM_KNOWN (LDPR_PREVAILING_DEF_IRONLY_EXP + 1)
70 /* All node orders are ofsetted by ORDER_BASE. */
71 static int order_base
;
73 /* Cgraph streaming is organized as set of record whose type
74 is indicated by a tag. */
77 /* Must leave 0 for the stopper. */
79 /* Cgraph node without body available. */
80 LTO_symtab_unavail_node
= 1,
81 /* Cgraph node with function body. */
82 LTO_symtab_analyzed_node
,
85 LTO_symtab_indirect_edge
,
90 /* Create a new symtab encoder.
91 if FOR_INPUT, the encoder allocate only datastructures needed
92 to read the symtab. */
95 lto_symtab_encoder_new (bool for_input
)
97 lto_symtab_encoder_t encoder
= XCNEW (struct lto_symtab_encoder_d
);
100 encoder
->map
= new hash_map
<symtab_node
*, size_t>;
101 encoder
->nodes
.create (0);
106 /* Delete ENCODER and its components. */
109 lto_symtab_encoder_delete (lto_symtab_encoder_t encoder
)
111 encoder
->nodes
.release ();
118 /* Return the existing reference number of NODE in the symtab encoder in
119 output block OB. Assign a new reference if this is the first time
123 lto_symtab_encoder_encode (lto_symtab_encoder_t encoder
,
130 lto_encoder_entry entry
= {node
, false, false, false};
132 ref
= encoder
->nodes
.length ();
133 encoder
->nodes
.safe_push (entry
);
137 size_t *slot
= encoder
->map
->get (node
);
140 lto_encoder_entry entry
= {node
, false, false, false};
141 ref
= encoder
->nodes
.length ();
143 encoder
->map
->put (node
, ref
+ 1);
144 encoder
->nodes
.safe_push (entry
);
152 /* Remove NODE from encoder. */
155 lto_symtab_encoder_delete_node (lto_symtab_encoder_t encoder
,
159 lto_encoder_entry last_node
;
161 size_t *slot
= encoder
->map
->get (node
);
162 if (slot
== NULL
|| !*slot
)
166 gcc_checking_assert (encoder
->nodes
[index
].node
== node
);
168 /* Remove from vector. We do this by swapping node with the last element
170 last_node
= encoder
->nodes
.pop ();
171 if (last_node
.node
!= node
)
173 gcc_assert (encoder
->map
->put (last_node
.node
, index
+ 1));
175 /* Move the last element to the original spot of NODE. */
176 encoder
->nodes
[index
] = last_node
;
179 /* Remove element from hash table. */
180 encoder
->map
->remove (node
);
185 /* Return TRUE if we should encode the body of NODE (if any). */
188 lto_symtab_encoder_encode_body_p (lto_symtab_encoder_t encoder
,
189 struct cgraph_node
*node
)
191 int index
= lto_symtab_encoder_lookup (encoder
, node
);
192 return encoder
->nodes
[index
].body
;
195 /* Specify that we encode the body of NODE in this partition. */
198 lto_set_symtab_encoder_encode_body (lto_symtab_encoder_t encoder
,
199 struct cgraph_node
*node
)
201 int index
= lto_symtab_encoder_encode (encoder
, node
);
202 gcc_checking_assert (encoder
->nodes
[index
].node
== node
);
203 encoder
->nodes
[index
].body
= true;
206 /* Return TRUE if we should encode initializer of NODE (if any). */
209 lto_symtab_encoder_encode_initializer_p (lto_symtab_encoder_t encoder
,
212 int index
= lto_symtab_encoder_lookup (encoder
, node
);
213 if (index
== LCC_NOT_FOUND
)
215 return encoder
->nodes
[index
].initializer
;
218 /* Specify that we should encode initializer of NODE (if any). */
221 lto_set_symtab_encoder_encode_initializer (lto_symtab_encoder_t encoder
,
224 int index
= lto_symtab_encoder_lookup (encoder
, node
);
225 encoder
->nodes
[index
].initializer
= true;
228 /* Return TRUE if NODE is in this partition. */
231 lto_symtab_encoder_in_partition_p (lto_symtab_encoder_t encoder
,
234 int index
= lto_symtab_encoder_lookup (encoder
, node
);
235 if (index
== LCC_NOT_FOUND
)
237 return encoder
->nodes
[index
].in_partition
;
240 /* Specify that NODE is in this partition. */
243 lto_set_symtab_encoder_in_partition (lto_symtab_encoder_t encoder
,
246 int index
= lto_symtab_encoder_encode (encoder
, node
);
247 encoder
->nodes
[index
].in_partition
= true;
250 /* Output the cgraph EDGE to OB using ENCODER. */
253 lto_output_edge (struct lto_simple_output_block
*ob
, struct cgraph_edge
*edge
,
254 lto_symtab_encoder_t encoder
)
260 if (edge
->indirect_unknown_callee
)
261 streamer_write_enum (ob
->main_stream
, LTO_symtab_tags
, LTO_symtab_last_tag
,
262 LTO_symtab_indirect_edge
);
264 streamer_write_enum (ob
->main_stream
, LTO_symtab_tags
, LTO_symtab_last_tag
,
267 ref
= lto_symtab_encoder_lookup (encoder
, edge
->caller
);
268 gcc_assert (ref
!= LCC_NOT_FOUND
);
269 streamer_write_hwi_stream (ob
->main_stream
, ref
);
271 if (!edge
->indirect_unknown_callee
)
273 ref
= lto_symtab_encoder_lookup (encoder
, edge
->callee
);
274 gcc_assert (ref
!= LCC_NOT_FOUND
);
275 streamer_write_hwi_stream (ob
->main_stream
, ref
);
278 streamer_write_gcov_count_stream (ob
->main_stream
, edge
->count
);
280 bp
= bitpack_create (ob
->main_stream
);
281 uid
= (!gimple_has_body_p (edge
->caller
->decl
)
282 ? edge
->lto_stmt_uid
: gimple_uid (edge
->call_stmt
) + 1);
283 bp_pack_enum (&bp
, cgraph_inline_failed_t
,
284 CIF_N_REASONS
, edge
->inline_failed
);
285 bp_pack_var_len_unsigned (&bp
, uid
);
286 bp_pack_var_len_unsigned (&bp
, edge
->frequency
);
287 bp_pack_value (&bp
, edge
->indirect_inlining_edge
, 1);
288 bp_pack_value (&bp
, edge
->speculative
, 1);
289 bp_pack_value (&bp
, edge
->call_stmt_cannot_inline_p
, 1);
290 bp_pack_value (&bp
, edge
->can_throw_external
, 1);
291 bp_pack_value (&bp
, edge
->in_polymorphic_cdtor
, 1);
292 if (edge
->indirect_unknown_callee
)
294 int flags
= edge
->indirect_info
->ecf_flags
;
295 bp_pack_value (&bp
, (flags
& ECF_CONST
) != 0, 1);
296 bp_pack_value (&bp
, (flags
& ECF_PURE
) != 0, 1);
297 bp_pack_value (&bp
, (flags
& ECF_NORETURN
) != 0, 1);
298 bp_pack_value (&bp
, (flags
& ECF_MALLOC
) != 0, 1);
299 bp_pack_value (&bp
, (flags
& ECF_NOTHROW
) != 0, 1);
300 bp_pack_value (&bp
, (flags
& ECF_RETURNS_TWICE
) != 0, 1);
301 /* Flags that should not appear on indirect calls. */
302 gcc_assert (!(flags
& (ECF_LOOPING_CONST_OR_PURE
308 streamer_write_bitpack (&bp
);
309 if (edge
->indirect_unknown_callee
)
311 streamer_write_hwi_stream (ob
->main_stream
,
312 edge
->indirect_info
->common_target_id
);
313 if (edge
->indirect_info
->common_target_id
)
314 streamer_write_hwi_stream
315 (ob
->main_stream
, edge
->indirect_info
->common_target_probability
);
319 /* Return if NODE contain references from other partitions. */
322 referenced_from_other_partition_p (symtab_node
*node
, lto_symtab_encoder_t encoder
)
325 struct ipa_ref
*ref
= NULL
;
327 for (i
= 0; node
->iterate_referring (i
, ref
); i
++)
329 /* Ignore references from non-offloadable nodes while streaming NODE into
330 offload LTO section. */
331 if (!ref
->referring
->need_lto_streaming
)
334 if (ref
->referring
->in_other_partition
335 || !lto_symtab_encoder_in_partition_p (encoder
, ref
->referring
))
341 /* Return true when node is reachable from other partition. */
344 reachable_from_other_partition_p (struct cgraph_node
*node
, lto_symtab_encoder_t encoder
)
346 struct cgraph_edge
*e
;
347 if (!node
->definition
)
349 if (node
->global
.inlined_to
)
351 for (e
= node
->callers
; e
; e
= e
->next_caller
)
353 /* Ignore references from non-offloadable nodes while streaming NODE into
354 offload LTO section. */
355 if (!e
->caller
->need_lto_streaming
)
358 if (e
->caller
->in_other_partition
359 || !lto_symtab_encoder_in_partition_p (encoder
, e
->caller
))
365 /* Return if NODE contain references from other partitions. */
368 referenced_from_this_partition_p (symtab_node
*node
,
369 lto_symtab_encoder_t encoder
)
372 struct ipa_ref
*ref
= NULL
;
374 for (i
= 0; node
->iterate_referring (i
, ref
); i
++)
375 if (lto_symtab_encoder_in_partition_p (encoder
, ref
->referring
))
380 /* Return true when node is reachable from other partition. */
383 reachable_from_this_partition_p (struct cgraph_node
*node
, lto_symtab_encoder_t encoder
)
385 struct cgraph_edge
*e
;
386 for (e
= node
->callers
; e
; e
= e
->next_caller
)
387 if (lto_symtab_encoder_in_partition_p (encoder
, e
->caller
))
392 /* Output the cgraph NODE to OB. ENCODER is used to find the
393 reference number of NODE->inlined_to. SET is the set of nodes we
394 are writing to the current file. If NODE is not in SET, then NODE
395 is a boundary of a cgraph_node_set and we pretend NODE just has a
396 decl and no callees. WRITTEN_DECLS is the set of FUNCTION_DECLs
397 that have had their callgraph node written so far. This is used to
398 determine if NODE is a clone of a previously written node. */
401 lto_output_node (struct lto_simple_output_block
*ob
, struct cgraph_node
*node
,
402 lto_symtab_encoder_t encoder
)
408 bool in_other_partition
= false;
409 struct cgraph_node
*clone_of
, *ultimate_clone_of
;
410 ipa_opt_pass_d
*pass
;
416 boundary_p
= !lto_symtab_encoder_in_partition_p (encoder
, node
);
418 if (node
->analyzed
&& (!boundary_p
|| node
->alias
|| node
->thunk
.thunk_p
))
419 tag
= LTO_symtab_analyzed_node
;
421 tag
= LTO_symtab_unavail_node
;
423 streamer_write_enum (ob
->main_stream
, LTO_symtab_tags
, LTO_symtab_last_tag
,
425 streamer_write_hwi_stream (ob
->main_stream
, node
->order
);
427 /* In WPA mode, we only output part of the call-graph. Also, we
428 fake cgraph node attributes. There are two cases that we care.
430 Boundary nodes: There are nodes that are not part of SET but are
431 called from within SET. We artificially make them look like
432 externally visible nodes with no function body.
434 Cherry-picked nodes: These are nodes we pulled from other
435 translation units into SET during IPA-inlining. We make them as
436 local static nodes to prevent clashes with other local statics. */
437 if (boundary_p
&& node
->analyzed
438 && node
->get_partitioning_class () == SYMBOL_PARTITION
)
440 /* Inline clones can not be part of boundary.
441 gcc_assert (!node->global.inlined_to);
443 FIXME: At the moment they can be, when partition contains an inline
444 clone that is clone of inline clone from outside partition. We can
445 reshape the clone tree and make other tree to be the root, but it
446 needs a bit extra work and will be promplty done by cgraph_remove_node
447 after reading back. */
448 in_other_partition
= 1;
451 clone_of
= node
->clone_of
;
453 && (ref
= lto_symtab_encoder_lookup (encoder
, clone_of
)) == LCC_NOT_FOUND
)
454 if (clone_of
->prev_sibling_clone
)
455 clone_of
= clone_of
->prev_sibling_clone
;
457 clone_of
= clone_of
->clone_of
;
459 /* See if body of the master function is output. If not, we are seeing only
460 an declaration and we do not need to pass down clone tree. */
461 ultimate_clone_of
= clone_of
;
462 while (ultimate_clone_of
&& ultimate_clone_of
->clone_of
)
463 ultimate_clone_of
= ultimate_clone_of
->clone_of
;
465 if (clone_of
&& !lto_symtab_encoder_encode_body_p (encoder
, ultimate_clone_of
))
468 if (tag
== LTO_symtab_analyzed_node
)
469 gcc_assert (clone_of
|| !node
->clone_of
);
471 streamer_write_hwi_stream (ob
->main_stream
, LCC_NOT_FOUND
);
473 streamer_write_hwi_stream (ob
->main_stream
, ref
);
476 lto_output_fn_decl_index (ob
->decl_state
, ob
->main_stream
, node
->decl
);
477 streamer_write_gcov_count_stream (ob
->main_stream
, node
->count
);
478 streamer_write_hwi_stream (ob
->main_stream
, node
->count_materialization_scale
);
480 streamer_write_hwi_stream (ob
->main_stream
,
481 node
->ipa_transforms_to_apply
.length ());
482 FOR_EACH_VEC_ELT (node
->ipa_transforms_to_apply
, i
, pass
)
483 streamer_write_hwi_stream (ob
->main_stream
, pass
->static_pass_number
);
485 if (tag
== LTO_symtab_analyzed_node
)
487 if (node
->global
.inlined_to
)
489 ref
= lto_symtab_encoder_lookup (encoder
, node
->global
.inlined_to
);
490 gcc_assert (ref
!= LCC_NOT_FOUND
);
495 streamer_write_hwi_stream (ob
->main_stream
, ref
);
498 group
= node
->get_comdat_group ();
500 comdat
= IDENTIFIER_POINTER (group
);
503 streamer_write_data_stream (ob
->main_stream
, comdat
, strlen (comdat
) + 1);
507 if (node
->same_comdat_group
&& !boundary_p
)
509 ref
= lto_symtab_encoder_lookup (encoder
,
510 node
->same_comdat_group
);
511 gcc_assert (ref
!= LCC_NOT_FOUND
);
515 streamer_write_hwi_stream (ob
->main_stream
, ref
);
518 section
= node
->get_section ();
522 streamer_write_hwi_stream (ob
->main_stream
, node
->tp_first_run
);
524 bp
= bitpack_create (ob
->main_stream
);
525 bp_pack_value (&bp
, node
->local
.local
, 1);
526 bp_pack_value (&bp
, node
->externally_visible
, 1);
527 bp_pack_value (&bp
, node
->no_reorder
, 1);
528 bp_pack_value (&bp
, node
->definition
, 1);
529 bp_pack_value (&bp
, node
->local
.versionable
, 1);
530 bp_pack_value (&bp
, node
->local
.can_change_signature
, 1);
531 bp_pack_value (&bp
, node
->local
.redefined_extern_inline
, 1);
532 bp_pack_value (&bp
, node
->force_output
, 1);
533 bp_pack_value (&bp
, node
->forced_by_abi
, 1);
534 bp_pack_value (&bp
, node
->unique_name
, 1);
535 bp_pack_value (&bp
, node
->body_removed
, 1);
536 bp_pack_value (&bp
, node
->implicit_section
, 1);
537 bp_pack_value (&bp
, node
->address_taken
, 1);
538 bp_pack_value (&bp
, tag
== LTO_symtab_analyzed_node
539 && node
->get_partitioning_class () == SYMBOL_PARTITION
540 && (reachable_from_other_partition_p (node
, encoder
)
541 || referenced_from_other_partition_p (node
, encoder
)), 1);
542 bp_pack_value (&bp
, node
->lowered
, 1);
543 bp_pack_value (&bp
, in_other_partition
, 1);
544 bp_pack_value (&bp
, node
->alias
, 1);
545 bp_pack_value (&bp
, node
->weakref
, 1);
546 bp_pack_value (&bp
, node
->frequency
, 2);
547 bp_pack_value (&bp
, node
->only_called_at_startup
, 1);
548 bp_pack_value (&bp
, node
->only_called_at_exit
, 1);
549 bp_pack_value (&bp
, node
->tm_clone
, 1);
550 bp_pack_value (&bp
, node
->calls_comdat_local
, 1);
551 bp_pack_value (&bp
, node
->icf_merged
, 1);
552 bp_pack_value (&bp
, node
->nonfreeing_fn
, 1);
553 bp_pack_value (&bp
, node
->thunk
.thunk_p
, 1);
554 bp_pack_value (&bp
, node
->parallelized_function
, 1);
555 bp_pack_enum (&bp
, ld_plugin_symbol_resolution
,
556 LDPR_NUM_KNOWN
, node
->resolution
);
557 bp_pack_value (&bp
, node
->instrumentation_clone
, 1);
558 bp_pack_value (&bp
, node
->split_part
, 1);
559 streamer_write_bitpack (&bp
);
560 streamer_write_data_stream (ob
->main_stream
, section
, strlen (section
) + 1);
562 if (node
->thunk
.thunk_p
)
564 streamer_write_uhwi_stream
566 1 + (node
->thunk
.this_adjusting
!= 0) * 2
567 + (node
->thunk
.virtual_offset_p
!= 0) * 4
568 + (node
->thunk
.add_pointer_bounds_args
!= 0) * 8);
569 streamer_write_uhwi_stream (ob
->main_stream
, node
->thunk
.fixed_offset
);
570 streamer_write_uhwi_stream (ob
->main_stream
, node
->thunk
.virtual_value
);
572 streamer_write_hwi_stream (ob
->main_stream
, node
->profile_id
);
573 if (DECL_STATIC_CONSTRUCTOR (node
->decl
))
574 streamer_write_hwi_stream (ob
->main_stream
, node
->get_init_priority ());
575 if (DECL_STATIC_DESTRUCTOR (node
->decl
))
576 streamer_write_hwi_stream (ob
->main_stream
, node
->get_fini_priority ());
578 if (node
->instrumentation_clone
)
579 lto_output_fn_decl_index (ob
->decl_state
, ob
->main_stream
, node
->orig_decl
);
582 /* Output the varpool NODE to OB.
583 If NODE is not in SET, then NODE is a boundary. */
586 lto_output_varpool_node (struct lto_simple_output_block
*ob
, varpool_node
*node
,
587 lto_symtab_encoder_t encoder
)
589 bool boundary_p
= !lto_symtab_encoder_in_partition_p (encoder
, node
);
590 bool encode_initializer_p
592 && lto_symtab_encoder_encode_initializer_p (encoder
, node
));
599 gcc_assert (!encode_initializer_p
|| node
->definition
);
600 gcc_assert (boundary_p
|| encode_initializer_p
);
602 streamer_write_enum (ob
->main_stream
, LTO_symtab_tags
, LTO_symtab_last_tag
,
603 LTO_symtab_variable
);
604 streamer_write_hwi_stream (ob
->main_stream
, node
->order
);
605 lto_output_var_decl_index (ob
->decl_state
, ob
->main_stream
, node
->decl
);
606 bp
= bitpack_create (ob
->main_stream
);
607 bp_pack_value (&bp
, node
->externally_visible
, 1);
608 bp_pack_value (&bp
, node
->no_reorder
, 1);
609 bp_pack_value (&bp
, node
->force_output
, 1);
610 bp_pack_value (&bp
, node
->forced_by_abi
, 1);
611 bp_pack_value (&bp
, node
->unique_name
, 1);
614 || (!encode_initializer_p
&& !node
->alias
&& node
->definition
),
616 bp_pack_value (&bp
, node
->implicit_section
, 1);
617 bp_pack_value (&bp
, node
->writeonly
, 1);
618 bp_pack_value (&bp
, node
->definition
&& (encode_initializer_p
|| node
->alias
),
620 bp_pack_value (&bp
, node
->alias
, 1);
621 bp_pack_value (&bp
, node
->weakref
, 1);
622 bp_pack_value (&bp
, node
->analyzed
&& !boundary_p
, 1);
623 gcc_assert (node
->definition
|| !node
->analyzed
);
624 /* Constant pool initializers can be de-unified into individual ltrans units.
625 FIXME: Alternatively at -Os we may want to avoid generating for them the local
626 labels and share them across LTRANS partitions. */
627 if (node
->get_partitioning_class () != SYMBOL_PARTITION
)
629 bp_pack_value (&bp
, 0, 1); /* used_from_other_parition. */
630 bp_pack_value (&bp
, 0, 1); /* in_other_partition. */
634 bp_pack_value (&bp
, node
->definition
635 && referenced_from_other_partition_p (node
, encoder
), 1);
636 bp_pack_value (&bp
, node
->analyzed
637 && boundary_p
&& !DECL_EXTERNAL (node
->decl
), 1);
638 /* in_other_partition. */
640 bp_pack_value (&bp
, node
->tls_model
, 3);
641 bp_pack_value (&bp
, node
->used_by_single_function
, 1);
642 bp_pack_value (&bp
, node
->need_bounds_init
, 1);
643 streamer_write_bitpack (&bp
);
645 group
= node
->get_comdat_group ();
647 comdat
= IDENTIFIER_POINTER (group
);
650 streamer_write_data_stream (ob
->main_stream
, comdat
, strlen (comdat
) + 1);
654 if (node
->same_comdat_group
&& !boundary_p
)
656 ref
= lto_symtab_encoder_lookup (encoder
,
657 node
->same_comdat_group
);
658 gcc_assert (ref
!= LCC_NOT_FOUND
);
662 streamer_write_hwi_stream (ob
->main_stream
, ref
);
665 section
= node
->get_section ();
668 streamer_write_data_stream (ob
->main_stream
, section
, strlen (section
) + 1);
670 streamer_write_enum (ob
->main_stream
, ld_plugin_symbol_resolution
,
671 LDPR_NUM_KNOWN
, node
->resolution
);
674 /* Output the varpool NODE to OB.
675 If NODE is not in SET, then NODE is a boundary. */
678 lto_output_ref (struct lto_simple_output_block
*ob
, struct ipa_ref
*ref
,
679 lto_symtab_encoder_t encoder
)
683 int uid
= ref
->lto_stmt_uid
;
684 struct cgraph_node
*node
;
686 bp
= bitpack_create (ob
->main_stream
);
687 bp_pack_value (&bp
, ref
->use
, 3);
688 bp_pack_value (&bp
, ref
->speculative
, 1);
689 streamer_write_bitpack (&bp
);
690 nref
= lto_symtab_encoder_lookup (encoder
, ref
->referred
);
691 gcc_assert (nref
!= LCC_NOT_FOUND
);
692 streamer_write_hwi_stream (ob
->main_stream
, nref
);
694 node
= dyn_cast
<cgraph_node
*> (ref
->referring
);
698 uid
= gimple_uid (ref
->stmt
) + 1;
699 streamer_write_hwi_stream (ob
->main_stream
, uid
);
703 /* Stream out profile_summary to OB. */
706 output_profile_summary (struct lto_simple_output_block
*ob
)
713 /* We do not output num and run_max, they are not used by
714 GCC profile feedback and they are difficult to merge from multiple
716 gcc_assert (profile_info
->runs
);
717 streamer_write_uhwi_stream (ob
->main_stream
, profile_info
->runs
);
718 streamer_write_gcov_count_stream (ob
->main_stream
, profile_info
->sum_max
);
720 /* sum_all is needed for computing the working set with the
722 streamer_write_gcov_count_stream (ob
->main_stream
, profile_info
->sum_all
);
724 /* Create and output a bitpack of non-zero histogram entries indices. */
725 bp
= bitpack_create (ob
->main_stream
);
726 for (h_ix
= 0; h_ix
< GCOV_HISTOGRAM_SIZE
; h_ix
++)
727 bp_pack_value (&bp
, profile_info
->histogram
[h_ix
].num_counters
> 0, 1);
728 streamer_write_bitpack (&bp
);
729 /* Now stream out only those non-zero entries. */
730 for (h_ix
= 0; h_ix
< GCOV_HISTOGRAM_SIZE
; h_ix
++)
732 if (!profile_info
->histogram
[h_ix
].num_counters
)
734 streamer_write_gcov_count_stream (ob
->main_stream
,
735 profile_info
->histogram
[h_ix
].num_counters
);
736 streamer_write_gcov_count_stream (ob
->main_stream
,
737 profile_info
->histogram
[h_ix
].min_value
);
738 streamer_write_gcov_count_stream (ob
->main_stream
,
739 profile_info
->histogram
[h_ix
].cum_value
);
741 /* IPA-profile computes hot bb threshold based on cumulated
742 whole program profile. We need to stream it down to ltrans. */
744 streamer_write_gcov_count_stream (ob
->main_stream
,
745 get_hot_bb_threshold ());
748 streamer_write_uhwi_stream (ob
->main_stream
, 0);
751 /* Output all callees or indirect outgoing edges. EDGE must be the first such
755 output_outgoing_cgraph_edges (struct cgraph_edge
*edge
,
756 struct lto_simple_output_block
*ob
,
757 lto_symtab_encoder_t encoder
)
762 /* Output edges in backward direction, so the reconstructed callgraph match
763 and it is easy to associate call sites in the IPA pass summaries. */
764 while (edge
->next_callee
)
765 edge
= edge
->next_callee
;
766 for (; edge
; edge
= edge
->prev_callee
)
767 lto_output_edge (ob
, edge
, encoder
);
770 /* Output the part of the cgraph in SET. */
773 output_refs (lto_symtab_encoder_t encoder
)
775 struct lto_simple_output_block
*ob
;
779 ob
= lto_create_simple_output_block (LTO_section_refs
);
781 for (int i
= 0; i
< lto_symtab_encoder_size (encoder
); i
++)
783 symtab_node
*node
= lto_symtab_encoder_deref (encoder
, i
);
785 /* IPA_REF_ALIAS and IPA_REF_CHKP references are always preserved
786 in the boundary. Alias node can't have other references and
787 can be always handled as if it's not in the boundary. */
788 if (!node
->alias
&& !lto_symtab_encoder_in_partition_p (encoder
, node
))
790 cgraph_node
*cnode
= dyn_cast
<cgraph_node
*> (node
);
791 /* Output IPA_REF_CHKP reference. */
793 && cnode
->instrumented_version
794 && !cnode
->instrumentation_clone
)
796 for (int i
= 0; node
->iterate_reference (i
, ref
); i
++)
797 if (ref
->use
== IPA_REF_CHKP
)
799 if (lto_symtab_encoder_lookup (encoder
, ref
->referred
)
802 int nref
= lto_symtab_encoder_lookup (encoder
, node
);
803 streamer_write_gcov_count_stream (ob
->main_stream
, 1);
804 streamer_write_uhwi_stream (ob
->main_stream
, nref
);
805 lto_output_ref (ob
, ref
, encoder
);
813 count
= node
->ref_list
.nreferences ();
816 streamer_write_gcov_count_stream (ob
->main_stream
, count
);
817 streamer_write_uhwi_stream (ob
->main_stream
,
818 lto_symtab_encoder_lookup (encoder
, node
));
819 for (int i
= 0; node
->iterate_reference (i
, ref
); i
++)
820 lto_output_ref (ob
, ref
, encoder
);
824 streamer_write_uhwi_stream (ob
->main_stream
, 0);
826 lto_destroy_simple_output_block (ob
);
829 /* Add NODE into encoder as well as nodes it is cloned from.
830 Do it in a way so clones appear first. */
833 add_node_to (lto_symtab_encoder_t encoder
, struct cgraph_node
*node
,
837 add_node_to (encoder
, node
->clone_of
, include_body
);
838 else if (include_body
)
839 lto_set_symtab_encoder_encode_body (encoder
, node
);
840 lto_symtab_encoder_encode (encoder
, node
);
843 /* Add all references in NODE to encoders. */
846 create_references (lto_symtab_encoder_t encoder
, symtab_node
*node
)
849 struct ipa_ref
*ref
= NULL
;
850 for (i
= 0; node
->iterate_reference (i
, ref
); i
++)
851 if (is_a
<cgraph_node
*> (ref
->referred
))
852 add_node_to (encoder
, dyn_cast
<cgraph_node
*> (ref
->referred
), false);
854 lto_symtab_encoder_encode (encoder
, ref
->referred
);
857 /* Select what needs to be streamed out. In regular lto mode stream everything.
858 In offload lto mode stream only nodes marked as offloadable. */
860 select_what_to_stream (void)
862 struct symtab_node
*snode
;
863 FOR_EACH_SYMBOL (snode
)
864 snode
->need_lto_streaming
= !lto_stream_offload_p
|| snode
->offloadable
;
867 /* Find all symbols we want to stream into given partition and insert them
870 The function actually replaces IN_ENCODER by new one. The reason is that
871 streaming code needs clone's origin to be streamed before clone. This
872 means that we need to insert the nodes in specific order. This order is
873 ignored by the partitioning logic earlier. */
876 compute_ltrans_boundary (lto_symtab_encoder_t in_encoder
)
878 struct cgraph_edge
*edge
;
880 lto_symtab_encoder_t encoder
;
881 lto_symtab_encoder_iterator lsei
;
882 hash_set
<void *> reachable_call_targets
;
884 encoder
= lto_symtab_encoder_new (false);
886 /* Go over all entries in the IN_ENCODER and duplicate them to
887 ENCODER. At the same time insert masters of clones so
888 every master appears before clone. */
889 for (lsei
= lsei_start_function_in_partition (in_encoder
);
890 !lsei_end_p (lsei
); lsei_next_function_in_partition (&lsei
))
892 struct cgraph_node
*node
= lsei_cgraph_node (lsei
);
893 if (!node
->need_lto_streaming
)
895 add_node_to (encoder
, node
, true);
896 lto_set_symtab_encoder_in_partition (encoder
, node
);
897 create_references (encoder
, node
);
898 /* For proper debug info, we need to ship the origins, too. */
899 if (DECL_ABSTRACT_ORIGIN (node
->decl
))
901 struct cgraph_node
*origin_node
902 = cgraph_node::get_create (DECL_ABSTRACT_ORIGIN (node
->decl
));
903 origin_node
->used_as_abstract_origin
= true;
904 add_node_to (encoder
, origin_node
, true);
907 for (lsei
= lsei_start_variable_in_partition (in_encoder
);
908 !lsei_end_p (lsei
); lsei_next_variable_in_partition (&lsei
))
910 varpool_node
*vnode
= lsei_varpool_node (lsei
);
912 if (!vnode
->need_lto_streaming
)
914 lto_set_symtab_encoder_in_partition (encoder
, vnode
);
915 lto_set_symtab_encoder_encode_initializer (encoder
, vnode
);
916 create_references (encoder
, vnode
);
917 /* For proper debug info, we need to ship the origins, too. */
918 if (DECL_ABSTRACT_ORIGIN (vnode
->decl
))
920 varpool_node
*origin_node
921 = varpool_node::get (DECL_ABSTRACT_ORIGIN (vnode
->decl
));
922 lto_set_symtab_encoder_in_partition (encoder
, origin_node
);
925 /* Pickle in also the initializer of all referenced readonly variables
926 to help folding. Constant pool variables are not shared, so we must
928 for (i
= 0; i
< lto_symtab_encoder_size (encoder
); i
++)
930 symtab_node
*node
= lto_symtab_encoder_deref (encoder
, i
);
931 if (varpool_node
*vnode
= dyn_cast
<varpool_node
*> (node
))
933 if (!lto_symtab_encoder_encode_initializer_p (encoder
,
935 && (((vnode
->ctor_useable_for_folding_p ()
936 && (!DECL_VIRTUAL_P (vnode
->decl
)
938 || flag_ltrans_devirtualize
))
939 || POINTER_BOUNDS_P (vnode
->decl
))))
941 lto_set_symtab_encoder_encode_initializer (encoder
, vnode
);
942 create_references (encoder
, vnode
);
947 /* Go over all the nodes again to include callees that are not in
949 for (lsei
= lsei_start_function_in_partition (encoder
);
950 !lsei_end_p (lsei
); lsei_next_function_in_partition (&lsei
))
952 struct cgraph_node
*node
= lsei_cgraph_node (lsei
);
953 for (edge
= node
->callees
; edge
; edge
= edge
->next_callee
)
955 struct cgraph_node
*callee
= edge
->callee
;
956 if (!lto_symtab_encoder_in_partition_p (encoder
, callee
))
958 /* We should have moved all the inlines. */
959 gcc_assert (!callee
->global
.inlined_to
);
960 add_node_to (encoder
, callee
, false);
963 /* Add all possible targets for late devirtualization. */
964 if (flag_ltrans_devirtualize
|| !flag_wpa
)
965 for (edge
= node
->indirect_calls
; edge
; edge
= edge
->next_callee
)
966 if (edge
->indirect_info
->polymorphic
)
971 vec
<cgraph_node
*>targets
972 = possible_polymorphic_call_targets
973 (edge
, &final
, &cache_token
);
974 if (!reachable_call_targets
.add (cache_token
))
976 for (i
= 0; i
< targets
.length (); i
++)
978 struct cgraph_node
*callee
= targets
[i
];
980 /* Adding an external declarations into the unit serves
981 no purpose and just increases its boundary. */
982 if (callee
->definition
983 && !lto_symtab_encoder_in_partition_p
986 gcc_assert (!callee
->global
.inlined_to
);
987 add_node_to (encoder
, callee
, false);
993 /* Be sure to also insert alias targert and thunk callees. These needs
994 to stay to aid local calling conventions. */
995 for (i
= 0; i
< lto_symtab_encoder_size (encoder
); i
++)
997 symtab_node
*node
= lto_symtab_encoder_deref (encoder
, i
);
998 cgraph_node
*cnode
= dyn_cast
<cgraph_node
*> (node
);
1000 if (node
->alias
&& node
->analyzed
)
1001 create_references (encoder
, node
);
1003 && cnode
->thunk
.thunk_p
)
1004 add_node_to (encoder
, cnode
->callees
->callee
, false);
1006 lto_symtab_encoder_delete (in_encoder
);
1010 /* Output the part of the symtab in SET and VSET. */
1013 output_symtab (void)
1015 struct cgraph_node
*node
;
1016 struct lto_simple_output_block
*ob
;
1018 lto_symtab_encoder_t encoder
;
1021 output_cgraph_opt_summary ();
1023 ob
= lto_create_simple_output_block (LTO_section_symtab_nodes
);
1025 output_profile_summary (ob
);
1027 /* An encoder for cgraph nodes should have been created by
1028 ipa_write_summaries_1. */
1029 gcc_assert (ob
->decl_state
->symtab_node_encoder
);
1030 encoder
= ob
->decl_state
->symtab_node_encoder
;
1032 /* Write out the nodes. We must first output a node and then its clones,
1033 otherwise at a time reading back the node there would be nothing to clone
1035 n_nodes
= lto_symtab_encoder_size (encoder
);
1036 for (i
= 0; i
< n_nodes
; i
++)
1038 symtab_node
*node
= lto_symtab_encoder_deref (encoder
, i
);
1039 if (cgraph_node
*cnode
= dyn_cast
<cgraph_node
*> (node
))
1040 lto_output_node (ob
, cnode
, encoder
);
1042 lto_output_varpool_node (ob
, dyn_cast
<varpool_node
*> (node
), encoder
);
1045 /* Go over the nodes in SET again to write edges. */
1046 for (int i
= 0; i
< lto_symtab_encoder_size (encoder
); i
++)
1048 node
= dyn_cast
<cgraph_node
*> (lto_symtab_encoder_deref (encoder
, i
));
1050 && (node
->thunk
.thunk_p
1051 || lto_symtab_encoder_in_partition_p (encoder
, node
)))
1053 output_outgoing_cgraph_edges (node
->callees
, ob
, encoder
);
1054 output_outgoing_cgraph_edges (node
->indirect_calls
, ob
, encoder
);
1058 streamer_write_uhwi_stream (ob
->main_stream
, 0);
1060 lto_destroy_simple_output_block (ob
);
1062 /* Emit toplevel asms.
1063 When doing WPA we must output every asm just once. Since we do not partition asm
1064 nodes at all, output them to first output. This is kind of hack, but should work
1066 if (!asm_nodes_output
)
1068 asm_nodes_output
= true;
1069 lto_output_toplevel_asms ();
1072 output_refs (encoder
);
1075 /* Return identifier encoded in IB as a plain string. */
1078 read_identifier (struct lto_input_block
*ib
)
1080 unsigned int len
= strnlen (ib
->data
+ ib
->p
, ib
->len
- ib
->p
- 1);
1083 if (ib
->data
[ib
->p
+ len
])
1084 lto_section_overrun (ib
);
1090 id
= get_identifier (ib
->data
+ ib
->p
);
1095 /* Return string encoded in IB, NULL if string is empty. */
1098 read_string (struct lto_input_block
*ib
)
1100 unsigned int len
= strnlen (ib
->data
+ ib
->p
, ib
->len
- ib
->p
- 1);
1103 if (ib
->data
[ib
->p
+ len
])
1104 lto_section_overrun (ib
);
1110 str
= ib
->data
+ ib
->p
;
1115 /* Output function/variable tables that will allow libgomp to look up offload
1117 OFFLOAD_FUNCS is filled in expand_omp_target, OFFLOAD_VARS is filled in
1118 varpool_node::get_create. In WHOPR (partitioned) mode during the WPA stage
1119 both OFFLOAD_FUNCS and OFFLOAD_VARS are filled by input_offload_tables. */
1122 output_offload_tables (void)
1124 if (vec_safe_is_empty (offload_funcs
) && vec_safe_is_empty (offload_vars
))
1127 struct lto_simple_output_block
*ob
1128 = lto_create_simple_output_block (LTO_section_offload_table
);
1130 for (unsigned i
= 0; i
< vec_safe_length (offload_funcs
); i
++)
1132 streamer_write_enum (ob
->main_stream
, LTO_symtab_tags
,
1133 LTO_symtab_last_tag
, LTO_symtab_unavail_node
);
1134 lto_output_fn_decl_index (ob
->decl_state
, ob
->main_stream
,
1135 (*offload_funcs
)[i
]);
1138 for (unsigned i
= 0; i
< vec_safe_length (offload_vars
); i
++)
1140 streamer_write_enum (ob
->main_stream
, LTO_symtab_tags
,
1141 LTO_symtab_last_tag
, LTO_symtab_variable
);
1142 lto_output_var_decl_index (ob
->decl_state
, ob
->main_stream
,
1143 (*offload_vars
)[i
]);
1146 streamer_write_uhwi_stream (ob
->main_stream
, 0);
1147 lto_destroy_simple_output_block (ob
);
1149 /* In WHOPR mode during the WPA stage the joint offload tables need to be
1150 streamed to one partition only. That's why we free offload_funcs and
1151 offload_vars after the first call of output_offload_tables. */
1154 vec_free (offload_funcs
);
1155 vec_free (offload_vars
);
1159 /* Overwrite the information in NODE based on FILE_DATA, TAG, FLAGS,
1160 STACK_SIZE, SELF_TIME and SELF_SIZE. This is called either to initialize
1161 NODE or to replace the values in it, for instance because the first
1162 time we saw it, the function body was not available but now it
1163 is. BP is a bitpack with all the bitflags for NODE read from the
1167 input_overwrite_node (struct lto_file_decl_data
*file_data
,
1168 struct cgraph_node
*node
,
1169 enum LTO_symtab_tags tag
,
1170 struct bitpack_d
*bp
)
1172 node
->aux
= (void *) tag
;
1173 node
->lto_file_data
= file_data
;
1175 node
->local
.local
= bp_unpack_value (bp
, 1);
1176 node
->externally_visible
= bp_unpack_value (bp
, 1);
1177 node
->no_reorder
= bp_unpack_value (bp
, 1);
1178 node
->definition
= bp_unpack_value (bp
, 1);
1179 node
->local
.versionable
= bp_unpack_value (bp
, 1);
1180 node
->local
.can_change_signature
= bp_unpack_value (bp
, 1);
1181 node
->local
.redefined_extern_inline
= bp_unpack_value (bp
, 1);
1182 node
->force_output
= bp_unpack_value (bp
, 1);
1183 node
->forced_by_abi
= bp_unpack_value (bp
, 1);
1184 node
->unique_name
= bp_unpack_value (bp
, 1);
1185 node
->body_removed
= bp_unpack_value (bp
, 1);
1186 node
->implicit_section
= bp_unpack_value (bp
, 1);
1187 node
->address_taken
= bp_unpack_value (bp
, 1);
1188 node
->used_from_other_partition
= bp_unpack_value (bp
, 1);
1189 node
->lowered
= bp_unpack_value (bp
, 1);
1190 node
->analyzed
= tag
== LTO_symtab_analyzed_node
;
1191 node
->in_other_partition
= bp_unpack_value (bp
, 1);
1192 if (node
->in_other_partition
1193 /* Avoid updating decl when we are seeing just inline clone.
1194 When inlining function that has functions already inlined into it,
1195 we produce clones of inline clones.
1197 WPA partitioning might put each clone into different unit and
1198 we might end up streaming inline clone from other partition
1199 to support clone we are interested in. */
1201 || node
->clone_of
->decl
!= node
->decl
))
1203 DECL_EXTERNAL (node
->decl
) = 1;
1204 TREE_STATIC (node
->decl
) = 0;
1206 node
->alias
= bp_unpack_value (bp
, 1);
1207 node
->weakref
= bp_unpack_value (bp
, 1);
1208 node
->frequency
= (enum node_frequency
)bp_unpack_value (bp
, 2);
1209 node
->only_called_at_startup
= bp_unpack_value (bp
, 1);
1210 node
->only_called_at_exit
= bp_unpack_value (bp
, 1);
1211 node
->tm_clone
= bp_unpack_value (bp
, 1);
1212 node
->calls_comdat_local
= bp_unpack_value (bp
, 1);
1213 node
->icf_merged
= bp_unpack_value (bp
, 1);
1214 node
->nonfreeing_fn
= bp_unpack_value (bp
, 1);
1215 node
->thunk
.thunk_p
= bp_unpack_value (bp
, 1);
1216 node
->parallelized_function
= bp_unpack_value (bp
, 1);
1217 node
->resolution
= bp_unpack_enum (bp
, ld_plugin_symbol_resolution
,
1219 node
->instrumentation_clone
= bp_unpack_value (bp
, 1);
1220 node
->split_part
= bp_unpack_value (bp
, 1);
1221 gcc_assert (flag_ltrans
1222 || (!node
->in_other_partition
1223 && !node
->used_from_other_partition
));
1226 /* Return string alias is alias of. */
1229 get_alias_symbol (tree decl
)
1231 tree alias
= lookup_attribute ("alias", DECL_ATTRIBUTES (decl
));
1232 return get_identifier (TREE_STRING_POINTER
1233 (TREE_VALUE (TREE_VALUE (alias
))));
1236 /* Read a node from input_block IB. TAG is the node's tag just read.
1237 Return the node read or overwriten. */
1239 static struct cgraph_node
*
1240 input_node (struct lto_file_decl_data
*file_data
,
1241 struct lto_input_block
*ib
,
1242 enum LTO_symtab_tags tag
,
1243 vec
<symtab_node
*> nodes
)
1245 gcc::pass_manager
*passes
= g
->get_passes ();
1247 struct cgraph_node
*node
;
1248 struct bitpack_d bp
;
1249 unsigned decl_index
;
1250 int ref
= LCC_NOT_FOUND
, ref2
= LCC_NOT_FOUND
;
1255 const char *section
;
1256 order
= streamer_read_hwi (ib
) + order_base
;
1257 clone_ref
= streamer_read_hwi (ib
);
1259 decl_index
= streamer_read_uhwi (ib
);
1260 fn_decl
= lto_file_decl_data_get_fn_decl (file_data
, decl_index
);
1262 if (clone_ref
!= LCC_NOT_FOUND
)
1264 node
= dyn_cast
<cgraph_node
*> (nodes
[clone_ref
])->create_clone (fn_decl
,
1265 0, CGRAPH_FREQ_BASE
, false,
1266 vNULL
, false, NULL
, NULL
);
1270 /* Declaration of functions can be already merged with a declaration
1271 from other input file. We keep cgraph unmerged until after streaming
1272 of ipa passes is done. Alays forcingly create a fresh node. */
1273 node
= symtab
->create_empty ();
1274 node
->decl
= fn_decl
;
1275 node
->register_symbol ();
1278 node
->order
= order
;
1279 if (order
>= symtab
->order
)
1280 symtab
->order
= order
+ 1;
1282 node
->count
= streamer_read_gcov_count (ib
);
1283 node
->count_materialization_scale
= streamer_read_hwi (ib
);
1285 count
= streamer_read_hwi (ib
);
1286 node
->ipa_transforms_to_apply
= vNULL
;
1287 for (i
= 0; i
< count
; i
++)
1290 int pid
= streamer_read_hwi (ib
);
1292 gcc_assert (pid
< passes
->passes_by_id_size
);
1293 pass
= passes
->passes_by_id
[pid
];
1294 node
->ipa_transforms_to_apply
.safe_push ((ipa_opt_pass_d
*) pass
);
1297 if (tag
== LTO_symtab_analyzed_node
)
1298 ref
= streamer_read_hwi (ib
);
1300 group
= read_identifier (ib
);
1302 ref2
= streamer_read_hwi (ib
);
1304 /* Make sure that we have not read this node before. Nodes that
1305 have already been read will have their tag stored in the 'aux'
1306 field. Since built-in functions can be referenced in multiple
1307 functions, they are expected to be read more than once. */
1308 if (node
->aux
&& !DECL_BUILT_IN (node
->decl
))
1309 internal_error ("bytecode stream: found multiple instances of cgraph "
1310 "node with uid %d", node
->uid
);
1312 node
->tp_first_run
= streamer_read_uhwi (ib
);
1314 bp
= streamer_read_bitpack (ib
);
1316 input_overwrite_node (file_data
, node
, tag
, &bp
);
1318 /* Store a reference for now, and fix up later to be a pointer. */
1319 node
->global
.inlined_to
= (cgraph_node
*) (intptr_t) ref
;
1323 node
->set_comdat_group (group
);
1324 /* Store a reference for now, and fix up later to be a pointer. */
1325 node
->same_comdat_group
= (symtab_node
*) (intptr_t) ref2
;
1328 node
->same_comdat_group
= (symtab_node
*) (intptr_t) LCC_NOT_FOUND
;
1329 section
= read_string (ib
);
1331 node
->set_section_for_node (section
);
1333 if (node
->thunk
.thunk_p
)
1335 int type
= streamer_read_uhwi (ib
);
1336 HOST_WIDE_INT fixed_offset
= streamer_read_uhwi (ib
);
1337 HOST_WIDE_INT virtual_value
= streamer_read_uhwi (ib
);
1339 node
->thunk
.fixed_offset
= fixed_offset
;
1340 node
->thunk
.this_adjusting
= (type
& 2);
1341 node
->thunk
.virtual_value
= virtual_value
;
1342 node
->thunk
.virtual_offset_p
= (type
& 4);
1343 node
->thunk
.add_pointer_bounds_args
= (type
& 8);
1345 if (node
->alias
&& !node
->analyzed
&& node
->weakref
)
1346 node
->alias_target
= get_alias_symbol (node
->decl
);
1347 node
->profile_id
= streamer_read_hwi (ib
);
1348 if (DECL_STATIC_CONSTRUCTOR (node
->decl
))
1349 node
->set_init_priority (streamer_read_hwi (ib
));
1350 if (DECL_STATIC_DESTRUCTOR (node
->decl
))
1351 node
->set_fini_priority (streamer_read_hwi (ib
));
1353 if (node
->instrumentation_clone
)
1355 decl_index
= streamer_read_uhwi (ib
);
1356 fn_decl
= lto_file_decl_data_get_fn_decl (file_data
, decl_index
);
1357 node
->orig_decl
= fn_decl
;
1363 /* Read a node from input_block IB. TAG is the node's tag just read.
1364 Return the node read or overwriten. */
1366 static varpool_node
*
1367 input_varpool_node (struct lto_file_decl_data
*file_data
,
1368 struct lto_input_block
*ib
)
1373 struct bitpack_d bp
;
1374 int ref
= LCC_NOT_FOUND
;
1377 const char *section
;
1379 order
= streamer_read_hwi (ib
) + order_base
;
1380 decl_index
= streamer_read_uhwi (ib
);
1381 var_decl
= lto_file_decl_data_get_var_decl (file_data
, decl_index
);
1383 /* Declaration of functions can be already merged with a declaration
1384 from other input file. We keep cgraph unmerged until after streaming
1385 of ipa passes is done. Alays forcingly create a fresh node. */
1386 node
= varpool_node::create_empty ();
1387 node
->decl
= var_decl
;
1388 node
->register_symbol ();
1390 node
->order
= order
;
1391 if (order
>= symtab
->order
)
1392 symtab
->order
= order
+ 1;
1393 node
->lto_file_data
= file_data
;
1395 bp
= streamer_read_bitpack (ib
);
1396 node
->externally_visible
= bp_unpack_value (&bp
, 1);
1397 node
->no_reorder
= bp_unpack_value (&bp
, 1);
1398 node
->force_output
= bp_unpack_value (&bp
, 1);
1399 node
->forced_by_abi
= bp_unpack_value (&bp
, 1);
1400 node
->unique_name
= bp_unpack_value (&bp
, 1);
1401 node
->body_removed
= bp_unpack_value (&bp
, 1);
1402 node
->implicit_section
= bp_unpack_value (&bp
, 1);
1403 node
->writeonly
= bp_unpack_value (&bp
, 1);
1404 node
->definition
= bp_unpack_value (&bp
, 1);
1405 node
->alias
= bp_unpack_value (&bp
, 1);
1406 node
->weakref
= bp_unpack_value (&bp
, 1);
1407 node
->analyzed
= bp_unpack_value (&bp
, 1);
1408 node
->used_from_other_partition
= bp_unpack_value (&bp
, 1);
1409 node
->in_other_partition
= bp_unpack_value (&bp
, 1);
1410 if (node
->in_other_partition
)
1412 DECL_EXTERNAL (node
->decl
) = 1;
1413 TREE_STATIC (node
->decl
) = 0;
1415 if (node
->alias
&& !node
->analyzed
&& node
->weakref
)
1416 node
->alias_target
= get_alias_symbol (node
->decl
);
1417 node
->tls_model
= (enum tls_model
)bp_unpack_value (&bp
, 3);
1418 node
->used_by_single_function
= (enum tls_model
)bp_unpack_value (&bp
, 1);
1419 node
->need_bounds_init
= bp_unpack_value (&bp
, 1);
1420 group
= read_identifier (ib
);
1423 node
->set_comdat_group (group
);
1424 ref
= streamer_read_hwi (ib
);
1425 /* Store a reference for now, and fix up later to be a pointer. */
1426 node
->same_comdat_group
= (symtab_node
*) (intptr_t) ref
;
1429 node
->same_comdat_group
= (symtab_node
*) (intptr_t) LCC_NOT_FOUND
;
1430 section
= read_string (ib
);
1432 node
->set_section_for_node (section
);
1433 node
->resolution
= streamer_read_enum (ib
, ld_plugin_symbol_resolution
,
1435 gcc_assert (flag_ltrans
1436 || (!node
->in_other_partition
1437 && !node
->used_from_other_partition
));
1442 /* Read a node from input_block IB. TAG is the node's tag just read.
1443 Return the node read or overwriten. */
1446 input_ref (struct lto_input_block
*ib
,
1447 symtab_node
*referring_node
,
1448 vec
<symtab_node
*> nodes
)
1450 symtab_node
*node
= NULL
;
1451 struct bitpack_d bp
;
1452 enum ipa_ref_use use
;
1454 struct ipa_ref
*ref
;
1456 bp
= streamer_read_bitpack (ib
);
1457 use
= (enum ipa_ref_use
) bp_unpack_value (&bp
, 3);
1458 speculative
= (enum ipa_ref_use
) bp_unpack_value (&bp
, 1);
1459 node
= nodes
[streamer_read_hwi (ib
)];
1460 ref
= referring_node
->create_reference (node
, use
);
1461 ref
->speculative
= speculative
;
1462 if (is_a
<cgraph_node
*> (referring_node
))
1463 ref
->lto_stmt_uid
= streamer_read_hwi (ib
);
1466 /* Read an edge from IB. NODES points to a vector of previously read nodes for
1467 decoding caller and callee of the edge to be read. If INDIRECT is true, the
1468 edge being read is indirect (in the sense that it has
1469 indirect_unknown_callee set). */
1472 input_edge (struct lto_input_block
*ib
, vec
<symtab_node
*> nodes
,
1475 struct cgraph_node
*caller
, *callee
;
1476 struct cgraph_edge
*edge
;
1477 unsigned int stmt_id
;
1480 cgraph_inline_failed_t inline_failed
;
1481 struct bitpack_d bp
;
1484 caller
= dyn_cast
<cgraph_node
*> (nodes
[streamer_read_hwi (ib
)]);
1485 if (caller
== NULL
|| caller
->decl
== NULL_TREE
)
1486 internal_error ("bytecode stream: no caller found while reading edge");
1490 callee
= dyn_cast
<cgraph_node
*> (nodes
[streamer_read_hwi (ib
)]);
1491 if (callee
== NULL
|| callee
->decl
== NULL_TREE
)
1492 internal_error ("bytecode stream: no callee found while reading edge");
1497 count
= streamer_read_gcov_count (ib
);
1499 bp
= streamer_read_bitpack (ib
);
1500 inline_failed
= bp_unpack_enum (&bp
, cgraph_inline_failed_t
, CIF_N_REASONS
);
1501 stmt_id
= bp_unpack_var_len_unsigned (&bp
);
1502 freq
= (int) bp_unpack_var_len_unsigned (&bp
);
1505 edge
= caller
->create_indirect_edge (NULL
, 0, count
, freq
);
1507 edge
= caller
->create_edge (callee
, NULL
, count
, freq
);
1509 edge
->indirect_inlining_edge
= bp_unpack_value (&bp
, 1);
1510 edge
->speculative
= bp_unpack_value (&bp
, 1);
1511 edge
->lto_stmt_uid
= stmt_id
;
1512 edge
->inline_failed
= inline_failed
;
1513 edge
->call_stmt_cannot_inline_p
= bp_unpack_value (&bp
, 1);
1514 edge
->can_throw_external
= bp_unpack_value (&bp
, 1);
1515 edge
->in_polymorphic_cdtor
= bp_unpack_value (&bp
, 1);
1518 if (bp_unpack_value (&bp
, 1))
1519 ecf_flags
|= ECF_CONST
;
1520 if (bp_unpack_value (&bp
, 1))
1521 ecf_flags
|= ECF_PURE
;
1522 if (bp_unpack_value (&bp
, 1))
1523 ecf_flags
|= ECF_NORETURN
;
1524 if (bp_unpack_value (&bp
, 1))
1525 ecf_flags
|= ECF_MALLOC
;
1526 if (bp_unpack_value (&bp
, 1))
1527 ecf_flags
|= ECF_NOTHROW
;
1528 if (bp_unpack_value (&bp
, 1))
1529 ecf_flags
|= ECF_RETURNS_TWICE
;
1530 edge
->indirect_info
->ecf_flags
= ecf_flags
;
1531 edge
->indirect_info
->common_target_id
= streamer_read_hwi (ib
);
1532 if (edge
->indirect_info
->common_target_id
)
1533 edge
->indirect_info
->common_target_probability
= streamer_read_hwi (ib
);
1538 /* Read a cgraph from IB using the info in FILE_DATA. */
1540 static vec
<symtab_node
*>
1541 input_cgraph_1 (struct lto_file_decl_data
*file_data
,
1542 struct lto_input_block
*ib
)
1544 enum LTO_symtab_tags tag
;
1545 vec
<symtab_node
*> nodes
= vNULL
;
1549 tag
= streamer_read_enum (ib
, LTO_symtab_tags
, LTO_symtab_last_tag
);
1550 order_base
= symtab
->order
;
1553 if (tag
== LTO_symtab_edge
)
1554 input_edge (ib
, nodes
, false);
1555 else if (tag
== LTO_symtab_indirect_edge
)
1556 input_edge (ib
, nodes
, true);
1557 else if (tag
== LTO_symtab_variable
)
1559 node
= input_varpool_node (file_data
, ib
);
1560 nodes
.safe_push (node
);
1561 lto_symtab_encoder_encode (file_data
->symtab_node_encoder
, node
);
1565 node
= input_node (file_data
, ib
, tag
, nodes
);
1566 if (node
== NULL
|| node
->decl
== NULL_TREE
)
1567 internal_error ("bytecode stream: found empty cgraph node");
1568 nodes
.safe_push (node
);
1569 lto_symtab_encoder_encode (file_data
->symtab_node_encoder
, node
);
1572 tag
= streamer_read_enum (ib
, LTO_symtab_tags
, LTO_symtab_last_tag
);
1575 lto_input_toplevel_asms (file_data
, order_base
);
1577 /* AUX pointers should be all non-zero for function nodes read from the stream. */
1578 #ifdef ENABLE_CHECKING
1579 FOR_EACH_VEC_ELT (nodes
, i
, node
)
1580 gcc_assert (node
->aux
|| !is_a
<cgraph_node
*> (node
));
1582 FOR_EACH_VEC_ELT (nodes
, i
, node
)
1585 if (cgraph_node
*cnode
= dyn_cast
<cgraph_node
*> (node
))
1587 ref
= (int) (intptr_t) cnode
->global
.inlined_to
;
1589 /* We share declaration of builtins, so we may read same node twice. */
1594 /* Fixup inlined_to from reference to pointer. */
1595 if (ref
!= LCC_NOT_FOUND
)
1596 dyn_cast
<cgraph_node
*> (node
)->global
.inlined_to
1597 = dyn_cast
<cgraph_node
*> (nodes
[ref
]);
1599 cnode
->global
.inlined_to
= NULL
;
1601 /* Compute instrumented_version. */
1602 if (cnode
->instrumentation_clone
)
1604 gcc_assert (cnode
->orig_decl
);
1606 cnode
->instrumented_version
= cgraph_node::get (cnode
->orig_decl
);
1607 if (cnode
->instrumented_version
)
1609 /* We may have multiple nodes for a single function which
1610 will be merged later. To have a proper merge we need
1611 to keep instrumentation_version reference between nodes
1612 consistent: each instrumented_version reference should
1613 have proper reverse reference. Thus don't break existing
1614 instrumented_version reference if it already exists. */
1615 if (cnode
->instrumented_version
->instrumented_version
)
1616 cnode
->instrumented_version
= NULL
;
1618 cnode
->instrumented_version
->instrumented_version
= cnode
;
1621 /* Restore decl names reference except for wrapper functions. */
1622 if (!chkp_wrap_function (cnode
->orig_decl
))
1624 tree name
= DECL_ASSEMBLER_NAME (cnode
->decl
);
1625 IDENTIFIER_TRANSPARENT_ALIAS (name
) = 1;
1626 TREE_CHAIN (name
) = DECL_ASSEMBLER_NAME (cnode
->orig_decl
);
1631 ref
= (int) (intptr_t) node
->same_comdat_group
;
1633 /* Fixup same_comdat_group from reference to pointer. */
1634 if (ref
!= LCC_NOT_FOUND
)
1635 node
->same_comdat_group
= nodes
[ref
];
1637 node
->same_comdat_group
= NULL
;
1639 FOR_EACH_VEC_ELT (nodes
, i
, node
)
1640 node
->aux
= is_a
<cgraph_node
*> (node
) ? (void *)1 : NULL
;
1644 /* Input ipa_refs. */
1647 input_refs (struct lto_input_block
*ib
,
1648 vec
<symtab_node
*> nodes
)
1655 count
= streamer_read_uhwi (ib
);
1658 idx
= streamer_read_uhwi (ib
);
1662 input_ref (ib
, node
, nodes
);
1669 static struct gcov_ctr_summary lto_gcov_summary
;
1671 /* Input profile_info from IB. */
1673 input_profile_summary (struct lto_input_block
*ib
,
1674 struct lto_file_decl_data
*file_data
)
1677 struct bitpack_d bp
;
1678 unsigned int runs
= streamer_read_uhwi (ib
);
1681 file_data
->profile_info
.runs
= runs
;
1682 file_data
->profile_info
.sum_max
= streamer_read_gcov_count (ib
);
1683 file_data
->profile_info
.sum_all
= streamer_read_gcov_count (ib
);
1685 memset (file_data
->profile_info
.histogram
, 0,
1686 sizeof (gcov_bucket_type
) * GCOV_HISTOGRAM_SIZE
);
1687 /* Input the bitpack of non-zero histogram indices. */
1688 bp
= streamer_read_bitpack (ib
);
1689 /* Read in and unpack the full bitpack, flagging non-zero
1690 histogram entries by setting the num_counters non-zero. */
1691 for (h_ix
= 0; h_ix
< GCOV_HISTOGRAM_SIZE
; h_ix
++)
1693 file_data
->profile_info
.histogram
[h_ix
].num_counters
1694 = bp_unpack_value (&bp
, 1);
1696 for (h_ix
= 0; h_ix
< GCOV_HISTOGRAM_SIZE
; h_ix
++)
1698 if (!file_data
->profile_info
.histogram
[h_ix
].num_counters
)
1701 file_data
->profile_info
.histogram
[h_ix
].num_counters
1702 = streamer_read_gcov_count (ib
);
1703 file_data
->profile_info
.histogram
[h_ix
].min_value
1704 = streamer_read_gcov_count (ib
);
1705 file_data
->profile_info
.histogram
[h_ix
].cum_value
1706 = streamer_read_gcov_count (ib
);
1708 /* IPA-profile computes hot bb threshold based on cumulated
1709 whole program profile. We need to stream it down to ltrans. */
1711 set_hot_bb_threshold (streamer_read_gcov_count (ib
));
1716 /* Rescale profile summaries to the same number of runs in the whole unit. */
1719 merge_profile_summaries (struct lto_file_decl_data
**file_data_vec
)
1721 struct lto_file_decl_data
*file_data
;
1722 unsigned int j
, h_ix
;
1723 gcov_unsigned_t max_runs
= 0;
1724 struct cgraph_node
*node
;
1725 struct cgraph_edge
*edge
;
1726 gcov_type saved_sum_all
= 0;
1727 gcov_ctr_summary
*saved_profile_info
= 0;
1728 int saved_scale
= 0;
1730 /* Find unit with maximal number of runs. If we ever get serious about
1731 roundoff errors, we might also consider computing smallest common
1733 for (j
= 0; (file_data
= file_data_vec
[j
]) != NULL
; j
++)
1734 if (max_runs
< file_data
->profile_info
.runs
)
1735 max_runs
= file_data
->profile_info
.runs
;
1740 /* Simple overflow check. We probably don't need to support that many train
1741 runs. Such a large value probably imply data corruption anyway. */
1742 if (max_runs
> INT_MAX
/ REG_BR_PROB_BASE
)
1744 sorry ("At most %i profile runs is supported. Perhaps corrupted profile?",
1745 INT_MAX
/ REG_BR_PROB_BASE
);
1749 profile_info
= <o_gcov_summary
;
1750 lto_gcov_summary
.runs
= max_runs
;
1751 lto_gcov_summary
.sum_max
= 0;
1752 memset (lto_gcov_summary
.histogram
, 0,
1753 sizeof (gcov_bucket_type
) * GCOV_HISTOGRAM_SIZE
);
1755 /* Rescale all units to the maximal number of runs.
1756 sum_max can not be easily merged, as we have no idea what files come from
1757 the same run. We do not use the info anyway, so leave it 0. */
1758 for (j
= 0; (file_data
= file_data_vec
[j
]) != NULL
; j
++)
1759 if (file_data
->profile_info
.runs
)
1761 int scale
= GCOV_COMPUTE_SCALE (max_runs
,
1762 file_data
->profile_info
.runs
);
1763 lto_gcov_summary
.sum_max
1764 = MAX (lto_gcov_summary
.sum_max
,
1765 apply_scale (file_data
->profile_info
.sum_max
, scale
));
1766 lto_gcov_summary
.sum_all
1767 = MAX (lto_gcov_summary
.sum_all
,
1768 apply_scale (file_data
->profile_info
.sum_all
, scale
));
1769 /* Save a pointer to the profile_info with the largest
1770 scaled sum_all and the scale for use in merging the
1772 if (!saved_profile_info
1773 || lto_gcov_summary
.sum_all
> saved_sum_all
)
1775 saved_profile_info
= &file_data
->profile_info
;
1776 saved_sum_all
= lto_gcov_summary
.sum_all
;
1777 saved_scale
= scale
;
1781 gcc_assert (saved_profile_info
);
1783 /* Scale up the histogram from the profile that had the largest
1784 scaled sum_all above. */
1785 for (h_ix
= 0; h_ix
< GCOV_HISTOGRAM_SIZE
; h_ix
++)
1787 /* Scale up the min value as we did the corresponding sum_all
1788 above. Use that to find the new histogram index. */
1789 gcov_type scaled_min
1790 = apply_scale (saved_profile_info
->histogram
[h_ix
].min_value
,
1792 /* The new index may be shared with another scaled histogram entry,
1793 so we need to account for a non-zero histogram entry at new_ix. */
1794 unsigned new_ix
= gcov_histo_index (scaled_min
);
1795 lto_gcov_summary
.histogram
[new_ix
].min_value
1796 = (lto_gcov_summary
.histogram
[new_ix
].num_counters
1797 ? MIN (lto_gcov_summary
.histogram
[new_ix
].min_value
, scaled_min
)
1799 /* Some of the scaled counter values would ostensibly need to be placed
1800 into different (larger) histogram buckets, but we keep things simple
1801 here and place the scaled cumulative counter value in the bucket
1802 corresponding to the scaled minimum counter value. */
1803 lto_gcov_summary
.histogram
[new_ix
].cum_value
1804 += apply_scale (saved_profile_info
->histogram
[h_ix
].cum_value
,
1806 lto_gcov_summary
.histogram
[new_ix
].num_counters
1807 += saved_profile_info
->histogram
[h_ix
].num_counters
;
1810 /* Watch roundoff errors. */
1811 if (lto_gcov_summary
.sum_max
< max_runs
)
1812 lto_gcov_summary
.sum_max
= max_runs
;
1814 /* If merging already happent at WPA time, we are done. */
1818 /* Now compute count_materialization_scale of each node.
1819 During LTRANS we already have values of count_materialization_scale
1820 computed, so just update them. */
1821 FOR_EACH_FUNCTION (node
)
1822 if (node
->lto_file_data
1823 && node
->lto_file_data
->profile_info
.runs
)
1827 scale
= RDIV (node
->count_materialization_scale
* max_runs
,
1828 node
->lto_file_data
->profile_info
.runs
);
1829 node
->count_materialization_scale
= scale
;
1831 fatal_error (input_location
, "Profile information in %s corrupted",
1832 file_data
->file_name
);
1834 if (scale
== REG_BR_PROB_BASE
)
1836 for (edge
= node
->callees
; edge
; edge
= edge
->next_callee
)
1837 edge
->count
= apply_scale (edge
->count
, scale
);
1838 node
->count
= apply_scale (node
->count
, scale
);
1842 /* Input and merge the symtab from each of the .o files passed to
1848 struct lto_file_decl_data
**file_data_vec
= lto_get_file_decl_data ();
1849 struct lto_file_decl_data
*file_data
;
1851 struct cgraph_node
*node
;
1853 while ((file_data
= file_data_vec
[j
++]))
1857 struct lto_input_block
*ib
;
1858 vec
<symtab_node
*> nodes
;
1860 ib
= lto_create_simple_input_block (file_data
, LTO_section_symtab_nodes
,
1863 fatal_error (input_location
,
1864 "cannot find LTO cgraph in %s", file_data
->file_name
);
1865 input_profile_summary (ib
, file_data
);
1866 file_data
->symtab_node_encoder
= lto_symtab_encoder_new (true);
1867 nodes
= input_cgraph_1 (file_data
, ib
);
1868 lto_destroy_simple_input_block (file_data
, LTO_section_symtab_nodes
,
1871 ib
= lto_create_simple_input_block (file_data
, LTO_section_refs
,
1874 fatal_error (input_location
, "cannot find LTO section refs in %s",
1875 file_data
->file_name
);
1876 input_refs (ib
, nodes
);
1877 lto_destroy_simple_input_block (file_data
, LTO_section_refs
,
1880 input_cgraph_opt_summary (nodes
);
1884 merge_profile_summaries (file_data_vec
);
1885 get_working_sets ();
1888 /* Clear out the aux field that was used to store enough state to
1889 tell which nodes should be overwritten. */
1890 FOR_EACH_FUNCTION (node
)
1892 /* Some nodes may have been created by cgraph_node. This
1893 happens when the callgraph contains nested functions. If the
1894 node for the parent function was never emitted to the gimple
1895 file, cgraph_node will create a node for it when setting the
1896 context of the nested function. */
1897 if (node
->lto_file_data
)
1902 /* Input function/variable tables that will allow libgomp to look up offload
1903 target code, and store them into OFFLOAD_FUNCS and OFFLOAD_VARS. */
1906 input_offload_tables (void)
1908 struct lto_file_decl_data
**file_data_vec
= lto_get_file_decl_data ();
1909 struct lto_file_decl_data
*file_data
;
1912 while ((file_data
= file_data_vec
[j
++]))
1916 struct lto_input_block
*ib
1917 = lto_create_simple_input_block (file_data
, LTO_section_offload_table
,
1922 enum LTO_symtab_tags tag
1923 = streamer_read_enum (ib
, LTO_symtab_tags
, LTO_symtab_last_tag
);
1926 if (tag
== LTO_symtab_unavail_node
)
1928 int decl_index
= streamer_read_uhwi (ib
);
1930 = lto_file_decl_data_get_fn_decl (file_data
, decl_index
);
1931 vec_safe_push (offload_funcs
, fn_decl
);
1933 else if (tag
== LTO_symtab_variable
)
1935 int decl_index
= streamer_read_uhwi (ib
);
1937 = lto_file_decl_data_get_var_decl (file_data
, decl_index
);
1938 vec_safe_push (offload_vars
, var_decl
);
1941 fatal_error (input_location
,
1942 "invalid offload table in %s", file_data
->file_name
);
1944 tag
= streamer_read_enum (ib
, LTO_symtab_tags
, LTO_symtab_last_tag
);
1947 lto_destroy_simple_input_block (file_data
, LTO_section_offload_table
,
1952 /* True when we need optimization summary for NODE. */
1955 output_cgraph_opt_summary_p (struct cgraph_node
*node
)
1957 return (node
->clone_of
1958 && (node
->clone
.tree_map
1959 || node
->clone
.args_to_skip
1960 || node
->clone
.combined_args_to_skip
));
1963 /* Output optimization summary for EDGE to OB. */
1965 output_edge_opt_summary (struct output_block
*ob ATTRIBUTE_UNUSED
,
1966 struct cgraph_edge
*edge ATTRIBUTE_UNUSED
)
1970 /* Output optimization summary for NODE to OB. */
1973 output_node_opt_summary (struct output_block
*ob
,
1974 struct cgraph_node
*node
,
1975 lto_symtab_encoder_t encoder
)
1979 struct ipa_replace_map
*map
;
1980 struct bitpack_d bp
;
1982 struct cgraph_edge
*e
;
1984 if (node
->clone
.args_to_skip
)
1986 streamer_write_uhwi (ob
, bitmap_count_bits (node
->clone
.args_to_skip
));
1987 EXECUTE_IF_SET_IN_BITMAP (node
->clone
.args_to_skip
, 0, index
, bi
)
1988 streamer_write_uhwi (ob
, index
);
1991 streamer_write_uhwi (ob
, 0);
1992 if (node
->clone
.combined_args_to_skip
)
1994 streamer_write_uhwi (ob
, bitmap_count_bits (node
->clone
.combined_args_to_skip
));
1995 EXECUTE_IF_SET_IN_BITMAP (node
->clone
.combined_args_to_skip
, 0, index
, bi
)
1996 streamer_write_uhwi (ob
, index
);
1999 streamer_write_uhwi (ob
, 0);
2000 streamer_write_uhwi (ob
, vec_safe_length (node
->clone
.tree_map
));
2001 FOR_EACH_VEC_SAFE_ELT (node
->clone
.tree_map
, i
, map
)
2003 /* At the moment we assume all old trees to be PARM_DECLs, because we have no
2004 mechanism to store function local declarations into summaries. */
2005 gcc_assert (!map
->old_tree
);
2006 streamer_write_uhwi (ob
, map
->parm_num
);
2007 gcc_assert (EXPR_LOCATION (map
->new_tree
) == UNKNOWN_LOCATION
);
2008 stream_write_tree (ob
, map
->new_tree
, true);
2009 bp
= bitpack_create (ob
->main_stream
);
2010 bp_pack_value (&bp
, map
->replace_p
, 1);
2011 bp_pack_value (&bp
, map
->ref_p
, 1);
2012 streamer_write_bitpack (&bp
);
2015 if (lto_symtab_encoder_in_partition_p (encoder
, node
))
2017 for (e
= node
->callees
; e
; e
= e
->next_callee
)
2018 output_edge_opt_summary (ob
, e
);
2019 for (e
= node
->indirect_calls
; e
; e
= e
->next_callee
)
2020 output_edge_opt_summary (ob
, e
);
2024 /* Output optimization summaries stored in callgraph.
2025 At the moment it is the clone info structure. */
2028 output_cgraph_opt_summary (void)
2031 lto_symtab_encoder_t encoder
;
2032 struct output_block
*ob
= create_output_block (LTO_section_cgraph_opt_sum
);
2036 encoder
= ob
->decl_state
->symtab_node_encoder
;
2037 n_nodes
= lto_symtab_encoder_size (encoder
);
2038 for (i
= 0; i
< n_nodes
; i
++)
2040 symtab_node
*node
= lto_symtab_encoder_deref (encoder
, i
);
2041 cgraph_node
*cnode
= dyn_cast
<cgraph_node
*> (node
);
2042 if (cnode
&& output_cgraph_opt_summary_p (cnode
))
2045 streamer_write_uhwi (ob
, count
);
2046 for (i
= 0; i
< n_nodes
; i
++)
2048 symtab_node
*node
= lto_symtab_encoder_deref (encoder
, i
);
2049 cgraph_node
*cnode
= dyn_cast
<cgraph_node
*> (node
);
2050 if (cnode
&& output_cgraph_opt_summary_p (cnode
))
2052 streamer_write_uhwi (ob
, i
);
2053 output_node_opt_summary (ob
, cnode
, encoder
);
2056 produce_asm (ob
, NULL
);
2057 destroy_output_block (ob
);
2060 /* Input optimisation summary of EDGE. */
2063 input_edge_opt_summary (struct cgraph_edge
*edge ATTRIBUTE_UNUSED
,
2064 struct lto_input_block
*ib_main ATTRIBUTE_UNUSED
)
2068 /* Input optimisation summary of NODE. */
2071 input_node_opt_summary (struct cgraph_node
*node
,
2072 struct lto_input_block
*ib_main
,
2073 struct data_in
*data_in
)
2078 struct bitpack_d bp
;
2079 struct cgraph_edge
*e
;
2081 count
= streamer_read_uhwi (ib_main
);
2083 node
->clone
.args_to_skip
= BITMAP_GGC_ALLOC ();
2084 for (i
= 0; i
< count
; i
++)
2086 bit
= streamer_read_uhwi (ib_main
);
2087 bitmap_set_bit (node
->clone
.args_to_skip
, bit
);
2089 count
= streamer_read_uhwi (ib_main
);
2091 node
->clone
.combined_args_to_skip
= BITMAP_GGC_ALLOC ();
2092 for (i
= 0; i
< count
; i
++)
2094 bit
= streamer_read_uhwi (ib_main
);
2095 bitmap_set_bit (node
->clone
.combined_args_to_skip
, bit
);
2097 count
= streamer_read_uhwi (ib_main
);
2098 for (i
= 0; i
< count
; i
++)
2100 struct ipa_replace_map
*map
= ggc_alloc
<ipa_replace_map
> ();
2102 vec_safe_push (node
->clone
.tree_map
, map
);
2103 map
->parm_num
= streamer_read_uhwi (ib_main
);
2104 map
->old_tree
= NULL
;
2105 map
->new_tree
= stream_read_tree (ib_main
, data_in
);
2106 bp
= streamer_read_bitpack (ib_main
);
2107 map
->replace_p
= bp_unpack_value (&bp
, 1);
2108 map
->ref_p
= bp_unpack_value (&bp
, 1);
2110 for (e
= node
->callees
; e
; e
= e
->next_callee
)
2111 input_edge_opt_summary (e
, ib_main
);
2112 for (e
= node
->indirect_calls
; e
; e
= e
->next_callee
)
2113 input_edge_opt_summary (e
, ib_main
);
2116 /* Read section in file FILE_DATA of length LEN with data DATA. */
2119 input_cgraph_opt_section (struct lto_file_decl_data
*file_data
,
2120 const char *data
, size_t len
,
2121 vec
<symtab_node
*> nodes
)
2123 const struct lto_function_header
*header
=
2124 (const struct lto_function_header
*) data
;
2125 const int cfg_offset
= sizeof (struct lto_function_header
);
2126 const int main_offset
= cfg_offset
+ header
->cfg_size
;
2127 const int string_offset
= main_offset
+ header
->main_size
;
2128 struct data_in
*data_in
;
2132 lto_input_block
ib_main ((const char *) data
+ main_offset
,
2133 header
->main_size
, file_data
->mode_table
);
2136 lto_data_in_create (file_data
, (const char *) data
+ string_offset
,
2137 header
->string_size
, vNULL
);
2138 count
= streamer_read_uhwi (&ib_main
);
2140 for (i
= 0; i
< count
; i
++)
2142 int ref
= streamer_read_uhwi (&ib_main
);
2143 input_node_opt_summary (dyn_cast
<cgraph_node
*> (nodes
[ref
]),
2146 lto_free_section_data (file_data
, LTO_section_cgraph_opt_sum
, NULL
, data
,
2148 lto_data_in_delete (data_in
);
2151 /* Input optimization summary of cgraph. */
2154 input_cgraph_opt_summary (vec
<symtab_node
*> nodes
)
2156 struct lto_file_decl_data
**file_data_vec
= lto_get_file_decl_data ();
2157 struct lto_file_decl_data
*file_data
;
2160 while ((file_data
= file_data_vec
[j
++]))
2164 lto_get_section_data (file_data
, LTO_section_cgraph_opt_sum
, NULL
,
2168 input_cgraph_opt_section (file_data
, data
, len
, nodes
);