PR target/60039
[official-gcc.git] / gcc / lto / lto-partition.c
blob1ee5fbb851e4fcb81f29b3be424254b4c0b72b53
1 /* LTO partitioning logic routines.
2 Copyright (C) 2009-2014 Free Software Foundation, Inc.
4 This file is part of GCC.
6 GCC is free software; you can redistribute it and/or modify it under
7 the terms of the GNU General Public License as published by the Free
8 Software Foundation; either version 3, or (at your option) any later
9 version.
11 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
12 WARRANTY; without even the implied warranty of MERCHANTABILITY or
13 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
14 for more details.
16 You should have received a copy of the GNU General Public License
17 along with GCC; see the file COPYING3. If not see
18 <http://www.gnu.org/licenses/>. */
20 #include "config.h"
21 #include "system.h"
22 #include "coretypes.h"
23 #include "toplev.h"
24 #include "tree.h"
25 #include "gcc-symtab.h"
26 #include "basic-block.h"
27 #include "tree-ssa-alias.h"
28 #include "internal-fn.h"
29 #include "gimple-expr.h"
30 #include "is-a.h"
31 #include "gimple.h"
32 #include "tm.h"
33 #include "cgraph.h"
34 #include "lto-streamer.h"
35 #include "timevar.h"
36 #include "params.h"
37 #include "ipa-inline.h"
38 #include "ipa-utils.h"
39 #include "lto-partition.h"
41 vec<ltrans_partition> ltrans_partitions;
43 static void add_symbol_to_partition (ltrans_partition part, symtab_node *node);
46 /* Create new partition with name NAME. */
48 static ltrans_partition
49 new_partition (const char *name)
51 ltrans_partition part = XCNEW (struct ltrans_partition_def);
52 part->encoder = lto_symtab_encoder_new (false);
53 part->name = name;
54 part->insns = 0;
55 ltrans_partitions.safe_push (part);
56 return part;
59 /* Free memory used by ltrans datastructures. */
61 void
62 free_ltrans_partitions (void)
64 unsigned int idx;
65 ltrans_partition part;
66 for (idx = 0; ltrans_partitions.iterate (idx, &part); idx++)
68 if (part->initializers_visited)
69 pointer_set_destroy (part->initializers_visited);
70 /* Symtab encoder is freed after streaming. */
71 free (part);
73 ltrans_partitions.release ();
76 /* Return true if symbol is already in some partition. */
78 static inline bool
79 symbol_partitioned_p (symtab_node *node)
81 return node->aux;
84 /* Add references into the partition. */
85 static void
86 add_references_to_partition (ltrans_partition part, symtab_node *node)
88 int i;
89 struct ipa_ref *ref;
91 /* Add all duplicated references to the partition. */
92 for (i = 0; ipa_ref_list_reference_iterate (&node->ref_list, i, ref); i++)
93 if (symtab_get_symbol_partitioning_class (ref->referred) == SYMBOL_DUPLICATE)
94 add_symbol_to_partition (part, ref->referred);
95 /* References to a readonly variable may be constant foled into its value.
96 Recursively look into the initializers of the constant variable and add
97 references, too. */
98 else if (is_a <varpool_node> (ref->referred)
99 && ctor_for_folding (ref->referred->decl) != error_mark_node
100 && !lto_symtab_encoder_in_partition_p (part->encoder, ref->referred))
102 if (!part->initializers_visited)
103 part->initializers_visited = pointer_set_create ();
104 if (!pointer_set_insert (part->initializers_visited, ref->referred))
105 add_references_to_partition (part, ref->referred);
109 /* Helper function for add_symbol_to_partition doing the actual dirty work
110 of adding NODE to PART. */
112 static bool
113 add_symbol_to_partition_1 (ltrans_partition part, symtab_node *node)
115 enum symbol_partitioning_class c = symtab_get_symbol_partitioning_class (node);
116 int i;
117 struct ipa_ref *ref;
118 symtab_node *node1;
120 /* If NODE is already there, we have nothing to do. */
121 if (lto_symtab_encoder_in_partition_p (part->encoder, node))
122 return true;
124 /* non-duplicated aliases or tunks of a duplicated symbol needs to be output
125 just once.
127 Be lax about comdats; they may or may not be duplicated and we may
128 end up in need to duplicate keyed comdat because it has unkeyed alias. */
129 if (c == SYMBOL_PARTITION && !DECL_COMDAT (node->decl)
130 && symbol_partitioned_p (node))
131 return false;
133 /* Be sure that we never try to duplicate partitioned symbol
134 or add external symbol. */
135 gcc_assert (c != SYMBOL_EXTERNAL
136 && (c == SYMBOL_DUPLICATE || !symbol_partitioned_p (node)));
138 lto_set_symtab_encoder_in_partition (part->encoder, node);
140 if (symbol_partitioned_p (node))
142 node->in_other_partition = 1;
143 if (cgraph_dump_file)
144 fprintf (cgraph_dump_file, "Symbol node %s now used in multiple partitions\n",
145 node->name ());
147 node->aux = (void *)((size_t)node->aux + 1);
149 if (cgraph_node *cnode = dyn_cast <cgraph_node> (node))
151 struct cgraph_edge *e;
152 if (!node->alias)
153 part->insns += inline_summary (cnode)->self_size;
155 /* Add all inline clones and callees that are duplicated. */
156 for (e = cnode->callees; e; e = e->next_callee)
157 if (!e->inline_failed)
158 add_symbol_to_partition_1 (part, e->callee);
159 else if (symtab_get_symbol_partitioning_class (e->callee) == SYMBOL_DUPLICATE)
160 add_symbol_to_partition (part, e->callee);
162 /* Add all thunks associated with the function. */
163 for (e = cnode->callers; e; e = e->next_caller)
164 if (e->caller->thunk.thunk_p)
165 add_symbol_to_partition_1 (part, e->caller);
168 add_references_to_partition (part, node);
170 /* Add all aliases associated with the symbol. */
171 for (i = 0; ipa_ref_list_referring_iterate (&node->ref_list, i, ref); i++)
172 if (ref->use == IPA_REF_ALIAS && !node->weakref)
173 add_symbol_to_partition_1 (part, ref->referring);
175 /* Ensure that SAME_COMDAT_GROUP lists all allways added in a group. */
176 if (node->same_comdat_group)
177 for (node1 = node->same_comdat_group;
178 node1 != node; node1 = node1->same_comdat_group)
179 if (!node->alias)
181 bool added = add_symbol_to_partition_1 (part, node1);
182 gcc_assert (added);
184 return true;
187 /* If symbol NODE is really part of other symbol's definition (i.e. it is
188 internal label, thunk, alias or so), return the outer symbol.
189 When add_symbol_to_partition_1 is called on the outer symbol it must
190 eventually add NODE, too. */
191 static symtab_node *
192 contained_in_symbol (symtab_node *node)
194 /* Weakrefs are never contained in anything. */
195 if (node->weakref)
196 return node;
197 if (cgraph_node *cnode = dyn_cast <cgraph_node> (node))
199 cnode = cgraph_function_node (cnode, NULL);
200 if (cnode->global.inlined_to)
201 cnode = cnode->global.inlined_to;
202 return cnode;
204 else if (varpool_node *vnode = dyn_cast <varpool_node> (node))
205 return varpool_variable_node (vnode, NULL);
206 return node;
209 /* Add symbol NODE to partition. When definition of NODE is part
210 of other symbol definition, add the other symbol, too. */
212 static void
213 add_symbol_to_partition (ltrans_partition part, symtab_node *node)
215 symtab_node *node1;
217 /* Verify that we do not try to duplicate something that can not be. */
218 gcc_checking_assert (symtab_get_symbol_partitioning_class (node) == SYMBOL_DUPLICATE
219 || !symbol_partitioned_p (node));
221 while ((node1 = contained_in_symbol (node)) != node)
222 node = node1;
224 /* If we have duplicated symbol contained in something we can not duplicate,
225 we are very badly screwed. The other way is possible, so we do not
226 assert this in add_symbol_to_partition_1.
228 Be lax about comdats; they may or may not be duplicated and we may
229 end up in need to duplicate keyed comdat because it has unkeyed alias. */
231 gcc_assert (symtab_get_symbol_partitioning_class (node) == SYMBOL_DUPLICATE
232 || DECL_COMDAT (node->decl)
233 || !symbol_partitioned_p (node));
235 add_symbol_to_partition_1 (part, node);
238 /* Undo all additions until number of cgraph nodes in PARITION is N_CGRAPH_NODES
239 and number of varpool nodes is N_VARPOOL_NODES. */
241 static void
242 undo_partition (ltrans_partition partition, unsigned int n_nodes)
244 while (lto_symtab_encoder_size (partition->encoder) > (int)n_nodes)
246 symtab_node *node = lto_symtab_encoder_deref (partition->encoder,
247 n_nodes);
248 cgraph_node *cnode;
250 /* After UNDO we no longer know what was visited. */
251 if (partition->initializers_visited)
252 pointer_set_destroy (partition->initializers_visited);
253 partition->initializers_visited = NULL;
255 if (!node->alias && (cnode = dyn_cast <cgraph_node> (node)))
256 partition->insns -= inline_summary (cnode)->self_size;
257 lto_symtab_encoder_delete_node (partition->encoder, node);
258 node->aux = (void *)((size_t)node->aux - 1);
262 /* Group cgrah nodes by input files. This is used mainly for testing
263 right now. */
265 void
266 lto_1_to_1_map (void)
268 symtab_node *node;
269 struct lto_file_decl_data *file_data;
270 struct pointer_map_t *pmap;
271 ltrans_partition partition;
272 void **slot;
273 int npartitions = 0;
275 pmap = pointer_map_create ();
277 FOR_EACH_SYMBOL (node)
279 if (symtab_get_symbol_partitioning_class (node) != SYMBOL_PARTITION
280 || symbol_partitioned_p (node))
281 continue;
283 file_data = node->lto_file_data;
285 if (file_data)
287 slot = pointer_map_contains (pmap, file_data);
288 if (slot)
289 partition = (ltrans_partition) *slot;
290 else
292 partition = new_partition (file_data->file_name);
293 slot = pointer_map_insert (pmap, file_data);
294 *slot = partition;
295 npartitions++;
298 else if (!file_data && ltrans_partitions.length ())
299 partition = ltrans_partitions[0];
300 else
302 partition = new_partition ("");
303 slot = pointer_map_insert (pmap, NULL);
304 *slot = partition;
305 npartitions++;
308 add_symbol_to_partition (partition, node);
311 /* If the cgraph is empty, create one cgraph node set so that there is still
312 an output file for any variables that need to be exported in a DSO. */
313 if (!npartitions)
314 new_partition ("empty");
316 pointer_map_destroy (pmap);
320 /* Maximal partitioning. Put every new symbol into new partition if possible. */
322 void
323 lto_max_map (void)
325 symtab_node *node;
326 ltrans_partition partition;
327 int npartitions = 0;
329 FOR_EACH_SYMBOL (node)
331 if (symtab_get_symbol_partitioning_class (node) != SYMBOL_PARTITION
332 || symbol_partitioned_p (node))
333 continue;
334 partition = new_partition (node->asm_name ());
335 add_symbol_to_partition (partition, node);
336 npartitions++;
338 if (!npartitions)
339 new_partition ("empty");
342 /* Helper function for qsort; sort nodes by order. */
343 static int
344 node_cmp (const void *pa, const void *pb)
346 const struct cgraph_node *a = *(const struct cgraph_node * const *) pa;
347 const struct cgraph_node *b = *(const struct cgraph_node * const *) pb;
349 /* Profile reorder flag enables function reordering based on first execution
350 of a function. All functions with profile are placed in ascending
351 order at the beginning. */
353 if (flag_profile_reorder_functions)
355 /* Functions with time profile are sorted in ascending order. */
356 if (a->tp_first_run && b->tp_first_run)
357 return a->tp_first_run != b->tp_first_run
358 ? a->tp_first_run - b->tp_first_run
359 : a->order - b->order;
361 /* Functions with time profile are sorted before the functions
362 that do not have the profile. */
363 if (a->tp_first_run || b->tp_first_run)
364 return b->tp_first_run - a->tp_first_run;
367 return b->order - a->order;
370 /* Helper function for qsort; sort nodes by order. */
371 static int
372 varpool_node_cmp (const void *pa, const void *pb)
374 const varpool_node *a = *(const varpool_node * const *) pa;
375 const varpool_node *b = *(const varpool_node * const *) pb;
376 return b->order - a->order;
379 /* Group cgraph nodes into equally-sized partitions.
381 The partitioning algorithm is simple: nodes are taken in predefined order.
382 The order corresponds to the order we want functions to have in the final
383 output. In the future this will be given by function reordering pass, but
384 at the moment we use the topological order, which is a good approximation.
386 The goal is to partition this linear order into intervals (partitions) so
387 that all the partitions have approximately the same size and the number of
388 callgraph or IPA reference edges crossing boundaries is minimal.
390 This is a lot faster (O(n) in size of callgraph) than algorithms doing
391 priority-based graph clustering that are generally O(n^2) and, since
392 WHOPR is designed to make things go well across partitions, it leads
393 to good results.
395 We compute the expected size of a partition as:
397 max (total_size / lto_partitions, min_partition_size)
399 We use dynamic expected size of partition so small programs are partitioned
400 into enough partitions to allow use of multiple CPUs, while large programs
401 are not partitioned too much. Creating too many partitions significantly
402 increases the streaming overhead.
404 In the future, we would like to bound the maximal size of partitions so as
405 to prevent the LTRANS stage from consuming too much memory. At the moment,
406 however, the WPA stage is the most memory intensive for large benchmarks,
407 since too many types and declarations are read into memory.
409 The function implements a simple greedy algorithm. Nodes are being added
410 to the current partition until after 3/4 of the expected partition size is
411 reached. Past this threshold, we keep track of boundary size (number of
412 edges going to other partitions) and continue adding functions until after
413 the current partition has grown to twice the expected partition size. Then
414 the process is undone to the point where the minimal ratio of boundary size
415 and in-partition calls was reached. */
417 void
418 lto_balanced_map (void)
420 int n_nodes = 0;
421 int n_varpool_nodes = 0, varpool_pos = 0, best_varpool_pos = 0;
422 struct cgraph_node **order = XNEWVEC (struct cgraph_node *, cgraph_max_uid);
423 varpool_node **varpool_order = NULL;
424 int i;
425 struct cgraph_node *node;
426 int total_size = 0, best_total_size = 0;
427 int partition_size;
428 ltrans_partition partition;
429 int last_visited_node = 0;
430 varpool_node *vnode;
431 int cost = 0, internal = 0;
432 int best_n_nodes = 0, best_i = 0, best_cost =
433 INT_MAX, best_internal = 0;
434 int npartitions;
435 int current_order = -1;
437 FOR_EACH_VARIABLE (vnode)
438 gcc_assert (!vnode->aux);
440 FOR_EACH_DEFINED_FUNCTION (node)
441 if (symtab_get_symbol_partitioning_class (node) == SYMBOL_PARTITION)
443 order[n_nodes++] = node;
444 if (!node->alias)
445 total_size += inline_summary (node)->size;
448 /* Streaming works best when the source units do not cross partition
449 boundaries much. This is because importing function from a source
450 unit tends to import a lot of global trees defined there. We should
451 get better about minimizing the function bounday, but until that
452 things works smoother if we order in source order. */
453 qsort (order, n_nodes, sizeof (struct cgraph_node *), node_cmp);
455 if (cgraph_dump_file)
456 for(i = 0; i < n_nodes; i++)
457 fprintf (cgraph_dump_file, "Balanced map symbol order:%s:%u\n", order[i]->name (), order[i]->tp_first_run);
459 if (!flag_toplevel_reorder)
461 FOR_EACH_VARIABLE (vnode)
462 if (symtab_get_symbol_partitioning_class (vnode) == SYMBOL_PARTITION)
463 n_varpool_nodes++;
464 varpool_order = XNEWVEC (varpool_node *, n_varpool_nodes);
466 n_varpool_nodes = 0;
467 FOR_EACH_VARIABLE (vnode)
468 if (symtab_get_symbol_partitioning_class (vnode) == SYMBOL_PARTITION)
469 varpool_order[n_varpool_nodes++] = vnode;
470 qsort (varpool_order, n_varpool_nodes, sizeof (varpool_node *),
471 varpool_node_cmp);
474 /* Compute partition size and create the first partition. */
475 partition_size = total_size / PARAM_VALUE (PARAM_LTO_PARTITIONS);
476 if (partition_size < PARAM_VALUE (MIN_PARTITION_SIZE))
477 partition_size = PARAM_VALUE (MIN_PARTITION_SIZE);
478 npartitions = 1;
479 partition = new_partition ("");
480 if (cgraph_dump_file)
481 fprintf (cgraph_dump_file, "Total unit size: %i, partition size: %i\n",
482 total_size, partition_size);
484 for (i = 0; i < n_nodes; i++)
486 if (symbol_partitioned_p (order[i]))
487 continue;
489 current_order = order[i]->order;
491 if (!flag_toplevel_reorder)
492 while (varpool_pos < n_varpool_nodes
493 && varpool_order[varpool_pos]->order < current_order)
495 if (!symbol_partitioned_p (varpool_order[varpool_pos]))
496 add_symbol_to_partition (partition, varpool_order[varpool_pos]);
497 varpool_pos++;
500 add_symbol_to_partition (partition, order[i]);
501 if (!order[i]->alias)
502 total_size -= inline_summary (order[i])->size;
505 /* Once we added a new node to the partition, we also want to add
506 all referenced variables unless they was already added into some
507 earlier partition.
508 add_symbol_to_partition adds possibly multiple nodes and
509 variables that are needed to satisfy needs of ORDER[i].
510 We remember last visited cgraph and varpool node from last iteration
511 of outer loop that allows us to process every new addition.
513 At the same time we compute size of the boundary into COST. Every
514 callgraph or IPA reference edge leaving the partition contributes into
515 COST. Every edge inside partition was earlier computed as one leaving
516 it and thus we need to subtract it from COST. */
517 while (last_visited_node < lto_symtab_encoder_size (partition->encoder))
519 struct ipa_ref_list *refs;
520 int j;
521 struct ipa_ref *ref;
522 symtab_node *snode = lto_symtab_encoder_deref (partition->encoder,
523 last_visited_node);
525 if (cgraph_node *node = dyn_cast <cgraph_node> (snode))
527 struct cgraph_edge *edge;
529 refs = &node->ref_list;
531 last_visited_node++;
533 gcc_assert (node->definition || node->weakref);
535 /* Compute boundary cost of callgraph edges. */
536 for (edge = node->callees; edge; edge = edge->next_callee)
537 if (edge->callee->definition)
539 int edge_cost = edge->frequency;
540 int index;
542 if (!edge_cost)
543 edge_cost = 1;
544 gcc_assert (edge_cost > 0);
545 index = lto_symtab_encoder_lookup (partition->encoder,
546 edge->callee);
547 if (index != LCC_NOT_FOUND
548 && index < last_visited_node - 1)
549 cost -= edge_cost, internal += edge_cost;
550 else
551 cost += edge_cost;
553 for (edge = node->callers; edge; edge = edge->next_caller)
555 int edge_cost = edge->frequency;
556 int index;
558 gcc_assert (edge->caller->definition);
559 if (!edge_cost)
560 edge_cost = 1;
561 gcc_assert (edge_cost > 0);
562 index = lto_symtab_encoder_lookup (partition->encoder,
563 edge->caller);
564 if (index != LCC_NOT_FOUND
565 && index < last_visited_node - 1)
566 cost -= edge_cost;
567 else
568 cost += edge_cost;
571 else
573 refs = &snode->ref_list;
574 last_visited_node++;
577 /* Compute boundary cost of IPA REF edges and at the same time look into
578 variables referenced from current partition and try to add them. */
579 for (j = 0; ipa_ref_list_reference_iterate (refs, j, ref); j++)
580 if (is_a <varpool_node> (ref->referred))
582 int index;
584 vnode = ipa_ref_varpool_node (ref);
585 if (!vnode->definition)
586 continue;
587 if (!symbol_partitioned_p (vnode) && flag_toplevel_reorder
588 && symtab_get_symbol_partitioning_class (vnode) == SYMBOL_PARTITION)
589 add_symbol_to_partition (partition, vnode);
590 index = lto_symtab_encoder_lookup (partition->encoder,
591 vnode);
592 if (index != LCC_NOT_FOUND
593 && index < last_visited_node - 1)
594 cost--, internal++;
595 else
596 cost++;
598 else
600 int index;
602 node = ipa_ref_node (ref);
603 if (!node->definition)
604 continue;
605 index = lto_symtab_encoder_lookup (partition->encoder,
606 node);
607 if (index != LCC_NOT_FOUND
608 && index < last_visited_node - 1)
609 cost--, internal++;
610 else
611 cost++;
613 for (j = 0; ipa_ref_list_referring_iterate (refs, j, ref); j++)
614 if (is_a <varpool_node> (ref->referring))
616 int index;
618 vnode = ipa_ref_referring_varpool_node (ref);
619 gcc_assert (vnode->definition);
620 /* It is better to couple variables with their users, because it allows them
621 to be removed. Coupling with objects they refer to only helps to reduce
622 number of symbols promoted to hidden. */
623 if (!symbol_partitioned_p (vnode) && flag_toplevel_reorder
624 && !varpool_can_remove_if_no_refs (vnode)
625 && symtab_get_symbol_partitioning_class (vnode) == SYMBOL_PARTITION)
626 add_symbol_to_partition (partition, vnode);
627 index = lto_symtab_encoder_lookup (partition->encoder,
628 vnode);
629 if (index != LCC_NOT_FOUND
630 && index < last_visited_node - 1)
631 cost--;
632 else
633 cost++;
635 else
637 int index;
639 node = ipa_ref_referring_node (ref);
640 gcc_assert (node->definition);
641 index = lto_symtab_encoder_lookup (partition->encoder,
642 node);
643 if (index != LCC_NOT_FOUND
644 && index < last_visited_node - 1)
645 cost--;
646 else
647 cost++;
651 /* If the partition is large enough, start looking for smallest boundary cost. */
652 if (partition->insns < partition_size * 3 / 4
653 || best_cost == INT_MAX
654 || ((!cost
655 || (best_internal * (HOST_WIDE_INT) cost
656 > (internal * (HOST_WIDE_INT)best_cost)))
657 && partition->insns < partition_size * 5 / 4))
659 best_cost = cost;
660 best_internal = internal;
661 best_i = i;
662 best_n_nodes = lto_symtab_encoder_size (partition->encoder);
663 best_total_size = total_size;
664 best_varpool_pos = varpool_pos;
666 if (cgraph_dump_file)
667 fprintf (cgraph_dump_file, "Step %i: added %s/%i, size %i, cost %i/%i "
668 "best %i/%i, step %i\n", i,
669 order[i]->name (), order[i]->order,
670 partition->insns, cost, internal,
671 best_cost, best_internal, best_i);
672 /* Partition is too large, unwind into step when best cost was reached and
673 start new partition. */
674 if (partition->insns > 2 * partition_size)
676 if (best_i != i)
678 if (cgraph_dump_file)
679 fprintf (cgraph_dump_file, "Unwinding %i insertions to step %i\n",
680 i - best_i, best_i);
681 undo_partition (partition, best_n_nodes);
682 varpool_pos = best_varpool_pos;
684 i = best_i;
685 /* When we are finished, avoid creating empty partition. */
686 while (i < n_nodes - 1 && symbol_partitioned_p (order[i + 1]))
687 i++;
688 if (i == n_nodes - 1)
689 break;
690 partition = new_partition ("");
691 last_visited_node = 0;
692 total_size = best_total_size;
693 cost = 0;
695 if (cgraph_dump_file)
696 fprintf (cgraph_dump_file, "New partition\n");
697 best_n_nodes = 0;
698 best_cost = INT_MAX;
700 /* Since the size of partitions is just approximate, update the size after
701 we finished current one. */
702 if (npartitions < PARAM_VALUE (PARAM_LTO_PARTITIONS))
703 partition_size = total_size
704 / (PARAM_VALUE (PARAM_LTO_PARTITIONS) - npartitions);
705 else
706 partition_size = INT_MAX;
708 if (partition_size < PARAM_VALUE (MIN_PARTITION_SIZE))
709 partition_size = PARAM_VALUE (MIN_PARTITION_SIZE);
710 npartitions ++;
714 /* Varables that are not reachable from the code go into last partition. */
715 if (flag_toplevel_reorder)
717 FOR_EACH_VARIABLE (vnode)
718 if (symtab_get_symbol_partitioning_class (vnode) == SYMBOL_PARTITION
719 && !symbol_partitioned_p (vnode))
720 add_symbol_to_partition (partition, vnode);
722 else
724 while (varpool_pos < n_varpool_nodes)
726 if (!symbol_partitioned_p (varpool_order[varpool_pos]))
727 add_symbol_to_partition (partition, varpool_order[varpool_pos]);
728 varpool_pos++;
730 free (varpool_order);
732 free (order);
735 /* Mangle NODE symbol name into a local name.
736 This is necessary to do
737 1) if two or more static vars of same assembler name
738 are merged into single ltrans unit.
739 2) if prevoiusly static var was promoted hidden to avoid possible conflict
740 with symbols defined out of the LTO world.
743 static bool
744 privatize_symbol_name (symtab_node *node)
746 tree decl = node->decl;
747 const char *name = IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl));
749 /* Our renaming machinery do not handle more than one change of assembler name.
750 We should not need more than one anyway. */
751 if (node->lto_file_data
752 && lto_get_decl_name_mapping (node->lto_file_data, name) != name)
754 if (cgraph_dump_file)
755 fprintf (cgraph_dump_file,
756 "Not privatizing symbol name: %s. It privatized already.\n",
757 name);
758 return false;
760 /* Avoid mangling of already mangled clones.
761 ??? should have a flag whether a symbol has a 'private' name already,
762 since we produce some symbols like that i.e. for global constructors
763 that are not really clones. */
764 if (node->unique_name)
766 if (cgraph_dump_file)
767 fprintf (cgraph_dump_file,
768 "Not privatizing symbol name: %s. Has unique name.\n",
769 name);
770 return false;
772 change_decl_assembler_name (decl, clone_function_name (decl, "lto_priv"));
773 if (node->lto_file_data)
774 lto_record_renamed_decl (node->lto_file_data, name,
775 IDENTIFIER_POINTER
776 (DECL_ASSEMBLER_NAME (decl)));
777 if (cgraph_dump_file)
778 fprintf (cgraph_dump_file,
779 "Privatizing symbol name: %s -> %s\n",
780 name, IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl)));
781 return true;
784 /* Promote variable VNODE to be static. */
786 static void
787 promote_symbol (symtab_node *node)
789 /* We already promoted ... */
790 if (DECL_VISIBILITY (node->decl) == VISIBILITY_HIDDEN
791 && DECL_VISIBILITY_SPECIFIED (node->decl)
792 && TREE_PUBLIC (node->decl))
793 return;
795 gcc_checking_assert (!TREE_PUBLIC (node->decl)
796 && !DECL_EXTERNAL (node->decl));
797 /* Be sure that newly public symbol does not conflict with anything already
798 defined by the non-LTO part. */
799 privatize_symbol_name (node);
800 TREE_PUBLIC (node->decl) = 1;
801 DECL_VISIBILITY (node->decl) = VISIBILITY_HIDDEN;
802 DECL_VISIBILITY_SPECIFIED (node->decl) = true;
803 if (cgraph_dump_file)
804 fprintf (cgraph_dump_file,
805 "Promoting as hidden: %s\n", node->name ());
808 /* Return true if NODE needs named section even if it won't land in the partition
809 symbol table.
810 FIXME: we should really not use named sections for inline clones and master clones. */
812 static bool
813 may_need_named_section_p (lto_symtab_encoder_t encoder, symtab_node *node)
815 struct cgraph_node *cnode = dyn_cast <cgraph_node> (node);
816 if (!cnode)
817 return false;
818 if (symtab_real_symbol_p (node))
819 return false;
820 return (!encoder
821 || (lto_symtab_encoder_lookup (encoder, node) != LCC_NOT_FOUND
822 && lto_symtab_encoder_encode_body_p (encoder,
823 cnode)));
826 /* If NODE represents a static variable. See if there are other variables
827 of the same name in partition ENCODER (or in whole compilation unit if
828 ENCODER is NULL) and if so, mangle the statics. Always mangle all
829 conflicting statics, so we reduce changes of silently miscompiling
830 asm statements referring to them by symbol name. */
832 static void
833 rename_statics (lto_symtab_encoder_t encoder, symtab_node *node)
835 tree decl = node->decl;
836 symtab_node *s;
837 tree name = DECL_ASSEMBLER_NAME (decl);
839 /* See if this is static symbol. */
840 if ((node->externally_visible
841 /* FIXME: externally_visible is somewhat illogically not set for
842 external symbols (i.e. those not defined). Remove this test
843 once this is fixed. */
844 || DECL_EXTERNAL (node->decl)
845 || !symtab_real_symbol_p (node))
846 && !may_need_named_section_p (encoder, node))
847 return;
849 /* Now walk symbols sharing the same name and see if there are any conflicts.
850 (all types of symbols counts here, since we can not have static of the
851 same name as external or public symbol.) */
852 for (s = symtab_node_for_asm (name);
853 s; s = s->next_sharing_asm_name)
854 if ((symtab_real_symbol_p (s) || may_need_named_section_p (encoder, s))
855 && s->decl != node->decl
856 && (!encoder
857 || lto_symtab_encoder_lookup (encoder, s) != LCC_NOT_FOUND))
858 break;
860 /* OK, no confict, so we have nothing to do. */
861 if (!s)
862 return;
864 if (cgraph_dump_file)
865 fprintf (cgraph_dump_file,
866 "Renaming statics with asm name: %s\n", node->name ());
868 /* Assign every symbol in the set that shares the same ASM name an unique
869 mangled name. */
870 for (s = symtab_node_for_asm (name); s;)
871 if (!s->externally_visible
872 && ((symtab_real_symbol_p (s)
873 && !DECL_EXTERNAL (node->decl)
874 && !TREE_PUBLIC (node->decl))
875 || may_need_named_section_p (encoder, s))
876 && (!encoder
877 || lto_symtab_encoder_lookup (encoder, s) != LCC_NOT_FOUND))
879 if (privatize_symbol_name (s))
880 /* Re-start from beginning since we do not know how many symbols changed a name. */
881 s = symtab_node_for_asm (name);
882 else s = s->next_sharing_asm_name;
884 else s = s->next_sharing_asm_name;
887 /* Find out all static decls that need to be promoted to global because
888 of cross file sharing. This function must be run in the WPA mode after
889 all inlinees are added. */
891 void
892 lto_promote_cross_file_statics (void)
894 unsigned i, n_sets;
896 gcc_assert (flag_wpa);
898 /* First compute boundaries. */
899 n_sets = ltrans_partitions.length ();
900 for (i = 0; i < n_sets; i++)
902 ltrans_partition part
903 = ltrans_partitions[i];
904 part->encoder = compute_ltrans_boundary (part->encoder);
907 /* Look at boundaries and promote symbols as needed. */
908 for (i = 0; i < n_sets; i++)
910 lto_symtab_encoder_iterator lsei;
911 lto_symtab_encoder_t encoder = ltrans_partitions[i]->encoder;
913 for (lsei = lsei_start (encoder); !lsei_end_p (lsei);
914 lsei_next (&lsei))
916 symtab_node *node = lsei_node (lsei);
918 /* If symbol is static, rename it if its assembler name clash with
919 anything else in this unit. */
920 rename_statics (encoder, node);
922 /* No need to promote if symbol already is externally visible ... */
923 if (node->externally_visible
924 /* ... or if it is part of current partition ... */
925 || lto_symtab_encoder_in_partition_p (encoder, node)
926 /* ... or if we do not partition it. This mean that it will
927 appear in every partition refernecing it. */
928 || symtab_get_symbol_partitioning_class (node) != SYMBOL_PARTITION)
929 continue;
931 promote_symbol (node);
936 /* Rename statics in the whole unit in the case that
937 we do -flto-partition=none. */
939 void
940 lto_promote_statics_nonwpa (void)
942 symtab_node *node;
943 FOR_EACH_SYMBOL (node)
944 rename_statics (NULL, node);