Allow gather loads to be used for grouped accesses
[official-gcc.git] / gcc / ira-color.c
blobc8b6ab4bcdf541b1385f2d75f4c7cb75461544fd
1 /* IRA allocation based on graph coloring.
2 Copyright (C) 2006-2018 Free Software Foundation, Inc.
3 Contributed by Vladimir Makarov <vmakarov@redhat.com>.
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify it under
8 the terms of the GNU General Public License as published by the Free
9 Software Foundation; either version 3, or (at your option) any later
10 version.
12 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
15 for more details.
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
21 #include "config.h"
22 #include "system.h"
23 #include "coretypes.h"
24 #include "backend.h"
25 #include "target.h"
26 #include "rtl.h"
27 #include "tree.h"
28 #include "predict.h"
29 #include "df.h"
30 #include "memmodel.h"
31 #include "tm_p.h"
32 #include "insn-config.h"
33 #include "regs.h"
34 #include "ira.h"
35 #include "ira-int.h"
36 #include "reload.h"
37 #include "cfgloop.h"
39 typedef struct allocno_hard_regs *allocno_hard_regs_t;
41 /* The structure contains information about hard registers can be
42 assigned to allocnos. Usually it is allocno profitable hard
43 registers but in some cases this set can be a bit different. Major
44 reason of the difference is a requirement to use hard register sets
45 that form a tree or a forest (set of trees), i.e. hard register set
46 of a node should contain hard register sets of its subnodes. */
47 struct allocno_hard_regs
49 /* Hard registers can be assigned to an allocno. */
50 HARD_REG_SET set;
51 /* Overall (spilling) cost of all allocnos with given register
52 set. */
53 int64_t cost;
56 typedef struct allocno_hard_regs_node *allocno_hard_regs_node_t;
58 /* A node representing allocno hard registers. Such nodes form a
59 forest (set of trees). Each subnode of given node in the forest
60 refers for hard register set (usually allocno profitable hard
61 register set) which is a subset of one referred from given
62 node. */
63 struct allocno_hard_regs_node
65 /* Set up number of the node in preorder traversing of the forest. */
66 int preorder_num;
67 /* Used for different calculation like finding conflict size of an
68 allocno. */
69 int check;
70 /* Used for calculation of conflict size of an allocno. The
71 conflict size of the allocno is maximal number of given allocno
72 hard registers needed for allocation of the conflicting allocnos.
73 Given allocno is trivially colored if this number plus the number
74 of hard registers needed for given allocno is not greater than
75 the number of given allocno hard register set. */
76 int conflict_size;
77 /* The number of hard registers given by member hard_regs. */
78 int hard_regs_num;
79 /* The following member is used to form the final forest. */
80 bool used_p;
81 /* Pointer to the corresponding profitable hard registers. */
82 allocno_hard_regs_t hard_regs;
83 /* Parent, first subnode, previous and next node with the same
84 parent in the forest. */
85 allocno_hard_regs_node_t parent, first, prev, next;
88 /* Info about changing hard reg costs of an allocno. */
89 struct update_cost_record
91 /* Hard regno for which we changed the cost. */
92 int hard_regno;
93 /* Divisor used when we changed the cost of HARD_REGNO. */
94 int divisor;
95 /* Next record for given allocno. */
96 struct update_cost_record *next;
99 /* To decrease footprint of ira_allocno structure we store all data
100 needed only for coloring in the following structure. */
101 struct allocno_color_data
103 /* TRUE value means that the allocno was not removed yet from the
104 conflicting graph during coloring. */
105 unsigned int in_graph_p : 1;
106 /* TRUE if it is put on the stack to make other allocnos
107 colorable. */
108 unsigned int may_be_spilled_p : 1;
109 /* TRUE if the allocno is trivially colorable. */
110 unsigned int colorable_p : 1;
111 /* Number of hard registers of the allocno class really
112 available for the allocno allocation. It is number of the
113 profitable hard regs. */
114 int available_regs_num;
115 /* Allocnos in a bucket (used in coloring) chained by the following
116 two members. */
117 ira_allocno_t next_bucket_allocno;
118 ira_allocno_t prev_bucket_allocno;
119 /* Used for temporary purposes. */
120 int temp;
121 /* Used to exclude repeated processing. */
122 int last_process;
123 /* Profitable hard regs available for this pseudo allocation. It
124 means that the set excludes unavailable hard regs and hard regs
125 conflicting with given pseudo. They should be of the allocno
126 class. */
127 HARD_REG_SET profitable_hard_regs;
128 /* The allocno hard registers node. */
129 allocno_hard_regs_node_t hard_regs_node;
130 /* Array of structures allocno_hard_regs_subnode representing
131 given allocno hard registers node (the 1st element in the array)
132 and all its subnodes in the tree (forest) of allocno hard
133 register nodes (see comments above). */
134 int hard_regs_subnodes_start;
135 /* The length of the previous array. */
136 int hard_regs_subnodes_num;
137 /* Records about updating allocno hard reg costs from copies. If
138 the allocno did not get expected hard register, these records are
139 used to restore original hard reg costs of allocnos connected to
140 this allocno by copies. */
141 struct update_cost_record *update_cost_records;
142 /* Threads. We collect allocnos connected by copies into threads
143 and try to assign hard regs to allocnos by threads. */
144 /* Allocno representing all thread. */
145 ira_allocno_t first_thread_allocno;
146 /* Allocnos in thread forms a cycle list through the following
147 member. */
148 ira_allocno_t next_thread_allocno;
149 /* All thread frequency. Defined only for first thread allocno. */
150 int thread_freq;
153 /* See above. */
154 typedef struct allocno_color_data *allocno_color_data_t;
156 /* Container for storing allocno data concerning coloring. */
157 static allocno_color_data_t allocno_color_data;
159 /* Macro to access the data concerning coloring. */
160 #define ALLOCNO_COLOR_DATA(a) ((allocno_color_data_t) ALLOCNO_ADD_DATA (a))
162 /* Used for finding allocno colorability to exclude repeated allocno
163 processing and for updating preferencing to exclude repeated
164 allocno processing during assignment. */
165 static int curr_allocno_process;
167 /* This file contains code for regional graph coloring, spill/restore
168 code placement optimization, and code helping the reload pass to do
169 a better job. */
171 /* Bitmap of allocnos which should be colored. */
172 static bitmap coloring_allocno_bitmap;
174 /* Bitmap of allocnos which should be taken into account during
175 coloring. In general case it contains allocnos from
176 coloring_allocno_bitmap plus other already colored conflicting
177 allocnos. */
178 static bitmap consideration_allocno_bitmap;
180 /* All allocnos sorted according their priorities. */
181 static ira_allocno_t *sorted_allocnos;
183 /* Vec representing the stack of allocnos used during coloring. */
184 static vec<ira_allocno_t> allocno_stack_vec;
186 /* Helper for qsort comparison callbacks - return a positive integer if
187 X > Y, or a negative value otherwise. Use a conditional expression
188 instead of a difference computation to insulate from possible overflow
189 issues, e.g. X - Y < 0 for some X > 0 and Y < 0. */
190 #define SORTGT(x,y) (((x) > (y)) ? 1 : -1)
194 /* Definition of vector of allocno hard registers. */
196 /* Vector of unique allocno hard registers. */
197 static vec<allocno_hard_regs_t> allocno_hard_regs_vec;
199 struct allocno_hard_regs_hasher : nofree_ptr_hash <allocno_hard_regs>
201 static inline hashval_t hash (const allocno_hard_regs *);
202 static inline bool equal (const allocno_hard_regs *,
203 const allocno_hard_regs *);
206 /* Returns hash value for allocno hard registers V. */
207 inline hashval_t
208 allocno_hard_regs_hasher::hash (const allocno_hard_regs *hv)
210 return iterative_hash (&hv->set, sizeof (HARD_REG_SET), 0);
213 /* Compares allocno hard registers V1 and V2. */
214 inline bool
215 allocno_hard_regs_hasher::equal (const allocno_hard_regs *hv1,
216 const allocno_hard_regs *hv2)
218 return hard_reg_set_equal_p (hv1->set, hv2->set);
221 /* Hash table of unique allocno hard registers. */
222 static hash_table<allocno_hard_regs_hasher> *allocno_hard_regs_htab;
224 /* Return allocno hard registers in the hash table equal to HV. */
225 static allocno_hard_regs_t
226 find_hard_regs (allocno_hard_regs_t hv)
228 return allocno_hard_regs_htab->find (hv);
231 /* Insert allocno hard registers HV in the hash table (if it is not
232 there yet) and return the value which in the table. */
233 static allocno_hard_regs_t
234 insert_hard_regs (allocno_hard_regs_t hv)
236 allocno_hard_regs **slot = allocno_hard_regs_htab->find_slot (hv, INSERT);
238 if (*slot == NULL)
239 *slot = hv;
240 return *slot;
243 /* Initialize data concerning allocno hard registers. */
244 static void
245 init_allocno_hard_regs (void)
247 allocno_hard_regs_vec.create (200);
248 allocno_hard_regs_htab
249 = new hash_table<allocno_hard_regs_hasher> (200);
252 /* Add (or update info about) allocno hard registers with SET and
253 COST. */
254 static allocno_hard_regs_t
255 add_allocno_hard_regs (HARD_REG_SET set, int64_t cost)
257 struct allocno_hard_regs temp;
258 allocno_hard_regs_t hv;
260 gcc_assert (! hard_reg_set_empty_p (set));
261 COPY_HARD_REG_SET (temp.set, set);
262 if ((hv = find_hard_regs (&temp)) != NULL)
263 hv->cost += cost;
264 else
266 hv = ((struct allocno_hard_regs *)
267 ira_allocate (sizeof (struct allocno_hard_regs)));
268 COPY_HARD_REG_SET (hv->set, set);
269 hv->cost = cost;
270 allocno_hard_regs_vec.safe_push (hv);
271 insert_hard_regs (hv);
273 return hv;
276 /* Finalize data concerning allocno hard registers. */
277 static void
278 finish_allocno_hard_regs (void)
280 int i;
281 allocno_hard_regs_t hv;
283 for (i = 0;
284 allocno_hard_regs_vec.iterate (i, &hv);
285 i++)
286 ira_free (hv);
287 delete allocno_hard_regs_htab;
288 allocno_hard_regs_htab = NULL;
289 allocno_hard_regs_vec.release ();
292 /* Sort hard regs according to their frequency of usage. */
293 static int
294 allocno_hard_regs_compare (const void *v1p, const void *v2p)
296 allocno_hard_regs_t hv1 = *(const allocno_hard_regs_t *) v1p;
297 allocno_hard_regs_t hv2 = *(const allocno_hard_regs_t *) v2p;
299 if (hv2->cost > hv1->cost)
300 return 1;
301 else if (hv2->cost < hv1->cost)
302 return -1;
303 else
304 return 0;
309 /* Used for finding a common ancestor of two allocno hard registers
310 nodes in the forest. We use the current value of
311 'node_check_tick' to mark all nodes from one node to the top and
312 then walking up from another node until we find a marked node.
314 It is also used to figure out allocno colorability as a mark that
315 we already reset value of member 'conflict_size' for the forest
316 node corresponding to the processed allocno. */
317 static int node_check_tick;
319 /* Roots of the forest containing hard register sets can be assigned
320 to allocnos. */
321 static allocno_hard_regs_node_t hard_regs_roots;
323 /* Definition of vector of allocno hard register nodes. */
325 /* Vector used to create the forest. */
326 static vec<allocno_hard_regs_node_t> hard_regs_node_vec;
328 /* Create and return allocno hard registers node containing allocno
329 hard registers HV. */
330 static allocno_hard_regs_node_t
331 create_new_allocno_hard_regs_node (allocno_hard_regs_t hv)
333 allocno_hard_regs_node_t new_node;
335 new_node = ((struct allocno_hard_regs_node *)
336 ira_allocate (sizeof (struct allocno_hard_regs_node)));
337 new_node->check = 0;
338 new_node->hard_regs = hv;
339 new_node->hard_regs_num = hard_reg_set_size (hv->set);
340 new_node->first = NULL;
341 new_node->used_p = false;
342 return new_node;
345 /* Add allocno hard registers node NEW_NODE to the forest on its level
346 given by ROOTS. */
347 static void
348 add_new_allocno_hard_regs_node_to_forest (allocno_hard_regs_node_t *roots,
349 allocno_hard_regs_node_t new_node)
351 new_node->next = *roots;
352 if (new_node->next != NULL)
353 new_node->next->prev = new_node;
354 new_node->prev = NULL;
355 *roots = new_node;
358 /* Add allocno hard registers HV (or its best approximation if it is
359 not possible) to the forest on its level given by ROOTS. */
360 static void
361 add_allocno_hard_regs_to_forest (allocno_hard_regs_node_t *roots,
362 allocno_hard_regs_t hv)
364 unsigned int i, start;
365 allocno_hard_regs_node_t node, prev, new_node;
366 HARD_REG_SET temp_set;
367 allocno_hard_regs_t hv2;
369 start = hard_regs_node_vec.length ();
370 for (node = *roots; node != NULL; node = node->next)
372 if (hard_reg_set_equal_p (hv->set, node->hard_regs->set))
373 return;
374 if (hard_reg_set_subset_p (hv->set, node->hard_regs->set))
376 add_allocno_hard_regs_to_forest (&node->first, hv);
377 return;
379 if (hard_reg_set_subset_p (node->hard_regs->set, hv->set))
380 hard_regs_node_vec.safe_push (node);
381 else if (hard_reg_set_intersect_p (hv->set, node->hard_regs->set))
383 COPY_HARD_REG_SET (temp_set, hv->set);
384 AND_HARD_REG_SET (temp_set, node->hard_regs->set);
385 hv2 = add_allocno_hard_regs (temp_set, hv->cost);
386 add_allocno_hard_regs_to_forest (&node->first, hv2);
389 if (hard_regs_node_vec.length ()
390 > start + 1)
392 /* Create a new node which contains nodes in hard_regs_node_vec. */
393 CLEAR_HARD_REG_SET (temp_set);
394 for (i = start;
395 i < hard_regs_node_vec.length ();
396 i++)
398 node = hard_regs_node_vec[i];
399 IOR_HARD_REG_SET (temp_set, node->hard_regs->set);
401 hv = add_allocno_hard_regs (temp_set, hv->cost);
402 new_node = create_new_allocno_hard_regs_node (hv);
403 prev = NULL;
404 for (i = start;
405 i < hard_regs_node_vec.length ();
406 i++)
408 node = hard_regs_node_vec[i];
409 if (node->prev == NULL)
410 *roots = node->next;
411 else
412 node->prev->next = node->next;
413 if (node->next != NULL)
414 node->next->prev = node->prev;
415 if (prev == NULL)
416 new_node->first = node;
417 else
418 prev->next = node;
419 node->prev = prev;
420 node->next = NULL;
421 prev = node;
423 add_new_allocno_hard_regs_node_to_forest (roots, new_node);
425 hard_regs_node_vec.truncate (start);
428 /* Add allocno hard registers nodes starting with the forest level
429 given by FIRST which contains biggest set inside SET. */
430 static void
431 collect_allocno_hard_regs_cover (allocno_hard_regs_node_t first,
432 HARD_REG_SET set)
434 allocno_hard_regs_node_t node;
436 ira_assert (first != NULL);
437 for (node = first; node != NULL; node = node->next)
438 if (hard_reg_set_subset_p (node->hard_regs->set, set))
439 hard_regs_node_vec.safe_push (node);
440 else if (hard_reg_set_intersect_p (set, node->hard_regs->set))
441 collect_allocno_hard_regs_cover (node->first, set);
444 /* Set up field parent as PARENT in all allocno hard registers nodes
445 in forest given by FIRST. */
446 static void
447 setup_allocno_hard_regs_nodes_parent (allocno_hard_regs_node_t first,
448 allocno_hard_regs_node_t parent)
450 allocno_hard_regs_node_t node;
452 for (node = first; node != NULL; node = node->next)
454 node->parent = parent;
455 setup_allocno_hard_regs_nodes_parent (node->first, node);
459 /* Return allocno hard registers node which is a first common ancestor
460 node of FIRST and SECOND in the forest. */
461 static allocno_hard_regs_node_t
462 first_common_ancestor_node (allocno_hard_regs_node_t first,
463 allocno_hard_regs_node_t second)
465 allocno_hard_regs_node_t node;
467 node_check_tick++;
468 for (node = first; node != NULL; node = node->parent)
469 node->check = node_check_tick;
470 for (node = second; node != NULL; node = node->parent)
471 if (node->check == node_check_tick)
472 return node;
473 return first_common_ancestor_node (second, first);
476 /* Print hard reg set SET to F. */
477 static void
478 print_hard_reg_set (FILE *f, HARD_REG_SET set, bool new_line_p)
480 int i, start;
482 for (start = -1, i = 0; i < FIRST_PSEUDO_REGISTER; i++)
484 if (TEST_HARD_REG_BIT (set, i))
486 if (i == 0 || ! TEST_HARD_REG_BIT (set, i - 1))
487 start = i;
489 if (start >= 0
490 && (i == FIRST_PSEUDO_REGISTER - 1 || ! TEST_HARD_REG_BIT (set, i)))
492 if (start == i - 1)
493 fprintf (f, " %d", start);
494 else if (start == i - 2)
495 fprintf (f, " %d %d", start, start + 1);
496 else
497 fprintf (f, " %d-%d", start, i - 1);
498 start = -1;
501 if (new_line_p)
502 fprintf (f, "\n");
505 /* Print allocno hard register subforest given by ROOTS and its LEVEL
506 to F. */
507 static void
508 print_hard_regs_subforest (FILE *f, allocno_hard_regs_node_t roots,
509 int level)
511 int i;
512 allocno_hard_regs_node_t node;
514 for (node = roots; node != NULL; node = node->next)
516 fprintf (f, " ");
517 for (i = 0; i < level * 2; i++)
518 fprintf (f, " ");
519 fprintf (f, "%d:(", node->preorder_num);
520 print_hard_reg_set (f, node->hard_regs->set, false);
521 fprintf (f, ")@%" PRId64"\n", node->hard_regs->cost);
522 print_hard_regs_subforest (f, node->first, level + 1);
526 /* Print the allocno hard register forest to F. */
527 static void
528 print_hard_regs_forest (FILE *f)
530 fprintf (f, " Hard reg set forest:\n");
531 print_hard_regs_subforest (f, hard_regs_roots, 1);
534 /* Print the allocno hard register forest to stderr. */
535 void
536 ira_debug_hard_regs_forest (void)
538 print_hard_regs_forest (stderr);
541 /* Remove unused allocno hard registers nodes from forest given by its
542 *ROOTS. */
543 static void
544 remove_unused_allocno_hard_regs_nodes (allocno_hard_regs_node_t *roots)
546 allocno_hard_regs_node_t node, prev, next, last;
548 for (prev = NULL, node = *roots; node != NULL; node = next)
550 next = node->next;
551 if (node->used_p)
553 remove_unused_allocno_hard_regs_nodes (&node->first);
554 prev = node;
556 else
558 for (last = node->first;
559 last != NULL && last->next != NULL;
560 last = last->next)
562 if (last != NULL)
564 if (prev == NULL)
565 *roots = node->first;
566 else
567 prev->next = node->first;
568 if (next != NULL)
569 next->prev = last;
570 last->next = next;
571 next = node->first;
573 else
575 if (prev == NULL)
576 *roots = next;
577 else
578 prev->next = next;
579 if (next != NULL)
580 next->prev = prev;
582 ira_free (node);
587 /* Set up fields preorder_num starting with START_NUM in all allocno
588 hard registers nodes in forest given by FIRST. Return biggest set
589 PREORDER_NUM increased by 1. */
590 static int
591 enumerate_allocno_hard_regs_nodes (allocno_hard_regs_node_t first,
592 allocno_hard_regs_node_t parent,
593 int start_num)
595 allocno_hard_regs_node_t node;
597 for (node = first; node != NULL; node = node->next)
599 node->preorder_num = start_num++;
600 node->parent = parent;
601 start_num = enumerate_allocno_hard_regs_nodes (node->first, node,
602 start_num);
604 return start_num;
607 /* Number of allocno hard registers nodes in the forest. */
608 static int allocno_hard_regs_nodes_num;
610 /* Table preorder number of allocno hard registers node in the forest
611 -> the allocno hard registers node. */
612 static allocno_hard_regs_node_t *allocno_hard_regs_nodes;
614 /* See below. */
615 typedef struct allocno_hard_regs_subnode *allocno_hard_regs_subnode_t;
617 /* The structure is used to describes all subnodes (not only immediate
618 ones) in the mentioned above tree for given allocno hard register
619 node. The usage of such data accelerates calculation of
620 colorability of given allocno. */
621 struct allocno_hard_regs_subnode
623 /* The conflict size of conflicting allocnos whose hard register
624 sets are equal sets (plus supersets if given node is given
625 allocno hard registers node) of one in the given node. */
626 int left_conflict_size;
627 /* The summary conflict size of conflicting allocnos whose hard
628 register sets are strict subsets of one in the given node.
629 Overall conflict size is
630 left_conflict_subnodes_size
631 + MIN (max_node_impact - left_conflict_subnodes_size,
632 left_conflict_size)
634 short left_conflict_subnodes_size;
635 short max_node_impact;
638 /* Container for hard regs subnodes of all allocnos. */
639 static allocno_hard_regs_subnode_t allocno_hard_regs_subnodes;
641 /* Table (preorder number of allocno hard registers node in the
642 forest, preorder number of allocno hard registers subnode) -> index
643 of the subnode relative to the node. -1 if it is not a
644 subnode. */
645 static int *allocno_hard_regs_subnode_index;
647 /* Setup arrays ALLOCNO_HARD_REGS_NODES and
648 ALLOCNO_HARD_REGS_SUBNODE_INDEX. */
649 static void
650 setup_allocno_hard_regs_subnode_index (allocno_hard_regs_node_t first)
652 allocno_hard_regs_node_t node, parent;
653 int index;
655 for (node = first; node != NULL; node = node->next)
657 allocno_hard_regs_nodes[node->preorder_num] = node;
658 for (parent = node; parent != NULL; parent = parent->parent)
660 index = parent->preorder_num * allocno_hard_regs_nodes_num;
661 allocno_hard_regs_subnode_index[index + node->preorder_num]
662 = node->preorder_num - parent->preorder_num;
664 setup_allocno_hard_regs_subnode_index (node->first);
668 /* Count all allocno hard registers nodes in tree ROOT. */
669 static int
670 get_allocno_hard_regs_subnodes_num (allocno_hard_regs_node_t root)
672 int len = 1;
674 for (root = root->first; root != NULL; root = root->next)
675 len += get_allocno_hard_regs_subnodes_num (root);
676 return len;
679 /* Build the forest of allocno hard registers nodes and assign each
680 allocno a node from the forest. */
681 static void
682 form_allocno_hard_regs_nodes_forest (void)
684 unsigned int i, j, size, len;
685 int start;
686 ira_allocno_t a;
687 allocno_hard_regs_t hv;
688 bitmap_iterator bi;
689 HARD_REG_SET temp;
690 allocno_hard_regs_node_t node, allocno_hard_regs_node;
691 allocno_color_data_t allocno_data;
693 node_check_tick = 0;
694 init_allocno_hard_regs ();
695 hard_regs_roots = NULL;
696 hard_regs_node_vec.create (100);
697 for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
698 if (! TEST_HARD_REG_BIT (ira_no_alloc_regs, i))
700 CLEAR_HARD_REG_SET (temp);
701 SET_HARD_REG_BIT (temp, i);
702 hv = add_allocno_hard_regs (temp, 0);
703 node = create_new_allocno_hard_regs_node (hv);
704 add_new_allocno_hard_regs_node_to_forest (&hard_regs_roots, node);
706 start = allocno_hard_regs_vec.length ();
707 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap, 0, i, bi)
709 a = ira_allocnos[i];
710 allocno_data = ALLOCNO_COLOR_DATA (a);
712 if (hard_reg_set_empty_p (allocno_data->profitable_hard_regs))
713 continue;
714 hv = (add_allocno_hard_regs
715 (allocno_data->profitable_hard_regs,
716 ALLOCNO_MEMORY_COST (a) - ALLOCNO_CLASS_COST (a)));
718 SET_HARD_REG_SET (temp);
719 AND_COMPL_HARD_REG_SET (temp, ira_no_alloc_regs);
720 add_allocno_hard_regs (temp, 0);
721 qsort (allocno_hard_regs_vec.address () + start,
722 allocno_hard_regs_vec.length () - start,
723 sizeof (allocno_hard_regs_t), allocno_hard_regs_compare);
724 for (i = start;
725 allocno_hard_regs_vec.iterate (i, &hv);
726 i++)
728 add_allocno_hard_regs_to_forest (&hard_regs_roots, hv);
729 ira_assert (hard_regs_node_vec.length () == 0);
731 /* We need to set up parent fields for right work of
732 first_common_ancestor_node. */
733 setup_allocno_hard_regs_nodes_parent (hard_regs_roots, NULL);
734 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap, 0, i, bi)
736 a = ira_allocnos[i];
737 allocno_data = ALLOCNO_COLOR_DATA (a);
738 if (hard_reg_set_empty_p (allocno_data->profitable_hard_regs))
739 continue;
740 hard_regs_node_vec.truncate (0);
741 collect_allocno_hard_regs_cover (hard_regs_roots,
742 allocno_data->profitable_hard_regs);
743 allocno_hard_regs_node = NULL;
744 for (j = 0; hard_regs_node_vec.iterate (j, &node); j++)
745 allocno_hard_regs_node
746 = (j == 0
747 ? node
748 : first_common_ancestor_node (node, allocno_hard_regs_node));
749 /* That is a temporary storage. */
750 allocno_hard_regs_node->used_p = true;
751 allocno_data->hard_regs_node = allocno_hard_regs_node;
753 ira_assert (hard_regs_roots->next == NULL);
754 hard_regs_roots->used_p = true;
755 remove_unused_allocno_hard_regs_nodes (&hard_regs_roots);
756 allocno_hard_regs_nodes_num
757 = enumerate_allocno_hard_regs_nodes (hard_regs_roots, NULL, 0);
758 allocno_hard_regs_nodes
759 = ((allocno_hard_regs_node_t *)
760 ira_allocate (allocno_hard_regs_nodes_num
761 * sizeof (allocno_hard_regs_node_t)));
762 size = allocno_hard_regs_nodes_num * allocno_hard_regs_nodes_num;
763 allocno_hard_regs_subnode_index
764 = (int *) ira_allocate (size * sizeof (int));
765 for (i = 0; i < size; i++)
766 allocno_hard_regs_subnode_index[i] = -1;
767 setup_allocno_hard_regs_subnode_index (hard_regs_roots);
768 start = 0;
769 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap, 0, i, bi)
771 a = ira_allocnos[i];
772 allocno_data = ALLOCNO_COLOR_DATA (a);
773 if (hard_reg_set_empty_p (allocno_data->profitable_hard_regs))
774 continue;
775 len = get_allocno_hard_regs_subnodes_num (allocno_data->hard_regs_node);
776 allocno_data->hard_regs_subnodes_start = start;
777 allocno_data->hard_regs_subnodes_num = len;
778 start += len;
780 allocno_hard_regs_subnodes
781 = ((allocno_hard_regs_subnode_t)
782 ira_allocate (sizeof (struct allocno_hard_regs_subnode) * start));
783 hard_regs_node_vec.release ();
786 /* Free tree of allocno hard registers nodes given by its ROOT. */
787 static void
788 finish_allocno_hard_regs_nodes_tree (allocno_hard_regs_node_t root)
790 allocno_hard_regs_node_t child, next;
792 for (child = root->first; child != NULL; child = next)
794 next = child->next;
795 finish_allocno_hard_regs_nodes_tree (child);
797 ira_free (root);
800 /* Finish work with the forest of allocno hard registers nodes. */
801 static void
802 finish_allocno_hard_regs_nodes_forest (void)
804 allocno_hard_regs_node_t node, next;
806 ira_free (allocno_hard_regs_subnodes);
807 for (node = hard_regs_roots; node != NULL; node = next)
809 next = node->next;
810 finish_allocno_hard_regs_nodes_tree (node);
812 ira_free (allocno_hard_regs_nodes);
813 ira_free (allocno_hard_regs_subnode_index);
814 finish_allocno_hard_regs ();
817 /* Set up left conflict sizes and left conflict subnodes sizes of hard
818 registers subnodes of allocno A. Return TRUE if allocno A is
819 trivially colorable. */
820 static bool
821 setup_left_conflict_sizes_p (ira_allocno_t a)
823 int i, k, nobj, start;
824 int conflict_size, left_conflict_subnodes_size, node_preorder_num;
825 allocno_color_data_t data;
826 HARD_REG_SET profitable_hard_regs;
827 allocno_hard_regs_subnode_t subnodes;
828 allocno_hard_regs_node_t node;
829 HARD_REG_SET node_set;
831 nobj = ALLOCNO_NUM_OBJECTS (a);
832 data = ALLOCNO_COLOR_DATA (a);
833 subnodes = allocno_hard_regs_subnodes + data->hard_regs_subnodes_start;
834 COPY_HARD_REG_SET (profitable_hard_regs, data->profitable_hard_regs);
835 node = data->hard_regs_node;
836 node_preorder_num = node->preorder_num;
837 COPY_HARD_REG_SET (node_set, node->hard_regs->set);
838 node_check_tick++;
839 for (k = 0; k < nobj; k++)
841 ira_object_t obj = ALLOCNO_OBJECT (a, k);
842 ira_object_t conflict_obj;
843 ira_object_conflict_iterator oci;
845 FOR_EACH_OBJECT_CONFLICT (obj, conflict_obj, oci)
847 int size;
848 ira_allocno_t conflict_a = OBJECT_ALLOCNO (conflict_obj);
849 allocno_hard_regs_node_t conflict_node, temp_node;
850 HARD_REG_SET conflict_node_set;
851 allocno_color_data_t conflict_data;
853 conflict_data = ALLOCNO_COLOR_DATA (conflict_a);
854 if (! ALLOCNO_COLOR_DATA (conflict_a)->in_graph_p
855 || ! hard_reg_set_intersect_p (profitable_hard_regs,
856 conflict_data
857 ->profitable_hard_regs))
858 continue;
859 conflict_node = conflict_data->hard_regs_node;
860 COPY_HARD_REG_SET (conflict_node_set, conflict_node->hard_regs->set);
861 if (hard_reg_set_subset_p (node_set, conflict_node_set))
862 temp_node = node;
863 else
865 ira_assert (hard_reg_set_subset_p (conflict_node_set, node_set));
866 temp_node = conflict_node;
868 if (temp_node->check != node_check_tick)
870 temp_node->check = node_check_tick;
871 temp_node->conflict_size = 0;
873 size = (ira_reg_class_max_nregs
874 [ALLOCNO_CLASS (conflict_a)][ALLOCNO_MODE (conflict_a)]);
875 if (ALLOCNO_NUM_OBJECTS (conflict_a) > 1)
876 /* We will deal with the subwords individually. */
877 size = 1;
878 temp_node->conflict_size += size;
881 for (i = 0; i < data->hard_regs_subnodes_num; i++)
883 allocno_hard_regs_node_t temp_node;
885 temp_node = allocno_hard_regs_nodes[i + node_preorder_num];
886 ira_assert (temp_node->preorder_num == i + node_preorder_num);
887 subnodes[i].left_conflict_size = (temp_node->check != node_check_tick
888 ? 0 : temp_node->conflict_size);
889 if (hard_reg_set_subset_p (temp_node->hard_regs->set,
890 profitable_hard_regs))
891 subnodes[i].max_node_impact = temp_node->hard_regs_num;
892 else
894 HARD_REG_SET temp_set;
895 int j, n, hard_regno;
896 enum reg_class aclass;
898 COPY_HARD_REG_SET (temp_set, temp_node->hard_regs->set);
899 AND_HARD_REG_SET (temp_set, profitable_hard_regs);
900 aclass = ALLOCNO_CLASS (a);
901 for (n = 0, j = ira_class_hard_regs_num[aclass] - 1; j >= 0; j--)
903 hard_regno = ira_class_hard_regs[aclass][j];
904 if (TEST_HARD_REG_BIT (temp_set, hard_regno))
905 n++;
907 subnodes[i].max_node_impact = n;
909 subnodes[i].left_conflict_subnodes_size = 0;
911 start = node_preorder_num * allocno_hard_regs_nodes_num;
912 for (i = data->hard_regs_subnodes_num - 1; i > 0; i--)
914 int size, parent_i;
915 allocno_hard_regs_node_t parent;
917 size = (subnodes[i].left_conflict_subnodes_size
918 + MIN (subnodes[i].max_node_impact
919 - subnodes[i].left_conflict_subnodes_size,
920 subnodes[i].left_conflict_size));
921 parent = allocno_hard_regs_nodes[i + node_preorder_num]->parent;
922 gcc_checking_assert(parent);
923 parent_i
924 = allocno_hard_regs_subnode_index[start + parent->preorder_num];
925 gcc_checking_assert(parent_i >= 0);
926 subnodes[parent_i].left_conflict_subnodes_size += size;
928 left_conflict_subnodes_size = subnodes[0].left_conflict_subnodes_size;
929 conflict_size
930 = (left_conflict_subnodes_size
931 + MIN (subnodes[0].max_node_impact - left_conflict_subnodes_size,
932 subnodes[0].left_conflict_size));
933 conflict_size += ira_reg_class_max_nregs[ALLOCNO_CLASS (a)][ALLOCNO_MODE (a)];
934 data->colorable_p = conflict_size <= data->available_regs_num;
935 return data->colorable_p;
938 /* Update left conflict sizes of hard registers subnodes of allocno A
939 after removing allocno REMOVED_A with SIZE from the conflict graph.
940 Return TRUE if A is trivially colorable. */
941 static bool
942 update_left_conflict_sizes_p (ira_allocno_t a,
943 ira_allocno_t removed_a, int size)
945 int i, conflict_size, before_conflict_size, diff, start;
946 int node_preorder_num, parent_i;
947 allocno_hard_regs_node_t node, removed_node, parent;
948 allocno_hard_regs_subnode_t subnodes;
949 allocno_color_data_t data = ALLOCNO_COLOR_DATA (a);
951 ira_assert (! data->colorable_p);
952 node = data->hard_regs_node;
953 node_preorder_num = node->preorder_num;
954 removed_node = ALLOCNO_COLOR_DATA (removed_a)->hard_regs_node;
955 ira_assert (hard_reg_set_subset_p (removed_node->hard_regs->set,
956 node->hard_regs->set)
957 || hard_reg_set_subset_p (node->hard_regs->set,
958 removed_node->hard_regs->set));
959 start = node_preorder_num * allocno_hard_regs_nodes_num;
960 i = allocno_hard_regs_subnode_index[start + removed_node->preorder_num];
961 if (i < 0)
962 i = 0;
963 subnodes = allocno_hard_regs_subnodes + data->hard_regs_subnodes_start;
964 before_conflict_size
965 = (subnodes[i].left_conflict_subnodes_size
966 + MIN (subnodes[i].max_node_impact
967 - subnodes[i].left_conflict_subnodes_size,
968 subnodes[i].left_conflict_size));
969 subnodes[i].left_conflict_size -= size;
970 for (;;)
972 conflict_size
973 = (subnodes[i].left_conflict_subnodes_size
974 + MIN (subnodes[i].max_node_impact
975 - subnodes[i].left_conflict_subnodes_size,
976 subnodes[i].left_conflict_size));
977 if ((diff = before_conflict_size - conflict_size) == 0)
978 break;
979 ira_assert (conflict_size < before_conflict_size);
980 parent = allocno_hard_regs_nodes[i + node_preorder_num]->parent;
981 if (parent == NULL)
982 break;
983 parent_i
984 = allocno_hard_regs_subnode_index[start + parent->preorder_num];
985 if (parent_i < 0)
986 break;
987 i = parent_i;
988 before_conflict_size
989 = (subnodes[i].left_conflict_subnodes_size
990 + MIN (subnodes[i].max_node_impact
991 - subnodes[i].left_conflict_subnodes_size,
992 subnodes[i].left_conflict_size));
993 subnodes[i].left_conflict_subnodes_size -= diff;
995 if (i != 0
996 || (conflict_size
997 + ira_reg_class_max_nregs[ALLOCNO_CLASS (a)][ALLOCNO_MODE (a)]
998 > data->available_regs_num))
999 return false;
1000 data->colorable_p = true;
1001 return true;
1004 /* Return true if allocno A has empty profitable hard regs. */
1005 static bool
1006 empty_profitable_hard_regs (ira_allocno_t a)
1008 allocno_color_data_t data = ALLOCNO_COLOR_DATA (a);
1010 return hard_reg_set_empty_p (data->profitable_hard_regs);
1013 /* Set up profitable hard registers for each allocno being
1014 colored. */
1015 static void
1016 setup_profitable_hard_regs (void)
1018 unsigned int i;
1019 int j, k, nobj, hard_regno, nregs, class_size;
1020 ira_allocno_t a;
1021 bitmap_iterator bi;
1022 enum reg_class aclass;
1023 machine_mode mode;
1024 allocno_color_data_t data;
1026 /* Initial set up from allocno classes and explicitly conflicting
1027 hard regs. */
1028 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap, 0, i, bi)
1030 a = ira_allocnos[i];
1031 if ((aclass = ALLOCNO_CLASS (a)) == NO_REGS)
1032 continue;
1033 data = ALLOCNO_COLOR_DATA (a);
1034 if (ALLOCNO_UPDATED_HARD_REG_COSTS (a) == NULL
1035 && ALLOCNO_CLASS_COST (a) > ALLOCNO_MEMORY_COST (a)
1036 /* Do not empty profitable regs for static chain pointer
1037 pseudo when non-local goto is used. */
1038 && ! non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a)))
1039 CLEAR_HARD_REG_SET (data->profitable_hard_regs);
1040 else
1042 mode = ALLOCNO_MODE (a);
1043 COPY_HARD_REG_SET (data->profitable_hard_regs,
1044 ira_useful_class_mode_regs[aclass][mode]);
1045 nobj = ALLOCNO_NUM_OBJECTS (a);
1046 for (k = 0; k < nobj; k++)
1048 ira_object_t obj = ALLOCNO_OBJECT (a, k);
1050 AND_COMPL_HARD_REG_SET (data->profitable_hard_regs,
1051 OBJECT_TOTAL_CONFLICT_HARD_REGS (obj));
1055 /* Exclude hard regs already assigned for conflicting objects. */
1056 EXECUTE_IF_SET_IN_BITMAP (consideration_allocno_bitmap, 0, i, bi)
1058 a = ira_allocnos[i];
1059 if ((aclass = ALLOCNO_CLASS (a)) == NO_REGS
1060 || ! ALLOCNO_ASSIGNED_P (a)
1061 || (hard_regno = ALLOCNO_HARD_REGNO (a)) < 0)
1062 continue;
1063 mode = ALLOCNO_MODE (a);
1064 nregs = hard_regno_nregs (hard_regno, mode);
1065 nobj = ALLOCNO_NUM_OBJECTS (a);
1066 for (k = 0; k < nobj; k++)
1068 ira_object_t obj = ALLOCNO_OBJECT (a, k);
1069 ira_object_t conflict_obj;
1070 ira_object_conflict_iterator oci;
1072 FOR_EACH_OBJECT_CONFLICT (obj, conflict_obj, oci)
1074 ira_allocno_t conflict_a = OBJECT_ALLOCNO (conflict_obj);
1076 /* We can process the conflict allocno repeatedly with
1077 the same result. */
1078 if (nregs == nobj && nregs > 1)
1080 int num = OBJECT_SUBWORD (conflict_obj);
1082 if (REG_WORDS_BIG_ENDIAN)
1083 CLEAR_HARD_REG_BIT
1084 (ALLOCNO_COLOR_DATA (conflict_a)->profitable_hard_regs,
1085 hard_regno + nobj - num - 1);
1086 else
1087 CLEAR_HARD_REG_BIT
1088 (ALLOCNO_COLOR_DATA (conflict_a)->profitable_hard_regs,
1089 hard_regno + num);
1091 else
1092 AND_COMPL_HARD_REG_SET
1093 (ALLOCNO_COLOR_DATA (conflict_a)->profitable_hard_regs,
1094 ira_reg_mode_hard_regset[hard_regno][mode]);
1098 /* Exclude too costly hard regs. */
1099 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap, 0, i, bi)
1101 int min_cost = INT_MAX;
1102 int *costs;
1104 a = ira_allocnos[i];
1105 if ((aclass = ALLOCNO_CLASS (a)) == NO_REGS
1106 || empty_profitable_hard_regs (a))
1107 continue;
1108 data = ALLOCNO_COLOR_DATA (a);
1109 mode = ALLOCNO_MODE (a);
1110 if ((costs = ALLOCNO_UPDATED_HARD_REG_COSTS (a)) != NULL
1111 || (costs = ALLOCNO_HARD_REG_COSTS (a)) != NULL)
1113 class_size = ira_class_hard_regs_num[aclass];
1114 for (j = 0; j < class_size; j++)
1116 hard_regno = ira_class_hard_regs[aclass][j];
1117 if (! TEST_HARD_REG_BIT (data->profitable_hard_regs,
1118 hard_regno))
1119 continue;
1120 if (ALLOCNO_UPDATED_MEMORY_COST (a) < costs[j]
1121 /* Do not remove HARD_REGNO for static chain pointer
1122 pseudo when non-local goto is used. */
1123 && ! non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a)))
1124 CLEAR_HARD_REG_BIT (data->profitable_hard_regs,
1125 hard_regno);
1126 else if (min_cost > costs[j])
1127 min_cost = costs[j];
1130 else if (ALLOCNO_UPDATED_MEMORY_COST (a)
1131 < ALLOCNO_UPDATED_CLASS_COST (a)
1132 /* Do not empty profitable regs for static chain
1133 pointer pseudo when non-local goto is used. */
1134 && ! non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a)))
1135 CLEAR_HARD_REG_SET (data->profitable_hard_regs);
1136 if (ALLOCNO_UPDATED_CLASS_COST (a) > min_cost)
1137 ALLOCNO_UPDATED_CLASS_COST (a) = min_cost;
1143 /* This page contains functions used to choose hard registers for
1144 allocnos. */
1146 /* Pool for update cost records. */
1147 static object_allocator<update_cost_record> update_cost_record_pool
1148 ("update cost records");
1150 /* Return new update cost record with given params. */
1151 static struct update_cost_record *
1152 get_update_cost_record (int hard_regno, int divisor,
1153 struct update_cost_record *next)
1155 struct update_cost_record *record;
1157 record = update_cost_record_pool.allocate ();
1158 record->hard_regno = hard_regno;
1159 record->divisor = divisor;
1160 record->next = next;
1161 return record;
1164 /* Free memory for all records in LIST. */
1165 static void
1166 free_update_cost_record_list (struct update_cost_record *list)
1168 struct update_cost_record *next;
1170 while (list != NULL)
1172 next = list->next;
1173 update_cost_record_pool.remove (list);
1174 list = next;
1178 /* Free memory allocated for all update cost records. */
1179 static void
1180 finish_update_cost_records (void)
1182 update_cost_record_pool.release ();
1185 /* Array whose element value is TRUE if the corresponding hard
1186 register was already allocated for an allocno. */
1187 static bool allocated_hardreg_p[FIRST_PSEUDO_REGISTER];
1189 /* Describes one element in a queue of allocnos whose costs need to be
1190 updated. Each allocno in the queue is known to have an allocno
1191 class. */
1192 struct update_cost_queue_elem
1194 /* This element is in the queue iff CHECK == update_cost_check. */
1195 int check;
1197 /* COST_HOP_DIVISOR**N, where N is the length of the shortest path
1198 connecting this allocno to the one being allocated. */
1199 int divisor;
1201 /* Allocno from which we are chaining costs of connected allocnos.
1202 It is used not go back in graph of allocnos connected by
1203 copies. */
1204 ira_allocno_t from;
1206 /* The next allocno in the queue, or null if this is the last element. */
1207 ira_allocno_t next;
1210 /* The first element in a queue of allocnos whose copy costs need to be
1211 updated. Null if the queue is empty. */
1212 static ira_allocno_t update_cost_queue;
1214 /* The last element in the queue described by update_cost_queue.
1215 Not valid if update_cost_queue is null. */
1216 static struct update_cost_queue_elem *update_cost_queue_tail;
1218 /* A pool of elements in the queue described by update_cost_queue.
1219 Elements are indexed by ALLOCNO_NUM. */
1220 static struct update_cost_queue_elem *update_cost_queue_elems;
1222 /* The current value of update_costs_from_copies call count. */
1223 static int update_cost_check;
1225 /* Allocate and initialize data necessary for function
1226 update_costs_from_copies. */
1227 static void
1228 initiate_cost_update (void)
1230 size_t size;
1232 size = ira_allocnos_num * sizeof (struct update_cost_queue_elem);
1233 update_cost_queue_elems
1234 = (struct update_cost_queue_elem *) ira_allocate (size);
1235 memset (update_cost_queue_elems, 0, size);
1236 update_cost_check = 0;
1239 /* Deallocate data used by function update_costs_from_copies. */
1240 static void
1241 finish_cost_update (void)
1243 ira_free (update_cost_queue_elems);
1244 finish_update_cost_records ();
1247 /* When we traverse allocnos to update hard register costs, the cost
1248 divisor will be multiplied by the following macro value for each
1249 hop from given allocno to directly connected allocnos. */
1250 #define COST_HOP_DIVISOR 4
1252 /* Start a new cost-updating pass. */
1253 static void
1254 start_update_cost (void)
1256 update_cost_check++;
1257 update_cost_queue = NULL;
1260 /* Add (ALLOCNO, FROM, DIVISOR) to the end of update_cost_queue, unless
1261 ALLOCNO is already in the queue, or has NO_REGS class. */
1262 static inline void
1263 queue_update_cost (ira_allocno_t allocno, ira_allocno_t from, int divisor)
1265 struct update_cost_queue_elem *elem;
1267 elem = &update_cost_queue_elems[ALLOCNO_NUM (allocno)];
1268 if (elem->check != update_cost_check
1269 && ALLOCNO_CLASS (allocno) != NO_REGS)
1271 elem->check = update_cost_check;
1272 elem->from = from;
1273 elem->divisor = divisor;
1274 elem->next = NULL;
1275 if (update_cost_queue == NULL)
1276 update_cost_queue = allocno;
1277 else
1278 update_cost_queue_tail->next = allocno;
1279 update_cost_queue_tail = elem;
1283 /* Try to remove the first element from update_cost_queue. Return
1284 false if the queue was empty, otherwise make (*ALLOCNO, *FROM,
1285 *DIVISOR) describe the removed element. */
1286 static inline bool
1287 get_next_update_cost (ira_allocno_t *allocno, ira_allocno_t *from, int *divisor)
1289 struct update_cost_queue_elem *elem;
1291 if (update_cost_queue == NULL)
1292 return false;
1294 *allocno = update_cost_queue;
1295 elem = &update_cost_queue_elems[ALLOCNO_NUM (*allocno)];
1296 *from = elem->from;
1297 *divisor = elem->divisor;
1298 update_cost_queue = elem->next;
1299 return true;
1302 /* Increase costs of HARD_REGNO by UPDATE_COST and conflict cost by
1303 UPDATE_CONFLICT_COST for ALLOCNO. Return true if we really
1304 modified the cost. */
1305 static bool
1306 update_allocno_cost (ira_allocno_t allocno, int hard_regno,
1307 int update_cost, int update_conflict_cost)
1309 int i;
1310 enum reg_class aclass = ALLOCNO_CLASS (allocno);
1312 i = ira_class_hard_reg_index[aclass][hard_regno];
1313 if (i < 0)
1314 return false;
1315 ira_allocate_and_set_or_copy_costs
1316 (&ALLOCNO_UPDATED_HARD_REG_COSTS (allocno), aclass,
1317 ALLOCNO_UPDATED_CLASS_COST (allocno),
1318 ALLOCNO_HARD_REG_COSTS (allocno));
1319 ira_allocate_and_set_or_copy_costs
1320 (&ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (allocno),
1321 aclass, 0, ALLOCNO_CONFLICT_HARD_REG_COSTS (allocno));
1322 ALLOCNO_UPDATED_HARD_REG_COSTS (allocno)[i] += update_cost;
1323 ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (allocno)[i] += update_conflict_cost;
1324 return true;
1327 /* Update (decrease if DECR_P) HARD_REGNO cost of allocnos connected
1328 by copies to ALLOCNO to increase chances to remove some copies as
1329 the result of subsequent assignment. Record cost updates if
1330 RECORD_P is true. */
1331 static void
1332 update_costs_from_allocno (ira_allocno_t allocno, int hard_regno,
1333 int divisor, bool decr_p, bool record_p)
1335 int cost, update_cost, update_conflict_cost;
1336 machine_mode mode;
1337 enum reg_class rclass, aclass;
1338 ira_allocno_t another_allocno, from = NULL;
1339 ira_copy_t cp, next_cp;
1341 rclass = REGNO_REG_CLASS (hard_regno);
1344 mode = ALLOCNO_MODE (allocno);
1345 ira_init_register_move_cost_if_necessary (mode);
1346 for (cp = ALLOCNO_COPIES (allocno); cp != NULL; cp = next_cp)
1348 if (cp->first == allocno)
1350 next_cp = cp->next_first_allocno_copy;
1351 another_allocno = cp->second;
1353 else if (cp->second == allocno)
1355 next_cp = cp->next_second_allocno_copy;
1356 another_allocno = cp->first;
1358 else
1359 gcc_unreachable ();
1361 if (another_allocno == from)
1362 continue;
1364 aclass = ALLOCNO_CLASS (another_allocno);
1365 if (! TEST_HARD_REG_BIT (reg_class_contents[aclass],
1366 hard_regno)
1367 || ALLOCNO_ASSIGNED_P (another_allocno))
1368 continue;
1370 /* If we have different modes use the smallest one. It is
1371 a sub-register move. It is hard to predict what LRA
1372 will reload (the pseudo or its sub-register) but LRA
1373 will try to minimize the data movement. Also for some
1374 register classes bigger modes might be invalid,
1375 e.g. DImode for AREG on x86. For such cases the
1376 register move cost will be maximal. */
1377 mode = narrower_subreg_mode (mode, ALLOCNO_MODE (cp->second));
1379 cost = (cp->second == allocno
1380 ? ira_register_move_cost[mode][rclass][aclass]
1381 : ira_register_move_cost[mode][aclass][rclass]);
1382 if (decr_p)
1383 cost = -cost;
1385 update_conflict_cost = update_cost = cp->freq * cost / divisor;
1387 if (ALLOCNO_COLOR_DATA (another_allocno) != NULL
1388 && (ALLOCNO_COLOR_DATA (allocno)->first_thread_allocno
1389 != ALLOCNO_COLOR_DATA (another_allocno)->first_thread_allocno))
1390 /* Decrease conflict cost of ANOTHER_ALLOCNO if it is not
1391 in the same allocation thread. */
1392 update_conflict_cost /= COST_HOP_DIVISOR;
1394 if (update_cost == 0)
1395 continue;
1397 if (! update_allocno_cost (another_allocno, hard_regno,
1398 update_cost, update_conflict_cost))
1399 continue;
1400 queue_update_cost (another_allocno, allocno, divisor * COST_HOP_DIVISOR);
1401 if (record_p && ALLOCNO_COLOR_DATA (another_allocno) != NULL)
1402 ALLOCNO_COLOR_DATA (another_allocno)->update_cost_records
1403 = get_update_cost_record (hard_regno, divisor,
1404 ALLOCNO_COLOR_DATA (another_allocno)
1405 ->update_cost_records);
1408 while (get_next_update_cost (&allocno, &from, &divisor));
1411 /* Decrease preferred ALLOCNO hard register costs and costs of
1412 allocnos connected to ALLOCNO through copy. */
1413 static void
1414 update_costs_from_prefs (ira_allocno_t allocno)
1416 ira_pref_t pref;
1418 start_update_cost ();
1419 for (pref = ALLOCNO_PREFS (allocno); pref != NULL; pref = pref->next_pref)
1420 update_costs_from_allocno (allocno, pref->hard_regno,
1421 COST_HOP_DIVISOR, true, true);
1424 /* Update (decrease if DECR_P) the cost of allocnos connected to
1425 ALLOCNO through copies to increase chances to remove some copies as
1426 the result of subsequent assignment. ALLOCNO was just assigned to
1427 a hard register. Record cost updates if RECORD_P is true. */
1428 static void
1429 update_costs_from_copies (ira_allocno_t allocno, bool decr_p, bool record_p)
1431 int hard_regno;
1433 hard_regno = ALLOCNO_HARD_REGNO (allocno);
1434 ira_assert (hard_regno >= 0 && ALLOCNO_CLASS (allocno) != NO_REGS);
1435 start_update_cost ();
1436 update_costs_from_allocno (allocno, hard_regno, 1, decr_p, record_p);
1439 /* Restore costs of allocnos connected to ALLOCNO by copies as it was
1440 before updating costs of these allocnos from given allocno. This
1441 is a wise thing to do as if given allocno did not get an expected
1442 hard reg, using smaller cost of the hard reg for allocnos connected
1443 by copies to given allocno becomes actually misleading. Free all
1444 update cost records for ALLOCNO as we don't need them anymore. */
1445 static void
1446 restore_costs_from_copies (ira_allocno_t allocno)
1448 struct update_cost_record *records, *curr;
1450 if (ALLOCNO_COLOR_DATA (allocno) == NULL)
1451 return;
1452 records = ALLOCNO_COLOR_DATA (allocno)->update_cost_records;
1453 start_update_cost ();
1454 for (curr = records; curr != NULL; curr = curr->next)
1455 update_costs_from_allocno (allocno, curr->hard_regno,
1456 curr->divisor, true, false);
1457 free_update_cost_record_list (records);
1458 ALLOCNO_COLOR_DATA (allocno)->update_cost_records = NULL;
1461 /* This function updates COSTS (decrease if DECR_P) for hard_registers
1462 of ACLASS by conflict costs of the unassigned allocnos
1463 connected by copies with allocnos in update_cost_queue. This
1464 update increases chances to remove some copies. */
1465 static void
1466 update_conflict_hard_regno_costs (int *costs, enum reg_class aclass,
1467 bool decr_p)
1469 int i, cost, class_size, freq, mult, div, divisor;
1470 int index, hard_regno;
1471 int *conflict_costs;
1472 bool cont_p;
1473 enum reg_class another_aclass;
1474 ira_allocno_t allocno, another_allocno, from;
1475 ira_copy_t cp, next_cp;
1477 while (get_next_update_cost (&allocno, &from, &divisor))
1478 for (cp = ALLOCNO_COPIES (allocno); cp != NULL; cp = next_cp)
1480 if (cp->first == allocno)
1482 next_cp = cp->next_first_allocno_copy;
1483 another_allocno = cp->second;
1485 else if (cp->second == allocno)
1487 next_cp = cp->next_second_allocno_copy;
1488 another_allocno = cp->first;
1490 else
1491 gcc_unreachable ();
1493 if (another_allocno == from)
1494 continue;
1496 another_aclass = ALLOCNO_CLASS (another_allocno);
1497 if (! ira_reg_classes_intersect_p[aclass][another_aclass]
1498 || ALLOCNO_ASSIGNED_P (another_allocno)
1499 || ALLOCNO_COLOR_DATA (another_allocno)->may_be_spilled_p)
1500 continue;
1501 class_size = ira_class_hard_regs_num[another_aclass];
1502 ira_allocate_and_copy_costs
1503 (&ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (another_allocno),
1504 another_aclass, ALLOCNO_CONFLICT_HARD_REG_COSTS (another_allocno));
1505 conflict_costs
1506 = ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (another_allocno);
1507 if (conflict_costs == NULL)
1508 cont_p = true;
1509 else
1511 mult = cp->freq;
1512 freq = ALLOCNO_FREQ (another_allocno);
1513 if (freq == 0)
1514 freq = 1;
1515 div = freq * divisor;
1516 cont_p = false;
1517 for (i = class_size - 1; i >= 0; i--)
1519 hard_regno = ira_class_hard_regs[another_aclass][i];
1520 ira_assert (hard_regno >= 0);
1521 index = ira_class_hard_reg_index[aclass][hard_regno];
1522 if (index < 0)
1523 continue;
1524 cost = (int) (((int64_t) conflict_costs [i] * mult) / div);
1525 if (cost == 0)
1526 continue;
1527 cont_p = true;
1528 if (decr_p)
1529 cost = -cost;
1530 costs[index] += cost;
1533 /* Probably 5 hops will be enough. */
1534 if (cont_p
1535 && divisor <= (COST_HOP_DIVISOR
1536 * COST_HOP_DIVISOR
1537 * COST_HOP_DIVISOR
1538 * COST_HOP_DIVISOR))
1539 queue_update_cost (another_allocno, allocno, divisor * COST_HOP_DIVISOR);
1543 /* Set up conflicting (through CONFLICT_REGS) for each object of
1544 allocno A and the start allocno profitable regs (through
1545 START_PROFITABLE_REGS). Remember that the start profitable regs
1546 exclude hard regs which can not hold value of mode of allocno A.
1547 This covers mostly cases when multi-register value should be
1548 aligned. */
1549 static inline void
1550 get_conflict_and_start_profitable_regs (ira_allocno_t a, bool retry_p,
1551 HARD_REG_SET *conflict_regs,
1552 HARD_REG_SET *start_profitable_regs)
1554 int i, nwords;
1555 ira_object_t obj;
1557 nwords = ALLOCNO_NUM_OBJECTS (a);
1558 for (i = 0; i < nwords; i++)
1560 obj = ALLOCNO_OBJECT (a, i);
1561 COPY_HARD_REG_SET (conflict_regs[i],
1562 OBJECT_TOTAL_CONFLICT_HARD_REGS (obj));
1564 if (retry_p)
1566 COPY_HARD_REG_SET (*start_profitable_regs,
1567 reg_class_contents[ALLOCNO_CLASS (a)]);
1568 AND_COMPL_HARD_REG_SET (*start_profitable_regs,
1569 ira_prohibited_class_mode_regs
1570 [ALLOCNO_CLASS (a)][ALLOCNO_MODE (a)]);
1572 else
1573 COPY_HARD_REG_SET (*start_profitable_regs,
1574 ALLOCNO_COLOR_DATA (a)->profitable_hard_regs);
1577 /* Return true if HARD_REGNO is ok for assigning to allocno A with
1578 PROFITABLE_REGS and whose objects have CONFLICT_REGS. */
1579 static inline bool
1580 check_hard_reg_p (ira_allocno_t a, int hard_regno,
1581 HARD_REG_SET *conflict_regs, HARD_REG_SET profitable_regs)
1583 int j, nwords, nregs;
1584 enum reg_class aclass;
1585 machine_mode mode;
1587 aclass = ALLOCNO_CLASS (a);
1588 mode = ALLOCNO_MODE (a);
1589 if (TEST_HARD_REG_BIT (ira_prohibited_class_mode_regs[aclass][mode],
1590 hard_regno))
1591 return false;
1592 /* Checking only profitable hard regs. */
1593 if (! TEST_HARD_REG_BIT (profitable_regs, hard_regno))
1594 return false;
1595 nregs = hard_regno_nregs (hard_regno, mode);
1596 nwords = ALLOCNO_NUM_OBJECTS (a);
1597 for (j = 0; j < nregs; j++)
1599 int k;
1600 int set_to_test_start = 0, set_to_test_end = nwords;
1602 if (nregs == nwords)
1604 if (REG_WORDS_BIG_ENDIAN)
1605 set_to_test_start = nwords - j - 1;
1606 else
1607 set_to_test_start = j;
1608 set_to_test_end = set_to_test_start + 1;
1610 for (k = set_to_test_start; k < set_to_test_end; k++)
1611 if (TEST_HARD_REG_BIT (conflict_regs[k], hard_regno + j))
1612 break;
1613 if (k != set_to_test_end)
1614 break;
1616 return j == nregs;
1619 /* Return number of registers needed to be saved and restored at
1620 function prologue/epilogue if we allocate HARD_REGNO to hold value
1621 of MODE. */
1622 static int
1623 calculate_saved_nregs (int hard_regno, machine_mode mode)
1625 int i;
1626 int nregs = 0;
1628 ira_assert (hard_regno >= 0);
1629 for (i = hard_regno_nregs (hard_regno, mode) - 1; i >= 0; i--)
1630 if (!allocated_hardreg_p[hard_regno + i]
1631 && !TEST_HARD_REG_BIT (call_used_reg_set, hard_regno + i)
1632 && !LOCAL_REGNO (hard_regno + i))
1633 nregs++;
1634 return nregs;
1637 /* Choose a hard register for allocno A. If RETRY_P is TRUE, it means
1638 that the function called from function
1639 `ira_reassign_conflict_allocnos' and `allocno_reload_assign'. In
1640 this case some allocno data are not defined or updated and we
1641 should not touch these data. The function returns true if we
1642 managed to assign a hard register to the allocno.
1644 To assign a hard register, first of all we calculate all conflict
1645 hard registers which can come from conflicting allocnos with
1646 already assigned hard registers. After that we find first free
1647 hard register with the minimal cost. During hard register cost
1648 calculation we take conflict hard register costs into account to
1649 give a chance for conflicting allocnos to get a better hard
1650 register in the future.
1652 If the best hard register cost is bigger than cost of memory usage
1653 for the allocno, we don't assign a hard register to given allocno
1654 at all.
1656 If we assign a hard register to the allocno, we update costs of the
1657 hard register for allocnos connected by copies to improve a chance
1658 to coalesce insns represented by the copies when we assign hard
1659 registers to the allocnos connected by the copies. */
1660 static bool
1661 assign_hard_reg (ira_allocno_t a, bool retry_p)
1663 HARD_REG_SET conflicting_regs[2], profitable_hard_regs;
1664 int i, j, hard_regno, best_hard_regno, class_size;
1665 int cost, mem_cost, min_cost, full_cost, min_full_cost, nwords, word;
1666 int *a_costs;
1667 enum reg_class aclass;
1668 machine_mode mode;
1669 static int costs[FIRST_PSEUDO_REGISTER], full_costs[FIRST_PSEUDO_REGISTER];
1670 int saved_nregs;
1671 enum reg_class rclass;
1672 int add_cost;
1673 #ifdef STACK_REGS
1674 bool no_stack_reg_p;
1675 #endif
1677 ira_assert (! ALLOCNO_ASSIGNED_P (a));
1678 get_conflict_and_start_profitable_regs (a, retry_p,
1679 conflicting_regs,
1680 &profitable_hard_regs);
1681 aclass = ALLOCNO_CLASS (a);
1682 class_size = ira_class_hard_regs_num[aclass];
1683 best_hard_regno = -1;
1684 memset (full_costs, 0, sizeof (int) * class_size);
1685 mem_cost = 0;
1686 memset (costs, 0, sizeof (int) * class_size);
1687 memset (full_costs, 0, sizeof (int) * class_size);
1688 #ifdef STACK_REGS
1689 no_stack_reg_p = false;
1690 #endif
1691 if (! retry_p)
1692 start_update_cost ();
1693 mem_cost += ALLOCNO_UPDATED_MEMORY_COST (a);
1695 ira_allocate_and_copy_costs (&ALLOCNO_UPDATED_HARD_REG_COSTS (a),
1696 aclass, ALLOCNO_HARD_REG_COSTS (a));
1697 a_costs = ALLOCNO_UPDATED_HARD_REG_COSTS (a);
1698 #ifdef STACK_REGS
1699 no_stack_reg_p = no_stack_reg_p || ALLOCNO_TOTAL_NO_STACK_REG_P (a);
1700 #endif
1701 cost = ALLOCNO_UPDATED_CLASS_COST (a);
1702 for (i = 0; i < class_size; i++)
1703 if (a_costs != NULL)
1705 costs[i] += a_costs[i];
1706 full_costs[i] += a_costs[i];
1708 else
1710 costs[i] += cost;
1711 full_costs[i] += cost;
1713 nwords = ALLOCNO_NUM_OBJECTS (a);
1714 curr_allocno_process++;
1715 for (word = 0; word < nwords; word++)
1717 ira_object_t conflict_obj;
1718 ira_object_t obj = ALLOCNO_OBJECT (a, word);
1719 ira_object_conflict_iterator oci;
1721 /* Take preferences of conflicting allocnos into account. */
1722 FOR_EACH_OBJECT_CONFLICT (obj, conflict_obj, oci)
1724 ira_allocno_t conflict_a = OBJECT_ALLOCNO (conflict_obj);
1725 enum reg_class conflict_aclass;
1726 allocno_color_data_t data = ALLOCNO_COLOR_DATA (conflict_a);
1728 /* Reload can give another class so we need to check all
1729 allocnos. */
1730 if (!retry_p
1731 && ((!ALLOCNO_ASSIGNED_P (conflict_a)
1732 || ALLOCNO_HARD_REGNO (conflict_a) < 0)
1733 && !(hard_reg_set_intersect_p
1734 (profitable_hard_regs,
1735 ALLOCNO_COLOR_DATA
1736 (conflict_a)->profitable_hard_regs))))
1738 /* All conflict allocnos are in consideration bitmap
1739 when retry_p is false. It might change in future and
1740 if it happens the assert will be broken. It means
1741 the code should be modified for the new
1742 assumptions. */
1743 ira_assert (bitmap_bit_p (consideration_allocno_bitmap,
1744 ALLOCNO_NUM (conflict_a)));
1745 continue;
1747 conflict_aclass = ALLOCNO_CLASS (conflict_a);
1748 ira_assert (ira_reg_classes_intersect_p
1749 [aclass][conflict_aclass]);
1750 if (ALLOCNO_ASSIGNED_P (conflict_a))
1752 hard_regno = ALLOCNO_HARD_REGNO (conflict_a);
1753 if (hard_regno >= 0
1754 && (ira_hard_reg_set_intersection_p
1755 (hard_regno, ALLOCNO_MODE (conflict_a),
1756 reg_class_contents[aclass])))
1758 int n_objects = ALLOCNO_NUM_OBJECTS (conflict_a);
1759 int conflict_nregs;
1761 mode = ALLOCNO_MODE (conflict_a);
1762 conflict_nregs = hard_regno_nregs (hard_regno, mode);
1763 if (conflict_nregs == n_objects && conflict_nregs > 1)
1765 int num = OBJECT_SUBWORD (conflict_obj);
1767 if (REG_WORDS_BIG_ENDIAN)
1768 SET_HARD_REG_BIT (conflicting_regs[word],
1769 hard_regno + n_objects - num - 1);
1770 else
1771 SET_HARD_REG_BIT (conflicting_regs[word],
1772 hard_regno + num);
1774 else
1775 IOR_HARD_REG_SET
1776 (conflicting_regs[word],
1777 ira_reg_mode_hard_regset[hard_regno][mode]);
1778 if (hard_reg_set_subset_p (profitable_hard_regs,
1779 conflicting_regs[word]))
1780 goto fail;
1783 else if (! retry_p
1784 && ! ALLOCNO_COLOR_DATA (conflict_a)->may_be_spilled_p
1785 /* Don't process the conflict allocno twice. */
1786 && (ALLOCNO_COLOR_DATA (conflict_a)->last_process
1787 != curr_allocno_process))
1789 int k, *conflict_costs;
1791 ALLOCNO_COLOR_DATA (conflict_a)->last_process
1792 = curr_allocno_process;
1793 ira_allocate_and_copy_costs
1794 (&ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (conflict_a),
1795 conflict_aclass,
1796 ALLOCNO_CONFLICT_HARD_REG_COSTS (conflict_a));
1797 conflict_costs
1798 = ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (conflict_a);
1799 if (conflict_costs != NULL)
1800 for (j = class_size - 1; j >= 0; j--)
1802 hard_regno = ira_class_hard_regs[aclass][j];
1803 ira_assert (hard_regno >= 0);
1804 k = ira_class_hard_reg_index[conflict_aclass][hard_regno];
1805 if (k < 0
1806 /* If HARD_REGNO is not available for CONFLICT_A,
1807 the conflict would be ignored, since HARD_REGNO
1808 will never be assigned to CONFLICT_A. */
1809 || !TEST_HARD_REG_BIT (data->profitable_hard_regs,
1810 hard_regno))
1811 continue;
1812 full_costs[j] -= conflict_costs[k];
1814 queue_update_cost (conflict_a, NULL, COST_HOP_DIVISOR);
1819 if (! retry_p)
1820 /* Take into account preferences of allocnos connected by copies to
1821 the conflict allocnos. */
1822 update_conflict_hard_regno_costs (full_costs, aclass, true);
1824 /* Take preferences of allocnos connected by copies into
1825 account. */
1826 if (! retry_p)
1828 start_update_cost ();
1829 queue_update_cost (a, NULL, COST_HOP_DIVISOR);
1830 update_conflict_hard_regno_costs (full_costs, aclass, false);
1832 min_cost = min_full_cost = INT_MAX;
1833 /* We don't care about giving callee saved registers to allocnos no
1834 living through calls because call clobbered registers are
1835 allocated first (it is usual practice to put them first in
1836 REG_ALLOC_ORDER). */
1837 mode = ALLOCNO_MODE (a);
1838 for (i = 0; i < class_size; i++)
1840 hard_regno = ira_class_hard_regs[aclass][i];
1841 #ifdef STACK_REGS
1842 if (no_stack_reg_p
1843 && FIRST_STACK_REG <= hard_regno && hard_regno <= LAST_STACK_REG)
1844 continue;
1845 #endif
1846 if (! check_hard_reg_p (a, hard_regno,
1847 conflicting_regs, profitable_hard_regs))
1848 continue;
1849 cost = costs[i];
1850 full_cost = full_costs[i];
1851 if (!HONOR_REG_ALLOC_ORDER)
1853 if ((saved_nregs = calculate_saved_nregs (hard_regno, mode)) != 0)
1854 /* We need to save/restore the hard register in
1855 epilogue/prologue. Therefore we increase the cost. */
1857 rclass = REGNO_REG_CLASS (hard_regno);
1858 add_cost = ((ira_memory_move_cost[mode][rclass][0]
1859 + ira_memory_move_cost[mode][rclass][1])
1860 * saved_nregs / hard_regno_nregs (hard_regno,
1861 mode) - 1);
1862 cost += add_cost;
1863 full_cost += add_cost;
1866 if (min_cost > cost)
1867 min_cost = cost;
1868 if (min_full_cost > full_cost)
1870 min_full_cost = full_cost;
1871 best_hard_regno = hard_regno;
1872 ira_assert (hard_regno >= 0);
1875 if (min_full_cost > mem_cost
1876 /* Do not spill static chain pointer pseudo when non-local goto
1877 is used. */
1878 && ! non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a)))
1880 if (! retry_p && internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
1881 fprintf (ira_dump_file, "(memory is more profitable %d vs %d) ",
1882 mem_cost, min_full_cost);
1883 best_hard_regno = -1;
1885 fail:
1886 if (best_hard_regno >= 0)
1888 for (i = hard_regno_nregs (best_hard_regno, mode) - 1; i >= 0; i--)
1889 allocated_hardreg_p[best_hard_regno + i] = true;
1891 if (! retry_p)
1892 restore_costs_from_copies (a);
1893 ALLOCNO_HARD_REGNO (a) = best_hard_regno;
1894 ALLOCNO_ASSIGNED_P (a) = true;
1895 if (best_hard_regno >= 0)
1896 update_costs_from_copies (a, true, ! retry_p);
1897 ira_assert (ALLOCNO_CLASS (a) == aclass);
1898 /* We don't need updated costs anymore. */
1899 ira_free_allocno_updated_costs (a);
1900 return best_hard_regno >= 0;
1905 /* An array used to sort copies. */
1906 static ira_copy_t *sorted_copies;
1908 /* If allocno A is a cap, return non-cap allocno from which A is
1909 created. Otherwise, return A. */
1910 static ira_allocno_t
1911 get_cap_member (ira_allocno_t a)
1913 ira_allocno_t member;
1915 while ((member = ALLOCNO_CAP_MEMBER (a)) != NULL)
1916 a = member;
1917 return a;
1920 /* Return TRUE if live ranges of allocnos A1 and A2 intersect. It is
1921 used to find a conflict for new allocnos or allocnos with the
1922 different allocno classes. */
1923 static bool
1924 allocnos_conflict_by_live_ranges_p (ira_allocno_t a1, ira_allocno_t a2)
1926 rtx reg1, reg2;
1927 int i, j;
1928 int n1 = ALLOCNO_NUM_OBJECTS (a1);
1929 int n2 = ALLOCNO_NUM_OBJECTS (a2);
1931 if (a1 == a2)
1932 return false;
1933 reg1 = regno_reg_rtx[ALLOCNO_REGNO (a1)];
1934 reg2 = regno_reg_rtx[ALLOCNO_REGNO (a2)];
1935 if (reg1 != NULL && reg2 != NULL
1936 && ORIGINAL_REGNO (reg1) == ORIGINAL_REGNO (reg2))
1937 return false;
1939 /* We don't keep live ranges for caps because they can be quite big.
1940 Use ranges of non-cap allocno from which caps are created. */
1941 a1 = get_cap_member (a1);
1942 a2 = get_cap_member (a2);
1943 for (i = 0; i < n1; i++)
1945 ira_object_t c1 = ALLOCNO_OBJECT (a1, i);
1947 for (j = 0; j < n2; j++)
1949 ira_object_t c2 = ALLOCNO_OBJECT (a2, j);
1951 if (ira_live_ranges_intersect_p (OBJECT_LIVE_RANGES (c1),
1952 OBJECT_LIVE_RANGES (c2)))
1953 return true;
1956 return false;
1959 /* The function is used to sort copies according to their execution
1960 frequencies. */
1961 static int
1962 copy_freq_compare_func (const void *v1p, const void *v2p)
1964 ira_copy_t cp1 = *(const ira_copy_t *) v1p, cp2 = *(const ira_copy_t *) v2p;
1965 int pri1, pri2;
1967 pri1 = cp1->freq;
1968 pri2 = cp2->freq;
1969 if (pri2 - pri1)
1970 return pri2 - pri1;
1972 /* If frequencies are equal, sort by copies, so that the results of
1973 qsort leave nothing to chance. */
1974 return cp1->num - cp2->num;
1979 /* Return true if any allocno from thread of A1 conflicts with any
1980 allocno from thread A2. */
1981 static bool
1982 allocno_thread_conflict_p (ira_allocno_t a1, ira_allocno_t a2)
1984 ira_allocno_t a, conflict_a;
1986 for (a = ALLOCNO_COLOR_DATA (a2)->next_thread_allocno;;
1987 a = ALLOCNO_COLOR_DATA (a)->next_thread_allocno)
1989 for (conflict_a = ALLOCNO_COLOR_DATA (a1)->next_thread_allocno;;
1990 conflict_a = ALLOCNO_COLOR_DATA (conflict_a)->next_thread_allocno)
1992 if (allocnos_conflict_by_live_ranges_p (a, conflict_a))
1993 return true;
1994 if (conflict_a == a1)
1995 break;
1997 if (a == a2)
1998 break;
2000 return false;
2003 /* Merge two threads given correspondingly by their first allocnos T1
2004 and T2 (more accurately merging T2 into T1). */
2005 static void
2006 merge_threads (ira_allocno_t t1, ira_allocno_t t2)
2008 ira_allocno_t a, next, last;
2010 gcc_assert (t1 != t2
2011 && ALLOCNO_COLOR_DATA (t1)->first_thread_allocno == t1
2012 && ALLOCNO_COLOR_DATA (t2)->first_thread_allocno == t2);
2013 for (last = t2, a = ALLOCNO_COLOR_DATA (t2)->next_thread_allocno;;
2014 a = ALLOCNO_COLOR_DATA (a)->next_thread_allocno)
2016 ALLOCNO_COLOR_DATA (a)->first_thread_allocno = t1;
2017 if (a == t2)
2018 break;
2019 last = a;
2021 next = ALLOCNO_COLOR_DATA (t1)->next_thread_allocno;
2022 ALLOCNO_COLOR_DATA (t1)->next_thread_allocno = t2;
2023 ALLOCNO_COLOR_DATA (last)->next_thread_allocno = next;
2024 ALLOCNO_COLOR_DATA (t1)->thread_freq += ALLOCNO_COLOR_DATA (t2)->thread_freq;
2027 /* Create threads by processing CP_NUM copies from sorted copies. We
2028 process the most expensive copies first. */
2029 static void
2030 form_threads_from_copies (int cp_num)
2032 ira_allocno_t a, thread1, thread2;
2033 ira_copy_t cp;
2034 int i, n;
2036 qsort (sorted_copies, cp_num, sizeof (ira_copy_t), copy_freq_compare_func);
2037 /* Form threads processing copies, most frequently executed
2038 first. */
2039 for (; cp_num != 0;)
2041 for (i = 0; i < cp_num; i++)
2043 cp = sorted_copies[i];
2044 thread1 = ALLOCNO_COLOR_DATA (cp->first)->first_thread_allocno;
2045 thread2 = ALLOCNO_COLOR_DATA (cp->second)->first_thread_allocno;
2046 if (thread1 == thread2)
2047 continue;
2048 if (! allocno_thread_conflict_p (thread1, thread2))
2050 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
2051 fprintf
2052 (ira_dump_file,
2053 " Forming thread by copy %d:a%dr%d-a%dr%d (freq=%d):\n",
2054 cp->num, ALLOCNO_NUM (cp->first), ALLOCNO_REGNO (cp->first),
2055 ALLOCNO_NUM (cp->second), ALLOCNO_REGNO (cp->second),
2056 cp->freq);
2057 merge_threads (thread1, thread2);
2058 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
2060 thread1 = ALLOCNO_COLOR_DATA (thread1)->first_thread_allocno;
2061 fprintf (ira_dump_file, " Result (freq=%d): a%dr%d(%d)",
2062 ALLOCNO_COLOR_DATA (thread1)->thread_freq,
2063 ALLOCNO_NUM (thread1), ALLOCNO_REGNO (thread1),
2064 ALLOCNO_FREQ (thread1));
2065 for (a = ALLOCNO_COLOR_DATA (thread1)->next_thread_allocno;
2066 a != thread1;
2067 a = ALLOCNO_COLOR_DATA (a)->next_thread_allocno)
2068 fprintf (ira_dump_file, " a%dr%d(%d)",
2069 ALLOCNO_NUM (a), ALLOCNO_REGNO (a),
2070 ALLOCNO_FREQ (a));
2071 fprintf (ira_dump_file, "\n");
2073 i++;
2074 break;
2077 /* Collect the rest of copies. */
2078 for (n = 0; i < cp_num; i++)
2080 cp = sorted_copies[i];
2081 if (ALLOCNO_COLOR_DATA (cp->first)->first_thread_allocno
2082 != ALLOCNO_COLOR_DATA (cp->second)->first_thread_allocno)
2083 sorted_copies[n++] = cp;
2085 cp_num = n;
2089 /* Create threads by processing copies of all alocnos from BUCKET. We
2090 process the most expensive copies first. */
2091 static void
2092 form_threads_from_bucket (ira_allocno_t bucket)
2094 ira_allocno_t a;
2095 ira_copy_t cp, next_cp;
2096 int cp_num = 0;
2098 for (a = bucket; a != NULL; a = ALLOCNO_COLOR_DATA (a)->next_bucket_allocno)
2100 for (cp = ALLOCNO_COPIES (a); cp != NULL; cp = next_cp)
2102 if (cp->first == a)
2104 next_cp = cp->next_first_allocno_copy;
2105 sorted_copies[cp_num++] = cp;
2107 else if (cp->second == a)
2108 next_cp = cp->next_second_allocno_copy;
2109 else
2110 gcc_unreachable ();
2113 form_threads_from_copies (cp_num);
2116 /* Create threads by processing copies of colorable allocno A. We
2117 process most expensive copies first. */
2118 static void
2119 form_threads_from_colorable_allocno (ira_allocno_t a)
2121 ira_allocno_t another_a;
2122 ira_copy_t cp, next_cp;
2123 int cp_num = 0;
2125 for (cp = ALLOCNO_COPIES (a); cp != NULL; cp = next_cp)
2127 if (cp->first == a)
2129 next_cp = cp->next_first_allocno_copy;
2130 another_a = cp->second;
2132 else if (cp->second == a)
2134 next_cp = cp->next_second_allocno_copy;
2135 another_a = cp->first;
2137 else
2138 gcc_unreachable ();
2139 if ((! ALLOCNO_COLOR_DATA (another_a)->in_graph_p
2140 && !ALLOCNO_COLOR_DATA (another_a)->may_be_spilled_p)
2141 || ALLOCNO_COLOR_DATA (another_a)->colorable_p)
2142 sorted_copies[cp_num++] = cp;
2144 form_threads_from_copies (cp_num);
2147 /* Form initial threads which contain only one allocno. */
2148 static void
2149 init_allocno_threads (void)
2151 ira_allocno_t a;
2152 unsigned int j;
2153 bitmap_iterator bi;
2155 EXECUTE_IF_SET_IN_BITMAP (consideration_allocno_bitmap, 0, j, bi)
2157 a = ira_allocnos[j];
2158 /* Set up initial thread data: */
2159 ALLOCNO_COLOR_DATA (a)->first_thread_allocno
2160 = ALLOCNO_COLOR_DATA (a)->next_thread_allocno = a;
2161 ALLOCNO_COLOR_DATA (a)->thread_freq = ALLOCNO_FREQ (a);
2167 /* This page contains the allocator based on the Chaitin-Briggs algorithm. */
2169 /* Bucket of allocnos that can colored currently without spilling. */
2170 static ira_allocno_t colorable_allocno_bucket;
2172 /* Bucket of allocnos that might be not colored currently without
2173 spilling. */
2174 static ira_allocno_t uncolorable_allocno_bucket;
2176 /* The current number of allocnos in the uncolorable_bucket. */
2177 static int uncolorable_allocnos_num;
2179 /* Return the current spill priority of allocno A. The less the
2180 number, the more preferable the allocno for spilling. */
2181 static inline int
2182 allocno_spill_priority (ira_allocno_t a)
2184 allocno_color_data_t data = ALLOCNO_COLOR_DATA (a);
2186 return (data->temp
2187 / (ALLOCNO_EXCESS_PRESSURE_POINTS_NUM (a)
2188 * ira_reg_class_max_nregs[ALLOCNO_CLASS (a)][ALLOCNO_MODE (a)]
2189 + 1));
2192 /* Add allocno A to bucket *BUCKET_PTR. A should be not in a bucket
2193 before the call. */
2194 static void
2195 add_allocno_to_bucket (ira_allocno_t a, ira_allocno_t *bucket_ptr)
2197 ira_allocno_t first_a;
2198 allocno_color_data_t data;
2200 if (bucket_ptr == &uncolorable_allocno_bucket
2201 && ALLOCNO_CLASS (a) != NO_REGS)
2203 uncolorable_allocnos_num++;
2204 ira_assert (uncolorable_allocnos_num > 0);
2206 first_a = *bucket_ptr;
2207 data = ALLOCNO_COLOR_DATA (a);
2208 data->next_bucket_allocno = first_a;
2209 data->prev_bucket_allocno = NULL;
2210 if (first_a != NULL)
2211 ALLOCNO_COLOR_DATA (first_a)->prev_bucket_allocno = a;
2212 *bucket_ptr = a;
2215 /* Compare two allocnos to define which allocno should be pushed first
2216 into the coloring stack. If the return is a negative number, the
2217 allocno given by the first parameter will be pushed first. In this
2218 case such allocno has less priority than the second one and the
2219 hard register will be assigned to it after assignment to the second
2220 one. As the result of such assignment order, the second allocno
2221 has a better chance to get the best hard register. */
2222 static int
2223 bucket_allocno_compare_func (const void *v1p, const void *v2p)
2225 ira_allocno_t a1 = *(const ira_allocno_t *) v1p;
2226 ira_allocno_t a2 = *(const ira_allocno_t *) v2p;
2227 int diff, freq1, freq2, a1_num, a2_num;
2228 ira_allocno_t t1 = ALLOCNO_COLOR_DATA (a1)->first_thread_allocno;
2229 ira_allocno_t t2 = ALLOCNO_COLOR_DATA (a2)->first_thread_allocno;
2230 int cl1 = ALLOCNO_CLASS (a1), cl2 = ALLOCNO_CLASS (a2);
2232 freq1 = ALLOCNO_COLOR_DATA (t1)->thread_freq;
2233 freq2 = ALLOCNO_COLOR_DATA (t2)->thread_freq;
2234 if ((diff = freq1 - freq2) != 0)
2235 return diff;
2237 if ((diff = ALLOCNO_NUM (t2) - ALLOCNO_NUM (t1)) != 0)
2238 return diff;
2240 /* Push pseudos requiring less hard registers first. It means that
2241 we will assign pseudos requiring more hard registers first
2242 avoiding creation small holes in free hard register file into
2243 which the pseudos requiring more hard registers can not fit. */
2244 if ((diff = (ira_reg_class_max_nregs[cl1][ALLOCNO_MODE (a1)]
2245 - ira_reg_class_max_nregs[cl2][ALLOCNO_MODE (a2)])) != 0)
2246 return diff;
2248 freq1 = ALLOCNO_FREQ (a1);
2249 freq2 = ALLOCNO_FREQ (a2);
2250 if ((diff = freq1 - freq2) != 0)
2251 return diff;
2253 a1_num = ALLOCNO_COLOR_DATA (a1)->available_regs_num;
2254 a2_num = ALLOCNO_COLOR_DATA (a2)->available_regs_num;
2255 if ((diff = a2_num - a1_num) != 0)
2256 return diff;
2257 return ALLOCNO_NUM (a2) - ALLOCNO_NUM (a1);
2260 /* Sort bucket *BUCKET_PTR and return the result through
2261 BUCKET_PTR. */
2262 static void
2263 sort_bucket (ira_allocno_t *bucket_ptr,
2264 int (*compare_func) (const void *, const void *))
2266 ira_allocno_t a, head;
2267 int n;
2269 for (n = 0, a = *bucket_ptr;
2270 a != NULL;
2271 a = ALLOCNO_COLOR_DATA (a)->next_bucket_allocno)
2272 sorted_allocnos[n++] = a;
2273 if (n <= 1)
2274 return;
2275 qsort (sorted_allocnos, n, sizeof (ira_allocno_t), compare_func);
2276 head = NULL;
2277 for (n--; n >= 0; n--)
2279 a = sorted_allocnos[n];
2280 ALLOCNO_COLOR_DATA (a)->next_bucket_allocno = head;
2281 ALLOCNO_COLOR_DATA (a)->prev_bucket_allocno = NULL;
2282 if (head != NULL)
2283 ALLOCNO_COLOR_DATA (head)->prev_bucket_allocno = a;
2284 head = a;
2286 *bucket_ptr = head;
2289 /* Add ALLOCNO to colorable bucket maintaining the order according
2290 their priority. ALLOCNO should be not in a bucket before the
2291 call. */
2292 static void
2293 add_allocno_to_ordered_colorable_bucket (ira_allocno_t allocno)
2295 ira_allocno_t before, after;
2297 form_threads_from_colorable_allocno (allocno);
2298 for (before = colorable_allocno_bucket, after = NULL;
2299 before != NULL;
2300 after = before,
2301 before = ALLOCNO_COLOR_DATA (before)->next_bucket_allocno)
2302 if (bucket_allocno_compare_func (&allocno, &before) < 0)
2303 break;
2304 ALLOCNO_COLOR_DATA (allocno)->next_bucket_allocno = before;
2305 ALLOCNO_COLOR_DATA (allocno)->prev_bucket_allocno = after;
2306 if (after == NULL)
2307 colorable_allocno_bucket = allocno;
2308 else
2309 ALLOCNO_COLOR_DATA (after)->next_bucket_allocno = allocno;
2310 if (before != NULL)
2311 ALLOCNO_COLOR_DATA (before)->prev_bucket_allocno = allocno;
2314 /* Delete ALLOCNO from bucket *BUCKET_PTR. It should be there before
2315 the call. */
2316 static void
2317 delete_allocno_from_bucket (ira_allocno_t allocno, ira_allocno_t *bucket_ptr)
2319 ira_allocno_t prev_allocno, next_allocno;
2321 if (bucket_ptr == &uncolorable_allocno_bucket
2322 && ALLOCNO_CLASS (allocno) != NO_REGS)
2324 uncolorable_allocnos_num--;
2325 ira_assert (uncolorable_allocnos_num >= 0);
2327 prev_allocno = ALLOCNO_COLOR_DATA (allocno)->prev_bucket_allocno;
2328 next_allocno = ALLOCNO_COLOR_DATA (allocno)->next_bucket_allocno;
2329 if (prev_allocno != NULL)
2330 ALLOCNO_COLOR_DATA (prev_allocno)->next_bucket_allocno = next_allocno;
2331 else
2333 ira_assert (*bucket_ptr == allocno);
2334 *bucket_ptr = next_allocno;
2336 if (next_allocno != NULL)
2337 ALLOCNO_COLOR_DATA (next_allocno)->prev_bucket_allocno = prev_allocno;
2340 /* Put allocno A onto the coloring stack without removing it from its
2341 bucket. Pushing allocno to the coloring stack can result in moving
2342 conflicting allocnos from the uncolorable bucket to the colorable
2343 one. */
2344 static void
2345 push_allocno_to_stack (ira_allocno_t a)
2347 enum reg_class aclass;
2348 allocno_color_data_t data, conflict_data;
2349 int size, i, n = ALLOCNO_NUM_OBJECTS (a);
2351 data = ALLOCNO_COLOR_DATA (a);
2352 data->in_graph_p = false;
2353 allocno_stack_vec.safe_push (a);
2354 aclass = ALLOCNO_CLASS (a);
2355 if (aclass == NO_REGS)
2356 return;
2357 size = ira_reg_class_max_nregs[aclass][ALLOCNO_MODE (a)];
2358 if (n > 1)
2360 /* We will deal with the subwords individually. */
2361 gcc_assert (size == ALLOCNO_NUM_OBJECTS (a));
2362 size = 1;
2364 for (i = 0; i < n; i++)
2366 ira_object_t obj = ALLOCNO_OBJECT (a, i);
2367 ira_object_t conflict_obj;
2368 ira_object_conflict_iterator oci;
2370 FOR_EACH_OBJECT_CONFLICT (obj, conflict_obj, oci)
2372 ira_allocno_t conflict_a = OBJECT_ALLOCNO (conflict_obj);
2374 conflict_data = ALLOCNO_COLOR_DATA (conflict_a);
2375 if (conflict_data->colorable_p
2376 || ! conflict_data->in_graph_p
2377 || ALLOCNO_ASSIGNED_P (conflict_a)
2378 || !(hard_reg_set_intersect_p
2379 (ALLOCNO_COLOR_DATA (a)->profitable_hard_regs,
2380 conflict_data->profitable_hard_regs)))
2381 continue;
2382 ira_assert (bitmap_bit_p (coloring_allocno_bitmap,
2383 ALLOCNO_NUM (conflict_a)));
2384 if (update_left_conflict_sizes_p (conflict_a, a, size))
2386 delete_allocno_from_bucket
2387 (conflict_a, &uncolorable_allocno_bucket);
2388 add_allocno_to_ordered_colorable_bucket (conflict_a);
2389 if (internal_flag_ira_verbose > 4 && ira_dump_file != NULL)
2391 fprintf (ira_dump_file, " Making");
2392 ira_print_expanded_allocno (conflict_a);
2393 fprintf (ira_dump_file, " colorable\n");
2401 /* Put ALLOCNO onto the coloring stack and remove it from its bucket.
2402 The allocno is in the colorable bucket if COLORABLE_P is TRUE. */
2403 static void
2404 remove_allocno_from_bucket_and_push (ira_allocno_t allocno, bool colorable_p)
2406 if (colorable_p)
2407 delete_allocno_from_bucket (allocno, &colorable_allocno_bucket);
2408 else
2409 delete_allocno_from_bucket (allocno, &uncolorable_allocno_bucket);
2410 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
2412 fprintf (ira_dump_file, " Pushing");
2413 ira_print_expanded_allocno (allocno);
2414 if (colorable_p)
2415 fprintf (ira_dump_file, "(cost %d)\n",
2416 ALLOCNO_COLOR_DATA (allocno)->temp);
2417 else
2418 fprintf (ira_dump_file, "(potential spill: %spri=%d, cost=%d)\n",
2419 ALLOCNO_BAD_SPILL_P (allocno) ? "bad spill, " : "",
2420 allocno_spill_priority (allocno),
2421 ALLOCNO_COLOR_DATA (allocno)->temp);
2423 if (! colorable_p)
2424 ALLOCNO_COLOR_DATA (allocno)->may_be_spilled_p = true;
2425 push_allocno_to_stack (allocno);
2428 /* Put all allocnos from colorable bucket onto the coloring stack. */
2429 static void
2430 push_only_colorable (void)
2432 form_threads_from_bucket (colorable_allocno_bucket);
2433 sort_bucket (&colorable_allocno_bucket, bucket_allocno_compare_func);
2434 for (;colorable_allocno_bucket != NULL;)
2435 remove_allocno_from_bucket_and_push (colorable_allocno_bucket, true);
2438 /* Return the frequency of exit edges (if EXIT_P) or entry from/to the
2439 loop given by its LOOP_NODE. */
2441 ira_loop_edge_freq (ira_loop_tree_node_t loop_node, int regno, bool exit_p)
2443 int freq, i;
2444 edge_iterator ei;
2445 edge e;
2446 vec<edge> edges;
2448 ira_assert (current_loops != NULL && loop_node->loop != NULL
2449 && (regno < 0 || regno >= FIRST_PSEUDO_REGISTER));
2450 freq = 0;
2451 if (! exit_p)
2453 FOR_EACH_EDGE (e, ei, loop_node->loop->header->preds)
2454 if (e->src != loop_node->loop->latch
2455 && (regno < 0
2456 || (bitmap_bit_p (df_get_live_out (e->src), regno)
2457 && bitmap_bit_p (df_get_live_in (e->dest), regno))))
2458 freq += EDGE_FREQUENCY (e);
2460 else
2462 edges = get_loop_exit_edges (loop_node->loop);
2463 FOR_EACH_VEC_ELT (edges, i, e)
2464 if (regno < 0
2465 || (bitmap_bit_p (df_get_live_out (e->src), regno)
2466 && bitmap_bit_p (df_get_live_in (e->dest), regno)))
2467 freq += EDGE_FREQUENCY (e);
2468 edges.release ();
2471 return REG_FREQ_FROM_EDGE_FREQ (freq);
2474 /* Calculate and return the cost of putting allocno A into memory. */
2475 static int
2476 calculate_allocno_spill_cost (ira_allocno_t a)
2478 int regno, cost;
2479 machine_mode mode;
2480 enum reg_class rclass;
2481 ira_allocno_t parent_allocno;
2482 ira_loop_tree_node_t parent_node, loop_node;
2484 regno = ALLOCNO_REGNO (a);
2485 cost = ALLOCNO_UPDATED_MEMORY_COST (a) - ALLOCNO_UPDATED_CLASS_COST (a);
2486 if (ALLOCNO_CAP (a) != NULL)
2487 return cost;
2488 loop_node = ALLOCNO_LOOP_TREE_NODE (a);
2489 if ((parent_node = loop_node->parent) == NULL)
2490 return cost;
2491 if ((parent_allocno = parent_node->regno_allocno_map[regno]) == NULL)
2492 return cost;
2493 mode = ALLOCNO_MODE (a);
2494 rclass = ALLOCNO_CLASS (a);
2495 if (ALLOCNO_HARD_REGNO (parent_allocno) < 0)
2496 cost -= (ira_memory_move_cost[mode][rclass][0]
2497 * ira_loop_edge_freq (loop_node, regno, true)
2498 + ira_memory_move_cost[mode][rclass][1]
2499 * ira_loop_edge_freq (loop_node, regno, false));
2500 else
2502 ira_init_register_move_cost_if_necessary (mode);
2503 cost += ((ira_memory_move_cost[mode][rclass][1]
2504 * ira_loop_edge_freq (loop_node, regno, true)
2505 + ira_memory_move_cost[mode][rclass][0]
2506 * ira_loop_edge_freq (loop_node, regno, false))
2507 - (ira_register_move_cost[mode][rclass][rclass]
2508 * (ira_loop_edge_freq (loop_node, regno, false)
2509 + ira_loop_edge_freq (loop_node, regno, true))));
2511 return cost;
2514 /* Used for sorting allocnos for spilling. */
2515 static inline int
2516 allocno_spill_priority_compare (ira_allocno_t a1, ira_allocno_t a2)
2518 int pri1, pri2, diff;
2520 /* Avoid spilling static chain pointer pseudo when non-local goto is
2521 used. */
2522 if (non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a1)))
2523 return 1;
2524 else if (non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a2)))
2525 return -1;
2526 if (ALLOCNO_BAD_SPILL_P (a1) && ! ALLOCNO_BAD_SPILL_P (a2))
2527 return 1;
2528 if (ALLOCNO_BAD_SPILL_P (a2) && ! ALLOCNO_BAD_SPILL_P (a1))
2529 return -1;
2530 pri1 = allocno_spill_priority (a1);
2531 pri2 = allocno_spill_priority (a2);
2532 if ((diff = pri1 - pri2) != 0)
2533 return diff;
2534 if ((diff
2535 = ALLOCNO_COLOR_DATA (a1)->temp - ALLOCNO_COLOR_DATA (a2)->temp) != 0)
2536 return diff;
2537 return ALLOCNO_NUM (a1) - ALLOCNO_NUM (a2);
2540 /* Used for sorting allocnos for spilling. */
2541 static int
2542 allocno_spill_sort_compare (const void *v1p, const void *v2p)
2544 ira_allocno_t p1 = *(const ira_allocno_t *) v1p;
2545 ira_allocno_t p2 = *(const ira_allocno_t *) v2p;
2547 return allocno_spill_priority_compare (p1, p2);
2550 /* Push allocnos to the coloring stack. The order of allocnos in the
2551 stack defines the order for the subsequent coloring. */
2552 static void
2553 push_allocnos_to_stack (void)
2555 ira_allocno_t a;
2556 int cost;
2558 /* Calculate uncolorable allocno spill costs. */
2559 for (a = uncolorable_allocno_bucket;
2560 a != NULL;
2561 a = ALLOCNO_COLOR_DATA (a)->next_bucket_allocno)
2562 if (ALLOCNO_CLASS (a) != NO_REGS)
2564 cost = calculate_allocno_spill_cost (a);
2565 /* ??? Remove cost of copies between the coalesced
2566 allocnos. */
2567 ALLOCNO_COLOR_DATA (a)->temp = cost;
2569 sort_bucket (&uncolorable_allocno_bucket, allocno_spill_sort_compare);
2570 for (;;)
2572 push_only_colorable ();
2573 a = uncolorable_allocno_bucket;
2574 if (a == NULL)
2575 break;
2576 remove_allocno_from_bucket_and_push (a, false);
2578 ira_assert (colorable_allocno_bucket == NULL
2579 && uncolorable_allocno_bucket == NULL);
2580 ira_assert (uncolorable_allocnos_num == 0);
2583 /* Pop the coloring stack and assign hard registers to the popped
2584 allocnos. */
2585 static void
2586 pop_allocnos_from_stack (void)
2588 ira_allocno_t allocno;
2589 enum reg_class aclass;
2591 for (;allocno_stack_vec.length () != 0;)
2593 allocno = allocno_stack_vec.pop ();
2594 aclass = ALLOCNO_CLASS (allocno);
2595 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
2597 fprintf (ira_dump_file, " Popping");
2598 ira_print_expanded_allocno (allocno);
2599 fprintf (ira_dump_file, " -- ");
2601 if (aclass == NO_REGS)
2603 ALLOCNO_HARD_REGNO (allocno) = -1;
2604 ALLOCNO_ASSIGNED_P (allocno) = true;
2605 ira_assert (ALLOCNO_UPDATED_HARD_REG_COSTS (allocno) == NULL);
2606 ira_assert
2607 (ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (allocno) == NULL);
2608 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
2609 fprintf (ira_dump_file, "assign memory\n");
2611 else if (assign_hard_reg (allocno, false))
2613 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
2614 fprintf (ira_dump_file, "assign reg %d\n",
2615 ALLOCNO_HARD_REGNO (allocno));
2617 else if (ALLOCNO_ASSIGNED_P (allocno))
2619 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
2620 fprintf (ira_dump_file, "spill%s\n",
2621 ALLOCNO_COLOR_DATA (allocno)->may_be_spilled_p
2622 ? "" : "!");
2624 ALLOCNO_COLOR_DATA (allocno)->in_graph_p = true;
2628 /* Set up number of available hard registers for allocno A. */
2629 static void
2630 setup_allocno_available_regs_num (ira_allocno_t a)
2632 int i, n, hard_regno, hard_regs_num, nwords;
2633 enum reg_class aclass;
2634 allocno_color_data_t data;
2636 aclass = ALLOCNO_CLASS (a);
2637 data = ALLOCNO_COLOR_DATA (a);
2638 data->available_regs_num = 0;
2639 if (aclass == NO_REGS)
2640 return;
2641 hard_regs_num = ira_class_hard_regs_num[aclass];
2642 nwords = ALLOCNO_NUM_OBJECTS (a);
2643 for (n = 0, i = hard_regs_num - 1; i >= 0; i--)
2645 hard_regno = ira_class_hard_regs[aclass][i];
2646 /* Checking only profitable hard regs. */
2647 if (TEST_HARD_REG_BIT (data->profitable_hard_regs, hard_regno))
2648 n++;
2650 data->available_regs_num = n;
2651 if (internal_flag_ira_verbose <= 2 || ira_dump_file == NULL)
2652 return;
2653 fprintf
2654 (ira_dump_file,
2655 " Allocno a%dr%d of %s(%d) has %d avail. regs ",
2656 ALLOCNO_NUM (a), ALLOCNO_REGNO (a),
2657 reg_class_names[aclass], ira_class_hard_regs_num[aclass], n);
2658 print_hard_reg_set (ira_dump_file, data->profitable_hard_regs, false);
2659 fprintf (ira_dump_file, ", %snode: ",
2660 hard_reg_set_equal_p (data->profitable_hard_regs,
2661 data->hard_regs_node->hard_regs->set)
2662 ? "" : "^");
2663 print_hard_reg_set (ira_dump_file,
2664 data->hard_regs_node->hard_regs->set, false);
2665 for (i = 0; i < nwords; i++)
2667 ira_object_t obj = ALLOCNO_OBJECT (a, i);
2669 if (nwords != 1)
2671 if (i != 0)
2672 fprintf (ira_dump_file, ", ");
2673 fprintf (ira_dump_file, " obj %d", i);
2675 fprintf (ira_dump_file, " (confl regs = ");
2676 print_hard_reg_set (ira_dump_file, OBJECT_TOTAL_CONFLICT_HARD_REGS (obj),
2677 false);
2678 fprintf (ira_dump_file, ")");
2680 fprintf (ira_dump_file, "\n");
2683 /* Put ALLOCNO in a bucket corresponding to its number and size of its
2684 conflicting allocnos and hard registers. */
2685 static void
2686 put_allocno_into_bucket (ira_allocno_t allocno)
2688 ALLOCNO_COLOR_DATA (allocno)->in_graph_p = true;
2689 setup_allocno_available_regs_num (allocno);
2690 if (setup_left_conflict_sizes_p (allocno))
2691 add_allocno_to_bucket (allocno, &colorable_allocno_bucket);
2692 else
2693 add_allocno_to_bucket (allocno, &uncolorable_allocno_bucket);
2696 /* Map: allocno number -> allocno priority. */
2697 static int *allocno_priorities;
2699 /* Set up priorities for N allocnos in array
2700 CONSIDERATION_ALLOCNOS. */
2701 static void
2702 setup_allocno_priorities (ira_allocno_t *consideration_allocnos, int n)
2704 int i, length, nrefs, priority, max_priority, mult;
2705 ira_allocno_t a;
2707 max_priority = 0;
2708 for (i = 0; i < n; i++)
2710 a = consideration_allocnos[i];
2711 nrefs = ALLOCNO_NREFS (a);
2712 ira_assert (nrefs >= 0);
2713 mult = floor_log2 (ALLOCNO_NREFS (a)) + 1;
2714 ira_assert (mult >= 0);
2715 allocno_priorities[ALLOCNO_NUM (a)]
2716 = priority
2717 = (mult
2718 * (ALLOCNO_MEMORY_COST (a) - ALLOCNO_CLASS_COST (a))
2719 * ira_reg_class_max_nregs[ALLOCNO_CLASS (a)][ALLOCNO_MODE (a)]);
2720 if (priority < 0)
2721 priority = -priority;
2722 if (max_priority < priority)
2723 max_priority = priority;
2725 mult = max_priority == 0 ? 1 : INT_MAX / max_priority;
2726 for (i = 0; i < n; i++)
2728 a = consideration_allocnos[i];
2729 length = ALLOCNO_EXCESS_PRESSURE_POINTS_NUM (a);
2730 if (ALLOCNO_NUM_OBJECTS (a) > 1)
2731 length /= ALLOCNO_NUM_OBJECTS (a);
2732 if (length <= 0)
2733 length = 1;
2734 allocno_priorities[ALLOCNO_NUM (a)]
2735 = allocno_priorities[ALLOCNO_NUM (a)] * mult / length;
2739 /* Sort allocnos according to the profit of usage of a hard register
2740 instead of memory for them. */
2741 static int
2742 allocno_cost_compare_func (const void *v1p, const void *v2p)
2744 ira_allocno_t p1 = *(const ira_allocno_t *) v1p;
2745 ira_allocno_t p2 = *(const ira_allocno_t *) v2p;
2746 int c1, c2;
2748 c1 = ALLOCNO_UPDATED_MEMORY_COST (p1) - ALLOCNO_UPDATED_CLASS_COST (p1);
2749 c2 = ALLOCNO_UPDATED_MEMORY_COST (p2) - ALLOCNO_UPDATED_CLASS_COST (p2);
2750 if (c1 - c2)
2751 return c1 - c2;
2753 /* If regs are equally good, sort by allocno numbers, so that the
2754 results of qsort leave nothing to chance. */
2755 return ALLOCNO_NUM (p1) - ALLOCNO_NUM (p2);
2758 /* Return savings on removed copies when ALLOCNO is assigned to
2759 HARD_REGNO. */
2760 static int
2761 allocno_copy_cost_saving (ira_allocno_t allocno, int hard_regno)
2763 int cost = 0;
2764 machine_mode allocno_mode = ALLOCNO_MODE (allocno);
2765 enum reg_class rclass;
2766 ira_copy_t cp, next_cp;
2768 rclass = REGNO_REG_CLASS (hard_regno);
2769 if (ira_reg_class_max_nregs[rclass][allocno_mode]
2770 > ira_class_hard_regs_num[rclass])
2771 /* For the above condition the cost can be wrong. Use the allocno
2772 class in this case. */
2773 rclass = ALLOCNO_CLASS (allocno);
2774 for (cp = ALLOCNO_COPIES (allocno); cp != NULL; cp = next_cp)
2776 if (cp->first == allocno)
2778 next_cp = cp->next_first_allocno_copy;
2779 if (ALLOCNO_HARD_REGNO (cp->second) != hard_regno)
2780 continue;
2782 else if (cp->second == allocno)
2784 next_cp = cp->next_second_allocno_copy;
2785 if (ALLOCNO_HARD_REGNO (cp->first) != hard_regno)
2786 continue;
2788 else
2789 gcc_unreachable ();
2790 cost += cp->freq * ira_register_move_cost[allocno_mode][rclass][rclass];
2792 return cost;
2795 /* We used Chaitin-Briggs coloring to assign as many pseudos as
2796 possible to hard registers. Let us try to improve allocation with
2797 cost point of view. This function improves the allocation by
2798 spilling some allocnos and assigning the freed hard registers to
2799 other allocnos if it decreases the overall allocation cost. */
2800 static void
2801 improve_allocation (void)
2803 unsigned int i;
2804 int j, k, n, hregno, conflict_hregno, base_cost, class_size, word, nwords;
2805 int check, spill_cost, min_cost, nregs, conflict_nregs, r, best;
2806 bool try_p;
2807 enum reg_class aclass;
2808 machine_mode mode;
2809 int *allocno_costs;
2810 int costs[FIRST_PSEUDO_REGISTER];
2811 HARD_REG_SET conflicting_regs[2], profitable_hard_regs;
2812 ira_allocno_t a;
2813 bitmap_iterator bi;
2815 /* Don't bother to optimize the code with static chain pointer and
2816 non-local goto in order not to spill the chain pointer
2817 pseudo. */
2818 if (cfun->static_chain_decl && crtl->has_nonlocal_goto)
2819 return;
2820 /* Clear counts used to process conflicting allocnos only once for
2821 each allocno. */
2822 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap, 0, i, bi)
2823 ALLOCNO_COLOR_DATA (ira_allocnos[i])->temp = 0;
2824 check = n = 0;
2825 /* Process each allocno and try to assign a hard register to it by
2826 spilling some its conflicting allocnos. */
2827 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap, 0, i, bi)
2829 a = ira_allocnos[i];
2830 ALLOCNO_COLOR_DATA (a)->temp = 0;
2831 if (empty_profitable_hard_regs (a))
2832 continue;
2833 check++;
2834 aclass = ALLOCNO_CLASS (a);
2835 allocno_costs = ALLOCNO_HARD_REG_COSTS (a);
2836 if ((hregno = ALLOCNO_HARD_REGNO (a)) < 0)
2837 base_cost = ALLOCNO_UPDATED_MEMORY_COST (a);
2838 else if (allocno_costs == NULL)
2839 /* It means that assigning a hard register is not profitable
2840 (we don't waste memory for hard register costs in this
2841 case). */
2842 continue;
2843 else
2844 base_cost = (allocno_costs[ira_class_hard_reg_index[aclass][hregno]]
2845 - allocno_copy_cost_saving (a, hregno));
2846 try_p = false;
2847 get_conflict_and_start_profitable_regs (a, false,
2848 conflicting_regs,
2849 &profitable_hard_regs);
2850 class_size = ira_class_hard_regs_num[aclass];
2851 /* Set up cost improvement for usage of each profitable hard
2852 register for allocno A. */
2853 for (j = 0; j < class_size; j++)
2855 hregno = ira_class_hard_regs[aclass][j];
2856 if (! check_hard_reg_p (a, hregno,
2857 conflicting_regs, profitable_hard_regs))
2858 continue;
2859 ira_assert (ira_class_hard_reg_index[aclass][hregno] == j);
2860 k = allocno_costs == NULL ? 0 : j;
2861 costs[hregno] = (allocno_costs == NULL
2862 ? ALLOCNO_UPDATED_CLASS_COST (a) : allocno_costs[k]);
2863 costs[hregno] -= allocno_copy_cost_saving (a, hregno);
2864 costs[hregno] -= base_cost;
2865 if (costs[hregno] < 0)
2866 try_p = true;
2868 if (! try_p)
2869 /* There is no chance to improve the allocation cost by
2870 assigning hard register to allocno A even without spilling
2871 conflicting allocnos. */
2872 continue;
2873 mode = ALLOCNO_MODE (a);
2874 nwords = ALLOCNO_NUM_OBJECTS (a);
2875 /* Process each allocno conflicting with A and update the cost
2876 improvement for profitable hard registers of A. To use a
2877 hard register for A we need to spill some conflicting
2878 allocnos and that creates penalty for the cost
2879 improvement. */
2880 for (word = 0; word < nwords; word++)
2882 ira_object_t conflict_obj;
2883 ira_object_t obj = ALLOCNO_OBJECT (a, word);
2884 ira_object_conflict_iterator oci;
2886 FOR_EACH_OBJECT_CONFLICT (obj, conflict_obj, oci)
2888 ira_allocno_t conflict_a = OBJECT_ALLOCNO (conflict_obj);
2890 if (ALLOCNO_COLOR_DATA (conflict_a)->temp == check)
2891 /* We already processed this conflicting allocno
2892 because we processed earlier another object of the
2893 conflicting allocno. */
2894 continue;
2895 ALLOCNO_COLOR_DATA (conflict_a)->temp = check;
2896 if ((conflict_hregno = ALLOCNO_HARD_REGNO (conflict_a)) < 0)
2897 continue;
2898 spill_cost = ALLOCNO_UPDATED_MEMORY_COST (conflict_a);
2899 k = (ira_class_hard_reg_index
2900 [ALLOCNO_CLASS (conflict_a)][conflict_hregno]);
2901 ira_assert (k >= 0);
2902 if ((allocno_costs = ALLOCNO_HARD_REG_COSTS (conflict_a))
2903 != NULL)
2904 spill_cost -= allocno_costs[k];
2905 else
2906 spill_cost -= ALLOCNO_UPDATED_CLASS_COST (conflict_a);
2907 spill_cost
2908 += allocno_copy_cost_saving (conflict_a, conflict_hregno);
2909 conflict_nregs = hard_regno_nregs (conflict_hregno,
2910 ALLOCNO_MODE (conflict_a));
2911 for (r = conflict_hregno;
2912 r >= 0 && (int) end_hard_regno (mode, r) > conflict_hregno;
2913 r--)
2914 if (check_hard_reg_p (a, r,
2915 conflicting_regs, profitable_hard_regs))
2916 costs[r] += spill_cost;
2917 for (r = conflict_hregno + 1;
2918 r < conflict_hregno + conflict_nregs;
2919 r++)
2920 if (check_hard_reg_p (a, r,
2921 conflicting_regs, profitable_hard_regs))
2922 costs[r] += spill_cost;
2925 min_cost = INT_MAX;
2926 best = -1;
2927 /* Now we choose hard register for A which results in highest
2928 allocation cost improvement. */
2929 for (j = 0; j < class_size; j++)
2931 hregno = ira_class_hard_regs[aclass][j];
2932 if (check_hard_reg_p (a, hregno,
2933 conflicting_regs, profitable_hard_regs)
2934 && min_cost > costs[hregno])
2936 best = hregno;
2937 min_cost = costs[hregno];
2940 if (min_cost >= 0)
2941 /* We are in a situation when assigning any hard register to A
2942 by spilling some conflicting allocnos does not improve the
2943 allocation cost. */
2944 continue;
2945 nregs = hard_regno_nregs (best, mode);
2946 /* Now spill conflicting allocnos which contain a hard register
2947 of A when we assign the best chosen hard register to it. */
2948 for (word = 0; word < nwords; word++)
2950 ira_object_t conflict_obj;
2951 ira_object_t obj = ALLOCNO_OBJECT (a, word);
2952 ira_object_conflict_iterator oci;
2954 FOR_EACH_OBJECT_CONFLICT (obj, conflict_obj, oci)
2956 ira_allocno_t conflict_a = OBJECT_ALLOCNO (conflict_obj);
2958 if ((conflict_hregno = ALLOCNO_HARD_REGNO (conflict_a)) < 0)
2959 continue;
2960 conflict_nregs = hard_regno_nregs (conflict_hregno,
2961 ALLOCNO_MODE (conflict_a));
2962 if (best + nregs <= conflict_hregno
2963 || conflict_hregno + conflict_nregs <= best)
2964 /* No intersection. */
2965 continue;
2966 ALLOCNO_HARD_REGNO (conflict_a) = -1;
2967 sorted_allocnos[n++] = conflict_a;
2968 if (internal_flag_ira_verbose > 2 && ira_dump_file != NULL)
2969 fprintf (ira_dump_file, "Spilling a%dr%d for a%dr%d\n",
2970 ALLOCNO_NUM (conflict_a), ALLOCNO_REGNO (conflict_a),
2971 ALLOCNO_NUM (a), ALLOCNO_REGNO (a));
2974 /* Assign the best chosen hard register to A. */
2975 ALLOCNO_HARD_REGNO (a) = best;
2976 if (internal_flag_ira_verbose > 2 && ira_dump_file != NULL)
2977 fprintf (ira_dump_file, "Assigning %d to a%dr%d\n",
2978 best, ALLOCNO_NUM (a), ALLOCNO_REGNO (a));
2980 if (n == 0)
2981 return;
2982 /* We spilled some allocnos to assign their hard registers to other
2983 allocnos. The spilled allocnos are now in array
2984 'sorted_allocnos'. There is still a possibility that some of the
2985 spilled allocnos can get hard registers. So let us try assign
2986 them hard registers again (just a reminder -- function
2987 'assign_hard_reg' assigns hard registers only if it is possible
2988 and profitable). We process the spilled allocnos with biggest
2989 benefit to get hard register first -- see function
2990 'allocno_cost_compare_func'. */
2991 qsort (sorted_allocnos, n, sizeof (ira_allocno_t),
2992 allocno_cost_compare_func);
2993 for (j = 0; j < n; j++)
2995 a = sorted_allocnos[j];
2996 ALLOCNO_ASSIGNED_P (a) = false;
2997 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
2999 fprintf (ira_dump_file, " ");
3000 ira_print_expanded_allocno (a);
3001 fprintf (ira_dump_file, " -- ");
3003 if (assign_hard_reg (a, false))
3005 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
3006 fprintf (ira_dump_file, "assign hard reg %d\n",
3007 ALLOCNO_HARD_REGNO (a));
3009 else
3011 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
3012 fprintf (ira_dump_file, "assign memory\n");
3017 /* Sort allocnos according to their priorities. */
3018 static int
3019 allocno_priority_compare_func (const void *v1p, const void *v2p)
3021 ira_allocno_t a1 = *(const ira_allocno_t *) v1p;
3022 ira_allocno_t a2 = *(const ira_allocno_t *) v2p;
3023 int pri1, pri2, diff;
3025 /* Assign hard reg to static chain pointer pseudo first when
3026 non-local goto is used. */
3027 if ((diff = (non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a2))
3028 - non_spilled_static_chain_regno_p (ALLOCNO_REGNO (a1)))) != 0)
3029 return diff;
3030 pri1 = allocno_priorities[ALLOCNO_NUM (a1)];
3031 pri2 = allocno_priorities[ALLOCNO_NUM (a2)];
3032 if (pri2 != pri1)
3033 return SORTGT (pri2, pri1);
3035 /* If regs are equally good, sort by allocnos, so that the results of
3036 qsort leave nothing to chance. */
3037 return ALLOCNO_NUM (a1) - ALLOCNO_NUM (a2);
3040 /* Chaitin-Briggs coloring for allocnos in COLORING_ALLOCNO_BITMAP
3041 taking into account allocnos in CONSIDERATION_ALLOCNO_BITMAP. */
3042 static void
3043 color_allocnos (void)
3045 unsigned int i, n;
3046 bitmap_iterator bi;
3047 ira_allocno_t a;
3049 setup_profitable_hard_regs ();
3050 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap, 0, i, bi)
3052 int l, nr;
3053 HARD_REG_SET conflict_hard_regs;
3054 allocno_color_data_t data;
3055 ira_pref_t pref, next_pref;
3057 a = ira_allocnos[i];
3058 nr = ALLOCNO_NUM_OBJECTS (a);
3059 CLEAR_HARD_REG_SET (conflict_hard_regs);
3060 for (l = 0; l < nr; l++)
3062 ira_object_t obj = ALLOCNO_OBJECT (a, l);
3063 IOR_HARD_REG_SET (conflict_hard_regs,
3064 OBJECT_CONFLICT_HARD_REGS (obj));
3066 data = ALLOCNO_COLOR_DATA (a);
3067 for (pref = ALLOCNO_PREFS (a); pref != NULL; pref = next_pref)
3069 next_pref = pref->next_pref;
3070 if (! ira_hard_reg_in_set_p (pref->hard_regno,
3071 ALLOCNO_MODE (a),
3072 data->profitable_hard_regs))
3073 ira_remove_pref (pref);
3076 if (flag_ira_algorithm == IRA_ALGORITHM_PRIORITY)
3078 n = 0;
3079 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap, 0, i, bi)
3081 a = ira_allocnos[i];
3082 if (ALLOCNO_CLASS (a) == NO_REGS)
3084 ALLOCNO_HARD_REGNO (a) = -1;
3085 ALLOCNO_ASSIGNED_P (a) = true;
3086 ira_assert (ALLOCNO_UPDATED_HARD_REG_COSTS (a) == NULL);
3087 ira_assert (ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (a) == NULL);
3088 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
3090 fprintf (ira_dump_file, " Spill");
3091 ira_print_expanded_allocno (a);
3092 fprintf (ira_dump_file, "\n");
3094 continue;
3096 sorted_allocnos[n++] = a;
3098 if (n != 0)
3100 setup_allocno_priorities (sorted_allocnos, n);
3101 qsort (sorted_allocnos, n, sizeof (ira_allocno_t),
3102 allocno_priority_compare_func);
3103 for (i = 0; i < n; i++)
3105 a = sorted_allocnos[i];
3106 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
3108 fprintf (ira_dump_file, " ");
3109 ira_print_expanded_allocno (a);
3110 fprintf (ira_dump_file, " -- ");
3112 if (assign_hard_reg (a, false))
3114 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
3115 fprintf (ira_dump_file, "assign hard reg %d\n",
3116 ALLOCNO_HARD_REGNO (a));
3118 else
3120 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
3121 fprintf (ira_dump_file, "assign memory\n");
3126 else
3128 form_allocno_hard_regs_nodes_forest ();
3129 if (internal_flag_ira_verbose > 2 && ira_dump_file != NULL)
3130 print_hard_regs_forest (ira_dump_file);
3131 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap, 0, i, bi)
3133 a = ira_allocnos[i];
3134 if (ALLOCNO_CLASS (a) != NO_REGS && ! empty_profitable_hard_regs (a))
3136 ALLOCNO_COLOR_DATA (a)->in_graph_p = true;
3137 update_costs_from_prefs (a);
3139 else
3141 ALLOCNO_HARD_REGNO (a) = -1;
3142 ALLOCNO_ASSIGNED_P (a) = true;
3143 /* We don't need updated costs anymore. */
3144 ira_free_allocno_updated_costs (a);
3145 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
3147 fprintf (ira_dump_file, " Spill");
3148 ira_print_expanded_allocno (a);
3149 fprintf (ira_dump_file, "\n");
3153 /* Put the allocnos into the corresponding buckets. */
3154 colorable_allocno_bucket = NULL;
3155 uncolorable_allocno_bucket = NULL;
3156 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap, 0, i, bi)
3158 a = ira_allocnos[i];
3159 if (ALLOCNO_COLOR_DATA (a)->in_graph_p)
3160 put_allocno_into_bucket (a);
3162 push_allocnos_to_stack ();
3163 pop_allocnos_from_stack ();
3164 finish_allocno_hard_regs_nodes_forest ();
3166 improve_allocation ();
3171 /* Output information about the loop given by its LOOP_TREE_NODE. */
3172 static void
3173 print_loop_title (ira_loop_tree_node_t loop_tree_node)
3175 unsigned int j;
3176 bitmap_iterator bi;
3177 ira_loop_tree_node_t subloop_node, dest_loop_node;
3178 edge e;
3179 edge_iterator ei;
3181 if (loop_tree_node->parent == NULL)
3182 fprintf (ira_dump_file,
3183 "\n Loop 0 (parent -1, header bb%d, depth 0)\n bbs:",
3184 NUM_FIXED_BLOCKS);
3185 else
3187 ira_assert (current_loops != NULL && loop_tree_node->loop != NULL);
3188 fprintf (ira_dump_file,
3189 "\n Loop %d (parent %d, header bb%d, depth %d)\n bbs:",
3190 loop_tree_node->loop_num, loop_tree_node->parent->loop_num,
3191 loop_tree_node->loop->header->index,
3192 loop_depth (loop_tree_node->loop));
3194 for (subloop_node = loop_tree_node->children;
3195 subloop_node != NULL;
3196 subloop_node = subloop_node->next)
3197 if (subloop_node->bb != NULL)
3199 fprintf (ira_dump_file, " %d", subloop_node->bb->index);
3200 FOR_EACH_EDGE (e, ei, subloop_node->bb->succs)
3201 if (e->dest != EXIT_BLOCK_PTR_FOR_FN (cfun)
3202 && ((dest_loop_node = IRA_BB_NODE (e->dest)->parent)
3203 != loop_tree_node))
3204 fprintf (ira_dump_file, "(->%d:l%d)",
3205 e->dest->index, dest_loop_node->loop_num);
3207 fprintf (ira_dump_file, "\n all:");
3208 EXECUTE_IF_SET_IN_BITMAP (loop_tree_node->all_allocnos, 0, j, bi)
3209 fprintf (ira_dump_file, " %dr%d", j, ALLOCNO_REGNO (ira_allocnos[j]));
3210 fprintf (ira_dump_file, "\n modified regnos:");
3211 EXECUTE_IF_SET_IN_BITMAP (loop_tree_node->modified_regnos, 0, j, bi)
3212 fprintf (ira_dump_file, " %d", j);
3213 fprintf (ira_dump_file, "\n border:");
3214 EXECUTE_IF_SET_IN_BITMAP (loop_tree_node->border_allocnos, 0, j, bi)
3215 fprintf (ira_dump_file, " %dr%d", j, ALLOCNO_REGNO (ira_allocnos[j]));
3216 fprintf (ira_dump_file, "\n Pressure:");
3217 for (j = 0; (int) j < ira_pressure_classes_num; j++)
3219 enum reg_class pclass;
3221 pclass = ira_pressure_classes[j];
3222 if (loop_tree_node->reg_pressure[pclass] == 0)
3223 continue;
3224 fprintf (ira_dump_file, " %s=%d", reg_class_names[pclass],
3225 loop_tree_node->reg_pressure[pclass]);
3227 fprintf (ira_dump_file, "\n");
3230 /* Color the allocnos inside loop (in the extreme case it can be all
3231 of the function) given the corresponding LOOP_TREE_NODE. The
3232 function is called for each loop during top-down traverse of the
3233 loop tree. */
3234 static void
3235 color_pass (ira_loop_tree_node_t loop_tree_node)
3237 int regno, hard_regno, index = -1, n;
3238 int cost, exit_freq, enter_freq;
3239 unsigned int j;
3240 bitmap_iterator bi;
3241 machine_mode mode;
3242 enum reg_class rclass, aclass, pclass;
3243 ira_allocno_t a, subloop_allocno;
3244 ira_loop_tree_node_t subloop_node;
3246 ira_assert (loop_tree_node->bb == NULL);
3247 if (internal_flag_ira_verbose > 1 && ira_dump_file != NULL)
3248 print_loop_title (loop_tree_node);
3250 bitmap_copy (coloring_allocno_bitmap, loop_tree_node->all_allocnos);
3251 bitmap_copy (consideration_allocno_bitmap, coloring_allocno_bitmap);
3252 n = 0;
3253 EXECUTE_IF_SET_IN_BITMAP (consideration_allocno_bitmap, 0, j, bi)
3255 a = ira_allocnos[j];
3256 n++;
3257 if (! ALLOCNO_ASSIGNED_P (a))
3258 continue;
3259 bitmap_clear_bit (coloring_allocno_bitmap, ALLOCNO_NUM (a));
3261 allocno_color_data
3262 = (allocno_color_data_t) ira_allocate (sizeof (struct allocno_color_data)
3263 * n);
3264 memset (allocno_color_data, 0, sizeof (struct allocno_color_data) * n);
3265 curr_allocno_process = 0;
3266 n = 0;
3267 EXECUTE_IF_SET_IN_BITMAP (consideration_allocno_bitmap, 0, j, bi)
3269 a = ira_allocnos[j];
3270 ALLOCNO_ADD_DATA (a) = allocno_color_data + n;
3271 n++;
3273 init_allocno_threads ();
3274 /* Color all mentioned allocnos including transparent ones. */
3275 color_allocnos ();
3276 /* Process caps. They are processed just once. */
3277 if (flag_ira_region == IRA_REGION_MIXED
3278 || flag_ira_region == IRA_REGION_ALL)
3279 EXECUTE_IF_SET_IN_BITMAP (loop_tree_node->all_allocnos, 0, j, bi)
3281 a = ira_allocnos[j];
3282 if (ALLOCNO_CAP_MEMBER (a) == NULL)
3283 continue;
3284 /* Remove from processing in the next loop. */
3285 bitmap_clear_bit (consideration_allocno_bitmap, j);
3286 rclass = ALLOCNO_CLASS (a);
3287 pclass = ira_pressure_class_translate[rclass];
3288 if (flag_ira_region == IRA_REGION_MIXED
3289 && (loop_tree_node->reg_pressure[pclass]
3290 <= ira_class_hard_regs_num[pclass]))
3292 mode = ALLOCNO_MODE (a);
3293 hard_regno = ALLOCNO_HARD_REGNO (a);
3294 if (hard_regno >= 0)
3296 index = ira_class_hard_reg_index[rclass][hard_regno];
3297 ira_assert (index >= 0);
3299 regno = ALLOCNO_REGNO (a);
3300 subloop_allocno = ALLOCNO_CAP_MEMBER (a);
3301 subloop_node = ALLOCNO_LOOP_TREE_NODE (subloop_allocno);
3302 ira_assert (!ALLOCNO_ASSIGNED_P (subloop_allocno));
3303 ALLOCNO_HARD_REGNO (subloop_allocno) = hard_regno;
3304 ALLOCNO_ASSIGNED_P (subloop_allocno) = true;
3305 if (hard_regno >= 0)
3306 update_costs_from_copies (subloop_allocno, true, true);
3307 /* We don't need updated costs anymore. */
3308 ira_free_allocno_updated_costs (subloop_allocno);
3311 /* Update costs of the corresponding allocnos (not caps) in the
3312 subloops. */
3313 for (subloop_node = loop_tree_node->subloops;
3314 subloop_node != NULL;
3315 subloop_node = subloop_node->subloop_next)
3317 ira_assert (subloop_node->bb == NULL);
3318 EXECUTE_IF_SET_IN_BITMAP (consideration_allocno_bitmap, 0, j, bi)
3320 a = ira_allocnos[j];
3321 ira_assert (ALLOCNO_CAP_MEMBER (a) == NULL);
3322 mode = ALLOCNO_MODE (a);
3323 rclass = ALLOCNO_CLASS (a);
3324 pclass = ira_pressure_class_translate[rclass];
3325 hard_regno = ALLOCNO_HARD_REGNO (a);
3326 /* Use hard register class here. ??? */
3327 if (hard_regno >= 0)
3329 index = ira_class_hard_reg_index[rclass][hard_regno];
3330 ira_assert (index >= 0);
3332 regno = ALLOCNO_REGNO (a);
3333 /* ??? conflict costs */
3334 subloop_allocno = subloop_node->regno_allocno_map[regno];
3335 if (subloop_allocno == NULL
3336 || ALLOCNO_CAP (subloop_allocno) != NULL)
3337 continue;
3338 ira_assert (ALLOCNO_CLASS (subloop_allocno) == rclass);
3339 ira_assert (bitmap_bit_p (subloop_node->all_allocnos,
3340 ALLOCNO_NUM (subloop_allocno)));
3341 if ((flag_ira_region == IRA_REGION_MIXED
3342 && (loop_tree_node->reg_pressure[pclass]
3343 <= ira_class_hard_regs_num[pclass]))
3344 || (pic_offset_table_rtx != NULL
3345 && regno == (int) REGNO (pic_offset_table_rtx))
3346 /* Avoid overlapped multi-registers. Moves between them
3347 might result in wrong code generation. */
3348 || (hard_regno >= 0
3349 && ira_reg_class_max_nregs[pclass][mode] > 1))
3351 if (! ALLOCNO_ASSIGNED_P (subloop_allocno))
3353 ALLOCNO_HARD_REGNO (subloop_allocno) = hard_regno;
3354 ALLOCNO_ASSIGNED_P (subloop_allocno) = true;
3355 if (hard_regno >= 0)
3356 update_costs_from_copies (subloop_allocno, true, true);
3357 /* We don't need updated costs anymore. */
3358 ira_free_allocno_updated_costs (subloop_allocno);
3360 continue;
3362 exit_freq = ira_loop_edge_freq (subloop_node, regno, true);
3363 enter_freq = ira_loop_edge_freq (subloop_node, regno, false);
3364 ira_assert (regno < ira_reg_equiv_len);
3365 if (ira_equiv_no_lvalue_p (regno))
3367 if (! ALLOCNO_ASSIGNED_P (subloop_allocno))
3369 ALLOCNO_HARD_REGNO (subloop_allocno) = hard_regno;
3370 ALLOCNO_ASSIGNED_P (subloop_allocno) = true;
3371 if (hard_regno >= 0)
3372 update_costs_from_copies (subloop_allocno, true, true);
3373 /* We don't need updated costs anymore. */
3374 ira_free_allocno_updated_costs (subloop_allocno);
3377 else if (hard_regno < 0)
3379 ALLOCNO_UPDATED_MEMORY_COST (subloop_allocno)
3380 -= ((ira_memory_move_cost[mode][rclass][1] * enter_freq)
3381 + (ira_memory_move_cost[mode][rclass][0] * exit_freq));
3383 else
3385 aclass = ALLOCNO_CLASS (subloop_allocno);
3386 ira_init_register_move_cost_if_necessary (mode);
3387 cost = (ira_register_move_cost[mode][rclass][rclass]
3388 * (exit_freq + enter_freq));
3389 ira_allocate_and_set_or_copy_costs
3390 (&ALLOCNO_UPDATED_HARD_REG_COSTS (subloop_allocno), aclass,
3391 ALLOCNO_UPDATED_CLASS_COST (subloop_allocno),
3392 ALLOCNO_HARD_REG_COSTS (subloop_allocno));
3393 ira_allocate_and_set_or_copy_costs
3394 (&ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (subloop_allocno),
3395 aclass, 0, ALLOCNO_CONFLICT_HARD_REG_COSTS (subloop_allocno));
3396 ALLOCNO_UPDATED_HARD_REG_COSTS (subloop_allocno)[index] -= cost;
3397 ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (subloop_allocno)[index]
3398 -= cost;
3399 if (ALLOCNO_UPDATED_CLASS_COST (subloop_allocno)
3400 > ALLOCNO_UPDATED_HARD_REG_COSTS (subloop_allocno)[index])
3401 ALLOCNO_UPDATED_CLASS_COST (subloop_allocno)
3402 = ALLOCNO_UPDATED_HARD_REG_COSTS (subloop_allocno)[index];
3403 ALLOCNO_UPDATED_MEMORY_COST (subloop_allocno)
3404 += (ira_memory_move_cost[mode][rclass][0] * enter_freq
3405 + ira_memory_move_cost[mode][rclass][1] * exit_freq);
3409 ira_free (allocno_color_data);
3410 EXECUTE_IF_SET_IN_BITMAP (consideration_allocno_bitmap, 0, j, bi)
3412 a = ira_allocnos[j];
3413 ALLOCNO_ADD_DATA (a) = NULL;
3417 /* Initialize the common data for coloring and calls functions to do
3418 Chaitin-Briggs and regional coloring. */
3419 static void
3420 do_coloring (void)
3422 coloring_allocno_bitmap = ira_allocate_bitmap ();
3423 if (internal_flag_ira_verbose > 0 && ira_dump_file != NULL)
3424 fprintf (ira_dump_file, "\n**** Allocnos coloring:\n\n");
3426 ira_traverse_loop_tree (false, ira_loop_tree_root, color_pass, NULL);
3428 if (internal_flag_ira_verbose > 1 && ira_dump_file != NULL)
3429 ira_print_disposition (ira_dump_file);
3431 ira_free_bitmap (coloring_allocno_bitmap);
3436 /* Move spill/restore code, which are to be generated in ira-emit.c,
3437 to less frequent points (if it is profitable) by reassigning some
3438 allocnos (in loop with subloops containing in another loop) to
3439 memory which results in longer live-range where the corresponding
3440 pseudo-registers will be in memory. */
3441 static void
3442 move_spill_restore (void)
3444 int cost, regno, hard_regno, hard_regno2, index;
3445 bool changed_p;
3446 int enter_freq, exit_freq;
3447 machine_mode mode;
3448 enum reg_class rclass;
3449 ira_allocno_t a, parent_allocno, subloop_allocno;
3450 ira_loop_tree_node_t parent, loop_node, subloop_node;
3451 ira_allocno_iterator ai;
3453 for (;;)
3455 changed_p = false;
3456 if (internal_flag_ira_verbose > 0 && ira_dump_file != NULL)
3457 fprintf (ira_dump_file, "New iteration of spill/restore move\n");
3458 FOR_EACH_ALLOCNO (a, ai)
3460 regno = ALLOCNO_REGNO (a);
3461 loop_node = ALLOCNO_LOOP_TREE_NODE (a);
3462 if (ALLOCNO_CAP_MEMBER (a) != NULL
3463 || ALLOCNO_CAP (a) != NULL
3464 || (hard_regno = ALLOCNO_HARD_REGNO (a)) < 0
3465 || loop_node->children == NULL
3466 /* don't do the optimization because it can create
3467 copies and the reload pass can spill the allocno set
3468 by copy although the allocno will not get memory
3469 slot. */
3470 || ira_equiv_no_lvalue_p (regno)
3471 || !bitmap_bit_p (loop_node->border_allocnos, ALLOCNO_NUM (a))
3472 /* Do not spill static chain pointer pseudo when
3473 non-local goto is used. */
3474 || non_spilled_static_chain_regno_p (regno))
3475 continue;
3476 mode = ALLOCNO_MODE (a);
3477 rclass = ALLOCNO_CLASS (a);
3478 index = ira_class_hard_reg_index[rclass][hard_regno];
3479 ira_assert (index >= 0);
3480 cost = (ALLOCNO_MEMORY_COST (a)
3481 - (ALLOCNO_HARD_REG_COSTS (a) == NULL
3482 ? ALLOCNO_CLASS_COST (a)
3483 : ALLOCNO_HARD_REG_COSTS (a)[index]));
3484 ira_init_register_move_cost_if_necessary (mode);
3485 for (subloop_node = loop_node->subloops;
3486 subloop_node != NULL;
3487 subloop_node = subloop_node->subloop_next)
3489 ira_assert (subloop_node->bb == NULL);
3490 subloop_allocno = subloop_node->regno_allocno_map[regno];
3491 if (subloop_allocno == NULL)
3492 continue;
3493 ira_assert (rclass == ALLOCNO_CLASS (subloop_allocno));
3494 /* We have accumulated cost. To get the real cost of
3495 allocno usage in the loop we should subtract costs of
3496 the subloop allocnos. */
3497 cost -= (ALLOCNO_MEMORY_COST (subloop_allocno)
3498 - (ALLOCNO_HARD_REG_COSTS (subloop_allocno) == NULL
3499 ? ALLOCNO_CLASS_COST (subloop_allocno)
3500 : ALLOCNO_HARD_REG_COSTS (subloop_allocno)[index]));
3501 exit_freq = ira_loop_edge_freq (subloop_node, regno, true);
3502 enter_freq = ira_loop_edge_freq (subloop_node, regno, false);
3503 if ((hard_regno2 = ALLOCNO_HARD_REGNO (subloop_allocno)) < 0)
3504 cost -= (ira_memory_move_cost[mode][rclass][0] * exit_freq
3505 + ira_memory_move_cost[mode][rclass][1] * enter_freq);
3506 else
3508 cost
3509 += (ira_memory_move_cost[mode][rclass][0] * exit_freq
3510 + ira_memory_move_cost[mode][rclass][1] * enter_freq);
3511 if (hard_regno2 != hard_regno)
3512 cost -= (ira_register_move_cost[mode][rclass][rclass]
3513 * (exit_freq + enter_freq));
3516 if ((parent = loop_node->parent) != NULL
3517 && (parent_allocno = parent->regno_allocno_map[regno]) != NULL)
3519 ira_assert (rclass == ALLOCNO_CLASS (parent_allocno));
3520 exit_freq = ira_loop_edge_freq (loop_node, regno, true);
3521 enter_freq = ira_loop_edge_freq (loop_node, regno, false);
3522 if ((hard_regno2 = ALLOCNO_HARD_REGNO (parent_allocno)) < 0)
3523 cost -= (ira_memory_move_cost[mode][rclass][0] * exit_freq
3524 + ira_memory_move_cost[mode][rclass][1] * enter_freq);
3525 else
3527 cost
3528 += (ira_memory_move_cost[mode][rclass][1] * exit_freq
3529 + ira_memory_move_cost[mode][rclass][0] * enter_freq);
3530 if (hard_regno2 != hard_regno)
3531 cost -= (ira_register_move_cost[mode][rclass][rclass]
3532 * (exit_freq + enter_freq));
3535 if (cost < 0)
3537 ALLOCNO_HARD_REGNO (a) = -1;
3538 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
3540 fprintf
3541 (ira_dump_file,
3542 " Moving spill/restore for a%dr%d up from loop %d",
3543 ALLOCNO_NUM (a), regno, loop_node->loop_num);
3544 fprintf (ira_dump_file, " - profit %d\n", -cost);
3546 changed_p = true;
3549 if (! changed_p)
3550 break;
3556 /* Update current hard reg costs and current conflict hard reg costs
3557 for allocno A. It is done by processing its copies containing
3558 other allocnos already assigned. */
3559 static void
3560 update_curr_costs (ira_allocno_t a)
3562 int i, hard_regno, cost;
3563 machine_mode mode;
3564 enum reg_class aclass, rclass;
3565 ira_allocno_t another_a;
3566 ira_copy_t cp, next_cp;
3568 ira_free_allocno_updated_costs (a);
3569 ira_assert (! ALLOCNO_ASSIGNED_P (a));
3570 aclass = ALLOCNO_CLASS (a);
3571 if (aclass == NO_REGS)
3572 return;
3573 mode = ALLOCNO_MODE (a);
3574 ira_init_register_move_cost_if_necessary (mode);
3575 for (cp = ALLOCNO_COPIES (a); cp != NULL; cp = next_cp)
3577 if (cp->first == a)
3579 next_cp = cp->next_first_allocno_copy;
3580 another_a = cp->second;
3582 else if (cp->second == a)
3584 next_cp = cp->next_second_allocno_copy;
3585 another_a = cp->first;
3587 else
3588 gcc_unreachable ();
3589 if (! ira_reg_classes_intersect_p[aclass][ALLOCNO_CLASS (another_a)]
3590 || ! ALLOCNO_ASSIGNED_P (another_a)
3591 || (hard_regno = ALLOCNO_HARD_REGNO (another_a)) < 0)
3592 continue;
3593 rclass = REGNO_REG_CLASS (hard_regno);
3594 i = ira_class_hard_reg_index[aclass][hard_regno];
3595 if (i < 0)
3596 continue;
3597 cost = (cp->first == a
3598 ? ira_register_move_cost[mode][rclass][aclass]
3599 : ira_register_move_cost[mode][aclass][rclass]);
3600 ira_allocate_and_set_or_copy_costs
3601 (&ALLOCNO_UPDATED_HARD_REG_COSTS (a), aclass, ALLOCNO_CLASS_COST (a),
3602 ALLOCNO_HARD_REG_COSTS (a));
3603 ira_allocate_and_set_or_copy_costs
3604 (&ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (a),
3605 aclass, 0, ALLOCNO_CONFLICT_HARD_REG_COSTS (a));
3606 ALLOCNO_UPDATED_HARD_REG_COSTS (a)[i] -= cp->freq * cost;
3607 ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (a)[i] -= cp->freq * cost;
3611 /* Try to assign hard registers to the unassigned allocnos and
3612 allocnos conflicting with them or conflicting with allocnos whose
3613 regno >= START_REGNO. The function is called after ira_flattening,
3614 so more allocnos (including ones created in ira-emit.c) will have a
3615 chance to get a hard register. We use simple assignment algorithm
3616 based on priorities. */
3617 void
3618 ira_reassign_conflict_allocnos (int start_regno)
3620 int i, allocnos_to_color_num;
3621 ira_allocno_t a;
3622 enum reg_class aclass;
3623 bitmap allocnos_to_color;
3624 ira_allocno_iterator ai;
3626 allocnos_to_color = ira_allocate_bitmap ();
3627 allocnos_to_color_num = 0;
3628 FOR_EACH_ALLOCNO (a, ai)
3630 int n = ALLOCNO_NUM_OBJECTS (a);
3632 if (! ALLOCNO_ASSIGNED_P (a)
3633 && ! bitmap_bit_p (allocnos_to_color, ALLOCNO_NUM (a)))
3635 if (ALLOCNO_CLASS (a) != NO_REGS)
3636 sorted_allocnos[allocnos_to_color_num++] = a;
3637 else
3639 ALLOCNO_ASSIGNED_P (a) = true;
3640 ALLOCNO_HARD_REGNO (a) = -1;
3641 ira_assert (ALLOCNO_UPDATED_HARD_REG_COSTS (a) == NULL);
3642 ira_assert (ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS (a) == NULL);
3644 bitmap_set_bit (allocnos_to_color, ALLOCNO_NUM (a));
3646 if (ALLOCNO_REGNO (a) < start_regno
3647 || (aclass = ALLOCNO_CLASS (a)) == NO_REGS)
3648 continue;
3649 for (i = 0; i < n; i++)
3651 ira_object_t obj = ALLOCNO_OBJECT (a, i);
3652 ira_object_t conflict_obj;
3653 ira_object_conflict_iterator oci;
3655 FOR_EACH_OBJECT_CONFLICT (obj, conflict_obj, oci)
3657 ira_allocno_t conflict_a = OBJECT_ALLOCNO (conflict_obj);
3659 ira_assert (ira_reg_classes_intersect_p
3660 [aclass][ALLOCNO_CLASS (conflict_a)]);
3661 if (!bitmap_set_bit (allocnos_to_color, ALLOCNO_NUM (conflict_a)))
3662 continue;
3663 sorted_allocnos[allocnos_to_color_num++] = conflict_a;
3667 ira_free_bitmap (allocnos_to_color);
3668 if (allocnos_to_color_num > 1)
3670 setup_allocno_priorities (sorted_allocnos, allocnos_to_color_num);
3671 qsort (sorted_allocnos, allocnos_to_color_num, sizeof (ira_allocno_t),
3672 allocno_priority_compare_func);
3674 for (i = 0; i < allocnos_to_color_num; i++)
3676 a = sorted_allocnos[i];
3677 ALLOCNO_ASSIGNED_P (a) = false;
3678 update_curr_costs (a);
3680 for (i = 0; i < allocnos_to_color_num; i++)
3682 a = sorted_allocnos[i];
3683 if (assign_hard_reg (a, true))
3685 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
3686 fprintf
3687 (ira_dump_file,
3688 " Secondary allocation: assign hard reg %d to reg %d\n",
3689 ALLOCNO_HARD_REGNO (a), ALLOCNO_REGNO (a));
3696 /* This page contains functions used to find conflicts using allocno
3697 live ranges. */
3699 #ifdef ENABLE_IRA_CHECKING
3701 /* Return TRUE if live ranges of pseudo-registers REGNO1 and REGNO2
3702 intersect. This should be used when there is only one region.
3703 Currently this is used during reload. */
3704 static bool
3705 conflict_by_live_ranges_p (int regno1, int regno2)
3707 ira_allocno_t a1, a2;
3709 ira_assert (regno1 >= FIRST_PSEUDO_REGISTER
3710 && regno2 >= FIRST_PSEUDO_REGISTER);
3711 /* Reg info calculated by dataflow infrastructure can be different
3712 from one calculated by regclass. */
3713 if ((a1 = ira_loop_tree_root->regno_allocno_map[regno1]) == NULL
3714 || (a2 = ira_loop_tree_root->regno_allocno_map[regno2]) == NULL)
3715 return false;
3716 return allocnos_conflict_by_live_ranges_p (a1, a2);
3719 #endif
3723 /* This page contains code to coalesce memory stack slots used by
3724 spilled allocnos. This results in smaller stack frame, better data
3725 locality, and in smaller code for some architectures like
3726 x86/x86_64 where insn size depends on address displacement value.
3727 On the other hand, it can worsen insn scheduling after the RA but
3728 in practice it is less important than smaller stack frames. */
3730 /* TRUE if we coalesced some allocnos. In other words, if we got
3731 loops formed by members first_coalesced_allocno and
3732 next_coalesced_allocno containing more one allocno. */
3733 static bool allocno_coalesced_p;
3735 /* Bitmap used to prevent a repeated allocno processing because of
3736 coalescing. */
3737 static bitmap processed_coalesced_allocno_bitmap;
3739 /* See below. */
3740 typedef struct coalesce_data *coalesce_data_t;
3742 /* To decrease footprint of ira_allocno structure we store all data
3743 needed only for coalescing in the following structure. */
3744 struct coalesce_data
3746 /* Coalesced allocnos form a cyclic list. One allocno given by
3747 FIRST represents all coalesced allocnos. The
3748 list is chained by NEXT. */
3749 ira_allocno_t first;
3750 ira_allocno_t next;
3751 int temp;
3754 /* Container for storing allocno data concerning coalescing. */
3755 static coalesce_data_t allocno_coalesce_data;
3757 /* Macro to access the data concerning coalescing. */
3758 #define ALLOCNO_COALESCE_DATA(a) ((coalesce_data_t) ALLOCNO_ADD_DATA (a))
3760 /* Merge two sets of coalesced allocnos given correspondingly by
3761 allocnos A1 and A2 (more accurately merging A2 set into A1
3762 set). */
3763 static void
3764 merge_allocnos (ira_allocno_t a1, ira_allocno_t a2)
3766 ira_allocno_t a, first, last, next;
3768 first = ALLOCNO_COALESCE_DATA (a1)->first;
3769 a = ALLOCNO_COALESCE_DATA (a2)->first;
3770 if (first == a)
3771 return;
3772 for (last = a2, a = ALLOCNO_COALESCE_DATA (a2)->next;;
3773 a = ALLOCNO_COALESCE_DATA (a)->next)
3775 ALLOCNO_COALESCE_DATA (a)->first = first;
3776 if (a == a2)
3777 break;
3778 last = a;
3780 next = allocno_coalesce_data[ALLOCNO_NUM (first)].next;
3781 allocno_coalesce_data[ALLOCNO_NUM (first)].next = a2;
3782 allocno_coalesce_data[ALLOCNO_NUM (last)].next = next;
3785 /* Return TRUE if there are conflicting allocnos from two sets of
3786 coalesced allocnos given correspondingly by allocnos A1 and A2. We
3787 use live ranges to find conflicts because conflicts are represented
3788 only for allocnos of the same allocno class and during the reload
3789 pass we coalesce allocnos for sharing stack memory slots. */
3790 static bool
3791 coalesced_allocno_conflict_p (ira_allocno_t a1, ira_allocno_t a2)
3793 ira_allocno_t a, conflict_a;
3795 if (allocno_coalesced_p)
3797 bitmap_clear (processed_coalesced_allocno_bitmap);
3798 for (a = ALLOCNO_COALESCE_DATA (a1)->next;;
3799 a = ALLOCNO_COALESCE_DATA (a)->next)
3801 bitmap_set_bit (processed_coalesced_allocno_bitmap, ALLOCNO_NUM (a));
3802 if (a == a1)
3803 break;
3806 for (a = ALLOCNO_COALESCE_DATA (a2)->next;;
3807 a = ALLOCNO_COALESCE_DATA (a)->next)
3809 for (conflict_a = ALLOCNO_COALESCE_DATA (a1)->next;;
3810 conflict_a = ALLOCNO_COALESCE_DATA (conflict_a)->next)
3812 if (allocnos_conflict_by_live_ranges_p (a, conflict_a))
3813 return true;
3814 if (conflict_a == a1)
3815 break;
3817 if (a == a2)
3818 break;
3820 return false;
3823 /* The major function for aggressive allocno coalescing. We coalesce
3824 only spilled allocnos. If some allocnos have been coalesced, we
3825 set up flag allocno_coalesced_p. */
3826 static void
3827 coalesce_allocnos (void)
3829 ira_allocno_t a;
3830 ira_copy_t cp, next_cp;
3831 unsigned int j;
3832 int i, n, cp_num, regno;
3833 bitmap_iterator bi;
3835 cp_num = 0;
3836 /* Collect copies. */
3837 EXECUTE_IF_SET_IN_BITMAP (coloring_allocno_bitmap, 0, j, bi)
3839 a = ira_allocnos[j];
3840 regno = ALLOCNO_REGNO (a);
3841 if (! ALLOCNO_ASSIGNED_P (a) || ALLOCNO_HARD_REGNO (a) >= 0
3842 || ira_equiv_no_lvalue_p (regno))
3843 continue;
3844 for (cp = ALLOCNO_COPIES (a); cp != NULL; cp = next_cp)
3846 if (cp->first == a)
3848 next_cp = cp->next_first_allocno_copy;
3849 regno = ALLOCNO_REGNO (cp->second);
3850 /* For priority coloring we coalesce allocnos only with
3851 the same allocno class not with intersected allocno
3852 classes as it were possible. It is done for
3853 simplicity. */
3854 if ((cp->insn != NULL || cp->constraint_p)
3855 && ALLOCNO_ASSIGNED_P (cp->second)
3856 && ALLOCNO_HARD_REGNO (cp->second) < 0
3857 && ! ira_equiv_no_lvalue_p (regno))
3858 sorted_copies[cp_num++] = cp;
3860 else if (cp->second == a)
3861 next_cp = cp->next_second_allocno_copy;
3862 else
3863 gcc_unreachable ();
3866 qsort (sorted_copies, cp_num, sizeof (ira_copy_t), copy_freq_compare_func);
3867 /* Coalesced copies, most frequently executed first. */
3868 for (; cp_num != 0;)
3870 for (i = 0; i < cp_num; i++)
3872 cp = sorted_copies[i];
3873 if (! coalesced_allocno_conflict_p (cp->first, cp->second))
3875 allocno_coalesced_p = true;
3876 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
3877 fprintf
3878 (ira_dump_file,
3879 " Coalescing copy %d:a%dr%d-a%dr%d (freq=%d)\n",
3880 cp->num, ALLOCNO_NUM (cp->first), ALLOCNO_REGNO (cp->first),
3881 ALLOCNO_NUM (cp->second), ALLOCNO_REGNO (cp->second),
3882 cp->freq);
3883 merge_allocnos (cp->first, cp->second);
3884 i++;
3885 break;
3888 /* Collect the rest of copies. */
3889 for (n = 0; i < cp_num; i++)
3891 cp = sorted_copies[i];
3892 if (allocno_coalesce_data[ALLOCNO_NUM (cp->first)].first
3893 != allocno_coalesce_data[ALLOCNO_NUM (cp->second)].first)
3894 sorted_copies[n++] = cp;
3896 cp_num = n;
3900 /* Usage cost and order number of coalesced allocno set to which
3901 given pseudo register belongs to. */
3902 static int *regno_coalesced_allocno_cost;
3903 static int *regno_coalesced_allocno_num;
3905 /* Sort pseudos according frequencies of coalesced allocno sets they
3906 belong to (putting most frequently ones first), and according to
3907 coalesced allocno set order numbers. */
3908 static int
3909 coalesced_pseudo_reg_freq_compare (const void *v1p, const void *v2p)
3911 const int regno1 = *(const int *) v1p;
3912 const int regno2 = *(const int *) v2p;
3913 int diff;
3915 if ((diff = (regno_coalesced_allocno_cost[regno2]
3916 - regno_coalesced_allocno_cost[regno1])) != 0)
3917 return diff;
3918 if ((diff = (regno_coalesced_allocno_num[regno1]
3919 - regno_coalesced_allocno_num[regno2])) != 0)
3920 return diff;
3921 return regno1 - regno2;
3924 /* Widest width in which each pseudo reg is referred to (via subreg).
3925 It is used for sorting pseudo registers. */
3926 static machine_mode *regno_max_ref_mode;
3928 /* Sort pseudos according their slot numbers (putting ones with
3929 smaller numbers first, or last when the frame pointer is not
3930 needed). */
3931 static int
3932 coalesced_pseudo_reg_slot_compare (const void *v1p, const void *v2p)
3934 const int regno1 = *(const int *) v1p;
3935 const int regno2 = *(const int *) v2p;
3936 ira_allocno_t a1 = ira_regno_allocno_map[regno1];
3937 ira_allocno_t a2 = ira_regno_allocno_map[regno2];
3938 int diff, slot_num1, slot_num2;
3939 machine_mode mode1, mode2;
3941 if (a1 == NULL || ALLOCNO_HARD_REGNO (a1) >= 0)
3943 if (a2 == NULL || ALLOCNO_HARD_REGNO (a2) >= 0)
3944 return regno1 - regno2;
3945 return 1;
3947 else if (a2 == NULL || ALLOCNO_HARD_REGNO (a2) >= 0)
3948 return -1;
3949 slot_num1 = -ALLOCNO_HARD_REGNO (a1);
3950 slot_num2 = -ALLOCNO_HARD_REGNO (a2);
3951 if ((diff = slot_num1 - slot_num2) != 0)
3952 return (frame_pointer_needed
3953 || (!FRAME_GROWS_DOWNWARD) == STACK_GROWS_DOWNWARD ? diff : -diff);
3954 mode1 = wider_subreg_mode (PSEUDO_REGNO_MODE (regno1),
3955 regno_max_ref_mode[regno1]);
3956 mode2 = wider_subreg_mode (PSEUDO_REGNO_MODE (regno2),
3957 regno_max_ref_mode[regno2]);
3958 if ((diff = compare_sizes_for_sort (GET_MODE_SIZE (mode2),
3959 GET_MODE_SIZE (mode1))) != 0)
3960 return diff;
3961 return regno1 - regno2;
3964 /* Setup REGNO_COALESCED_ALLOCNO_COST and REGNO_COALESCED_ALLOCNO_NUM
3965 for coalesced allocno sets containing allocnos with their regnos
3966 given in array PSEUDO_REGNOS of length N. */
3967 static void
3968 setup_coalesced_allocno_costs_and_nums (int *pseudo_regnos, int n)
3970 int i, num, regno, cost;
3971 ira_allocno_t allocno, a;
3973 for (num = i = 0; i < n; i++)
3975 regno = pseudo_regnos[i];
3976 allocno = ira_regno_allocno_map[regno];
3977 if (allocno == NULL)
3979 regno_coalesced_allocno_cost[regno] = 0;
3980 regno_coalesced_allocno_num[regno] = ++num;
3981 continue;
3983 if (ALLOCNO_COALESCE_DATA (allocno)->first != allocno)
3984 continue;
3985 num++;
3986 for (cost = 0, a = ALLOCNO_COALESCE_DATA (allocno)->next;;
3987 a = ALLOCNO_COALESCE_DATA (a)->next)
3989 cost += ALLOCNO_FREQ (a);
3990 if (a == allocno)
3991 break;
3993 for (a = ALLOCNO_COALESCE_DATA (allocno)->next;;
3994 a = ALLOCNO_COALESCE_DATA (a)->next)
3996 regno_coalesced_allocno_num[ALLOCNO_REGNO (a)] = num;
3997 regno_coalesced_allocno_cost[ALLOCNO_REGNO (a)] = cost;
3998 if (a == allocno)
3999 break;
4004 /* Collect spilled allocnos representing coalesced allocno sets (the
4005 first coalesced allocno). The collected allocnos are returned
4006 through array SPILLED_COALESCED_ALLOCNOS. The function returns the
4007 number of the collected allocnos. The allocnos are given by their
4008 regnos in array PSEUDO_REGNOS of length N. */
4009 static int
4010 collect_spilled_coalesced_allocnos (int *pseudo_regnos, int n,
4011 ira_allocno_t *spilled_coalesced_allocnos)
4013 int i, num, regno;
4014 ira_allocno_t allocno;
4016 for (num = i = 0; i < n; i++)
4018 regno = pseudo_regnos[i];
4019 allocno = ira_regno_allocno_map[regno];
4020 if (allocno == NULL || ALLOCNO_HARD_REGNO (allocno) >= 0
4021 || ALLOCNO_COALESCE_DATA (allocno)->first != allocno)
4022 continue;
4023 spilled_coalesced_allocnos[num++] = allocno;
4025 return num;
4028 /* Array of live ranges of size IRA_ALLOCNOS_NUM. Live range for
4029 given slot contains live ranges of coalesced allocnos assigned to
4030 given slot. */
4031 static live_range_t *slot_coalesced_allocnos_live_ranges;
4033 /* Return TRUE if coalesced allocnos represented by ALLOCNO has live
4034 ranges intersected with live ranges of coalesced allocnos assigned
4035 to slot with number N. */
4036 static bool
4037 slot_coalesced_allocno_live_ranges_intersect_p (ira_allocno_t allocno, int n)
4039 ira_allocno_t a;
4041 for (a = ALLOCNO_COALESCE_DATA (allocno)->next;;
4042 a = ALLOCNO_COALESCE_DATA (a)->next)
4044 int i;
4045 int nr = ALLOCNO_NUM_OBJECTS (a);
4046 gcc_assert (ALLOCNO_CAP_MEMBER (a) == NULL);
4047 for (i = 0; i < nr; i++)
4049 ira_object_t obj = ALLOCNO_OBJECT (a, i);
4051 if (ira_live_ranges_intersect_p
4052 (slot_coalesced_allocnos_live_ranges[n],
4053 OBJECT_LIVE_RANGES (obj)))
4054 return true;
4056 if (a == allocno)
4057 break;
4059 return false;
4062 /* Update live ranges of slot to which coalesced allocnos represented
4063 by ALLOCNO were assigned. */
4064 static void
4065 setup_slot_coalesced_allocno_live_ranges (ira_allocno_t allocno)
4067 int i, n;
4068 ira_allocno_t a;
4069 live_range_t r;
4071 n = ALLOCNO_COALESCE_DATA (allocno)->temp;
4072 for (a = ALLOCNO_COALESCE_DATA (allocno)->next;;
4073 a = ALLOCNO_COALESCE_DATA (a)->next)
4075 int nr = ALLOCNO_NUM_OBJECTS (a);
4076 gcc_assert (ALLOCNO_CAP_MEMBER (a) == NULL);
4077 for (i = 0; i < nr; i++)
4079 ira_object_t obj = ALLOCNO_OBJECT (a, i);
4081 r = ira_copy_live_range_list (OBJECT_LIVE_RANGES (obj));
4082 slot_coalesced_allocnos_live_ranges[n]
4083 = ira_merge_live_ranges
4084 (slot_coalesced_allocnos_live_ranges[n], r);
4086 if (a == allocno)
4087 break;
4091 /* We have coalesced allocnos involving in copies. Coalesce allocnos
4092 further in order to share the same memory stack slot. Allocnos
4093 representing sets of allocnos coalesced before the call are given
4094 in array SPILLED_COALESCED_ALLOCNOS of length NUM. Return TRUE if
4095 some allocnos were coalesced in the function. */
4096 static bool
4097 coalesce_spill_slots (ira_allocno_t *spilled_coalesced_allocnos, int num)
4099 int i, j, n, last_coalesced_allocno_num;
4100 ira_allocno_t allocno, a;
4101 bool merged_p = false;
4102 bitmap set_jump_crosses = regstat_get_setjmp_crosses ();
4104 slot_coalesced_allocnos_live_ranges
4105 = (live_range_t *) ira_allocate (sizeof (live_range_t) * ira_allocnos_num);
4106 memset (slot_coalesced_allocnos_live_ranges, 0,
4107 sizeof (live_range_t) * ira_allocnos_num);
4108 last_coalesced_allocno_num = 0;
4109 /* Coalesce non-conflicting spilled allocnos preferring most
4110 frequently used. */
4111 for (i = 0; i < num; i++)
4113 allocno = spilled_coalesced_allocnos[i];
4114 if (ALLOCNO_COALESCE_DATA (allocno)->first != allocno
4115 || bitmap_bit_p (set_jump_crosses, ALLOCNO_REGNO (allocno))
4116 || ira_equiv_no_lvalue_p (ALLOCNO_REGNO (allocno)))
4117 continue;
4118 for (j = 0; j < i; j++)
4120 a = spilled_coalesced_allocnos[j];
4121 n = ALLOCNO_COALESCE_DATA (a)->temp;
4122 if (ALLOCNO_COALESCE_DATA (a)->first == a
4123 && ! bitmap_bit_p (set_jump_crosses, ALLOCNO_REGNO (a))
4124 && ! ira_equiv_no_lvalue_p (ALLOCNO_REGNO (a))
4125 && ! slot_coalesced_allocno_live_ranges_intersect_p (allocno, n))
4126 break;
4128 if (j >= i)
4130 /* No coalescing: set up number for coalesced allocnos
4131 represented by ALLOCNO. */
4132 ALLOCNO_COALESCE_DATA (allocno)->temp = last_coalesced_allocno_num++;
4133 setup_slot_coalesced_allocno_live_ranges (allocno);
4135 else
4137 allocno_coalesced_p = true;
4138 merged_p = true;
4139 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
4140 fprintf (ira_dump_file,
4141 " Coalescing spilled allocnos a%dr%d->a%dr%d\n",
4142 ALLOCNO_NUM (allocno), ALLOCNO_REGNO (allocno),
4143 ALLOCNO_NUM (a), ALLOCNO_REGNO (a));
4144 ALLOCNO_COALESCE_DATA (allocno)->temp
4145 = ALLOCNO_COALESCE_DATA (a)->temp;
4146 setup_slot_coalesced_allocno_live_ranges (allocno);
4147 merge_allocnos (a, allocno);
4148 ira_assert (ALLOCNO_COALESCE_DATA (a)->first == a);
4151 for (i = 0; i < ira_allocnos_num; i++)
4152 ira_finish_live_range_list (slot_coalesced_allocnos_live_ranges[i]);
4153 ira_free (slot_coalesced_allocnos_live_ranges);
4154 return merged_p;
4157 /* Sort pseudo-register numbers in array PSEUDO_REGNOS of length N for
4158 subsequent assigning stack slots to them in the reload pass. To do
4159 this we coalesce spilled allocnos first to decrease the number of
4160 memory-memory move insns. This function is called by the
4161 reload. */
4162 void
4163 ira_sort_regnos_for_alter_reg (int *pseudo_regnos, int n,
4164 machine_mode *reg_max_ref_mode)
4166 int max_regno = max_reg_num ();
4167 int i, regno, num, slot_num;
4168 ira_allocno_t allocno, a;
4169 ira_allocno_iterator ai;
4170 ira_allocno_t *spilled_coalesced_allocnos;
4172 ira_assert (! ira_use_lra_p);
4174 /* Set up allocnos can be coalesced. */
4175 coloring_allocno_bitmap = ira_allocate_bitmap ();
4176 for (i = 0; i < n; i++)
4178 regno = pseudo_regnos[i];
4179 allocno = ira_regno_allocno_map[regno];
4180 if (allocno != NULL)
4181 bitmap_set_bit (coloring_allocno_bitmap, ALLOCNO_NUM (allocno));
4183 allocno_coalesced_p = false;
4184 processed_coalesced_allocno_bitmap = ira_allocate_bitmap ();
4185 allocno_coalesce_data
4186 = (coalesce_data_t) ira_allocate (sizeof (struct coalesce_data)
4187 * ira_allocnos_num);
4188 /* Initialize coalesce data for allocnos. */
4189 FOR_EACH_ALLOCNO (a, ai)
4191 ALLOCNO_ADD_DATA (a) = allocno_coalesce_data + ALLOCNO_NUM (a);
4192 ALLOCNO_COALESCE_DATA (a)->first = a;
4193 ALLOCNO_COALESCE_DATA (a)->next = a;
4195 coalesce_allocnos ();
4196 ira_free_bitmap (coloring_allocno_bitmap);
4197 regno_coalesced_allocno_cost
4198 = (int *) ira_allocate (max_regno * sizeof (int));
4199 regno_coalesced_allocno_num
4200 = (int *) ira_allocate (max_regno * sizeof (int));
4201 memset (regno_coalesced_allocno_num, 0, max_regno * sizeof (int));
4202 setup_coalesced_allocno_costs_and_nums (pseudo_regnos, n);
4203 /* Sort regnos according frequencies of the corresponding coalesced
4204 allocno sets. */
4205 qsort (pseudo_regnos, n, sizeof (int), coalesced_pseudo_reg_freq_compare);
4206 spilled_coalesced_allocnos
4207 = (ira_allocno_t *) ira_allocate (ira_allocnos_num
4208 * sizeof (ira_allocno_t));
4209 /* Collect allocnos representing the spilled coalesced allocno
4210 sets. */
4211 num = collect_spilled_coalesced_allocnos (pseudo_regnos, n,
4212 spilled_coalesced_allocnos);
4213 if (flag_ira_share_spill_slots
4214 && coalesce_spill_slots (spilled_coalesced_allocnos, num))
4216 setup_coalesced_allocno_costs_and_nums (pseudo_regnos, n);
4217 qsort (pseudo_regnos, n, sizeof (int),
4218 coalesced_pseudo_reg_freq_compare);
4219 num = collect_spilled_coalesced_allocnos (pseudo_regnos, n,
4220 spilled_coalesced_allocnos);
4222 ira_free_bitmap (processed_coalesced_allocno_bitmap);
4223 allocno_coalesced_p = false;
4224 /* Assign stack slot numbers to spilled allocno sets, use smaller
4225 numbers for most frequently used coalesced allocnos. -1 is
4226 reserved for dynamic search of stack slots for pseudos spilled by
4227 the reload. */
4228 slot_num = 1;
4229 for (i = 0; i < num; i++)
4231 allocno = spilled_coalesced_allocnos[i];
4232 if (ALLOCNO_COALESCE_DATA (allocno)->first != allocno
4233 || ALLOCNO_HARD_REGNO (allocno) >= 0
4234 || ira_equiv_no_lvalue_p (ALLOCNO_REGNO (allocno)))
4235 continue;
4236 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
4237 fprintf (ira_dump_file, " Slot %d (freq,size):", slot_num);
4238 slot_num++;
4239 for (a = ALLOCNO_COALESCE_DATA (allocno)->next;;
4240 a = ALLOCNO_COALESCE_DATA (a)->next)
4242 ira_assert (ALLOCNO_HARD_REGNO (a) < 0);
4243 ALLOCNO_HARD_REGNO (a) = -slot_num;
4244 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
4246 machine_mode mode = wider_subreg_mode
4247 (PSEUDO_REGNO_MODE (ALLOCNO_REGNO (a)),
4248 reg_max_ref_mode[ALLOCNO_REGNO (a)]);
4249 fprintf (ira_dump_file, " a%dr%d(%d,",
4250 ALLOCNO_NUM (a), ALLOCNO_REGNO (a), ALLOCNO_FREQ (a));
4251 print_dec (GET_MODE_SIZE (mode), ira_dump_file, SIGNED);
4252 fprintf (ira_dump_file, ")\n");
4255 if (a == allocno)
4256 break;
4258 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
4259 fprintf (ira_dump_file, "\n");
4261 ira_spilled_reg_stack_slots_num = slot_num - 1;
4262 ira_free (spilled_coalesced_allocnos);
4263 /* Sort regnos according the slot numbers. */
4264 regno_max_ref_mode = reg_max_ref_mode;
4265 qsort (pseudo_regnos, n, sizeof (int), coalesced_pseudo_reg_slot_compare);
4266 FOR_EACH_ALLOCNO (a, ai)
4267 ALLOCNO_ADD_DATA (a) = NULL;
4268 ira_free (allocno_coalesce_data);
4269 ira_free (regno_coalesced_allocno_num);
4270 ira_free (regno_coalesced_allocno_cost);
4275 /* This page contains code used by the reload pass to improve the
4276 final code. */
4278 /* The function is called from reload to mark changes in the
4279 allocation of REGNO made by the reload. Remember that reg_renumber
4280 reflects the change result. */
4281 void
4282 ira_mark_allocation_change (int regno)
4284 ira_allocno_t a = ira_regno_allocno_map[regno];
4285 int old_hard_regno, hard_regno, cost;
4286 enum reg_class aclass = ALLOCNO_CLASS (a);
4288 ira_assert (a != NULL);
4289 hard_regno = reg_renumber[regno];
4290 if ((old_hard_regno = ALLOCNO_HARD_REGNO (a)) == hard_regno)
4291 return;
4292 if (old_hard_regno < 0)
4293 cost = -ALLOCNO_MEMORY_COST (a);
4294 else
4296 ira_assert (ira_class_hard_reg_index[aclass][old_hard_regno] >= 0);
4297 cost = -(ALLOCNO_HARD_REG_COSTS (a) == NULL
4298 ? ALLOCNO_CLASS_COST (a)
4299 : ALLOCNO_HARD_REG_COSTS (a)
4300 [ira_class_hard_reg_index[aclass][old_hard_regno]]);
4301 update_costs_from_copies (a, false, false);
4303 ira_overall_cost -= cost;
4304 ALLOCNO_HARD_REGNO (a) = hard_regno;
4305 if (hard_regno < 0)
4307 ALLOCNO_HARD_REGNO (a) = -1;
4308 cost += ALLOCNO_MEMORY_COST (a);
4310 else if (ira_class_hard_reg_index[aclass][hard_regno] >= 0)
4312 cost += (ALLOCNO_HARD_REG_COSTS (a) == NULL
4313 ? ALLOCNO_CLASS_COST (a)
4314 : ALLOCNO_HARD_REG_COSTS (a)
4315 [ira_class_hard_reg_index[aclass][hard_regno]]);
4316 update_costs_from_copies (a, true, false);
4318 else
4319 /* Reload changed class of the allocno. */
4320 cost = 0;
4321 ira_overall_cost += cost;
4324 /* This function is called when reload deletes memory-memory move. In
4325 this case we marks that the allocation of the corresponding
4326 allocnos should be not changed in future. Otherwise we risk to get
4327 a wrong code. */
4328 void
4329 ira_mark_memory_move_deletion (int dst_regno, int src_regno)
4331 ira_allocno_t dst = ira_regno_allocno_map[dst_regno];
4332 ira_allocno_t src = ira_regno_allocno_map[src_regno];
4334 ira_assert (dst != NULL && src != NULL
4335 && ALLOCNO_HARD_REGNO (dst) < 0
4336 && ALLOCNO_HARD_REGNO (src) < 0);
4337 ALLOCNO_DONT_REASSIGN_P (dst) = true;
4338 ALLOCNO_DONT_REASSIGN_P (src) = true;
4341 /* Try to assign a hard register (except for FORBIDDEN_REGS) to
4342 allocno A and return TRUE in the case of success. */
4343 static bool
4344 allocno_reload_assign (ira_allocno_t a, HARD_REG_SET forbidden_regs)
4346 int hard_regno;
4347 enum reg_class aclass;
4348 int regno = ALLOCNO_REGNO (a);
4349 HARD_REG_SET saved[2];
4350 int i, n;
4352 n = ALLOCNO_NUM_OBJECTS (a);
4353 for (i = 0; i < n; i++)
4355 ira_object_t obj = ALLOCNO_OBJECT (a, i);
4356 COPY_HARD_REG_SET (saved[i], OBJECT_TOTAL_CONFLICT_HARD_REGS (obj));
4357 IOR_HARD_REG_SET (OBJECT_TOTAL_CONFLICT_HARD_REGS (obj), forbidden_regs);
4358 if (! flag_caller_saves && ALLOCNO_CALLS_CROSSED_NUM (a) != 0)
4359 IOR_HARD_REG_SET (OBJECT_TOTAL_CONFLICT_HARD_REGS (obj),
4360 call_used_reg_set);
4362 ALLOCNO_ASSIGNED_P (a) = false;
4363 aclass = ALLOCNO_CLASS (a);
4364 update_curr_costs (a);
4365 assign_hard_reg (a, true);
4366 hard_regno = ALLOCNO_HARD_REGNO (a);
4367 reg_renumber[regno] = hard_regno;
4368 if (hard_regno < 0)
4369 ALLOCNO_HARD_REGNO (a) = -1;
4370 else
4372 ira_assert (ira_class_hard_reg_index[aclass][hard_regno] >= 0);
4373 ira_overall_cost
4374 -= (ALLOCNO_MEMORY_COST (a)
4375 - (ALLOCNO_HARD_REG_COSTS (a) == NULL
4376 ? ALLOCNO_CLASS_COST (a)
4377 : ALLOCNO_HARD_REG_COSTS (a)[ira_class_hard_reg_index
4378 [aclass][hard_regno]]));
4379 if (ALLOCNO_CALLS_CROSSED_NUM (a) != 0
4380 && ira_hard_reg_set_intersection_p (hard_regno, ALLOCNO_MODE (a),
4381 call_used_reg_set))
4383 ira_assert (flag_caller_saves);
4384 caller_save_needed = 1;
4388 /* If we found a hard register, modify the RTL for the pseudo
4389 register to show the hard register, and mark the pseudo register
4390 live. */
4391 if (reg_renumber[regno] >= 0)
4393 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
4394 fprintf (ira_dump_file, ": reassign to %d\n", reg_renumber[regno]);
4395 SET_REGNO (regno_reg_rtx[regno], reg_renumber[regno]);
4396 mark_home_live (regno);
4398 else if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
4399 fprintf (ira_dump_file, "\n");
4400 for (i = 0; i < n; i++)
4402 ira_object_t obj = ALLOCNO_OBJECT (a, i);
4403 COPY_HARD_REG_SET (OBJECT_TOTAL_CONFLICT_HARD_REGS (obj), saved[i]);
4405 return reg_renumber[regno] >= 0;
4408 /* Sort pseudos according their usage frequencies (putting most
4409 frequently ones first). */
4410 static int
4411 pseudo_reg_compare (const void *v1p, const void *v2p)
4413 int regno1 = *(const int *) v1p;
4414 int regno2 = *(const int *) v2p;
4415 int diff;
4417 if ((diff = REG_FREQ (regno2) - REG_FREQ (regno1)) != 0)
4418 return diff;
4419 return regno1 - regno2;
4422 /* Try to allocate hard registers to SPILLED_PSEUDO_REGS (there are
4423 NUM of them) or spilled pseudos conflicting with pseudos in
4424 SPILLED_PSEUDO_REGS. Return TRUE and update SPILLED, if the
4425 allocation has been changed. The function doesn't use
4426 BAD_SPILL_REGS and hard registers in PSEUDO_FORBIDDEN_REGS and
4427 PSEUDO_PREVIOUS_REGS for the corresponding pseudos. The function
4428 is called by the reload pass at the end of each reload
4429 iteration. */
4430 bool
4431 ira_reassign_pseudos (int *spilled_pseudo_regs, int num,
4432 HARD_REG_SET bad_spill_regs,
4433 HARD_REG_SET *pseudo_forbidden_regs,
4434 HARD_REG_SET *pseudo_previous_regs,
4435 bitmap spilled)
4437 int i, n, regno;
4438 bool changed_p;
4439 ira_allocno_t a;
4440 HARD_REG_SET forbidden_regs;
4441 bitmap temp = BITMAP_ALLOC (NULL);
4443 /* Add pseudos which conflict with pseudos already in
4444 SPILLED_PSEUDO_REGS to SPILLED_PSEUDO_REGS. This is preferable
4445 to allocating in two steps as some of the conflicts might have
4446 a higher priority than the pseudos passed in SPILLED_PSEUDO_REGS. */
4447 for (i = 0; i < num; i++)
4448 bitmap_set_bit (temp, spilled_pseudo_regs[i]);
4450 for (i = 0, n = num; i < n; i++)
4452 int nr, j;
4453 int regno = spilled_pseudo_regs[i];
4454 bitmap_set_bit (temp, regno);
4456 a = ira_regno_allocno_map[regno];
4457 nr = ALLOCNO_NUM_OBJECTS (a);
4458 for (j = 0; j < nr; j++)
4460 ira_object_t conflict_obj;
4461 ira_object_t obj = ALLOCNO_OBJECT (a, j);
4462 ira_object_conflict_iterator oci;
4464 FOR_EACH_OBJECT_CONFLICT (obj, conflict_obj, oci)
4466 ira_allocno_t conflict_a = OBJECT_ALLOCNO (conflict_obj);
4467 if (ALLOCNO_HARD_REGNO (conflict_a) < 0
4468 && ! ALLOCNO_DONT_REASSIGN_P (conflict_a)
4469 && bitmap_set_bit (temp, ALLOCNO_REGNO (conflict_a)))
4471 spilled_pseudo_regs[num++] = ALLOCNO_REGNO (conflict_a);
4472 /* ?!? This seems wrong. */
4473 bitmap_set_bit (consideration_allocno_bitmap,
4474 ALLOCNO_NUM (conflict_a));
4480 if (num > 1)
4481 qsort (spilled_pseudo_regs, num, sizeof (int), pseudo_reg_compare);
4482 changed_p = false;
4483 /* Try to assign hard registers to pseudos from
4484 SPILLED_PSEUDO_REGS. */
4485 for (i = 0; i < num; i++)
4487 regno = spilled_pseudo_regs[i];
4488 COPY_HARD_REG_SET (forbidden_regs, bad_spill_regs);
4489 IOR_HARD_REG_SET (forbidden_regs, pseudo_forbidden_regs[regno]);
4490 IOR_HARD_REG_SET (forbidden_regs, pseudo_previous_regs[regno]);
4491 gcc_assert (reg_renumber[regno] < 0);
4492 a = ira_regno_allocno_map[regno];
4493 ira_mark_allocation_change (regno);
4494 ira_assert (reg_renumber[regno] < 0);
4495 if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
4496 fprintf (ira_dump_file,
4497 " Try Assign %d(a%d), cost=%d", regno, ALLOCNO_NUM (a),
4498 ALLOCNO_MEMORY_COST (a)
4499 - ALLOCNO_CLASS_COST (a));
4500 allocno_reload_assign (a, forbidden_regs);
4501 if (reg_renumber[regno] >= 0)
4503 CLEAR_REGNO_REG_SET (spilled, regno);
4504 changed_p = true;
4507 BITMAP_FREE (temp);
4508 return changed_p;
4511 /* The function is called by reload and returns already allocated
4512 stack slot (if any) for REGNO with given INHERENT_SIZE and
4513 TOTAL_SIZE. In the case of failure to find a slot which can be
4514 used for REGNO, the function returns NULL. */
4516 ira_reuse_stack_slot (int regno, poly_uint64 inherent_size,
4517 poly_uint64 total_size)
4519 unsigned int i;
4520 int slot_num, best_slot_num;
4521 int cost, best_cost;
4522 ira_copy_t cp, next_cp;
4523 ira_allocno_t another_allocno, allocno = ira_regno_allocno_map[regno];
4524 rtx x;
4525 bitmap_iterator bi;
4526 struct ira_spilled_reg_stack_slot *slot = NULL;
4528 ira_assert (! ira_use_lra_p);
4530 ira_assert (known_eq (inherent_size, PSEUDO_REGNO_BYTES (regno))
4531 && known_le (inherent_size, total_size)
4532 && ALLOCNO_HARD_REGNO (allocno) < 0);
4533 if (! flag_ira_share_spill_slots)
4534 return NULL_RTX;
4535 slot_num = -ALLOCNO_HARD_REGNO (allocno) - 2;
4536 if (slot_num != -1)
4538 slot = &ira_spilled_reg_stack_slots[slot_num];
4539 x = slot->mem;
4541 else
4543 best_cost = best_slot_num = -1;
4544 x = NULL_RTX;
4545 /* It means that the pseudo was spilled in the reload pass, try
4546 to reuse a slot. */
4547 for (slot_num = 0;
4548 slot_num < ira_spilled_reg_stack_slots_num;
4549 slot_num++)
4551 slot = &ira_spilled_reg_stack_slots[slot_num];
4552 if (slot->mem == NULL_RTX)
4553 continue;
4554 if (maybe_lt (slot->width, total_size)
4555 || maybe_lt (GET_MODE_SIZE (GET_MODE (slot->mem)), inherent_size))
4556 continue;
4558 EXECUTE_IF_SET_IN_BITMAP (&slot->spilled_regs,
4559 FIRST_PSEUDO_REGISTER, i, bi)
4561 another_allocno = ira_regno_allocno_map[i];
4562 if (allocnos_conflict_by_live_ranges_p (allocno,
4563 another_allocno))
4564 goto cont;
4566 for (cost = 0, cp = ALLOCNO_COPIES (allocno);
4567 cp != NULL;
4568 cp = next_cp)
4570 if (cp->first == allocno)
4572 next_cp = cp->next_first_allocno_copy;
4573 another_allocno = cp->second;
4575 else if (cp->second == allocno)
4577 next_cp = cp->next_second_allocno_copy;
4578 another_allocno = cp->first;
4580 else
4581 gcc_unreachable ();
4582 if (cp->insn == NULL_RTX)
4583 continue;
4584 if (bitmap_bit_p (&slot->spilled_regs,
4585 ALLOCNO_REGNO (another_allocno)))
4586 cost += cp->freq;
4588 if (cost > best_cost)
4590 best_cost = cost;
4591 best_slot_num = slot_num;
4593 cont:
4596 if (best_cost >= 0)
4598 slot_num = best_slot_num;
4599 slot = &ira_spilled_reg_stack_slots[slot_num];
4600 SET_REGNO_REG_SET (&slot->spilled_regs, regno);
4601 x = slot->mem;
4602 ALLOCNO_HARD_REGNO (allocno) = -slot_num - 2;
4605 if (x != NULL_RTX)
4607 ira_assert (known_ge (slot->width, total_size));
4608 #ifdef ENABLE_IRA_CHECKING
4609 EXECUTE_IF_SET_IN_BITMAP (&slot->spilled_regs,
4610 FIRST_PSEUDO_REGISTER, i, bi)
4612 ira_assert (! conflict_by_live_ranges_p (regno, i));
4614 #endif
4615 SET_REGNO_REG_SET (&slot->spilled_regs, regno);
4616 if (internal_flag_ira_verbose > 3 && ira_dump_file)
4618 fprintf (ira_dump_file, " Assigning %d(freq=%d) slot %d of",
4619 regno, REG_FREQ (regno), slot_num);
4620 EXECUTE_IF_SET_IN_BITMAP (&slot->spilled_regs,
4621 FIRST_PSEUDO_REGISTER, i, bi)
4623 if ((unsigned) regno != i)
4624 fprintf (ira_dump_file, " %d", i);
4626 fprintf (ira_dump_file, "\n");
4629 return x;
4632 /* This is called by reload every time a new stack slot X with
4633 TOTAL_SIZE was allocated for REGNO. We store this info for
4634 subsequent ira_reuse_stack_slot calls. */
4635 void
4636 ira_mark_new_stack_slot (rtx x, int regno, poly_uint64 total_size)
4638 struct ira_spilled_reg_stack_slot *slot;
4639 int slot_num;
4640 ira_allocno_t allocno;
4642 ira_assert (! ira_use_lra_p);
4644 ira_assert (known_le (PSEUDO_REGNO_BYTES (regno), total_size));
4645 allocno = ira_regno_allocno_map[regno];
4646 slot_num = -ALLOCNO_HARD_REGNO (allocno) - 2;
4647 if (slot_num == -1)
4649 slot_num = ira_spilled_reg_stack_slots_num++;
4650 ALLOCNO_HARD_REGNO (allocno) = -slot_num - 2;
4652 slot = &ira_spilled_reg_stack_slots[slot_num];
4653 INIT_REG_SET (&slot->spilled_regs);
4654 SET_REGNO_REG_SET (&slot->spilled_regs, regno);
4655 slot->mem = x;
4656 slot->width = total_size;
4657 if (internal_flag_ira_verbose > 3 && ira_dump_file)
4658 fprintf (ira_dump_file, " Assigning %d(freq=%d) a new slot %d\n",
4659 regno, REG_FREQ (regno), slot_num);
4663 /* Return spill cost for pseudo-registers whose numbers are in array
4664 REGNOS (with a negative number as an end marker) for reload with
4665 given IN and OUT for INSN. Return also number points (through
4666 EXCESS_PRESSURE_LIVE_LENGTH) where the pseudo-register lives and
4667 the register pressure is high, number of references of the
4668 pseudo-registers (through NREFS), number of callee-clobbered
4669 hard-registers occupied by the pseudo-registers (through
4670 CALL_USED_COUNT), and the first hard regno occupied by the
4671 pseudo-registers (through FIRST_HARD_REGNO). */
4672 static int
4673 calculate_spill_cost (int *regnos, rtx in, rtx out, rtx_insn *insn,
4674 int *excess_pressure_live_length,
4675 int *nrefs, int *call_used_count, int *first_hard_regno)
4677 int i, cost, regno, hard_regno, j, count, saved_cost, nregs;
4678 bool in_p, out_p;
4679 int length;
4680 ira_allocno_t a;
4682 *nrefs = 0;
4683 for (length = count = cost = i = 0;; i++)
4685 regno = regnos[i];
4686 if (regno < 0)
4687 break;
4688 *nrefs += REG_N_REFS (regno);
4689 hard_regno = reg_renumber[regno];
4690 ira_assert (hard_regno >= 0);
4691 a = ira_regno_allocno_map[regno];
4692 length += ALLOCNO_EXCESS_PRESSURE_POINTS_NUM (a) / ALLOCNO_NUM_OBJECTS (a);
4693 cost += ALLOCNO_MEMORY_COST (a) - ALLOCNO_CLASS_COST (a);
4694 nregs = hard_regno_nregs (hard_regno, ALLOCNO_MODE (a));
4695 for (j = 0; j < nregs; j++)
4696 if (! TEST_HARD_REG_BIT (call_used_reg_set, hard_regno + j))
4697 break;
4698 if (j == nregs)
4699 count++;
4700 in_p = in && REG_P (in) && (int) REGNO (in) == hard_regno;
4701 out_p = out && REG_P (out) && (int) REGNO (out) == hard_regno;
4702 if ((in_p || out_p)
4703 && find_regno_note (insn, REG_DEAD, hard_regno) != NULL_RTX)
4705 saved_cost = 0;
4706 if (in_p)
4707 saved_cost += ira_memory_move_cost
4708 [ALLOCNO_MODE (a)][ALLOCNO_CLASS (a)][1];
4709 if (out_p)
4710 saved_cost
4711 += ira_memory_move_cost
4712 [ALLOCNO_MODE (a)][ALLOCNO_CLASS (a)][0];
4713 cost -= REG_FREQ_FROM_BB (BLOCK_FOR_INSN (insn)) * saved_cost;
4716 *excess_pressure_live_length = length;
4717 *call_used_count = count;
4718 hard_regno = -1;
4719 if (regnos[0] >= 0)
4721 hard_regno = reg_renumber[regnos[0]];
4723 *first_hard_regno = hard_regno;
4724 return cost;
4727 /* Return TRUE if spilling pseudo-registers whose numbers are in array
4728 REGNOS is better than spilling pseudo-registers with numbers in
4729 OTHER_REGNOS for reload with given IN and OUT for INSN. The
4730 function used by the reload pass to make better register spilling
4731 decisions. */
4732 bool
4733 ira_better_spill_reload_regno_p (int *regnos, int *other_regnos,
4734 rtx in, rtx out, rtx_insn *insn)
4736 int cost, other_cost;
4737 int length, other_length;
4738 int nrefs, other_nrefs;
4739 int call_used_count, other_call_used_count;
4740 int hard_regno, other_hard_regno;
4742 cost = calculate_spill_cost (regnos, in, out, insn,
4743 &length, &nrefs, &call_used_count, &hard_regno);
4744 other_cost = calculate_spill_cost (other_regnos, in, out, insn,
4745 &other_length, &other_nrefs,
4746 &other_call_used_count,
4747 &other_hard_regno);
4748 if (nrefs == 0 && other_nrefs != 0)
4749 return true;
4750 if (nrefs != 0 && other_nrefs == 0)
4751 return false;
4752 if (cost != other_cost)
4753 return cost < other_cost;
4754 if (length != other_length)
4755 return length > other_length;
4756 #ifdef REG_ALLOC_ORDER
4757 if (hard_regno >= 0 && other_hard_regno >= 0)
4758 return (inv_reg_alloc_order[hard_regno]
4759 < inv_reg_alloc_order[other_hard_regno]);
4760 #else
4761 if (call_used_count != other_call_used_count)
4762 return call_used_count > other_call_used_count;
4763 #endif
4764 return false;
4769 /* Allocate and initialize data necessary for assign_hard_reg. */
4770 void
4771 ira_initiate_assign (void)
4773 sorted_allocnos
4774 = (ira_allocno_t *) ira_allocate (sizeof (ira_allocno_t)
4775 * ira_allocnos_num);
4776 consideration_allocno_bitmap = ira_allocate_bitmap ();
4777 initiate_cost_update ();
4778 allocno_priorities = (int *) ira_allocate (sizeof (int) * ira_allocnos_num);
4779 sorted_copies = (ira_copy_t *) ira_allocate (ira_copies_num
4780 * sizeof (ira_copy_t));
4783 /* Deallocate data used by assign_hard_reg. */
4784 void
4785 ira_finish_assign (void)
4787 ira_free (sorted_allocnos);
4788 ira_free_bitmap (consideration_allocno_bitmap);
4789 finish_cost_update ();
4790 ira_free (allocno_priorities);
4791 ira_free (sorted_copies);
4796 /* Entry function doing color-based register allocation. */
4797 static void
4798 color (void)
4800 allocno_stack_vec.create (ira_allocnos_num);
4801 memset (allocated_hardreg_p, 0, sizeof (allocated_hardreg_p));
4802 ira_initiate_assign ();
4803 do_coloring ();
4804 ira_finish_assign ();
4805 allocno_stack_vec.release ();
4806 move_spill_restore ();
4811 /* This page contains a simple register allocator without usage of
4812 allocno conflicts. This is used for fast allocation for -O0. */
4814 /* Do register allocation by not using allocno conflicts. It uses
4815 only allocno live ranges. The algorithm is close to Chow's
4816 priority coloring. */
4817 static void
4818 fast_allocation (void)
4820 int i, j, k, num, class_size, hard_regno;
4821 #ifdef STACK_REGS
4822 bool no_stack_reg_p;
4823 #endif
4824 enum reg_class aclass;
4825 machine_mode mode;
4826 ira_allocno_t a;
4827 ira_allocno_iterator ai;
4828 live_range_t r;
4829 HARD_REG_SET conflict_hard_regs, *used_hard_regs;
4831 sorted_allocnos = (ira_allocno_t *) ira_allocate (sizeof (ira_allocno_t)
4832 * ira_allocnos_num);
4833 num = 0;
4834 FOR_EACH_ALLOCNO (a, ai)
4835 sorted_allocnos[num++] = a;
4836 allocno_priorities = (int *) ira_allocate (sizeof (int) * ira_allocnos_num);
4837 setup_allocno_priorities (sorted_allocnos, num);
4838 used_hard_regs = (HARD_REG_SET *) ira_allocate (sizeof (HARD_REG_SET)
4839 * ira_max_point);
4840 for (i = 0; i < ira_max_point; i++)
4841 CLEAR_HARD_REG_SET (used_hard_regs[i]);
4842 qsort (sorted_allocnos, num, sizeof (ira_allocno_t),
4843 allocno_priority_compare_func);
4844 for (i = 0; i < num; i++)
4846 int nr, l;
4848 a = sorted_allocnos[i];
4849 nr = ALLOCNO_NUM_OBJECTS (a);
4850 CLEAR_HARD_REG_SET (conflict_hard_regs);
4851 for (l = 0; l < nr; l++)
4853 ira_object_t obj = ALLOCNO_OBJECT (a, l);
4854 IOR_HARD_REG_SET (conflict_hard_regs,
4855 OBJECT_CONFLICT_HARD_REGS (obj));
4856 for (r = OBJECT_LIVE_RANGES (obj); r != NULL; r = r->next)
4857 for (j = r->start; j <= r->finish; j++)
4858 IOR_HARD_REG_SET (conflict_hard_regs, used_hard_regs[j]);
4860 aclass = ALLOCNO_CLASS (a);
4861 ALLOCNO_ASSIGNED_P (a) = true;
4862 ALLOCNO_HARD_REGNO (a) = -1;
4863 if (hard_reg_set_subset_p (reg_class_contents[aclass],
4864 conflict_hard_regs))
4865 continue;
4866 mode = ALLOCNO_MODE (a);
4867 #ifdef STACK_REGS
4868 no_stack_reg_p = ALLOCNO_NO_STACK_REG_P (a);
4869 #endif
4870 class_size = ira_class_hard_regs_num[aclass];
4871 for (j = 0; j < class_size; j++)
4873 hard_regno = ira_class_hard_regs[aclass][j];
4874 #ifdef STACK_REGS
4875 if (no_stack_reg_p && FIRST_STACK_REG <= hard_regno
4876 && hard_regno <= LAST_STACK_REG)
4877 continue;
4878 #endif
4879 if (ira_hard_reg_set_intersection_p (hard_regno, mode, conflict_hard_regs)
4880 || (TEST_HARD_REG_BIT
4881 (ira_prohibited_class_mode_regs[aclass][mode], hard_regno)))
4882 continue;
4883 ALLOCNO_HARD_REGNO (a) = hard_regno;
4884 for (l = 0; l < nr; l++)
4886 ira_object_t obj = ALLOCNO_OBJECT (a, l);
4887 for (r = OBJECT_LIVE_RANGES (obj); r != NULL; r = r->next)
4888 for (k = r->start; k <= r->finish; k++)
4889 IOR_HARD_REG_SET (used_hard_regs[k],
4890 ira_reg_mode_hard_regset[hard_regno][mode]);
4892 break;
4895 ira_free (sorted_allocnos);
4896 ira_free (used_hard_regs);
4897 ira_free (allocno_priorities);
4898 if (internal_flag_ira_verbose > 1 && ira_dump_file != NULL)
4899 ira_print_disposition (ira_dump_file);
4904 /* Entry function doing coloring. */
4905 void
4906 ira_color (void)
4908 ira_allocno_t a;
4909 ira_allocno_iterator ai;
4911 /* Setup updated costs. */
4912 FOR_EACH_ALLOCNO (a, ai)
4914 ALLOCNO_UPDATED_MEMORY_COST (a) = ALLOCNO_MEMORY_COST (a);
4915 ALLOCNO_UPDATED_CLASS_COST (a) = ALLOCNO_CLASS_COST (a);
4917 if (ira_conflicts_p)
4918 color ();
4919 else
4920 fast_allocation ();