2015-01-14 Sandra Loosemore <sandra@codesourcery.com>
[official-gcc.git] / gcc / lra-assigns.c
blobc8a2fb138a2d6dd46d063ebc327c0ea85178a452
1 /* Assign reload pseudos.
2 Copyright (C) 2010-2015 Free Software Foundation, Inc.
3 Contributed by Vladimir Makarov <vmakarov@redhat.com>.
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify it under
8 the terms of the GNU General Public License as published by the Free
9 Software Foundation; either version 3, or (at your option) any later
10 version.
12 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
15 for more details.
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
22 /* This file's main objective is to assign hard registers to reload
23 pseudos. It also tries to allocate hard registers to other
24 pseudos, but at a lower priority than the reload pseudos. The pass
25 does not transform the RTL.
27 We must allocate a hard register to every reload pseudo. We try to
28 increase the chances of finding a viable allocation by assigning
29 the pseudos in order of fewest available hard registers first. If
30 we still fail to find a hard register, we spill other (non-reload)
31 pseudos in order to make room.
33 find_hard_regno_for finds hard registers for allocation without
34 spilling. spill_for does the same with spilling. Both functions
35 use a cost model to determine the most profitable choice of hard
36 and spill registers.
38 Once we have finished allocating reload pseudos, we also try to
39 assign registers to other (non-reload) pseudos. This is useful if
40 hard registers were freed up by the spilling just described.
42 We try to assign hard registers by collecting pseudos into threads.
43 These threads contain reload and inheritance pseudos that are
44 connected by copies (move insns). Doing this improves the chances
45 of pseudos in the thread getting the same hard register and, as a
46 result, of allowing some move insns to be deleted.
48 When we assign a hard register to a pseudo, we decrease the cost of
49 using the same hard register for pseudos that are connected by
50 copies.
52 If two hard registers have the same frequency-derived cost, we
53 prefer hard registers with higher priorities. The mapping of
54 registers to priorities is controlled by the register_priority
55 target hook. For example, x86-64 has a few register priorities:
56 hard registers with and without REX prefixes have different
57 priorities. This permits us to generate smaller code as insns
58 without REX prefixes are shorter.
60 If a few hard registers are still equally good for the assignment,
61 we choose the least used hard register. It is called leveling and
62 may be profitable for some targets.
64 Only insns with changed allocation pseudos are processed on the
65 next constraint pass.
67 The pseudo live-ranges are used to find conflicting pseudos.
69 For understanding the code, it is important to keep in mind that
70 inheritance, split, and reload pseudos created since last
71 constraint pass have regno >= lra_constraint_new_regno_start.
72 Inheritance and split pseudos created on any pass are in the
73 corresponding bitmaps. Inheritance and split pseudos since the
74 last constraint pass have also the corresponding non-negative
75 restore_regno. */
77 #include "config.h"
78 #include "system.h"
79 #include "coretypes.h"
80 #include "tm.h"
81 #include "hard-reg-set.h"
82 #include "rtl.h"
83 #include "rtl-error.h"
84 #include "tm_p.h"
85 #include "target.h"
86 #include "insn-config.h"
87 #include "recog.h"
88 #include "output.h"
89 #include "regs.h"
90 #include "hashtab.h"
91 #include "hash-set.h"
92 #include "vec.h"
93 #include "machmode.h"
94 #include "input.h"
95 #include "function.h"
96 #include "symtab.h"
97 #include "expr.h"
98 #include "predict.h"
99 #include "dominance.h"
100 #include "cfg.h"
101 #include "basic-block.h"
102 #include "except.h"
103 #include "df.h"
104 #include "ira.h"
105 #include "sparseset.h"
106 #include "params.h"
107 #include "lra-int.h"
109 /* Current iteration number of the pass and current iteration number
110 of the pass after the latest spill pass when any former reload
111 pseudo was spilled. */
112 int lra_assignment_iter;
113 int lra_assignment_iter_after_spill;
115 /* Flag of spilling former reload pseudos on this pass. */
116 static bool former_reload_pseudo_spill_p;
118 /* Array containing corresponding values of function
119 lra_get_allocno_class. It is used to speed up the code. */
120 static enum reg_class *regno_allocno_class_array;
122 /* Information about the thread to which a pseudo belongs. Threads are
123 a set of connected reload and inheritance pseudos with the same set of
124 available hard registers. Lone registers belong to their own threads. */
125 struct regno_assign_info
127 /* First/next pseudo of the same thread. */
128 int first, next;
129 /* Frequency of the thread (execution frequency of only reload
130 pseudos in the thread when the thread contains a reload pseudo).
131 Defined only for the first thread pseudo. */
132 int freq;
135 /* Map regno to the corresponding regno assignment info. */
136 static struct regno_assign_info *regno_assign_info;
138 /* All inherited, subreg or optional pseudos created before last spill
139 sub-pass. Such pseudos are permitted to get memory instead of hard
140 regs. */
141 static bitmap_head non_reload_pseudos;
143 /* Process a pseudo copy with execution frequency COPY_FREQ connecting
144 REGNO1 and REGNO2 to form threads. */
145 static void
146 process_copy_to_form_thread (int regno1, int regno2, int copy_freq)
148 int last, regno1_first, regno2_first;
150 lra_assert (regno1 >= lra_constraint_new_regno_start
151 && regno2 >= lra_constraint_new_regno_start);
152 regno1_first = regno_assign_info[regno1].first;
153 regno2_first = regno_assign_info[regno2].first;
154 if (regno1_first != regno2_first)
156 for (last = regno2_first;
157 regno_assign_info[last].next >= 0;
158 last = regno_assign_info[last].next)
159 regno_assign_info[last].first = regno1_first;
160 regno_assign_info[last].first = regno1_first;
161 regno_assign_info[last].next = regno_assign_info[regno1_first].next;
162 regno_assign_info[regno1_first].next = regno2_first;
163 regno_assign_info[regno1_first].freq
164 += regno_assign_info[regno2_first].freq;
166 regno_assign_info[regno1_first].freq -= 2 * copy_freq;
167 lra_assert (regno_assign_info[regno1_first].freq >= 0);
170 /* Initialize REGNO_ASSIGN_INFO and form threads. */
171 static void
172 init_regno_assign_info (void)
174 int i, regno1, regno2, max_regno = max_reg_num ();
175 lra_copy_t cp;
177 regno_assign_info = XNEWVEC (struct regno_assign_info, max_regno);
178 for (i = FIRST_PSEUDO_REGISTER; i < max_regno; i++)
180 regno_assign_info[i].first = i;
181 regno_assign_info[i].next = -1;
182 regno_assign_info[i].freq = lra_reg_info[i].freq;
184 /* Form the threads. */
185 for (i = 0; (cp = lra_get_copy (i)) != NULL; i++)
186 if ((regno1 = cp->regno1) >= lra_constraint_new_regno_start
187 && (regno2 = cp->regno2) >= lra_constraint_new_regno_start
188 && reg_renumber[regno1] < 0 && lra_reg_info[regno1].nrefs != 0
189 && reg_renumber[regno2] < 0 && lra_reg_info[regno2].nrefs != 0
190 && (ira_class_hard_regs_num[regno_allocno_class_array[regno1]]
191 == ira_class_hard_regs_num[regno_allocno_class_array[regno2]]))
192 process_copy_to_form_thread (regno1, regno2, cp->freq);
195 /* Free REGNO_ASSIGN_INFO. */
196 static void
197 finish_regno_assign_info (void)
199 free (regno_assign_info);
202 /* The function is used to sort *reload* and *inheritance* pseudos to
203 try to assign them hard registers. We put pseudos from the same
204 thread always nearby. */
205 static int
206 reload_pseudo_compare_func (const void *v1p, const void *v2p)
208 int r1 = *(const int *) v1p, r2 = *(const int *) v2p;
209 enum reg_class cl1 = regno_allocno_class_array[r1];
210 enum reg_class cl2 = regno_allocno_class_array[r2];
211 int diff;
213 lra_assert (r1 >= lra_constraint_new_regno_start
214 && r2 >= lra_constraint_new_regno_start);
216 /* Prefer to assign reload registers with smaller classes first to
217 guarantee assignment to all reload registers. */
218 if ((diff = (ira_class_hard_regs_num[cl1]
219 - ira_class_hard_regs_num[cl2])) != 0)
220 return diff;
221 if ((diff
222 = (ira_reg_class_max_nregs[cl2][lra_reg_info[r2].biggest_mode]
223 - ira_reg_class_max_nregs[cl1][lra_reg_info[r1].biggest_mode])) != 0
224 /* The code below executes rarely as nregs == 1 in most cases.
225 So we should not worry about using faster data structures to
226 check reload pseudos. */
227 && ! bitmap_bit_p (&non_reload_pseudos, r1)
228 && ! bitmap_bit_p (&non_reload_pseudos, r2))
229 return diff;
230 if ((diff = (regno_assign_info[regno_assign_info[r2].first].freq
231 - regno_assign_info[regno_assign_info[r1].first].freq)) != 0)
232 return diff;
233 /* Allocate bigger pseudos first to avoid register file
234 fragmentation. */
235 if ((diff
236 = (ira_reg_class_max_nregs[cl2][lra_reg_info[r2].biggest_mode]
237 - ira_reg_class_max_nregs[cl1][lra_reg_info[r1].biggest_mode])) != 0)
238 return diff;
239 /* Put pseudos from the thread nearby. */
240 if ((diff = regno_assign_info[r1].first - regno_assign_info[r2].first) != 0)
241 return diff;
242 /* If regs are equally good, sort by their numbers, so that the
243 results of qsort leave nothing to chance. */
244 return r1 - r2;
247 /* The function is used to sort *non-reload* pseudos to try to assign
248 them hard registers. The order calculation is simpler than in the
249 previous function and based on the pseudo frequency usage. */
250 static int
251 pseudo_compare_func (const void *v1p, const void *v2p)
253 int r1 = *(const int *) v1p, r2 = *(const int *) v2p;
254 int diff;
256 /* Prefer to assign more frequently used registers first. */
257 if ((diff = lra_reg_info[r2].freq - lra_reg_info[r1].freq) != 0)
258 return diff;
260 /* If regs are equally good, sort by their numbers, so that the
261 results of qsort leave nothing to chance. */
262 return r1 - r2;
265 /* Arrays of size LRA_LIVE_MAX_POINT mapping a program point to the
266 pseudo live ranges with given start point. We insert only live
267 ranges of pseudos interesting for assignment purposes. They are
268 reload pseudos and pseudos assigned to hard registers. */
269 static lra_live_range_t *start_point_ranges;
271 /* Used as a flag that a live range is not inserted in the start point
272 chain. */
273 static struct lra_live_range not_in_chain_mark;
275 /* Create and set up START_POINT_RANGES. */
276 static void
277 create_live_range_start_chains (void)
279 int i, max_regno;
280 lra_live_range_t r;
282 start_point_ranges = XCNEWVEC (lra_live_range_t, lra_live_max_point);
283 max_regno = max_reg_num ();
284 for (i = FIRST_PSEUDO_REGISTER; i < max_regno; i++)
285 if (i >= lra_constraint_new_regno_start || reg_renumber[i] >= 0)
287 for (r = lra_reg_info[i].live_ranges; r != NULL; r = r->next)
289 r->start_next = start_point_ranges[r->start];
290 start_point_ranges[r->start] = r;
293 else
295 for (r = lra_reg_info[i].live_ranges; r != NULL; r = r->next)
296 r->start_next = &not_in_chain_mark;
300 /* Insert live ranges of pseudo REGNO into start chains if they are
301 not there yet. */
302 static void
303 insert_in_live_range_start_chain (int regno)
305 lra_live_range_t r = lra_reg_info[regno].live_ranges;
307 if (r->start_next != &not_in_chain_mark)
308 return;
309 for (; r != NULL; r = r->next)
311 r->start_next = start_point_ranges[r->start];
312 start_point_ranges[r->start] = r;
316 /* Free START_POINT_RANGES. */
317 static void
318 finish_live_range_start_chains (void)
320 gcc_assert (start_point_ranges != NULL);
321 free (start_point_ranges);
322 start_point_ranges = NULL;
325 /* Map: program point -> bitmap of all pseudos living at the point and
326 assigned to hard registers. */
327 static bitmap_head *live_hard_reg_pseudos;
328 static bitmap_obstack live_hard_reg_pseudos_bitmap_obstack;
330 /* reg_renumber corresponding to pseudos marked in
331 live_hard_reg_pseudos. reg_renumber might be not matched to
332 live_hard_reg_pseudos but live_pseudos_reg_renumber always reflects
333 live_hard_reg_pseudos. */
334 static int *live_pseudos_reg_renumber;
336 /* Sparseset used to calculate living hard reg pseudos for some program
337 point range. */
338 static sparseset live_range_hard_reg_pseudos;
340 /* Sparseset used to calculate living reload/inheritance pseudos for
341 some program point range. */
342 static sparseset live_range_reload_inheritance_pseudos;
344 /* Allocate and initialize the data about living pseudos at program
345 points. */
346 static void
347 init_lives (void)
349 int i, max_regno = max_reg_num ();
351 live_range_hard_reg_pseudos = sparseset_alloc (max_regno);
352 live_range_reload_inheritance_pseudos = sparseset_alloc (max_regno);
353 live_hard_reg_pseudos = XNEWVEC (bitmap_head, lra_live_max_point);
354 bitmap_obstack_initialize (&live_hard_reg_pseudos_bitmap_obstack);
355 for (i = 0; i < lra_live_max_point; i++)
356 bitmap_initialize (&live_hard_reg_pseudos[i],
357 &live_hard_reg_pseudos_bitmap_obstack);
358 live_pseudos_reg_renumber = XNEWVEC (int, max_regno);
359 for (i = 0; i < max_regno; i++)
360 live_pseudos_reg_renumber[i] = -1;
363 /* Free the data about living pseudos at program points. */
364 static void
365 finish_lives (void)
367 sparseset_free (live_range_hard_reg_pseudos);
368 sparseset_free (live_range_reload_inheritance_pseudos);
369 free (live_hard_reg_pseudos);
370 bitmap_obstack_release (&live_hard_reg_pseudos_bitmap_obstack);
371 free (live_pseudos_reg_renumber);
374 /* Update the LIVE_HARD_REG_PSEUDOS and LIVE_PSEUDOS_REG_RENUMBER
375 entries for pseudo REGNO. Assume that the register has been
376 spilled if FREE_P, otherwise assume that it has been assigned
377 reg_renumber[REGNO] (if >= 0). We also insert the pseudo live
378 ranges in the start chains when it is assumed to be assigned to a
379 hard register because we use the chains of pseudos assigned to hard
380 registers during allocation. */
381 static void
382 update_lives (int regno, bool free_p)
384 int p;
385 lra_live_range_t r;
387 if (reg_renumber[regno] < 0)
388 return;
389 live_pseudos_reg_renumber[regno] = free_p ? -1 : reg_renumber[regno];
390 for (r = lra_reg_info[regno].live_ranges; r != NULL; r = r->next)
392 for (p = r->start; p <= r->finish; p++)
393 if (free_p)
394 bitmap_clear_bit (&live_hard_reg_pseudos[p], regno);
395 else
397 bitmap_set_bit (&live_hard_reg_pseudos[p], regno);
398 insert_in_live_range_start_chain (regno);
403 /* Sparseset used to calculate reload pseudos conflicting with a given
404 pseudo when we are trying to find a hard register for the given
405 pseudo. */
406 static sparseset conflict_reload_and_inheritance_pseudos;
408 /* Map: program point -> bitmap of all reload and inheritance pseudos
409 living at the point. */
410 static bitmap_head *live_reload_and_inheritance_pseudos;
411 static bitmap_obstack live_reload_and_inheritance_pseudos_bitmap_obstack;
413 /* Allocate and initialize data about living reload pseudos at any
414 given program point. */
415 static void
416 init_live_reload_and_inheritance_pseudos (void)
418 int i, p, max_regno = max_reg_num ();
419 lra_live_range_t r;
421 conflict_reload_and_inheritance_pseudos = sparseset_alloc (max_regno);
422 live_reload_and_inheritance_pseudos = XNEWVEC (bitmap_head, lra_live_max_point);
423 bitmap_obstack_initialize (&live_reload_and_inheritance_pseudos_bitmap_obstack);
424 for (p = 0; p < lra_live_max_point; p++)
425 bitmap_initialize (&live_reload_and_inheritance_pseudos[p],
426 &live_reload_and_inheritance_pseudos_bitmap_obstack);
427 for (i = lra_constraint_new_regno_start; i < max_regno; i++)
429 for (r = lra_reg_info[i].live_ranges; r != NULL; r = r->next)
430 for (p = r->start; p <= r->finish; p++)
431 bitmap_set_bit (&live_reload_and_inheritance_pseudos[p], i);
435 /* Finalize data about living reload pseudos at any given program
436 point. */
437 static void
438 finish_live_reload_and_inheritance_pseudos (void)
440 sparseset_free (conflict_reload_and_inheritance_pseudos);
441 free (live_reload_and_inheritance_pseudos);
442 bitmap_obstack_release (&live_reload_and_inheritance_pseudos_bitmap_obstack);
445 /* The value used to check that cost of given hard reg is really
446 defined currently. */
447 static int curr_hard_regno_costs_check = 0;
448 /* Array used to check that cost of the corresponding hard reg (the
449 array element index) is really defined currently. */
450 static int hard_regno_costs_check[FIRST_PSEUDO_REGISTER];
451 /* The current costs of allocation of hard regs. Defined only if the
452 value of the corresponding element of the previous array is equal to
453 CURR_HARD_REGNO_COSTS_CHECK. */
454 static int hard_regno_costs[FIRST_PSEUDO_REGISTER];
456 /* Adjust cost of HARD_REGNO by INCR. Reset the cost first if it is
457 not defined yet. */
458 static inline void
459 adjust_hard_regno_cost (int hard_regno, int incr)
461 if (hard_regno_costs_check[hard_regno] != curr_hard_regno_costs_check)
462 hard_regno_costs[hard_regno] = 0;
463 hard_regno_costs_check[hard_regno] = curr_hard_regno_costs_check;
464 hard_regno_costs[hard_regno] += incr;
467 /* Try to find a free hard register for pseudo REGNO. Return the
468 hard register on success and set *COST to the cost of using
469 that register. (If several registers have equal cost, the one with
470 the highest priority wins.) Return -1 on failure.
472 If FIRST_P, return the first available hard reg ignoring other
473 criteria, e.g. allocation cost. This approach results in less hard
474 reg pool fragmentation and permit to allocate hard regs to reload
475 pseudos in complicated situations where pseudo sizes are different.
477 If TRY_ONLY_HARD_REGNO >= 0, consider only that hard register,
478 otherwise consider all hard registers in REGNO's class. */
479 static int
480 find_hard_regno_for (int regno, int *cost, int try_only_hard_regno,
481 bool first_p)
483 HARD_REG_SET conflict_set;
484 int best_cost = INT_MAX, best_priority = INT_MIN, best_usage = INT_MAX;
485 lra_live_range_t r;
486 int p, i, j, rclass_size, best_hard_regno, priority, hard_regno;
487 int hr, conflict_hr, nregs;
488 machine_mode biggest_mode;
489 unsigned int k, conflict_regno;
490 int offset, val, biggest_nregs, nregs_diff;
491 enum reg_class rclass;
492 bitmap_iterator bi;
493 bool *rclass_intersect_p;
494 HARD_REG_SET impossible_start_hard_regs, available_regs;
496 COPY_HARD_REG_SET (conflict_set, lra_no_alloc_regs);
497 rclass = regno_allocno_class_array[regno];
498 rclass_intersect_p = ira_reg_classes_intersect_p[rclass];
499 curr_hard_regno_costs_check++;
500 sparseset_clear (conflict_reload_and_inheritance_pseudos);
501 sparseset_clear (live_range_hard_reg_pseudos);
502 IOR_HARD_REG_SET (conflict_set, lra_reg_info[regno].conflict_hard_regs);
503 biggest_mode = lra_reg_info[regno].biggest_mode;
504 for (r = lra_reg_info[regno].live_ranges; r != NULL; r = r->next)
506 EXECUTE_IF_SET_IN_BITMAP (&live_hard_reg_pseudos[r->start], 0, k, bi)
507 if (rclass_intersect_p[regno_allocno_class_array[k]])
508 sparseset_set_bit (live_range_hard_reg_pseudos, k);
509 EXECUTE_IF_SET_IN_BITMAP (&live_reload_and_inheritance_pseudos[r->start],
510 0, k, bi)
511 if (lra_reg_info[k].preferred_hard_regno1 >= 0
512 && live_pseudos_reg_renumber[k] < 0
513 && rclass_intersect_p[regno_allocno_class_array[k]])
514 sparseset_set_bit (conflict_reload_and_inheritance_pseudos, k);
515 for (p = r->start + 1; p <= r->finish; p++)
517 lra_live_range_t r2;
519 for (r2 = start_point_ranges[p];
520 r2 != NULL;
521 r2 = r2->start_next)
523 if (r2->regno >= lra_constraint_new_regno_start
524 && lra_reg_info[r2->regno].preferred_hard_regno1 >= 0
525 && live_pseudos_reg_renumber[r2->regno] < 0
526 && rclass_intersect_p[regno_allocno_class_array[r2->regno]])
527 sparseset_set_bit (conflict_reload_and_inheritance_pseudos,
528 r2->regno);
529 if (live_pseudos_reg_renumber[r2->regno] >= 0
530 && rclass_intersect_p[regno_allocno_class_array[r2->regno]])
531 sparseset_set_bit (live_range_hard_reg_pseudos, r2->regno);
535 if ((hard_regno = lra_reg_info[regno].preferred_hard_regno1) >= 0)
537 adjust_hard_regno_cost
538 (hard_regno, -lra_reg_info[regno].preferred_hard_regno_profit1);
539 if ((hard_regno = lra_reg_info[regno].preferred_hard_regno2) >= 0)
540 adjust_hard_regno_cost
541 (hard_regno, -lra_reg_info[regno].preferred_hard_regno_profit2);
543 #ifdef STACK_REGS
544 if (lra_reg_info[regno].no_stack_p)
545 for (i = FIRST_STACK_REG; i <= LAST_STACK_REG; i++)
546 SET_HARD_REG_BIT (conflict_set, i);
547 #endif
548 sparseset_clear_bit (conflict_reload_and_inheritance_pseudos, regno);
549 val = lra_reg_info[regno].val;
550 offset = lra_reg_info[regno].offset;
551 CLEAR_HARD_REG_SET (impossible_start_hard_regs);
552 EXECUTE_IF_SET_IN_SPARSESET (live_range_hard_reg_pseudos, conflict_regno)
553 if (lra_reg_val_equal_p (conflict_regno, val, offset))
555 conflict_hr = live_pseudos_reg_renumber[conflict_regno];
556 nregs = (hard_regno_nregs[conflict_hr]
557 [lra_reg_info[conflict_regno].biggest_mode]);
558 /* Remember about multi-register pseudos. For example, 2 hard
559 register pseudos can start on the same hard register but can
560 not start on HR and HR+1/HR-1. */
561 for (hr = conflict_hr + 1;
562 hr < FIRST_PSEUDO_REGISTER && hr < conflict_hr + nregs;
563 hr++)
564 SET_HARD_REG_BIT (impossible_start_hard_regs, hr);
565 for (hr = conflict_hr - 1;
566 hr >= 0 && hr + hard_regno_nregs[hr][biggest_mode] > conflict_hr;
567 hr--)
568 SET_HARD_REG_BIT (impossible_start_hard_regs, hr);
570 else
572 add_to_hard_reg_set (&conflict_set,
573 lra_reg_info[conflict_regno].biggest_mode,
574 live_pseudos_reg_renumber[conflict_regno]);
575 if (hard_reg_set_subset_p (reg_class_contents[rclass],
576 conflict_set))
577 return -1;
579 EXECUTE_IF_SET_IN_SPARSESET (conflict_reload_and_inheritance_pseudos,
580 conflict_regno)
581 if (!lra_reg_val_equal_p (conflict_regno, val, offset))
583 lra_assert (live_pseudos_reg_renumber[conflict_regno] < 0);
584 if ((hard_regno
585 = lra_reg_info[conflict_regno].preferred_hard_regno1) >= 0)
587 adjust_hard_regno_cost
588 (hard_regno,
589 lra_reg_info[conflict_regno].preferred_hard_regno_profit1);
590 if ((hard_regno
591 = lra_reg_info[conflict_regno].preferred_hard_regno2) >= 0)
592 adjust_hard_regno_cost
593 (hard_regno,
594 lra_reg_info[conflict_regno].preferred_hard_regno_profit2);
597 /* Make sure that all registers in a multi-word pseudo belong to the
598 required class. */
599 IOR_COMPL_HARD_REG_SET (conflict_set, reg_class_contents[rclass]);
600 lra_assert (rclass != NO_REGS);
601 rclass_size = ira_class_hard_regs_num[rclass];
602 best_hard_regno = -1;
603 hard_regno = ira_class_hard_regs[rclass][0];
604 biggest_nregs = hard_regno_nregs[hard_regno][biggest_mode];
605 nregs_diff = (biggest_nregs
606 - hard_regno_nregs[hard_regno][PSEUDO_REGNO_MODE (regno)]);
607 COPY_HARD_REG_SET (available_regs, reg_class_contents[rclass]);
608 AND_COMPL_HARD_REG_SET (available_regs, lra_no_alloc_regs);
609 for (i = 0; i < rclass_size; i++)
611 if (try_only_hard_regno >= 0)
612 hard_regno = try_only_hard_regno;
613 else
614 hard_regno = ira_class_hard_regs[rclass][i];
615 if (! overlaps_hard_reg_set_p (conflict_set,
616 PSEUDO_REGNO_MODE (regno), hard_regno)
617 /* We can not use prohibited_class_mode_regs because it is
618 not defined for all classes. */
619 && HARD_REGNO_MODE_OK (hard_regno, PSEUDO_REGNO_MODE (regno))
620 && ! TEST_HARD_REG_BIT (impossible_start_hard_regs, hard_regno)
621 && (nregs_diff == 0
622 || (WORDS_BIG_ENDIAN
623 ? (hard_regno - nregs_diff >= 0
624 && TEST_HARD_REG_BIT (available_regs,
625 hard_regno - nregs_diff))
626 : TEST_HARD_REG_BIT (available_regs,
627 hard_regno + nregs_diff))))
629 if (hard_regno_costs_check[hard_regno]
630 != curr_hard_regno_costs_check)
632 hard_regno_costs_check[hard_regno] = curr_hard_regno_costs_check;
633 hard_regno_costs[hard_regno] = 0;
635 for (j = 0;
636 j < hard_regno_nregs[hard_regno][PSEUDO_REGNO_MODE (regno)];
637 j++)
638 if (! TEST_HARD_REG_BIT (call_used_reg_set, hard_regno + j)
639 && ! df_regs_ever_live_p (hard_regno + j))
640 /* It needs save restore. */
641 hard_regno_costs[hard_regno]
642 += (2
643 * REG_FREQ_FROM_BB (ENTRY_BLOCK_PTR_FOR_FN (cfun)->next_bb)
644 + 1);
645 priority = targetm.register_priority (hard_regno);
646 if (best_hard_regno < 0 || hard_regno_costs[hard_regno] < best_cost
647 || (hard_regno_costs[hard_regno] == best_cost
648 && (priority > best_priority
649 || (targetm.register_usage_leveling_p ()
650 && priority == best_priority
651 && best_usage > lra_hard_reg_usage[hard_regno]))))
653 best_hard_regno = hard_regno;
654 best_cost = hard_regno_costs[hard_regno];
655 best_priority = priority;
656 best_usage = lra_hard_reg_usage[hard_regno];
659 if (try_only_hard_regno >= 0 || (first_p && best_hard_regno >= 0))
660 break;
662 if (best_hard_regno >= 0)
663 *cost = best_cost - lra_reg_info[regno].freq;
664 return best_hard_regno;
667 /* Current value used for checking elements in
668 update_hard_regno_preference_check. */
669 static int curr_update_hard_regno_preference_check;
670 /* If an element value is equal to the above variable value, then the
671 corresponding regno has been processed for preference
672 propagation. */
673 static int *update_hard_regno_preference_check;
675 /* Update the preference for using HARD_REGNO for pseudos that are
676 connected directly or indirectly with REGNO. Apply divisor DIV
677 to any preference adjustments.
679 The more indirectly a pseudo is connected, the smaller its effect
680 should be. We therefore increase DIV on each "hop". */
681 static void
682 update_hard_regno_preference (int regno, int hard_regno, int div)
684 int another_regno, cost;
685 lra_copy_t cp, next_cp;
687 /* Search depth 5 seems to be enough. */
688 if (div > (1 << 5))
689 return;
690 for (cp = lra_reg_info[regno].copies; cp != NULL; cp = next_cp)
692 if (cp->regno1 == regno)
694 next_cp = cp->regno1_next;
695 another_regno = cp->regno2;
697 else if (cp->regno2 == regno)
699 next_cp = cp->regno2_next;
700 another_regno = cp->regno1;
702 else
703 gcc_unreachable ();
704 if (reg_renumber[another_regno] < 0
705 && (update_hard_regno_preference_check[another_regno]
706 != curr_update_hard_regno_preference_check))
708 update_hard_regno_preference_check[another_regno]
709 = curr_update_hard_regno_preference_check;
710 cost = cp->freq < div ? 1 : cp->freq / div;
711 lra_setup_reload_pseudo_preferenced_hard_reg
712 (another_regno, hard_regno, cost);
713 update_hard_regno_preference (another_regno, hard_regno, div * 2);
718 /* Return prefix title for pseudo REGNO. */
719 static const char *
720 pseudo_prefix_title (int regno)
722 return
723 (regno < lra_constraint_new_regno_start ? ""
724 : bitmap_bit_p (&lra_inheritance_pseudos, regno) ? "inheritance "
725 : bitmap_bit_p (&lra_split_regs, regno) ? "split "
726 : bitmap_bit_p (&lra_optional_reload_pseudos, regno) ? "optional reload "
727 : bitmap_bit_p (&lra_subreg_reload_pseudos, regno) ? "subreg reload "
728 : "reload ");
731 /* Update REG_RENUMBER and other pseudo preferences by assignment of
732 HARD_REGNO to pseudo REGNO and print about it if PRINT_P. */
733 void
734 lra_setup_reg_renumber (int regno, int hard_regno, bool print_p)
736 int i, hr;
738 /* We can not just reassign hard register. */
739 lra_assert (hard_regno < 0 || reg_renumber[regno] < 0);
740 if ((hr = hard_regno) < 0)
741 hr = reg_renumber[regno];
742 reg_renumber[regno] = hard_regno;
743 lra_assert (hr >= 0);
744 for (i = 0; i < hard_regno_nregs[hr][PSEUDO_REGNO_MODE (regno)]; i++)
745 if (hard_regno < 0)
746 lra_hard_reg_usage[hr + i] -= lra_reg_info[regno].freq;
747 else
748 lra_hard_reg_usage[hr + i] += lra_reg_info[regno].freq;
749 if (print_p && lra_dump_file != NULL)
750 fprintf (lra_dump_file, " Assign %d to %sr%d (freq=%d)\n",
751 reg_renumber[regno], pseudo_prefix_title (regno),
752 regno, lra_reg_info[regno].freq);
753 if (hard_regno >= 0)
755 curr_update_hard_regno_preference_check++;
756 update_hard_regno_preference (regno, hard_regno, 1);
760 /* Pseudos which occur in insns containing a particular pseudo. */
761 static bitmap_head insn_conflict_pseudos;
763 /* Bitmaps used to contain spill pseudos for given pseudo hard regno
764 and best spill pseudos for given pseudo (and best hard regno). */
765 static bitmap_head spill_pseudos_bitmap, best_spill_pseudos_bitmap;
767 /* Current pseudo check for validity of elements in
768 TRY_HARD_REG_PSEUDOS. */
769 static int curr_pseudo_check;
770 /* Array used for validity of elements in TRY_HARD_REG_PSEUDOS. */
771 static int try_hard_reg_pseudos_check[FIRST_PSEUDO_REGISTER];
772 /* Pseudos who hold given hard register at the considered points. */
773 static bitmap_head try_hard_reg_pseudos[FIRST_PSEUDO_REGISTER];
775 /* Set up try_hard_reg_pseudos for given program point P and class
776 RCLASS. Those are pseudos living at P and assigned to a hard
777 register of RCLASS. In other words, those are pseudos which can be
778 spilled to assign a hard register of RCLASS to a pseudo living at
779 P. */
780 static void
781 setup_try_hard_regno_pseudos (int p, enum reg_class rclass)
783 int i, hard_regno;
784 machine_mode mode;
785 unsigned int spill_regno;
786 bitmap_iterator bi;
788 /* Find what pseudos could be spilled. */
789 EXECUTE_IF_SET_IN_BITMAP (&live_hard_reg_pseudos[p], 0, spill_regno, bi)
791 mode = PSEUDO_REGNO_MODE (spill_regno);
792 hard_regno = live_pseudos_reg_renumber[spill_regno];
793 if (overlaps_hard_reg_set_p (reg_class_contents[rclass],
794 mode, hard_regno))
796 for (i = hard_regno_nregs[hard_regno][mode] - 1; i >= 0; i--)
798 if (try_hard_reg_pseudos_check[hard_regno + i]
799 != curr_pseudo_check)
801 try_hard_reg_pseudos_check[hard_regno + i]
802 = curr_pseudo_check;
803 bitmap_clear (&try_hard_reg_pseudos[hard_regno + i]);
805 bitmap_set_bit (&try_hard_reg_pseudos[hard_regno + i],
806 spill_regno);
812 /* Assign temporarily HARD_REGNO to pseudo REGNO. Temporary
813 assignment means that we might undo the data change. */
814 static void
815 assign_temporarily (int regno, int hard_regno)
817 int p;
818 lra_live_range_t r;
820 for (r = lra_reg_info[regno].live_ranges; r != NULL; r = r->next)
822 for (p = r->start; p <= r->finish; p++)
823 if (hard_regno < 0)
824 bitmap_clear_bit (&live_hard_reg_pseudos[p], regno);
825 else
827 bitmap_set_bit (&live_hard_reg_pseudos[p], regno);
828 insert_in_live_range_start_chain (regno);
831 live_pseudos_reg_renumber[regno] = hard_regno;
834 /* Array used for sorting reload pseudos for subsequent allocation
835 after spilling some pseudo. */
836 static int *sorted_reload_pseudos;
838 /* Spill some pseudos for a reload pseudo REGNO and return hard
839 register which should be used for pseudo after spilling. The
840 function adds spilled pseudos to SPILLED_PSEUDO_BITMAP. When we
841 choose hard register (and pseudos occupying the hard registers and
842 to be spilled), we take into account not only how REGNO will
843 benefit from the spills but also how other reload pseudos not yet
844 assigned to hard registers benefit from the spills too. In very
845 rare cases, the function can fail and return -1.
847 If FIRST_P, return the first available hard reg ignoring other
848 criteria, e.g. allocation cost and cost of spilling non-reload
849 pseudos. This approach results in less hard reg pool fragmentation
850 and permit to allocate hard regs to reload pseudos in complicated
851 situations where pseudo sizes are different. */
852 static int
853 spill_for (int regno, bitmap spilled_pseudo_bitmap, bool first_p)
855 int i, j, n, p, hard_regno, best_hard_regno, cost, best_cost, rclass_size;
856 int reload_hard_regno, reload_cost;
857 machine_mode mode;
858 enum reg_class rclass;
859 unsigned int spill_regno, reload_regno, uid;
860 int insn_pseudos_num, best_insn_pseudos_num;
861 lra_live_range_t r;
862 bitmap_iterator bi;
864 rclass = regno_allocno_class_array[regno];
865 lra_assert (reg_renumber[regno] < 0 && rclass != NO_REGS);
866 bitmap_clear (&insn_conflict_pseudos);
867 bitmap_clear (&best_spill_pseudos_bitmap);
868 EXECUTE_IF_SET_IN_BITMAP (&lra_reg_info[regno].insn_bitmap, 0, uid, bi)
870 struct lra_insn_reg *ir;
872 for (ir = lra_get_insn_regs (uid); ir != NULL; ir = ir->next)
873 if (ir->regno >= FIRST_PSEUDO_REGISTER)
874 bitmap_set_bit (&insn_conflict_pseudos, ir->regno);
876 best_hard_regno = -1;
877 best_cost = INT_MAX;
878 best_insn_pseudos_num = INT_MAX;
879 rclass_size = ira_class_hard_regs_num[rclass];
880 mode = PSEUDO_REGNO_MODE (regno);
881 /* Invalidate try_hard_reg_pseudos elements. */
882 curr_pseudo_check++;
883 for (r = lra_reg_info[regno].live_ranges; r != NULL; r = r->next)
884 for (p = r->start; p <= r->finish; p++)
885 setup_try_hard_regno_pseudos (p, rclass);
886 for (i = 0; i < rclass_size; i++)
888 hard_regno = ira_class_hard_regs[rclass][i];
889 bitmap_clear (&spill_pseudos_bitmap);
890 for (j = hard_regno_nregs[hard_regno][mode] - 1; j >= 0; j--)
892 if (try_hard_reg_pseudos_check[hard_regno + j] != curr_pseudo_check)
893 continue;
894 lra_assert (!bitmap_empty_p (&try_hard_reg_pseudos[hard_regno + j]));
895 bitmap_ior_into (&spill_pseudos_bitmap,
896 &try_hard_reg_pseudos[hard_regno + j]);
898 /* Spill pseudos. */
899 EXECUTE_IF_SET_IN_BITMAP (&spill_pseudos_bitmap, 0, spill_regno, bi)
900 if ((pic_offset_table_rtx != NULL
901 && spill_regno == REGNO (pic_offset_table_rtx))
902 || ((int) spill_regno >= lra_constraint_new_regno_start
903 && ! bitmap_bit_p (&lra_inheritance_pseudos, spill_regno)
904 && ! bitmap_bit_p (&lra_split_regs, spill_regno)
905 && ! bitmap_bit_p (&lra_subreg_reload_pseudos, spill_regno)
906 && ! bitmap_bit_p (&lra_optional_reload_pseudos, spill_regno)))
907 goto fail;
908 insn_pseudos_num = 0;
909 if (lra_dump_file != NULL)
910 fprintf (lra_dump_file, " Trying %d:", hard_regno);
911 sparseset_clear (live_range_reload_inheritance_pseudos);
912 EXECUTE_IF_SET_IN_BITMAP (&spill_pseudos_bitmap, 0, spill_regno, bi)
914 if (bitmap_bit_p (&insn_conflict_pseudos, spill_regno))
915 insn_pseudos_num++;
916 for (r = lra_reg_info[spill_regno].live_ranges;
917 r != NULL;
918 r = r->next)
920 for (p = r->start; p <= r->finish; p++)
922 lra_live_range_t r2;
924 for (r2 = start_point_ranges[p];
925 r2 != NULL;
926 r2 = r2->start_next)
927 if (r2->regno >= lra_constraint_new_regno_start)
928 sparseset_set_bit (live_range_reload_inheritance_pseudos,
929 r2->regno);
933 n = 0;
934 if (sparseset_cardinality (live_range_reload_inheritance_pseudos)
935 <= (unsigned)LRA_MAX_CONSIDERED_RELOAD_PSEUDOS)
936 EXECUTE_IF_SET_IN_SPARSESET (live_range_reload_inheritance_pseudos,
937 reload_regno)
938 if ((int) reload_regno != regno
939 && (ira_reg_classes_intersect_p
940 [rclass][regno_allocno_class_array[reload_regno]])
941 && live_pseudos_reg_renumber[reload_regno] < 0
942 && find_hard_regno_for (reload_regno, &cost, -1, first_p) < 0)
943 sorted_reload_pseudos[n++] = reload_regno;
944 EXECUTE_IF_SET_IN_BITMAP (&spill_pseudos_bitmap, 0, spill_regno, bi)
946 update_lives (spill_regno, true);
947 if (lra_dump_file != NULL)
948 fprintf (lra_dump_file, " spill %d(freq=%d)",
949 spill_regno, lra_reg_info[spill_regno].freq);
951 hard_regno = find_hard_regno_for (regno, &cost, -1, first_p);
952 if (hard_regno >= 0)
954 assign_temporarily (regno, hard_regno);
955 qsort (sorted_reload_pseudos, n, sizeof (int),
956 reload_pseudo_compare_func);
957 for (j = 0; j < n; j++)
959 reload_regno = sorted_reload_pseudos[j];
960 lra_assert (live_pseudos_reg_renumber[reload_regno] < 0);
961 if ((reload_hard_regno
962 = find_hard_regno_for (reload_regno,
963 &reload_cost, -1, first_p)) >= 0)
965 if (lra_dump_file != NULL)
966 fprintf (lra_dump_file, " assign %d(cost=%d)",
967 reload_regno, reload_cost);
968 assign_temporarily (reload_regno, reload_hard_regno);
969 cost += reload_cost;
972 EXECUTE_IF_SET_IN_BITMAP (&spill_pseudos_bitmap, 0, spill_regno, bi)
974 rtx_insn_list *x;
976 cost += lra_reg_info[spill_regno].freq;
977 if (ira_reg_equiv[spill_regno].memory != NULL
978 || ira_reg_equiv[spill_regno].constant != NULL)
979 for (x = ira_reg_equiv[spill_regno].init_insns;
980 x != NULL;
981 x = x->next ())
982 cost -= REG_FREQ_FROM_BB (BLOCK_FOR_INSN (x->insn ()));
984 if (best_insn_pseudos_num > insn_pseudos_num
985 || (best_insn_pseudos_num == insn_pseudos_num
986 && best_cost > cost))
988 best_insn_pseudos_num = insn_pseudos_num;
989 best_cost = cost;
990 best_hard_regno = hard_regno;
991 bitmap_copy (&best_spill_pseudos_bitmap, &spill_pseudos_bitmap);
992 if (lra_dump_file != NULL)
993 fprintf (lra_dump_file, " Now best %d(cost=%d)\n",
994 hard_regno, cost);
996 assign_temporarily (regno, -1);
997 for (j = 0; j < n; j++)
999 reload_regno = sorted_reload_pseudos[j];
1000 if (live_pseudos_reg_renumber[reload_regno] >= 0)
1001 assign_temporarily (reload_regno, -1);
1004 if (lra_dump_file != NULL)
1005 fprintf (lra_dump_file, "\n");
1006 /* Restore the live hard reg pseudo info for spilled pseudos. */
1007 EXECUTE_IF_SET_IN_BITMAP (&spill_pseudos_bitmap, 0, spill_regno, bi)
1008 update_lives (spill_regno, false);
1009 fail:
1012 /* Spill: */
1013 EXECUTE_IF_SET_IN_BITMAP (&best_spill_pseudos_bitmap, 0, spill_regno, bi)
1015 if ((int) spill_regno >= lra_constraint_new_regno_start)
1016 former_reload_pseudo_spill_p = true;
1017 if (lra_dump_file != NULL)
1018 fprintf (lra_dump_file, " Spill %sr%d(hr=%d, freq=%d) for r%d\n",
1019 pseudo_prefix_title (spill_regno),
1020 spill_regno, reg_renumber[spill_regno],
1021 lra_reg_info[spill_regno].freq, regno);
1022 update_lives (spill_regno, true);
1023 lra_setup_reg_renumber (spill_regno, -1, false);
1025 bitmap_ior_into (spilled_pseudo_bitmap, &best_spill_pseudos_bitmap);
1026 return best_hard_regno;
1029 /* Assign HARD_REGNO to REGNO. */
1030 static void
1031 assign_hard_regno (int hard_regno, int regno)
1033 int i;
1035 lra_assert (hard_regno >= 0);
1036 lra_setup_reg_renumber (regno, hard_regno, true);
1037 update_lives (regno, false);
1038 for (i = 0;
1039 i < hard_regno_nregs[hard_regno][lra_reg_info[regno].biggest_mode];
1040 i++)
1041 df_set_regs_ever_live (hard_regno + i, true);
1044 /* Array used for sorting different pseudos. */
1045 static int *sorted_pseudos;
1047 /* The constraints pass is allowed to create equivalences between
1048 pseudos that make the current allocation "incorrect" (in the sense
1049 that pseudos are assigned to hard registers from their own conflict
1050 sets). The global variable lra_risky_transformations_p says
1051 whether this might have happened.
1053 Process pseudos assigned to hard registers (less frequently used
1054 first), spill if a conflict is found, and mark the spilled pseudos
1055 in SPILLED_PSEUDO_BITMAP. Set up LIVE_HARD_REG_PSEUDOS from
1056 pseudos, assigned to hard registers. */
1057 static void
1058 setup_live_pseudos_and_spill_after_risky_transforms (bitmap
1059 spilled_pseudo_bitmap)
1061 int p, i, j, n, regno, hard_regno;
1062 unsigned int k, conflict_regno;
1063 int val, offset;
1064 HARD_REG_SET conflict_set;
1065 machine_mode mode;
1066 lra_live_range_t r;
1067 bitmap_iterator bi;
1068 int max_regno = max_reg_num ();
1070 if (! lra_risky_transformations_p)
1072 for (i = FIRST_PSEUDO_REGISTER; i < max_regno; i++)
1073 if (reg_renumber[i] >= 0 && lra_reg_info[i].nrefs > 0)
1074 update_lives (i, false);
1075 return;
1077 for (n = 0, i = FIRST_PSEUDO_REGISTER; i < max_regno; i++)
1078 if ((pic_offset_table_rtx == NULL_RTX
1079 || i != (int) REGNO (pic_offset_table_rtx))
1080 && reg_renumber[i] >= 0 && lra_reg_info[i].nrefs > 0)
1081 sorted_pseudos[n++] = i;
1082 qsort (sorted_pseudos, n, sizeof (int), pseudo_compare_func);
1083 if (pic_offset_table_rtx != NULL_RTX
1084 && (regno = REGNO (pic_offset_table_rtx)) >= FIRST_PSEUDO_REGISTER
1085 && reg_renumber[regno] >= 0 && lra_reg_info[regno].nrefs > 0)
1086 sorted_pseudos[n++] = regno;
1087 for (i = n - 1; i >= 0; i--)
1089 regno = sorted_pseudos[i];
1090 hard_regno = reg_renumber[regno];
1091 lra_assert (hard_regno >= 0);
1092 mode = lra_reg_info[regno].biggest_mode;
1093 sparseset_clear (live_range_hard_reg_pseudos);
1094 for (r = lra_reg_info[regno].live_ranges; r != NULL; r = r->next)
1096 EXECUTE_IF_SET_IN_BITMAP (&live_hard_reg_pseudos[r->start], 0, k, bi)
1097 sparseset_set_bit (live_range_hard_reg_pseudos, k);
1098 for (p = r->start + 1; p <= r->finish; p++)
1100 lra_live_range_t r2;
1102 for (r2 = start_point_ranges[p];
1103 r2 != NULL;
1104 r2 = r2->start_next)
1105 if (live_pseudos_reg_renumber[r2->regno] >= 0)
1106 sparseset_set_bit (live_range_hard_reg_pseudos, r2->regno);
1109 COPY_HARD_REG_SET (conflict_set, lra_no_alloc_regs);
1110 IOR_HARD_REG_SET (conflict_set, lra_reg_info[regno].conflict_hard_regs);
1111 val = lra_reg_info[regno].val;
1112 offset = lra_reg_info[regno].offset;
1113 EXECUTE_IF_SET_IN_SPARSESET (live_range_hard_reg_pseudos, conflict_regno)
1114 if (!lra_reg_val_equal_p (conflict_regno, val, offset)
1115 /* If it is multi-register pseudos they should start on
1116 the same hard register. */
1117 || hard_regno != reg_renumber[conflict_regno])
1118 add_to_hard_reg_set (&conflict_set,
1119 lra_reg_info[conflict_regno].biggest_mode,
1120 reg_renumber[conflict_regno]);
1121 if (! overlaps_hard_reg_set_p (conflict_set, mode, hard_regno))
1123 update_lives (regno, false);
1124 continue;
1126 bitmap_set_bit (spilled_pseudo_bitmap, regno);
1127 for (j = 0;
1128 j < hard_regno_nregs[hard_regno][PSEUDO_REGNO_MODE (regno)];
1129 j++)
1130 lra_hard_reg_usage[hard_regno + j] -= lra_reg_info[regno].freq;
1131 reg_renumber[regno] = -1;
1132 if (regno >= lra_constraint_new_regno_start)
1133 former_reload_pseudo_spill_p = true;
1134 if (lra_dump_file != NULL)
1135 fprintf (lra_dump_file, " Spill r%d after risky transformations\n",
1136 regno);
1140 /* Improve allocation by assigning the same hard regno of inheritance
1141 pseudos to the connected pseudos. We need this because inheritance
1142 pseudos are allocated after reload pseudos in the thread and when
1143 we assign a hard register to a reload pseudo we don't know yet that
1144 the connected inheritance pseudos can get the same hard register.
1145 Add pseudos with changed allocation to bitmap CHANGED_PSEUDOS. */
1146 static void
1147 improve_inheritance (bitmap changed_pseudos)
1149 unsigned int k;
1150 int regno, another_regno, hard_regno, another_hard_regno, cost, i, n;
1151 lra_copy_t cp, next_cp;
1152 bitmap_iterator bi;
1154 if (lra_inheritance_iter > LRA_MAX_INHERITANCE_PASSES)
1155 return;
1156 n = 0;
1157 EXECUTE_IF_SET_IN_BITMAP (&lra_inheritance_pseudos, 0, k, bi)
1158 if (reg_renumber[k] >= 0 && lra_reg_info[k].nrefs != 0)
1159 sorted_pseudos[n++] = k;
1160 qsort (sorted_pseudos, n, sizeof (int), pseudo_compare_func);
1161 for (i = 0; i < n; i++)
1163 regno = sorted_pseudos[i];
1164 hard_regno = reg_renumber[regno];
1165 lra_assert (hard_regno >= 0);
1166 for (cp = lra_reg_info[regno].copies; cp != NULL; cp = next_cp)
1168 if (cp->regno1 == regno)
1170 next_cp = cp->regno1_next;
1171 another_regno = cp->regno2;
1173 else if (cp->regno2 == regno)
1175 next_cp = cp->regno2_next;
1176 another_regno = cp->regno1;
1178 else
1179 gcc_unreachable ();
1180 /* Don't change reload pseudo allocation. It might have
1181 this allocation for a purpose and changing it can result
1182 in LRA cycling. */
1183 if ((another_regno < lra_constraint_new_regno_start
1184 || bitmap_bit_p (&lra_inheritance_pseudos, another_regno))
1185 && (another_hard_regno = reg_renumber[another_regno]) >= 0
1186 && another_hard_regno != hard_regno)
1188 if (lra_dump_file != NULL)
1189 fprintf
1190 (lra_dump_file,
1191 " Improving inheritance for %d(%d) and %d(%d)...\n",
1192 regno, hard_regno, another_regno, another_hard_regno);
1193 update_lives (another_regno, true);
1194 lra_setup_reg_renumber (another_regno, -1, false);
1195 if (hard_regno == find_hard_regno_for (another_regno, &cost,
1196 hard_regno, false))
1197 assign_hard_regno (hard_regno, another_regno);
1198 else
1199 assign_hard_regno (another_hard_regno, another_regno);
1200 bitmap_set_bit (changed_pseudos, another_regno);
1207 /* Bitmap finally containing all pseudos spilled on this assignment
1208 pass. */
1209 static bitmap_head all_spilled_pseudos;
1210 /* All pseudos whose allocation was changed. */
1211 static bitmap_head changed_pseudo_bitmap;
1214 /* Add to LIVE_RANGE_HARD_REG_PSEUDOS all pseudos conflicting with
1215 REGNO and whose hard regs can be assigned to REGNO. */
1216 static void
1217 find_all_spills_for (int regno)
1219 int p;
1220 lra_live_range_t r;
1221 unsigned int k;
1222 bitmap_iterator bi;
1223 enum reg_class rclass;
1224 bool *rclass_intersect_p;
1226 rclass = regno_allocno_class_array[regno];
1227 rclass_intersect_p = ira_reg_classes_intersect_p[rclass];
1228 for (r = lra_reg_info[regno].live_ranges; r != NULL; r = r->next)
1230 EXECUTE_IF_SET_IN_BITMAP (&live_hard_reg_pseudos[r->start], 0, k, bi)
1231 if (rclass_intersect_p[regno_allocno_class_array[k]])
1232 sparseset_set_bit (live_range_hard_reg_pseudos, k);
1233 for (p = r->start + 1; p <= r->finish; p++)
1235 lra_live_range_t r2;
1237 for (r2 = start_point_ranges[p];
1238 r2 != NULL;
1239 r2 = r2->start_next)
1241 if (live_pseudos_reg_renumber[r2->regno] >= 0
1242 && rclass_intersect_p[regno_allocno_class_array[r2->regno]])
1243 sparseset_set_bit (live_range_hard_reg_pseudos, r2->regno);
1249 /* Assign hard registers to reload pseudos and other pseudos. */
1250 static void
1251 assign_by_spills (void)
1253 int i, n, nfails, iter, regno, hard_regno, cost, restore_regno;
1254 rtx_insn *insn;
1255 bitmap_head changed_insns, do_not_assign_nonreload_pseudos;
1256 unsigned int u, conflict_regno;
1257 bitmap_iterator bi;
1258 bool reload_p;
1259 int max_regno = max_reg_num ();
1261 for (n = 0, i = lra_constraint_new_regno_start; i < max_regno; i++)
1262 if (reg_renumber[i] < 0 && lra_reg_info[i].nrefs != 0
1263 && regno_allocno_class_array[i] != NO_REGS)
1264 sorted_pseudos[n++] = i;
1265 bitmap_initialize (&insn_conflict_pseudos, &reg_obstack);
1266 bitmap_initialize (&spill_pseudos_bitmap, &reg_obstack);
1267 bitmap_initialize (&best_spill_pseudos_bitmap, &reg_obstack);
1268 update_hard_regno_preference_check = XCNEWVEC (int, max_regno);
1269 curr_update_hard_regno_preference_check = 0;
1270 memset (try_hard_reg_pseudos_check, 0, sizeof (try_hard_reg_pseudos_check));
1271 for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
1272 bitmap_initialize (&try_hard_reg_pseudos[i], &reg_obstack);
1273 curr_pseudo_check = 0;
1274 bitmap_initialize (&changed_insns, &reg_obstack);
1275 bitmap_initialize (&non_reload_pseudos, &reg_obstack);
1276 bitmap_ior (&non_reload_pseudos, &lra_inheritance_pseudos, &lra_split_regs);
1277 bitmap_ior_into (&non_reload_pseudos, &lra_subreg_reload_pseudos);
1278 bitmap_ior_into (&non_reload_pseudos, &lra_optional_reload_pseudos);
1279 for (iter = 0; iter <= 1; iter++)
1281 qsort (sorted_pseudos, n, sizeof (int), reload_pseudo_compare_func);
1282 nfails = 0;
1283 for (i = 0; i < n; i++)
1285 regno = sorted_pseudos[i];
1286 if (lra_dump_file != NULL)
1287 fprintf (lra_dump_file, " Assigning to %d "
1288 "(cl=%s, orig=%d, freq=%d, tfirst=%d, tfreq=%d)...\n",
1289 regno, reg_class_names[regno_allocno_class_array[regno]],
1290 ORIGINAL_REGNO (regno_reg_rtx[regno]),
1291 lra_reg_info[regno].freq, regno_assign_info[regno].first,
1292 regno_assign_info[regno_assign_info[regno].first].freq);
1293 hard_regno = find_hard_regno_for (regno, &cost, -1, iter == 1);
1294 reload_p = ! bitmap_bit_p (&non_reload_pseudos, regno);
1295 if (hard_regno < 0 && reload_p)
1296 hard_regno = spill_for (regno, &all_spilled_pseudos, iter == 1);
1297 if (hard_regno < 0)
1299 if (reload_p)
1300 sorted_pseudos[nfails++] = regno;
1302 else
1304 /* This register might have been spilled by the previous
1305 pass. Indicate that it is no longer spilled. */
1306 bitmap_clear_bit (&all_spilled_pseudos, regno);
1307 assign_hard_regno (hard_regno, regno);
1308 if (! reload_p)
1309 /* As non-reload pseudo assignment is changed we
1310 should reconsider insns referring for the
1311 pseudo. */
1312 bitmap_set_bit (&changed_pseudo_bitmap, regno);
1315 if (nfails == 0)
1316 break;
1317 if (iter > 0)
1319 /* We did not assign hard regs to reload pseudos after two iterations.
1320 Either it's an asm and something is wrong with the constraints, or
1321 we have run out of spill registers; error out in either case. */
1322 bool asm_p = false;
1323 bitmap_head failed_reload_insns;
1325 bitmap_initialize (&failed_reload_insns, &reg_obstack);
1326 for (i = 0; i < nfails; i++)
1328 regno = sorted_pseudos[i];
1329 bitmap_ior_into (&failed_reload_insns,
1330 &lra_reg_info[regno].insn_bitmap);
1331 /* Assign an arbitrary hard register of regno class to
1332 avoid further trouble with this insn. */
1333 bitmap_clear_bit (&all_spilled_pseudos, regno);
1334 assign_hard_regno
1335 (ira_class_hard_regs[regno_allocno_class_array[regno]][0],
1336 regno);
1338 EXECUTE_IF_SET_IN_BITMAP (&failed_reload_insns, 0, u, bi)
1340 insn = lra_insn_recog_data[u]->insn;
1341 if (asm_noperands (PATTERN (insn)) >= 0)
1343 asm_p = true;
1344 error_for_asm (insn,
1345 "%<asm%> operand has impossible constraints");
1346 /* Avoid further trouble with this insn.
1347 For asm goto, instead of fixing up all the edges
1348 just clear the template and clear input operands
1349 (asm goto doesn't have any output operands). */
1350 if (JUMP_P (insn))
1352 rtx asm_op = extract_asm_operands (PATTERN (insn));
1353 ASM_OPERANDS_TEMPLATE (asm_op) = ggc_strdup ("");
1354 ASM_OPERANDS_INPUT_VEC (asm_op) = rtvec_alloc (0);
1355 ASM_OPERANDS_INPUT_CONSTRAINT_VEC (asm_op) = rtvec_alloc (0);
1356 lra_update_insn_regno_info (insn);
1358 else
1360 PATTERN (insn) = gen_rtx_USE (VOIDmode, const0_rtx);
1361 lra_set_insn_deleted (insn);
1364 else if (!asm_p)
1366 error ("unable to find a register to spill");
1367 fatal_insn ("this is the insn:", insn);
1370 break;
1372 /* This is a very rare event. We can not assign a hard register
1373 to reload pseudo because the hard register was assigned to
1374 another reload pseudo on a previous assignment pass. For x86
1375 example, on the 1st pass we assigned CX (although another
1376 hard register could be used for this) to reload pseudo in an
1377 insn, on the 2nd pass we need CX (and only this) hard
1378 register for a new reload pseudo in the same insn. Another
1379 possible situation may occur in assigning to multi-regs
1380 reload pseudos when hard regs pool is too fragmented even
1381 after spilling non-reload pseudos.
1383 We should do something radical here to succeed. Here we
1384 spill *all* conflicting pseudos and reassign them. */
1385 if (lra_dump_file != NULL)
1386 fprintf (lra_dump_file, " 2nd iter for reload pseudo assignments:\n");
1387 sparseset_clear (live_range_hard_reg_pseudos);
1388 for (i = 0; i < nfails; i++)
1390 if (lra_dump_file != NULL)
1391 fprintf (lra_dump_file, " Reload r%d assignment failure\n",
1392 sorted_pseudos[i]);
1393 find_all_spills_for (sorted_pseudos[i]);
1395 EXECUTE_IF_SET_IN_SPARSESET (live_range_hard_reg_pseudos, conflict_regno)
1397 if ((int) conflict_regno >= lra_constraint_new_regno_start)
1399 sorted_pseudos[nfails++] = conflict_regno;
1400 former_reload_pseudo_spill_p = true;
1402 if (lra_dump_file != NULL)
1403 fprintf (lra_dump_file, " Spill %s r%d(hr=%d, freq=%d)\n",
1404 pseudo_prefix_title (conflict_regno), conflict_regno,
1405 reg_renumber[conflict_regno],
1406 lra_reg_info[conflict_regno].freq);
1407 update_lives (conflict_regno, true);
1408 lra_setup_reg_renumber (conflict_regno, -1, false);
1410 n = nfails;
1412 improve_inheritance (&changed_pseudo_bitmap);
1413 bitmap_clear (&non_reload_pseudos);
1414 bitmap_clear (&changed_insns);
1415 if (! lra_simple_p)
1417 /* We should not assign to original pseudos of inheritance
1418 pseudos or split pseudos if any its inheritance pseudo did
1419 not get hard register or any its split pseudo was not split
1420 because undo inheritance/split pass will extend live range of
1421 such inheritance or split pseudos. */
1422 bitmap_initialize (&do_not_assign_nonreload_pseudos, &reg_obstack);
1423 EXECUTE_IF_SET_IN_BITMAP (&lra_inheritance_pseudos, 0, u, bi)
1424 if ((restore_regno = lra_reg_info[u].restore_regno) >= 0
1425 && reg_renumber[u] < 0
1426 && bitmap_bit_p (&lra_inheritance_pseudos, u))
1427 bitmap_set_bit (&do_not_assign_nonreload_pseudos, restore_regno);
1428 EXECUTE_IF_SET_IN_BITMAP (&lra_split_regs, 0, u, bi)
1429 if ((restore_regno = lra_reg_info[u].restore_regno) >= 0
1430 && reg_renumber[u] >= 0)
1431 bitmap_set_bit (&do_not_assign_nonreload_pseudos, restore_regno);
1432 for (n = 0, i = FIRST_PSEUDO_REGISTER; i < max_regno; i++)
1433 if (((i < lra_constraint_new_regno_start
1434 && ! bitmap_bit_p (&do_not_assign_nonreload_pseudos, i))
1435 || (bitmap_bit_p (&lra_inheritance_pseudos, i)
1436 && lra_reg_info[i].restore_regno >= 0)
1437 || (bitmap_bit_p (&lra_split_regs, i)
1438 && lra_reg_info[i].restore_regno >= 0)
1439 || bitmap_bit_p (&lra_subreg_reload_pseudos, i)
1440 || bitmap_bit_p (&lra_optional_reload_pseudos, i))
1441 && reg_renumber[i] < 0 && lra_reg_info[i].nrefs != 0
1442 && regno_allocno_class_array[i] != NO_REGS)
1443 sorted_pseudos[n++] = i;
1444 bitmap_clear (&do_not_assign_nonreload_pseudos);
1445 if (n != 0 && lra_dump_file != NULL)
1446 fprintf (lra_dump_file, " Reassigning non-reload pseudos\n");
1447 qsort (sorted_pseudos, n, sizeof (int), pseudo_compare_func);
1448 for (i = 0; i < n; i++)
1450 regno = sorted_pseudos[i];
1451 hard_regno = find_hard_regno_for (regno, &cost, -1, false);
1452 if (hard_regno >= 0)
1454 assign_hard_regno (hard_regno, regno);
1455 /* We change allocation for non-reload pseudo on this
1456 iteration -- mark the pseudo for invalidation of used
1457 alternatives of insns containing the pseudo. */
1458 bitmap_set_bit (&changed_pseudo_bitmap, regno);
1460 else
1462 enum reg_class rclass = lra_get_allocno_class (regno);
1463 enum reg_class spill_class;
1465 if (targetm.spill_class == NULL
1466 || lra_reg_info[regno].restore_regno < 0
1467 || ! bitmap_bit_p (&lra_inheritance_pseudos, regno)
1468 || (spill_class
1469 = ((enum reg_class)
1470 targetm.spill_class
1471 ((reg_class_t) rclass,
1472 PSEUDO_REGNO_MODE (regno)))) == NO_REGS)
1473 continue;
1474 regno_allocno_class_array[regno] = spill_class;
1475 hard_regno = find_hard_regno_for (regno, &cost, -1, false);
1476 if (hard_regno < 0)
1477 regno_allocno_class_array[regno] = rclass;
1478 else
1480 setup_reg_classes
1481 (regno, spill_class, spill_class, spill_class);
1482 assign_hard_regno (hard_regno, regno);
1483 bitmap_set_bit (&changed_pseudo_bitmap, regno);
1488 free (update_hard_regno_preference_check);
1489 bitmap_clear (&best_spill_pseudos_bitmap);
1490 bitmap_clear (&spill_pseudos_bitmap);
1491 bitmap_clear (&insn_conflict_pseudos);
1495 /* Entry function to assign hard registers to new reload pseudos
1496 starting with LRA_CONSTRAINT_NEW_REGNO_START (by possible spilling
1497 of old pseudos) and possibly to the old pseudos. The function adds
1498 what insns to process for the next constraint pass. Those are all
1499 insns who contains non-reload and non-inheritance pseudos with
1500 changed allocation.
1502 Return true if we did not spill any non-reload and non-inheritance
1503 pseudos. */
1504 bool
1505 lra_assign (void)
1507 int i;
1508 unsigned int u;
1509 bitmap_iterator bi;
1510 bitmap_head insns_to_process;
1511 bool no_spills_p;
1512 int max_regno = max_reg_num ();
1514 timevar_push (TV_LRA_ASSIGN);
1515 lra_assignment_iter++;
1516 if (lra_dump_file != NULL)
1517 fprintf (lra_dump_file, "\n********** Assignment #%d: **********\n\n",
1518 lra_assignment_iter);
1519 init_lives ();
1520 sorted_pseudos = XNEWVEC (int, max_regno);
1521 sorted_reload_pseudos = XNEWVEC (int, max_regno);
1522 regno_allocno_class_array = XNEWVEC (enum reg_class, max_regno);
1523 for (i = FIRST_PSEUDO_REGISTER; i < max_regno; i++)
1524 regno_allocno_class_array[i] = lra_get_allocno_class (i);
1525 former_reload_pseudo_spill_p = false;
1526 init_regno_assign_info ();
1527 bitmap_initialize (&all_spilled_pseudos, &reg_obstack);
1528 create_live_range_start_chains ();
1529 setup_live_pseudos_and_spill_after_risky_transforms (&all_spilled_pseudos);
1530 #ifdef ENABLE_CHECKING
1531 if (!flag_ipa_ra)
1532 for (i = FIRST_PSEUDO_REGISTER; i < max_regno; i++)
1533 if (lra_reg_info[i].nrefs != 0 && reg_renumber[i] >= 0
1534 && lra_reg_info[i].call_p
1535 && overlaps_hard_reg_set_p (call_used_reg_set,
1536 PSEUDO_REGNO_MODE (i), reg_renumber[i]))
1537 gcc_unreachable ();
1538 #endif
1539 /* Setup insns to process on the next constraint pass. */
1540 bitmap_initialize (&changed_pseudo_bitmap, &reg_obstack);
1541 init_live_reload_and_inheritance_pseudos ();
1542 assign_by_spills ();
1543 finish_live_reload_and_inheritance_pseudos ();
1544 bitmap_ior_into (&changed_pseudo_bitmap, &all_spilled_pseudos);
1545 no_spills_p = true;
1546 EXECUTE_IF_SET_IN_BITMAP (&all_spilled_pseudos, 0, u, bi)
1547 /* We ignore spilled pseudos created on last inheritance pass
1548 because they will be removed. */
1549 if (lra_reg_info[u].restore_regno < 0)
1551 no_spills_p = false;
1552 break;
1554 finish_live_range_start_chains ();
1555 bitmap_clear (&all_spilled_pseudos);
1556 bitmap_initialize (&insns_to_process, &reg_obstack);
1557 EXECUTE_IF_SET_IN_BITMAP (&changed_pseudo_bitmap, 0, u, bi)
1558 bitmap_ior_into (&insns_to_process, &lra_reg_info[u].insn_bitmap);
1559 bitmap_clear (&changed_pseudo_bitmap);
1560 EXECUTE_IF_SET_IN_BITMAP (&insns_to_process, 0, u, bi)
1562 lra_push_insn_by_uid (u);
1563 /* Invalidate alternatives for insn should be processed. */
1564 lra_set_used_insn_alternative_by_uid (u, -1);
1566 bitmap_clear (&insns_to_process);
1567 finish_regno_assign_info ();
1568 free (regno_allocno_class_array);
1569 free (sorted_pseudos);
1570 free (sorted_reload_pseudos);
1571 finish_lives ();
1572 timevar_pop (TV_LRA_ASSIGN);
1573 if (former_reload_pseudo_spill_p)
1574 lra_assignment_iter_after_spill++;
1575 if (lra_assignment_iter_after_spill > LRA_MAX_ASSIGNMENT_ITERATION_NUMBER)
1576 internal_error
1577 ("Maximum number of LRA assignment passes is achieved (%d)\n",
1578 LRA_MAX_ASSIGNMENT_ITERATION_NUMBER);
1579 return no_spills_p;