Relocation (= move+destroy)
[official-gcc.git] / gcc / postreload.c
blob56cb14dc6760fc2b57d52c855234402a85067252
1 /* Perform simple optimizations to clean up the result of reload.
2 Copyright (C) 1987-2018 Free Software Foundation, Inc.
4 This file is part of GCC.
6 GCC is free software; you can redistribute it and/or modify it under
7 the terms of the GNU General Public License as published by the Free
8 Software Foundation; either version 3, or (at your option) any later
9 version.
11 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
12 WARRANTY; without even the implied warranty of MERCHANTABILITY or
13 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
14 for more details.
16 You should have received a copy of the GNU General Public License
17 along with GCC; see the file COPYING3. If not see
18 <http://www.gnu.org/licenses/>. */
20 #include "config.h"
21 #include "system.h"
22 #include "coretypes.h"
23 #include "backend.h"
24 #include "target.h"
25 #include "rtl.h"
26 #include "tree.h"
27 #include "predict.h"
28 #include "df.h"
29 #include "memmodel.h"
30 #include "tm_p.h"
31 #include "optabs.h"
32 #include "regs.h"
33 #include "emit-rtl.h"
34 #include "recog.h"
36 #include "cfgrtl.h"
37 #include "cfgbuild.h"
38 #include "cfgcleanup.h"
39 #include "reload.h"
40 #include "cselib.h"
41 #include "tree-pass.h"
42 #include "dbgcnt.h"
44 static int reload_cse_noop_set_p (rtx);
45 static bool reload_cse_simplify (rtx_insn *, rtx);
46 static void reload_cse_regs_1 (void);
47 static int reload_cse_simplify_set (rtx, rtx_insn *);
48 static int reload_cse_simplify_operands (rtx_insn *, rtx);
50 static void reload_combine (void);
51 static void reload_combine_note_use (rtx *, rtx_insn *, int, rtx);
52 static void reload_combine_note_store (rtx, const_rtx, void *);
54 static bool reload_cse_move2add (rtx_insn *);
55 static void move2add_note_store (rtx, const_rtx, void *);
57 /* Call cse / combine like post-reload optimization phases.
58 FIRST is the first instruction. */
60 static void
61 reload_cse_regs (rtx_insn *first ATTRIBUTE_UNUSED)
63 bool moves_converted;
64 reload_cse_regs_1 ();
65 reload_combine ();
66 moves_converted = reload_cse_move2add (first);
67 if (flag_expensive_optimizations)
69 if (moves_converted)
70 reload_combine ();
71 reload_cse_regs_1 ();
75 /* See whether a single set SET is a noop. */
76 static int
77 reload_cse_noop_set_p (rtx set)
79 if (cselib_reg_set_mode (SET_DEST (set)) != GET_MODE (SET_DEST (set)))
80 return 0;
82 return rtx_equal_for_cselib_p (SET_DEST (set), SET_SRC (set));
85 /* Try to simplify INSN. Return true if the CFG may have changed. */
86 static bool
87 reload_cse_simplify (rtx_insn *insn, rtx testreg)
89 rtx body = PATTERN (insn);
90 basic_block insn_bb = BLOCK_FOR_INSN (insn);
91 unsigned insn_bb_succs = EDGE_COUNT (insn_bb->succs);
93 /* If NO_FUNCTION_CSE has been set by the target, then we should not try
94 to cse function calls. */
95 if (NO_FUNCTION_CSE && CALL_P (insn))
96 return false;
98 if (GET_CODE (body) == SET)
100 int count = 0;
102 /* Simplify even if we may think it is a no-op.
103 We may think a memory load of a value smaller than WORD_SIZE
104 is redundant because we haven't taken into account possible
105 implicit extension. reload_cse_simplify_set() will bring
106 this out, so it's safer to simplify before we delete. */
107 count += reload_cse_simplify_set (body, insn);
109 if (!count && reload_cse_noop_set_p (body))
111 if (check_for_inc_dec (insn))
112 delete_insn_and_edges (insn);
113 /* We're done with this insn. */
114 goto done;
117 if (count > 0)
118 apply_change_group ();
119 else
120 reload_cse_simplify_operands (insn, testreg);
122 else if (GET_CODE (body) == PARALLEL)
124 int i;
125 int count = 0;
126 rtx value = NULL_RTX;
128 /* Registers mentioned in the clobber list for an asm cannot be reused
129 within the body of the asm. Invalidate those registers now so that
130 we don't try to substitute values for them. */
131 if (asm_noperands (body) >= 0)
133 for (i = XVECLEN (body, 0) - 1; i >= 0; --i)
135 rtx part = XVECEXP (body, 0, i);
136 /* asms can only have full clobbers, not clobber_highs. */
137 gcc_assert (GET_CODE (part) != CLOBBER_HIGH);
138 if (GET_CODE (part) == CLOBBER && REG_P (XEXP (part, 0)))
139 cselib_invalidate_rtx (XEXP (part, 0));
143 /* If every action in a PARALLEL is a noop, we can delete
144 the entire PARALLEL. */
145 for (i = XVECLEN (body, 0) - 1; i >= 0; --i)
147 rtx part = XVECEXP (body, 0, i);
148 if (GET_CODE (part) == SET)
150 if (! reload_cse_noop_set_p (part))
151 break;
152 if (REG_P (SET_DEST (part))
153 && REG_FUNCTION_VALUE_P (SET_DEST (part)))
155 if (value)
156 break;
157 value = SET_DEST (part);
160 else if (GET_CODE (part) != CLOBBER
161 && GET_CODE (part) != CLOBBER_HIGH
162 && GET_CODE (part) != USE)
163 break;
166 if (i < 0)
168 if (check_for_inc_dec (insn))
169 delete_insn_and_edges (insn);
170 /* We're done with this insn. */
171 goto done;
174 /* It's not a no-op, but we can try to simplify it. */
175 for (i = XVECLEN (body, 0) - 1; i >= 0; --i)
176 if (GET_CODE (XVECEXP (body, 0, i)) == SET)
177 count += reload_cse_simplify_set (XVECEXP (body, 0, i), insn);
179 if (count > 0)
180 apply_change_group ();
181 else
182 reload_cse_simplify_operands (insn, testreg);
185 done:
186 return (EDGE_COUNT (insn_bb->succs) != insn_bb_succs);
189 /* Do a very simple CSE pass over the hard registers.
191 This function detects no-op moves where we happened to assign two
192 different pseudo-registers to the same hard register, and then
193 copied one to the other. Reload will generate a useless
194 instruction copying a register to itself.
196 This function also detects cases where we load a value from memory
197 into two different registers, and (if memory is more expensive than
198 registers) changes it to simply copy the first register into the
199 second register.
201 Another optimization is performed that scans the operands of each
202 instruction to see whether the value is already available in a
203 hard register. It then replaces the operand with the hard register
204 if possible, much like an optional reload would. */
206 static void
207 reload_cse_regs_1 (void)
209 bool cfg_changed = false;
210 basic_block bb;
211 rtx_insn *insn;
212 rtx testreg = gen_rtx_REG (word_mode, LAST_VIRTUAL_REGISTER + 1);
214 cselib_init (CSELIB_RECORD_MEMORY);
215 init_alias_analysis ();
217 FOR_EACH_BB_FN (bb, cfun)
218 FOR_BB_INSNS (bb, insn)
220 if (INSN_P (insn))
221 cfg_changed |= reload_cse_simplify (insn, testreg);
223 cselib_process_insn (insn);
226 /* Clean up. */
227 end_alias_analysis ();
228 cselib_finish ();
229 if (cfg_changed)
230 cleanup_cfg (0);
233 /* Try to simplify a single SET instruction. SET is the set pattern.
234 INSN is the instruction it came from.
235 This function only handles one case: if we set a register to a value
236 which is not a register, we try to find that value in some other register
237 and change the set into a register copy. */
239 static int
240 reload_cse_simplify_set (rtx set, rtx_insn *insn)
242 int did_change = 0;
243 int dreg;
244 rtx src;
245 reg_class_t dclass;
246 int old_cost;
247 cselib_val *val;
248 struct elt_loc_list *l;
249 enum rtx_code extend_op = UNKNOWN;
250 bool speed = optimize_bb_for_speed_p (BLOCK_FOR_INSN (insn));
252 dreg = true_regnum (SET_DEST (set));
253 if (dreg < 0)
254 return 0;
256 src = SET_SRC (set);
257 if (side_effects_p (src) || true_regnum (src) >= 0)
258 return 0;
260 dclass = REGNO_REG_CLASS (dreg);
262 /* When replacing a memory with a register, we need to honor assumptions
263 that combine made wrt the contents of sign bits. We'll do this by
264 generating an extend instruction instead of a reg->reg copy. Thus
265 the destination must be a register that we can widen. */
266 if (MEM_P (src)
267 && (extend_op = load_extend_op (GET_MODE (src))) != UNKNOWN
268 && !REG_P (SET_DEST (set)))
269 return 0;
271 val = cselib_lookup (src, GET_MODE (SET_DEST (set)), 0, VOIDmode);
272 if (! val)
273 return 0;
275 /* If memory loads are cheaper than register copies, don't change them. */
276 if (MEM_P (src))
277 old_cost = memory_move_cost (GET_MODE (src), dclass, true);
278 else if (REG_P (src))
279 old_cost = register_move_cost (GET_MODE (src),
280 REGNO_REG_CLASS (REGNO (src)), dclass);
281 else
282 old_cost = set_src_cost (src, GET_MODE (SET_DEST (set)), speed);
284 for (l = val->locs; l; l = l->next)
286 rtx this_rtx = l->loc;
287 int this_cost;
289 if (CONSTANT_P (this_rtx) && ! references_value_p (this_rtx, 0))
291 if (extend_op != UNKNOWN)
293 wide_int result;
295 if (!CONST_SCALAR_INT_P (this_rtx))
296 continue;
298 switch (extend_op)
300 case ZERO_EXTEND:
301 result = wide_int::from (rtx_mode_t (this_rtx,
302 GET_MODE (src)),
303 BITS_PER_WORD, UNSIGNED);
304 break;
305 case SIGN_EXTEND:
306 result = wide_int::from (rtx_mode_t (this_rtx,
307 GET_MODE (src)),
308 BITS_PER_WORD, SIGNED);
309 break;
310 default:
311 gcc_unreachable ();
313 this_rtx = immed_wide_int_const (result, word_mode);
316 this_cost = set_src_cost (this_rtx, GET_MODE (SET_DEST (set)), speed);
318 else if (REG_P (this_rtx))
320 if (extend_op != UNKNOWN)
322 this_rtx = gen_rtx_fmt_e (extend_op, word_mode, this_rtx);
323 this_cost = set_src_cost (this_rtx, word_mode, speed);
325 else
326 this_cost = register_move_cost (GET_MODE (this_rtx),
327 REGNO_REG_CLASS (REGNO (this_rtx)),
328 dclass);
330 else
331 continue;
333 /* If equal costs, prefer registers over anything else. That
334 tends to lead to smaller instructions on some machines. */
335 if (this_cost < old_cost
336 || (this_cost == old_cost
337 && REG_P (this_rtx)
338 && !REG_P (SET_SRC (set))))
340 if (extend_op != UNKNOWN
341 && REG_CAN_CHANGE_MODE_P (REGNO (SET_DEST (set)),
342 GET_MODE (SET_DEST (set)), word_mode))
344 rtx wide_dest = gen_rtx_REG (word_mode, REGNO (SET_DEST (set)));
345 ORIGINAL_REGNO (wide_dest) = ORIGINAL_REGNO (SET_DEST (set));
346 validate_change (insn, &SET_DEST (set), wide_dest, 1);
349 validate_unshare_change (insn, &SET_SRC (set), this_rtx, 1);
350 old_cost = this_cost, did_change = 1;
354 return did_change;
357 /* Try to replace operands in INSN with equivalent values that are already
358 in registers. This can be viewed as optional reloading.
360 For each non-register operand in the insn, see if any hard regs are
361 known to be equivalent to that operand. Record the alternatives which
362 can accept these hard registers. Among all alternatives, select the
363 ones which are better or equal to the one currently matching, where
364 "better" is in terms of '?' and '!' constraints. Among the remaining
365 alternatives, select the one which replaces most operands with
366 hard registers. */
368 static int
369 reload_cse_simplify_operands (rtx_insn *insn, rtx testreg)
371 int i, j;
373 /* For each operand, all registers that are equivalent to it. */
374 HARD_REG_SET equiv_regs[MAX_RECOG_OPERANDS];
376 const char *constraints[MAX_RECOG_OPERANDS];
378 /* Vector recording how bad an alternative is. */
379 int *alternative_reject;
380 /* Vector recording how many registers can be introduced by choosing
381 this alternative. */
382 int *alternative_nregs;
383 /* Array of vectors recording, for each operand and each alternative,
384 which hard register to substitute, or -1 if the operand should be
385 left as it is. */
386 int *op_alt_regno[MAX_RECOG_OPERANDS];
387 /* Array of alternatives, sorted in order of decreasing desirability. */
388 int *alternative_order;
390 extract_constrain_insn (insn);
392 if (recog_data.n_alternatives == 0 || recog_data.n_operands == 0)
393 return 0;
395 alternative_reject = XALLOCAVEC (int, recog_data.n_alternatives);
396 alternative_nregs = XALLOCAVEC (int, recog_data.n_alternatives);
397 alternative_order = XALLOCAVEC (int, recog_data.n_alternatives);
398 memset (alternative_reject, 0, recog_data.n_alternatives * sizeof (int));
399 memset (alternative_nregs, 0, recog_data.n_alternatives * sizeof (int));
401 /* For each operand, find out which regs are equivalent. */
402 for (i = 0; i < recog_data.n_operands; i++)
404 cselib_val *v;
405 struct elt_loc_list *l;
406 rtx op;
408 CLEAR_HARD_REG_SET (equiv_regs[i]);
410 /* cselib blows up on CODE_LABELs. Trying to fix that doesn't seem
411 right, so avoid the problem here. Similarly NOTE_INSN_DELETED_LABEL.
412 Likewise if we have a constant and the insn pattern doesn't tell us
413 the mode we need. */
414 if (LABEL_P (recog_data.operand[i])
415 || (NOTE_P (recog_data.operand[i])
416 && NOTE_KIND (recog_data.operand[i]) == NOTE_INSN_DELETED_LABEL)
417 || (CONSTANT_P (recog_data.operand[i])
418 && recog_data.operand_mode[i] == VOIDmode))
419 continue;
421 op = recog_data.operand[i];
422 if (MEM_P (op) && load_extend_op (GET_MODE (op)) != UNKNOWN)
424 rtx set = single_set (insn);
426 /* We might have multiple sets, some of which do implicit
427 extension. Punt on this for now. */
428 if (! set)
429 continue;
430 /* If the destination is also a MEM or a STRICT_LOW_PART, no
431 extension applies.
432 Also, if there is an explicit extension, we don't have to
433 worry about an implicit one. */
434 else if (MEM_P (SET_DEST (set))
435 || GET_CODE (SET_DEST (set)) == STRICT_LOW_PART
436 || GET_CODE (SET_SRC (set)) == ZERO_EXTEND
437 || GET_CODE (SET_SRC (set)) == SIGN_EXTEND)
438 ; /* Continue ordinary processing. */
439 /* If the register cannot change mode to word_mode, it follows that
440 it cannot have been used in word_mode. */
441 else if (REG_P (SET_DEST (set))
442 && !REG_CAN_CHANGE_MODE_P (REGNO (SET_DEST (set)),
443 GET_MODE (SET_DEST (set)),
444 word_mode))
445 ; /* Continue ordinary processing. */
446 /* If this is a straight load, make the extension explicit. */
447 else if (REG_P (SET_DEST (set))
448 && recog_data.n_operands == 2
449 && SET_SRC (set) == op
450 && SET_DEST (set) == recog_data.operand[1-i])
452 validate_change (insn, recog_data.operand_loc[i],
453 gen_rtx_fmt_e (load_extend_op (GET_MODE (op)),
454 word_mode, op),
456 validate_change (insn, recog_data.operand_loc[1-i],
457 gen_rtx_REG (word_mode, REGNO (SET_DEST (set))),
459 if (! apply_change_group ())
460 return 0;
461 return reload_cse_simplify_operands (insn, testreg);
463 else
464 /* ??? There might be arithmetic operations with memory that are
465 safe to optimize, but is it worth the trouble? */
466 continue;
469 if (side_effects_p (op))
470 continue;
471 v = cselib_lookup (op, recog_data.operand_mode[i], 0, VOIDmode);
472 if (! v)
473 continue;
475 for (l = v->locs; l; l = l->next)
476 if (REG_P (l->loc))
477 SET_HARD_REG_BIT (equiv_regs[i], REGNO (l->loc));
480 alternative_mask preferred = get_preferred_alternatives (insn);
481 for (i = 0; i < recog_data.n_operands; i++)
483 machine_mode mode;
484 int regno;
485 const char *p;
487 op_alt_regno[i] = XALLOCAVEC (int, recog_data.n_alternatives);
488 for (j = 0; j < recog_data.n_alternatives; j++)
489 op_alt_regno[i][j] = -1;
491 p = constraints[i] = recog_data.constraints[i];
492 mode = recog_data.operand_mode[i];
494 /* Add the reject values for each alternative given by the constraints
495 for this operand. */
496 j = 0;
497 while (*p != '\0')
499 char c = *p++;
500 if (c == ',')
501 j++;
502 else if (c == '?')
503 alternative_reject[j] += 3;
504 else if (c == '!')
505 alternative_reject[j] += 300;
508 /* We won't change operands which are already registers. We
509 also don't want to modify output operands. */
510 regno = true_regnum (recog_data.operand[i]);
511 if (regno >= 0
512 || constraints[i][0] == '='
513 || constraints[i][0] == '+')
514 continue;
516 for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
518 enum reg_class rclass = NO_REGS;
520 if (! TEST_HARD_REG_BIT (equiv_regs[i], regno))
521 continue;
523 set_mode_and_regno (testreg, mode, regno);
525 /* We found a register equal to this operand. Now look for all
526 alternatives that can accept this register and have not been
527 assigned a register they can use yet. */
528 j = 0;
529 p = constraints[i];
530 for (;;)
532 char c = *p;
534 switch (c)
536 case 'g':
537 rclass = reg_class_subunion[rclass][GENERAL_REGS];
538 break;
540 default:
541 rclass
542 = (reg_class_subunion
543 [rclass]
544 [reg_class_for_constraint (lookup_constraint (p))]);
545 break;
547 case ',': case '\0':
548 /* See if REGNO fits this alternative, and set it up as the
549 replacement register if we don't have one for this
550 alternative yet and the operand being replaced is not
551 a cheap CONST_INT. */
552 if (op_alt_regno[i][j] == -1
553 && TEST_BIT (preferred, j)
554 && reg_fits_class_p (testreg, rclass, 0, mode)
555 && (!CONST_INT_P (recog_data.operand[i])
556 || (set_src_cost (recog_data.operand[i], mode,
557 optimize_bb_for_speed_p
558 (BLOCK_FOR_INSN (insn)))
559 > set_src_cost (testreg, mode,
560 optimize_bb_for_speed_p
561 (BLOCK_FOR_INSN (insn))))))
563 alternative_nregs[j]++;
564 op_alt_regno[i][j] = regno;
566 j++;
567 rclass = NO_REGS;
568 break;
570 p += CONSTRAINT_LEN (c, p);
572 if (c == '\0')
573 break;
578 /* Record all alternatives which are better or equal to the currently
579 matching one in the alternative_order array. */
580 for (i = j = 0; i < recog_data.n_alternatives; i++)
581 if (alternative_reject[i] <= alternative_reject[which_alternative])
582 alternative_order[j++] = i;
583 recog_data.n_alternatives = j;
585 /* Sort it. Given a small number of alternatives, a dumb algorithm
586 won't hurt too much. */
587 for (i = 0; i < recog_data.n_alternatives - 1; i++)
589 int best = i;
590 int best_reject = alternative_reject[alternative_order[i]];
591 int best_nregs = alternative_nregs[alternative_order[i]];
593 for (j = i + 1; j < recog_data.n_alternatives; j++)
595 int this_reject = alternative_reject[alternative_order[j]];
596 int this_nregs = alternative_nregs[alternative_order[j]];
598 if (this_reject < best_reject
599 || (this_reject == best_reject && this_nregs > best_nregs))
601 best = j;
602 best_reject = this_reject;
603 best_nregs = this_nregs;
607 std::swap (alternative_order[best], alternative_order[i]);
610 /* Substitute the operands as determined by op_alt_regno for the best
611 alternative. */
612 j = alternative_order[0];
614 for (i = 0; i < recog_data.n_operands; i++)
616 machine_mode mode = recog_data.operand_mode[i];
617 if (op_alt_regno[i][j] == -1)
618 continue;
620 validate_change (insn, recog_data.operand_loc[i],
621 gen_rtx_REG (mode, op_alt_regno[i][j]), 1);
624 for (i = recog_data.n_dups - 1; i >= 0; i--)
626 int op = recog_data.dup_num[i];
627 machine_mode mode = recog_data.operand_mode[op];
629 if (op_alt_regno[op][j] == -1)
630 continue;
632 validate_change (insn, recog_data.dup_loc[i],
633 gen_rtx_REG (mode, op_alt_regno[op][j]), 1);
636 return apply_change_group ();
639 /* If reload couldn't use reg+reg+offset addressing, try to use reg+reg
640 addressing now.
641 This code might also be useful when reload gave up on reg+reg addressing
642 because of clashes between the return register and INDEX_REG_CLASS. */
644 /* The maximum number of uses of a register we can keep track of to
645 replace them with reg+reg addressing. */
646 #define RELOAD_COMBINE_MAX_USES 16
648 /* Describes a recorded use of a register. */
649 struct reg_use
651 /* The insn where a register has been used. */
652 rtx_insn *insn;
653 /* Points to the memory reference enclosing the use, if any, NULL_RTX
654 otherwise. */
655 rtx containing_mem;
656 /* Location of the register within INSN. */
657 rtx *usep;
658 /* The reverse uid of the insn. */
659 int ruid;
662 /* If the register is used in some unknown fashion, USE_INDEX is negative.
663 If it is dead, USE_INDEX is RELOAD_COMBINE_MAX_USES, and STORE_RUID
664 indicates where it is first set or clobbered.
665 Otherwise, USE_INDEX is the index of the last encountered use of the
666 register (which is first among these we have seen since we scan backwards).
667 USE_RUID indicates the first encountered, i.e. last, of these uses.
668 If ALL_OFFSETS_MATCH is true, all encountered uses were inside a PLUS
669 with a constant offset; OFFSET contains this constant in that case.
670 STORE_RUID is always meaningful if we only want to use a value in a
671 register in a different place: it denotes the next insn in the insn
672 stream (i.e. the last encountered) that sets or clobbers the register.
673 REAL_STORE_RUID is similar, but clobbers are ignored when updating it.
674 EXPR is the expression used when storing the register. */
675 static struct
677 struct reg_use reg_use[RELOAD_COMBINE_MAX_USES];
678 rtx offset;
679 int use_index;
680 int store_ruid;
681 int real_store_ruid;
682 int use_ruid;
683 bool all_offsets_match;
684 rtx expr;
685 } reg_state[FIRST_PSEUDO_REGISTER];
687 /* Reverse linear uid. This is increased in reload_combine while scanning
688 the instructions from last to first. It is used to set last_label_ruid
689 and the store_ruid / use_ruid fields in reg_state. */
690 static int reload_combine_ruid;
692 /* The RUID of the last label we encountered in reload_combine. */
693 static int last_label_ruid;
695 /* The RUID of the last jump we encountered in reload_combine. */
696 static int last_jump_ruid;
698 /* The register numbers of the first and last index register. A value of
699 -1 in LAST_INDEX_REG indicates that we've previously computed these
700 values and found no suitable index registers. */
701 static int first_index_reg = -1;
702 static int last_index_reg;
704 #define LABEL_LIVE(LABEL) \
705 (label_live[CODE_LABEL_NUMBER (LABEL) - min_labelno])
707 /* Subroutine of reload_combine_split_ruids, called to fix up a single
708 ruid pointed to by *PRUID if it is higher than SPLIT_RUID. */
710 static inline void
711 reload_combine_split_one_ruid (int *pruid, int split_ruid)
713 if (*pruid > split_ruid)
714 (*pruid)++;
717 /* Called when we insert a new insn in a position we've already passed in
718 the scan. Examine all our state, increasing all ruids that are higher
719 than SPLIT_RUID by one in order to make room for a new insn. */
721 static void
722 reload_combine_split_ruids (int split_ruid)
724 unsigned i;
726 reload_combine_split_one_ruid (&reload_combine_ruid, split_ruid);
727 reload_combine_split_one_ruid (&last_label_ruid, split_ruid);
728 reload_combine_split_one_ruid (&last_jump_ruid, split_ruid);
730 for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
732 int j, idx = reg_state[i].use_index;
733 reload_combine_split_one_ruid (&reg_state[i].use_ruid, split_ruid);
734 reload_combine_split_one_ruid (&reg_state[i].store_ruid, split_ruid);
735 reload_combine_split_one_ruid (&reg_state[i].real_store_ruid,
736 split_ruid);
737 if (idx < 0)
738 continue;
739 for (j = idx; j < RELOAD_COMBINE_MAX_USES; j++)
741 reload_combine_split_one_ruid (&reg_state[i].reg_use[j].ruid,
742 split_ruid);
747 /* Called when we are about to rescan a previously encountered insn with
748 reload_combine_note_use after modifying some part of it. This clears all
749 information about uses in that particular insn. */
751 static void
752 reload_combine_purge_insn_uses (rtx_insn *insn)
754 unsigned i;
756 for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
758 int j, k, idx = reg_state[i].use_index;
759 if (idx < 0)
760 continue;
761 j = k = RELOAD_COMBINE_MAX_USES;
762 while (j-- > idx)
764 if (reg_state[i].reg_use[j].insn != insn)
766 k--;
767 if (k != j)
768 reg_state[i].reg_use[k] = reg_state[i].reg_use[j];
771 reg_state[i].use_index = k;
775 /* Called when we need to forget about all uses of REGNO after an insn
776 which is identified by RUID. */
778 static void
779 reload_combine_purge_reg_uses_after_ruid (unsigned regno, int ruid)
781 int j, k, idx = reg_state[regno].use_index;
782 if (idx < 0)
783 return;
784 j = k = RELOAD_COMBINE_MAX_USES;
785 while (j-- > idx)
787 if (reg_state[regno].reg_use[j].ruid >= ruid)
789 k--;
790 if (k != j)
791 reg_state[regno].reg_use[k] = reg_state[regno].reg_use[j];
794 reg_state[regno].use_index = k;
797 /* Find the use of REGNO with the ruid that is highest among those
798 lower than RUID_LIMIT, and return it if it is the only use of this
799 reg in the insn. Return NULL otherwise. */
801 static struct reg_use *
802 reload_combine_closest_single_use (unsigned regno, int ruid_limit)
804 int i, best_ruid = 0;
805 int use_idx = reg_state[regno].use_index;
806 struct reg_use *retval;
808 if (use_idx < 0)
809 return NULL;
810 retval = NULL;
811 for (i = use_idx; i < RELOAD_COMBINE_MAX_USES; i++)
813 struct reg_use *use = reg_state[regno].reg_use + i;
814 int this_ruid = use->ruid;
815 if (this_ruid >= ruid_limit)
816 continue;
817 if (this_ruid > best_ruid)
819 best_ruid = this_ruid;
820 retval = use;
822 else if (this_ruid == best_ruid)
823 retval = NULL;
825 if (last_label_ruid >= best_ruid)
826 return NULL;
827 return retval;
830 /* After we've moved an add insn, fix up any debug insns that occur
831 between the old location of the add and the new location. REG is
832 the destination register of the add insn; REPLACEMENT is the
833 SET_SRC of the add. FROM and TO specify the range in which we
834 should make this change on debug insns. */
836 static void
837 fixup_debug_insns (rtx reg, rtx replacement, rtx_insn *from, rtx_insn *to)
839 rtx_insn *insn;
840 for (insn = from; insn != to; insn = NEXT_INSN (insn))
842 rtx t;
844 if (!DEBUG_BIND_INSN_P (insn))
845 continue;
847 t = INSN_VAR_LOCATION_LOC (insn);
848 t = simplify_replace_rtx (t, reg, replacement);
849 validate_change (insn, &INSN_VAR_LOCATION_LOC (insn), t, 0);
853 /* Subroutine of reload_combine_recognize_const_pattern. Try to replace REG
854 with SRC in the insn described by USE, taking costs into account. Return
855 true if we made the replacement. */
857 static bool
858 try_replace_in_use (struct reg_use *use, rtx reg, rtx src)
860 rtx_insn *use_insn = use->insn;
861 rtx mem = use->containing_mem;
862 bool speed = optimize_bb_for_speed_p (BLOCK_FOR_INSN (use_insn));
864 if (mem != NULL_RTX)
866 addr_space_t as = MEM_ADDR_SPACE (mem);
867 rtx oldaddr = XEXP (mem, 0);
868 rtx newaddr = NULL_RTX;
869 int old_cost = address_cost (oldaddr, GET_MODE (mem), as, speed);
870 int new_cost;
872 newaddr = simplify_replace_rtx (oldaddr, reg, src);
873 if (memory_address_addr_space_p (GET_MODE (mem), newaddr, as))
875 XEXP (mem, 0) = newaddr;
876 new_cost = address_cost (newaddr, GET_MODE (mem), as, speed);
877 XEXP (mem, 0) = oldaddr;
878 if (new_cost <= old_cost
879 && validate_change (use_insn,
880 &XEXP (mem, 0), newaddr, 0))
881 return true;
884 else
886 rtx new_set = single_set (use_insn);
887 if (new_set
888 && REG_P (SET_DEST (new_set))
889 && GET_CODE (SET_SRC (new_set)) == PLUS
890 && REG_P (XEXP (SET_SRC (new_set), 0))
891 && CONSTANT_P (XEXP (SET_SRC (new_set), 1)))
893 rtx new_src;
894 machine_mode mode = GET_MODE (SET_DEST (new_set));
895 int old_cost = set_src_cost (SET_SRC (new_set), mode, speed);
897 gcc_assert (rtx_equal_p (XEXP (SET_SRC (new_set), 0), reg));
898 new_src = simplify_replace_rtx (SET_SRC (new_set), reg, src);
900 if (set_src_cost (new_src, mode, speed) <= old_cost
901 && validate_change (use_insn, &SET_SRC (new_set),
902 new_src, 0))
903 return true;
906 return false;
909 /* Called by reload_combine when scanning INSN. This function tries to detect
910 patterns where a constant is added to a register, and the result is used
911 in an address.
912 Return true if no further processing is needed on INSN; false if it wasn't
913 recognized and should be handled normally. */
915 static bool
916 reload_combine_recognize_const_pattern (rtx_insn *insn)
918 int from_ruid = reload_combine_ruid;
919 rtx set, pat, reg, src, addreg;
920 unsigned int regno;
921 struct reg_use *use;
922 bool must_move_add;
923 rtx_insn *add_moved_after_insn = NULL;
924 int add_moved_after_ruid = 0;
925 int clobbered_regno = -1;
927 set = single_set (insn);
928 if (set == NULL_RTX)
929 return false;
931 reg = SET_DEST (set);
932 src = SET_SRC (set);
933 if (!REG_P (reg)
934 || REG_NREGS (reg) != 1
935 || GET_MODE (reg) != Pmode
936 || reg == stack_pointer_rtx)
937 return false;
939 regno = REGNO (reg);
941 /* We look for a REG1 = REG2 + CONSTANT insn, followed by either
942 uses of REG1 inside an address, or inside another add insn. If
943 possible and profitable, merge the addition into subsequent
944 uses. */
945 if (GET_CODE (src) != PLUS
946 || !REG_P (XEXP (src, 0))
947 || !CONSTANT_P (XEXP (src, 1)))
948 return false;
950 addreg = XEXP (src, 0);
951 must_move_add = rtx_equal_p (reg, addreg);
953 pat = PATTERN (insn);
954 if (must_move_add && set != pat)
956 /* We have to be careful when moving the add; apart from the
957 single_set there may also be clobbers. Recognize one special
958 case, that of one clobber alongside the set (likely a clobber
959 of the CC register). */
960 gcc_assert (GET_CODE (PATTERN (insn)) == PARALLEL);
961 if (XVECLEN (pat, 0) != 2 || XVECEXP (pat, 0, 0) != set
962 || GET_CODE (XVECEXP (pat, 0, 1)) != CLOBBER
963 || !REG_P (XEXP (XVECEXP (pat, 0, 1), 0)))
964 return false;
965 clobbered_regno = REGNO (XEXP (XVECEXP (pat, 0, 1), 0));
970 use = reload_combine_closest_single_use (regno, from_ruid);
972 if (use)
973 /* Start the search for the next use from here. */
974 from_ruid = use->ruid;
976 if (use && GET_MODE (*use->usep) == Pmode)
978 bool delete_add = false;
979 rtx_insn *use_insn = use->insn;
980 int use_ruid = use->ruid;
982 /* Avoid moving the add insn past a jump. */
983 if (must_move_add && use_ruid <= last_jump_ruid)
984 break;
986 /* If the add clobbers another hard reg in parallel, don't move
987 it past a real set of this hard reg. */
988 if (must_move_add && clobbered_regno >= 0
989 && reg_state[clobbered_regno].real_store_ruid >= use_ruid)
990 break;
992 /* Do not separate cc0 setter and cc0 user on HAVE_cc0 targets. */
993 if (HAVE_cc0 && must_move_add && sets_cc0_p (PATTERN (use_insn)))
994 break;
996 gcc_assert (reg_state[regno].store_ruid <= use_ruid);
997 /* Avoid moving a use of ADDREG past a point where it is stored. */
998 if (reg_state[REGNO (addreg)].store_ruid > use_ruid)
999 break;
1001 /* We also must not move the addition past an insn that sets
1002 the same register, unless we can combine two add insns. */
1003 if (must_move_add && reg_state[regno].store_ruid == use_ruid)
1005 if (use->containing_mem == NULL_RTX)
1006 delete_add = true;
1007 else
1008 break;
1011 if (try_replace_in_use (use, reg, src))
1013 reload_combine_purge_insn_uses (use_insn);
1014 reload_combine_note_use (&PATTERN (use_insn), use_insn,
1015 use_ruid, NULL_RTX);
1017 if (delete_add)
1019 fixup_debug_insns (reg, src, insn, use_insn);
1020 delete_insn (insn);
1021 return true;
1023 if (must_move_add)
1025 add_moved_after_insn = use_insn;
1026 add_moved_after_ruid = use_ruid;
1028 continue;
1031 /* If we get here, we couldn't handle this use. */
1032 if (must_move_add)
1033 break;
1035 while (use);
1037 if (!must_move_add || add_moved_after_insn == NULL_RTX)
1038 /* Process the add normally. */
1039 return false;
1041 fixup_debug_insns (reg, src, insn, add_moved_after_insn);
1043 reorder_insns (insn, insn, add_moved_after_insn);
1044 reload_combine_purge_reg_uses_after_ruid (regno, add_moved_after_ruid);
1045 reload_combine_split_ruids (add_moved_after_ruid - 1);
1046 reload_combine_note_use (&PATTERN (insn), insn,
1047 add_moved_after_ruid, NULL_RTX);
1048 reg_state[regno].store_ruid = add_moved_after_ruid;
1050 return true;
1053 /* Called by reload_combine when scanning INSN. Try to detect a pattern we
1054 can handle and improve. Return true if no further processing is needed on
1055 INSN; false if it wasn't recognized and should be handled normally. */
1057 static bool
1058 reload_combine_recognize_pattern (rtx_insn *insn)
1060 rtx set, reg, src;
1062 set = single_set (insn);
1063 if (set == NULL_RTX)
1064 return false;
1066 reg = SET_DEST (set);
1067 src = SET_SRC (set);
1068 if (!REG_P (reg) || REG_NREGS (reg) != 1)
1069 return false;
1071 unsigned int regno = REGNO (reg);
1072 machine_mode mode = GET_MODE (reg);
1074 if (reg_state[regno].use_index < 0
1075 || reg_state[regno].use_index >= RELOAD_COMBINE_MAX_USES)
1076 return false;
1078 for (int i = reg_state[regno].use_index;
1079 i < RELOAD_COMBINE_MAX_USES; i++)
1081 struct reg_use *use = reg_state[regno].reg_use + i;
1082 if (GET_MODE (*use->usep) != mode)
1083 return false;
1086 /* Look for (set (REGX) (CONST_INT))
1087 (set (REGX) (PLUS (REGX) (REGY)))
1089 ... (MEM (REGX)) ...
1090 and convert it to
1091 (set (REGZ) (CONST_INT))
1093 ... (MEM (PLUS (REGZ) (REGY)))... .
1095 First, check that we have (set (REGX) (PLUS (REGX) (REGY)))
1096 and that we know all uses of REGX before it dies.
1097 Also, explicitly check that REGX != REGY; our life information
1098 does not yet show whether REGY changes in this insn. */
1100 if (GET_CODE (src) == PLUS
1101 && reg_state[regno].all_offsets_match
1102 && last_index_reg != -1
1103 && REG_P (XEXP (src, 1))
1104 && rtx_equal_p (XEXP (src, 0), reg)
1105 && !rtx_equal_p (XEXP (src, 1), reg)
1106 && last_label_ruid < reg_state[regno].use_ruid)
1108 rtx base = XEXP (src, 1);
1109 rtx_insn *prev = prev_nonnote_nondebug_insn (insn);
1110 rtx prev_set = prev ? single_set (prev) : NULL_RTX;
1111 rtx index_reg = NULL_RTX;
1112 rtx reg_sum = NULL_RTX;
1113 int i;
1115 /* Now we need to set INDEX_REG to an index register (denoted as
1116 REGZ in the illustration above) and REG_SUM to the expression
1117 register+register that we want to use to substitute uses of REG
1118 (typically in MEMs) with. First check REG and BASE for being
1119 index registers; we can use them even if they are not dead. */
1120 if (TEST_HARD_REG_BIT (reg_class_contents[INDEX_REG_CLASS], regno)
1121 || TEST_HARD_REG_BIT (reg_class_contents[INDEX_REG_CLASS],
1122 REGNO (base)))
1124 index_reg = reg;
1125 reg_sum = src;
1127 else
1129 /* Otherwise, look for a free index register. Since we have
1130 checked above that neither REG nor BASE are index registers,
1131 if we find anything at all, it will be different from these
1132 two registers. */
1133 for (i = first_index_reg; i <= last_index_reg; i++)
1135 if (TEST_HARD_REG_BIT (reg_class_contents[INDEX_REG_CLASS], i)
1136 && reg_state[i].use_index == RELOAD_COMBINE_MAX_USES
1137 && reg_state[i].store_ruid <= reg_state[regno].use_ruid
1138 && (call_used_regs[i] || df_regs_ever_live_p (i))
1139 && (!frame_pointer_needed || i != HARD_FRAME_POINTER_REGNUM)
1140 && !fixed_regs[i] && !global_regs[i]
1141 && hard_regno_nregs (i, GET_MODE (reg)) == 1
1142 && targetm.hard_regno_scratch_ok (i))
1144 index_reg = gen_rtx_REG (GET_MODE (reg), i);
1145 reg_sum = gen_rtx_PLUS (GET_MODE (reg), index_reg, base);
1146 break;
1151 /* Check that PREV_SET is indeed (set (REGX) (CONST_INT)) and that
1152 (REGY), i.e. BASE, is not clobbered before the last use we'll
1153 create. */
1154 if (reg_sum
1155 && prev_set
1156 && CONST_INT_P (SET_SRC (prev_set))
1157 && rtx_equal_p (SET_DEST (prev_set), reg)
1158 && (reg_state[REGNO (base)].store_ruid
1159 <= reg_state[regno].use_ruid))
1161 /* Change destination register and, if necessary, the constant
1162 value in PREV, the constant loading instruction. */
1163 validate_change (prev, &SET_DEST (prev_set), index_reg, 1);
1164 if (reg_state[regno].offset != const0_rtx)
1166 HOST_WIDE_INT c
1167 = trunc_int_for_mode (UINTVAL (SET_SRC (prev_set))
1168 + UINTVAL (reg_state[regno].offset),
1169 GET_MODE (index_reg));
1170 validate_change (prev, &SET_SRC (prev_set), GEN_INT (c), 1);
1173 /* Now for every use of REG that we have recorded, replace REG
1174 with REG_SUM. */
1175 for (i = reg_state[regno].use_index;
1176 i < RELOAD_COMBINE_MAX_USES; i++)
1177 validate_unshare_change (reg_state[regno].reg_use[i].insn,
1178 reg_state[regno].reg_use[i].usep,
1179 /* Each change must have its own
1180 replacement. */
1181 reg_sum, 1);
1183 if (apply_change_group ())
1185 struct reg_use *lowest_ruid = NULL;
1187 /* For every new use of REG_SUM, we have to record the use
1188 of BASE therein, i.e. operand 1. */
1189 for (i = reg_state[regno].use_index;
1190 i < RELOAD_COMBINE_MAX_USES; i++)
1192 struct reg_use *use = reg_state[regno].reg_use + i;
1193 reload_combine_note_use (&XEXP (*use->usep, 1), use->insn,
1194 use->ruid, use->containing_mem);
1195 if (lowest_ruid == NULL || use->ruid < lowest_ruid->ruid)
1196 lowest_ruid = use;
1199 fixup_debug_insns (reg, reg_sum, insn, lowest_ruid->insn);
1201 /* Delete the reg-reg addition. */
1202 delete_insn (insn);
1204 if (reg_state[regno].offset != const0_rtx
1205 /* Previous REG_EQUIV / REG_EQUAL notes for PREV
1206 are now invalid. */
1207 && remove_reg_equal_equiv_notes (prev))
1208 df_notes_rescan (prev);
1210 reg_state[regno].use_index = RELOAD_COMBINE_MAX_USES;
1211 return true;
1215 return false;
1218 static void
1219 reload_combine (void)
1221 rtx_insn *insn, *prev;
1222 basic_block bb;
1223 unsigned int r;
1224 int min_labelno, n_labels;
1225 HARD_REG_SET ever_live_at_start, *label_live;
1227 /* To avoid wasting too much time later searching for an index register,
1228 determine the minimum and maximum index register numbers. */
1229 if (INDEX_REG_CLASS == NO_REGS)
1230 last_index_reg = -1;
1231 else if (first_index_reg == -1 && last_index_reg == 0)
1233 for (r = 0; r < FIRST_PSEUDO_REGISTER; r++)
1234 if (TEST_HARD_REG_BIT (reg_class_contents[INDEX_REG_CLASS], r))
1236 if (first_index_reg == -1)
1237 first_index_reg = r;
1239 last_index_reg = r;
1242 /* If no index register is available, we can quit now. Set LAST_INDEX_REG
1243 to -1 so we'll know to quit early the next time we get here. */
1244 if (first_index_reg == -1)
1246 last_index_reg = -1;
1247 return;
1251 /* Set up LABEL_LIVE and EVER_LIVE_AT_START. The register lifetime
1252 information is a bit fuzzy immediately after reload, but it's
1253 still good enough to determine which registers are live at a jump
1254 destination. */
1255 min_labelno = get_first_label_num ();
1256 n_labels = max_label_num () - min_labelno;
1257 label_live = XNEWVEC (HARD_REG_SET, n_labels);
1258 CLEAR_HARD_REG_SET (ever_live_at_start);
1260 FOR_EACH_BB_REVERSE_FN (bb, cfun)
1262 insn = BB_HEAD (bb);
1263 if (LABEL_P (insn))
1265 HARD_REG_SET live;
1266 bitmap live_in = df_get_live_in (bb);
1268 REG_SET_TO_HARD_REG_SET (live, live_in);
1269 compute_use_by_pseudos (&live, live_in);
1270 COPY_HARD_REG_SET (LABEL_LIVE (insn), live);
1271 IOR_HARD_REG_SET (ever_live_at_start, live);
1275 /* Initialize last_label_ruid, reload_combine_ruid and reg_state. */
1276 last_label_ruid = last_jump_ruid = reload_combine_ruid = 0;
1277 for (r = 0; r < FIRST_PSEUDO_REGISTER; r++)
1279 reg_state[r].store_ruid = 0;
1280 reg_state[r].real_store_ruid = 0;
1281 if (fixed_regs[r])
1282 reg_state[r].use_index = -1;
1283 else
1284 reg_state[r].use_index = RELOAD_COMBINE_MAX_USES;
1287 for (insn = get_last_insn (); insn; insn = prev)
1289 bool control_flow_insn;
1290 rtx note;
1292 prev = PREV_INSN (insn);
1294 /* We cannot do our optimization across labels. Invalidating all the use
1295 information we have would be costly, so we just note where the label
1296 is and then later disable any optimization that would cross it. */
1297 if (LABEL_P (insn))
1298 last_label_ruid = reload_combine_ruid;
1299 else if (BARRIER_P (insn))
1301 /* Crossing a barrier resets all the use information. */
1302 for (r = 0; r < FIRST_PSEUDO_REGISTER; r++)
1303 if (! fixed_regs[r])
1304 reg_state[r].use_index = RELOAD_COMBINE_MAX_USES;
1306 else if (INSN_P (insn) && volatile_insn_p (PATTERN (insn)))
1307 /* Optimizations across insns being marked as volatile must be
1308 prevented. All the usage information is invalidated
1309 here. */
1310 for (r = 0; r < FIRST_PSEUDO_REGISTER; r++)
1311 if (! fixed_regs[r]
1312 && reg_state[r].use_index != RELOAD_COMBINE_MAX_USES)
1313 reg_state[r].use_index = -1;
1315 if (! NONDEBUG_INSN_P (insn))
1316 continue;
1318 reload_combine_ruid++;
1320 control_flow_insn = control_flow_insn_p (insn);
1321 if (control_flow_insn)
1322 last_jump_ruid = reload_combine_ruid;
1324 if (reload_combine_recognize_const_pattern (insn)
1325 || reload_combine_recognize_pattern (insn))
1326 continue;
1328 note_stores (PATTERN (insn), reload_combine_note_store, NULL);
1330 if (CALL_P (insn))
1332 rtx link;
1333 HARD_REG_SET used_regs;
1335 get_call_reg_set_usage (insn, &used_regs, call_used_reg_set);
1337 for (r = 0; r < FIRST_PSEUDO_REGISTER; r++)
1338 if (TEST_HARD_REG_BIT (used_regs, r))
1340 reg_state[r].use_index = RELOAD_COMBINE_MAX_USES;
1341 reg_state[r].store_ruid = reload_combine_ruid;
1344 for (link = CALL_INSN_FUNCTION_USAGE (insn); link;
1345 link = XEXP (link, 1))
1347 rtx setuse = XEXP (link, 0);
1348 rtx usage_rtx = XEXP (setuse, 0);
1349 /* We could support CLOBBER_HIGH and treat it in the same way as
1350 HARD_REGNO_CALL_PART_CLOBBERED, but no port needs that yet. */
1351 gcc_assert (GET_CODE (setuse) != CLOBBER_HIGH);
1353 if ((GET_CODE (setuse) == USE || GET_CODE (setuse) == CLOBBER)
1354 && REG_P (usage_rtx))
1356 unsigned int end_regno = END_REGNO (usage_rtx);
1357 for (unsigned int i = REGNO (usage_rtx); i < end_regno; ++i)
1358 if (GET_CODE (XEXP (link, 0)) == CLOBBER)
1360 reg_state[i].use_index = RELOAD_COMBINE_MAX_USES;
1361 reg_state[i].store_ruid = reload_combine_ruid;
1363 else
1364 reg_state[i].use_index = -1;
1369 if (control_flow_insn && !ANY_RETURN_P (PATTERN (insn)))
1371 /* Non-spill registers might be used at the call destination in
1372 some unknown fashion, so we have to mark the unknown use. */
1373 HARD_REG_SET *live;
1375 if ((condjump_p (insn) || condjump_in_parallel_p (insn))
1376 && JUMP_LABEL (insn))
1378 if (ANY_RETURN_P (JUMP_LABEL (insn)))
1379 live = NULL;
1380 else
1381 live = &LABEL_LIVE (JUMP_LABEL (insn));
1383 else
1384 live = &ever_live_at_start;
1386 if (live)
1387 for (r = 0; r < FIRST_PSEUDO_REGISTER; r++)
1388 if (TEST_HARD_REG_BIT (*live, r))
1389 reg_state[r].use_index = -1;
1392 reload_combine_note_use (&PATTERN (insn), insn, reload_combine_ruid,
1393 NULL_RTX);
1395 for (note = REG_NOTES (insn); note; note = XEXP (note, 1))
1397 if (REG_NOTE_KIND (note) == REG_INC && REG_P (XEXP (note, 0)))
1399 int regno = REGNO (XEXP (note, 0));
1400 reg_state[regno].store_ruid = reload_combine_ruid;
1401 reg_state[regno].real_store_ruid = reload_combine_ruid;
1402 reg_state[regno].use_index = -1;
1407 free (label_live);
1410 /* Check if DST is a register or a subreg of a register; if it is,
1411 update store_ruid, real_store_ruid and use_index in the reg_state
1412 structure accordingly. Called via note_stores from reload_combine. */
1414 static void
1415 reload_combine_note_store (rtx dst, const_rtx set, void *data ATTRIBUTE_UNUSED)
1417 int regno = 0;
1418 int i;
1419 machine_mode mode = GET_MODE (dst);
1421 if (GET_CODE (dst) == SUBREG)
1423 regno = subreg_regno_offset (REGNO (SUBREG_REG (dst)),
1424 GET_MODE (SUBREG_REG (dst)),
1425 SUBREG_BYTE (dst),
1426 GET_MODE (dst));
1427 dst = SUBREG_REG (dst);
1430 /* Some targets do argument pushes without adding REG_INC notes. */
1432 if (MEM_P (dst))
1434 dst = XEXP (dst, 0);
1435 if (GET_CODE (dst) == PRE_INC || GET_CODE (dst) == POST_INC
1436 || GET_CODE (dst) == PRE_DEC || GET_CODE (dst) == POST_DEC
1437 || GET_CODE (dst) == PRE_MODIFY || GET_CODE (dst) == POST_MODIFY)
1439 unsigned int end_regno = END_REGNO (XEXP (dst, 0));
1440 for (unsigned int i = REGNO (XEXP (dst, 0)); i < end_regno; ++i)
1442 /* We could probably do better, but for now mark the register
1443 as used in an unknown fashion and set/clobbered at this
1444 insn. */
1445 reg_state[i].use_index = -1;
1446 reg_state[i].store_ruid = reload_combine_ruid;
1447 reg_state[i].real_store_ruid = reload_combine_ruid;
1450 else
1451 return;
1454 if (!REG_P (dst))
1455 return;
1456 regno += REGNO (dst);
1458 /* note_stores might have stripped a STRICT_LOW_PART, so we have to be
1459 careful with registers / register parts that are not full words.
1460 Similarly for ZERO_EXTRACT. */
1461 if (GET_CODE (SET_DEST (set)) == ZERO_EXTRACT
1462 || GET_CODE (SET_DEST (set)) == STRICT_LOW_PART)
1464 for (i = end_hard_regno (mode, regno) - 1; i >= regno; i--)
1466 reg_state[i].use_index = -1;
1467 reg_state[i].store_ruid = reload_combine_ruid;
1468 reg_state[i].real_store_ruid = reload_combine_ruid;
1471 else
1473 for (i = end_hard_regno (mode, regno) - 1; i >= regno; i--)
1475 reg_state[i].store_ruid = reload_combine_ruid;
1476 if (GET_CODE (set) == SET)
1477 reg_state[i].real_store_ruid = reload_combine_ruid;
1478 reg_state[i].use_index = RELOAD_COMBINE_MAX_USES;
1483 /* XP points to a piece of rtl that has to be checked for any uses of
1484 registers.
1485 *XP is the pattern of INSN, or a part of it.
1486 Called from reload_combine, and recursively by itself. */
1487 static void
1488 reload_combine_note_use (rtx *xp, rtx_insn *insn, int ruid, rtx containing_mem)
1490 rtx x = *xp;
1491 enum rtx_code code = x->code;
1492 const char *fmt;
1493 int i, j;
1494 rtx offset = const0_rtx; /* For the REG case below. */
1496 switch (code)
1498 case SET:
1499 if (REG_P (SET_DEST (x)))
1501 reload_combine_note_use (&SET_SRC (x), insn, ruid, NULL_RTX);
1502 return;
1504 break;
1506 case USE:
1507 /* If this is the USE of a return value, we can't change it. */
1508 if (REG_P (XEXP (x, 0)) && REG_FUNCTION_VALUE_P (XEXP (x, 0)))
1510 /* Mark the return register as used in an unknown fashion. */
1511 rtx reg = XEXP (x, 0);
1512 unsigned int end_regno = END_REGNO (reg);
1513 for (unsigned int regno = REGNO (reg); regno < end_regno; ++regno)
1514 reg_state[regno].use_index = -1;
1515 return;
1517 break;
1519 case CLOBBER:
1520 if (REG_P (SET_DEST (x)))
1522 /* No spurious CLOBBERs of pseudo registers may remain. */
1523 gcc_assert (REGNO (SET_DEST (x)) < FIRST_PSEUDO_REGISTER);
1524 return;
1526 break;
1528 case CLOBBER_HIGH:
1529 gcc_assert (REG_P (SET_DEST (x)));
1530 return;
1532 case PLUS:
1533 /* We are interested in (plus (reg) (const_int)) . */
1534 if (!REG_P (XEXP (x, 0))
1535 || !CONST_INT_P (XEXP (x, 1)))
1536 break;
1537 offset = XEXP (x, 1);
1538 x = XEXP (x, 0);
1539 /* Fall through. */
1540 case REG:
1542 int regno = REGNO (x);
1543 int use_index;
1544 int nregs;
1546 /* No spurious USEs of pseudo registers may remain. */
1547 gcc_assert (regno < FIRST_PSEUDO_REGISTER);
1549 nregs = REG_NREGS (x);
1551 /* We can't substitute into multi-hard-reg uses. */
1552 if (nregs > 1)
1554 while (--nregs >= 0)
1555 reg_state[regno + nregs].use_index = -1;
1556 return;
1559 /* We may be called to update uses in previously seen insns.
1560 Don't add uses beyond the last store we saw. */
1561 if (ruid < reg_state[regno].store_ruid)
1562 return;
1564 /* If this register is already used in some unknown fashion, we
1565 can't do anything.
1566 If we decrement the index from zero to -1, we can't store more
1567 uses, so this register becomes used in an unknown fashion. */
1568 use_index = --reg_state[regno].use_index;
1569 if (use_index < 0)
1570 return;
1572 if (use_index == RELOAD_COMBINE_MAX_USES - 1)
1574 /* This is the first use of this register we have seen since we
1575 marked it as dead. */
1576 reg_state[regno].offset = offset;
1577 reg_state[regno].all_offsets_match = true;
1578 reg_state[regno].use_ruid = ruid;
1580 else
1582 if (reg_state[regno].use_ruid > ruid)
1583 reg_state[regno].use_ruid = ruid;
1585 if (! rtx_equal_p (offset, reg_state[regno].offset))
1586 reg_state[regno].all_offsets_match = false;
1589 reg_state[regno].reg_use[use_index].insn = insn;
1590 reg_state[regno].reg_use[use_index].ruid = ruid;
1591 reg_state[regno].reg_use[use_index].containing_mem = containing_mem;
1592 reg_state[regno].reg_use[use_index].usep = xp;
1593 return;
1596 case MEM:
1597 containing_mem = x;
1598 break;
1600 default:
1601 break;
1604 /* Recursively process the components of X. */
1605 fmt = GET_RTX_FORMAT (code);
1606 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
1608 if (fmt[i] == 'e')
1609 reload_combine_note_use (&XEXP (x, i), insn, ruid, containing_mem);
1610 else if (fmt[i] == 'E')
1612 for (j = XVECLEN (x, i) - 1; j >= 0; j--)
1613 reload_combine_note_use (&XVECEXP (x, i, j), insn, ruid,
1614 containing_mem);
1619 /* See if we can reduce the cost of a constant by replacing a move
1620 with an add. We track situations in which a register is set to a
1621 constant or to a register plus a constant. */
1622 /* We cannot do our optimization across labels. Invalidating all the
1623 information about register contents we have would be costly, so we
1624 use move2add_last_label_luid to note where the label is and then
1625 later disable any optimization that would cross it.
1626 reg_offset[n] / reg_base_reg[n] / reg_symbol_ref[n] / reg_mode[n]
1627 are only valid if reg_set_luid[n] is greater than
1628 move2add_last_label_luid.
1629 For a set that established a new (potential) base register with
1630 non-constant value, we use move2add_luid from the place where the
1631 setting insn is encountered; registers based off that base then
1632 get the same reg_set_luid. Constants all get
1633 move2add_last_label_luid + 1 as their reg_set_luid. */
1634 static int reg_set_luid[FIRST_PSEUDO_REGISTER];
1636 /* If reg_base_reg[n] is negative, register n has been set to
1637 reg_offset[n] or reg_symbol_ref[n] + reg_offset[n] in mode reg_mode[n].
1638 If reg_base_reg[n] is non-negative, register n has been set to the
1639 sum of reg_offset[n] and the value of register reg_base_reg[n]
1640 before reg_set_luid[n], calculated in mode reg_mode[n] .
1641 For multi-hard-register registers, all but the first one are
1642 recorded as BLKmode in reg_mode. Setting reg_mode to VOIDmode
1643 marks it as invalid. */
1644 static HOST_WIDE_INT reg_offset[FIRST_PSEUDO_REGISTER];
1645 static int reg_base_reg[FIRST_PSEUDO_REGISTER];
1646 static rtx reg_symbol_ref[FIRST_PSEUDO_REGISTER];
1647 static machine_mode reg_mode[FIRST_PSEUDO_REGISTER];
1649 /* move2add_luid is linearly increased while scanning the instructions
1650 from first to last. It is used to set reg_set_luid in
1651 reload_cse_move2add and move2add_note_store. */
1652 static int move2add_luid;
1654 /* move2add_last_label_luid is set whenever a label is found. Labels
1655 invalidate all previously collected reg_offset data. */
1656 static int move2add_last_label_luid;
1658 /* ??? We don't know how zero / sign extension is handled, hence we
1659 can't go from a narrower to a wider mode. */
1660 #define MODES_OK_FOR_MOVE2ADD(OUTMODE, INMODE) \
1661 (GET_MODE_SIZE (OUTMODE) == GET_MODE_SIZE (INMODE) \
1662 || (GET_MODE_SIZE (OUTMODE) <= GET_MODE_SIZE (INMODE) \
1663 && TRULY_NOOP_TRUNCATION_MODES_P (OUTMODE, INMODE)))
1665 /* Record that REG is being set to a value with the mode of REG. */
1667 static void
1668 move2add_record_mode (rtx reg)
1670 int regno, nregs;
1671 machine_mode mode = GET_MODE (reg);
1673 if (GET_CODE (reg) == SUBREG)
1675 regno = subreg_regno (reg);
1676 nregs = subreg_nregs (reg);
1678 else if (REG_P (reg))
1680 regno = REGNO (reg);
1681 nregs = REG_NREGS (reg);
1683 else
1684 gcc_unreachable ();
1685 for (int i = nregs - 1; i > 0; i--)
1686 reg_mode[regno + i] = BLKmode;
1687 reg_mode[regno] = mode;
1690 /* Record that REG is being set to the sum of SYM and OFF. */
1692 static void
1693 move2add_record_sym_value (rtx reg, rtx sym, rtx off)
1695 int regno = REGNO (reg);
1697 move2add_record_mode (reg);
1698 reg_set_luid[regno] = move2add_luid;
1699 reg_base_reg[regno] = -1;
1700 reg_symbol_ref[regno] = sym;
1701 reg_offset[regno] = INTVAL (off);
1704 /* Check if REGNO contains a valid value in MODE. */
1706 static bool
1707 move2add_valid_value_p (int regno, scalar_int_mode mode)
1709 if (reg_set_luid[regno] <= move2add_last_label_luid)
1710 return false;
1712 if (mode != reg_mode[regno])
1714 scalar_int_mode old_mode;
1715 if (!is_a <scalar_int_mode> (reg_mode[regno], &old_mode)
1716 || !MODES_OK_FOR_MOVE2ADD (mode, old_mode))
1717 return false;
1718 /* The value loaded into regno in reg_mode[regno] is also valid in
1719 mode after truncation only if (REG:mode regno) is the lowpart of
1720 (REG:reg_mode[regno] regno). Now, for big endian, the starting
1721 regno of the lowpart might be different. */
1722 poly_int64 s_off = subreg_lowpart_offset (mode, old_mode);
1723 s_off = subreg_regno_offset (regno, old_mode, s_off, mode);
1724 if (maybe_ne (s_off, 0))
1725 /* We could in principle adjust regno, check reg_mode[regno] to be
1726 BLKmode, and return s_off to the caller (vs. -1 for failure),
1727 but we currently have no callers that could make use of this
1728 information. */
1729 return false;
1732 for (int i = end_hard_regno (mode, regno) - 1; i > regno; i--)
1733 if (reg_mode[i] != BLKmode)
1734 return false;
1735 return true;
1738 /* This function is called with INSN that sets REG (of mode MODE)
1739 to (SYM + OFF), while REG is known to already have value (SYM + offset).
1740 This function tries to change INSN into an add instruction
1741 (set (REG) (plus (REG) (OFF - offset))) using the known value.
1742 It also updates the information about REG's known value.
1743 Return true if we made a change. */
1745 static bool
1746 move2add_use_add2_insn (scalar_int_mode mode, rtx reg, rtx sym, rtx off,
1747 rtx_insn *insn)
1749 rtx pat = PATTERN (insn);
1750 rtx src = SET_SRC (pat);
1751 int regno = REGNO (reg);
1752 rtx new_src = gen_int_mode (UINTVAL (off) - reg_offset[regno], mode);
1753 bool speed = optimize_bb_for_speed_p (BLOCK_FOR_INSN (insn));
1754 bool changed = false;
1756 /* (set (reg) (plus (reg) (const_int 0))) is not canonical;
1757 use (set (reg) (reg)) instead.
1758 We don't delete this insn, nor do we convert it into a
1759 note, to avoid losing register notes or the return
1760 value flag. jump2 already knows how to get rid of
1761 no-op moves. */
1762 if (new_src == const0_rtx)
1764 /* If the constants are different, this is a
1765 truncation, that, if turned into (set (reg)
1766 (reg)), would be discarded. Maybe we should
1767 try a truncMN pattern? */
1768 if (INTVAL (off) == reg_offset [regno])
1769 changed = validate_change (insn, &SET_SRC (pat), reg, 0);
1771 else
1773 struct full_rtx_costs oldcst, newcst;
1774 rtx tem = gen_rtx_PLUS (mode, reg, new_src);
1776 get_full_set_rtx_cost (pat, &oldcst);
1777 SET_SRC (pat) = tem;
1778 get_full_set_rtx_cost (pat, &newcst);
1779 SET_SRC (pat) = src;
1781 if (costs_lt_p (&newcst, &oldcst, speed)
1782 && have_add2_insn (reg, new_src))
1783 changed = validate_change (insn, &SET_SRC (pat), tem, 0);
1784 else if (sym == NULL_RTX && mode != BImode)
1786 scalar_int_mode narrow_mode;
1787 FOR_EACH_MODE_UNTIL (narrow_mode, mode)
1789 if (have_insn_for (STRICT_LOW_PART, narrow_mode)
1790 && ((reg_offset[regno] & ~GET_MODE_MASK (narrow_mode))
1791 == (INTVAL (off) & ~GET_MODE_MASK (narrow_mode))))
1793 rtx narrow_reg = gen_lowpart_common (narrow_mode, reg);
1794 rtx narrow_src = gen_int_mode (INTVAL (off),
1795 narrow_mode);
1796 rtx new_set
1797 = gen_rtx_SET (gen_rtx_STRICT_LOW_PART (VOIDmode,
1798 narrow_reg),
1799 narrow_src);
1800 get_full_set_rtx_cost (new_set, &newcst);
1801 if (costs_lt_p (&newcst, &oldcst, speed))
1803 changed = validate_change (insn, &PATTERN (insn),
1804 new_set, 0);
1805 if (changed)
1806 break;
1812 move2add_record_sym_value (reg, sym, off);
1813 return changed;
1817 /* This function is called with INSN that sets REG (of mode MODE) to
1818 (SYM + OFF), but REG doesn't have known value (SYM + offset). This
1819 function tries to find another register which is known to already have
1820 value (SYM + offset) and change INSN into an add instruction
1821 (set (REG) (plus (the found register) (OFF - offset))) if such
1822 a register is found. It also updates the information about
1823 REG's known value.
1824 Return true iff we made a change. */
1826 static bool
1827 move2add_use_add3_insn (scalar_int_mode mode, rtx reg, rtx sym, rtx off,
1828 rtx_insn *insn)
1830 rtx pat = PATTERN (insn);
1831 rtx src = SET_SRC (pat);
1832 int regno = REGNO (reg);
1833 int min_regno = 0;
1834 bool speed = optimize_bb_for_speed_p (BLOCK_FOR_INSN (insn));
1835 int i;
1836 bool changed = false;
1837 struct full_rtx_costs oldcst, newcst, mincst;
1838 rtx plus_expr;
1840 init_costs_to_max (&mincst);
1841 get_full_set_rtx_cost (pat, &oldcst);
1843 plus_expr = gen_rtx_PLUS (GET_MODE (reg), reg, const0_rtx);
1844 SET_SRC (pat) = plus_expr;
1846 for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
1847 if (move2add_valid_value_p (i, mode)
1848 && reg_base_reg[i] < 0
1849 && reg_symbol_ref[i] != NULL_RTX
1850 && rtx_equal_p (sym, reg_symbol_ref[i]))
1852 rtx new_src = gen_int_mode (UINTVAL (off) - reg_offset[i],
1853 GET_MODE (reg));
1854 /* (set (reg) (plus (reg) (const_int 0))) is not canonical;
1855 use (set (reg) (reg)) instead.
1856 We don't delete this insn, nor do we convert it into a
1857 note, to avoid losing register notes or the return
1858 value flag. jump2 already knows how to get rid of
1859 no-op moves. */
1860 if (new_src == const0_rtx)
1862 init_costs_to_zero (&mincst);
1863 min_regno = i;
1864 break;
1866 else
1868 XEXP (plus_expr, 1) = new_src;
1869 get_full_set_rtx_cost (pat, &newcst);
1871 if (costs_lt_p (&newcst, &mincst, speed))
1873 mincst = newcst;
1874 min_regno = i;
1878 SET_SRC (pat) = src;
1880 if (costs_lt_p (&mincst, &oldcst, speed))
1882 rtx tem;
1884 tem = gen_rtx_REG (GET_MODE (reg), min_regno);
1885 if (i != min_regno)
1887 rtx new_src = gen_int_mode (UINTVAL (off) - reg_offset[min_regno],
1888 GET_MODE (reg));
1889 tem = gen_rtx_PLUS (GET_MODE (reg), tem, new_src);
1891 if (validate_change (insn, &SET_SRC (pat), tem, 0))
1892 changed = true;
1894 reg_set_luid[regno] = move2add_luid;
1895 move2add_record_sym_value (reg, sym, off);
1896 return changed;
1899 /* Convert move insns with constant inputs to additions if they are cheaper.
1900 Return true if any changes were made. */
1901 static bool
1902 reload_cse_move2add (rtx_insn *first)
1904 int i;
1905 rtx_insn *insn;
1906 bool changed = false;
1908 for (i = FIRST_PSEUDO_REGISTER - 1; i >= 0; i--)
1910 reg_set_luid[i] = 0;
1911 reg_offset[i] = 0;
1912 reg_base_reg[i] = 0;
1913 reg_symbol_ref[i] = NULL_RTX;
1914 reg_mode[i] = VOIDmode;
1917 move2add_last_label_luid = 0;
1918 move2add_luid = 2;
1919 for (insn = first; insn; insn = NEXT_INSN (insn), move2add_luid++)
1921 rtx pat, note;
1923 if (LABEL_P (insn))
1925 move2add_last_label_luid = move2add_luid;
1926 /* We're going to increment move2add_luid twice after a
1927 label, so that we can use move2add_last_label_luid + 1 as
1928 the luid for constants. */
1929 move2add_luid++;
1930 continue;
1932 if (! INSN_P (insn))
1933 continue;
1934 pat = PATTERN (insn);
1935 /* For simplicity, we only perform this optimization on
1936 straightforward SETs. */
1937 scalar_int_mode mode;
1938 if (GET_CODE (pat) == SET
1939 && REG_P (SET_DEST (pat))
1940 && is_a <scalar_int_mode> (GET_MODE (SET_DEST (pat)), &mode))
1942 rtx reg = SET_DEST (pat);
1943 int regno = REGNO (reg);
1944 rtx src = SET_SRC (pat);
1946 /* Check if we have valid information on the contents of this
1947 register in the mode of REG. */
1948 if (move2add_valid_value_p (regno, mode)
1949 && dbg_cnt (cse2_move2add))
1951 /* Try to transform (set (REGX) (CONST_INT A))
1953 (set (REGX) (CONST_INT B))
1955 (set (REGX) (CONST_INT A))
1957 (set (REGX) (plus (REGX) (CONST_INT B-A)))
1959 (set (REGX) (CONST_INT A))
1961 (set (STRICT_LOW_PART (REGX)) (CONST_INT B))
1964 if (CONST_INT_P (src)
1965 && reg_base_reg[regno] < 0
1966 && reg_symbol_ref[regno] == NULL_RTX)
1968 changed |= move2add_use_add2_insn (mode, reg, NULL_RTX,
1969 src, insn);
1970 continue;
1973 /* Try to transform (set (REGX) (REGY))
1974 (set (REGX) (PLUS (REGX) (CONST_INT A)))
1976 (set (REGX) (REGY))
1977 (set (REGX) (PLUS (REGX) (CONST_INT B)))
1979 (set (REGX) (REGY))
1980 (set (REGX) (PLUS (REGX) (CONST_INT A)))
1982 (set (REGX) (plus (REGX) (CONST_INT B-A))) */
1983 else if (REG_P (src)
1984 && reg_set_luid[regno] == reg_set_luid[REGNO (src)]
1985 && reg_base_reg[regno] == reg_base_reg[REGNO (src)]
1986 && move2add_valid_value_p (REGNO (src), mode))
1988 rtx_insn *next = next_nonnote_nondebug_insn (insn);
1989 rtx set = NULL_RTX;
1990 if (next)
1991 set = single_set (next);
1992 if (set
1993 && SET_DEST (set) == reg
1994 && GET_CODE (SET_SRC (set)) == PLUS
1995 && XEXP (SET_SRC (set), 0) == reg
1996 && CONST_INT_P (XEXP (SET_SRC (set), 1)))
1998 rtx src3 = XEXP (SET_SRC (set), 1);
1999 unsigned HOST_WIDE_INT added_offset = UINTVAL (src3);
2000 HOST_WIDE_INT base_offset = reg_offset[REGNO (src)];
2001 HOST_WIDE_INT regno_offset = reg_offset[regno];
2002 rtx new_src =
2003 gen_int_mode (added_offset
2004 + base_offset
2005 - regno_offset,
2006 mode);
2007 bool success = false;
2008 bool speed = optimize_bb_for_speed_p (BLOCK_FOR_INSN (insn));
2010 if (new_src == const0_rtx)
2011 /* See above why we create (set (reg) (reg)) here. */
2012 success
2013 = validate_change (next, &SET_SRC (set), reg, 0);
2014 else
2016 rtx old_src = SET_SRC (set);
2017 struct full_rtx_costs oldcst, newcst;
2018 rtx tem = gen_rtx_PLUS (mode, reg, new_src);
2020 get_full_set_rtx_cost (set, &oldcst);
2021 SET_SRC (set) = tem;
2022 get_full_set_src_cost (tem, mode, &newcst);
2023 SET_SRC (set) = old_src;
2024 costs_add_n_insns (&oldcst, 1);
2026 if (costs_lt_p (&newcst, &oldcst, speed)
2027 && have_add2_insn (reg, new_src))
2029 rtx newpat = gen_rtx_SET (reg, tem);
2030 success
2031 = validate_change (next, &PATTERN (next),
2032 newpat, 0);
2035 if (success)
2036 delete_insn (insn);
2037 changed |= success;
2038 insn = next;
2039 move2add_record_mode (reg);
2040 reg_offset[regno]
2041 = trunc_int_for_mode (added_offset + base_offset,
2042 mode);
2043 continue;
2048 /* Try to transform
2049 (set (REGX) (CONST (PLUS (SYMBOL_REF) (CONST_INT A))))
2051 (set (REGY) (CONST (PLUS (SYMBOL_REF) (CONST_INT B))))
2053 (set (REGX) (CONST (PLUS (SYMBOL_REF) (CONST_INT A))))
2055 (set (REGY) (CONST (PLUS (REGX) (CONST_INT B-A)))) */
2056 if ((GET_CODE (src) == SYMBOL_REF
2057 || (GET_CODE (src) == CONST
2058 && GET_CODE (XEXP (src, 0)) == PLUS
2059 && GET_CODE (XEXP (XEXP (src, 0), 0)) == SYMBOL_REF
2060 && CONST_INT_P (XEXP (XEXP (src, 0), 1))))
2061 && dbg_cnt (cse2_move2add))
2063 rtx sym, off;
2065 if (GET_CODE (src) == SYMBOL_REF)
2067 sym = src;
2068 off = const0_rtx;
2070 else
2072 sym = XEXP (XEXP (src, 0), 0);
2073 off = XEXP (XEXP (src, 0), 1);
2076 /* If the reg already contains the value which is sum of
2077 sym and some constant value, we can use an add2 insn. */
2078 if (move2add_valid_value_p (regno, mode)
2079 && reg_base_reg[regno] < 0
2080 && reg_symbol_ref[regno] != NULL_RTX
2081 && rtx_equal_p (sym, reg_symbol_ref[regno]))
2082 changed |= move2add_use_add2_insn (mode, reg, sym, off, insn);
2084 /* Otherwise, we have to find a register whose value is sum
2085 of sym and some constant value. */
2086 else
2087 changed |= move2add_use_add3_insn (mode, reg, sym, off, insn);
2089 continue;
2093 for (note = REG_NOTES (insn); note; note = XEXP (note, 1))
2095 if (REG_NOTE_KIND (note) == REG_INC
2096 && REG_P (XEXP (note, 0)))
2098 /* Reset the information about this register. */
2099 int regno = REGNO (XEXP (note, 0));
2100 if (regno < FIRST_PSEUDO_REGISTER)
2102 move2add_record_mode (XEXP (note, 0));
2103 reg_mode[regno] = VOIDmode;
2107 note_stores (PATTERN (insn), move2add_note_store, insn);
2109 /* If INSN is a conditional branch, we try to extract an
2110 implicit set out of it. */
2111 if (any_condjump_p (insn))
2113 rtx cnd = fis_get_condition (insn);
2115 if (cnd != NULL_RTX
2116 && GET_CODE (cnd) == NE
2117 && REG_P (XEXP (cnd, 0))
2118 && !reg_set_p (XEXP (cnd, 0), insn)
2119 /* The following two checks, which are also in
2120 move2add_note_store, are intended to reduce the
2121 number of calls to gen_rtx_SET to avoid memory
2122 allocation if possible. */
2123 && SCALAR_INT_MODE_P (GET_MODE (XEXP (cnd, 0)))
2124 && REG_NREGS (XEXP (cnd, 0)) == 1
2125 && CONST_INT_P (XEXP (cnd, 1)))
2127 rtx implicit_set =
2128 gen_rtx_SET (XEXP (cnd, 0), XEXP (cnd, 1));
2129 move2add_note_store (SET_DEST (implicit_set), implicit_set, insn);
2133 /* If this is a CALL_INSN, all call used registers are stored with
2134 unknown values. */
2135 if (CALL_P (insn))
2137 rtx link;
2139 for (i = FIRST_PSEUDO_REGISTER - 1; i >= 0; i--)
2141 if (call_used_regs[i])
2142 /* Reset the information about this register. */
2143 reg_mode[i] = VOIDmode;
2146 for (link = CALL_INSN_FUNCTION_USAGE (insn); link;
2147 link = XEXP (link, 1))
2149 rtx setuse = XEXP (link, 0);
2150 rtx usage_rtx = XEXP (setuse, 0);
2151 /* CALL_INSN_FUNCTION_USAGEs can only have full clobbers, not
2152 clobber_highs. */
2153 gcc_assert (GET_CODE (setuse) != CLOBBER_HIGH);
2154 if (GET_CODE (setuse) == CLOBBER
2155 && REG_P (usage_rtx))
2157 unsigned int end_regno = END_REGNO (usage_rtx);
2158 for (unsigned int r = REGNO (usage_rtx); r < end_regno; ++r)
2159 /* Reset the information about this register. */
2160 reg_mode[r] = VOIDmode;
2165 return changed;
2168 /* SET is a SET or CLOBBER that sets DST. DATA is the insn which
2169 contains SET.
2170 Update reg_set_luid, reg_offset and reg_base_reg accordingly.
2171 Called from reload_cse_move2add via note_stores. */
2173 static void
2174 move2add_note_store (rtx dst, const_rtx set, void *data)
2176 rtx_insn *insn = (rtx_insn *) data;
2177 unsigned int regno = 0;
2178 scalar_int_mode mode;
2180 /* Some targets do argument pushes without adding REG_INC notes. */
2182 if (MEM_P (dst))
2184 dst = XEXP (dst, 0);
2185 if (GET_CODE (dst) == PRE_INC || GET_CODE (dst) == POST_INC
2186 || GET_CODE (dst) == PRE_DEC || GET_CODE (dst) == POST_DEC)
2187 reg_mode[REGNO (XEXP (dst, 0))] = VOIDmode;
2188 return;
2191 if (GET_CODE (dst) == SUBREG)
2192 regno = subreg_regno (dst);
2193 else if (REG_P (dst))
2194 regno = REGNO (dst);
2195 else
2196 return;
2198 if (!is_a <scalar_int_mode> (GET_MODE (dst), &mode))
2199 goto invalidate;
2201 if (GET_CODE (set) == SET)
2203 rtx note, sym = NULL_RTX;
2204 rtx off;
2206 note = find_reg_equal_equiv_note (insn);
2207 if (note && GET_CODE (XEXP (note, 0)) == SYMBOL_REF)
2209 sym = XEXP (note, 0);
2210 off = const0_rtx;
2212 else if (note && GET_CODE (XEXP (note, 0)) == CONST
2213 && GET_CODE (XEXP (XEXP (note, 0), 0)) == PLUS
2214 && GET_CODE (XEXP (XEXP (XEXP (note, 0), 0), 0)) == SYMBOL_REF
2215 && CONST_INT_P (XEXP (XEXP (XEXP (note, 0), 0), 1)))
2217 sym = XEXP (XEXP (XEXP (note, 0), 0), 0);
2218 off = XEXP (XEXP (XEXP (note, 0), 0), 1);
2221 if (sym != NULL_RTX)
2223 move2add_record_sym_value (dst, sym, off);
2224 return;
2228 if (GET_CODE (set) == SET
2229 && GET_CODE (SET_DEST (set)) != ZERO_EXTRACT
2230 && GET_CODE (SET_DEST (set)) != STRICT_LOW_PART)
2232 rtx src = SET_SRC (set);
2233 rtx base_reg;
2234 unsigned HOST_WIDE_INT offset;
2235 int base_regno;
2237 switch (GET_CODE (src))
2239 case PLUS:
2240 if (REG_P (XEXP (src, 0)))
2242 base_reg = XEXP (src, 0);
2244 if (CONST_INT_P (XEXP (src, 1)))
2245 offset = UINTVAL (XEXP (src, 1));
2246 else if (REG_P (XEXP (src, 1))
2247 && move2add_valid_value_p (REGNO (XEXP (src, 1)), mode))
2249 if (reg_base_reg[REGNO (XEXP (src, 1))] < 0
2250 && reg_symbol_ref[REGNO (XEXP (src, 1))] == NULL_RTX)
2251 offset = reg_offset[REGNO (XEXP (src, 1))];
2252 /* Maybe the first register is known to be a
2253 constant. */
2254 else if (move2add_valid_value_p (REGNO (base_reg), mode)
2255 && reg_base_reg[REGNO (base_reg)] < 0
2256 && reg_symbol_ref[REGNO (base_reg)] == NULL_RTX)
2258 offset = reg_offset[REGNO (base_reg)];
2259 base_reg = XEXP (src, 1);
2261 else
2262 goto invalidate;
2264 else
2265 goto invalidate;
2267 break;
2270 goto invalidate;
2272 case REG:
2273 base_reg = src;
2274 offset = 0;
2275 break;
2277 case CONST_INT:
2278 /* Start tracking the register as a constant. */
2279 reg_base_reg[regno] = -1;
2280 reg_symbol_ref[regno] = NULL_RTX;
2281 reg_offset[regno] = INTVAL (SET_SRC (set));
2282 /* We assign the same luid to all registers set to constants. */
2283 reg_set_luid[regno] = move2add_last_label_luid + 1;
2284 move2add_record_mode (dst);
2285 return;
2287 default:
2288 goto invalidate;
2291 base_regno = REGNO (base_reg);
2292 /* If information about the base register is not valid, set it
2293 up as a new base register, pretending its value is known
2294 starting from the current insn. */
2295 if (!move2add_valid_value_p (base_regno, mode))
2297 reg_base_reg[base_regno] = base_regno;
2298 reg_symbol_ref[base_regno] = NULL_RTX;
2299 reg_offset[base_regno] = 0;
2300 reg_set_luid[base_regno] = move2add_luid;
2301 gcc_assert (GET_MODE (base_reg) == mode);
2302 move2add_record_mode (base_reg);
2305 /* Copy base information from our base register. */
2306 reg_set_luid[regno] = reg_set_luid[base_regno];
2307 reg_base_reg[regno] = reg_base_reg[base_regno];
2308 reg_symbol_ref[regno] = reg_symbol_ref[base_regno];
2310 /* Compute the sum of the offsets or constants. */
2311 reg_offset[regno]
2312 = trunc_int_for_mode (offset + reg_offset[base_regno], mode);
2314 move2add_record_mode (dst);
2316 else if (GET_CODE (set) == CLOBBER_HIGH)
2318 /* Only invalidate if actually clobbered. */
2319 if (reg_mode[regno] == BLKmode
2320 || reg_is_clobbered_by_clobber_high (regno, reg_mode[regno], dst))
2321 goto invalidate;
2323 else
2325 invalidate:
2326 /* Invalidate the contents of the register. */
2327 move2add_record_mode (dst);
2328 reg_mode[regno] = VOIDmode;
2332 namespace {
2334 const pass_data pass_data_postreload_cse =
2336 RTL_PASS, /* type */
2337 "postreload", /* name */
2338 OPTGROUP_NONE, /* optinfo_flags */
2339 TV_RELOAD_CSE_REGS, /* tv_id */
2340 0, /* properties_required */
2341 0, /* properties_provided */
2342 0, /* properties_destroyed */
2343 0, /* todo_flags_start */
2344 TODO_df_finish, /* todo_flags_finish */
2347 class pass_postreload_cse : public rtl_opt_pass
2349 public:
2350 pass_postreload_cse (gcc::context *ctxt)
2351 : rtl_opt_pass (pass_data_postreload_cse, ctxt)
2354 /* opt_pass methods: */
2355 virtual bool gate (function *) { return (optimize > 0 && reload_completed); }
2357 virtual unsigned int execute (function *);
2359 }; // class pass_postreload_cse
2361 unsigned int
2362 pass_postreload_cse::execute (function *fun)
2364 if (!dbg_cnt (postreload_cse))
2365 return 0;
2367 /* Do a very simple CSE pass over just the hard registers. */
2368 reload_cse_regs (get_insns ());
2369 /* Reload_cse_regs can eliminate potentially-trapping MEMs.
2370 Remove any EH edges associated with them. */
2371 if (fun->can_throw_non_call_exceptions
2372 && purge_all_dead_edges ())
2373 cleanup_cfg (0);
2375 return 0;
2378 } // anon namespace
2380 rtl_opt_pass *
2381 make_pass_postreload_cse (gcc::context *ctxt)
2383 return new pass_postreload_cse (ctxt);