* rw.po: Remove.
[official-gcc.git] / gcc / rtlanal.c
blobb16da4a0956fe4c58327cde1654e6407e136c665
1 /* Analyze RTL for GNU compiler.
2 Copyright (C) 1987, 1988, 1992, 1993, 1994, 1995, 1996, 1997, 1998,
3 1999, 2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007 Free Software
4 Foundation, Inc.
6 This file is part of GCC.
8 GCC is free software; you can redistribute it and/or modify it under
9 the terms of the GNU General Public License as published by the Free
10 Software Foundation; either version 3, or (at your option) any later
11 version.
13 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
14 WARRANTY; without even the implied warranty of MERCHANTABILITY or
15 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
16 for more details.
18 You should have received a copy of the GNU General Public License
19 along with GCC; see the file COPYING3. If not see
20 <http://www.gnu.org/licenses/>. */
23 #include "config.h"
24 #include "system.h"
25 #include "coretypes.h"
26 #include "tm.h"
27 #include "toplev.h"
28 #include "rtl.h"
29 #include "hard-reg-set.h"
30 #include "insn-config.h"
31 #include "recog.h"
32 #include "target.h"
33 #include "output.h"
34 #include "tm_p.h"
35 #include "flags.h"
36 #include "real.h"
37 #include "regs.h"
38 #include "function.h"
40 /* Forward declarations */
41 static void set_of_1 (rtx, rtx, void *);
42 static bool covers_regno_p (rtx, unsigned int);
43 static bool covers_regno_no_parallel_p (rtx, unsigned int);
44 static int rtx_referenced_p_1 (rtx *, void *);
45 static int computed_jump_p_1 (rtx);
46 static void parms_set (rtx, rtx, void *);
48 static unsigned HOST_WIDE_INT cached_nonzero_bits (rtx, enum machine_mode,
49 rtx, enum machine_mode,
50 unsigned HOST_WIDE_INT);
51 static unsigned HOST_WIDE_INT nonzero_bits1 (rtx, enum machine_mode, rtx,
52 enum machine_mode,
53 unsigned HOST_WIDE_INT);
54 static unsigned int cached_num_sign_bit_copies (rtx, enum machine_mode, rtx,
55 enum machine_mode,
56 unsigned int);
57 static unsigned int num_sign_bit_copies1 (rtx, enum machine_mode, rtx,
58 enum machine_mode, unsigned int);
60 /* Offset of the first 'e', 'E' or 'V' operand for each rtx code, or
61 -1 if a code has no such operand. */
62 static int non_rtx_starting_operands[NUM_RTX_CODE];
64 /* Bit flags that specify the machine subtype we are compiling for.
65 Bits are tested using macros TARGET_... defined in the tm.h file
66 and set by `-m...' switches. Must be defined in rtlanal.c. */
68 int target_flags;
70 /* Truncation narrows the mode from SOURCE mode to DESTINATION mode.
71 If TARGET_MODE_REP_EXTENDED (DESTINATION, DESTINATION_REP) is
72 SIGN_EXTEND then while narrowing we also have to enforce the
73 representation and sign-extend the value to mode DESTINATION_REP.
75 If the value is already sign-extended to DESTINATION_REP mode we
76 can just switch to DESTINATION mode on it. For each pair of
77 integral modes SOURCE and DESTINATION, when truncating from SOURCE
78 to DESTINATION, NUM_SIGN_BIT_COPIES_IN_REP[SOURCE][DESTINATION]
79 contains the number of high-order bits in SOURCE that have to be
80 copies of the sign-bit so that we can do this mode-switch to
81 DESTINATION. */
83 static unsigned int
84 num_sign_bit_copies_in_rep[MAX_MODE_INT + 1][MAX_MODE_INT + 1];
86 /* Return 1 if the value of X is unstable
87 (would be different at a different point in the program).
88 The frame pointer, arg pointer, etc. are considered stable
89 (within one function) and so is anything marked `unchanging'. */
91 int
92 rtx_unstable_p (rtx x)
94 RTX_CODE code = GET_CODE (x);
95 int i;
96 const char *fmt;
98 switch (code)
100 case MEM:
101 return !MEM_READONLY_P (x) || rtx_unstable_p (XEXP (x, 0));
103 case CONST:
104 case CONST_INT:
105 case CONST_DOUBLE:
106 case CONST_VECTOR:
107 case SYMBOL_REF:
108 case LABEL_REF:
109 return 0;
111 case REG:
112 /* As in rtx_varies_p, we have to use the actual rtx, not reg number. */
113 if (x == frame_pointer_rtx || x == hard_frame_pointer_rtx
114 /* The arg pointer varies if it is not a fixed register. */
115 || (x == arg_pointer_rtx && fixed_regs[ARG_POINTER_REGNUM]))
116 return 0;
117 #ifndef PIC_OFFSET_TABLE_REG_CALL_CLOBBERED
118 /* ??? When call-clobbered, the value is stable modulo the restore
119 that must happen after a call. This currently screws up local-alloc
120 into believing that the restore is not needed. */
121 if (x == pic_offset_table_rtx)
122 return 0;
123 #endif
124 return 1;
126 case ASM_OPERANDS:
127 if (MEM_VOLATILE_P (x))
128 return 1;
130 /* Fall through. */
132 default:
133 break;
136 fmt = GET_RTX_FORMAT (code);
137 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
138 if (fmt[i] == 'e')
140 if (rtx_unstable_p (XEXP (x, i)))
141 return 1;
143 else if (fmt[i] == 'E')
145 int j;
146 for (j = 0; j < XVECLEN (x, i); j++)
147 if (rtx_unstable_p (XVECEXP (x, i, j)))
148 return 1;
151 return 0;
154 /* Return 1 if X has a value that can vary even between two
155 executions of the program. 0 means X can be compared reliably
156 against certain constants or near-constants.
157 FOR_ALIAS is nonzero if we are called from alias analysis; if it is
158 zero, we are slightly more conservative.
159 The frame pointer and the arg pointer are considered constant. */
162 rtx_varies_p (rtx x, int for_alias)
164 RTX_CODE code;
165 int i;
166 const char *fmt;
168 if (!x)
169 return 0;
171 code = GET_CODE (x);
172 switch (code)
174 case MEM:
175 return !MEM_READONLY_P (x) || rtx_varies_p (XEXP (x, 0), for_alias);
177 case CONST:
178 case CONST_INT:
179 case CONST_DOUBLE:
180 case CONST_VECTOR:
181 case SYMBOL_REF:
182 case LABEL_REF:
183 return 0;
185 case REG:
186 /* Note that we have to test for the actual rtx used for the frame
187 and arg pointers and not just the register number in case we have
188 eliminated the frame and/or arg pointer and are using it
189 for pseudos. */
190 if (x == frame_pointer_rtx || x == hard_frame_pointer_rtx
191 /* The arg pointer varies if it is not a fixed register. */
192 || (x == arg_pointer_rtx && fixed_regs[ARG_POINTER_REGNUM]))
193 return 0;
194 if (x == pic_offset_table_rtx
195 #ifdef PIC_OFFSET_TABLE_REG_CALL_CLOBBERED
196 /* ??? When call-clobbered, the value is stable modulo the restore
197 that must happen after a call. This currently screws up
198 local-alloc into believing that the restore is not needed, so we
199 must return 0 only if we are called from alias analysis. */
200 && for_alias
201 #endif
203 return 0;
204 return 1;
206 case LO_SUM:
207 /* The operand 0 of a LO_SUM is considered constant
208 (in fact it is related specifically to operand 1)
209 during alias analysis. */
210 return (! for_alias && rtx_varies_p (XEXP (x, 0), for_alias))
211 || rtx_varies_p (XEXP (x, 1), for_alias);
213 case ASM_OPERANDS:
214 if (MEM_VOLATILE_P (x))
215 return 1;
217 /* Fall through. */
219 default:
220 break;
223 fmt = GET_RTX_FORMAT (code);
224 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
225 if (fmt[i] == 'e')
227 if (rtx_varies_p (XEXP (x, i), for_alias))
228 return 1;
230 else if (fmt[i] == 'E')
232 int j;
233 for (j = 0; j < XVECLEN (x, i); j++)
234 if (rtx_varies_p (XVECEXP (x, i, j), for_alias))
235 return 1;
238 return 0;
241 /* Return nonzero if the use of X as an address in a MEM can cause a trap.
242 MODE is the mode of the MEM (not that of X) and UNALIGNED_MEMS controls
243 whether nonzero is returned for unaligned memory accesses on strict
244 alignment machines. */
246 static int
247 rtx_addr_can_trap_p_1 (rtx x, enum machine_mode mode, bool unaligned_mems)
249 enum rtx_code code = GET_CODE (x);
251 switch (code)
253 case SYMBOL_REF:
254 return SYMBOL_REF_WEAK (x);
256 case LABEL_REF:
257 return 0;
259 case REG:
260 /* As in rtx_varies_p, we have to use the actual rtx, not reg number. */
261 if (x == frame_pointer_rtx || x == hard_frame_pointer_rtx
262 || x == stack_pointer_rtx
263 /* The arg pointer varies if it is not a fixed register. */
264 || (x == arg_pointer_rtx && fixed_regs[ARG_POINTER_REGNUM]))
265 return 0;
266 /* All of the virtual frame registers are stack references. */
267 if (REGNO (x) >= FIRST_VIRTUAL_REGISTER
268 && REGNO (x) <= LAST_VIRTUAL_REGISTER)
269 return 0;
270 return 1;
272 case CONST:
273 return rtx_addr_can_trap_p_1 (XEXP (x, 0), mode, unaligned_mems);
275 case PLUS:
276 /* An address is assumed not to trap if:
277 - it is an address that can't trap plus a constant integer,
278 with the proper remainder modulo the mode size if we are
279 considering unaligned memory references. */
280 if (!rtx_addr_can_trap_p_1 (XEXP (x, 0), mode, unaligned_mems)
281 && GET_CODE (XEXP (x, 1)) == CONST_INT)
283 HOST_WIDE_INT offset;
285 if (!STRICT_ALIGNMENT
286 || !unaligned_mems
287 || GET_MODE_SIZE (mode) == 0)
288 return 0;
290 offset = INTVAL (XEXP (x, 1));
292 #ifdef SPARC_STACK_BOUNDARY_HACK
293 /* ??? The SPARC port may claim a STACK_BOUNDARY higher than
294 the real alignment of %sp. However, when it does this, the
295 alignment of %sp+STACK_POINTER_OFFSET is STACK_BOUNDARY. */
296 if (SPARC_STACK_BOUNDARY_HACK
297 && (XEXP (x, 0) == stack_pointer_rtx
298 || XEXP (x, 0) == hard_frame_pointer_rtx))
299 offset -= STACK_POINTER_OFFSET;
300 #endif
302 return offset % GET_MODE_SIZE (mode) != 0;
305 /* - or it is the pic register plus a constant. */
306 if (XEXP (x, 0) == pic_offset_table_rtx && CONSTANT_P (XEXP (x, 1)))
307 return 0;
309 return 1;
311 case LO_SUM:
312 case PRE_MODIFY:
313 return rtx_addr_can_trap_p_1 (XEXP (x, 1), mode, unaligned_mems);
315 case PRE_DEC:
316 case PRE_INC:
317 case POST_DEC:
318 case POST_INC:
319 case POST_MODIFY:
320 return rtx_addr_can_trap_p_1 (XEXP (x, 0), mode, unaligned_mems);
322 default:
323 break;
326 /* If it isn't one of the case above, it can cause a trap. */
327 return 1;
330 /* Return nonzero if the use of X as an address in a MEM can cause a trap. */
333 rtx_addr_can_trap_p (rtx x)
335 return rtx_addr_can_trap_p_1 (x, VOIDmode, false);
338 /* Return true if X is an address that is known to not be zero. */
340 bool
341 nonzero_address_p (rtx x)
343 enum rtx_code code = GET_CODE (x);
345 switch (code)
347 case SYMBOL_REF:
348 return !SYMBOL_REF_WEAK (x);
350 case LABEL_REF:
351 return true;
353 case REG:
354 /* As in rtx_varies_p, we have to use the actual rtx, not reg number. */
355 if (x == frame_pointer_rtx || x == hard_frame_pointer_rtx
356 || x == stack_pointer_rtx
357 || (x == arg_pointer_rtx && fixed_regs[ARG_POINTER_REGNUM]))
358 return true;
359 /* All of the virtual frame registers are stack references. */
360 if (REGNO (x) >= FIRST_VIRTUAL_REGISTER
361 && REGNO (x) <= LAST_VIRTUAL_REGISTER)
362 return true;
363 return false;
365 case CONST:
366 return nonzero_address_p (XEXP (x, 0));
368 case PLUS:
369 if (GET_CODE (XEXP (x, 1)) == CONST_INT)
370 return nonzero_address_p (XEXP (x, 0));
371 /* Handle PIC references. */
372 else if (XEXP (x, 0) == pic_offset_table_rtx
373 && CONSTANT_P (XEXP (x, 1)))
374 return true;
375 return false;
377 case PRE_MODIFY:
378 /* Similar to the above; allow positive offsets. Further, since
379 auto-inc is only allowed in memories, the register must be a
380 pointer. */
381 if (GET_CODE (XEXP (x, 1)) == CONST_INT
382 && INTVAL (XEXP (x, 1)) > 0)
383 return true;
384 return nonzero_address_p (XEXP (x, 0));
386 case PRE_INC:
387 /* Similarly. Further, the offset is always positive. */
388 return true;
390 case PRE_DEC:
391 case POST_DEC:
392 case POST_INC:
393 case POST_MODIFY:
394 return nonzero_address_p (XEXP (x, 0));
396 case LO_SUM:
397 return nonzero_address_p (XEXP (x, 1));
399 default:
400 break;
403 /* If it isn't one of the case above, might be zero. */
404 return false;
407 /* Return 1 if X refers to a memory location whose address
408 cannot be compared reliably with constant addresses,
409 or if X refers to a BLKmode memory object.
410 FOR_ALIAS is nonzero if we are called from alias analysis; if it is
411 zero, we are slightly more conservative. */
414 rtx_addr_varies_p (rtx x, int for_alias)
416 enum rtx_code code;
417 int i;
418 const char *fmt;
420 if (x == 0)
421 return 0;
423 code = GET_CODE (x);
424 if (code == MEM)
425 return GET_MODE (x) == BLKmode || rtx_varies_p (XEXP (x, 0), for_alias);
427 fmt = GET_RTX_FORMAT (code);
428 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
429 if (fmt[i] == 'e')
431 if (rtx_addr_varies_p (XEXP (x, i), for_alias))
432 return 1;
434 else if (fmt[i] == 'E')
436 int j;
437 for (j = 0; j < XVECLEN (x, i); j++)
438 if (rtx_addr_varies_p (XVECEXP (x, i, j), for_alias))
439 return 1;
441 return 0;
444 /* Return the value of the integer term in X, if one is apparent;
445 otherwise return 0.
446 Only obvious integer terms are detected.
447 This is used in cse.c with the `related_value' field. */
449 HOST_WIDE_INT
450 get_integer_term (rtx x)
452 if (GET_CODE (x) == CONST)
453 x = XEXP (x, 0);
455 if (GET_CODE (x) == MINUS
456 && GET_CODE (XEXP (x, 1)) == CONST_INT)
457 return - INTVAL (XEXP (x, 1));
458 if (GET_CODE (x) == PLUS
459 && GET_CODE (XEXP (x, 1)) == CONST_INT)
460 return INTVAL (XEXP (x, 1));
461 return 0;
464 /* If X is a constant, return the value sans apparent integer term;
465 otherwise return 0.
466 Only obvious integer terms are detected. */
469 get_related_value (rtx x)
471 if (GET_CODE (x) != CONST)
472 return 0;
473 x = XEXP (x, 0);
474 if (GET_CODE (x) == PLUS
475 && GET_CODE (XEXP (x, 1)) == CONST_INT)
476 return XEXP (x, 0);
477 else if (GET_CODE (x) == MINUS
478 && GET_CODE (XEXP (x, 1)) == CONST_INT)
479 return XEXP (x, 0);
480 return 0;
483 /* Return the number of places FIND appears within X. If COUNT_DEST is
484 zero, we do not count occurrences inside the destination of a SET. */
487 count_occurrences (rtx x, rtx find, int count_dest)
489 int i, j;
490 enum rtx_code code;
491 const char *format_ptr;
492 int count;
494 if (x == find)
495 return 1;
497 code = GET_CODE (x);
499 switch (code)
501 case REG:
502 case CONST_INT:
503 case CONST_DOUBLE:
504 case CONST_VECTOR:
505 case SYMBOL_REF:
506 case CODE_LABEL:
507 case PC:
508 case CC0:
509 return 0;
511 case MEM:
512 if (MEM_P (find) && rtx_equal_p (x, find))
513 return 1;
514 break;
516 case SET:
517 if (SET_DEST (x) == find && ! count_dest)
518 return count_occurrences (SET_SRC (x), find, count_dest);
519 break;
521 default:
522 break;
525 format_ptr = GET_RTX_FORMAT (code);
526 count = 0;
528 for (i = 0; i < GET_RTX_LENGTH (code); i++)
530 switch (*format_ptr++)
532 case 'e':
533 count += count_occurrences (XEXP (x, i), find, count_dest);
534 break;
536 case 'E':
537 for (j = 0; j < XVECLEN (x, i); j++)
538 count += count_occurrences (XVECEXP (x, i, j), find, count_dest);
539 break;
542 return count;
545 /* Nonzero if register REG appears somewhere within IN.
546 Also works if REG is not a register; in this case it checks
547 for a subexpression of IN that is Lisp "equal" to REG. */
550 reg_mentioned_p (rtx reg, rtx in)
552 const char *fmt;
553 int i;
554 enum rtx_code code;
556 if (in == 0)
557 return 0;
559 if (reg == in)
560 return 1;
562 if (GET_CODE (in) == LABEL_REF)
563 return reg == XEXP (in, 0);
565 code = GET_CODE (in);
567 switch (code)
569 /* Compare registers by number. */
570 case REG:
571 return REG_P (reg) && REGNO (in) == REGNO (reg);
573 /* These codes have no constituent expressions
574 and are unique. */
575 case SCRATCH:
576 case CC0:
577 case PC:
578 return 0;
580 case CONST_INT:
581 case CONST_VECTOR:
582 case CONST_DOUBLE:
583 /* These are kept unique for a given value. */
584 return 0;
586 default:
587 break;
590 if (GET_CODE (reg) == code && rtx_equal_p (reg, in))
591 return 1;
593 fmt = GET_RTX_FORMAT (code);
595 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
597 if (fmt[i] == 'E')
599 int j;
600 for (j = XVECLEN (in, i) - 1; j >= 0; j--)
601 if (reg_mentioned_p (reg, XVECEXP (in, i, j)))
602 return 1;
604 else if (fmt[i] == 'e'
605 && reg_mentioned_p (reg, XEXP (in, i)))
606 return 1;
608 return 0;
611 /* Return 1 if in between BEG and END, exclusive of BEG and END, there is
612 no CODE_LABEL insn. */
615 no_labels_between_p (rtx beg, rtx end)
617 rtx p;
618 if (beg == end)
619 return 0;
620 for (p = NEXT_INSN (beg); p != end; p = NEXT_INSN (p))
621 if (LABEL_P (p))
622 return 0;
623 return 1;
626 /* Nonzero if register REG is used in an insn between
627 FROM_INSN and TO_INSN (exclusive of those two). */
630 reg_used_between_p (rtx reg, rtx from_insn, rtx to_insn)
632 rtx insn;
634 if (from_insn == to_insn)
635 return 0;
637 for (insn = NEXT_INSN (from_insn); insn != to_insn; insn = NEXT_INSN (insn))
638 if (INSN_P (insn)
639 && (reg_overlap_mentioned_p (reg, PATTERN (insn))
640 || (CALL_P (insn) && find_reg_fusage (insn, USE, reg))))
641 return 1;
642 return 0;
645 /* Nonzero if the old value of X, a register, is referenced in BODY. If X
646 is entirely replaced by a new value and the only use is as a SET_DEST,
647 we do not consider it a reference. */
650 reg_referenced_p (rtx x, rtx body)
652 int i;
654 switch (GET_CODE (body))
656 case SET:
657 if (reg_overlap_mentioned_p (x, SET_SRC (body)))
658 return 1;
660 /* If the destination is anything other than CC0, PC, a REG or a SUBREG
661 of a REG that occupies all of the REG, the insn references X if
662 it is mentioned in the destination. */
663 if (GET_CODE (SET_DEST (body)) != CC0
664 && GET_CODE (SET_DEST (body)) != PC
665 && !REG_P (SET_DEST (body))
666 && ! (GET_CODE (SET_DEST (body)) == SUBREG
667 && REG_P (SUBREG_REG (SET_DEST (body)))
668 && (((GET_MODE_SIZE (GET_MODE (SUBREG_REG (SET_DEST (body))))
669 + (UNITS_PER_WORD - 1)) / UNITS_PER_WORD)
670 == ((GET_MODE_SIZE (GET_MODE (SET_DEST (body)))
671 + (UNITS_PER_WORD - 1)) / UNITS_PER_WORD)))
672 && reg_overlap_mentioned_p (x, SET_DEST (body)))
673 return 1;
674 return 0;
676 case ASM_OPERANDS:
677 for (i = ASM_OPERANDS_INPUT_LENGTH (body) - 1; i >= 0; i--)
678 if (reg_overlap_mentioned_p (x, ASM_OPERANDS_INPUT (body, i)))
679 return 1;
680 return 0;
682 case CALL:
683 case USE:
684 case IF_THEN_ELSE:
685 return reg_overlap_mentioned_p (x, body);
687 case TRAP_IF:
688 return reg_overlap_mentioned_p (x, TRAP_CONDITION (body));
690 case PREFETCH:
691 return reg_overlap_mentioned_p (x, XEXP (body, 0));
693 case UNSPEC:
694 case UNSPEC_VOLATILE:
695 for (i = XVECLEN (body, 0) - 1; i >= 0; i--)
696 if (reg_overlap_mentioned_p (x, XVECEXP (body, 0, i)))
697 return 1;
698 return 0;
700 case PARALLEL:
701 for (i = XVECLEN (body, 0) - 1; i >= 0; i--)
702 if (reg_referenced_p (x, XVECEXP (body, 0, i)))
703 return 1;
704 return 0;
706 case CLOBBER:
707 if (MEM_P (XEXP (body, 0)))
708 if (reg_overlap_mentioned_p (x, XEXP (XEXP (body, 0), 0)))
709 return 1;
710 return 0;
712 case COND_EXEC:
713 if (reg_overlap_mentioned_p (x, COND_EXEC_TEST (body)))
714 return 1;
715 return reg_referenced_p (x, COND_EXEC_CODE (body));
717 default:
718 return 0;
722 /* Nonzero if register REG is set or clobbered in an insn between
723 FROM_INSN and TO_INSN (exclusive of those two). */
726 reg_set_between_p (rtx reg, rtx from_insn, rtx to_insn)
728 rtx insn;
730 if (from_insn == to_insn)
731 return 0;
733 for (insn = NEXT_INSN (from_insn); insn != to_insn; insn = NEXT_INSN (insn))
734 if (INSN_P (insn) && reg_set_p (reg, insn))
735 return 1;
736 return 0;
739 /* Internals of reg_set_between_p. */
741 reg_set_p (rtx reg, rtx insn)
743 /* We can be passed an insn or part of one. If we are passed an insn,
744 check if a side-effect of the insn clobbers REG. */
745 if (INSN_P (insn)
746 && (FIND_REG_INC_NOTE (insn, reg)
747 || (CALL_P (insn)
748 && ((REG_P (reg)
749 && REGNO (reg) < FIRST_PSEUDO_REGISTER
750 && TEST_HARD_REG_BIT (regs_invalidated_by_call,
751 REGNO (reg)))
752 || MEM_P (reg)
753 || find_reg_fusage (insn, CLOBBER, reg)))))
754 return 1;
756 return set_of (reg, insn) != NULL_RTX;
759 /* Similar to reg_set_between_p, but check all registers in X. Return 0
760 only if none of them are modified between START and END. Return 1 if
761 X contains a MEM; this routine does usememory aliasing. */
764 modified_between_p (rtx x, rtx start, rtx end)
766 enum rtx_code code = GET_CODE (x);
767 const char *fmt;
768 int i, j;
769 rtx insn;
771 if (start == end)
772 return 0;
774 switch (code)
776 case CONST_INT:
777 case CONST_DOUBLE:
778 case CONST_VECTOR:
779 case CONST:
780 case SYMBOL_REF:
781 case LABEL_REF:
782 return 0;
784 case PC:
785 case CC0:
786 return 1;
788 case MEM:
789 if (modified_between_p (XEXP (x, 0), start, end))
790 return 1;
791 if (MEM_READONLY_P (x))
792 return 0;
793 for (insn = NEXT_INSN (start); insn != end; insn = NEXT_INSN (insn))
794 if (memory_modified_in_insn_p (x, insn))
795 return 1;
796 return 0;
797 break;
799 case REG:
800 return reg_set_between_p (x, start, end);
802 default:
803 break;
806 fmt = GET_RTX_FORMAT (code);
807 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
809 if (fmt[i] == 'e' && modified_between_p (XEXP (x, i), start, end))
810 return 1;
812 else if (fmt[i] == 'E')
813 for (j = XVECLEN (x, i) - 1; j >= 0; j--)
814 if (modified_between_p (XVECEXP (x, i, j), start, end))
815 return 1;
818 return 0;
821 /* Similar to reg_set_p, but check all registers in X. Return 0 only if none
822 of them are modified in INSN. Return 1 if X contains a MEM; this routine
823 does use memory aliasing. */
826 modified_in_p (rtx x, rtx insn)
828 enum rtx_code code = GET_CODE (x);
829 const char *fmt;
830 int i, j;
832 switch (code)
834 case CONST_INT:
835 case CONST_DOUBLE:
836 case CONST_VECTOR:
837 case CONST:
838 case SYMBOL_REF:
839 case LABEL_REF:
840 return 0;
842 case PC:
843 case CC0:
844 return 1;
846 case MEM:
847 if (modified_in_p (XEXP (x, 0), insn))
848 return 1;
849 if (MEM_READONLY_P (x))
850 return 0;
851 if (memory_modified_in_insn_p (x, insn))
852 return 1;
853 return 0;
854 break;
856 case REG:
857 return reg_set_p (x, insn);
859 default:
860 break;
863 fmt = GET_RTX_FORMAT (code);
864 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
866 if (fmt[i] == 'e' && modified_in_p (XEXP (x, i), insn))
867 return 1;
869 else if (fmt[i] == 'E')
870 for (j = XVECLEN (x, i) - 1; j >= 0; j--)
871 if (modified_in_p (XVECEXP (x, i, j), insn))
872 return 1;
875 return 0;
878 /* Helper function for set_of. */
879 struct set_of_data
881 rtx found;
882 rtx pat;
885 static void
886 set_of_1 (rtx x, rtx pat, void *data1)
888 struct set_of_data *data = (struct set_of_data *) (data1);
889 if (rtx_equal_p (x, data->pat)
890 || (!MEM_P (x) && reg_overlap_mentioned_p (data->pat, x)))
891 data->found = pat;
894 /* Give an INSN, return a SET or CLOBBER expression that does modify PAT
895 (either directly or via STRICT_LOW_PART and similar modifiers). */
897 set_of (rtx pat, rtx insn)
899 struct set_of_data data;
900 data.found = NULL_RTX;
901 data.pat = pat;
902 note_stores (INSN_P (insn) ? PATTERN (insn) : insn, set_of_1, &data);
903 return data.found;
906 /* Given an INSN, return a SET expression if this insn has only a single SET.
907 It may also have CLOBBERs, USEs, or SET whose output
908 will not be used, which we ignore. */
911 single_set_2 (rtx insn, rtx pat)
913 rtx set = NULL;
914 int set_verified = 1;
915 int i;
917 if (GET_CODE (pat) == PARALLEL)
919 for (i = 0; i < XVECLEN (pat, 0); i++)
921 rtx sub = XVECEXP (pat, 0, i);
922 switch (GET_CODE (sub))
924 case USE:
925 case CLOBBER:
926 break;
928 case SET:
929 /* We can consider insns having multiple sets, where all
930 but one are dead as single set insns. In common case
931 only single set is present in the pattern so we want
932 to avoid checking for REG_UNUSED notes unless necessary.
934 When we reach set first time, we just expect this is
935 the single set we are looking for and only when more
936 sets are found in the insn, we check them. */
937 if (!set_verified)
939 if (find_reg_note (insn, REG_UNUSED, SET_DEST (set))
940 && !side_effects_p (set))
941 set = NULL;
942 else
943 set_verified = 1;
945 if (!set)
946 set = sub, set_verified = 0;
947 else if (!find_reg_note (insn, REG_UNUSED, SET_DEST (sub))
948 || side_effects_p (sub))
949 return NULL_RTX;
950 break;
952 default:
953 return NULL_RTX;
957 return set;
960 /* Given an INSN, return nonzero if it has more than one SET, else return
961 zero. */
964 multiple_sets (rtx insn)
966 int found;
967 int i;
969 /* INSN must be an insn. */
970 if (! INSN_P (insn))
971 return 0;
973 /* Only a PARALLEL can have multiple SETs. */
974 if (GET_CODE (PATTERN (insn)) == PARALLEL)
976 for (i = 0, found = 0; i < XVECLEN (PATTERN (insn), 0); i++)
977 if (GET_CODE (XVECEXP (PATTERN (insn), 0, i)) == SET)
979 /* If we have already found a SET, then return now. */
980 if (found)
981 return 1;
982 else
983 found = 1;
987 /* Either zero or one SET. */
988 return 0;
991 /* Return nonzero if the destination of SET equals the source
992 and there are no side effects. */
995 set_noop_p (rtx set)
997 rtx src = SET_SRC (set);
998 rtx dst = SET_DEST (set);
1000 if (dst == pc_rtx && src == pc_rtx)
1001 return 1;
1003 if (MEM_P (dst) && MEM_P (src))
1004 return rtx_equal_p (dst, src) && !side_effects_p (dst);
1006 if (GET_CODE (dst) == ZERO_EXTRACT)
1007 return rtx_equal_p (XEXP (dst, 0), src)
1008 && ! BYTES_BIG_ENDIAN && XEXP (dst, 2) == const0_rtx
1009 && !side_effects_p (src);
1011 if (GET_CODE (dst) == STRICT_LOW_PART)
1012 dst = XEXP (dst, 0);
1014 if (GET_CODE (src) == SUBREG && GET_CODE (dst) == SUBREG)
1016 if (SUBREG_BYTE (src) != SUBREG_BYTE (dst))
1017 return 0;
1018 src = SUBREG_REG (src);
1019 dst = SUBREG_REG (dst);
1022 return (REG_P (src) && REG_P (dst)
1023 && REGNO (src) == REGNO (dst));
1026 /* Return nonzero if an insn consists only of SETs, each of which only sets a
1027 value to itself. */
1030 noop_move_p (rtx insn)
1032 rtx pat = PATTERN (insn);
1034 if (INSN_CODE (insn) == NOOP_MOVE_INSN_CODE)
1035 return 1;
1037 /* Insns carrying these notes are useful later on. */
1038 if (find_reg_note (insn, REG_EQUAL, NULL_RTX))
1039 return 0;
1041 /* For now treat an insn with a REG_RETVAL note as a
1042 a special insn which should not be considered a no-op. */
1043 if (find_reg_note (insn, REG_RETVAL, NULL_RTX))
1044 return 0;
1046 if (GET_CODE (pat) == SET && set_noop_p (pat))
1047 return 1;
1049 if (GET_CODE (pat) == PARALLEL)
1051 int i;
1052 /* If nothing but SETs of registers to themselves,
1053 this insn can also be deleted. */
1054 for (i = 0; i < XVECLEN (pat, 0); i++)
1056 rtx tem = XVECEXP (pat, 0, i);
1058 if (GET_CODE (tem) == USE
1059 || GET_CODE (tem) == CLOBBER)
1060 continue;
1062 if (GET_CODE (tem) != SET || ! set_noop_p (tem))
1063 return 0;
1066 return 1;
1068 return 0;
1072 /* Return the last thing that X was assigned from before *PINSN. If VALID_TO
1073 is not NULL_RTX then verify that the object is not modified up to VALID_TO.
1074 If the object was modified, if we hit a partial assignment to X, or hit a
1075 CODE_LABEL first, return X. If we found an assignment, update *PINSN to
1076 point to it. ALLOW_HWREG is set to 1 if hardware registers are allowed to
1077 be the src. */
1080 find_last_value (rtx x, rtx *pinsn, rtx valid_to, int allow_hwreg)
1082 rtx p;
1084 for (p = PREV_INSN (*pinsn); p && !LABEL_P (p);
1085 p = PREV_INSN (p))
1086 if (INSN_P (p))
1088 rtx set = single_set (p);
1089 rtx note = find_reg_note (p, REG_EQUAL, NULL_RTX);
1091 if (set && rtx_equal_p (x, SET_DEST (set)))
1093 rtx src = SET_SRC (set);
1095 if (note && GET_CODE (XEXP (note, 0)) != EXPR_LIST)
1096 src = XEXP (note, 0);
1098 if ((valid_to == NULL_RTX
1099 || ! modified_between_p (src, PREV_INSN (p), valid_to))
1100 /* Reject hard registers because we don't usually want
1101 to use them; we'd rather use a pseudo. */
1102 && (! (REG_P (src)
1103 && REGNO (src) < FIRST_PSEUDO_REGISTER) || allow_hwreg))
1105 *pinsn = p;
1106 return src;
1110 /* If set in non-simple way, we don't have a value. */
1111 if (reg_set_p (x, p))
1112 break;
1115 return x;
1118 /* Return nonzero if register in range [REGNO, ENDREGNO)
1119 appears either explicitly or implicitly in X
1120 other than being stored into.
1122 References contained within the substructure at LOC do not count.
1123 LOC may be zero, meaning don't ignore anything. */
1126 refers_to_regno_p (unsigned int regno, unsigned int endregno, rtx x,
1127 rtx *loc)
1129 int i;
1130 unsigned int x_regno;
1131 RTX_CODE code;
1132 const char *fmt;
1134 repeat:
1135 /* The contents of a REG_NONNEG note is always zero, so we must come here
1136 upon repeat in case the last REG_NOTE is a REG_NONNEG note. */
1137 if (x == 0)
1138 return 0;
1140 code = GET_CODE (x);
1142 switch (code)
1144 case REG:
1145 x_regno = REGNO (x);
1147 /* If we modifying the stack, frame, or argument pointer, it will
1148 clobber a virtual register. In fact, we could be more precise,
1149 but it isn't worth it. */
1150 if ((x_regno == STACK_POINTER_REGNUM
1151 #if FRAME_POINTER_REGNUM != ARG_POINTER_REGNUM
1152 || x_regno == ARG_POINTER_REGNUM
1153 #endif
1154 || x_regno == FRAME_POINTER_REGNUM)
1155 && regno >= FIRST_VIRTUAL_REGISTER && regno <= LAST_VIRTUAL_REGISTER)
1156 return 1;
1158 return (endregno > x_regno
1159 && regno < x_regno + (x_regno < FIRST_PSEUDO_REGISTER
1160 ? hard_regno_nregs[x_regno][GET_MODE (x)]
1161 : 1));
1163 case SUBREG:
1164 /* If this is a SUBREG of a hard reg, we can see exactly which
1165 registers are being modified. Otherwise, handle normally. */
1166 if (REG_P (SUBREG_REG (x))
1167 && REGNO (SUBREG_REG (x)) < FIRST_PSEUDO_REGISTER)
1169 unsigned int inner_regno = subreg_regno (x);
1170 unsigned int inner_endregno
1171 = inner_regno + (inner_regno < FIRST_PSEUDO_REGISTER
1172 ? hard_regno_nregs[inner_regno][GET_MODE (x)] : 1);
1174 return endregno > inner_regno && regno < inner_endregno;
1176 break;
1178 case CLOBBER:
1179 case SET:
1180 if (&SET_DEST (x) != loc
1181 /* Note setting a SUBREG counts as referring to the REG it is in for
1182 a pseudo but not for hard registers since we can
1183 treat each word individually. */
1184 && ((GET_CODE (SET_DEST (x)) == SUBREG
1185 && loc != &SUBREG_REG (SET_DEST (x))
1186 && REG_P (SUBREG_REG (SET_DEST (x)))
1187 && REGNO (SUBREG_REG (SET_DEST (x))) >= FIRST_PSEUDO_REGISTER
1188 && refers_to_regno_p (regno, endregno,
1189 SUBREG_REG (SET_DEST (x)), loc))
1190 || (!REG_P (SET_DEST (x))
1191 && refers_to_regno_p (regno, endregno, SET_DEST (x), loc))))
1192 return 1;
1194 if (code == CLOBBER || loc == &SET_SRC (x))
1195 return 0;
1196 x = SET_SRC (x);
1197 goto repeat;
1199 default:
1200 break;
1203 /* X does not match, so try its subexpressions. */
1205 fmt = GET_RTX_FORMAT (code);
1206 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
1208 if (fmt[i] == 'e' && loc != &XEXP (x, i))
1210 if (i == 0)
1212 x = XEXP (x, 0);
1213 goto repeat;
1215 else
1216 if (refers_to_regno_p (regno, endregno, XEXP (x, i), loc))
1217 return 1;
1219 else if (fmt[i] == 'E')
1221 int j;
1222 for (j = XVECLEN (x, i) - 1; j >= 0; j--)
1223 if (loc != &XVECEXP (x, i, j)
1224 && refers_to_regno_p (regno, endregno, XVECEXP (x, i, j), loc))
1225 return 1;
1228 return 0;
1231 /* Nonzero if modifying X will affect IN. If X is a register or a SUBREG,
1232 we check if any register number in X conflicts with the relevant register
1233 numbers. If X is a constant, return 0. If X is a MEM, return 1 iff IN
1234 contains a MEM (we don't bother checking for memory addresses that can't
1235 conflict because we expect this to be a rare case. */
1238 reg_overlap_mentioned_p (rtx x, rtx in)
1240 unsigned int regno, endregno;
1242 /* If either argument is a constant, then modifying X can not
1243 affect IN. Here we look at IN, we can profitably combine
1244 CONSTANT_P (x) with the switch statement below. */
1245 if (CONSTANT_P (in))
1246 return 0;
1248 recurse:
1249 switch (GET_CODE (x))
1251 case STRICT_LOW_PART:
1252 case ZERO_EXTRACT:
1253 case SIGN_EXTRACT:
1254 /* Overly conservative. */
1255 x = XEXP (x, 0);
1256 goto recurse;
1258 case SUBREG:
1259 regno = REGNO (SUBREG_REG (x));
1260 if (regno < FIRST_PSEUDO_REGISTER)
1261 regno = subreg_regno (x);
1262 goto do_reg;
1264 case REG:
1265 regno = REGNO (x);
1266 do_reg:
1267 endregno = regno + (regno < FIRST_PSEUDO_REGISTER
1268 ? hard_regno_nregs[regno][GET_MODE (x)] : 1);
1269 return refers_to_regno_p (regno, endregno, in, (rtx*) 0);
1271 case MEM:
1273 const char *fmt;
1274 int i;
1276 if (MEM_P (in))
1277 return 1;
1279 fmt = GET_RTX_FORMAT (GET_CODE (in));
1280 for (i = GET_RTX_LENGTH (GET_CODE (in)) - 1; i >= 0; i--)
1281 if (fmt[i] == 'e')
1283 if (reg_overlap_mentioned_p (x, XEXP (in, i)))
1284 return 1;
1286 else if (fmt[i] == 'E')
1288 int j;
1289 for (j = XVECLEN (in, i) - 1; j >= 0; --j)
1290 if (reg_overlap_mentioned_p (x, XVECEXP (in, i, j)))
1291 return 1;
1294 return 0;
1297 case SCRATCH:
1298 case PC:
1299 case CC0:
1300 return reg_mentioned_p (x, in);
1302 case PARALLEL:
1304 int i;
1306 /* If any register in here refers to it we return true. */
1307 for (i = XVECLEN (x, 0) - 1; i >= 0; i--)
1308 if (XEXP (XVECEXP (x, 0, i), 0) != 0
1309 && reg_overlap_mentioned_p (XEXP (XVECEXP (x, 0, i), 0), in))
1310 return 1;
1311 return 0;
1314 default:
1315 gcc_assert (CONSTANT_P (x));
1316 return 0;
1320 /* Call FUN on each register or MEM that is stored into or clobbered by X.
1321 (X would be the pattern of an insn).
1322 FUN receives two arguments:
1323 the REG, MEM, CC0 or PC being stored in or clobbered,
1324 the SET or CLOBBER rtx that does the store.
1326 If the item being stored in or clobbered is a SUBREG of a hard register,
1327 the SUBREG will be passed. */
1329 void
1330 note_stores (rtx x, void (*fun) (rtx, rtx, void *), void *data)
1332 int i;
1334 if (GET_CODE (x) == COND_EXEC)
1335 x = COND_EXEC_CODE (x);
1337 if (GET_CODE (x) == SET || GET_CODE (x) == CLOBBER)
1339 rtx dest = SET_DEST (x);
1341 while ((GET_CODE (dest) == SUBREG
1342 && (!REG_P (SUBREG_REG (dest))
1343 || REGNO (SUBREG_REG (dest)) >= FIRST_PSEUDO_REGISTER))
1344 || GET_CODE (dest) == ZERO_EXTRACT
1345 || GET_CODE (dest) == STRICT_LOW_PART)
1346 dest = XEXP (dest, 0);
1348 /* If we have a PARALLEL, SET_DEST is a list of EXPR_LIST expressions,
1349 each of whose first operand is a register. */
1350 if (GET_CODE (dest) == PARALLEL)
1352 for (i = XVECLEN (dest, 0) - 1; i >= 0; i--)
1353 if (XEXP (XVECEXP (dest, 0, i), 0) != 0)
1354 (*fun) (XEXP (XVECEXP (dest, 0, i), 0), x, data);
1356 else
1357 (*fun) (dest, x, data);
1360 else if (GET_CODE (x) == PARALLEL)
1361 for (i = XVECLEN (x, 0) - 1; i >= 0; i--)
1362 note_stores (XVECEXP (x, 0, i), fun, data);
1365 /* Like notes_stores, but call FUN for each expression that is being
1366 referenced in PBODY, a pointer to the PATTERN of an insn. We only call
1367 FUN for each expression, not any interior subexpressions. FUN receives a
1368 pointer to the expression and the DATA passed to this function.
1370 Note that this is not quite the same test as that done in reg_referenced_p
1371 since that considers something as being referenced if it is being
1372 partially set, while we do not. */
1374 void
1375 note_uses (rtx *pbody, void (*fun) (rtx *, void *), void *data)
1377 rtx body = *pbody;
1378 int i;
1380 switch (GET_CODE (body))
1382 case COND_EXEC:
1383 (*fun) (&COND_EXEC_TEST (body), data);
1384 note_uses (&COND_EXEC_CODE (body), fun, data);
1385 return;
1387 case PARALLEL:
1388 for (i = XVECLEN (body, 0) - 1; i >= 0; i--)
1389 note_uses (&XVECEXP (body, 0, i), fun, data);
1390 return;
1392 case USE:
1393 (*fun) (&XEXP (body, 0), data);
1394 return;
1396 case ASM_OPERANDS:
1397 for (i = ASM_OPERANDS_INPUT_LENGTH (body) - 1; i >= 0; i--)
1398 (*fun) (&ASM_OPERANDS_INPUT (body, i), data);
1399 return;
1401 case TRAP_IF:
1402 (*fun) (&TRAP_CONDITION (body), data);
1403 return;
1405 case PREFETCH:
1406 (*fun) (&XEXP (body, 0), data);
1407 return;
1409 case UNSPEC:
1410 case UNSPEC_VOLATILE:
1411 for (i = XVECLEN (body, 0) - 1; i >= 0; i--)
1412 (*fun) (&XVECEXP (body, 0, i), data);
1413 return;
1415 case CLOBBER:
1416 if (MEM_P (XEXP (body, 0)))
1417 (*fun) (&XEXP (XEXP (body, 0), 0), data);
1418 return;
1420 case SET:
1422 rtx dest = SET_DEST (body);
1424 /* For sets we replace everything in source plus registers in memory
1425 expression in store and operands of a ZERO_EXTRACT. */
1426 (*fun) (&SET_SRC (body), data);
1428 if (GET_CODE (dest) == ZERO_EXTRACT)
1430 (*fun) (&XEXP (dest, 1), data);
1431 (*fun) (&XEXP (dest, 2), data);
1434 while (GET_CODE (dest) == SUBREG || GET_CODE (dest) == STRICT_LOW_PART)
1435 dest = XEXP (dest, 0);
1437 if (MEM_P (dest))
1438 (*fun) (&XEXP (dest, 0), data);
1440 return;
1442 default:
1443 /* All the other possibilities never store. */
1444 (*fun) (pbody, data);
1445 return;
1449 /* Return nonzero if X's old contents don't survive after INSN.
1450 This will be true if X is (cc0) or if X is a register and
1451 X dies in INSN or because INSN entirely sets X.
1453 "Entirely set" means set directly and not through a SUBREG, or
1454 ZERO_EXTRACT, so no trace of the old contents remains.
1455 Likewise, REG_INC does not count.
1457 REG may be a hard or pseudo reg. Renumbering is not taken into account,
1458 but for this use that makes no difference, since regs don't overlap
1459 during their lifetimes. Therefore, this function may be used
1460 at any time after deaths have been computed (in flow.c).
1462 If REG is a hard reg that occupies multiple machine registers, this
1463 function will only return 1 if each of those registers will be replaced
1464 by INSN. */
1467 dead_or_set_p (rtx insn, rtx x)
1469 unsigned int regno, last_regno;
1470 unsigned int i;
1472 /* Can't use cc0_rtx below since this file is used by genattrtab.c. */
1473 if (GET_CODE (x) == CC0)
1474 return 1;
1476 gcc_assert (REG_P (x));
1478 regno = REGNO (x);
1479 last_regno = (regno >= FIRST_PSEUDO_REGISTER ? regno
1480 : regno + hard_regno_nregs[regno][GET_MODE (x)] - 1);
1482 for (i = regno; i <= last_regno; i++)
1483 if (! dead_or_set_regno_p (insn, i))
1484 return 0;
1486 return 1;
1489 /* Return TRUE iff DEST is a register or subreg of a register and
1490 doesn't change the number of words of the inner register, and any
1491 part of the register is TEST_REGNO. */
1493 static bool
1494 covers_regno_no_parallel_p (rtx dest, unsigned int test_regno)
1496 unsigned int regno, endregno;
1498 if (GET_CODE (dest) == SUBREG
1499 && (((GET_MODE_SIZE (GET_MODE (dest))
1500 + UNITS_PER_WORD - 1) / UNITS_PER_WORD)
1501 == ((GET_MODE_SIZE (GET_MODE (SUBREG_REG (dest)))
1502 + UNITS_PER_WORD - 1) / UNITS_PER_WORD)))
1503 dest = SUBREG_REG (dest);
1505 if (!REG_P (dest))
1506 return false;
1508 regno = REGNO (dest);
1509 endregno = (regno >= FIRST_PSEUDO_REGISTER ? regno + 1
1510 : regno + hard_regno_nregs[regno][GET_MODE (dest)]);
1511 return (test_regno >= regno && test_regno < endregno);
1514 /* Like covers_regno_no_parallel_p, but also handles PARALLELs where
1515 any member matches the covers_regno_no_parallel_p criteria. */
1517 static bool
1518 covers_regno_p (rtx dest, unsigned int test_regno)
1520 if (GET_CODE (dest) == PARALLEL)
1522 /* Some targets place small structures in registers for return
1523 values of functions, and those registers are wrapped in
1524 PARALLELs that we may see as the destination of a SET. */
1525 int i;
1527 for (i = XVECLEN (dest, 0) - 1; i >= 0; i--)
1529 rtx inner = XEXP (XVECEXP (dest, 0, i), 0);
1530 if (inner != NULL_RTX
1531 && covers_regno_no_parallel_p (inner, test_regno))
1532 return true;
1535 return false;
1537 else
1538 return covers_regno_no_parallel_p (dest, test_regno);
1541 /* Utility function for dead_or_set_p to check an individual register. Also
1542 called from flow.c. */
1545 dead_or_set_regno_p (rtx insn, unsigned int test_regno)
1547 rtx pattern;
1549 /* See if there is a death note for something that includes TEST_REGNO. */
1550 if (find_regno_note (insn, REG_DEAD, test_regno))
1551 return 1;
1553 if (CALL_P (insn)
1554 && find_regno_fusage (insn, CLOBBER, test_regno))
1555 return 1;
1557 pattern = PATTERN (insn);
1559 if (GET_CODE (pattern) == COND_EXEC)
1560 pattern = COND_EXEC_CODE (pattern);
1562 if (GET_CODE (pattern) == SET)
1563 return covers_regno_p (SET_DEST (pattern), test_regno);
1564 else if (GET_CODE (pattern) == PARALLEL)
1566 int i;
1568 for (i = XVECLEN (pattern, 0) - 1; i >= 0; i--)
1570 rtx body = XVECEXP (pattern, 0, i);
1572 if (GET_CODE (body) == COND_EXEC)
1573 body = COND_EXEC_CODE (body);
1575 if ((GET_CODE (body) == SET || GET_CODE (body) == CLOBBER)
1576 && covers_regno_p (SET_DEST (body), test_regno))
1577 return 1;
1581 return 0;
1584 /* Return the reg-note of kind KIND in insn INSN, if there is one.
1585 If DATUM is nonzero, look for one whose datum is DATUM. */
1588 find_reg_note (rtx insn, enum reg_note kind, rtx datum)
1590 rtx link;
1592 gcc_assert (insn);
1594 /* Ignore anything that is not an INSN, JUMP_INSN or CALL_INSN. */
1595 if (! INSN_P (insn))
1596 return 0;
1597 if (datum == 0)
1599 for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
1600 if (REG_NOTE_KIND (link) == kind)
1601 return link;
1602 return 0;
1605 for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
1606 if (REG_NOTE_KIND (link) == kind && datum == XEXP (link, 0))
1607 return link;
1608 return 0;
1611 /* Return the reg-note of kind KIND in insn INSN which applies to register
1612 number REGNO, if any. Return 0 if there is no such reg-note. Note that
1613 the REGNO of this NOTE need not be REGNO if REGNO is a hard register;
1614 it might be the case that the note overlaps REGNO. */
1617 find_regno_note (rtx insn, enum reg_note kind, unsigned int regno)
1619 rtx link;
1621 /* Ignore anything that is not an INSN, JUMP_INSN or CALL_INSN. */
1622 if (! INSN_P (insn))
1623 return 0;
1625 for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
1626 if (REG_NOTE_KIND (link) == kind
1627 /* Verify that it is a register, so that scratch and MEM won't cause a
1628 problem here. */
1629 && REG_P (XEXP (link, 0))
1630 && REGNO (XEXP (link, 0)) <= regno
1631 && ((REGNO (XEXP (link, 0))
1632 + (REGNO (XEXP (link, 0)) >= FIRST_PSEUDO_REGISTER ? 1
1633 : hard_regno_nregs[REGNO (XEXP (link, 0))]
1634 [GET_MODE (XEXP (link, 0))]))
1635 > regno))
1636 return link;
1637 return 0;
1640 /* Return a REG_EQUIV or REG_EQUAL note if insn has only a single set and
1641 has such a note. */
1644 find_reg_equal_equiv_note (rtx insn)
1646 rtx link;
1648 if (!INSN_P (insn))
1649 return 0;
1650 for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
1651 if (REG_NOTE_KIND (link) == REG_EQUAL
1652 || REG_NOTE_KIND (link) == REG_EQUIV)
1654 if (single_set (insn) == 0)
1655 return 0;
1656 return link;
1658 return NULL;
1661 /* Return true if DATUM, or any overlap of DATUM, of kind CODE is found
1662 in the CALL_INSN_FUNCTION_USAGE information of INSN. */
1665 find_reg_fusage (rtx insn, enum rtx_code code, rtx datum)
1667 /* If it's not a CALL_INSN, it can't possibly have a
1668 CALL_INSN_FUNCTION_USAGE field, so don't bother checking. */
1669 if (!CALL_P (insn))
1670 return 0;
1672 gcc_assert (datum);
1674 if (!REG_P (datum))
1676 rtx link;
1678 for (link = CALL_INSN_FUNCTION_USAGE (insn);
1679 link;
1680 link = XEXP (link, 1))
1681 if (GET_CODE (XEXP (link, 0)) == code
1682 && rtx_equal_p (datum, XEXP (XEXP (link, 0), 0)))
1683 return 1;
1685 else
1687 unsigned int regno = REGNO (datum);
1689 /* CALL_INSN_FUNCTION_USAGE information cannot contain references
1690 to pseudo registers, so don't bother checking. */
1692 if (regno < FIRST_PSEUDO_REGISTER)
1694 unsigned int end_regno
1695 = regno + hard_regno_nregs[regno][GET_MODE (datum)];
1696 unsigned int i;
1698 for (i = regno; i < end_regno; i++)
1699 if (find_regno_fusage (insn, code, i))
1700 return 1;
1704 return 0;
1707 /* Return true if REGNO, or any overlap of REGNO, of kind CODE is found
1708 in the CALL_INSN_FUNCTION_USAGE information of INSN. */
1711 find_regno_fusage (rtx insn, enum rtx_code code, unsigned int regno)
1713 rtx link;
1715 /* CALL_INSN_FUNCTION_USAGE information cannot contain references
1716 to pseudo registers, so don't bother checking. */
1718 if (regno >= FIRST_PSEUDO_REGISTER
1719 || !CALL_P (insn) )
1720 return 0;
1722 for (link = CALL_INSN_FUNCTION_USAGE (insn); link; link = XEXP (link, 1))
1724 unsigned int regnote;
1725 rtx op, reg;
1727 if (GET_CODE (op = XEXP (link, 0)) == code
1728 && REG_P (reg = XEXP (op, 0))
1729 && (regnote = REGNO (reg)) <= regno
1730 && regnote + hard_regno_nregs[regnote][GET_MODE (reg)] > regno)
1731 return 1;
1734 return 0;
1737 /* Return true if INSN is a call to a pure function. */
1740 pure_call_p (rtx insn)
1742 rtx link;
1744 if (!CALL_P (insn) || ! CONST_OR_PURE_CALL_P (insn))
1745 return 0;
1747 /* Look for the note that differentiates const and pure functions. */
1748 for (link = CALL_INSN_FUNCTION_USAGE (insn); link; link = XEXP (link, 1))
1750 rtx u, m;
1752 if (GET_CODE (u = XEXP (link, 0)) == USE
1753 && MEM_P (m = XEXP (u, 0)) && GET_MODE (m) == BLKmode
1754 && GET_CODE (XEXP (m, 0)) == SCRATCH)
1755 return 1;
1758 return 0;
1761 /* Remove register note NOTE from the REG_NOTES of INSN. */
1763 void
1764 remove_note (rtx insn, rtx note)
1766 rtx link;
1768 if (note == NULL_RTX)
1769 return;
1771 if (REG_NOTES (insn) == note)
1773 REG_NOTES (insn) = XEXP (note, 1);
1774 return;
1777 for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
1778 if (XEXP (link, 1) == note)
1780 XEXP (link, 1) = XEXP (note, 1);
1781 return;
1784 gcc_unreachable ();
1787 /* Search LISTP (an EXPR_LIST) for an entry whose first operand is NODE and
1788 return 1 if it is found. A simple equality test is used to determine if
1789 NODE matches. */
1792 in_expr_list_p (rtx listp, rtx node)
1794 rtx x;
1796 for (x = listp; x; x = XEXP (x, 1))
1797 if (node == XEXP (x, 0))
1798 return 1;
1800 return 0;
1803 /* Search LISTP (an EXPR_LIST) for an entry whose first operand is NODE and
1804 remove that entry from the list if it is found.
1806 A simple equality test is used to determine if NODE matches. */
1808 void
1809 remove_node_from_expr_list (rtx node, rtx *listp)
1811 rtx temp = *listp;
1812 rtx prev = NULL_RTX;
1814 while (temp)
1816 if (node == XEXP (temp, 0))
1818 /* Splice the node out of the list. */
1819 if (prev)
1820 XEXP (prev, 1) = XEXP (temp, 1);
1821 else
1822 *listp = XEXP (temp, 1);
1824 return;
1827 prev = temp;
1828 temp = XEXP (temp, 1);
1832 /* Nonzero if X contains any volatile instructions. These are instructions
1833 which may cause unpredictable machine state instructions, and thus no
1834 instructions should be moved or combined across them. This includes
1835 only volatile asms and UNSPEC_VOLATILE instructions. */
1838 volatile_insn_p (rtx x)
1840 RTX_CODE code;
1842 code = GET_CODE (x);
1843 switch (code)
1845 case LABEL_REF:
1846 case SYMBOL_REF:
1847 case CONST_INT:
1848 case CONST:
1849 case CONST_DOUBLE:
1850 case CONST_VECTOR:
1851 case CC0:
1852 case PC:
1853 case REG:
1854 case SCRATCH:
1855 case CLOBBER:
1856 case ADDR_VEC:
1857 case ADDR_DIFF_VEC:
1858 case CALL:
1859 case MEM:
1860 return 0;
1862 case UNSPEC_VOLATILE:
1863 /* case TRAP_IF: This isn't clear yet. */
1864 return 1;
1866 case ASM_INPUT:
1867 case ASM_OPERANDS:
1868 if (MEM_VOLATILE_P (x))
1869 return 1;
1871 default:
1872 break;
1875 /* Recursively scan the operands of this expression. */
1878 const char *fmt = GET_RTX_FORMAT (code);
1879 int i;
1881 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
1883 if (fmt[i] == 'e')
1885 if (volatile_insn_p (XEXP (x, i)))
1886 return 1;
1888 else if (fmt[i] == 'E')
1890 int j;
1891 for (j = 0; j < XVECLEN (x, i); j++)
1892 if (volatile_insn_p (XVECEXP (x, i, j)))
1893 return 1;
1897 return 0;
1900 /* Nonzero if X contains any volatile memory references
1901 UNSPEC_VOLATILE operations or volatile ASM_OPERANDS expressions. */
1904 volatile_refs_p (rtx x)
1906 RTX_CODE code;
1908 code = GET_CODE (x);
1909 switch (code)
1911 case LABEL_REF:
1912 case SYMBOL_REF:
1913 case CONST_INT:
1914 case CONST:
1915 case CONST_DOUBLE:
1916 case CONST_VECTOR:
1917 case CC0:
1918 case PC:
1919 case REG:
1920 case SCRATCH:
1921 case CLOBBER:
1922 case ADDR_VEC:
1923 case ADDR_DIFF_VEC:
1924 return 0;
1926 case UNSPEC_VOLATILE:
1927 return 1;
1929 case MEM:
1930 case ASM_INPUT:
1931 case ASM_OPERANDS:
1932 if (MEM_VOLATILE_P (x))
1933 return 1;
1935 default:
1936 break;
1939 /* Recursively scan the operands of this expression. */
1942 const char *fmt = GET_RTX_FORMAT (code);
1943 int i;
1945 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
1947 if (fmt[i] == 'e')
1949 if (volatile_refs_p (XEXP (x, i)))
1950 return 1;
1952 else if (fmt[i] == 'E')
1954 int j;
1955 for (j = 0; j < XVECLEN (x, i); j++)
1956 if (volatile_refs_p (XVECEXP (x, i, j)))
1957 return 1;
1961 return 0;
1964 /* Similar to above, except that it also rejects register pre- and post-
1965 incrementing. */
1968 side_effects_p (rtx x)
1970 RTX_CODE code;
1972 code = GET_CODE (x);
1973 switch (code)
1975 case LABEL_REF:
1976 case SYMBOL_REF:
1977 case CONST_INT:
1978 case CONST:
1979 case CONST_DOUBLE:
1980 case CONST_VECTOR:
1981 case CC0:
1982 case PC:
1983 case REG:
1984 case SCRATCH:
1985 case ADDR_VEC:
1986 case ADDR_DIFF_VEC:
1987 return 0;
1989 case CLOBBER:
1990 /* Reject CLOBBER with a non-VOID mode. These are made by combine.c
1991 when some combination can't be done. If we see one, don't think
1992 that we can simplify the expression. */
1993 return (GET_MODE (x) != VOIDmode);
1995 case PRE_INC:
1996 case PRE_DEC:
1997 case POST_INC:
1998 case POST_DEC:
1999 case PRE_MODIFY:
2000 case POST_MODIFY:
2001 case CALL:
2002 case UNSPEC_VOLATILE:
2003 /* case TRAP_IF: This isn't clear yet. */
2004 return 1;
2006 case MEM:
2007 case ASM_INPUT:
2008 case ASM_OPERANDS:
2009 if (MEM_VOLATILE_P (x))
2010 return 1;
2012 default:
2013 break;
2016 /* Recursively scan the operands of this expression. */
2019 const char *fmt = GET_RTX_FORMAT (code);
2020 int i;
2022 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
2024 if (fmt[i] == 'e')
2026 if (side_effects_p (XEXP (x, i)))
2027 return 1;
2029 else if (fmt[i] == 'E')
2031 int j;
2032 for (j = 0; j < XVECLEN (x, i); j++)
2033 if (side_effects_p (XVECEXP (x, i, j)))
2034 return 1;
2038 return 0;
2041 enum may_trap_p_flags
2043 MTP_UNALIGNED_MEMS = 1,
2044 MTP_AFTER_MOVE = 2
2046 /* Return nonzero if evaluating rtx X might cause a trap.
2047 (FLAGS & MTP_UNALIGNED_MEMS) controls whether nonzero is returned for
2048 unaligned memory accesses on strict alignment machines. If
2049 (FLAGS & AFTER_MOVE) is true, returns nonzero even in case the expression
2050 cannot trap at its current location, but it might become trapping if moved
2051 elsewhere. */
2053 static int
2054 may_trap_p_1 (rtx x, unsigned flags)
2056 int i;
2057 enum rtx_code code;
2058 const char *fmt;
2059 bool unaligned_mems = (flags & MTP_UNALIGNED_MEMS) != 0;
2061 if (x == 0)
2062 return 0;
2063 code = GET_CODE (x);
2064 switch (code)
2066 /* Handle these cases quickly. */
2067 case CONST_INT:
2068 case CONST_DOUBLE:
2069 case CONST_VECTOR:
2070 case SYMBOL_REF:
2071 case LABEL_REF:
2072 case CONST:
2073 case PC:
2074 case CC0:
2075 case REG:
2076 case SCRATCH:
2077 return 0;
2079 case ASM_INPUT:
2080 case UNSPEC_VOLATILE:
2081 case TRAP_IF:
2082 return 1;
2084 case ASM_OPERANDS:
2085 return MEM_VOLATILE_P (x);
2087 /* Memory ref can trap unless it's a static var or a stack slot. */
2088 case MEM:
2089 if (/* MEM_NOTRAP_P only relates to the actual position of the memory
2090 reference; moving it out of condition might cause its address
2091 become invalid. */
2092 !(flags & MTP_AFTER_MOVE)
2093 && MEM_NOTRAP_P (x)
2094 && (!STRICT_ALIGNMENT || !unaligned_mems))
2095 return 0;
2096 return
2097 rtx_addr_can_trap_p_1 (XEXP (x, 0), GET_MODE (x), unaligned_mems);
2099 /* Division by a non-constant might trap. */
2100 case DIV:
2101 case MOD:
2102 case UDIV:
2103 case UMOD:
2104 if (HONOR_SNANS (GET_MODE (x)))
2105 return 1;
2106 if (SCALAR_FLOAT_MODE_P (GET_MODE (x)))
2107 return flag_trapping_math;
2108 if (!CONSTANT_P (XEXP (x, 1)) || (XEXP (x, 1) == const0_rtx))
2109 return 1;
2110 break;
2112 case EXPR_LIST:
2113 /* An EXPR_LIST is used to represent a function call. This
2114 certainly may trap. */
2115 return 1;
2117 case GE:
2118 case GT:
2119 case LE:
2120 case LT:
2121 case LTGT:
2122 case COMPARE:
2123 /* Some floating point comparisons may trap. */
2124 if (!flag_trapping_math)
2125 break;
2126 /* ??? There is no machine independent way to check for tests that trap
2127 when COMPARE is used, though many targets do make this distinction.
2128 For instance, sparc uses CCFPE for compares which generate exceptions
2129 and CCFP for compares which do not generate exceptions. */
2130 if (HONOR_NANS (GET_MODE (x)))
2131 return 1;
2132 /* But often the compare has some CC mode, so check operand
2133 modes as well. */
2134 if (HONOR_NANS (GET_MODE (XEXP (x, 0)))
2135 || HONOR_NANS (GET_MODE (XEXP (x, 1))))
2136 return 1;
2137 break;
2139 case EQ:
2140 case NE:
2141 if (HONOR_SNANS (GET_MODE (x)))
2142 return 1;
2143 /* Often comparison is CC mode, so check operand modes. */
2144 if (HONOR_SNANS (GET_MODE (XEXP (x, 0)))
2145 || HONOR_SNANS (GET_MODE (XEXP (x, 1))))
2146 return 1;
2147 break;
2149 case FIX:
2150 /* Conversion of floating point might trap. */
2151 if (flag_trapping_math && HONOR_NANS (GET_MODE (XEXP (x, 0))))
2152 return 1;
2153 break;
2155 case NEG:
2156 case ABS:
2157 case SUBREG:
2158 /* These operations don't trap even with floating point. */
2159 break;
2161 default:
2162 /* Any floating arithmetic may trap. */
2163 if (SCALAR_FLOAT_MODE_P (GET_MODE (x))
2164 && flag_trapping_math)
2165 return 1;
2168 fmt = GET_RTX_FORMAT (code);
2169 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
2171 if (fmt[i] == 'e')
2173 if (may_trap_p_1 (XEXP (x, i), flags))
2174 return 1;
2176 else if (fmt[i] == 'E')
2178 int j;
2179 for (j = 0; j < XVECLEN (x, i); j++)
2180 if (may_trap_p_1 (XVECEXP (x, i, j), flags))
2181 return 1;
2184 return 0;
2187 /* Return nonzero if evaluating rtx X might cause a trap. */
2190 may_trap_p (rtx x)
2192 return may_trap_p_1 (x, 0);
2195 /* Return nonzero if evaluating rtx X might cause a trap, when the expression
2196 is moved from its current location by some optimization. */
2199 may_trap_after_code_motion_p (rtx x)
2201 return may_trap_p_1 (x, MTP_AFTER_MOVE);
2204 /* Same as above, but additionally return nonzero if evaluating rtx X might
2205 cause a fault. We define a fault for the purpose of this function as a
2206 erroneous execution condition that cannot be encountered during the normal
2207 execution of a valid program; the typical example is an unaligned memory
2208 access on a strict alignment machine. The compiler guarantees that it
2209 doesn't generate code that will fault from a valid program, but this
2210 guarantee doesn't mean anything for individual instructions. Consider
2211 the following example:
2213 struct S { int d; union { char *cp; int *ip; }; };
2215 int foo(struct S *s)
2217 if (s->d == 1)
2218 return *s->ip;
2219 else
2220 return *s->cp;
2223 on a strict alignment machine. In a valid program, foo will never be
2224 invoked on a structure for which d is equal to 1 and the underlying
2225 unique field of the union not aligned on a 4-byte boundary, but the
2226 expression *s->ip might cause a fault if considered individually.
2228 At the RTL level, potentially problematic expressions will almost always
2229 verify may_trap_p; for example, the above dereference can be emitted as
2230 (mem:SI (reg:P)) and this expression is may_trap_p for a generic register.
2231 However, suppose that foo is inlined in a caller that causes s->cp to
2232 point to a local character variable and guarantees that s->d is not set
2233 to 1; foo may have been effectively translated into pseudo-RTL as:
2235 if ((reg:SI) == 1)
2236 (set (reg:SI) (mem:SI (%fp - 7)))
2237 else
2238 (set (reg:QI) (mem:QI (%fp - 7)))
2240 Now (mem:SI (%fp - 7)) is considered as not may_trap_p since it is a
2241 memory reference to a stack slot, but it will certainly cause a fault
2242 on a strict alignment machine. */
2245 may_trap_or_fault_p (rtx x)
2247 return may_trap_p_1 (x, MTP_UNALIGNED_MEMS);
2250 /* Return nonzero if X contains a comparison that is not either EQ or NE,
2251 i.e., an inequality. */
2254 inequality_comparisons_p (rtx x)
2256 const char *fmt;
2257 int len, i;
2258 enum rtx_code code = GET_CODE (x);
2260 switch (code)
2262 case REG:
2263 case SCRATCH:
2264 case PC:
2265 case CC0:
2266 case CONST_INT:
2267 case CONST_DOUBLE:
2268 case CONST_VECTOR:
2269 case CONST:
2270 case LABEL_REF:
2271 case SYMBOL_REF:
2272 return 0;
2274 case LT:
2275 case LTU:
2276 case GT:
2277 case GTU:
2278 case LE:
2279 case LEU:
2280 case GE:
2281 case GEU:
2282 return 1;
2284 default:
2285 break;
2288 len = GET_RTX_LENGTH (code);
2289 fmt = GET_RTX_FORMAT (code);
2291 for (i = 0; i < len; i++)
2293 if (fmt[i] == 'e')
2295 if (inequality_comparisons_p (XEXP (x, i)))
2296 return 1;
2298 else if (fmt[i] == 'E')
2300 int j;
2301 for (j = XVECLEN (x, i) - 1; j >= 0; j--)
2302 if (inequality_comparisons_p (XVECEXP (x, i, j)))
2303 return 1;
2307 return 0;
2310 /* Replace any occurrence of FROM in X with TO. The function does
2311 not enter into CONST_DOUBLE for the replace.
2313 Note that copying is not done so X must not be shared unless all copies
2314 are to be modified. */
2317 replace_rtx (rtx x, rtx from, rtx to)
2319 int i, j;
2320 const char *fmt;
2322 /* The following prevents loops occurrence when we change MEM in
2323 CONST_DOUBLE onto the same CONST_DOUBLE. */
2324 if (x != 0 && GET_CODE (x) == CONST_DOUBLE)
2325 return x;
2327 if (x == from)
2328 return to;
2330 /* Allow this function to make replacements in EXPR_LISTs. */
2331 if (x == 0)
2332 return 0;
2334 if (GET_CODE (x) == SUBREG)
2336 rtx new = replace_rtx (SUBREG_REG (x), from, to);
2338 if (GET_CODE (new) == CONST_INT)
2340 x = simplify_subreg (GET_MODE (x), new,
2341 GET_MODE (SUBREG_REG (x)),
2342 SUBREG_BYTE (x));
2343 gcc_assert (x);
2345 else
2346 SUBREG_REG (x) = new;
2348 return x;
2350 else if (GET_CODE (x) == ZERO_EXTEND)
2352 rtx new = replace_rtx (XEXP (x, 0), from, to);
2354 if (GET_CODE (new) == CONST_INT)
2356 x = simplify_unary_operation (ZERO_EXTEND, GET_MODE (x),
2357 new, GET_MODE (XEXP (x, 0)));
2358 gcc_assert (x);
2360 else
2361 XEXP (x, 0) = new;
2363 return x;
2366 fmt = GET_RTX_FORMAT (GET_CODE (x));
2367 for (i = GET_RTX_LENGTH (GET_CODE (x)) - 1; i >= 0; i--)
2369 if (fmt[i] == 'e')
2370 XEXP (x, i) = replace_rtx (XEXP (x, i), from, to);
2371 else if (fmt[i] == 'E')
2372 for (j = XVECLEN (x, i) - 1; j >= 0; j--)
2373 XVECEXP (x, i, j) = replace_rtx (XVECEXP (x, i, j), from, to);
2376 return x;
2379 /* Replace occurrences of the old label in *X with the new one.
2380 DATA is a REPLACE_LABEL_DATA containing the old and new labels. */
2383 replace_label (rtx *x, void *data)
2385 rtx l = *x;
2386 rtx old_label = ((replace_label_data *) data)->r1;
2387 rtx new_label = ((replace_label_data *) data)->r2;
2388 bool update_label_nuses = ((replace_label_data *) data)->update_label_nuses;
2390 if (l == NULL_RTX)
2391 return 0;
2393 if (GET_CODE (l) == SYMBOL_REF
2394 && CONSTANT_POOL_ADDRESS_P (l))
2396 rtx c = get_pool_constant (l);
2397 if (rtx_referenced_p (old_label, c))
2399 rtx new_c, new_l;
2400 replace_label_data *d = (replace_label_data *) data;
2402 /* Create a copy of constant C; replace the label inside
2403 but do not update LABEL_NUSES because uses in constant pool
2404 are not counted. */
2405 new_c = copy_rtx (c);
2406 d->update_label_nuses = false;
2407 for_each_rtx (&new_c, replace_label, data);
2408 d->update_label_nuses = update_label_nuses;
2410 /* Add the new constant NEW_C to constant pool and replace
2411 the old reference to constant by new reference. */
2412 new_l = XEXP (force_const_mem (get_pool_mode (l), new_c), 0);
2413 *x = replace_rtx (l, l, new_l);
2415 return 0;
2418 /* If this is a JUMP_INSN, then we also need to fix the JUMP_LABEL
2419 field. This is not handled by for_each_rtx because it doesn't
2420 handle unprinted ('0') fields. */
2421 if (JUMP_P (l) && JUMP_LABEL (l) == old_label)
2422 JUMP_LABEL (l) = new_label;
2424 if ((GET_CODE (l) == LABEL_REF
2425 || GET_CODE (l) == INSN_LIST)
2426 && XEXP (l, 0) == old_label)
2428 XEXP (l, 0) = new_label;
2429 if (update_label_nuses)
2431 ++LABEL_NUSES (new_label);
2432 --LABEL_NUSES (old_label);
2434 return 0;
2437 return 0;
2440 /* When *BODY is equal to X or X is directly referenced by *BODY
2441 return nonzero, thus FOR_EACH_RTX stops traversing and returns nonzero
2442 too, otherwise FOR_EACH_RTX continues traversing *BODY. */
2444 static int
2445 rtx_referenced_p_1 (rtx *body, void *x)
2447 rtx y = (rtx) x;
2449 if (*body == NULL_RTX)
2450 return y == NULL_RTX;
2452 /* Return true if a label_ref *BODY refers to label Y. */
2453 if (GET_CODE (*body) == LABEL_REF && LABEL_P (y))
2454 return XEXP (*body, 0) == y;
2456 /* If *BODY is a reference to pool constant traverse the constant. */
2457 if (GET_CODE (*body) == SYMBOL_REF
2458 && CONSTANT_POOL_ADDRESS_P (*body))
2459 return rtx_referenced_p (y, get_pool_constant (*body));
2461 /* By default, compare the RTL expressions. */
2462 return rtx_equal_p (*body, y);
2465 /* Return true if X is referenced in BODY. */
2468 rtx_referenced_p (rtx x, rtx body)
2470 return for_each_rtx (&body, rtx_referenced_p_1, x);
2473 /* If INSN is a tablejump return true and store the label (before jump table) to
2474 *LABELP and the jump table to *TABLEP. LABELP and TABLEP may be NULL. */
2476 bool
2477 tablejump_p (rtx insn, rtx *labelp, rtx *tablep)
2479 rtx label, table;
2481 if (JUMP_P (insn)
2482 && (label = JUMP_LABEL (insn)) != NULL_RTX
2483 && (table = next_active_insn (label)) != NULL_RTX
2484 && JUMP_P (table)
2485 && (GET_CODE (PATTERN (table)) == ADDR_VEC
2486 || GET_CODE (PATTERN (table)) == ADDR_DIFF_VEC))
2488 if (labelp)
2489 *labelp = label;
2490 if (tablep)
2491 *tablep = table;
2492 return true;
2494 return false;
2497 /* A subroutine of computed_jump_p, return 1 if X contains a REG or MEM or
2498 constant that is not in the constant pool and not in the condition
2499 of an IF_THEN_ELSE. */
2501 static int
2502 computed_jump_p_1 (rtx x)
2504 enum rtx_code code = GET_CODE (x);
2505 int i, j;
2506 const char *fmt;
2508 switch (code)
2510 case LABEL_REF:
2511 case PC:
2512 return 0;
2514 case CONST:
2515 case CONST_INT:
2516 case CONST_DOUBLE:
2517 case CONST_VECTOR:
2518 case SYMBOL_REF:
2519 case REG:
2520 return 1;
2522 case MEM:
2523 return ! (GET_CODE (XEXP (x, 0)) == SYMBOL_REF
2524 && CONSTANT_POOL_ADDRESS_P (XEXP (x, 0)));
2526 case IF_THEN_ELSE:
2527 return (computed_jump_p_1 (XEXP (x, 1))
2528 || computed_jump_p_1 (XEXP (x, 2)));
2530 default:
2531 break;
2534 fmt = GET_RTX_FORMAT (code);
2535 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
2537 if (fmt[i] == 'e'
2538 && computed_jump_p_1 (XEXP (x, i)))
2539 return 1;
2541 else if (fmt[i] == 'E')
2542 for (j = 0; j < XVECLEN (x, i); j++)
2543 if (computed_jump_p_1 (XVECEXP (x, i, j)))
2544 return 1;
2547 return 0;
2550 /* Return nonzero if INSN is an indirect jump (aka computed jump).
2552 Tablejumps and casesi insns are not considered indirect jumps;
2553 we can recognize them by a (use (label_ref)). */
2556 computed_jump_p (rtx insn)
2558 int i;
2559 if (JUMP_P (insn))
2561 rtx pat = PATTERN (insn);
2563 if (find_reg_note (insn, REG_LABEL, NULL_RTX))
2564 return 0;
2565 else if (GET_CODE (pat) == PARALLEL)
2567 int len = XVECLEN (pat, 0);
2568 int has_use_labelref = 0;
2570 for (i = len - 1; i >= 0; i--)
2571 if (GET_CODE (XVECEXP (pat, 0, i)) == USE
2572 && (GET_CODE (XEXP (XVECEXP (pat, 0, i), 0))
2573 == LABEL_REF))
2574 has_use_labelref = 1;
2576 if (! has_use_labelref)
2577 for (i = len - 1; i >= 0; i--)
2578 if (GET_CODE (XVECEXP (pat, 0, i)) == SET
2579 && SET_DEST (XVECEXP (pat, 0, i)) == pc_rtx
2580 && computed_jump_p_1 (SET_SRC (XVECEXP (pat, 0, i))))
2581 return 1;
2583 else if (GET_CODE (pat) == SET
2584 && SET_DEST (pat) == pc_rtx
2585 && computed_jump_p_1 (SET_SRC (pat)))
2586 return 1;
2588 return 0;
2591 /* Optimized loop of for_each_rtx, trying to avoid useless recursive
2592 calls. Processes the subexpressions of EXP and passes them to F. */
2593 static int
2594 for_each_rtx_1 (rtx exp, int n, rtx_function f, void *data)
2596 int result, i, j;
2597 const char *format = GET_RTX_FORMAT (GET_CODE (exp));
2598 rtx *x;
2600 for (; format[n] != '\0'; n++)
2602 switch (format[n])
2604 case 'e':
2605 /* Call F on X. */
2606 x = &XEXP (exp, n);
2607 result = (*f) (x, data);
2608 if (result == -1)
2609 /* Do not traverse sub-expressions. */
2610 continue;
2611 else if (result != 0)
2612 /* Stop the traversal. */
2613 return result;
2615 if (*x == NULL_RTX)
2616 /* There are no sub-expressions. */
2617 continue;
2619 i = non_rtx_starting_operands[GET_CODE (*x)];
2620 if (i >= 0)
2622 result = for_each_rtx_1 (*x, i, f, data);
2623 if (result != 0)
2624 return result;
2626 break;
2628 case 'V':
2629 case 'E':
2630 if (XVEC (exp, n) == 0)
2631 continue;
2632 for (j = 0; j < XVECLEN (exp, n); ++j)
2634 /* Call F on X. */
2635 x = &XVECEXP (exp, n, j);
2636 result = (*f) (x, data);
2637 if (result == -1)
2638 /* Do not traverse sub-expressions. */
2639 continue;
2640 else if (result != 0)
2641 /* Stop the traversal. */
2642 return result;
2644 if (*x == NULL_RTX)
2645 /* There are no sub-expressions. */
2646 continue;
2648 i = non_rtx_starting_operands[GET_CODE (*x)];
2649 if (i >= 0)
2651 result = for_each_rtx_1 (*x, i, f, data);
2652 if (result != 0)
2653 return result;
2656 break;
2658 default:
2659 /* Nothing to do. */
2660 break;
2664 return 0;
2667 /* Traverse X via depth-first search, calling F for each
2668 sub-expression (including X itself). F is also passed the DATA.
2669 If F returns -1, do not traverse sub-expressions, but continue
2670 traversing the rest of the tree. If F ever returns any other
2671 nonzero value, stop the traversal, and return the value returned
2672 by F. Otherwise, return 0. This function does not traverse inside
2673 tree structure that contains RTX_EXPRs, or into sub-expressions
2674 whose format code is `0' since it is not known whether or not those
2675 codes are actually RTL.
2677 This routine is very general, and could (should?) be used to
2678 implement many of the other routines in this file. */
2681 for_each_rtx (rtx *x, rtx_function f, void *data)
2683 int result;
2684 int i;
2686 /* Call F on X. */
2687 result = (*f) (x, data);
2688 if (result == -1)
2689 /* Do not traverse sub-expressions. */
2690 return 0;
2691 else if (result != 0)
2692 /* Stop the traversal. */
2693 return result;
2695 if (*x == NULL_RTX)
2696 /* There are no sub-expressions. */
2697 return 0;
2699 i = non_rtx_starting_operands[GET_CODE (*x)];
2700 if (i < 0)
2701 return 0;
2703 return for_each_rtx_1 (*x, i, f, data);
2707 /* Searches X for any reference to REGNO, returning the rtx of the
2708 reference found if any. Otherwise, returns NULL_RTX. */
2711 regno_use_in (unsigned int regno, rtx x)
2713 const char *fmt;
2714 int i, j;
2715 rtx tem;
2717 if (REG_P (x) && REGNO (x) == regno)
2718 return x;
2720 fmt = GET_RTX_FORMAT (GET_CODE (x));
2721 for (i = GET_RTX_LENGTH (GET_CODE (x)) - 1; i >= 0; i--)
2723 if (fmt[i] == 'e')
2725 if ((tem = regno_use_in (regno, XEXP (x, i))))
2726 return tem;
2728 else if (fmt[i] == 'E')
2729 for (j = XVECLEN (x, i) - 1; j >= 0; j--)
2730 if ((tem = regno_use_in (regno , XVECEXP (x, i, j))))
2731 return tem;
2734 return NULL_RTX;
2737 /* Return a value indicating whether OP, an operand of a commutative
2738 operation, is preferred as the first or second operand. The higher
2739 the value, the stronger the preference for being the first operand.
2740 We use negative values to indicate a preference for the first operand
2741 and positive values for the second operand. */
2744 commutative_operand_precedence (rtx op)
2746 enum rtx_code code = GET_CODE (op);
2748 /* Constants always come the second operand. Prefer "nice" constants. */
2749 if (code == CONST_INT)
2750 return -7;
2751 if (code == CONST_DOUBLE)
2752 return -6;
2753 op = avoid_constant_pool_reference (op);
2754 code = GET_CODE (op);
2756 switch (GET_RTX_CLASS (code))
2758 case RTX_CONST_OBJ:
2759 if (code == CONST_INT)
2760 return -5;
2761 if (code == CONST_DOUBLE)
2762 return -4;
2763 return -3;
2765 case RTX_EXTRA:
2766 /* SUBREGs of objects should come second. */
2767 if (code == SUBREG && OBJECT_P (SUBREG_REG (op)))
2768 return -2;
2770 if (!CONSTANT_P (op))
2771 return 0;
2772 else
2773 /* As for RTX_CONST_OBJ. */
2774 return -3;
2776 case RTX_OBJ:
2777 /* Complex expressions should be the first, so decrease priority
2778 of objects. */
2779 return -1;
2781 case RTX_COMM_ARITH:
2782 /* Prefer operands that are themselves commutative to be first.
2783 This helps to make things linear. In particular,
2784 (and (and (reg) (reg)) (not (reg))) is canonical. */
2785 return 4;
2787 case RTX_BIN_ARITH:
2788 /* If only one operand is a binary expression, it will be the first
2789 operand. In particular, (plus (minus (reg) (reg)) (neg (reg)))
2790 is canonical, although it will usually be further simplified. */
2791 return 2;
2793 case RTX_UNARY:
2794 /* Then prefer NEG and NOT. */
2795 if (code == NEG || code == NOT)
2796 return 1;
2798 default:
2799 return 0;
2803 /* Return 1 iff it is necessary to swap operands of commutative operation
2804 in order to canonicalize expression. */
2807 swap_commutative_operands_p (rtx x, rtx y)
2809 return (commutative_operand_precedence (x)
2810 < commutative_operand_precedence (y));
2813 /* Return 1 if X is an autoincrement side effect and the register is
2814 not the stack pointer. */
2816 auto_inc_p (rtx x)
2818 switch (GET_CODE (x))
2820 case PRE_INC:
2821 case POST_INC:
2822 case PRE_DEC:
2823 case POST_DEC:
2824 case PRE_MODIFY:
2825 case POST_MODIFY:
2826 /* There are no REG_INC notes for SP. */
2827 if (XEXP (x, 0) != stack_pointer_rtx)
2828 return 1;
2829 default:
2830 break;
2832 return 0;
2835 /* Return nonzero if IN contains a piece of rtl that has the address LOC. */
2837 loc_mentioned_in_p (rtx *loc, rtx in)
2839 enum rtx_code code = GET_CODE (in);
2840 const char *fmt = GET_RTX_FORMAT (code);
2841 int i, j;
2843 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
2845 if (loc == &in->u.fld[i].rt_rtx)
2846 return 1;
2847 if (fmt[i] == 'e')
2849 if (loc_mentioned_in_p (loc, XEXP (in, i)))
2850 return 1;
2852 else if (fmt[i] == 'E')
2853 for (j = XVECLEN (in, i) - 1; j >= 0; j--)
2854 if (loc_mentioned_in_p (loc, XVECEXP (in, i, j)))
2855 return 1;
2857 return 0;
2860 /* Helper function for subreg_lsb. Given a subreg's OUTER_MODE, INNER_MODE,
2861 and SUBREG_BYTE, return the bit offset where the subreg begins
2862 (counting from the least significant bit of the operand). */
2864 unsigned int
2865 subreg_lsb_1 (enum machine_mode outer_mode,
2866 enum machine_mode inner_mode,
2867 unsigned int subreg_byte)
2869 unsigned int bitpos;
2870 unsigned int byte;
2871 unsigned int word;
2873 /* A paradoxical subreg begins at bit position 0. */
2874 if (GET_MODE_BITSIZE (outer_mode) > GET_MODE_BITSIZE (inner_mode))
2875 return 0;
2877 if (WORDS_BIG_ENDIAN != BYTES_BIG_ENDIAN)
2878 /* If the subreg crosses a word boundary ensure that
2879 it also begins and ends on a word boundary. */
2880 gcc_assert (!((subreg_byte % UNITS_PER_WORD
2881 + GET_MODE_SIZE (outer_mode)) > UNITS_PER_WORD
2882 && (subreg_byte % UNITS_PER_WORD
2883 || GET_MODE_SIZE (outer_mode) % UNITS_PER_WORD)));
2885 if (WORDS_BIG_ENDIAN)
2886 word = (GET_MODE_SIZE (inner_mode)
2887 - (subreg_byte + GET_MODE_SIZE (outer_mode))) / UNITS_PER_WORD;
2888 else
2889 word = subreg_byte / UNITS_PER_WORD;
2890 bitpos = word * BITS_PER_WORD;
2892 if (BYTES_BIG_ENDIAN)
2893 byte = (GET_MODE_SIZE (inner_mode)
2894 - (subreg_byte + GET_MODE_SIZE (outer_mode))) % UNITS_PER_WORD;
2895 else
2896 byte = subreg_byte % UNITS_PER_WORD;
2897 bitpos += byte * BITS_PER_UNIT;
2899 return bitpos;
2902 /* Given a subreg X, return the bit offset where the subreg begins
2903 (counting from the least significant bit of the reg). */
2905 unsigned int
2906 subreg_lsb (rtx x)
2908 return subreg_lsb_1 (GET_MODE (x), GET_MODE (SUBREG_REG (x)),
2909 SUBREG_BYTE (x));
2912 /* This function returns the regno offset of a subreg expression.
2913 xregno - A regno of an inner hard subreg_reg (or what will become one).
2914 xmode - The mode of xregno.
2915 offset - The byte offset.
2916 ymode - The mode of a top level SUBREG (or what may become one).
2917 RETURN - The regno offset which would be used. */
2918 unsigned int
2919 subreg_regno_offset (unsigned int xregno, enum machine_mode xmode,
2920 unsigned int offset, enum machine_mode ymode)
2922 int nregs_xmode, nregs_ymode;
2923 int mode_multiple, nregs_multiple;
2924 int y_offset;
2926 gcc_assert (xregno < FIRST_PSEUDO_REGISTER);
2928 /* Adjust nregs_xmode to allow for 'holes'. */
2929 if (HARD_REGNO_NREGS_HAS_PADDING (xregno, xmode))
2930 nregs_xmode = HARD_REGNO_NREGS_WITH_PADDING (xregno, xmode);
2931 else
2932 nregs_xmode = hard_regno_nregs[xregno][xmode];
2934 nregs_ymode = hard_regno_nregs[xregno][ymode];
2936 /* If this is a big endian paradoxical subreg, which uses more actual
2937 hard registers than the original register, we must return a negative
2938 offset so that we find the proper highpart of the register. */
2939 if (offset == 0
2940 && nregs_ymode > nregs_xmode
2941 && (GET_MODE_SIZE (ymode) > UNITS_PER_WORD
2942 ? WORDS_BIG_ENDIAN : BYTES_BIG_ENDIAN))
2943 return nregs_xmode - nregs_ymode;
2945 if (offset == 0 || nregs_xmode == nregs_ymode)
2946 return 0;
2948 /* Size of ymode must not be greater than the size of xmode. */
2949 mode_multiple = GET_MODE_SIZE (xmode) / GET_MODE_SIZE (ymode);
2950 gcc_assert (mode_multiple != 0);
2952 y_offset = offset / GET_MODE_SIZE (ymode);
2953 nregs_multiple = nregs_xmode / nregs_ymode;
2954 return (y_offset / (mode_multiple / nregs_multiple)) * nregs_ymode;
2957 /* This function returns true when the offset is representable via
2958 subreg_offset in the given regno.
2959 xregno - A regno of an inner hard subreg_reg (or what will become one).
2960 xmode - The mode of xregno.
2961 offset - The byte offset.
2962 ymode - The mode of a top level SUBREG (or what may become one).
2963 RETURN - Whether the offset is representable. */
2964 bool
2965 subreg_offset_representable_p (unsigned int xregno, enum machine_mode xmode,
2966 unsigned int offset, enum machine_mode ymode)
2968 int nregs_xmode, nregs_ymode;
2969 int mode_multiple, nregs_multiple;
2970 int y_offset;
2971 int regsize_xmode, regsize_ymode;
2973 gcc_assert (xregno < FIRST_PSEUDO_REGISTER);
2975 /* If there are holes in a non-scalar mode in registers, we expect
2976 that it is made up of its units concatenated together. */
2977 if (HARD_REGNO_NREGS_HAS_PADDING (xregno, xmode))
2979 enum machine_mode xmode_unit;
2981 nregs_xmode = HARD_REGNO_NREGS_WITH_PADDING (xregno, xmode);
2982 if (GET_MODE_INNER (xmode) == VOIDmode)
2983 xmode_unit = xmode;
2984 else
2985 xmode_unit = GET_MODE_INNER (xmode);
2986 gcc_assert (HARD_REGNO_NREGS_HAS_PADDING (xregno, xmode_unit));
2987 gcc_assert (nregs_xmode
2988 == (GET_MODE_NUNITS (xmode)
2989 * HARD_REGNO_NREGS_WITH_PADDING (xregno, xmode_unit)));
2990 gcc_assert (hard_regno_nregs[xregno][xmode]
2991 == (hard_regno_nregs[xregno][xmode_unit]
2992 * GET_MODE_NUNITS (xmode)));
2994 /* You can only ask for a SUBREG of a value with holes in the middle
2995 if you don't cross the holes. (Such a SUBREG should be done by
2996 picking a different register class, or doing it in memory if
2997 necessary.) An example of a value with holes is XCmode on 32-bit
2998 x86 with -m128bit-long-double; it's represented in 6 32-bit registers,
2999 3 for each part, but in memory it's two 128-bit parts.
3000 Padding is assumed to be at the end (not necessarily the 'high part')
3001 of each unit. */
3002 if ((offset / GET_MODE_SIZE (xmode_unit) + 1
3003 < GET_MODE_NUNITS (xmode))
3004 && (offset / GET_MODE_SIZE (xmode_unit)
3005 != ((offset + GET_MODE_SIZE (ymode) - 1)
3006 / GET_MODE_SIZE (xmode_unit))))
3007 return false;
3009 else
3010 nregs_xmode = hard_regno_nregs[xregno][xmode];
3012 nregs_ymode = hard_regno_nregs[xregno][ymode];
3014 /* Paradoxical subregs are otherwise valid. */
3015 if (offset == 0
3016 && nregs_ymode > nregs_xmode
3017 && (GET_MODE_SIZE (ymode) > UNITS_PER_WORD
3018 ? WORDS_BIG_ENDIAN : BYTES_BIG_ENDIAN))
3019 return true;
3021 /* If registers store different numbers of bits in the different
3022 modes, we cannot generally form this subreg. */
3023 regsize_xmode = GET_MODE_SIZE (xmode) / nregs_xmode;
3024 regsize_ymode = GET_MODE_SIZE (ymode) / nregs_ymode;
3025 if (regsize_xmode > regsize_ymode && nregs_ymode > 1)
3026 return false;
3027 if (regsize_ymode > regsize_xmode && nregs_xmode > 1)
3028 return false;
3030 /* Lowpart subregs are otherwise valid. */
3031 if (offset == subreg_lowpart_offset (ymode, xmode))
3032 return true;
3034 /* This should always pass, otherwise we don't know how to verify
3035 the constraint. These conditions may be relaxed but
3036 subreg_regno_offset would need to be redesigned. */
3037 gcc_assert ((GET_MODE_SIZE (xmode) % GET_MODE_SIZE (ymode)) == 0);
3038 gcc_assert ((nregs_xmode % nregs_ymode) == 0);
3040 /* The XMODE value can be seen as a vector of NREGS_XMODE
3041 values. The subreg must represent a lowpart of given field.
3042 Compute what field it is. */
3043 offset -= subreg_lowpart_offset (ymode,
3044 mode_for_size (GET_MODE_BITSIZE (xmode)
3045 / nregs_xmode,
3046 MODE_INT, 0));
3048 /* Size of ymode must not be greater than the size of xmode. */
3049 mode_multiple = GET_MODE_SIZE (xmode) / GET_MODE_SIZE (ymode);
3050 gcc_assert (mode_multiple != 0);
3052 y_offset = offset / GET_MODE_SIZE (ymode);
3053 nregs_multiple = nregs_xmode / nregs_ymode;
3055 gcc_assert ((offset % GET_MODE_SIZE (ymode)) == 0);
3056 gcc_assert ((mode_multiple % nregs_multiple) == 0);
3058 return (!(y_offset % (mode_multiple / nregs_multiple)));
3061 /* Return the final regno that a subreg expression refers to. */
3062 unsigned int
3063 subreg_regno (rtx x)
3065 unsigned int ret;
3066 rtx subreg = SUBREG_REG (x);
3067 int regno = REGNO (subreg);
3069 ret = regno + subreg_regno_offset (regno,
3070 GET_MODE (subreg),
3071 SUBREG_BYTE (x),
3072 GET_MODE (x));
3073 return ret;
3076 struct parms_set_data
3078 int nregs;
3079 HARD_REG_SET regs;
3082 /* Helper function for noticing stores to parameter registers. */
3083 static void
3084 parms_set (rtx x, rtx pat ATTRIBUTE_UNUSED, void *data)
3086 struct parms_set_data *d = data;
3087 if (REG_P (x) && REGNO (x) < FIRST_PSEUDO_REGISTER
3088 && TEST_HARD_REG_BIT (d->regs, REGNO (x)))
3090 CLEAR_HARD_REG_BIT (d->regs, REGNO (x));
3091 d->nregs--;
3095 /* Look backward for first parameter to be loaded.
3096 Note that loads of all parameters will not necessarily be
3097 found if CSE has eliminated some of them (e.g., an argument
3098 to the outer function is passed down as a parameter).
3099 Do not skip BOUNDARY. */
3101 find_first_parameter_load (rtx call_insn, rtx boundary)
3103 struct parms_set_data parm;
3104 rtx p, before, first_set;
3106 /* Since different machines initialize their parameter registers
3107 in different orders, assume nothing. Collect the set of all
3108 parameter registers. */
3109 CLEAR_HARD_REG_SET (parm.regs);
3110 parm.nregs = 0;
3111 for (p = CALL_INSN_FUNCTION_USAGE (call_insn); p; p = XEXP (p, 1))
3112 if (GET_CODE (XEXP (p, 0)) == USE
3113 && REG_P (XEXP (XEXP (p, 0), 0)))
3115 gcc_assert (REGNO (XEXP (XEXP (p, 0), 0)) < FIRST_PSEUDO_REGISTER);
3117 /* We only care about registers which can hold function
3118 arguments. */
3119 if (!FUNCTION_ARG_REGNO_P (REGNO (XEXP (XEXP (p, 0), 0))))
3120 continue;
3122 SET_HARD_REG_BIT (parm.regs, REGNO (XEXP (XEXP (p, 0), 0)));
3123 parm.nregs++;
3125 before = call_insn;
3126 first_set = call_insn;
3128 /* Search backward for the first set of a register in this set. */
3129 while (parm.nregs && before != boundary)
3131 before = PREV_INSN (before);
3133 /* It is possible that some loads got CSEed from one call to
3134 another. Stop in that case. */
3135 if (CALL_P (before))
3136 break;
3138 /* Our caller needs either ensure that we will find all sets
3139 (in case code has not been optimized yet), or take care
3140 for possible labels in a way by setting boundary to preceding
3141 CODE_LABEL. */
3142 if (LABEL_P (before))
3144 gcc_assert (before == boundary);
3145 break;
3148 if (INSN_P (before))
3150 int nregs_old = parm.nregs;
3151 note_stores (PATTERN (before), parms_set, &parm);
3152 /* If we found something that did not set a parameter reg,
3153 we're done. Do not keep going, as that might result
3154 in hoisting an insn before the setting of a pseudo
3155 that is used by the hoisted insn. */
3156 if (nregs_old != parm.nregs)
3157 first_set = before;
3158 else
3159 break;
3162 return first_set;
3165 /* Return true if we should avoid inserting code between INSN and preceding
3166 call instruction. */
3168 bool
3169 keep_with_call_p (rtx insn)
3171 rtx set;
3173 if (INSN_P (insn) && (set = single_set (insn)) != NULL)
3175 if (REG_P (SET_DEST (set))
3176 && REGNO (SET_DEST (set)) < FIRST_PSEUDO_REGISTER
3177 && fixed_regs[REGNO (SET_DEST (set))]
3178 && general_operand (SET_SRC (set), VOIDmode))
3179 return true;
3180 if (REG_P (SET_SRC (set))
3181 && FUNCTION_VALUE_REGNO_P (REGNO (SET_SRC (set)))
3182 && REG_P (SET_DEST (set))
3183 && REGNO (SET_DEST (set)) >= FIRST_PSEUDO_REGISTER)
3184 return true;
3185 /* There may be a stack pop just after the call and before the store
3186 of the return register. Search for the actual store when deciding
3187 if we can break or not. */
3188 if (SET_DEST (set) == stack_pointer_rtx)
3190 rtx i2 = next_nonnote_insn (insn);
3191 if (i2 && keep_with_call_p (i2))
3192 return true;
3195 return false;
3198 /* Return true if LABEL is a target of JUMP_INSN. This applies only
3199 to non-complex jumps. That is, direct unconditional, conditional,
3200 and tablejumps, but not computed jumps or returns. It also does
3201 not apply to the fallthru case of a conditional jump. */
3203 bool
3204 label_is_jump_target_p (rtx label, rtx jump_insn)
3206 rtx tmp = JUMP_LABEL (jump_insn);
3208 if (label == tmp)
3209 return true;
3211 if (tablejump_p (jump_insn, NULL, &tmp))
3213 rtvec vec = XVEC (PATTERN (tmp),
3214 GET_CODE (PATTERN (tmp)) == ADDR_DIFF_VEC);
3215 int i, veclen = GET_NUM_ELEM (vec);
3217 for (i = 0; i < veclen; ++i)
3218 if (XEXP (RTVEC_ELT (vec, i), 0) == label)
3219 return true;
3222 return false;
3226 /* Return an estimate of the cost of computing rtx X.
3227 One use is in cse, to decide which expression to keep in the hash table.
3228 Another is in rtl generation, to pick the cheapest way to multiply.
3229 Other uses like the latter are expected in the future. */
3232 rtx_cost (rtx x, enum rtx_code outer_code ATTRIBUTE_UNUSED)
3234 int i, j;
3235 enum rtx_code code;
3236 const char *fmt;
3237 int total;
3239 if (x == 0)
3240 return 0;
3242 /* Compute the default costs of certain things.
3243 Note that targetm.rtx_costs can override the defaults. */
3245 code = GET_CODE (x);
3246 switch (code)
3248 case MULT:
3249 total = COSTS_N_INSNS (5);
3250 break;
3251 case DIV:
3252 case UDIV:
3253 case MOD:
3254 case UMOD:
3255 total = COSTS_N_INSNS (7);
3256 break;
3257 case USE:
3258 /* Used in combine.c as a marker. */
3259 total = 0;
3260 break;
3261 default:
3262 total = COSTS_N_INSNS (1);
3265 switch (code)
3267 case REG:
3268 return 0;
3270 case SUBREG:
3271 total = 0;
3272 /* If we can't tie these modes, make this expensive. The larger
3273 the mode, the more expensive it is. */
3274 if (! MODES_TIEABLE_P (GET_MODE (x), GET_MODE (SUBREG_REG (x))))
3275 return COSTS_N_INSNS (2
3276 + GET_MODE_SIZE (GET_MODE (x)) / UNITS_PER_WORD);
3277 break;
3279 default:
3280 if (targetm.rtx_costs (x, code, outer_code, &total))
3281 return total;
3282 break;
3285 /* Sum the costs of the sub-rtx's, plus cost of this operation,
3286 which is already in total. */
3288 fmt = GET_RTX_FORMAT (code);
3289 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
3290 if (fmt[i] == 'e')
3291 total += rtx_cost (XEXP (x, i), code);
3292 else if (fmt[i] == 'E')
3293 for (j = 0; j < XVECLEN (x, i); j++)
3294 total += rtx_cost (XVECEXP (x, i, j), code);
3296 return total;
3299 /* Return cost of address expression X.
3300 Expect that X is properly formed address reference. */
3303 address_cost (rtx x, enum machine_mode mode)
3305 /* We may be asked for cost of various unusual addresses, such as operands
3306 of push instruction. It is not worthwhile to complicate writing
3307 of the target hook by such cases. */
3309 if (!memory_address_p (mode, x))
3310 return 1000;
3312 return targetm.address_cost (x);
3315 /* If the target doesn't override, compute the cost as with arithmetic. */
3318 default_address_cost (rtx x)
3320 return rtx_cost (x, MEM);
3324 unsigned HOST_WIDE_INT
3325 nonzero_bits (rtx x, enum machine_mode mode)
3327 return cached_nonzero_bits (x, mode, NULL_RTX, VOIDmode, 0);
3330 unsigned int
3331 num_sign_bit_copies (rtx x, enum machine_mode mode)
3333 return cached_num_sign_bit_copies (x, mode, NULL_RTX, VOIDmode, 0);
3336 /* The function cached_nonzero_bits is a wrapper around nonzero_bits1.
3337 It avoids exponential behavior in nonzero_bits1 when X has
3338 identical subexpressions on the first or the second level. */
3340 static unsigned HOST_WIDE_INT
3341 cached_nonzero_bits (rtx x, enum machine_mode mode, rtx known_x,
3342 enum machine_mode known_mode,
3343 unsigned HOST_WIDE_INT known_ret)
3345 if (x == known_x && mode == known_mode)
3346 return known_ret;
3348 /* Try to find identical subexpressions. If found call
3349 nonzero_bits1 on X with the subexpressions as KNOWN_X and the
3350 precomputed value for the subexpression as KNOWN_RET. */
3352 if (ARITHMETIC_P (x))
3354 rtx x0 = XEXP (x, 0);
3355 rtx x1 = XEXP (x, 1);
3357 /* Check the first level. */
3358 if (x0 == x1)
3359 return nonzero_bits1 (x, mode, x0, mode,
3360 cached_nonzero_bits (x0, mode, known_x,
3361 known_mode, known_ret));
3363 /* Check the second level. */
3364 if (ARITHMETIC_P (x0)
3365 && (x1 == XEXP (x0, 0) || x1 == XEXP (x0, 1)))
3366 return nonzero_bits1 (x, mode, x1, mode,
3367 cached_nonzero_bits (x1, mode, known_x,
3368 known_mode, known_ret));
3370 if (ARITHMETIC_P (x1)
3371 && (x0 == XEXP (x1, 0) || x0 == XEXP (x1, 1)))
3372 return nonzero_bits1 (x, mode, x0, mode,
3373 cached_nonzero_bits (x0, mode, known_x,
3374 known_mode, known_ret));
3377 return nonzero_bits1 (x, mode, known_x, known_mode, known_ret);
3380 /* We let num_sign_bit_copies recur into nonzero_bits as that is useful.
3381 We don't let nonzero_bits recur into num_sign_bit_copies, because that
3382 is less useful. We can't allow both, because that results in exponential
3383 run time recursion. There is a nullstone testcase that triggered
3384 this. This macro avoids accidental uses of num_sign_bit_copies. */
3385 #define cached_num_sign_bit_copies sorry_i_am_preventing_exponential_behavior
3387 /* Given an expression, X, compute which bits in X can be nonzero.
3388 We don't care about bits outside of those defined in MODE.
3390 For most X this is simply GET_MODE_MASK (GET_MODE (MODE)), but if X is
3391 an arithmetic operation, we can do better. */
3393 static unsigned HOST_WIDE_INT
3394 nonzero_bits1 (rtx x, enum machine_mode mode, rtx known_x,
3395 enum machine_mode known_mode,
3396 unsigned HOST_WIDE_INT known_ret)
3398 unsigned HOST_WIDE_INT nonzero = GET_MODE_MASK (mode);
3399 unsigned HOST_WIDE_INT inner_nz;
3400 enum rtx_code code;
3401 unsigned int mode_width = GET_MODE_BITSIZE (mode);
3403 /* For floating-point values, assume all bits are needed. */
3404 if (FLOAT_MODE_P (GET_MODE (x)) || FLOAT_MODE_P (mode))
3405 return nonzero;
3407 /* If X is wider than MODE, use its mode instead. */
3408 if (GET_MODE_BITSIZE (GET_MODE (x)) > mode_width)
3410 mode = GET_MODE (x);
3411 nonzero = GET_MODE_MASK (mode);
3412 mode_width = GET_MODE_BITSIZE (mode);
3415 if (mode_width > HOST_BITS_PER_WIDE_INT)
3416 /* Our only callers in this case look for single bit values. So
3417 just return the mode mask. Those tests will then be false. */
3418 return nonzero;
3420 #ifndef WORD_REGISTER_OPERATIONS
3421 /* If MODE is wider than X, but both are a single word for both the host
3422 and target machines, we can compute this from which bits of the
3423 object might be nonzero in its own mode, taking into account the fact
3424 that on many CISC machines, accessing an object in a wider mode
3425 causes the high-order bits to become undefined. So they are
3426 not known to be zero. */
3428 if (GET_MODE (x) != VOIDmode && GET_MODE (x) != mode
3429 && GET_MODE_BITSIZE (GET_MODE (x)) <= BITS_PER_WORD
3430 && GET_MODE_BITSIZE (GET_MODE (x)) <= HOST_BITS_PER_WIDE_INT
3431 && GET_MODE_BITSIZE (mode) > GET_MODE_BITSIZE (GET_MODE (x)))
3433 nonzero &= cached_nonzero_bits (x, GET_MODE (x),
3434 known_x, known_mode, known_ret);
3435 nonzero |= GET_MODE_MASK (mode) & ~GET_MODE_MASK (GET_MODE (x));
3436 return nonzero;
3438 #endif
3440 code = GET_CODE (x);
3441 switch (code)
3443 case REG:
3444 #if defined(POINTERS_EXTEND_UNSIGNED) && !defined(HAVE_ptr_extend)
3445 /* If pointers extend unsigned and this is a pointer in Pmode, say that
3446 all the bits above ptr_mode are known to be zero. */
3447 if (POINTERS_EXTEND_UNSIGNED && GET_MODE (x) == Pmode
3448 && REG_POINTER (x))
3449 nonzero &= GET_MODE_MASK (ptr_mode);
3450 #endif
3452 /* Include declared information about alignment of pointers. */
3453 /* ??? We don't properly preserve REG_POINTER changes across
3454 pointer-to-integer casts, so we can't trust it except for
3455 things that we know must be pointers. See execute/960116-1.c. */
3456 if ((x == stack_pointer_rtx
3457 || x == frame_pointer_rtx
3458 || x == arg_pointer_rtx)
3459 && REGNO_POINTER_ALIGN (REGNO (x)))
3461 unsigned HOST_WIDE_INT alignment
3462 = REGNO_POINTER_ALIGN (REGNO (x)) / BITS_PER_UNIT;
3464 #ifdef PUSH_ROUNDING
3465 /* If PUSH_ROUNDING is defined, it is possible for the
3466 stack to be momentarily aligned only to that amount,
3467 so we pick the least alignment. */
3468 if (x == stack_pointer_rtx && PUSH_ARGS)
3469 alignment = MIN ((unsigned HOST_WIDE_INT) PUSH_ROUNDING (1),
3470 alignment);
3471 #endif
3473 nonzero &= ~(alignment - 1);
3477 unsigned HOST_WIDE_INT nonzero_for_hook = nonzero;
3478 rtx new = rtl_hooks.reg_nonzero_bits (x, mode, known_x,
3479 known_mode, known_ret,
3480 &nonzero_for_hook);
3482 if (new)
3483 nonzero_for_hook &= cached_nonzero_bits (new, mode, known_x,
3484 known_mode, known_ret);
3486 return nonzero_for_hook;
3489 case CONST_INT:
3490 #ifdef SHORT_IMMEDIATES_SIGN_EXTEND
3491 /* If X is negative in MODE, sign-extend the value. */
3492 if (INTVAL (x) > 0 && mode_width < BITS_PER_WORD
3493 && 0 != (INTVAL (x) & ((HOST_WIDE_INT) 1 << (mode_width - 1))))
3494 return (INTVAL (x) | ((HOST_WIDE_INT) (-1) << mode_width));
3495 #endif
3497 return INTVAL (x);
3499 case MEM:
3500 #ifdef LOAD_EXTEND_OP
3501 /* In many, if not most, RISC machines, reading a byte from memory
3502 zeros the rest of the register. Noticing that fact saves a lot
3503 of extra zero-extends. */
3504 if (LOAD_EXTEND_OP (GET_MODE (x)) == ZERO_EXTEND)
3505 nonzero &= GET_MODE_MASK (GET_MODE (x));
3506 #endif
3507 break;
3509 case EQ: case NE:
3510 case UNEQ: case LTGT:
3511 case GT: case GTU: case UNGT:
3512 case LT: case LTU: case UNLT:
3513 case GE: case GEU: case UNGE:
3514 case LE: case LEU: case UNLE:
3515 case UNORDERED: case ORDERED:
3516 /* If this produces an integer result, we know which bits are set.
3517 Code here used to clear bits outside the mode of X, but that is
3518 now done above. */
3519 /* Mind that MODE is the mode the caller wants to look at this
3520 operation in, and not the actual operation mode. We can wind
3521 up with (subreg:DI (gt:V4HI x y)), and we don't have anything
3522 that describes the results of a vector compare. */
3523 if (GET_MODE_CLASS (GET_MODE (x)) == MODE_INT
3524 && mode_width <= HOST_BITS_PER_WIDE_INT)
3525 nonzero = STORE_FLAG_VALUE;
3526 break;
3528 case NEG:
3529 #if 0
3530 /* Disabled to avoid exponential mutual recursion between nonzero_bits
3531 and num_sign_bit_copies. */
3532 if (num_sign_bit_copies (XEXP (x, 0), GET_MODE (x))
3533 == GET_MODE_BITSIZE (GET_MODE (x)))
3534 nonzero = 1;
3535 #endif
3537 if (GET_MODE_SIZE (GET_MODE (x)) < mode_width)
3538 nonzero |= (GET_MODE_MASK (mode) & ~GET_MODE_MASK (GET_MODE (x)));
3539 break;
3541 case ABS:
3542 #if 0
3543 /* Disabled to avoid exponential mutual recursion between nonzero_bits
3544 and num_sign_bit_copies. */
3545 if (num_sign_bit_copies (XEXP (x, 0), GET_MODE (x))
3546 == GET_MODE_BITSIZE (GET_MODE (x)))
3547 nonzero = 1;
3548 #endif
3549 break;
3551 case TRUNCATE:
3552 nonzero &= (cached_nonzero_bits (XEXP (x, 0), mode,
3553 known_x, known_mode, known_ret)
3554 & GET_MODE_MASK (mode));
3555 break;
3557 case ZERO_EXTEND:
3558 nonzero &= cached_nonzero_bits (XEXP (x, 0), mode,
3559 known_x, known_mode, known_ret);
3560 if (GET_MODE (XEXP (x, 0)) != VOIDmode)
3561 nonzero &= GET_MODE_MASK (GET_MODE (XEXP (x, 0)));
3562 break;
3564 case SIGN_EXTEND:
3565 /* If the sign bit is known clear, this is the same as ZERO_EXTEND.
3566 Otherwise, show all the bits in the outer mode but not the inner
3567 may be nonzero. */
3568 inner_nz = cached_nonzero_bits (XEXP (x, 0), mode,
3569 known_x, known_mode, known_ret);
3570 if (GET_MODE (XEXP (x, 0)) != VOIDmode)
3572 inner_nz &= GET_MODE_MASK (GET_MODE (XEXP (x, 0)));
3573 if (inner_nz
3574 & (((HOST_WIDE_INT) 1
3575 << (GET_MODE_BITSIZE (GET_MODE (XEXP (x, 0))) - 1))))
3576 inner_nz |= (GET_MODE_MASK (mode)
3577 & ~GET_MODE_MASK (GET_MODE (XEXP (x, 0))));
3580 nonzero &= inner_nz;
3581 break;
3583 case AND:
3584 nonzero &= cached_nonzero_bits (XEXP (x, 0), mode,
3585 known_x, known_mode, known_ret)
3586 & cached_nonzero_bits (XEXP (x, 1), mode,
3587 known_x, known_mode, known_ret);
3588 break;
3590 case XOR: case IOR:
3591 case UMIN: case UMAX: case SMIN: case SMAX:
3593 unsigned HOST_WIDE_INT nonzero0 =
3594 cached_nonzero_bits (XEXP (x, 0), mode,
3595 known_x, known_mode, known_ret);
3597 /* Don't call nonzero_bits for the second time if it cannot change
3598 anything. */
3599 if ((nonzero & nonzero0) != nonzero)
3600 nonzero &= nonzero0
3601 | cached_nonzero_bits (XEXP (x, 1), mode,
3602 known_x, known_mode, known_ret);
3604 break;
3606 case PLUS: case MINUS:
3607 case MULT:
3608 case DIV: case UDIV:
3609 case MOD: case UMOD:
3610 /* We can apply the rules of arithmetic to compute the number of
3611 high- and low-order zero bits of these operations. We start by
3612 computing the width (position of the highest-order nonzero bit)
3613 and the number of low-order zero bits for each value. */
3615 unsigned HOST_WIDE_INT nz0 =
3616 cached_nonzero_bits (XEXP (x, 0), mode,
3617 known_x, known_mode, known_ret);
3618 unsigned HOST_WIDE_INT nz1 =
3619 cached_nonzero_bits (XEXP (x, 1), mode,
3620 known_x, known_mode, known_ret);
3621 int sign_index = GET_MODE_BITSIZE (GET_MODE (x)) - 1;
3622 int width0 = floor_log2 (nz0) + 1;
3623 int width1 = floor_log2 (nz1) + 1;
3624 int low0 = floor_log2 (nz0 & -nz0);
3625 int low1 = floor_log2 (nz1 & -nz1);
3626 HOST_WIDE_INT op0_maybe_minusp
3627 = (nz0 & ((HOST_WIDE_INT) 1 << sign_index));
3628 HOST_WIDE_INT op1_maybe_minusp
3629 = (nz1 & ((HOST_WIDE_INT) 1 << sign_index));
3630 unsigned int result_width = mode_width;
3631 int result_low = 0;
3633 switch (code)
3635 case PLUS:
3636 result_width = MAX (width0, width1) + 1;
3637 result_low = MIN (low0, low1);
3638 break;
3639 case MINUS:
3640 result_low = MIN (low0, low1);
3641 break;
3642 case MULT:
3643 result_width = width0 + width1;
3644 result_low = low0 + low1;
3645 break;
3646 case DIV:
3647 if (width1 == 0)
3648 break;
3649 if (! op0_maybe_minusp && ! op1_maybe_minusp)
3650 result_width = width0;
3651 break;
3652 case UDIV:
3653 if (width1 == 0)
3654 break;
3655 result_width = width0;
3656 break;
3657 case MOD:
3658 if (width1 == 0)
3659 break;
3660 if (! op0_maybe_minusp && ! op1_maybe_minusp)
3661 result_width = MIN (width0, width1);
3662 result_low = MIN (low0, low1);
3663 break;
3664 case UMOD:
3665 if (width1 == 0)
3666 break;
3667 result_width = MIN (width0, width1);
3668 result_low = MIN (low0, low1);
3669 break;
3670 default:
3671 gcc_unreachable ();
3674 if (result_width < mode_width)
3675 nonzero &= ((HOST_WIDE_INT) 1 << result_width) - 1;
3677 if (result_low > 0)
3678 nonzero &= ~(((HOST_WIDE_INT) 1 << result_low) - 1);
3680 #ifdef POINTERS_EXTEND_UNSIGNED
3681 /* If pointers extend unsigned and this is an addition or subtraction
3682 to a pointer in Pmode, all the bits above ptr_mode are known to be
3683 zero. */
3684 if (POINTERS_EXTEND_UNSIGNED > 0 && GET_MODE (x) == Pmode
3685 && (code == PLUS || code == MINUS)
3686 && REG_P (XEXP (x, 0)) && REG_POINTER (XEXP (x, 0)))
3687 nonzero &= GET_MODE_MASK (ptr_mode);
3688 #endif
3690 break;
3692 case ZERO_EXTRACT:
3693 if (GET_CODE (XEXP (x, 1)) == CONST_INT
3694 && INTVAL (XEXP (x, 1)) < HOST_BITS_PER_WIDE_INT)
3695 nonzero &= ((HOST_WIDE_INT) 1 << INTVAL (XEXP (x, 1))) - 1;
3696 break;
3698 case SUBREG:
3699 /* If this is a SUBREG formed for a promoted variable that has
3700 been zero-extended, we know that at least the high-order bits
3701 are zero, though others might be too. */
3703 if (SUBREG_PROMOTED_VAR_P (x) && SUBREG_PROMOTED_UNSIGNED_P (x) > 0)
3704 nonzero = GET_MODE_MASK (GET_MODE (x))
3705 & cached_nonzero_bits (SUBREG_REG (x), GET_MODE (x),
3706 known_x, known_mode, known_ret);
3708 /* If the inner mode is a single word for both the host and target
3709 machines, we can compute this from which bits of the inner
3710 object might be nonzero. */
3711 if (GET_MODE_BITSIZE (GET_MODE (SUBREG_REG (x))) <= BITS_PER_WORD
3712 && (GET_MODE_BITSIZE (GET_MODE (SUBREG_REG (x)))
3713 <= HOST_BITS_PER_WIDE_INT))
3715 nonzero &= cached_nonzero_bits (SUBREG_REG (x), mode,
3716 known_x, known_mode, known_ret);
3718 #if defined (WORD_REGISTER_OPERATIONS) && defined (LOAD_EXTEND_OP)
3719 /* If this is a typical RISC machine, we only have to worry
3720 about the way loads are extended. */
3721 if ((LOAD_EXTEND_OP (GET_MODE (SUBREG_REG (x))) == SIGN_EXTEND
3722 ? (((nonzero
3723 & (((unsigned HOST_WIDE_INT) 1
3724 << (GET_MODE_BITSIZE (GET_MODE (SUBREG_REG (x))) - 1))))
3725 != 0))
3726 : LOAD_EXTEND_OP (GET_MODE (SUBREG_REG (x))) != ZERO_EXTEND)
3727 || !MEM_P (SUBREG_REG (x)))
3728 #endif
3730 /* On many CISC machines, accessing an object in a wider mode
3731 causes the high-order bits to become undefined. So they are
3732 not known to be zero. */
3733 if (GET_MODE_SIZE (GET_MODE (x))
3734 > GET_MODE_SIZE (GET_MODE (SUBREG_REG (x))))
3735 nonzero |= (GET_MODE_MASK (GET_MODE (x))
3736 & ~GET_MODE_MASK (GET_MODE (SUBREG_REG (x))));
3739 break;
3741 case ASHIFTRT:
3742 case LSHIFTRT:
3743 case ASHIFT:
3744 case ROTATE:
3745 /* The nonzero bits are in two classes: any bits within MODE
3746 that aren't in GET_MODE (x) are always significant. The rest of the
3747 nonzero bits are those that are significant in the operand of
3748 the shift when shifted the appropriate number of bits. This
3749 shows that high-order bits are cleared by the right shift and
3750 low-order bits by left shifts. */
3751 if (GET_CODE (XEXP (x, 1)) == CONST_INT
3752 && INTVAL (XEXP (x, 1)) >= 0
3753 && INTVAL (XEXP (x, 1)) < HOST_BITS_PER_WIDE_INT)
3755 enum machine_mode inner_mode = GET_MODE (x);
3756 unsigned int width = GET_MODE_BITSIZE (inner_mode);
3757 int count = INTVAL (XEXP (x, 1));
3758 unsigned HOST_WIDE_INT mode_mask = GET_MODE_MASK (inner_mode);
3759 unsigned HOST_WIDE_INT op_nonzero =
3760 cached_nonzero_bits (XEXP (x, 0), mode,
3761 known_x, known_mode, known_ret);
3762 unsigned HOST_WIDE_INT inner = op_nonzero & mode_mask;
3763 unsigned HOST_WIDE_INT outer = 0;
3765 if (mode_width > width)
3766 outer = (op_nonzero & nonzero & ~mode_mask);
3768 if (code == LSHIFTRT)
3769 inner >>= count;
3770 else if (code == ASHIFTRT)
3772 inner >>= count;
3774 /* If the sign bit may have been nonzero before the shift, we
3775 need to mark all the places it could have been copied to
3776 by the shift as possibly nonzero. */
3777 if (inner & ((HOST_WIDE_INT) 1 << (width - 1 - count)))
3778 inner |= (((HOST_WIDE_INT) 1 << count) - 1) << (width - count);
3780 else if (code == ASHIFT)
3781 inner <<= count;
3782 else
3783 inner = ((inner << (count % width)
3784 | (inner >> (width - (count % width)))) & mode_mask);
3786 nonzero &= (outer | inner);
3788 break;
3790 case FFS:
3791 case POPCOUNT:
3792 /* This is at most the number of bits in the mode. */
3793 nonzero = ((HOST_WIDE_INT) 2 << (floor_log2 (mode_width))) - 1;
3794 break;
3796 case CLZ:
3797 /* If CLZ has a known value at zero, then the nonzero bits are
3798 that value, plus the number of bits in the mode minus one. */
3799 if (CLZ_DEFINED_VALUE_AT_ZERO (mode, nonzero))
3800 nonzero |= ((HOST_WIDE_INT) 1 << (floor_log2 (mode_width))) - 1;
3801 else
3802 nonzero = -1;
3803 break;
3805 case CTZ:
3806 /* If CTZ has a known value at zero, then the nonzero bits are
3807 that value, plus the number of bits in the mode minus one. */
3808 if (CTZ_DEFINED_VALUE_AT_ZERO (mode, nonzero))
3809 nonzero |= ((HOST_WIDE_INT) 1 << (floor_log2 (mode_width))) - 1;
3810 else
3811 nonzero = -1;
3812 break;
3814 case PARITY:
3815 nonzero = 1;
3816 break;
3818 case IF_THEN_ELSE:
3820 unsigned HOST_WIDE_INT nonzero_true =
3821 cached_nonzero_bits (XEXP (x, 1), mode,
3822 known_x, known_mode, known_ret);
3824 /* Don't call nonzero_bits for the second time if it cannot change
3825 anything. */
3826 if ((nonzero & nonzero_true) != nonzero)
3827 nonzero &= nonzero_true
3828 | cached_nonzero_bits (XEXP (x, 2), mode,
3829 known_x, known_mode, known_ret);
3831 break;
3833 default:
3834 break;
3837 return nonzero;
3840 /* See the macro definition above. */
3841 #undef cached_num_sign_bit_copies
3844 /* The function cached_num_sign_bit_copies is a wrapper around
3845 num_sign_bit_copies1. It avoids exponential behavior in
3846 num_sign_bit_copies1 when X has identical subexpressions on the
3847 first or the second level. */
3849 static unsigned int
3850 cached_num_sign_bit_copies (rtx x, enum machine_mode mode, rtx known_x,
3851 enum machine_mode known_mode,
3852 unsigned int known_ret)
3854 if (x == known_x && mode == known_mode)
3855 return known_ret;
3857 /* Try to find identical subexpressions. If found call
3858 num_sign_bit_copies1 on X with the subexpressions as KNOWN_X and
3859 the precomputed value for the subexpression as KNOWN_RET. */
3861 if (ARITHMETIC_P (x))
3863 rtx x0 = XEXP (x, 0);
3864 rtx x1 = XEXP (x, 1);
3866 /* Check the first level. */
3867 if (x0 == x1)
3868 return
3869 num_sign_bit_copies1 (x, mode, x0, mode,
3870 cached_num_sign_bit_copies (x0, mode, known_x,
3871 known_mode,
3872 known_ret));
3874 /* Check the second level. */
3875 if (ARITHMETIC_P (x0)
3876 && (x1 == XEXP (x0, 0) || x1 == XEXP (x0, 1)))
3877 return
3878 num_sign_bit_copies1 (x, mode, x1, mode,
3879 cached_num_sign_bit_copies (x1, mode, known_x,
3880 known_mode,
3881 known_ret));
3883 if (ARITHMETIC_P (x1)
3884 && (x0 == XEXP (x1, 0) || x0 == XEXP (x1, 1)))
3885 return
3886 num_sign_bit_copies1 (x, mode, x0, mode,
3887 cached_num_sign_bit_copies (x0, mode, known_x,
3888 known_mode,
3889 known_ret));
3892 return num_sign_bit_copies1 (x, mode, known_x, known_mode, known_ret);
3895 /* Return the number of bits at the high-order end of X that are known to
3896 be equal to the sign bit. X will be used in mode MODE; if MODE is
3897 VOIDmode, X will be used in its own mode. The returned value will always
3898 be between 1 and the number of bits in MODE. */
3900 static unsigned int
3901 num_sign_bit_copies1 (rtx x, enum machine_mode mode, rtx known_x,
3902 enum machine_mode known_mode,
3903 unsigned int known_ret)
3905 enum rtx_code code = GET_CODE (x);
3906 unsigned int bitwidth = GET_MODE_BITSIZE (mode);
3907 int num0, num1, result;
3908 unsigned HOST_WIDE_INT nonzero;
3910 /* If we weren't given a mode, use the mode of X. If the mode is still
3911 VOIDmode, we don't know anything. Likewise if one of the modes is
3912 floating-point. */
3914 if (mode == VOIDmode)
3915 mode = GET_MODE (x);
3917 if (mode == VOIDmode || FLOAT_MODE_P (mode) || FLOAT_MODE_P (GET_MODE (x)))
3918 return 1;
3920 /* For a smaller object, just ignore the high bits. */
3921 if (bitwidth < GET_MODE_BITSIZE (GET_MODE (x)))
3923 num0 = cached_num_sign_bit_copies (x, GET_MODE (x),
3924 known_x, known_mode, known_ret);
3925 return MAX (1,
3926 num0 - (int) (GET_MODE_BITSIZE (GET_MODE (x)) - bitwidth));
3929 if (GET_MODE (x) != VOIDmode && bitwidth > GET_MODE_BITSIZE (GET_MODE (x)))
3931 #ifndef WORD_REGISTER_OPERATIONS
3932 /* If this machine does not do all register operations on the entire
3933 register and MODE is wider than the mode of X, we can say nothing
3934 at all about the high-order bits. */
3935 return 1;
3936 #else
3937 /* Likewise on machines that do, if the mode of the object is smaller
3938 than a word and loads of that size don't sign extend, we can say
3939 nothing about the high order bits. */
3940 if (GET_MODE_BITSIZE (GET_MODE (x)) < BITS_PER_WORD
3941 #ifdef LOAD_EXTEND_OP
3942 && LOAD_EXTEND_OP (GET_MODE (x)) != SIGN_EXTEND
3943 #endif
3945 return 1;
3946 #endif
3949 switch (code)
3951 case REG:
3953 #if defined(POINTERS_EXTEND_UNSIGNED) && !defined(HAVE_ptr_extend)
3954 /* If pointers extend signed and this is a pointer in Pmode, say that
3955 all the bits above ptr_mode are known to be sign bit copies. */
3956 if (! POINTERS_EXTEND_UNSIGNED && GET_MODE (x) == Pmode && mode == Pmode
3957 && REG_POINTER (x))
3958 return GET_MODE_BITSIZE (Pmode) - GET_MODE_BITSIZE (ptr_mode) + 1;
3959 #endif
3962 unsigned int copies_for_hook = 1, copies = 1;
3963 rtx new = rtl_hooks.reg_num_sign_bit_copies (x, mode, known_x,
3964 known_mode, known_ret,
3965 &copies_for_hook);
3967 if (new)
3968 copies = cached_num_sign_bit_copies (new, mode, known_x,
3969 known_mode, known_ret);
3971 if (copies > 1 || copies_for_hook > 1)
3972 return MAX (copies, copies_for_hook);
3974 /* Else, use nonzero_bits to guess num_sign_bit_copies (see below). */
3976 break;
3978 case MEM:
3979 #ifdef LOAD_EXTEND_OP
3980 /* Some RISC machines sign-extend all loads of smaller than a word. */
3981 if (LOAD_EXTEND_OP (GET_MODE (x)) == SIGN_EXTEND)
3982 return MAX (1, ((int) bitwidth
3983 - (int) GET_MODE_BITSIZE (GET_MODE (x)) + 1));
3984 #endif
3985 break;
3987 case CONST_INT:
3988 /* If the constant is negative, take its 1's complement and remask.
3989 Then see how many zero bits we have. */
3990 nonzero = INTVAL (x) & GET_MODE_MASK (mode);
3991 if (bitwidth <= HOST_BITS_PER_WIDE_INT
3992 && (nonzero & ((HOST_WIDE_INT) 1 << (bitwidth - 1))) != 0)
3993 nonzero = (~nonzero) & GET_MODE_MASK (mode);
3995 return (nonzero == 0 ? bitwidth : bitwidth - floor_log2 (nonzero) - 1);
3997 case SUBREG:
3998 /* If this is a SUBREG for a promoted object that is sign-extended
3999 and we are looking at it in a wider mode, we know that at least the
4000 high-order bits are known to be sign bit copies. */
4002 if (SUBREG_PROMOTED_VAR_P (x) && ! SUBREG_PROMOTED_UNSIGNED_P (x))
4004 num0 = cached_num_sign_bit_copies (SUBREG_REG (x), mode,
4005 known_x, known_mode, known_ret);
4006 return MAX ((int) bitwidth
4007 - (int) GET_MODE_BITSIZE (GET_MODE (x)) + 1,
4008 num0);
4011 /* For a smaller object, just ignore the high bits. */
4012 if (bitwidth <= GET_MODE_BITSIZE (GET_MODE (SUBREG_REG (x))))
4014 num0 = cached_num_sign_bit_copies (SUBREG_REG (x), VOIDmode,
4015 known_x, known_mode, known_ret);
4016 return MAX (1, (num0
4017 - (int) (GET_MODE_BITSIZE (GET_MODE (SUBREG_REG (x)))
4018 - bitwidth)));
4021 #ifdef WORD_REGISTER_OPERATIONS
4022 #ifdef LOAD_EXTEND_OP
4023 /* For paradoxical SUBREGs on machines where all register operations
4024 affect the entire register, just look inside. Note that we are
4025 passing MODE to the recursive call, so the number of sign bit copies
4026 will remain relative to that mode, not the inner mode. */
4028 /* This works only if loads sign extend. Otherwise, if we get a
4029 reload for the inner part, it may be loaded from the stack, and
4030 then we lose all sign bit copies that existed before the store
4031 to the stack. */
4033 if ((GET_MODE_SIZE (GET_MODE (x))
4034 > GET_MODE_SIZE (GET_MODE (SUBREG_REG (x))))
4035 && LOAD_EXTEND_OP (GET_MODE (SUBREG_REG (x))) == SIGN_EXTEND
4036 && MEM_P (SUBREG_REG (x)))
4037 return cached_num_sign_bit_copies (SUBREG_REG (x), mode,
4038 known_x, known_mode, known_ret);
4039 #endif
4040 #endif
4041 break;
4043 case SIGN_EXTRACT:
4044 if (GET_CODE (XEXP (x, 1)) == CONST_INT)
4045 return MAX (1, (int) bitwidth - INTVAL (XEXP (x, 1)));
4046 break;
4048 case SIGN_EXTEND:
4049 return (bitwidth - GET_MODE_BITSIZE (GET_MODE (XEXP (x, 0)))
4050 + cached_num_sign_bit_copies (XEXP (x, 0), VOIDmode,
4051 known_x, known_mode, known_ret));
4053 case TRUNCATE:
4054 /* For a smaller object, just ignore the high bits. */
4055 num0 = cached_num_sign_bit_copies (XEXP (x, 0), VOIDmode,
4056 known_x, known_mode, known_ret);
4057 return MAX (1, (num0 - (int) (GET_MODE_BITSIZE (GET_MODE (XEXP (x, 0)))
4058 - bitwidth)));
4060 case NOT:
4061 return cached_num_sign_bit_copies (XEXP (x, 0), mode,
4062 known_x, known_mode, known_ret);
4064 case ROTATE: case ROTATERT:
4065 /* If we are rotating left by a number of bits less than the number
4066 of sign bit copies, we can just subtract that amount from the
4067 number. */
4068 if (GET_CODE (XEXP (x, 1)) == CONST_INT
4069 && INTVAL (XEXP (x, 1)) >= 0
4070 && INTVAL (XEXP (x, 1)) < (int) bitwidth)
4072 num0 = cached_num_sign_bit_copies (XEXP (x, 0), mode,
4073 known_x, known_mode, known_ret);
4074 return MAX (1, num0 - (code == ROTATE ? INTVAL (XEXP (x, 1))
4075 : (int) bitwidth - INTVAL (XEXP (x, 1))));
4077 break;
4079 case NEG:
4080 /* In general, this subtracts one sign bit copy. But if the value
4081 is known to be positive, the number of sign bit copies is the
4082 same as that of the input. Finally, if the input has just one bit
4083 that might be nonzero, all the bits are copies of the sign bit. */
4084 num0 = cached_num_sign_bit_copies (XEXP (x, 0), mode,
4085 known_x, known_mode, known_ret);
4086 if (bitwidth > HOST_BITS_PER_WIDE_INT)
4087 return num0 > 1 ? num0 - 1 : 1;
4089 nonzero = nonzero_bits (XEXP (x, 0), mode);
4090 if (nonzero == 1)
4091 return bitwidth;
4093 if (num0 > 1
4094 && (((HOST_WIDE_INT) 1 << (bitwidth - 1)) & nonzero))
4095 num0--;
4097 return num0;
4099 case IOR: case AND: case XOR:
4100 case SMIN: case SMAX: case UMIN: case UMAX:
4101 /* Logical operations will preserve the number of sign-bit copies.
4102 MIN and MAX operations always return one of the operands. */
4103 num0 = cached_num_sign_bit_copies (XEXP (x, 0), mode,
4104 known_x, known_mode, known_ret);
4105 num1 = cached_num_sign_bit_copies (XEXP (x, 1), mode,
4106 known_x, known_mode, known_ret);
4107 return MIN (num0, num1);
4109 case PLUS: case MINUS:
4110 /* For addition and subtraction, we can have a 1-bit carry. However,
4111 if we are subtracting 1 from a positive number, there will not
4112 be such a carry. Furthermore, if the positive number is known to
4113 be 0 or 1, we know the result is either -1 or 0. */
4115 if (code == PLUS && XEXP (x, 1) == constm1_rtx
4116 && bitwidth <= HOST_BITS_PER_WIDE_INT)
4118 nonzero = nonzero_bits (XEXP (x, 0), mode);
4119 if ((((HOST_WIDE_INT) 1 << (bitwidth - 1)) & nonzero) == 0)
4120 return (nonzero == 1 || nonzero == 0 ? bitwidth
4121 : bitwidth - floor_log2 (nonzero) - 1);
4124 num0 = cached_num_sign_bit_copies (XEXP (x, 0), mode,
4125 known_x, known_mode, known_ret);
4126 num1 = cached_num_sign_bit_copies (XEXP (x, 1), mode,
4127 known_x, known_mode, known_ret);
4128 result = MAX (1, MIN (num0, num1) - 1);
4130 #ifdef POINTERS_EXTEND_UNSIGNED
4131 /* If pointers extend signed and this is an addition or subtraction
4132 to a pointer in Pmode, all the bits above ptr_mode are known to be
4133 sign bit copies. */
4134 if (! POINTERS_EXTEND_UNSIGNED && GET_MODE (x) == Pmode
4135 && (code == PLUS || code == MINUS)
4136 && REG_P (XEXP (x, 0)) && REG_POINTER (XEXP (x, 0)))
4137 result = MAX ((int) (GET_MODE_BITSIZE (Pmode)
4138 - GET_MODE_BITSIZE (ptr_mode) + 1),
4139 result);
4140 #endif
4141 return result;
4143 case MULT:
4144 /* The number of bits of the product is the sum of the number of
4145 bits of both terms. However, unless one of the terms if known
4146 to be positive, we must allow for an additional bit since negating
4147 a negative number can remove one sign bit copy. */
4149 num0 = cached_num_sign_bit_copies (XEXP (x, 0), mode,
4150 known_x, known_mode, known_ret);
4151 num1 = cached_num_sign_bit_copies (XEXP (x, 1), mode,
4152 known_x, known_mode, known_ret);
4154 result = bitwidth - (bitwidth - num0) - (bitwidth - num1);
4155 if (result > 0
4156 && (bitwidth > HOST_BITS_PER_WIDE_INT
4157 || (((nonzero_bits (XEXP (x, 0), mode)
4158 & ((HOST_WIDE_INT) 1 << (bitwidth - 1))) != 0)
4159 && ((nonzero_bits (XEXP (x, 1), mode)
4160 & ((HOST_WIDE_INT) 1 << (bitwidth - 1))) != 0))))
4161 result--;
4163 return MAX (1, result);
4165 case UDIV:
4166 /* The result must be <= the first operand. If the first operand
4167 has the high bit set, we know nothing about the number of sign
4168 bit copies. */
4169 if (bitwidth > HOST_BITS_PER_WIDE_INT)
4170 return 1;
4171 else if ((nonzero_bits (XEXP (x, 0), mode)
4172 & ((HOST_WIDE_INT) 1 << (bitwidth - 1))) != 0)
4173 return 1;
4174 else
4175 return cached_num_sign_bit_copies (XEXP (x, 0), mode,
4176 known_x, known_mode, known_ret);
4178 case UMOD:
4179 /* The result must be <= the second operand. */
4180 return cached_num_sign_bit_copies (XEXP (x, 1), mode,
4181 known_x, known_mode, known_ret);
4183 case DIV:
4184 /* Similar to unsigned division, except that we have to worry about
4185 the case where the divisor is negative, in which case we have
4186 to add 1. */
4187 result = cached_num_sign_bit_copies (XEXP (x, 0), mode,
4188 known_x, known_mode, known_ret);
4189 if (result > 1
4190 && (bitwidth > HOST_BITS_PER_WIDE_INT
4191 || (nonzero_bits (XEXP (x, 1), mode)
4192 & ((HOST_WIDE_INT) 1 << (bitwidth - 1))) != 0))
4193 result--;
4195 return result;
4197 case MOD:
4198 result = cached_num_sign_bit_copies (XEXP (x, 1), mode,
4199 known_x, known_mode, known_ret);
4200 if (result > 1
4201 && (bitwidth > HOST_BITS_PER_WIDE_INT
4202 || (nonzero_bits (XEXP (x, 1), mode)
4203 & ((HOST_WIDE_INT) 1 << (bitwidth - 1))) != 0))
4204 result--;
4206 return result;
4208 case ASHIFTRT:
4209 /* Shifts by a constant add to the number of bits equal to the
4210 sign bit. */
4211 num0 = cached_num_sign_bit_copies (XEXP (x, 0), mode,
4212 known_x, known_mode, known_ret);
4213 if (GET_CODE (XEXP (x, 1)) == CONST_INT
4214 && INTVAL (XEXP (x, 1)) > 0)
4215 num0 = MIN ((int) bitwidth, num0 + INTVAL (XEXP (x, 1)));
4217 return num0;
4219 case ASHIFT:
4220 /* Left shifts destroy copies. */
4221 if (GET_CODE (XEXP (x, 1)) != CONST_INT
4222 || INTVAL (XEXP (x, 1)) < 0
4223 || INTVAL (XEXP (x, 1)) >= (int) bitwidth)
4224 return 1;
4226 num0 = cached_num_sign_bit_copies (XEXP (x, 0), mode,
4227 known_x, known_mode, known_ret);
4228 return MAX (1, num0 - INTVAL (XEXP (x, 1)));
4230 case IF_THEN_ELSE:
4231 num0 = cached_num_sign_bit_copies (XEXP (x, 1), mode,
4232 known_x, known_mode, known_ret);
4233 num1 = cached_num_sign_bit_copies (XEXP (x, 2), mode,
4234 known_x, known_mode, known_ret);
4235 return MIN (num0, num1);
4237 case EQ: case NE: case GE: case GT: case LE: case LT:
4238 case UNEQ: case LTGT: case UNGE: case UNGT: case UNLE: case UNLT:
4239 case GEU: case GTU: case LEU: case LTU:
4240 case UNORDERED: case ORDERED:
4241 /* If the constant is negative, take its 1's complement and remask.
4242 Then see how many zero bits we have. */
4243 nonzero = STORE_FLAG_VALUE;
4244 if (bitwidth <= HOST_BITS_PER_WIDE_INT
4245 && (nonzero & ((HOST_WIDE_INT) 1 << (bitwidth - 1))) != 0)
4246 nonzero = (~nonzero) & GET_MODE_MASK (mode);
4248 return (nonzero == 0 ? bitwidth : bitwidth - floor_log2 (nonzero) - 1);
4250 default:
4251 break;
4254 /* If we haven't been able to figure it out by one of the above rules,
4255 see if some of the high-order bits are known to be zero. If so,
4256 count those bits and return one less than that amount. If we can't
4257 safely compute the mask for this mode, always return BITWIDTH. */
4259 bitwidth = GET_MODE_BITSIZE (mode);
4260 if (bitwidth > HOST_BITS_PER_WIDE_INT)
4261 return 1;
4263 nonzero = nonzero_bits (x, mode);
4264 return nonzero & ((HOST_WIDE_INT) 1 << (bitwidth - 1))
4265 ? 1 : bitwidth - floor_log2 (nonzero) - 1;
4268 /* Calculate the rtx_cost of a single instruction. A return value of
4269 zero indicates an instruction pattern without a known cost. */
4272 insn_rtx_cost (rtx pat)
4274 int i, cost;
4275 rtx set;
4277 /* Extract the single set rtx from the instruction pattern.
4278 We can't use single_set since we only have the pattern. */
4279 if (GET_CODE (pat) == SET)
4280 set = pat;
4281 else if (GET_CODE (pat) == PARALLEL)
4283 set = NULL_RTX;
4284 for (i = 0; i < XVECLEN (pat, 0); i++)
4286 rtx x = XVECEXP (pat, 0, i);
4287 if (GET_CODE (x) == SET)
4289 if (set)
4290 return 0;
4291 set = x;
4294 if (!set)
4295 return 0;
4297 else
4298 return 0;
4300 cost = rtx_cost (SET_SRC (set), SET);
4301 return cost > 0 ? cost : COSTS_N_INSNS (1);
4304 /* Given an insn INSN and condition COND, return the condition in a
4305 canonical form to simplify testing by callers. Specifically:
4307 (1) The code will always be a comparison operation (EQ, NE, GT, etc.).
4308 (2) Both operands will be machine operands; (cc0) will have been replaced.
4309 (3) If an operand is a constant, it will be the second operand.
4310 (4) (LE x const) will be replaced with (LT x <const+1>) and similarly
4311 for GE, GEU, and LEU.
4313 If the condition cannot be understood, or is an inequality floating-point
4314 comparison which needs to be reversed, 0 will be returned.
4316 If REVERSE is nonzero, then reverse the condition prior to canonizing it.
4318 If EARLIEST is nonzero, it is a pointer to a place where the earliest
4319 insn used in locating the condition was found. If a replacement test
4320 of the condition is desired, it should be placed in front of that
4321 insn and we will be sure that the inputs are still valid.
4323 If WANT_REG is nonzero, we wish the condition to be relative to that
4324 register, if possible. Therefore, do not canonicalize the condition
4325 further. If ALLOW_CC_MODE is nonzero, allow the condition returned
4326 to be a compare to a CC mode register.
4328 If VALID_AT_INSN_P, the condition must be valid at both *EARLIEST
4329 and at INSN. */
4332 canonicalize_condition (rtx insn, rtx cond, int reverse, rtx *earliest,
4333 rtx want_reg, int allow_cc_mode, int valid_at_insn_p)
4335 enum rtx_code code;
4336 rtx prev = insn;
4337 rtx set;
4338 rtx tem;
4339 rtx op0, op1;
4340 int reverse_code = 0;
4341 enum machine_mode mode;
4342 basic_block bb = BLOCK_FOR_INSN (insn);
4344 code = GET_CODE (cond);
4345 mode = GET_MODE (cond);
4346 op0 = XEXP (cond, 0);
4347 op1 = XEXP (cond, 1);
4349 if (reverse)
4350 code = reversed_comparison_code (cond, insn);
4351 if (code == UNKNOWN)
4352 return 0;
4354 if (earliest)
4355 *earliest = insn;
4357 /* If we are comparing a register with zero, see if the register is set
4358 in the previous insn to a COMPARE or a comparison operation. Perform
4359 the same tests as a function of STORE_FLAG_VALUE as find_comparison_args
4360 in cse.c */
4362 while ((GET_RTX_CLASS (code) == RTX_COMPARE
4363 || GET_RTX_CLASS (code) == RTX_COMM_COMPARE)
4364 && op1 == CONST0_RTX (GET_MODE (op0))
4365 && op0 != want_reg)
4367 /* Set nonzero when we find something of interest. */
4368 rtx x = 0;
4370 #ifdef HAVE_cc0
4371 /* If comparison with cc0, import actual comparison from compare
4372 insn. */
4373 if (op0 == cc0_rtx)
4375 if ((prev = prev_nonnote_insn (prev)) == 0
4376 || !NONJUMP_INSN_P (prev)
4377 || (set = single_set (prev)) == 0
4378 || SET_DEST (set) != cc0_rtx)
4379 return 0;
4381 op0 = SET_SRC (set);
4382 op1 = CONST0_RTX (GET_MODE (op0));
4383 if (earliest)
4384 *earliest = prev;
4386 #endif
4388 /* If this is a COMPARE, pick up the two things being compared. */
4389 if (GET_CODE (op0) == COMPARE)
4391 op1 = XEXP (op0, 1);
4392 op0 = XEXP (op0, 0);
4393 continue;
4395 else if (!REG_P (op0))
4396 break;
4398 /* Go back to the previous insn. Stop if it is not an INSN. We also
4399 stop if it isn't a single set or if it has a REG_INC note because
4400 we don't want to bother dealing with it. */
4402 if ((prev = prev_nonnote_insn (prev)) == 0
4403 || !NONJUMP_INSN_P (prev)
4404 || FIND_REG_INC_NOTE (prev, NULL_RTX)
4405 /* In cfglayout mode, there do not have to be labels at the
4406 beginning of a block, or jumps at the end, so the previous
4407 conditions would not stop us when we reach bb boundary. */
4408 || BLOCK_FOR_INSN (prev) != bb)
4409 break;
4411 set = set_of (op0, prev);
4413 if (set
4414 && (GET_CODE (set) != SET
4415 || !rtx_equal_p (SET_DEST (set), op0)))
4416 break;
4418 /* If this is setting OP0, get what it sets it to if it looks
4419 relevant. */
4420 if (set)
4422 enum machine_mode inner_mode = GET_MODE (SET_DEST (set));
4423 #ifdef FLOAT_STORE_FLAG_VALUE
4424 REAL_VALUE_TYPE fsfv;
4425 #endif
4427 /* ??? We may not combine comparisons done in a CCmode with
4428 comparisons not done in a CCmode. This is to aid targets
4429 like Alpha that have an IEEE compliant EQ instruction, and
4430 a non-IEEE compliant BEQ instruction. The use of CCmode is
4431 actually artificial, simply to prevent the combination, but
4432 should not affect other platforms.
4434 However, we must allow VOIDmode comparisons to match either
4435 CCmode or non-CCmode comparison, because some ports have
4436 modeless comparisons inside branch patterns.
4438 ??? This mode check should perhaps look more like the mode check
4439 in simplify_comparison in combine. */
4441 if ((GET_CODE (SET_SRC (set)) == COMPARE
4442 || (((code == NE
4443 || (code == LT
4444 && GET_MODE_CLASS (inner_mode) == MODE_INT
4445 && (GET_MODE_BITSIZE (inner_mode)
4446 <= HOST_BITS_PER_WIDE_INT)
4447 && (STORE_FLAG_VALUE
4448 & ((HOST_WIDE_INT) 1
4449 << (GET_MODE_BITSIZE (inner_mode) - 1))))
4450 #ifdef FLOAT_STORE_FLAG_VALUE
4451 || (code == LT
4452 && SCALAR_FLOAT_MODE_P (inner_mode)
4453 && (fsfv = FLOAT_STORE_FLAG_VALUE (inner_mode),
4454 REAL_VALUE_NEGATIVE (fsfv)))
4455 #endif
4457 && COMPARISON_P (SET_SRC (set))))
4458 && (((GET_MODE_CLASS (mode) == MODE_CC)
4459 == (GET_MODE_CLASS (inner_mode) == MODE_CC))
4460 || mode == VOIDmode || inner_mode == VOIDmode))
4461 x = SET_SRC (set);
4462 else if (((code == EQ
4463 || (code == GE
4464 && (GET_MODE_BITSIZE (inner_mode)
4465 <= HOST_BITS_PER_WIDE_INT)
4466 && GET_MODE_CLASS (inner_mode) == MODE_INT
4467 && (STORE_FLAG_VALUE
4468 & ((HOST_WIDE_INT) 1
4469 << (GET_MODE_BITSIZE (inner_mode) - 1))))
4470 #ifdef FLOAT_STORE_FLAG_VALUE
4471 || (code == GE
4472 && SCALAR_FLOAT_MODE_P (inner_mode)
4473 && (fsfv = FLOAT_STORE_FLAG_VALUE (inner_mode),
4474 REAL_VALUE_NEGATIVE (fsfv)))
4475 #endif
4477 && COMPARISON_P (SET_SRC (set))
4478 && (((GET_MODE_CLASS (mode) == MODE_CC)
4479 == (GET_MODE_CLASS (inner_mode) == MODE_CC))
4480 || mode == VOIDmode || inner_mode == VOIDmode))
4483 reverse_code = 1;
4484 x = SET_SRC (set);
4486 else
4487 break;
4490 else if (reg_set_p (op0, prev))
4491 /* If this sets OP0, but not directly, we have to give up. */
4492 break;
4494 if (x)
4496 /* If the caller is expecting the condition to be valid at INSN,
4497 make sure X doesn't change before INSN. */
4498 if (valid_at_insn_p)
4499 if (modified_in_p (x, prev) || modified_between_p (x, prev, insn))
4500 break;
4501 if (COMPARISON_P (x))
4502 code = GET_CODE (x);
4503 if (reverse_code)
4505 code = reversed_comparison_code (x, prev);
4506 if (code == UNKNOWN)
4507 return 0;
4508 reverse_code = 0;
4511 op0 = XEXP (x, 0), op1 = XEXP (x, 1);
4512 if (earliest)
4513 *earliest = prev;
4517 /* If constant is first, put it last. */
4518 if (CONSTANT_P (op0))
4519 code = swap_condition (code), tem = op0, op0 = op1, op1 = tem;
4521 /* If OP0 is the result of a comparison, we weren't able to find what
4522 was really being compared, so fail. */
4523 if (!allow_cc_mode
4524 && GET_MODE_CLASS (GET_MODE (op0)) == MODE_CC)
4525 return 0;
4527 /* Canonicalize any ordered comparison with integers involving equality
4528 if we can do computations in the relevant mode and we do not
4529 overflow. */
4531 if (GET_MODE_CLASS (GET_MODE (op0)) != MODE_CC
4532 && GET_CODE (op1) == CONST_INT
4533 && GET_MODE (op0) != VOIDmode
4534 && GET_MODE_BITSIZE (GET_MODE (op0)) <= HOST_BITS_PER_WIDE_INT)
4536 HOST_WIDE_INT const_val = INTVAL (op1);
4537 unsigned HOST_WIDE_INT uconst_val = const_val;
4538 unsigned HOST_WIDE_INT max_val
4539 = (unsigned HOST_WIDE_INT) GET_MODE_MASK (GET_MODE (op0));
4541 switch (code)
4543 case LE:
4544 if ((unsigned HOST_WIDE_INT) const_val != max_val >> 1)
4545 code = LT, op1 = gen_int_mode (const_val + 1, GET_MODE (op0));
4546 break;
4548 /* When cross-compiling, const_val might be sign-extended from
4549 BITS_PER_WORD to HOST_BITS_PER_WIDE_INT */
4550 case GE:
4551 if ((HOST_WIDE_INT) (const_val & max_val)
4552 != (((HOST_WIDE_INT) 1
4553 << (GET_MODE_BITSIZE (GET_MODE (op0)) - 1))))
4554 code = GT, op1 = gen_int_mode (const_val - 1, GET_MODE (op0));
4555 break;
4557 case LEU:
4558 if (uconst_val < max_val)
4559 code = LTU, op1 = gen_int_mode (uconst_val + 1, GET_MODE (op0));
4560 break;
4562 case GEU:
4563 if (uconst_val != 0)
4564 code = GTU, op1 = gen_int_mode (uconst_val - 1, GET_MODE (op0));
4565 break;
4567 default:
4568 break;
4572 /* Never return CC0; return zero instead. */
4573 if (CC0_P (op0))
4574 return 0;
4576 return gen_rtx_fmt_ee (code, VOIDmode, op0, op1);
4579 /* Given a jump insn JUMP, return the condition that will cause it to branch
4580 to its JUMP_LABEL. If the condition cannot be understood, or is an
4581 inequality floating-point comparison which needs to be reversed, 0 will
4582 be returned.
4584 If EARLIEST is nonzero, it is a pointer to a place where the earliest
4585 insn used in locating the condition was found. If a replacement test
4586 of the condition is desired, it should be placed in front of that
4587 insn and we will be sure that the inputs are still valid. If EARLIEST
4588 is null, the returned condition will be valid at INSN.
4590 If ALLOW_CC_MODE is nonzero, allow the condition returned to be a
4591 compare CC mode register.
4593 VALID_AT_INSN_P is the same as for canonicalize_condition. */
4596 get_condition (rtx jump, rtx *earliest, int allow_cc_mode, int valid_at_insn_p)
4598 rtx cond;
4599 int reverse;
4600 rtx set;
4602 /* If this is not a standard conditional jump, we can't parse it. */
4603 if (!JUMP_P (jump)
4604 || ! any_condjump_p (jump))
4605 return 0;
4606 set = pc_set (jump);
4608 cond = XEXP (SET_SRC (set), 0);
4610 /* If this branches to JUMP_LABEL when the condition is false, reverse
4611 the condition. */
4612 reverse
4613 = GET_CODE (XEXP (SET_SRC (set), 2)) == LABEL_REF
4614 && XEXP (XEXP (SET_SRC (set), 2), 0) == JUMP_LABEL (jump);
4616 return canonicalize_condition (jump, cond, reverse, earliest, NULL_RTX,
4617 allow_cc_mode, valid_at_insn_p);
4620 /* Initialize the table NUM_SIGN_BIT_COPIES_IN_REP based on
4621 TARGET_MODE_REP_EXTENDED.
4623 Note that we assume that the property of
4624 TARGET_MODE_REP_EXTENDED(B, C) is sticky to the integral modes
4625 narrower than mode B. I.e., if A is a mode narrower than B then in
4626 order to be able to operate on it in mode B, mode A needs to
4627 satisfy the requirements set by the representation of mode B. */
4629 static void
4630 init_num_sign_bit_copies_in_rep (void)
4632 enum machine_mode mode, in_mode;
4634 for (in_mode = GET_CLASS_NARROWEST_MODE (MODE_INT); in_mode != VOIDmode;
4635 in_mode = GET_MODE_WIDER_MODE (mode))
4636 for (mode = GET_CLASS_NARROWEST_MODE (MODE_INT); mode != in_mode;
4637 mode = GET_MODE_WIDER_MODE (mode))
4639 enum machine_mode i;
4641 /* Currently, it is assumed that TARGET_MODE_REP_EXTENDED
4642 extends to the next widest mode. */
4643 gcc_assert (targetm.mode_rep_extended (mode, in_mode) == UNKNOWN
4644 || GET_MODE_WIDER_MODE (mode) == in_mode);
4646 /* We are in in_mode. Count how many bits outside of mode
4647 have to be copies of the sign-bit. */
4648 for (i = mode; i != in_mode; i = GET_MODE_WIDER_MODE (i))
4650 enum machine_mode wider = GET_MODE_WIDER_MODE (i);
4652 if (targetm.mode_rep_extended (i, wider) == SIGN_EXTEND
4653 /* We can only check sign-bit copies starting from the
4654 top-bit. In order to be able to check the bits we
4655 have already seen we pretend that subsequent bits
4656 have to be sign-bit copies too. */
4657 || num_sign_bit_copies_in_rep [in_mode][mode])
4658 num_sign_bit_copies_in_rep [in_mode][mode]
4659 += GET_MODE_BITSIZE (wider) - GET_MODE_BITSIZE (i);
4664 /* Suppose that truncation from the machine mode of X to MODE is not a
4665 no-op. See if there is anything special about X so that we can
4666 assume it already contains a truncated value of MODE. */
4668 bool
4669 truncated_to_mode (enum machine_mode mode, rtx x)
4671 /* This register has already been used in MODE without explicit
4672 truncation. */
4673 if (REG_P (x) && rtl_hooks.reg_truncated_to_mode (mode, x))
4674 return true;
4676 /* See if we already satisfy the requirements of MODE. If yes we
4677 can just switch to MODE. */
4678 if (num_sign_bit_copies_in_rep[GET_MODE (x)][mode]
4679 && (num_sign_bit_copies (x, GET_MODE (x))
4680 >= num_sign_bit_copies_in_rep[GET_MODE (x)][mode] + 1))
4681 return true;
4683 return false;
4686 /* Initialize non_rtx_starting_operands, which is used to speed up
4687 for_each_rtx. */
4688 void
4689 init_rtlanal (void)
4691 int i;
4692 for (i = 0; i < NUM_RTX_CODE; i++)
4694 const char *format = GET_RTX_FORMAT (i);
4695 const char *first = strpbrk (format, "eEV");
4696 non_rtx_starting_operands[i] = first ? first - format : -1;
4699 init_num_sign_bit_copies_in_rep ();
4702 /* Check whether this is a constant pool constant. */
4703 bool
4704 constant_pool_constant_p (rtx x)
4706 x = avoid_constant_pool_reference (x);
4707 return GET_CODE (x) == CONST_DOUBLE;