1 /* Fold a constant sub-tree into a single node for C-compiler
2 Copyright (C) 1987-2016 Free Software Foundation, Inc.
4 This file is part of GCC.
6 GCC is free software; you can redistribute it and/or modify it under
7 the terms of the GNU General Public License as published by the Free
8 Software Foundation; either version 3, or (at your option) any later
11 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
12 WARRANTY; without even the implied warranty of MERCHANTABILITY or
13 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
16 You should have received a copy of the GNU General Public License
17 along with GCC; see the file COPYING3. If not see
18 <http://www.gnu.org/licenses/>. */
20 /*@@ This file should be rewritten to use an arbitrary precision
21 @@ representation for "struct tree_int_cst" and "struct tree_real_cst".
22 @@ Perhaps the routines could also be used for bc/dc, and made a lib.
23 @@ The routines that translate from the ap rep should
24 @@ warn if precision et. al. is lost.
25 @@ This would also make life easier when this technology is used
26 @@ for cross-compilers. */
28 /* The entry points in this file are fold, size_int_wide and size_binop.
30 fold takes a tree as argument and returns a simplified tree.
32 size_binop takes a tree code for an arithmetic operation
33 and two operands that are trees, and produces a tree for the
34 result, assuming the type comes from `sizetype'.
36 size_int takes an integer value, and creates a tree constant
37 with type from `sizetype'.
39 Note: Since the folders get called on non-gimple code as well as
40 gimple code, we need to handle GIMPLE tuples as well as their
41 corresponding tree equivalents. */
45 #include "coretypes.h"
53 #include "tree-ssa-operands.h"
54 #include "optabs-query.h"
56 #include "diagnostic-core.h"
59 #include "fold-const.h"
60 #include "fold-const-call.h"
61 #include "stor-layout.h"
63 #include "tree-iterator.h"
66 #include "langhooks.h"
71 #include "generic-match.h"
72 #include "gimple-fold.h"
74 #include "tree-into-ssa.h"
76 #include "case-cfn-macros.h"
77 #include "stringpool.h"
78 #include "tree-ssanames.h"
80 #ifndef LOAD_EXTEND_OP
81 #define LOAD_EXTEND_OP(M) UNKNOWN
84 /* Nonzero if we are folding constants inside an initializer; zero
86 int folding_initializer
= 0;
88 /* The following constants represent a bit based encoding of GCC's
89 comparison operators. This encoding simplifies transformations
90 on relational comparison operators, such as AND and OR. */
91 enum comparison_code
{
110 static bool negate_expr_p (tree
);
111 static tree
negate_expr (tree
);
112 static tree
split_tree (location_t
, tree
, tree
, enum tree_code
,
113 tree
*, tree
*, tree
*, int);
114 static tree
associate_trees (location_t
, tree
, tree
, enum tree_code
, tree
);
115 static enum comparison_code
comparison_to_compcode (enum tree_code
);
116 static enum tree_code
compcode_to_comparison (enum comparison_code
);
117 static int operand_equal_for_comparison_p (tree
, tree
, tree
);
118 static int twoval_comparison_p (tree
, tree
*, tree
*, int *);
119 static tree
eval_subst (location_t
, tree
, tree
, tree
, tree
, tree
);
120 static tree
make_bit_field_ref (location_t
, tree
, tree
,
121 HOST_WIDE_INT
, HOST_WIDE_INT
, int, int);
122 static tree
optimize_bit_field_compare (location_t
, enum tree_code
,
124 static tree
decode_field_reference (location_t
, tree
, HOST_WIDE_INT
*,
126 machine_mode
*, int *, int *, int *,
128 static int simple_operand_p (const_tree
);
129 static bool simple_operand_p_2 (tree
);
130 static tree
range_binop (enum tree_code
, tree
, tree
, int, tree
, int);
131 static tree
range_predecessor (tree
);
132 static tree
range_successor (tree
);
133 static tree
fold_range_test (location_t
, enum tree_code
, tree
, tree
, tree
);
134 static tree
fold_cond_expr_with_comparison (location_t
, tree
, tree
, tree
, tree
);
135 static tree
unextend (tree
, int, int, tree
);
136 static tree
optimize_minmax_comparison (location_t
, enum tree_code
,
138 static tree
extract_muldiv (tree
, tree
, enum tree_code
, tree
, bool *);
139 static tree
extract_muldiv_1 (tree
, tree
, enum tree_code
, tree
, bool *);
140 static tree
fold_binary_op_with_conditional_arg (location_t
,
141 enum tree_code
, tree
,
144 static tree
fold_div_compare (location_t
, enum tree_code
, tree
, tree
, tree
);
145 static bool reorder_operands_p (const_tree
, const_tree
);
146 static tree
fold_negate_const (tree
, tree
);
147 static tree
fold_not_const (const_tree
, tree
);
148 static tree
fold_relational_const (enum tree_code
, tree
, tree
, tree
);
149 static tree
fold_convert_const (enum tree_code
, tree
, tree
);
150 static tree
fold_view_convert_expr (tree
, tree
);
151 static bool vec_cst_ctor_to_array (tree
, tree
*);
154 /* Return EXPR_LOCATION of T if it is not UNKNOWN_LOCATION.
155 Otherwise, return LOC. */
158 expr_location_or (tree t
, location_t loc
)
160 location_t tloc
= EXPR_LOCATION (t
);
161 return tloc
== UNKNOWN_LOCATION
? loc
: tloc
;
164 /* Similar to protected_set_expr_location, but never modify x in place,
165 if location can and needs to be set, unshare it. */
168 protected_set_expr_location_unshare (tree x
, location_t loc
)
170 if (CAN_HAVE_LOCATION_P (x
)
171 && EXPR_LOCATION (x
) != loc
172 && !(TREE_CODE (x
) == SAVE_EXPR
173 || TREE_CODE (x
) == TARGET_EXPR
174 || TREE_CODE (x
) == BIND_EXPR
))
177 SET_EXPR_LOCATION (x
, loc
);
182 /* If ARG2 divides ARG1 with zero remainder, carries out the exact
183 division and returns the quotient. Otherwise returns
187 div_if_zero_remainder (const_tree arg1
, const_tree arg2
)
191 if (wi::multiple_of_p (wi::to_widest (arg1
), wi::to_widest (arg2
),
193 return wide_int_to_tree (TREE_TYPE (arg1
), quo
);
198 /* This is nonzero if we should defer warnings about undefined
199 overflow. This facility exists because these warnings are a
200 special case. The code to estimate loop iterations does not want
201 to issue any warnings, since it works with expressions which do not
202 occur in user code. Various bits of cleanup code call fold(), but
203 only use the result if it has certain characteristics (e.g., is a
204 constant); that code only wants to issue a warning if the result is
207 static int fold_deferring_overflow_warnings
;
209 /* If a warning about undefined overflow is deferred, this is the
210 warning. Note that this may cause us to turn two warnings into
211 one, but that is fine since it is sufficient to only give one
212 warning per expression. */
214 static const char* fold_deferred_overflow_warning
;
216 /* If a warning about undefined overflow is deferred, this is the
217 level at which the warning should be emitted. */
219 static enum warn_strict_overflow_code fold_deferred_overflow_code
;
221 /* Start deferring overflow warnings. We could use a stack here to
222 permit nested calls, but at present it is not necessary. */
225 fold_defer_overflow_warnings (void)
227 ++fold_deferring_overflow_warnings
;
230 /* Stop deferring overflow warnings. If there is a pending warning,
231 and ISSUE is true, then issue the warning if appropriate. STMT is
232 the statement with which the warning should be associated (used for
233 location information); STMT may be NULL. CODE is the level of the
234 warning--a warn_strict_overflow_code value. This function will use
235 the smaller of CODE and the deferred code when deciding whether to
236 issue the warning. CODE may be zero to mean to always use the
240 fold_undefer_overflow_warnings (bool issue
, const gimple
*stmt
, int code
)
245 gcc_assert (fold_deferring_overflow_warnings
> 0);
246 --fold_deferring_overflow_warnings
;
247 if (fold_deferring_overflow_warnings
> 0)
249 if (fold_deferred_overflow_warning
!= NULL
251 && code
< (int) fold_deferred_overflow_code
)
252 fold_deferred_overflow_code
= (enum warn_strict_overflow_code
) code
;
256 warnmsg
= fold_deferred_overflow_warning
;
257 fold_deferred_overflow_warning
= NULL
;
259 if (!issue
|| warnmsg
== NULL
)
262 if (gimple_no_warning_p (stmt
))
265 /* Use the smallest code level when deciding to issue the
267 if (code
== 0 || code
> (int) fold_deferred_overflow_code
)
268 code
= fold_deferred_overflow_code
;
270 if (!issue_strict_overflow_warning (code
))
274 locus
= input_location
;
276 locus
= gimple_location (stmt
);
277 warning_at (locus
, OPT_Wstrict_overflow
, "%s", warnmsg
);
280 /* Stop deferring overflow warnings, ignoring any deferred
284 fold_undefer_and_ignore_overflow_warnings (void)
286 fold_undefer_overflow_warnings (false, NULL
, 0);
289 /* Whether we are deferring overflow warnings. */
292 fold_deferring_overflow_warnings_p (void)
294 return fold_deferring_overflow_warnings
> 0;
297 /* This is called when we fold something based on the fact that signed
298 overflow is undefined. */
301 fold_overflow_warning (const char* gmsgid
, enum warn_strict_overflow_code wc
)
303 if (fold_deferring_overflow_warnings
> 0)
305 if (fold_deferred_overflow_warning
== NULL
306 || wc
< fold_deferred_overflow_code
)
308 fold_deferred_overflow_warning
= gmsgid
;
309 fold_deferred_overflow_code
= wc
;
312 else if (issue_strict_overflow_warning (wc
))
313 warning (OPT_Wstrict_overflow
, gmsgid
);
316 /* Return true if the built-in mathematical function specified by CODE
317 is odd, i.e. -f(x) == f(-x). */
320 negate_mathfn_p (combined_fn fn
)
353 return !flag_rounding_math
;
361 /* Check whether we may negate an integer constant T without causing
365 may_negate_without_overflow_p (const_tree t
)
369 gcc_assert (TREE_CODE (t
) == INTEGER_CST
);
371 type
= TREE_TYPE (t
);
372 if (TYPE_UNSIGNED (type
))
375 return !wi::only_sign_bit_p (t
);
378 /* Determine whether an expression T can be cheaply negated using
379 the function negate_expr without introducing undefined overflow. */
382 negate_expr_p (tree t
)
389 type
= TREE_TYPE (t
);
392 switch (TREE_CODE (t
))
395 if (INTEGRAL_TYPE_P (type
) && TYPE_OVERFLOW_WRAPS (type
))
398 /* Check that -CST will not overflow type. */
399 return may_negate_without_overflow_p (t
);
401 return (INTEGRAL_TYPE_P (type
)
402 && TYPE_OVERFLOW_WRAPS (type
));
408 return !TYPE_OVERFLOW_SANITIZED (type
);
411 /* We want to canonicalize to positive real constants. Pretend
412 that only negative ones can be easily negated. */
413 return REAL_VALUE_NEGATIVE (TREE_REAL_CST (t
));
416 return negate_expr_p (TREE_REALPART (t
))
417 && negate_expr_p (TREE_IMAGPART (t
));
421 if (FLOAT_TYPE_P (TREE_TYPE (type
)) || TYPE_OVERFLOW_WRAPS (type
))
424 int count
= TYPE_VECTOR_SUBPARTS (type
), i
;
426 for (i
= 0; i
< count
; i
++)
427 if (!negate_expr_p (VECTOR_CST_ELT (t
, i
)))
434 return negate_expr_p (TREE_OPERAND (t
, 0))
435 && negate_expr_p (TREE_OPERAND (t
, 1));
438 return negate_expr_p (TREE_OPERAND (t
, 0));
441 if (HONOR_SIGN_DEPENDENT_ROUNDING (element_mode (type
))
442 || HONOR_SIGNED_ZEROS (element_mode (type
))
443 || (INTEGRAL_TYPE_P (type
)
444 && ! TYPE_OVERFLOW_WRAPS (type
)))
446 /* -(A + B) -> (-B) - A. */
447 if (negate_expr_p (TREE_OPERAND (t
, 1))
448 && reorder_operands_p (TREE_OPERAND (t
, 0),
449 TREE_OPERAND (t
, 1)))
451 /* -(A + B) -> (-A) - B. */
452 return negate_expr_p (TREE_OPERAND (t
, 0));
455 /* We can't turn -(A-B) into B-A when we honor signed zeros. */
456 return !HONOR_SIGN_DEPENDENT_ROUNDING (element_mode (type
))
457 && !HONOR_SIGNED_ZEROS (element_mode (type
))
458 && (! INTEGRAL_TYPE_P (type
)
459 || TYPE_OVERFLOW_WRAPS (type
))
460 && reorder_operands_p (TREE_OPERAND (t
, 0),
461 TREE_OPERAND (t
, 1));
464 if (TYPE_UNSIGNED (type
))
466 /* INT_MIN/n * n doesn't overflow while negating one operand it does
467 if n is a power of two. */
468 if (INTEGRAL_TYPE_P (TREE_TYPE (t
))
469 && ! TYPE_OVERFLOW_WRAPS (TREE_TYPE (t
))
470 && ! ((TREE_CODE (TREE_OPERAND (t
, 0)) == INTEGER_CST
471 && ! integer_pow2p (TREE_OPERAND (t
, 0)))
472 || (TREE_CODE (TREE_OPERAND (t
, 1)) == INTEGER_CST
473 && ! integer_pow2p (TREE_OPERAND (t
, 1)))))
479 if (! HONOR_SIGN_DEPENDENT_ROUNDING (element_mode (TREE_TYPE (t
))))
480 return negate_expr_p (TREE_OPERAND (t
, 1))
481 || negate_expr_p (TREE_OPERAND (t
, 0));
487 if (TYPE_UNSIGNED (type
))
489 if (negate_expr_p (TREE_OPERAND (t
, 0)))
491 /* In general we can't negate B in A / B, because if A is INT_MIN and
492 B is 1, we may turn this into INT_MIN / -1 which is undefined
493 and actually traps on some architectures. */
494 if (! INTEGRAL_TYPE_P (TREE_TYPE (t
))
495 || TYPE_OVERFLOW_WRAPS (TREE_TYPE (t
))
496 || (TREE_CODE (TREE_OPERAND (t
, 1)) == INTEGER_CST
497 && ! integer_onep (TREE_OPERAND (t
, 1))))
498 return negate_expr_p (TREE_OPERAND (t
, 1));
502 /* Negate -((double)float) as (double)(-float). */
503 if (TREE_CODE (type
) == REAL_TYPE
)
505 tree tem
= strip_float_extensions (t
);
507 return negate_expr_p (tem
);
512 /* Negate -f(x) as f(-x). */
513 if (negate_mathfn_p (get_call_combined_fn (t
)))
514 return negate_expr_p (CALL_EXPR_ARG (t
, 0));
518 /* Optimize -((int)x >> 31) into (unsigned)x >> 31 for int. */
519 if (TREE_CODE (TREE_OPERAND (t
, 1)) == INTEGER_CST
)
521 tree op1
= TREE_OPERAND (t
, 1);
522 if (wi::eq_p (op1
, TYPE_PRECISION (type
) - 1))
533 /* Given T, an expression, return a folded tree for -T or NULL_TREE, if no
534 simplification is possible.
535 If negate_expr_p would return true for T, NULL_TREE will never be
539 fold_negate_expr (location_t loc
, tree t
)
541 tree type
= TREE_TYPE (t
);
544 switch (TREE_CODE (t
))
546 /* Convert - (~A) to A + 1. */
548 if (INTEGRAL_TYPE_P (type
))
549 return fold_build2_loc (loc
, PLUS_EXPR
, type
, TREE_OPERAND (t
, 0),
550 build_one_cst (type
));
554 tem
= fold_negate_const (t
, type
);
555 if (TREE_OVERFLOW (tem
) == TREE_OVERFLOW (t
)
556 || (ANY_INTEGRAL_TYPE_P (type
)
557 && !TYPE_OVERFLOW_TRAPS (type
)
558 && TYPE_OVERFLOW_WRAPS (type
))
559 || (flag_sanitize
& SANITIZE_SI_OVERFLOW
) == 0)
564 tem
= fold_negate_const (t
, type
);
568 tem
= fold_negate_const (t
, type
);
573 tree rpart
= fold_negate_expr (loc
, TREE_REALPART (t
));
574 tree ipart
= fold_negate_expr (loc
, TREE_IMAGPART (t
));
576 return build_complex (type
, rpart
, ipart
);
582 int count
= TYPE_VECTOR_SUBPARTS (type
), i
;
583 tree
*elts
= XALLOCAVEC (tree
, count
);
585 for (i
= 0; i
< count
; i
++)
587 elts
[i
] = fold_negate_expr (loc
, VECTOR_CST_ELT (t
, i
));
588 if (elts
[i
] == NULL_TREE
)
592 return build_vector (type
, elts
);
596 if (negate_expr_p (t
))
597 return fold_build2_loc (loc
, COMPLEX_EXPR
, type
,
598 fold_negate_expr (loc
, TREE_OPERAND (t
, 0)),
599 fold_negate_expr (loc
, TREE_OPERAND (t
, 1)));
603 if (negate_expr_p (t
))
604 return fold_build1_loc (loc
, CONJ_EXPR
, type
,
605 fold_negate_expr (loc
, TREE_OPERAND (t
, 0)));
609 if (!TYPE_OVERFLOW_SANITIZED (type
))
610 return TREE_OPERAND (t
, 0);
614 if (!HONOR_SIGN_DEPENDENT_ROUNDING (element_mode (type
))
615 && !HONOR_SIGNED_ZEROS (element_mode (type
)))
617 /* -(A + B) -> (-B) - A. */
618 if (negate_expr_p (TREE_OPERAND (t
, 1))
619 && reorder_operands_p (TREE_OPERAND (t
, 0),
620 TREE_OPERAND (t
, 1)))
622 tem
= negate_expr (TREE_OPERAND (t
, 1));
623 return fold_build2_loc (loc
, MINUS_EXPR
, type
,
624 tem
, TREE_OPERAND (t
, 0));
627 /* -(A + B) -> (-A) - B. */
628 if (negate_expr_p (TREE_OPERAND (t
, 0)))
630 tem
= negate_expr (TREE_OPERAND (t
, 0));
631 return fold_build2_loc (loc
, MINUS_EXPR
, type
,
632 tem
, TREE_OPERAND (t
, 1));
638 /* - (A - B) -> B - A */
639 if (!HONOR_SIGN_DEPENDENT_ROUNDING (element_mode (type
))
640 && !HONOR_SIGNED_ZEROS (element_mode (type
))
641 && reorder_operands_p (TREE_OPERAND (t
, 0), TREE_OPERAND (t
, 1)))
642 return fold_build2_loc (loc
, MINUS_EXPR
, type
,
643 TREE_OPERAND (t
, 1), TREE_OPERAND (t
, 0));
647 if (TYPE_UNSIGNED (type
))
653 if (! HONOR_SIGN_DEPENDENT_ROUNDING (element_mode (type
)))
655 tem
= TREE_OPERAND (t
, 1);
656 if (negate_expr_p (tem
))
657 return fold_build2_loc (loc
, TREE_CODE (t
), type
,
658 TREE_OPERAND (t
, 0), negate_expr (tem
));
659 tem
= TREE_OPERAND (t
, 0);
660 if (negate_expr_p (tem
))
661 return fold_build2_loc (loc
, TREE_CODE (t
), type
,
662 negate_expr (tem
), TREE_OPERAND (t
, 1));
669 if (TYPE_UNSIGNED (type
))
671 if (negate_expr_p (TREE_OPERAND (t
, 0)))
672 return fold_build2_loc (loc
, TREE_CODE (t
), type
,
673 negate_expr (TREE_OPERAND (t
, 0)),
674 TREE_OPERAND (t
, 1));
675 /* In general we can't negate B in A / B, because if A is INT_MIN and
676 B is 1, we may turn this into INT_MIN / -1 which is undefined
677 and actually traps on some architectures. */
678 if ((! INTEGRAL_TYPE_P (TREE_TYPE (t
))
679 || TYPE_OVERFLOW_WRAPS (TREE_TYPE (t
))
680 || (TREE_CODE (TREE_OPERAND (t
, 1)) == INTEGER_CST
681 && ! integer_onep (TREE_OPERAND (t
, 1))))
682 && negate_expr_p (TREE_OPERAND (t
, 1)))
683 return fold_build2_loc (loc
, TREE_CODE (t
), type
,
685 negate_expr (TREE_OPERAND (t
, 1)));
689 /* Convert -((double)float) into (double)(-float). */
690 if (TREE_CODE (type
) == REAL_TYPE
)
692 tem
= strip_float_extensions (t
);
693 if (tem
!= t
&& negate_expr_p (tem
))
694 return fold_convert_loc (loc
, type
, negate_expr (tem
));
699 /* Negate -f(x) as f(-x). */
700 if (negate_mathfn_p (get_call_combined_fn (t
))
701 && negate_expr_p (CALL_EXPR_ARG (t
, 0)))
705 fndecl
= get_callee_fndecl (t
);
706 arg
= negate_expr (CALL_EXPR_ARG (t
, 0));
707 return build_call_expr_loc (loc
, fndecl
, 1, arg
);
712 /* Optimize -((int)x >> 31) into (unsigned)x >> 31 for int. */
713 if (TREE_CODE (TREE_OPERAND (t
, 1)) == INTEGER_CST
)
715 tree op1
= TREE_OPERAND (t
, 1);
716 if (wi::eq_p (op1
, TYPE_PRECISION (type
) - 1))
718 tree ntype
= TYPE_UNSIGNED (type
)
719 ? signed_type_for (type
)
720 : unsigned_type_for (type
);
721 tree temp
= fold_convert_loc (loc
, ntype
, TREE_OPERAND (t
, 0));
722 temp
= fold_build2_loc (loc
, RSHIFT_EXPR
, ntype
, temp
, op1
);
723 return fold_convert_loc (loc
, type
, temp
);
735 /* Like fold_negate_expr, but return a NEGATE_EXPR tree, if T can not be
736 negated in a simpler way. Also allow for T to be NULL_TREE, in which case
748 loc
= EXPR_LOCATION (t
);
749 type
= TREE_TYPE (t
);
752 tem
= fold_negate_expr (loc
, t
);
754 tem
= build1_loc (loc
, NEGATE_EXPR
, TREE_TYPE (t
), t
);
755 return fold_convert_loc (loc
, type
, tem
);
758 /* Split a tree IN into a constant, literal and variable parts that could be
759 combined with CODE to make IN. "constant" means an expression with
760 TREE_CONSTANT but that isn't an actual constant. CODE must be a
761 commutative arithmetic operation. Store the constant part into *CONP,
762 the literal in *LITP and return the variable part. If a part isn't
763 present, set it to null. If the tree does not decompose in this way,
764 return the entire tree as the variable part and the other parts as null.
766 If CODE is PLUS_EXPR we also split trees that use MINUS_EXPR. In that
767 case, we negate an operand that was subtracted. Except if it is a
768 literal for which we use *MINUS_LITP instead.
770 If NEGATE_P is true, we are negating all of IN, again except a literal
771 for which we use *MINUS_LITP instead. If a variable part is of pointer
772 type, it is negated after converting to TYPE. This prevents us from
773 generating illegal MINUS pointer expression. LOC is the location of
774 the converted variable part.
776 If IN is itself a literal or constant, return it as appropriate.
778 Note that we do not guarantee that any of the three values will be the
779 same type as IN, but they will have the same signedness and mode. */
782 split_tree (location_t loc
, tree in
, tree type
, enum tree_code code
,
783 tree
*conp
, tree
*litp
, tree
*minus_litp
, int negate_p
)
791 /* Strip any conversions that don't change the machine mode or signedness. */
792 STRIP_SIGN_NOPS (in
);
794 if (TREE_CODE (in
) == INTEGER_CST
|| TREE_CODE (in
) == REAL_CST
795 || TREE_CODE (in
) == FIXED_CST
)
797 else if (TREE_CODE (in
) == code
798 || ((! FLOAT_TYPE_P (TREE_TYPE (in
)) || flag_associative_math
)
799 && ! SAT_FIXED_POINT_TYPE_P (TREE_TYPE (in
))
800 /* We can associate addition and subtraction together (even
801 though the C standard doesn't say so) for integers because
802 the value is not affected. For reals, the value might be
803 affected, so we can't. */
804 && ((code
== PLUS_EXPR
&& TREE_CODE (in
) == MINUS_EXPR
)
805 || (code
== MINUS_EXPR
&& TREE_CODE (in
) == PLUS_EXPR
))))
807 tree op0
= TREE_OPERAND (in
, 0);
808 tree op1
= TREE_OPERAND (in
, 1);
809 int neg1_p
= TREE_CODE (in
) == MINUS_EXPR
;
810 int neg_litp_p
= 0, neg_conp_p
= 0, neg_var_p
= 0;
812 /* First see if either of the operands is a literal, then a constant. */
813 if (TREE_CODE (op0
) == INTEGER_CST
|| TREE_CODE (op0
) == REAL_CST
814 || TREE_CODE (op0
) == FIXED_CST
)
815 *litp
= op0
, op0
= 0;
816 else if (TREE_CODE (op1
) == INTEGER_CST
|| TREE_CODE (op1
) == REAL_CST
817 || TREE_CODE (op1
) == FIXED_CST
)
818 *litp
= op1
, neg_litp_p
= neg1_p
, op1
= 0;
820 if (op0
!= 0 && TREE_CONSTANT (op0
))
821 *conp
= op0
, op0
= 0;
822 else if (op1
!= 0 && TREE_CONSTANT (op1
))
823 *conp
= op1
, neg_conp_p
= neg1_p
, op1
= 0;
825 /* If we haven't dealt with either operand, this is not a case we can
826 decompose. Otherwise, VAR is either of the ones remaining, if any. */
827 if (op0
!= 0 && op1
!= 0)
832 var
= op1
, neg_var_p
= neg1_p
;
834 /* Now do any needed negations. */
836 *minus_litp
= *litp
, *litp
= 0;
838 *conp
= negate_expr (*conp
);
841 /* Convert to TYPE before negating a pointer type expr. */
842 if (var
&& POINTER_TYPE_P (TREE_TYPE (var
)))
843 var
= fold_convert_loc (loc
, type
, var
);
844 var
= negate_expr (var
);
847 else if (TREE_CODE (in
) == BIT_NOT_EXPR
848 && code
== PLUS_EXPR
)
850 /* -X - 1 is folded to ~X, undo that here. */
851 *minus_litp
= build_one_cst (TREE_TYPE (in
));
852 var
= negate_expr (TREE_OPERAND (in
, 0));
854 else if (TREE_CONSTANT (in
))
862 *minus_litp
= *litp
, *litp
= 0;
863 else if (*minus_litp
)
864 *litp
= *minus_litp
, *minus_litp
= 0;
865 *conp
= negate_expr (*conp
);
866 /* Convert to TYPE before negating a pointer type expr. */
867 if (var
&& POINTER_TYPE_P (TREE_TYPE (var
)))
868 var
= fold_convert_loc (loc
, type
, var
);
869 var
= negate_expr (var
);
875 /* Re-associate trees split by the above function. T1 and T2 are
876 either expressions to associate or null. Return the new
877 expression, if any. LOC is the location of the new expression. If
878 we build an operation, do it in TYPE and with CODE. */
881 associate_trees (location_t loc
, tree t1
, tree t2
, enum tree_code code
, tree type
)
888 /* If either input is CODE, a PLUS_EXPR, or a MINUS_EXPR, don't
889 try to fold this since we will have infinite recursion. But do
890 deal with any NEGATE_EXPRs. */
891 if (TREE_CODE (t1
) == code
|| TREE_CODE (t2
) == code
892 || TREE_CODE (t1
) == MINUS_EXPR
|| TREE_CODE (t2
) == MINUS_EXPR
)
894 if (code
== PLUS_EXPR
)
896 if (TREE_CODE (t1
) == NEGATE_EXPR
)
897 return build2_loc (loc
, MINUS_EXPR
, type
,
898 fold_convert_loc (loc
, type
, t2
),
899 fold_convert_loc (loc
, type
,
900 TREE_OPERAND (t1
, 0)));
901 else if (TREE_CODE (t2
) == NEGATE_EXPR
)
902 return build2_loc (loc
, MINUS_EXPR
, type
,
903 fold_convert_loc (loc
, type
, t1
),
904 fold_convert_loc (loc
, type
,
905 TREE_OPERAND (t2
, 0)));
906 else if (integer_zerop (t2
))
907 return fold_convert_loc (loc
, type
, t1
);
909 else if (code
== MINUS_EXPR
)
911 if (integer_zerop (t2
))
912 return fold_convert_loc (loc
, type
, t1
);
915 return build2_loc (loc
, code
, type
, fold_convert_loc (loc
, type
, t1
),
916 fold_convert_loc (loc
, type
, t2
));
919 return fold_build2_loc (loc
, code
, type
, fold_convert_loc (loc
, type
, t1
),
920 fold_convert_loc (loc
, type
, t2
));
923 /* Check whether TYPE1 and TYPE2 are equivalent integer types, suitable
924 for use in int_const_binop, size_binop and size_diffop. */
927 int_binop_types_match_p (enum tree_code code
, const_tree type1
, const_tree type2
)
929 if (!INTEGRAL_TYPE_P (type1
) && !POINTER_TYPE_P (type1
))
931 if (!INTEGRAL_TYPE_P (type2
) && !POINTER_TYPE_P (type2
))
946 return TYPE_UNSIGNED (type1
) == TYPE_UNSIGNED (type2
)
947 && TYPE_PRECISION (type1
) == TYPE_PRECISION (type2
)
948 && TYPE_MODE (type1
) == TYPE_MODE (type2
);
952 /* Combine two integer constants ARG1 and ARG2 under operation CODE
953 to produce a new constant. Return NULL_TREE if we don't know how
954 to evaluate CODE at compile-time. */
957 int_const_binop_1 (enum tree_code code
, const_tree arg1
, const_tree parg2
,
962 tree type
= TREE_TYPE (arg1
);
963 signop sign
= TYPE_SIGN (type
);
964 bool overflow
= false;
966 wide_int arg2
= wide_int::from (parg2
, TYPE_PRECISION (type
),
967 TYPE_SIGN (TREE_TYPE (parg2
)));
972 res
= wi::bit_or (arg1
, arg2
);
976 res
= wi::bit_xor (arg1
, arg2
);
980 res
= wi::bit_and (arg1
, arg2
);
985 if (wi::neg_p (arg2
))
988 if (code
== RSHIFT_EXPR
)
994 if (code
== RSHIFT_EXPR
)
995 /* It's unclear from the C standard whether shifts can overflow.
996 The following code ignores overflow; perhaps a C standard
997 interpretation ruling is needed. */
998 res
= wi::rshift (arg1
, arg2
, sign
);
1000 res
= wi::lshift (arg1
, arg2
);
1005 if (wi::neg_p (arg2
))
1008 if (code
== RROTATE_EXPR
)
1009 code
= LROTATE_EXPR
;
1011 code
= RROTATE_EXPR
;
1014 if (code
== RROTATE_EXPR
)
1015 res
= wi::rrotate (arg1
, arg2
);
1017 res
= wi::lrotate (arg1
, arg2
);
1021 res
= wi::add (arg1
, arg2
, sign
, &overflow
);
1025 res
= wi::sub (arg1
, arg2
, sign
, &overflow
);
1029 res
= wi::mul (arg1
, arg2
, sign
, &overflow
);
1032 case MULT_HIGHPART_EXPR
:
1033 res
= wi::mul_high (arg1
, arg2
, sign
);
1036 case TRUNC_DIV_EXPR
:
1037 case EXACT_DIV_EXPR
:
1040 res
= wi::div_trunc (arg1
, arg2
, sign
, &overflow
);
1043 case FLOOR_DIV_EXPR
:
1046 res
= wi::div_floor (arg1
, arg2
, sign
, &overflow
);
1052 res
= wi::div_ceil (arg1
, arg2
, sign
, &overflow
);
1055 case ROUND_DIV_EXPR
:
1058 res
= wi::div_round (arg1
, arg2
, sign
, &overflow
);
1061 case TRUNC_MOD_EXPR
:
1064 res
= wi::mod_trunc (arg1
, arg2
, sign
, &overflow
);
1067 case FLOOR_MOD_EXPR
:
1070 res
= wi::mod_floor (arg1
, arg2
, sign
, &overflow
);
1076 res
= wi::mod_ceil (arg1
, arg2
, sign
, &overflow
);
1079 case ROUND_MOD_EXPR
:
1082 res
= wi::mod_round (arg1
, arg2
, sign
, &overflow
);
1086 res
= wi::min (arg1
, arg2
, sign
);
1090 res
= wi::max (arg1
, arg2
, sign
);
1097 t
= force_fit_type (type
, res
, overflowable
,
1098 (((sign
== SIGNED
|| overflowable
== -1)
1100 | TREE_OVERFLOW (arg1
) | TREE_OVERFLOW (parg2
)));
1106 int_const_binop (enum tree_code code
, const_tree arg1
, const_tree arg2
)
1108 return int_const_binop_1 (code
, arg1
, arg2
, 1);
1111 /* Combine two constants ARG1 and ARG2 under operation CODE to produce a new
1112 constant. We assume ARG1 and ARG2 have the same data type, or at least
1113 are the same kind of constant and the same machine mode. Return zero if
1114 combining the constants is not allowed in the current operating mode. */
1117 const_binop (enum tree_code code
, tree arg1
, tree arg2
)
1119 /* Sanity check for the recursive cases. */
1126 if (TREE_CODE (arg1
) == INTEGER_CST
&& TREE_CODE (arg2
) == INTEGER_CST
)
1128 if (code
== POINTER_PLUS_EXPR
)
1129 return int_const_binop (PLUS_EXPR
,
1130 arg1
, fold_convert (TREE_TYPE (arg1
), arg2
));
1132 return int_const_binop (code
, arg1
, arg2
);
1135 if (TREE_CODE (arg1
) == REAL_CST
&& TREE_CODE (arg2
) == REAL_CST
)
1140 REAL_VALUE_TYPE value
;
1141 REAL_VALUE_TYPE result
;
1145 /* The following codes are handled by real_arithmetic. */
1160 d1
= TREE_REAL_CST (arg1
);
1161 d2
= TREE_REAL_CST (arg2
);
1163 type
= TREE_TYPE (arg1
);
1164 mode
= TYPE_MODE (type
);
1166 /* Don't perform operation if we honor signaling NaNs and
1167 either operand is a signaling NaN. */
1168 if (HONOR_SNANS (mode
)
1169 && (REAL_VALUE_ISSIGNALING_NAN (d1
)
1170 || REAL_VALUE_ISSIGNALING_NAN (d2
)))
1173 /* Don't perform operation if it would raise a division
1174 by zero exception. */
1175 if (code
== RDIV_EXPR
1176 && real_equal (&d2
, &dconst0
)
1177 && (flag_trapping_math
|| ! MODE_HAS_INFINITIES (mode
)))
1180 /* If either operand is a NaN, just return it. Otherwise, set up
1181 for floating-point trap; we return an overflow. */
1182 if (REAL_VALUE_ISNAN (d1
))
1184 /* Make resulting NaN value to be qNaN when flag_signaling_nans
1187 t
= build_real (type
, d1
);
1190 else if (REAL_VALUE_ISNAN (d2
))
1192 /* Make resulting NaN value to be qNaN when flag_signaling_nans
1195 t
= build_real (type
, d2
);
1199 inexact
= real_arithmetic (&value
, code
, &d1
, &d2
);
1200 real_convert (&result
, mode
, &value
);
1202 /* Don't constant fold this floating point operation if
1203 the result has overflowed and flag_trapping_math. */
1204 if (flag_trapping_math
1205 && MODE_HAS_INFINITIES (mode
)
1206 && REAL_VALUE_ISINF (result
)
1207 && !REAL_VALUE_ISINF (d1
)
1208 && !REAL_VALUE_ISINF (d2
))
1211 /* Don't constant fold this floating point operation if the
1212 result may dependent upon the run-time rounding mode and
1213 flag_rounding_math is set, or if GCC's software emulation
1214 is unable to accurately represent the result. */
1215 if ((flag_rounding_math
1216 || (MODE_COMPOSITE_P (mode
) && !flag_unsafe_math_optimizations
))
1217 && (inexact
|| !real_identical (&result
, &value
)))
1220 t
= build_real (type
, result
);
1222 TREE_OVERFLOW (t
) = TREE_OVERFLOW (arg1
) | TREE_OVERFLOW (arg2
);
1226 if (TREE_CODE (arg1
) == FIXED_CST
)
1228 FIXED_VALUE_TYPE f1
;
1229 FIXED_VALUE_TYPE f2
;
1230 FIXED_VALUE_TYPE result
;
1235 /* The following codes are handled by fixed_arithmetic. */
1241 case TRUNC_DIV_EXPR
:
1242 if (TREE_CODE (arg2
) != FIXED_CST
)
1244 f2
= TREE_FIXED_CST (arg2
);
1250 if (TREE_CODE (arg2
) != INTEGER_CST
)
1253 f2
.data
.high
= w2
.elt (1);
1254 f2
.data
.low
= w2
.elt (0);
1263 f1
= TREE_FIXED_CST (arg1
);
1264 type
= TREE_TYPE (arg1
);
1265 sat_p
= TYPE_SATURATING (type
);
1266 overflow_p
= fixed_arithmetic (&result
, code
, &f1
, &f2
, sat_p
);
1267 t
= build_fixed (type
, result
);
1268 /* Propagate overflow flags. */
1269 if (overflow_p
| TREE_OVERFLOW (arg1
) | TREE_OVERFLOW (arg2
))
1270 TREE_OVERFLOW (t
) = 1;
1274 if (TREE_CODE (arg1
) == COMPLEX_CST
&& TREE_CODE (arg2
) == COMPLEX_CST
)
1276 tree type
= TREE_TYPE (arg1
);
1277 tree r1
= TREE_REALPART (arg1
);
1278 tree i1
= TREE_IMAGPART (arg1
);
1279 tree r2
= TREE_REALPART (arg2
);
1280 tree i2
= TREE_IMAGPART (arg2
);
1287 real
= const_binop (code
, r1
, r2
);
1288 imag
= const_binop (code
, i1
, i2
);
1292 if (COMPLEX_FLOAT_TYPE_P (type
))
1293 return do_mpc_arg2 (arg1
, arg2
, type
,
1294 /* do_nonfinite= */ folding_initializer
,
1297 real
= const_binop (MINUS_EXPR
,
1298 const_binop (MULT_EXPR
, r1
, r2
),
1299 const_binop (MULT_EXPR
, i1
, i2
));
1300 imag
= const_binop (PLUS_EXPR
,
1301 const_binop (MULT_EXPR
, r1
, i2
),
1302 const_binop (MULT_EXPR
, i1
, r2
));
1306 if (COMPLEX_FLOAT_TYPE_P (type
))
1307 return do_mpc_arg2 (arg1
, arg2
, type
,
1308 /* do_nonfinite= */ folding_initializer
,
1311 case TRUNC_DIV_EXPR
:
1313 case FLOOR_DIV_EXPR
:
1314 case ROUND_DIV_EXPR
:
1315 if (flag_complex_method
== 0)
1317 /* Keep this algorithm in sync with
1318 tree-complex.c:expand_complex_div_straight().
1320 Expand complex division to scalars, straightforward algorithm.
1321 a / b = ((ar*br + ai*bi)/t) + i((ai*br - ar*bi)/t)
1325 = const_binop (PLUS_EXPR
,
1326 const_binop (MULT_EXPR
, r2
, r2
),
1327 const_binop (MULT_EXPR
, i2
, i2
));
1329 = const_binop (PLUS_EXPR
,
1330 const_binop (MULT_EXPR
, r1
, r2
),
1331 const_binop (MULT_EXPR
, i1
, i2
));
1333 = const_binop (MINUS_EXPR
,
1334 const_binop (MULT_EXPR
, i1
, r2
),
1335 const_binop (MULT_EXPR
, r1
, i2
));
1337 real
= const_binop (code
, t1
, magsquared
);
1338 imag
= const_binop (code
, t2
, magsquared
);
1342 /* Keep this algorithm in sync with
1343 tree-complex.c:expand_complex_div_wide().
1345 Expand complex division to scalars, modified algorithm to minimize
1346 overflow with wide input ranges. */
1347 tree compare
= fold_build2 (LT_EXPR
, boolean_type_node
,
1348 fold_abs_const (r2
, TREE_TYPE (type
)),
1349 fold_abs_const (i2
, TREE_TYPE (type
)));
1351 if (integer_nonzerop (compare
))
1353 /* In the TRUE branch, we compute
1355 div = (br * ratio) + bi;
1356 tr = (ar * ratio) + ai;
1357 ti = (ai * ratio) - ar;
1360 tree ratio
= const_binop (code
, r2
, i2
);
1361 tree div
= const_binop (PLUS_EXPR
, i2
,
1362 const_binop (MULT_EXPR
, r2
, ratio
));
1363 real
= const_binop (MULT_EXPR
, r1
, ratio
);
1364 real
= const_binop (PLUS_EXPR
, real
, i1
);
1365 real
= const_binop (code
, real
, div
);
1367 imag
= const_binop (MULT_EXPR
, i1
, ratio
);
1368 imag
= const_binop (MINUS_EXPR
, imag
, r1
);
1369 imag
= const_binop (code
, imag
, div
);
1373 /* In the FALSE branch, we compute
1375 divisor = (d * ratio) + c;
1376 tr = (b * ratio) + a;
1377 ti = b - (a * ratio);
1380 tree ratio
= const_binop (code
, i2
, r2
);
1381 tree div
= const_binop (PLUS_EXPR
, r2
,
1382 const_binop (MULT_EXPR
, i2
, ratio
));
1384 real
= const_binop (MULT_EXPR
, i1
, ratio
);
1385 real
= const_binop (PLUS_EXPR
, real
, r1
);
1386 real
= const_binop (code
, real
, div
);
1388 imag
= const_binop (MULT_EXPR
, r1
, ratio
);
1389 imag
= const_binop (MINUS_EXPR
, i1
, imag
);
1390 imag
= const_binop (code
, imag
, div
);
1400 return build_complex (type
, real
, imag
);
1403 if (TREE_CODE (arg1
) == VECTOR_CST
1404 && TREE_CODE (arg2
) == VECTOR_CST
)
1406 tree type
= TREE_TYPE (arg1
);
1407 int count
= TYPE_VECTOR_SUBPARTS (type
), i
;
1408 tree
*elts
= XALLOCAVEC (tree
, count
);
1410 for (i
= 0; i
< count
; i
++)
1412 tree elem1
= VECTOR_CST_ELT (arg1
, i
);
1413 tree elem2
= VECTOR_CST_ELT (arg2
, i
);
1415 elts
[i
] = const_binop (code
, elem1
, elem2
);
1417 /* It is possible that const_binop cannot handle the given
1418 code and return NULL_TREE */
1419 if (elts
[i
] == NULL_TREE
)
1423 return build_vector (type
, elts
);
1426 /* Shifts allow a scalar offset for a vector. */
1427 if (TREE_CODE (arg1
) == VECTOR_CST
1428 && TREE_CODE (arg2
) == INTEGER_CST
)
1430 tree type
= TREE_TYPE (arg1
);
1431 int count
= TYPE_VECTOR_SUBPARTS (type
), i
;
1432 tree
*elts
= XALLOCAVEC (tree
, count
);
1434 for (i
= 0; i
< count
; i
++)
1436 tree elem1
= VECTOR_CST_ELT (arg1
, i
);
1438 elts
[i
] = const_binop (code
, elem1
, arg2
);
1440 /* It is possible that const_binop cannot handle the given
1441 code and return NULL_TREE. */
1442 if (elts
[i
] == NULL_TREE
)
1446 return build_vector (type
, elts
);
1451 /* Overload that adds a TYPE parameter to be able to dispatch
1452 to fold_relational_const. */
1455 const_binop (enum tree_code code
, tree type
, tree arg1
, tree arg2
)
1457 if (TREE_CODE_CLASS (code
) == tcc_comparison
)
1458 return fold_relational_const (code
, type
, arg1
, arg2
);
1460 /* ??? Until we make the const_binop worker take the type of the
1461 result as argument put those cases that need it here. */
1465 if ((TREE_CODE (arg1
) == REAL_CST
1466 && TREE_CODE (arg2
) == REAL_CST
)
1467 || (TREE_CODE (arg1
) == INTEGER_CST
1468 && TREE_CODE (arg2
) == INTEGER_CST
))
1469 return build_complex (type
, arg1
, arg2
);
1472 case VEC_PACK_TRUNC_EXPR
:
1473 case VEC_PACK_FIX_TRUNC_EXPR
:
1475 unsigned int nelts
= TYPE_VECTOR_SUBPARTS (type
), i
;
1478 gcc_assert (TYPE_VECTOR_SUBPARTS (TREE_TYPE (arg1
)) == nelts
/ 2
1479 && TYPE_VECTOR_SUBPARTS (TREE_TYPE (arg2
)) == nelts
/ 2);
1480 if (TREE_CODE (arg1
) != VECTOR_CST
1481 || TREE_CODE (arg2
) != VECTOR_CST
)
1484 elts
= XALLOCAVEC (tree
, nelts
);
1485 if (!vec_cst_ctor_to_array (arg1
, elts
)
1486 || !vec_cst_ctor_to_array (arg2
, elts
+ nelts
/ 2))
1489 for (i
= 0; i
< nelts
; i
++)
1491 elts
[i
] = fold_convert_const (code
== VEC_PACK_TRUNC_EXPR
1492 ? NOP_EXPR
: FIX_TRUNC_EXPR
,
1493 TREE_TYPE (type
), elts
[i
]);
1494 if (elts
[i
] == NULL_TREE
|| !CONSTANT_CLASS_P (elts
[i
]))
1498 return build_vector (type
, elts
);
1501 case VEC_WIDEN_MULT_LO_EXPR
:
1502 case VEC_WIDEN_MULT_HI_EXPR
:
1503 case VEC_WIDEN_MULT_EVEN_EXPR
:
1504 case VEC_WIDEN_MULT_ODD_EXPR
:
1506 unsigned int nelts
= TYPE_VECTOR_SUBPARTS (type
);
1507 unsigned int out
, ofs
, scale
;
1510 gcc_assert (TYPE_VECTOR_SUBPARTS (TREE_TYPE (arg1
)) == nelts
* 2
1511 && TYPE_VECTOR_SUBPARTS (TREE_TYPE (arg2
)) == nelts
* 2);
1512 if (TREE_CODE (arg1
) != VECTOR_CST
|| TREE_CODE (arg2
) != VECTOR_CST
)
1515 elts
= XALLOCAVEC (tree
, nelts
* 4);
1516 if (!vec_cst_ctor_to_array (arg1
, elts
)
1517 || !vec_cst_ctor_to_array (arg2
, elts
+ nelts
* 2))
1520 if (code
== VEC_WIDEN_MULT_LO_EXPR
)
1521 scale
= 0, ofs
= BYTES_BIG_ENDIAN
? nelts
: 0;
1522 else if (code
== VEC_WIDEN_MULT_HI_EXPR
)
1523 scale
= 0, ofs
= BYTES_BIG_ENDIAN
? 0 : nelts
;
1524 else if (code
== VEC_WIDEN_MULT_EVEN_EXPR
)
1526 else /* if (code == VEC_WIDEN_MULT_ODD_EXPR) */
1529 for (out
= 0; out
< nelts
; out
++)
1531 unsigned int in1
= (out
<< scale
) + ofs
;
1532 unsigned int in2
= in1
+ nelts
* 2;
1535 t1
= fold_convert_const (NOP_EXPR
, TREE_TYPE (type
), elts
[in1
]);
1536 t2
= fold_convert_const (NOP_EXPR
, TREE_TYPE (type
), elts
[in2
]);
1538 if (t1
== NULL_TREE
|| t2
== NULL_TREE
)
1540 elts
[out
] = const_binop (MULT_EXPR
, t1
, t2
);
1541 if (elts
[out
] == NULL_TREE
|| !CONSTANT_CLASS_P (elts
[out
]))
1545 return build_vector (type
, elts
);
1551 if (TREE_CODE_CLASS (code
) != tcc_binary
)
1554 /* Make sure type and arg0 have the same saturating flag. */
1555 gcc_checking_assert (TYPE_SATURATING (type
)
1556 == TYPE_SATURATING (TREE_TYPE (arg1
)));
1558 return const_binop (code
, arg1
, arg2
);
1561 /* Compute CODE ARG1 with resulting type TYPE with ARG1 being constant.
1562 Return zero if computing the constants is not possible. */
1565 const_unop (enum tree_code code
, tree type
, tree arg0
)
1567 /* Don't perform the operation, other than NEGATE and ABS, if
1568 flag_signaling_nans is on and the operand is a signaling NaN. */
1569 if (TREE_CODE (arg0
) == REAL_CST
1570 && HONOR_SNANS (TYPE_MODE (TREE_TYPE (arg0
)))
1571 && REAL_VALUE_ISSIGNALING_NAN (TREE_REAL_CST (arg0
))
1572 && code
!= NEGATE_EXPR
1573 && code
!= ABS_EXPR
)
1580 case FIX_TRUNC_EXPR
:
1581 case FIXED_CONVERT_EXPR
:
1582 return fold_convert_const (code
, type
, arg0
);
1584 case ADDR_SPACE_CONVERT_EXPR
:
1585 /* If the source address is 0, and the source address space
1586 cannot have a valid object at 0, fold to dest type null. */
1587 if (integer_zerop (arg0
)
1588 && !(targetm
.addr_space
.zero_address_valid
1589 (TYPE_ADDR_SPACE (TREE_TYPE (TREE_TYPE (arg0
))))))
1590 return fold_convert_const (code
, type
, arg0
);
1593 case VIEW_CONVERT_EXPR
:
1594 return fold_view_convert_expr (type
, arg0
);
1598 /* Can't call fold_negate_const directly here as that doesn't
1599 handle all cases and we might not be able to negate some
1601 tree tem
= fold_negate_expr (UNKNOWN_LOCATION
, arg0
);
1602 if (tem
&& CONSTANT_CLASS_P (tem
))
1608 if (TREE_CODE (arg0
) == INTEGER_CST
|| TREE_CODE (arg0
) == REAL_CST
)
1609 return fold_abs_const (arg0
, type
);
1613 if (TREE_CODE (arg0
) == COMPLEX_CST
)
1615 tree ipart
= fold_negate_const (TREE_IMAGPART (arg0
),
1617 return build_complex (type
, TREE_REALPART (arg0
), ipart
);
1622 if (TREE_CODE (arg0
) == INTEGER_CST
)
1623 return fold_not_const (arg0
, type
);
1624 /* Perform BIT_NOT_EXPR on each element individually. */
1625 else if (TREE_CODE (arg0
) == VECTOR_CST
)
1629 unsigned count
= VECTOR_CST_NELTS (arg0
), i
;
1631 elements
= XALLOCAVEC (tree
, count
);
1632 for (i
= 0; i
< count
; i
++)
1634 elem
= VECTOR_CST_ELT (arg0
, i
);
1635 elem
= const_unop (BIT_NOT_EXPR
, TREE_TYPE (type
), elem
);
1636 if (elem
== NULL_TREE
)
1641 return build_vector (type
, elements
);
1645 case TRUTH_NOT_EXPR
:
1646 if (TREE_CODE (arg0
) == INTEGER_CST
)
1647 return constant_boolean_node (integer_zerop (arg0
), type
);
1651 if (TREE_CODE (arg0
) == COMPLEX_CST
)
1652 return fold_convert (type
, TREE_REALPART (arg0
));
1656 if (TREE_CODE (arg0
) == COMPLEX_CST
)
1657 return fold_convert (type
, TREE_IMAGPART (arg0
));
1660 case VEC_UNPACK_LO_EXPR
:
1661 case VEC_UNPACK_HI_EXPR
:
1662 case VEC_UNPACK_FLOAT_LO_EXPR
:
1663 case VEC_UNPACK_FLOAT_HI_EXPR
:
1665 unsigned int nelts
= TYPE_VECTOR_SUBPARTS (type
), i
;
1667 enum tree_code subcode
;
1669 gcc_assert (TYPE_VECTOR_SUBPARTS (TREE_TYPE (arg0
)) == nelts
* 2);
1670 if (TREE_CODE (arg0
) != VECTOR_CST
)
1673 elts
= XALLOCAVEC (tree
, nelts
* 2);
1674 if (!vec_cst_ctor_to_array (arg0
, elts
))
1677 if ((!BYTES_BIG_ENDIAN
) ^ (code
== VEC_UNPACK_LO_EXPR
1678 || code
== VEC_UNPACK_FLOAT_LO_EXPR
))
1681 if (code
== VEC_UNPACK_LO_EXPR
|| code
== VEC_UNPACK_HI_EXPR
)
1684 subcode
= FLOAT_EXPR
;
1686 for (i
= 0; i
< nelts
; i
++)
1688 elts
[i
] = fold_convert_const (subcode
, TREE_TYPE (type
), elts
[i
]);
1689 if (elts
[i
] == NULL_TREE
|| !CONSTANT_CLASS_P (elts
[i
]))
1693 return build_vector (type
, elts
);
1696 case REDUC_MIN_EXPR
:
1697 case REDUC_MAX_EXPR
:
1698 case REDUC_PLUS_EXPR
:
1700 unsigned int nelts
, i
;
1702 enum tree_code subcode
;
1704 if (TREE_CODE (arg0
) != VECTOR_CST
)
1706 nelts
= TYPE_VECTOR_SUBPARTS (TREE_TYPE (arg0
));
1708 elts
= XALLOCAVEC (tree
, nelts
);
1709 if (!vec_cst_ctor_to_array (arg0
, elts
))
1714 case REDUC_MIN_EXPR
: subcode
= MIN_EXPR
; break;
1715 case REDUC_MAX_EXPR
: subcode
= MAX_EXPR
; break;
1716 case REDUC_PLUS_EXPR
: subcode
= PLUS_EXPR
; break;
1717 default: gcc_unreachable ();
1720 for (i
= 1; i
< nelts
; i
++)
1722 elts
[0] = const_binop (subcode
, elts
[0], elts
[i
]);
1723 if (elts
[0] == NULL_TREE
|| !CONSTANT_CLASS_P (elts
[0]))
1737 /* Create a sizetype INT_CST node with NUMBER sign extended. KIND
1738 indicates which particular sizetype to create. */
1741 size_int_kind (HOST_WIDE_INT number
, enum size_type_kind kind
)
1743 return build_int_cst (sizetype_tab
[(int) kind
], number
);
1746 /* Combine operands OP1 and OP2 with arithmetic operation CODE. CODE
1747 is a tree code. The type of the result is taken from the operands.
1748 Both must be equivalent integer types, ala int_binop_types_match_p.
1749 If the operands are constant, so is the result. */
1752 size_binop_loc (location_t loc
, enum tree_code code
, tree arg0
, tree arg1
)
1754 tree type
= TREE_TYPE (arg0
);
1756 if (arg0
== error_mark_node
|| arg1
== error_mark_node
)
1757 return error_mark_node
;
1759 gcc_assert (int_binop_types_match_p (code
, TREE_TYPE (arg0
),
1762 /* Handle the special case of two integer constants faster. */
1763 if (TREE_CODE (arg0
) == INTEGER_CST
&& TREE_CODE (arg1
) == INTEGER_CST
)
1765 /* And some specific cases even faster than that. */
1766 if (code
== PLUS_EXPR
)
1768 if (integer_zerop (arg0
) && !TREE_OVERFLOW (arg0
))
1770 if (integer_zerop (arg1
) && !TREE_OVERFLOW (arg1
))
1773 else if (code
== MINUS_EXPR
)
1775 if (integer_zerop (arg1
) && !TREE_OVERFLOW (arg1
))
1778 else if (code
== MULT_EXPR
)
1780 if (integer_onep (arg0
) && !TREE_OVERFLOW (arg0
))
1784 /* Handle general case of two integer constants. For sizetype
1785 constant calculations we always want to know about overflow,
1786 even in the unsigned case. */
1787 return int_const_binop_1 (code
, arg0
, arg1
, -1);
1790 return fold_build2_loc (loc
, code
, type
, arg0
, arg1
);
1793 /* Given two values, either both of sizetype or both of bitsizetype,
1794 compute the difference between the two values. Return the value
1795 in signed type corresponding to the type of the operands. */
1798 size_diffop_loc (location_t loc
, tree arg0
, tree arg1
)
1800 tree type
= TREE_TYPE (arg0
);
1803 gcc_assert (int_binop_types_match_p (MINUS_EXPR
, TREE_TYPE (arg0
),
1806 /* If the type is already signed, just do the simple thing. */
1807 if (!TYPE_UNSIGNED (type
))
1808 return size_binop_loc (loc
, MINUS_EXPR
, arg0
, arg1
);
1810 if (type
== sizetype
)
1812 else if (type
== bitsizetype
)
1813 ctype
= sbitsizetype
;
1815 ctype
= signed_type_for (type
);
1817 /* If either operand is not a constant, do the conversions to the signed
1818 type and subtract. The hardware will do the right thing with any
1819 overflow in the subtraction. */
1820 if (TREE_CODE (arg0
) != INTEGER_CST
|| TREE_CODE (arg1
) != INTEGER_CST
)
1821 return size_binop_loc (loc
, MINUS_EXPR
,
1822 fold_convert_loc (loc
, ctype
, arg0
),
1823 fold_convert_loc (loc
, ctype
, arg1
));
1825 /* If ARG0 is larger than ARG1, subtract and return the result in CTYPE.
1826 Otherwise, subtract the other way, convert to CTYPE (we know that can't
1827 overflow) and negate (which can't either). Special-case a result
1828 of zero while we're here. */
1829 if (tree_int_cst_equal (arg0
, arg1
))
1830 return build_int_cst (ctype
, 0);
1831 else if (tree_int_cst_lt (arg1
, arg0
))
1832 return fold_convert_loc (loc
, ctype
,
1833 size_binop_loc (loc
, MINUS_EXPR
, arg0
, arg1
));
1835 return size_binop_loc (loc
, MINUS_EXPR
, build_int_cst (ctype
, 0),
1836 fold_convert_loc (loc
, ctype
,
1837 size_binop_loc (loc
,
1842 /* A subroutine of fold_convert_const handling conversions of an
1843 INTEGER_CST to another integer type. */
1846 fold_convert_const_int_from_int (tree type
, const_tree arg1
)
1848 /* Given an integer constant, make new constant with new type,
1849 appropriately sign-extended or truncated. Use widest_int
1850 so that any extension is done according ARG1's type. */
1851 return force_fit_type (type
, wi::to_widest (arg1
),
1852 !POINTER_TYPE_P (TREE_TYPE (arg1
)),
1853 TREE_OVERFLOW (arg1
));
1856 /* A subroutine of fold_convert_const handling conversions a REAL_CST
1857 to an integer type. */
1860 fold_convert_const_int_from_real (enum tree_code code
, tree type
, const_tree arg1
)
1862 bool overflow
= false;
1865 /* The following code implements the floating point to integer
1866 conversion rules required by the Java Language Specification,
1867 that IEEE NaNs are mapped to zero and values that overflow
1868 the target precision saturate, i.e. values greater than
1869 INT_MAX are mapped to INT_MAX, and values less than INT_MIN
1870 are mapped to INT_MIN. These semantics are allowed by the
1871 C and C++ standards that simply state that the behavior of
1872 FP-to-integer conversion is unspecified upon overflow. */
1876 REAL_VALUE_TYPE x
= TREE_REAL_CST (arg1
);
1880 case FIX_TRUNC_EXPR
:
1881 real_trunc (&r
, VOIDmode
, &x
);
1888 /* If R is NaN, return zero and show we have an overflow. */
1889 if (REAL_VALUE_ISNAN (r
))
1892 val
= wi::zero (TYPE_PRECISION (type
));
1895 /* See if R is less than the lower bound or greater than the
1900 tree lt
= TYPE_MIN_VALUE (type
);
1901 REAL_VALUE_TYPE l
= real_value_from_int_cst (NULL_TREE
, lt
);
1902 if (real_less (&r
, &l
))
1911 tree ut
= TYPE_MAX_VALUE (type
);
1914 REAL_VALUE_TYPE u
= real_value_from_int_cst (NULL_TREE
, ut
);
1915 if (real_less (&u
, &r
))
1924 val
= real_to_integer (&r
, &overflow
, TYPE_PRECISION (type
));
1926 t
= force_fit_type (type
, val
, -1, overflow
| TREE_OVERFLOW (arg1
));
1930 /* A subroutine of fold_convert_const handling conversions of a
1931 FIXED_CST to an integer type. */
1934 fold_convert_const_int_from_fixed (tree type
, const_tree arg1
)
1937 double_int temp
, temp_trunc
;
1940 /* Right shift FIXED_CST to temp by fbit. */
1941 temp
= TREE_FIXED_CST (arg1
).data
;
1942 mode
= TREE_FIXED_CST (arg1
).mode
;
1943 if (GET_MODE_FBIT (mode
) < HOST_BITS_PER_DOUBLE_INT
)
1945 temp
= temp
.rshift (GET_MODE_FBIT (mode
),
1946 HOST_BITS_PER_DOUBLE_INT
,
1947 SIGNED_FIXED_POINT_MODE_P (mode
));
1949 /* Left shift temp to temp_trunc by fbit. */
1950 temp_trunc
= temp
.lshift (GET_MODE_FBIT (mode
),
1951 HOST_BITS_PER_DOUBLE_INT
,
1952 SIGNED_FIXED_POINT_MODE_P (mode
));
1956 temp
= double_int_zero
;
1957 temp_trunc
= double_int_zero
;
1960 /* If FIXED_CST is negative, we need to round the value toward 0.
1961 By checking if the fractional bits are not zero to add 1 to temp. */
1962 if (SIGNED_FIXED_POINT_MODE_P (mode
)
1963 && temp_trunc
.is_negative ()
1964 && TREE_FIXED_CST (arg1
).data
!= temp_trunc
)
1965 temp
+= double_int_one
;
1967 /* Given a fixed-point constant, make new constant with new type,
1968 appropriately sign-extended or truncated. */
1969 t
= force_fit_type (type
, temp
, -1,
1970 (temp
.is_negative ()
1971 && (TYPE_UNSIGNED (type
)
1972 < TYPE_UNSIGNED (TREE_TYPE (arg1
))))
1973 | TREE_OVERFLOW (arg1
));
1978 /* A subroutine of fold_convert_const handling conversions a REAL_CST
1979 to another floating point type. */
1982 fold_convert_const_real_from_real (tree type
, const_tree arg1
)
1984 REAL_VALUE_TYPE value
;
1987 /* Don't perform the operation if flag_signaling_nans is on
1988 and the operand is a signaling NaN. */
1989 if (HONOR_SNANS (TYPE_MODE (TREE_TYPE (arg1
)))
1990 && REAL_VALUE_ISSIGNALING_NAN (TREE_REAL_CST (arg1
)))
1993 real_convert (&value
, TYPE_MODE (type
), &TREE_REAL_CST (arg1
));
1994 t
= build_real (type
, value
);
1996 /* If converting an infinity or NAN to a representation that doesn't
1997 have one, set the overflow bit so that we can produce some kind of
1998 error message at the appropriate point if necessary. It's not the
1999 most user-friendly message, but it's better than nothing. */
2000 if (REAL_VALUE_ISINF (TREE_REAL_CST (arg1
))
2001 && !MODE_HAS_INFINITIES (TYPE_MODE (type
)))
2002 TREE_OVERFLOW (t
) = 1;
2003 else if (REAL_VALUE_ISNAN (TREE_REAL_CST (arg1
))
2004 && !MODE_HAS_NANS (TYPE_MODE (type
)))
2005 TREE_OVERFLOW (t
) = 1;
2006 /* Regular overflow, conversion produced an infinity in a mode that
2007 can't represent them. */
2008 else if (!MODE_HAS_INFINITIES (TYPE_MODE (type
))
2009 && REAL_VALUE_ISINF (value
)
2010 && !REAL_VALUE_ISINF (TREE_REAL_CST (arg1
)))
2011 TREE_OVERFLOW (t
) = 1;
2013 TREE_OVERFLOW (t
) = TREE_OVERFLOW (arg1
);
2017 /* A subroutine of fold_convert_const handling conversions a FIXED_CST
2018 to a floating point type. */
2021 fold_convert_const_real_from_fixed (tree type
, const_tree arg1
)
2023 REAL_VALUE_TYPE value
;
2026 real_convert_from_fixed (&value
, TYPE_MODE (type
), &TREE_FIXED_CST (arg1
));
2027 t
= build_real (type
, value
);
2029 TREE_OVERFLOW (t
) = TREE_OVERFLOW (arg1
);
2033 /* A subroutine of fold_convert_const handling conversions a FIXED_CST
2034 to another fixed-point type. */
2037 fold_convert_const_fixed_from_fixed (tree type
, const_tree arg1
)
2039 FIXED_VALUE_TYPE value
;
2043 overflow_p
= fixed_convert (&value
, TYPE_MODE (type
), &TREE_FIXED_CST (arg1
),
2044 TYPE_SATURATING (type
));
2045 t
= build_fixed (type
, value
);
2047 /* Propagate overflow flags. */
2048 if (overflow_p
| TREE_OVERFLOW (arg1
))
2049 TREE_OVERFLOW (t
) = 1;
2053 /* A subroutine of fold_convert_const handling conversions an INTEGER_CST
2054 to a fixed-point type. */
2057 fold_convert_const_fixed_from_int (tree type
, const_tree arg1
)
2059 FIXED_VALUE_TYPE value
;
2064 gcc_assert (TREE_INT_CST_NUNITS (arg1
) <= 2);
2066 di
.low
= TREE_INT_CST_ELT (arg1
, 0);
2067 if (TREE_INT_CST_NUNITS (arg1
) == 1)
2068 di
.high
= (HOST_WIDE_INT
) di
.low
< 0 ? (HOST_WIDE_INT
) -1 : 0;
2070 di
.high
= TREE_INT_CST_ELT (arg1
, 1);
2072 overflow_p
= fixed_convert_from_int (&value
, TYPE_MODE (type
), di
,
2073 TYPE_UNSIGNED (TREE_TYPE (arg1
)),
2074 TYPE_SATURATING (type
));
2075 t
= build_fixed (type
, value
);
2077 /* Propagate overflow flags. */
2078 if (overflow_p
| TREE_OVERFLOW (arg1
))
2079 TREE_OVERFLOW (t
) = 1;
2083 /* A subroutine of fold_convert_const handling conversions a REAL_CST
2084 to a fixed-point type. */
2087 fold_convert_const_fixed_from_real (tree type
, const_tree arg1
)
2089 FIXED_VALUE_TYPE value
;
2093 overflow_p
= fixed_convert_from_real (&value
, TYPE_MODE (type
),
2094 &TREE_REAL_CST (arg1
),
2095 TYPE_SATURATING (type
));
2096 t
= build_fixed (type
, value
);
2098 /* Propagate overflow flags. */
2099 if (overflow_p
| TREE_OVERFLOW (arg1
))
2100 TREE_OVERFLOW (t
) = 1;
2104 /* Attempt to fold type conversion operation CODE of expression ARG1 to
2105 type TYPE. If no simplification can be done return NULL_TREE. */
2108 fold_convert_const (enum tree_code code
, tree type
, tree arg1
)
2110 if (TREE_TYPE (arg1
) == type
)
2113 if (POINTER_TYPE_P (type
) || INTEGRAL_TYPE_P (type
)
2114 || TREE_CODE (type
) == OFFSET_TYPE
)
2116 if (TREE_CODE (arg1
) == INTEGER_CST
)
2117 return fold_convert_const_int_from_int (type
, arg1
);
2118 else if (TREE_CODE (arg1
) == REAL_CST
)
2119 return fold_convert_const_int_from_real (code
, type
, arg1
);
2120 else if (TREE_CODE (arg1
) == FIXED_CST
)
2121 return fold_convert_const_int_from_fixed (type
, arg1
);
2123 else if (TREE_CODE (type
) == REAL_TYPE
)
2125 if (TREE_CODE (arg1
) == INTEGER_CST
)
2126 return build_real_from_int_cst (type
, arg1
);
2127 else if (TREE_CODE (arg1
) == REAL_CST
)
2128 return fold_convert_const_real_from_real (type
, arg1
);
2129 else if (TREE_CODE (arg1
) == FIXED_CST
)
2130 return fold_convert_const_real_from_fixed (type
, arg1
);
2132 else if (TREE_CODE (type
) == FIXED_POINT_TYPE
)
2134 if (TREE_CODE (arg1
) == FIXED_CST
)
2135 return fold_convert_const_fixed_from_fixed (type
, arg1
);
2136 else if (TREE_CODE (arg1
) == INTEGER_CST
)
2137 return fold_convert_const_fixed_from_int (type
, arg1
);
2138 else if (TREE_CODE (arg1
) == REAL_CST
)
2139 return fold_convert_const_fixed_from_real (type
, arg1
);
2141 else if (TREE_CODE (type
) == VECTOR_TYPE
)
2143 if (TREE_CODE (arg1
) == VECTOR_CST
2144 && TYPE_VECTOR_SUBPARTS (type
) == VECTOR_CST_NELTS (arg1
))
2146 int len
= TYPE_VECTOR_SUBPARTS (type
);
2147 tree elttype
= TREE_TYPE (type
);
2148 tree
*v
= XALLOCAVEC (tree
, len
);
2149 for (int i
= 0; i
< len
; ++i
)
2151 tree elt
= VECTOR_CST_ELT (arg1
, i
);
2152 tree cvt
= fold_convert_const (code
, elttype
, elt
);
2153 if (cvt
== NULL_TREE
)
2157 return build_vector (type
, v
);
2163 /* Construct a vector of zero elements of vector type TYPE. */
2166 build_zero_vector (tree type
)
2170 t
= fold_convert_const (NOP_EXPR
, TREE_TYPE (type
), integer_zero_node
);
2171 return build_vector_from_val (type
, t
);
2174 /* Returns true, if ARG is convertible to TYPE using a NOP_EXPR. */
2177 fold_convertible_p (const_tree type
, const_tree arg
)
2179 tree orig
= TREE_TYPE (arg
);
2184 if (TREE_CODE (arg
) == ERROR_MARK
2185 || TREE_CODE (type
) == ERROR_MARK
2186 || TREE_CODE (orig
) == ERROR_MARK
)
2189 if (TYPE_MAIN_VARIANT (type
) == TYPE_MAIN_VARIANT (orig
))
2192 switch (TREE_CODE (type
))
2194 case INTEGER_TYPE
: case ENUMERAL_TYPE
: case BOOLEAN_TYPE
:
2195 case POINTER_TYPE
: case REFERENCE_TYPE
:
2197 return (INTEGRAL_TYPE_P (orig
) || POINTER_TYPE_P (orig
)
2198 || TREE_CODE (orig
) == OFFSET_TYPE
);
2201 case FIXED_POINT_TYPE
:
2205 return TREE_CODE (type
) == TREE_CODE (orig
);
2212 /* Convert expression ARG to type TYPE. Used by the middle-end for
2213 simple conversions in preference to calling the front-end's convert. */
2216 fold_convert_loc (location_t loc
, tree type
, tree arg
)
2218 tree orig
= TREE_TYPE (arg
);
2224 if (TREE_CODE (arg
) == ERROR_MARK
2225 || TREE_CODE (type
) == ERROR_MARK
2226 || TREE_CODE (orig
) == ERROR_MARK
)
2227 return error_mark_node
;
2229 switch (TREE_CODE (type
))
2232 case REFERENCE_TYPE
:
2233 /* Handle conversions between pointers to different address spaces. */
2234 if (POINTER_TYPE_P (orig
)
2235 && (TYPE_ADDR_SPACE (TREE_TYPE (type
))
2236 != TYPE_ADDR_SPACE (TREE_TYPE (orig
))))
2237 return fold_build1_loc (loc
, ADDR_SPACE_CONVERT_EXPR
, type
, arg
);
2240 case INTEGER_TYPE
: case ENUMERAL_TYPE
: case BOOLEAN_TYPE
:
2242 if (TREE_CODE (arg
) == INTEGER_CST
)
2244 tem
= fold_convert_const (NOP_EXPR
, type
, arg
);
2245 if (tem
!= NULL_TREE
)
2248 if (INTEGRAL_TYPE_P (orig
) || POINTER_TYPE_P (orig
)
2249 || TREE_CODE (orig
) == OFFSET_TYPE
)
2250 return fold_build1_loc (loc
, NOP_EXPR
, type
, arg
);
2251 if (TREE_CODE (orig
) == COMPLEX_TYPE
)
2252 return fold_convert_loc (loc
, type
,
2253 fold_build1_loc (loc
, REALPART_EXPR
,
2254 TREE_TYPE (orig
), arg
));
2255 gcc_assert (TREE_CODE (orig
) == VECTOR_TYPE
2256 && tree_int_cst_equal (TYPE_SIZE (type
), TYPE_SIZE (orig
)));
2257 return fold_build1_loc (loc
, VIEW_CONVERT_EXPR
, type
, arg
);
2260 if (TREE_CODE (arg
) == INTEGER_CST
)
2262 tem
= fold_convert_const (FLOAT_EXPR
, type
, arg
);
2263 if (tem
!= NULL_TREE
)
2266 else if (TREE_CODE (arg
) == REAL_CST
)
2268 tem
= fold_convert_const (NOP_EXPR
, type
, arg
);
2269 if (tem
!= NULL_TREE
)
2272 else if (TREE_CODE (arg
) == FIXED_CST
)
2274 tem
= fold_convert_const (FIXED_CONVERT_EXPR
, type
, arg
);
2275 if (tem
!= NULL_TREE
)
2279 switch (TREE_CODE (orig
))
2282 case BOOLEAN_TYPE
: case ENUMERAL_TYPE
:
2283 case POINTER_TYPE
: case REFERENCE_TYPE
:
2284 return fold_build1_loc (loc
, FLOAT_EXPR
, type
, arg
);
2287 return fold_build1_loc (loc
, NOP_EXPR
, type
, arg
);
2289 case FIXED_POINT_TYPE
:
2290 return fold_build1_loc (loc
, FIXED_CONVERT_EXPR
, type
, arg
);
2293 tem
= fold_build1_loc (loc
, REALPART_EXPR
, TREE_TYPE (orig
), arg
);
2294 return fold_convert_loc (loc
, type
, tem
);
2300 case FIXED_POINT_TYPE
:
2301 if (TREE_CODE (arg
) == FIXED_CST
|| TREE_CODE (arg
) == INTEGER_CST
2302 || TREE_CODE (arg
) == REAL_CST
)
2304 tem
= fold_convert_const (FIXED_CONVERT_EXPR
, type
, arg
);
2305 if (tem
!= NULL_TREE
)
2306 goto fold_convert_exit
;
2309 switch (TREE_CODE (orig
))
2311 case FIXED_POINT_TYPE
:
2316 return fold_build1_loc (loc
, FIXED_CONVERT_EXPR
, type
, arg
);
2319 tem
= fold_build1_loc (loc
, REALPART_EXPR
, TREE_TYPE (orig
), arg
);
2320 return fold_convert_loc (loc
, type
, tem
);
2327 switch (TREE_CODE (orig
))
2330 case BOOLEAN_TYPE
: case ENUMERAL_TYPE
:
2331 case POINTER_TYPE
: case REFERENCE_TYPE
:
2333 case FIXED_POINT_TYPE
:
2334 return fold_build2_loc (loc
, COMPLEX_EXPR
, type
,
2335 fold_convert_loc (loc
, TREE_TYPE (type
), arg
),
2336 fold_convert_loc (loc
, TREE_TYPE (type
),
2337 integer_zero_node
));
2342 if (TREE_CODE (arg
) == COMPLEX_EXPR
)
2344 rpart
= fold_convert_loc (loc
, TREE_TYPE (type
),
2345 TREE_OPERAND (arg
, 0));
2346 ipart
= fold_convert_loc (loc
, TREE_TYPE (type
),
2347 TREE_OPERAND (arg
, 1));
2348 return fold_build2_loc (loc
, COMPLEX_EXPR
, type
, rpart
, ipart
);
2351 arg
= save_expr (arg
);
2352 rpart
= fold_build1_loc (loc
, REALPART_EXPR
, TREE_TYPE (orig
), arg
);
2353 ipart
= fold_build1_loc (loc
, IMAGPART_EXPR
, TREE_TYPE (orig
), arg
);
2354 rpart
= fold_convert_loc (loc
, TREE_TYPE (type
), rpart
);
2355 ipart
= fold_convert_loc (loc
, TREE_TYPE (type
), ipart
);
2356 return fold_build2_loc (loc
, COMPLEX_EXPR
, type
, rpart
, ipart
);
2364 if (integer_zerop (arg
))
2365 return build_zero_vector (type
);
2366 gcc_assert (tree_int_cst_equal (TYPE_SIZE (type
), TYPE_SIZE (orig
)));
2367 gcc_assert (INTEGRAL_TYPE_P (orig
) || POINTER_TYPE_P (orig
)
2368 || TREE_CODE (orig
) == VECTOR_TYPE
);
2369 return fold_build1_loc (loc
, VIEW_CONVERT_EXPR
, type
, arg
);
2372 tem
= fold_ignored_result (arg
);
2373 return fold_build1_loc (loc
, NOP_EXPR
, type
, tem
);
2376 if (TYPE_MAIN_VARIANT (type
) == TYPE_MAIN_VARIANT (orig
))
2377 return fold_build1_loc (loc
, NOP_EXPR
, type
, arg
);
2381 protected_set_expr_location_unshare (tem
, loc
);
2385 /* Return false if expr can be assumed not to be an lvalue, true
2389 maybe_lvalue_p (const_tree x
)
2391 /* We only need to wrap lvalue tree codes. */
2392 switch (TREE_CODE (x
))
2405 case ARRAY_RANGE_REF
:
2411 case PREINCREMENT_EXPR
:
2412 case PREDECREMENT_EXPR
:
2414 case TRY_CATCH_EXPR
:
2415 case WITH_CLEANUP_EXPR
:
2424 /* Assume the worst for front-end tree codes. */
2425 if ((int)TREE_CODE (x
) >= NUM_TREE_CODES
)
2433 /* Return an expr equal to X but certainly not valid as an lvalue. */
2436 non_lvalue_loc (location_t loc
, tree x
)
2438 /* While we are in GIMPLE, NON_LVALUE_EXPR doesn't mean anything to
2443 if (! maybe_lvalue_p (x
))
2445 return build1_loc (loc
, NON_LVALUE_EXPR
, TREE_TYPE (x
), x
);
2448 /* When pedantic, return an expr equal to X but certainly not valid as a
2449 pedantic lvalue. Otherwise, return X. */
2452 pedantic_non_lvalue_loc (location_t loc
, tree x
)
2454 return protected_set_expr_location_unshare (x
, loc
);
2457 /* Given a tree comparison code, return the code that is the logical inverse.
2458 It is generally not safe to do this for floating-point comparisons, except
2459 for EQ_EXPR, NE_EXPR, ORDERED_EXPR and UNORDERED_EXPR, so we return
2460 ERROR_MARK in this case. */
2463 invert_tree_comparison (enum tree_code code
, bool honor_nans
)
2465 if (honor_nans
&& flag_trapping_math
&& code
!= EQ_EXPR
&& code
!= NE_EXPR
2466 && code
!= ORDERED_EXPR
&& code
!= UNORDERED_EXPR
)
2476 return honor_nans
? UNLE_EXPR
: LE_EXPR
;
2478 return honor_nans
? UNLT_EXPR
: LT_EXPR
;
2480 return honor_nans
? UNGE_EXPR
: GE_EXPR
;
2482 return honor_nans
? UNGT_EXPR
: GT_EXPR
;
2496 return UNORDERED_EXPR
;
2497 case UNORDERED_EXPR
:
2498 return ORDERED_EXPR
;
2504 /* Similar, but return the comparison that results if the operands are
2505 swapped. This is safe for floating-point. */
2508 swap_tree_comparison (enum tree_code code
)
2515 case UNORDERED_EXPR
:
2541 /* Convert a comparison tree code from an enum tree_code representation
2542 into a compcode bit-based encoding. This function is the inverse of
2543 compcode_to_comparison. */
2545 static enum comparison_code
2546 comparison_to_compcode (enum tree_code code
)
2563 return COMPCODE_ORD
;
2564 case UNORDERED_EXPR
:
2565 return COMPCODE_UNORD
;
2567 return COMPCODE_UNLT
;
2569 return COMPCODE_UNEQ
;
2571 return COMPCODE_UNLE
;
2573 return COMPCODE_UNGT
;
2575 return COMPCODE_LTGT
;
2577 return COMPCODE_UNGE
;
2583 /* Convert a compcode bit-based encoding of a comparison operator back
2584 to GCC's enum tree_code representation. This function is the
2585 inverse of comparison_to_compcode. */
2587 static enum tree_code
2588 compcode_to_comparison (enum comparison_code code
)
2605 return ORDERED_EXPR
;
2606 case COMPCODE_UNORD
:
2607 return UNORDERED_EXPR
;
2625 /* Return a tree for the comparison which is the combination of
2626 doing the AND or OR (depending on CODE) of the two operations LCODE
2627 and RCODE on the identical operands LL_ARG and LR_ARG. Take into account
2628 the possibility of trapping if the mode has NaNs, and return NULL_TREE
2629 if this makes the transformation invalid. */
2632 combine_comparisons (location_t loc
,
2633 enum tree_code code
, enum tree_code lcode
,
2634 enum tree_code rcode
, tree truth_type
,
2635 tree ll_arg
, tree lr_arg
)
2637 bool honor_nans
= HONOR_NANS (ll_arg
);
2638 enum comparison_code lcompcode
= comparison_to_compcode (lcode
);
2639 enum comparison_code rcompcode
= comparison_to_compcode (rcode
);
2644 case TRUTH_AND_EXPR
: case TRUTH_ANDIF_EXPR
:
2645 compcode
= lcompcode
& rcompcode
;
2648 case TRUTH_OR_EXPR
: case TRUTH_ORIF_EXPR
:
2649 compcode
= lcompcode
| rcompcode
;
2658 /* Eliminate unordered comparisons, as well as LTGT and ORD
2659 which are not used unless the mode has NaNs. */
2660 compcode
&= ~COMPCODE_UNORD
;
2661 if (compcode
== COMPCODE_LTGT
)
2662 compcode
= COMPCODE_NE
;
2663 else if (compcode
== COMPCODE_ORD
)
2664 compcode
= COMPCODE_TRUE
;
2666 else if (flag_trapping_math
)
2668 /* Check that the original operation and the optimized ones will trap
2669 under the same condition. */
2670 bool ltrap
= (lcompcode
& COMPCODE_UNORD
) == 0
2671 && (lcompcode
!= COMPCODE_EQ
)
2672 && (lcompcode
!= COMPCODE_ORD
);
2673 bool rtrap
= (rcompcode
& COMPCODE_UNORD
) == 0
2674 && (rcompcode
!= COMPCODE_EQ
)
2675 && (rcompcode
!= COMPCODE_ORD
);
2676 bool trap
= (compcode
& COMPCODE_UNORD
) == 0
2677 && (compcode
!= COMPCODE_EQ
)
2678 && (compcode
!= COMPCODE_ORD
);
2680 /* In a short-circuited boolean expression the LHS might be
2681 such that the RHS, if evaluated, will never trap. For
2682 example, in ORD (x, y) && (x < y), we evaluate the RHS only
2683 if neither x nor y is NaN. (This is a mixed blessing: for
2684 example, the expression above will never trap, hence
2685 optimizing it to x < y would be invalid). */
2686 if ((code
== TRUTH_ORIF_EXPR
&& (lcompcode
& COMPCODE_UNORD
))
2687 || (code
== TRUTH_ANDIF_EXPR
&& !(lcompcode
& COMPCODE_UNORD
)))
2690 /* If the comparison was short-circuited, and only the RHS
2691 trapped, we may now generate a spurious trap. */
2693 && (code
== TRUTH_ANDIF_EXPR
|| code
== TRUTH_ORIF_EXPR
))
2696 /* If we changed the conditions that cause a trap, we lose. */
2697 if ((ltrap
|| rtrap
) != trap
)
2701 if (compcode
== COMPCODE_TRUE
)
2702 return constant_boolean_node (true, truth_type
);
2703 else if (compcode
== COMPCODE_FALSE
)
2704 return constant_boolean_node (false, truth_type
);
2707 enum tree_code tcode
;
2709 tcode
= compcode_to_comparison ((enum comparison_code
) compcode
);
2710 return fold_build2_loc (loc
, tcode
, truth_type
, ll_arg
, lr_arg
);
2714 /* Return nonzero if two operands (typically of the same tree node)
2715 are necessarily equal. FLAGS modifies behavior as follows:
2717 If OEP_ONLY_CONST is set, only return nonzero for constants.
2718 This function tests whether the operands are indistinguishable;
2719 it does not test whether they are equal using C's == operation.
2720 The distinction is important for IEEE floating point, because
2721 (1) -0.0 and 0.0 are distinguishable, but -0.0==0.0, and
2722 (2) two NaNs may be indistinguishable, but NaN!=NaN.
2724 If OEP_ONLY_CONST is unset, a VAR_DECL is considered equal to itself
2725 even though it may hold multiple values during a function.
2726 This is because a GCC tree node guarantees that nothing else is
2727 executed between the evaluation of its "operands" (which may often
2728 be evaluated in arbitrary order). Hence if the operands themselves
2729 don't side-effect, the VAR_DECLs, PARM_DECLs etc... must hold the
2730 same value in each operand/subexpression. Hence leaving OEP_ONLY_CONST
2731 unset means assuming isochronic (or instantaneous) tree equivalence.
2732 Unless comparing arbitrary expression trees, such as from different
2733 statements, this flag can usually be left unset.
2735 If OEP_PURE_SAME is set, then pure functions with identical arguments
2736 are considered the same. It is used when the caller has other ways
2737 to ensure that global memory is unchanged in between.
2739 If OEP_ADDRESS_OF is set, we are actually comparing addresses of objects,
2740 not values of expressions.
2742 Unless OEP_MATCH_SIDE_EFFECTS is set, the function returns false on
2743 any operand with side effect. This is unnecesarily conservative in the
2744 case we know that arg0 and arg1 are in disjoint code paths (such as in
2745 ?: operator). In addition OEP_MATCH_SIDE_EFFECTS is used when comparing
2746 addresses with TREE_CONSTANT flag set so we know that &var == &var
2747 even if var is volatile. */
2750 operand_equal_p (const_tree arg0
, const_tree arg1
, unsigned int flags
)
2752 /* If either is ERROR_MARK, they aren't equal. */
2753 if (TREE_CODE (arg0
) == ERROR_MARK
|| TREE_CODE (arg1
) == ERROR_MARK
2754 || TREE_TYPE (arg0
) == error_mark_node
2755 || TREE_TYPE (arg1
) == error_mark_node
)
2758 /* Similar, if either does not have a type (like a released SSA name),
2759 they aren't equal. */
2760 if (!TREE_TYPE (arg0
) || !TREE_TYPE (arg1
))
2763 /* We cannot consider pointers to different address space equal. */
2764 if (POINTER_TYPE_P (TREE_TYPE (arg0
))
2765 && POINTER_TYPE_P (TREE_TYPE (arg1
))
2766 && (TYPE_ADDR_SPACE (TREE_TYPE (TREE_TYPE (arg0
)))
2767 != TYPE_ADDR_SPACE (TREE_TYPE (TREE_TYPE (arg1
)))))
2770 /* Check equality of integer constants before bailing out due to
2771 precision differences. */
2772 if (TREE_CODE (arg0
) == INTEGER_CST
&& TREE_CODE (arg1
) == INTEGER_CST
)
2774 /* Address of INTEGER_CST is not defined; check that we did not forget
2775 to drop the OEP_ADDRESS_OF flags. */
2776 gcc_checking_assert (!(flags
& OEP_ADDRESS_OF
));
2777 return tree_int_cst_equal (arg0
, arg1
);
2780 if (!(flags
& OEP_ADDRESS_OF
))
2782 /* If both types don't have the same signedness, then we can't consider
2783 them equal. We must check this before the STRIP_NOPS calls
2784 because they may change the signedness of the arguments. As pointers
2785 strictly don't have a signedness, require either two pointers or
2786 two non-pointers as well. */
2787 if (TYPE_UNSIGNED (TREE_TYPE (arg0
)) != TYPE_UNSIGNED (TREE_TYPE (arg1
))
2788 || POINTER_TYPE_P (TREE_TYPE (arg0
))
2789 != POINTER_TYPE_P (TREE_TYPE (arg1
)))
2792 /* If both types don't have the same precision, then it is not safe
2794 if (element_precision (TREE_TYPE (arg0
))
2795 != element_precision (TREE_TYPE (arg1
)))
2802 /* FIXME: Fortran FE currently produce ADDR_EXPR of NOP_EXPR. Enable the
2803 sanity check once the issue is solved. */
2805 /* Addresses of conversions and SSA_NAMEs (and many other things)
2806 are not defined. Check that we did not forget to drop the
2807 OEP_ADDRESS_OF/OEP_CONSTANT_ADDRESS_OF flags. */
2808 gcc_checking_assert (!CONVERT_EXPR_P (arg0
) && !CONVERT_EXPR_P (arg1
)
2809 && TREE_CODE (arg0
) != SSA_NAME
);
2812 /* In case both args are comparisons but with different comparison
2813 code, try to swap the comparison operands of one arg to produce
2814 a match and compare that variant. */
2815 if (TREE_CODE (arg0
) != TREE_CODE (arg1
)
2816 && COMPARISON_CLASS_P (arg0
)
2817 && COMPARISON_CLASS_P (arg1
))
2819 enum tree_code swap_code
= swap_tree_comparison (TREE_CODE (arg1
));
2821 if (TREE_CODE (arg0
) == swap_code
)
2822 return operand_equal_p (TREE_OPERAND (arg0
, 0),
2823 TREE_OPERAND (arg1
, 1), flags
)
2824 && operand_equal_p (TREE_OPERAND (arg0
, 1),
2825 TREE_OPERAND (arg1
, 0), flags
);
2828 if (TREE_CODE (arg0
) != TREE_CODE (arg1
))
2830 /* NOP_EXPR and CONVERT_EXPR are considered equal. */
2831 if (CONVERT_EXPR_P (arg0
) && CONVERT_EXPR_P (arg1
))
2833 else if (flags
& OEP_ADDRESS_OF
)
2835 /* If we are interested in comparing addresses ignore
2836 MEM_REF wrappings of the base that can appear just for
2838 if (TREE_CODE (arg0
) == MEM_REF
2840 && TREE_CODE (TREE_OPERAND (arg0
, 0)) == ADDR_EXPR
2841 && TREE_OPERAND (TREE_OPERAND (arg0
, 0), 0) == arg1
2842 && integer_zerop (TREE_OPERAND (arg0
, 1)))
2844 else if (TREE_CODE (arg1
) == MEM_REF
2846 && TREE_CODE (TREE_OPERAND (arg1
, 0)) == ADDR_EXPR
2847 && TREE_OPERAND (TREE_OPERAND (arg1
, 0), 0) == arg0
2848 && integer_zerop (TREE_OPERAND (arg1
, 1)))
2856 /* When not checking adddresses, this is needed for conversions and for
2857 COMPONENT_REF. Might as well play it safe and always test this. */
2858 if (TREE_CODE (TREE_TYPE (arg0
)) == ERROR_MARK
2859 || TREE_CODE (TREE_TYPE (arg1
)) == ERROR_MARK
2860 || (TYPE_MODE (TREE_TYPE (arg0
)) != TYPE_MODE (TREE_TYPE (arg1
))
2861 && !(flags
& OEP_ADDRESS_OF
)))
2864 /* If ARG0 and ARG1 are the same SAVE_EXPR, they are necessarily equal.
2865 We don't care about side effects in that case because the SAVE_EXPR
2866 takes care of that for us. In all other cases, two expressions are
2867 equal if they have no side effects. If we have two identical
2868 expressions with side effects that should be treated the same due
2869 to the only side effects being identical SAVE_EXPR's, that will
2870 be detected in the recursive calls below.
2871 If we are taking an invariant address of two identical objects
2872 they are necessarily equal as well. */
2873 if (arg0
== arg1
&& ! (flags
& OEP_ONLY_CONST
)
2874 && (TREE_CODE (arg0
) == SAVE_EXPR
2875 || (flags
& OEP_MATCH_SIDE_EFFECTS
)
2876 || (! TREE_SIDE_EFFECTS (arg0
) && ! TREE_SIDE_EFFECTS (arg1
))))
2879 /* Next handle constant cases, those for which we can return 1 even
2880 if ONLY_CONST is set. */
2881 if (TREE_CONSTANT (arg0
) && TREE_CONSTANT (arg1
))
2882 switch (TREE_CODE (arg0
))
2885 return tree_int_cst_equal (arg0
, arg1
);
2888 return FIXED_VALUES_IDENTICAL (TREE_FIXED_CST (arg0
),
2889 TREE_FIXED_CST (arg1
));
2892 if (real_identical (&TREE_REAL_CST (arg0
), &TREE_REAL_CST (arg1
)))
2896 if (!HONOR_SIGNED_ZEROS (arg0
))
2898 /* If we do not distinguish between signed and unsigned zero,
2899 consider them equal. */
2900 if (real_zerop (arg0
) && real_zerop (arg1
))
2909 if (VECTOR_CST_NELTS (arg0
) != VECTOR_CST_NELTS (arg1
))
2912 for (i
= 0; i
< VECTOR_CST_NELTS (arg0
); ++i
)
2914 if (!operand_equal_p (VECTOR_CST_ELT (arg0
, i
),
2915 VECTOR_CST_ELT (arg1
, i
), flags
))
2922 return (operand_equal_p (TREE_REALPART (arg0
), TREE_REALPART (arg1
),
2924 && operand_equal_p (TREE_IMAGPART (arg0
), TREE_IMAGPART (arg1
),
2928 return (TREE_STRING_LENGTH (arg0
) == TREE_STRING_LENGTH (arg1
)
2929 && ! memcmp (TREE_STRING_POINTER (arg0
),
2930 TREE_STRING_POINTER (arg1
),
2931 TREE_STRING_LENGTH (arg0
)));
2934 gcc_checking_assert (!(flags
& OEP_ADDRESS_OF
));
2935 return operand_equal_p (TREE_OPERAND (arg0
, 0), TREE_OPERAND (arg1
, 0),
2936 flags
| OEP_ADDRESS_OF
2937 | OEP_MATCH_SIDE_EFFECTS
);
2939 /* In GIMPLE empty constructors are allowed in initializers of
2941 return (!vec_safe_length (CONSTRUCTOR_ELTS (arg0
))
2942 && !vec_safe_length (CONSTRUCTOR_ELTS (arg1
)));
2947 if (flags
& OEP_ONLY_CONST
)
2950 /* Define macros to test an operand from arg0 and arg1 for equality and a
2951 variant that allows null and views null as being different from any
2952 non-null value. In the latter case, if either is null, the both
2953 must be; otherwise, do the normal comparison. */
2954 #define OP_SAME(N) operand_equal_p (TREE_OPERAND (arg0, N), \
2955 TREE_OPERAND (arg1, N), flags)
2957 #define OP_SAME_WITH_NULL(N) \
2958 ((!TREE_OPERAND (arg0, N) || !TREE_OPERAND (arg1, N)) \
2959 ? TREE_OPERAND (arg0, N) == TREE_OPERAND (arg1, N) : OP_SAME (N))
2961 switch (TREE_CODE_CLASS (TREE_CODE (arg0
)))
2964 /* Two conversions are equal only if signedness and modes match. */
2965 switch (TREE_CODE (arg0
))
2968 case FIX_TRUNC_EXPR
:
2969 if (TYPE_UNSIGNED (TREE_TYPE (arg0
))
2970 != TYPE_UNSIGNED (TREE_TYPE (arg1
)))
2980 case tcc_comparison
:
2982 if (OP_SAME (0) && OP_SAME (1))
2985 /* For commutative ops, allow the other order. */
2986 return (commutative_tree_code (TREE_CODE (arg0
))
2987 && operand_equal_p (TREE_OPERAND (arg0
, 0),
2988 TREE_OPERAND (arg1
, 1), flags
)
2989 && operand_equal_p (TREE_OPERAND (arg0
, 1),
2990 TREE_OPERAND (arg1
, 0), flags
));
2993 /* If either of the pointer (or reference) expressions we are
2994 dereferencing contain a side effect, these cannot be equal,
2995 but their addresses can be. */
2996 if ((flags
& OEP_MATCH_SIDE_EFFECTS
) == 0
2997 && (TREE_SIDE_EFFECTS (arg0
)
2998 || TREE_SIDE_EFFECTS (arg1
)))
3001 switch (TREE_CODE (arg0
))
3004 if (!(flags
& OEP_ADDRESS_OF
)
3005 && (TYPE_ALIGN (TREE_TYPE (arg0
))
3006 != TYPE_ALIGN (TREE_TYPE (arg1
))))
3008 flags
&= ~OEP_ADDRESS_OF
;
3012 /* Require the same offset. */
3013 if (!operand_equal_p (TYPE_SIZE (TREE_TYPE (arg0
)),
3014 TYPE_SIZE (TREE_TYPE (arg1
)),
3015 flags
& ~OEP_ADDRESS_OF
))
3020 case VIEW_CONVERT_EXPR
:
3023 case TARGET_MEM_REF
:
3025 if (!(flags
& OEP_ADDRESS_OF
))
3027 /* Require equal access sizes */
3028 if (TYPE_SIZE (TREE_TYPE (arg0
)) != TYPE_SIZE (TREE_TYPE (arg1
))
3029 && (!TYPE_SIZE (TREE_TYPE (arg0
))
3030 || !TYPE_SIZE (TREE_TYPE (arg1
))
3031 || !operand_equal_p (TYPE_SIZE (TREE_TYPE (arg0
)),
3032 TYPE_SIZE (TREE_TYPE (arg1
)),
3035 /* Verify that access happens in similar types. */
3036 if (!types_compatible_p (TREE_TYPE (arg0
), TREE_TYPE (arg1
)))
3038 /* Verify that accesses are TBAA compatible. */
3039 if (!alias_ptr_types_compatible_p
3040 (TREE_TYPE (TREE_OPERAND (arg0
, 1)),
3041 TREE_TYPE (TREE_OPERAND (arg1
, 1)))
3042 || (MR_DEPENDENCE_CLIQUE (arg0
)
3043 != MR_DEPENDENCE_CLIQUE (arg1
))
3044 || (MR_DEPENDENCE_BASE (arg0
)
3045 != MR_DEPENDENCE_BASE (arg1
)))
3047 /* Verify that alignment is compatible. */
3048 if (TYPE_ALIGN (TREE_TYPE (arg0
))
3049 != TYPE_ALIGN (TREE_TYPE (arg1
)))
3052 flags
&= ~OEP_ADDRESS_OF
;
3053 return (OP_SAME (0) && OP_SAME (1)
3054 /* TARGET_MEM_REF require equal extra operands. */
3055 && (TREE_CODE (arg0
) != TARGET_MEM_REF
3056 || (OP_SAME_WITH_NULL (2)
3057 && OP_SAME_WITH_NULL (3)
3058 && OP_SAME_WITH_NULL (4))));
3061 case ARRAY_RANGE_REF
:
3064 flags
&= ~OEP_ADDRESS_OF
;
3065 /* Compare the array index by value if it is constant first as we
3066 may have different types but same value here. */
3067 return ((tree_int_cst_equal (TREE_OPERAND (arg0
, 1),
3068 TREE_OPERAND (arg1
, 1))
3070 && OP_SAME_WITH_NULL (2)
3071 && OP_SAME_WITH_NULL (3)
3072 /* Compare low bound and element size as with OEP_ADDRESS_OF
3073 we have to account for the offset of the ref. */
3074 && (TREE_TYPE (TREE_OPERAND (arg0
, 0))
3075 == TREE_TYPE (TREE_OPERAND (arg1
, 0))
3076 || (operand_equal_p (array_ref_low_bound
3077 (CONST_CAST_TREE (arg0
)),
3079 (CONST_CAST_TREE (arg1
)), flags
)
3080 && operand_equal_p (array_ref_element_size
3081 (CONST_CAST_TREE (arg0
)),
3082 array_ref_element_size
3083 (CONST_CAST_TREE (arg1
)),
3087 /* Handle operand 2 the same as for ARRAY_REF. Operand 0
3088 may be NULL when we're called to compare MEM_EXPRs. */
3089 if (!OP_SAME_WITH_NULL (0)
3092 flags
&= ~OEP_ADDRESS_OF
;
3093 return OP_SAME_WITH_NULL (2);
3098 flags
&= ~OEP_ADDRESS_OF
;
3099 return OP_SAME (1) && OP_SAME (2);
3105 case tcc_expression
:
3106 switch (TREE_CODE (arg0
))
3109 /* Be sure we pass right ADDRESS_OF flag. */
3110 gcc_checking_assert (!(flags
& OEP_ADDRESS_OF
));
3111 return operand_equal_p (TREE_OPERAND (arg0
, 0),
3112 TREE_OPERAND (arg1
, 0),
3113 flags
| OEP_ADDRESS_OF
);
3115 case TRUTH_NOT_EXPR
:
3118 case TRUTH_ANDIF_EXPR
:
3119 case TRUTH_ORIF_EXPR
:
3120 return OP_SAME (0) && OP_SAME (1);
3123 case WIDEN_MULT_PLUS_EXPR
:
3124 case WIDEN_MULT_MINUS_EXPR
:
3127 /* The multiplcation operands are commutative. */
3130 case TRUTH_AND_EXPR
:
3132 case TRUTH_XOR_EXPR
:
3133 if (OP_SAME (0) && OP_SAME (1))
3136 /* Otherwise take into account this is a commutative operation. */
3137 return (operand_equal_p (TREE_OPERAND (arg0
, 0),
3138 TREE_OPERAND (arg1
, 1), flags
)
3139 && operand_equal_p (TREE_OPERAND (arg0
, 1),
3140 TREE_OPERAND (arg1
, 0), flags
));
3143 if (! OP_SAME (1) || ! OP_SAME (2))
3145 flags
&= ~OEP_ADDRESS_OF
;
3150 return OP_SAME (0) && OP_SAME (1) && OP_SAME (2);
3157 switch (TREE_CODE (arg0
))
3160 if ((CALL_EXPR_FN (arg0
) == NULL_TREE
)
3161 != (CALL_EXPR_FN (arg1
) == NULL_TREE
))
3162 /* If not both CALL_EXPRs are either internal or normal function
3163 functions, then they are not equal. */
3165 else if (CALL_EXPR_FN (arg0
) == NULL_TREE
)
3167 /* If the CALL_EXPRs call different internal functions, then they
3169 if (CALL_EXPR_IFN (arg0
) != CALL_EXPR_IFN (arg1
))
3174 /* If the CALL_EXPRs call different functions, then they are not
3176 if (! operand_equal_p (CALL_EXPR_FN (arg0
), CALL_EXPR_FN (arg1
),
3181 /* FIXME: We could skip this test for OEP_MATCH_SIDE_EFFECTS. */
3183 unsigned int cef
= call_expr_flags (arg0
);
3184 if (flags
& OEP_PURE_SAME
)
3185 cef
&= ECF_CONST
| ECF_PURE
;
3192 /* Now see if all the arguments are the same. */
3194 const_call_expr_arg_iterator iter0
, iter1
;
3196 for (a0
= first_const_call_expr_arg (arg0
, &iter0
),
3197 a1
= first_const_call_expr_arg (arg1
, &iter1
);
3199 a0
= next_const_call_expr_arg (&iter0
),
3200 a1
= next_const_call_expr_arg (&iter1
))
3201 if (! operand_equal_p (a0
, a1
, flags
))
3204 /* If we get here and both argument lists are exhausted
3205 then the CALL_EXPRs are equal. */
3206 return ! (a0
|| a1
);
3212 case tcc_declaration
:
3213 /* Consider __builtin_sqrt equal to sqrt. */
3214 return (TREE_CODE (arg0
) == FUNCTION_DECL
3215 && DECL_BUILT_IN (arg0
) && DECL_BUILT_IN (arg1
)
3216 && DECL_BUILT_IN_CLASS (arg0
) == DECL_BUILT_IN_CLASS (arg1
)
3217 && DECL_FUNCTION_CODE (arg0
) == DECL_FUNCTION_CODE (arg1
));
3219 case tcc_exceptional
:
3220 if (TREE_CODE (arg0
) == CONSTRUCTOR
)
3222 /* In GIMPLE constructors are used only to build vectors from
3223 elements. Individual elements in the constructor must be
3224 indexed in increasing order and form an initial sequence.
3226 We make no effort to compare constructors in generic.
3227 (see sem_variable::equals in ipa-icf which can do so for
3229 if (!VECTOR_TYPE_P (TREE_TYPE (arg0
))
3230 || !VECTOR_TYPE_P (TREE_TYPE (arg1
)))
3233 /* Be sure that vectors constructed have the same representation.
3234 We only tested element precision and modes to match.
3235 Vectors may be BLKmode and thus also check that the number of
3237 if (TYPE_VECTOR_SUBPARTS (TREE_TYPE (arg0
))
3238 != TYPE_VECTOR_SUBPARTS (TREE_TYPE (arg1
)))
3241 vec
<constructor_elt
, va_gc
> *v0
= CONSTRUCTOR_ELTS (arg0
);
3242 vec
<constructor_elt
, va_gc
> *v1
= CONSTRUCTOR_ELTS (arg1
);
3243 unsigned int len
= vec_safe_length (v0
);
3245 if (len
!= vec_safe_length (v1
))
3248 for (unsigned int i
= 0; i
< len
; i
++)
3250 constructor_elt
*c0
= &(*v0
)[i
];
3251 constructor_elt
*c1
= &(*v1
)[i
];
3253 if (!operand_equal_p (c0
->value
, c1
->value
, flags
)
3254 /* In GIMPLE the indexes can be either NULL or matching i.
3255 Double check this so we won't get false
3256 positives for GENERIC. */
3258 && (TREE_CODE (c0
->index
) != INTEGER_CST
3259 || !compare_tree_int (c0
->index
, i
)))
3261 && (TREE_CODE (c1
->index
) != INTEGER_CST
3262 || !compare_tree_int (c1
->index
, i
))))
3274 #undef OP_SAME_WITH_NULL
3277 /* Similar to operand_equal_p, but see if ARG0 might have been made by
3278 shorten_compare from ARG1 when ARG1 was being compared with OTHER.
3280 When in doubt, return 0. */
3283 operand_equal_for_comparison_p (tree arg0
, tree arg1
, tree other
)
3285 int unsignedp1
, unsignedpo
;
3286 tree primarg0
, primarg1
, primother
;
3287 unsigned int correct_width
;
3289 if (operand_equal_p (arg0
, arg1
, 0))
3292 if (! INTEGRAL_TYPE_P (TREE_TYPE (arg0
))
3293 || ! INTEGRAL_TYPE_P (TREE_TYPE (arg1
)))
3296 /* Discard any conversions that don't change the modes of ARG0 and ARG1
3297 and see if the inner values are the same. This removes any
3298 signedness comparison, which doesn't matter here. */
3299 primarg0
= arg0
, primarg1
= arg1
;
3300 STRIP_NOPS (primarg0
);
3301 STRIP_NOPS (primarg1
);
3302 if (operand_equal_p (primarg0
, primarg1
, 0))
3305 /* Duplicate what shorten_compare does to ARG1 and see if that gives the
3306 actual comparison operand, ARG0.
3308 First throw away any conversions to wider types
3309 already present in the operands. */
3311 primarg1
= get_narrower (arg1
, &unsignedp1
);
3312 primother
= get_narrower (other
, &unsignedpo
);
3314 correct_width
= TYPE_PRECISION (TREE_TYPE (arg1
));
3315 if (unsignedp1
== unsignedpo
3316 && TYPE_PRECISION (TREE_TYPE (primarg1
)) < correct_width
3317 && TYPE_PRECISION (TREE_TYPE (primother
)) < correct_width
)
3319 tree type
= TREE_TYPE (arg0
);
3321 /* Make sure shorter operand is extended the right way
3322 to match the longer operand. */
3323 primarg1
= fold_convert (signed_or_unsigned_type_for
3324 (unsignedp1
, TREE_TYPE (primarg1
)), primarg1
);
3326 if (operand_equal_p (arg0
, fold_convert (type
, primarg1
), 0))
3333 /* See if ARG is an expression that is either a comparison or is performing
3334 arithmetic on comparisons. The comparisons must only be comparing
3335 two different values, which will be stored in *CVAL1 and *CVAL2; if
3336 they are nonzero it means that some operands have already been found.
3337 No variables may be used anywhere else in the expression except in the
3338 comparisons. If SAVE_P is true it means we removed a SAVE_EXPR around
3339 the expression and save_expr needs to be called with CVAL1 and CVAL2.
3341 If this is true, return 1. Otherwise, return zero. */
3344 twoval_comparison_p (tree arg
, tree
*cval1
, tree
*cval2
, int *save_p
)
3346 enum tree_code code
= TREE_CODE (arg
);
3347 enum tree_code_class tclass
= TREE_CODE_CLASS (code
);
3349 /* We can handle some of the tcc_expression cases here. */
3350 if (tclass
== tcc_expression
&& code
== TRUTH_NOT_EXPR
)
3352 else if (tclass
== tcc_expression
3353 && (code
== TRUTH_ANDIF_EXPR
|| code
== TRUTH_ORIF_EXPR
3354 || code
== COMPOUND_EXPR
))
3355 tclass
= tcc_binary
;
3357 else if (tclass
== tcc_expression
&& code
== SAVE_EXPR
3358 && ! TREE_SIDE_EFFECTS (TREE_OPERAND (arg
, 0)))
3360 /* If we've already found a CVAL1 or CVAL2, this expression is
3361 two complex to handle. */
3362 if (*cval1
|| *cval2
)
3372 return twoval_comparison_p (TREE_OPERAND (arg
, 0), cval1
, cval2
, save_p
);
3375 return (twoval_comparison_p (TREE_OPERAND (arg
, 0), cval1
, cval2
, save_p
)
3376 && twoval_comparison_p (TREE_OPERAND (arg
, 1),
3377 cval1
, cval2
, save_p
));
3382 case tcc_expression
:
3383 if (code
== COND_EXPR
)
3384 return (twoval_comparison_p (TREE_OPERAND (arg
, 0),
3385 cval1
, cval2
, save_p
)
3386 && twoval_comparison_p (TREE_OPERAND (arg
, 1),
3387 cval1
, cval2
, save_p
)
3388 && twoval_comparison_p (TREE_OPERAND (arg
, 2),
3389 cval1
, cval2
, save_p
));
3392 case tcc_comparison
:
3393 /* First see if we can handle the first operand, then the second. For
3394 the second operand, we know *CVAL1 can't be zero. It must be that
3395 one side of the comparison is each of the values; test for the
3396 case where this isn't true by failing if the two operands
3399 if (operand_equal_p (TREE_OPERAND (arg
, 0),
3400 TREE_OPERAND (arg
, 1), 0))
3404 *cval1
= TREE_OPERAND (arg
, 0);
3405 else if (operand_equal_p (*cval1
, TREE_OPERAND (arg
, 0), 0))
3407 else if (*cval2
== 0)
3408 *cval2
= TREE_OPERAND (arg
, 0);
3409 else if (operand_equal_p (*cval2
, TREE_OPERAND (arg
, 0), 0))
3414 if (operand_equal_p (*cval1
, TREE_OPERAND (arg
, 1), 0))
3416 else if (*cval2
== 0)
3417 *cval2
= TREE_OPERAND (arg
, 1);
3418 else if (operand_equal_p (*cval2
, TREE_OPERAND (arg
, 1), 0))
3430 /* ARG is a tree that is known to contain just arithmetic operations and
3431 comparisons. Evaluate the operations in the tree substituting NEW0 for
3432 any occurrence of OLD0 as an operand of a comparison and likewise for
3436 eval_subst (location_t loc
, tree arg
, tree old0
, tree new0
,
3437 tree old1
, tree new1
)
3439 tree type
= TREE_TYPE (arg
);
3440 enum tree_code code
= TREE_CODE (arg
);
3441 enum tree_code_class tclass
= TREE_CODE_CLASS (code
);
3443 /* We can handle some of the tcc_expression cases here. */
3444 if (tclass
== tcc_expression
&& code
== TRUTH_NOT_EXPR
)
3446 else if (tclass
== tcc_expression
3447 && (code
== TRUTH_ANDIF_EXPR
|| code
== TRUTH_ORIF_EXPR
))
3448 tclass
= tcc_binary
;
3453 return fold_build1_loc (loc
, code
, type
,
3454 eval_subst (loc
, TREE_OPERAND (arg
, 0),
3455 old0
, new0
, old1
, new1
));
3458 return fold_build2_loc (loc
, code
, type
,
3459 eval_subst (loc
, TREE_OPERAND (arg
, 0),
3460 old0
, new0
, old1
, new1
),
3461 eval_subst (loc
, TREE_OPERAND (arg
, 1),
3462 old0
, new0
, old1
, new1
));
3464 case tcc_expression
:
3468 return eval_subst (loc
, TREE_OPERAND (arg
, 0), old0
, new0
,
3472 return eval_subst (loc
, TREE_OPERAND (arg
, 1), old0
, new0
,
3476 return fold_build3_loc (loc
, code
, type
,
3477 eval_subst (loc
, TREE_OPERAND (arg
, 0),
3478 old0
, new0
, old1
, new1
),
3479 eval_subst (loc
, TREE_OPERAND (arg
, 1),
3480 old0
, new0
, old1
, new1
),
3481 eval_subst (loc
, TREE_OPERAND (arg
, 2),
3482 old0
, new0
, old1
, new1
));
3486 /* Fall through - ??? */
3488 case tcc_comparison
:
3490 tree arg0
= TREE_OPERAND (arg
, 0);
3491 tree arg1
= TREE_OPERAND (arg
, 1);
3493 /* We need to check both for exact equality and tree equality. The
3494 former will be true if the operand has a side-effect. In that
3495 case, we know the operand occurred exactly once. */
3497 if (arg0
== old0
|| operand_equal_p (arg0
, old0
, 0))
3499 else if (arg0
== old1
|| operand_equal_p (arg0
, old1
, 0))
3502 if (arg1
== old0
|| operand_equal_p (arg1
, old0
, 0))
3504 else if (arg1
== old1
|| operand_equal_p (arg1
, old1
, 0))
3507 return fold_build2_loc (loc
, code
, type
, arg0
, arg1
);
3515 /* Return a tree for the case when the result of an expression is RESULT
3516 converted to TYPE and OMITTED was previously an operand of the expression
3517 but is now not needed (e.g., we folded OMITTED * 0).
3519 If OMITTED has side effects, we must evaluate it. Otherwise, just do
3520 the conversion of RESULT to TYPE. */
3523 omit_one_operand_loc (location_t loc
, tree type
, tree result
, tree omitted
)
3525 tree t
= fold_convert_loc (loc
, type
, result
);
3527 /* If the resulting operand is an empty statement, just return the omitted
3528 statement casted to void. */
3529 if (IS_EMPTY_STMT (t
) && TREE_SIDE_EFFECTS (omitted
))
3530 return build1_loc (loc
, NOP_EXPR
, void_type_node
,
3531 fold_ignored_result (omitted
));
3533 if (TREE_SIDE_EFFECTS (omitted
))
3534 return build2_loc (loc
, COMPOUND_EXPR
, type
,
3535 fold_ignored_result (omitted
), t
);
3537 return non_lvalue_loc (loc
, t
);
3540 /* Return a tree for the case when the result of an expression is RESULT
3541 converted to TYPE and OMITTED1 and OMITTED2 were previously operands
3542 of the expression but are now not needed.
3544 If OMITTED1 or OMITTED2 has side effects, they must be evaluated.
3545 If both OMITTED1 and OMITTED2 have side effects, OMITTED1 is
3546 evaluated before OMITTED2. Otherwise, if neither has side effects,
3547 just do the conversion of RESULT to TYPE. */
3550 omit_two_operands_loc (location_t loc
, tree type
, tree result
,
3551 tree omitted1
, tree omitted2
)
3553 tree t
= fold_convert_loc (loc
, type
, result
);
3555 if (TREE_SIDE_EFFECTS (omitted2
))
3556 t
= build2_loc (loc
, COMPOUND_EXPR
, type
, omitted2
, t
);
3557 if (TREE_SIDE_EFFECTS (omitted1
))
3558 t
= build2_loc (loc
, COMPOUND_EXPR
, type
, omitted1
, t
);
3560 return TREE_CODE (t
) != COMPOUND_EXPR
? non_lvalue_loc (loc
, t
) : t
;
3564 /* Return a simplified tree node for the truth-negation of ARG. This
3565 never alters ARG itself. We assume that ARG is an operation that
3566 returns a truth value (0 or 1).
3568 FIXME: one would think we would fold the result, but it causes
3569 problems with the dominator optimizer. */
3572 fold_truth_not_expr (location_t loc
, tree arg
)
3574 tree type
= TREE_TYPE (arg
);
3575 enum tree_code code
= TREE_CODE (arg
);
3576 location_t loc1
, loc2
;
3578 /* If this is a comparison, we can simply invert it, except for
3579 floating-point non-equality comparisons, in which case we just
3580 enclose a TRUTH_NOT_EXPR around what we have. */
3582 if (TREE_CODE_CLASS (code
) == tcc_comparison
)
3584 tree op_type
= TREE_TYPE (TREE_OPERAND (arg
, 0));
3585 if (FLOAT_TYPE_P (op_type
)
3586 && flag_trapping_math
3587 && code
!= ORDERED_EXPR
&& code
!= UNORDERED_EXPR
3588 && code
!= NE_EXPR
&& code
!= EQ_EXPR
)
3591 code
= invert_tree_comparison (code
, HONOR_NANS (op_type
));
3592 if (code
== ERROR_MARK
)
3595 tree ret
= build2_loc (loc
, code
, type
, TREE_OPERAND (arg
, 0),
3596 TREE_OPERAND (arg
, 1));
3597 if (TREE_NO_WARNING (arg
))
3598 TREE_NO_WARNING (ret
) = 1;
3605 return constant_boolean_node (integer_zerop (arg
), type
);
3607 case TRUTH_AND_EXPR
:
3608 loc1
= expr_location_or (TREE_OPERAND (arg
, 0), loc
);
3609 loc2
= expr_location_or (TREE_OPERAND (arg
, 1), loc
);
3610 return build2_loc (loc
, TRUTH_OR_EXPR
, type
,
3611 invert_truthvalue_loc (loc1
, TREE_OPERAND (arg
, 0)),
3612 invert_truthvalue_loc (loc2
, TREE_OPERAND (arg
, 1)));
3615 loc1
= expr_location_or (TREE_OPERAND (arg
, 0), loc
);
3616 loc2
= expr_location_or (TREE_OPERAND (arg
, 1), loc
);
3617 return build2_loc (loc
, TRUTH_AND_EXPR
, type
,
3618 invert_truthvalue_loc (loc1
, TREE_OPERAND (arg
, 0)),
3619 invert_truthvalue_loc (loc2
, TREE_OPERAND (arg
, 1)));
3621 case TRUTH_XOR_EXPR
:
3622 /* Here we can invert either operand. We invert the first operand
3623 unless the second operand is a TRUTH_NOT_EXPR in which case our
3624 result is the XOR of the first operand with the inside of the
3625 negation of the second operand. */
3627 if (TREE_CODE (TREE_OPERAND (arg
, 1)) == TRUTH_NOT_EXPR
)
3628 return build2_loc (loc
, TRUTH_XOR_EXPR
, type
, TREE_OPERAND (arg
, 0),
3629 TREE_OPERAND (TREE_OPERAND (arg
, 1), 0));
3631 return build2_loc (loc
, TRUTH_XOR_EXPR
, type
,
3632 invert_truthvalue_loc (loc
, TREE_OPERAND (arg
, 0)),
3633 TREE_OPERAND (arg
, 1));
3635 case TRUTH_ANDIF_EXPR
:
3636 loc1
= expr_location_or (TREE_OPERAND (arg
, 0), loc
);
3637 loc2
= expr_location_or (TREE_OPERAND (arg
, 1), loc
);
3638 return build2_loc (loc
, TRUTH_ORIF_EXPR
, type
,
3639 invert_truthvalue_loc (loc1
, TREE_OPERAND (arg
, 0)),
3640 invert_truthvalue_loc (loc2
, TREE_OPERAND (arg
, 1)));
3642 case TRUTH_ORIF_EXPR
:
3643 loc1
= expr_location_or (TREE_OPERAND (arg
, 0), loc
);
3644 loc2
= expr_location_or (TREE_OPERAND (arg
, 1), loc
);
3645 return build2_loc (loc
, TRUTH_ANDIF_EXPR
, type
,
3646 invert_truthvalue_loc (loc1
, TREE_OPERAND (arg
, 0)),
3647 invert_truthvalue_loc (loc2
, TREE_OPERAND (arg
, 1)));
3649 case TRUTH_NOT_EXPR
:
3650 return TREE_OPERAND (arg
, 0);
3654 tree arg1
= TREE_OPERAND (arg
, 1);
3655 tree arg2
= TREE_OPERAND (arg
, 2);
3657 loc1
= expr_location_or (TREE_OPERAND (arg
, 1), loc
);
3658 loc2
= expr_location_or (TREE_OPERAND (arg
, 2), loc
);
3660 /* A COND_EXPR may have a throw as one operand, which
3661 then has void type. Just leave void operands
3663 return build3_loc (loc
, COND_EXPR
, type
, TREE_OPERAND (arg
, 0),
3664 VOID_TYPE_P (TREE_TYPE (arg1
))
3665 ? arg1
: invert_truthvalue_loc (loc1
, arg1
),
3666 VOID_TYPE_P (TREE_TYPE (arg2
))
3667 ? arg2
: invert_truthvalue_loc (loc2
, arg2
));
3671 loc1
= expr_location_or (TREE_OPERAND (arg
, 1), loc
);
3672 return build2_loc (loc
, COMPOUND_EXPR
, type
,
3673 TREE_OPERAND (arg
, 0),
3674 invert_truthvalue_loc (loc1
, TREE_OPERAND (arg
, 1)));
3676 case NON_LVALUE_EXPR
:
3677 loc1
= expr_location_or (TREE_OPERAND (arg
, 0), loc
);
3678 return invert_truthvalue_loc (loc1
, TREE_OPERAND (arg
, 0));
3681 if (TREE_CODE (TREE_TYPE (arg
)) == BOOLEAN_TYPE
)
3682 return build1_loc (loc
, TRUTH_NOT_EXPR
, type
, arg
);
3684 /* ... fall through ... */
3687 loc1
= expr_location_or (TREE_OPERAND (arg
, 0), loc
);
3688 return build1_loc (loc
, TREE_CODE (arg
), type
,
3689 invert_truthvalue_loc (loc1
, TREE_OPERAND (arg
, 0)));
3692 if (!integer_onep (TREE_OPERAND (arg
, 1)))
3694 return build2_loc (loc
, EQ_EXPR
, type
, arg
, build_int_cst (type
, 0));
3697 return build1_loc (loc
, TRUTH_NOT_EXPR
, type
, arg
);
3699 case CLEANUP_POINT_EXPR
:
3700 loc1
= expr_location_or (TREE_OPERAND (arg
, 0), loc
);
3701 return build1_loc (loc
, CLEANUP_POINT_EXPR
, type
,
3702 invert_truthvalue_loc (loc1
, TREE_OPERAND (arg
, 0)));
3709 /* Fold the truth-negation of ARG. This never alters ARG itself. We
3710 assume that ARG is an operation that returns a truth value (0 or 1
3711 for scalars, 0 or -1 for vectors). Return the folded expression if
3712 folding is successful. Otherwise, return NULL_TREE. */
3715 fold_invert_truthvalue (location_t loc
, tree arg
)
3717 tree type
= TREE_TYPE (arg
);
3718 return fold_unary_loc (loc
, VECTOR_TYPE_P (type
)
3724 /* Return a simplified tree node for the truth-negation of ARG. This
3725 never alters ARG itself. We assume that ARG is an operation that
3726 returns a truth value (0 or 1 for scalars, 0 or -1 for vectors). */
3729 invert_truthvalue_loc (location_t loc
, tree arg
)
3731 if (TREE_CODE (arg
) == ERROR_MARK
)
3734 tree type
= TREE_TYPE (arg
);
3735 return fold_build1_loc (loc
, VECTOR_TYPE_P (type
)
3741 /* Knowing that ARG0 and ARG1 are both RDIV_EXPRs, simplify a binary operation
3742 with code CODE. This optimization is unsafe. */
3744 distribute_real_division (location_t loc
, enum tree_code code
, tree type
,
3745 tree arg0
, tree arg1
)
3747 bool mul0
= TREE_CODE (arg0
) == MULT_EXPR
;
3748 bool mul1
= TREE_CODE (arg1
) == MULT_EXPR
;
3750 /* (A / C) +- (B / C) -> (A +- B) / C. */
3752 && operand_equal_p (TREE_OPERAND (arg0
, 1),
3753 TREE_OPERAND (arg1
, 1), 0))
3754 return fold_build2_loc (loc
, mul0
? MULT_EXPR
: RDIV_EXPR
, type
,
3755 fold_build2_loc (loc
, code
, type
,
3756 TREE_OPERAND (arg0
, 0),
3757 TREE_OPERAND (arg1
, 0)),
3758 TREE_OPERAND (arg0
, 1));
3760 /* (A / C1) +- (A / C2) -> A * (1 / C1 +- 1 / C2). */
3761 if (operand_equal_p (TREE_OPERAND (arg0
, 0),
3762 TREE_OPERAND (arg1
, 0), 0)
3763 && TREE_CODE (TREE_OPERAND (arg0
, 1)) == REAL_CST
3764 && TREE_CODE (TREE_OPERAND (arg1
, 1)) == REAL_CST
)
3766 REAL_VALUE_TYPE r0
, r1
;
3767 r0
= TREE_REAL_CST (TREE_OPERAND (arg0
, 1));
3768 r1
= TREE_REAL_CST (TREE_OPERAND (arg1
, 1));
3770 real_arithmetic (&r0
, RDIV_EXPR
, &dconst1
, &r0
);
3772 real_arithmetic (&r1
, RDIV_EXPR
, &dconst1
, &r1
);
3773 real_arithmetic (&r0
, code
, &r0
, &r1
);
3774 return fold_build2_loc (loc
, MULT_EXPR
, type
,
3775 TREE_OPERAND (arg0
, 0),
3776 build_real (type
, r0
));
3782 /* Return a BIT_FIELD_REF of type TYPE to refer to BITSIZE bits of INNER
3783 starting at BITPOS. The field is unsigned if UNSIGNEDP is nonzero
3784 and uses reverse storage order if REVERSEP is nonzero. */
3787 make_bit_field_ref (location_t loc
, tree inner
, tree type
,
3788 HOST_WIDE_INT bitsize
, HOST_WIDE_INT bitpos
,
3789 int unsignedp
, int reversep
)
3791 tree result
, bftype
;
3793 if (bitpos
== 0 && !reversep
)
3795 tree size
= TYPE_SIZE (TREE_TYPE (inner
));
3796 if ((INTEGRAL_TYPE_P (TREE_TYPE (inner
))
3797 || POINTER_TYPE_P (TREE_TYPE (inner
)))
3798 && tree_fits_shwi_p (size
)
3799 && tree_to_shwi (size
) == bitsize
)
3800 return fold_convert_loc (loc
, type
, inner
);
3804 if (TYPE_PRECISION (bftype
) != bitsize
3805 || TYPE_UNSIGNED (bftype
) == !unsignedp
)
3806 bftype
= build_nonstandard_integer_type (bitsize
, 0);
3808 result
= build3_loc (loc
, BIT_FIELD_REF
, bftype
, inner
,
3809 size_int (bitsize
), bitsize_int (bitpos
));
3810 REF_REVERSE_STORAGE_ORDER (result
) = reversep
;
3813 result
= fold_convert_loc (loc
, type
, result
);
3818 /* Optimize a bit-field compare.
3820 There are two cases: First is a compare against a constant and the
3821 second is a comparison of two items where the fields are at the same
3822 bit position relative to the start of a chunk (byte, halfword, word)
3823 large enough to contain it. In these cases we can avoid the shift
3824 implicit in bitfield extractions.
3826 For constants, we emit a compare of the shifted constant with the
3827 BIT_AND_EXPR of a mask and a byte, halfword, or word of the operand being
3828 compared. For two fields at the same position, we do the ANDs with the
3829 similar mask and compare the result of the ANDs.
3831 CODE is the comparison code, known to be either NE_EXPR or EQ_EXPR.
3832 COMPARE_TYPE is the type of the comparison, and LHS and RHS
3833 are the left and right operands of the comparison, respectively.
3835 If the optimization described above can be done, we return the resulting
3836 tree. Otherwise we return zero. */
3839 optimize_bit_field_compare (location_t loc
, enum tree_code code
,
3840 tree compare_type
, tree lhs
, tree rhs
)
3842 HOST_WIDE_INT lbitpos
, lbitsize
, rbitpos
, rbitsize
, nbitpos
, nbitsize
;
3843 tree type
= TREE_TYPE (lhs
);
3845 int const_p
= TREE_CODE (rhs
) == INTEGER_CST
;
3846 machine_mode lmode
, rmode
, nmode
;
3847 int lunsignedp
, runsignedp
;
3848 int lreversep
, rreversep
;
3849 int lvolatilep
= 0, rvolatilep
= 0;
3850 tree linner
, rinner
= NULL_TREE
;
3854 /* Get all the information about the extractions being done. If the bit size
3855 if the same as the size of the underlying object, we aren't doing an
3856 extraction at all and so can do nothing. We also don't want to
3857 do anything if the inner expression is a PLACEHOLDER_EXPR since we
3858 then will no longer be able to replace it. */
3859 linner
= get_inner_reference (lhs
, &lbitsize
, &lbitpos
, &offset
, &lmode
,
3860 &lunsignedp
, &lreversep
, &lvolatilep
, false);
3861 if (linner
== lhs
|| lbitsize
== GET_MODE_BITSIZE (lmode
) || lbitsize
< 0
3862 || offset
!= 0 || TREE_CODE (linner
) == PLACEHOLDER_EXPR
|| lvolatilep
)
3866 rreversep
= lreversep
;
3869 /* If this is not a constant, we can only do something if bit positions,
3870 sizes, signedness and storage order are the same. */
3872 = get_inner_reference (rhs
, &rbitsize
, &rbitpos
, &offset
, &rmode
,
3873 &runsignedp
, &rreversep
, &rvolatilep
, false);
3875 if (rinner
== rhs
|| lbitpos
!= rbitpos
|| lbitsize
!= rbitsize
3876 || lunsignedp
!= runsignedp
|| lreversep
!= rreversep
|| offset
!= 0
3877 || TREE_CODE (rinner
) == PLACEHOLDER_EXPR
|| rvolatilep
)
3881 /* See if we can find a mode to refer to this field. We should be able to,
3882 but fail if we can't. */
3883 nmode
= get_best_mode (lbitsize
, lbitpos
, 0, 0,
3884 const_p
? TYPE_ALIGN (TREE_TYPE (linner
))
3885 : MIN (TYPE_ALIGN (TREE_TYPE (linner
)),
3886 TYPE_ALIGN (TREE_TYPE (rinner
))),
3888 if (nmode
== VOIDmode
)
3891 /* Set signed and unsigned types of the precision of this mode for the
3893 unsigned_type
= lang_hooks
.types
.type_for_mode (nmode
, 1);
3895 /* Compute the bit position and size for the new reference and our offset
3896 within it. If the new reference is the same size as the original, we
3897 won't optimize anything, so return zero. */
3898 nbitsize
= GET_MODE_BITSIZE (nmode
);
3899 nbitpos
= lbitpos
& ~ (nbitsize
- 1);
3901 if (nbitsize
== lbitsize
)
3904 if (lreversep
? !BYTES_BIG_ENDIAN
: BYTES_BIG_ENDIAN
)
3905 lbitpos
= nbitsize
- lbitsize
- lbitpos
;
3907 /* Make the mask to be used against the extracted field. */
3908 mask
= build_int_cst_type (unsigned_type
, -1);
3909 mask
= const_binop (LSHIFT_EXPR
, mask
, size_int (nbitsize
- lbitsize
));
3910 mask
= const_binop (RSHIFT_EXPR
, mask
,
3911 size_int (nbitsize
- lbitsize
- lbitpos
));
3914 /* If not comparing with constant, just rework the comparison
3916 return fold_build2_loc (loc
, code
, compare_type
,
3917 fold_build2_loc (loc
, BIT_AND_EXPR
, unsigned_type
,
3918 make_bit_field_ref (loc
, linner
,
3923 fold_build2_loc (loc
, BIT_AND_EXPR
, unsigned_type
,
3924 make_bit_field_ref (loc
, rinner
,
3930 /* Otherwise, we are handling the constant case. See if the constant is too
3931 big for the field. Warn and return a tree for 0 (false) if so. We do
3932 this not only for its own sake, but to avoid having to test for this
3933 error case below. If we didn't, we might generate wrong code.
3935 For unsigned fields, the constant shifted right by the field length should
3936 be all zero. For signed fields, the high-order bits should agree with
3941 if (wi::lrshift (rhs
, lbitsize
) != 0)
3943 warning (0, "comparison is always %d due to width of bit-field",
3945 return constant_boolean_node (code
== NE_EXPR
, compare_type
);
3950 wide_int tem
= wi::arshift (rhs
, lbitsize
- 1);
3951 if (tem
!= 0 && tem
!= -1)
3953 warning (0, "comparison is always %d due to width of bit-field",
3955 return constant_boolean_node (code
== NE_EXPR
, compare_type
);
3959 /* Single-bit compares should always be against zero. */
3960 if (lbitsize
== 1 && ! integer_zerop (rhs
))
3962 code
= code
== EQ_EXPR
? NE_EXPR
: EQ_EXPR
;
3963 rhs
= build_int_cst (type
, 0);
3966 /* Make a new bitfield reference, shift the constant over the
3967 appropriate number of bits and mask it with the computed mask
3968 (in case this was a signed field). If we changed it, make a new one. */
3969 lhs
= make_bit_field_ref (loc
, linner
, unsigned_type
, nbitsize
, nbitpos
, 1,
3972 rhs
= const_binop (BIT_AND_EXPR
,
3973 const_binop (LSHIFT_EXPR
,
3974 fold_convert_loc (loc
, unsigned_type
, rhs
),
3975 size_int (lbitpos
)),
3978 lhs
= build2_loc (loc
, code
, compare_type
,
3979 build2 (BIT_AND_EXPR
, unsigned_type
, lhs
, mask
), rhs
);
3983 /* Subroutine for fold_truth_andor_1: decode a field reference.
3985 If EXP is a comparison reference, we return the innermost reference.
3987 *PBITSIZE is set to the number of bits in the reference, *PBITPOS is
3988 set to the starting bit number.
3990 If the innermost field can be completely contained in a mode-sized
3991 unit, *PMODE is set to that mode. Otherwise, it is set to VOIDmode.
3993 *PVOLATILEP is set to 1 if the any expression encountered is volatile;
3994 otherwise it is not changed.
3996 *PUNSIGNEDP is set to the signedness of the field.
3998 *PREVERSEP is set to the storage order of the field.
4000 *PMASK is set to the mask used. This is either contained in a
4001 BIT_AND_EXPR or derived from the width of the field.
4003 *PAND_MASK is set to the mask found in a BIT_AND_EXPR, if any.
4005 Return 0 if this is not a component reference or is one that we can't
4006 do anything with. */
4009 decode_field_reference (location_t loc
, tree exp
, HOST_WIDE_INT
*pbitsize
,
4010 HOST_WIDE_INT
*pbitpos
, machine_mode
*pmode
,
4011 int *punsignedp
, int *preversep
, int *pvolatilep
,
4012 tree
*pmask
, tree
*pand_mask
)
4014 tree outer_type
= 0;
4016 tree mask
, inner
, offset
;
4018 unsigned int precision
;
4020 /* All the optimizations using this function assume integer fields.
4021 There are problems with FP fields since the type_for_size call
4022 below can fail for, e.g., XFmode. */
4023 if (! INTEGRAL_TYPE_P (TREE_TYPE (exp
)))
4026 /* We are interested in the bare arrangement of bits, so strip everything
4027 that doesn't affect the machine mode. However, record the type of the
4028 outermost expression if it may matter below. */
4029 if (CONVERT_EXPR_P (exp
)
4030 || TREE_CODE (exp
) == NON_LVALUE_EXPR
)
4031 outer_type
= TREE_TYPE (exp
);
4034 if (TREE_CODE (exp
) == BIT_AND_EXPR
)
4036 and_mask
= TREE_OPERAND (exp
, 1);
4037 exp
= TREE_OPERAND (exp
, 0);
4038 STRIP_NOPS (exp
); STRIP_NOPS (and_mask
);
4039 if (TREE_CODE (and_mask
) != INTEGER_CST
)
4043 inner
= get_inner_reference (exp
, pbitsize
, pbitpos
, &offset
, pmode
,
4044 punsignedp
, preversep
, pvolatilep
, false);
4045 if ((inner
== exp
&& and_mask
== 0)
4046 || *pbitsize
< 0 || offset
!= 0
4047 || TREE_CODE (inner
) == PLACEHOLDER_EXPR
)
4050 /* If the number of bits in the reference is the same as the bitsize of
4051 the outer type, then the outer type gives the signedness. Otherwise
4052 (in case of a small bitfield) the signedness is unchanged. */
4053 if (outer_type
&& *pbitsize
== TYPE_PRECISION (outer_type
))
4054 *punsignedp
= TYPE_UNSIGNED (outer_type
);
4056 /* Compute the mask to access the bitfield. */
4057 unsigned_type
= lang_hooks
.types
.type_for_size (*pbitsize
, 1);
4058 precision
= TYPE_PRECISION (unsigned_type
);
4060 mask
= build_int_cst_type (unsigned_type
, -1);
4062 mask
= const_binop (LSHIFT_EXPR
, mask
, size_int (precision
- *pbitsize
));
4063 mask
= const_binop (RSHIFT_EXPR
, mask
, size_int (precision
- *pbitsize
));
4065 /* Merge it with the mask we found in the BIT_AND_EXPR, if any. */
4067 mask
= fold_build2_loc (loc
, BIT_AND_EXPR
, unsigned_type
,
4068 fold_convert_loc (loc
, unsigned_type
, and_mask
), mask
);
4071 *pand_mask
= and_mask
;
4075 /* Return nonzero if MASK represents a mask of SIZE ones in the low-order
4076 bit positions and MASK is SIGNED. */
4079 all_ones_mask_p (const_tree mask
, unsigned int size
)
4081 tree type
= TREE_TYPE (mask
);
4082 unsigned int precision
= TYPE_PRECISION (type
);
4084 /* If this function returns true when the type of the mask is
4085 UNSIGNED, then there will be errors. In particular see
4086 gcc.c-torture/execute/990326-1.c. There does not appear to be
4087 any documentation paper trail as to why this is so. But the pre
4088 wide-int worked with that restriction and it has been preserved
4090 if (size
> precision
|| TYPE_SIGN (type
) == UNSIGNED
)
4093 return wi::mask (size
, false, precision
) == mask
;
4096 /* Subroutine for fold: determine if VAL is the INTEGER_CONST that
4097 represents the sign bit of EXP's type. If EXP represents a sign
4098 or zero extension, also test VAL against the unextended type.
4099 The return value is the (sub)expression whose sign bit is VAL,
4100 or NULL_TREE otherwise. */
4103 sign_bit_p (tree exp
, const_tree val
)
4108 /* Tree EXP must have an integral type. */
4109 t
= TREE_TYPE (exp
);
4110 if (! INTEGRAL_TYPE_P (t
))
4113 /* Tree VAL must be an integer constant. */
4114 if (TREE_CODE (val
) != INTEGER_CST
4115 || TREE_OVERFLOW (val
))
4118 width
= TYPE_PRECISION (t
);
4119 if (wi::only_sign_bit_p (val
, width
))
4122 /* Handle extension from a narrower type. */
4123 if (TREE_CODE (exp
) == NOP_EXPR
4124 && TYPE_PRECISION (TREE_TYPE (TREE_OPERAND (exp
, 0))) < width
)
4125 return sign_bit_p (TREE_OPERAND (exp
, 0), val
);
4130 /* Subroutine for fold_truth_andor_1: determine if an operand is simple enough
4131 to be evaluated unconditionally. */
4134 simple_operand_p (const_tree exp
)
4136 /* Strip any conversions that don't change the machine mode. */
4139 return (CONSTANT_CLASS_P (exp
)
4140 || TREE_CODE (exp
) == SSA_NAME
4142 && ! TREE_ADDRESSABLE (exp
)
4143 && ! TREE_THIS_VOLATILE (exp
)
4144 && ! DECL_NONLOCAL (exp
)
4145 /* Don't regard global variables as simple. They may be
4146 allocated in ways unknown to the compiler (shared memory,
4147 #pragma weak, etc). */
4148 && ! TREE_PUBLIC (exp
)
4149 && ! DECL_EXTERNAL (exp
)
4150 /* Weakrefs are not safe to be read, since they can be NULL.
4151 They are !TREE_PUBLIC && !DECL_EXTERNAL but still
4152 have DECL_WEAK flag set. */
4153 && (! VAR_OR_FUNCTION_DECL_P (exp
) || ! DECL_WEAK (exp
))
4154 /* Loading a static variable is unduly expensive, but global
4155 registers aren't expensive. */
4156 && (! TREE_STATIC (exp
) || DECL_REGISTER (exp
))));
4159 /* Subroutine for fold_truth_andor: determine if an operand is simple enough
4160 to be evaluated unconditionally.
4161 I addition to simple_operand_p, we assume that comparisons, conversions,
4162 and logic-not operations are simple, if their operands are simple, too. */
4165 simple_operand_p_2 (tree exp
)
4167 enum tree_code code
;
4169 if (TREE_SIDE_EFFECTS (exp
)
4170 || tree_could_trap_p (exp
))
4173 while (CONVERT_EXPR_P (exp
))
4174 exp
= TREE_OPERAND (exp
, 0);
4176 code
= TREE_CODE (exp
);
4178 if (TREE_CODE_CLASS (code
) == tcc_comparison
)
4179 return (simple_operand_p (TREE_OPERAND (exp
, 0))
4180 && simple_operand_p (TREE_OPERAND (exp
, 1)));
4182 if (code
== TRUTH_NOT_EXPR
)
4183 return simple_operand_p_2 (TREE_OPERAND (exp
, 0));
4185 return simple_operand_p (exp
);
4189 /* The following functions are subroutines to fold_range_test and allow it to
4190 try to change a logical combination of comparisons into a range test.
4193 X == 2 || X == 3 || X == 4 || X == 5
4197 (unsigned) (X - 2) <= 3
4199 We describe each set of comparisons as being either inside or outside
4200 a range, using a variable named like IN_P, and then describe the
4201 range with a lower and upper bound. If one of the bounds is omitted,
4202 it represents either the highest or lowest value of the type.
4204 In the comments below, we represent a range by two numbers in brackets
4205 preceded by a "+" to designate being inside that range, or a "-" to
4206 designate being outside that range, so the condition can be inverted by
4207 flipping the prefix. An omitted bound is represented by a "-". For
4208 example, "- [-, 10]" means being outside the range starting at the lowest
4209 possible value and ending at 10, in other words, being greater than 10.
4210 The range "+ [-, -]" is always true and hence the range "- [-, -]" is
4213 We set up things so that the missing bounds are handled in a consistent
4214 manner so neither a missing bound nor "true" and "false" need to be
4215 handled using a special case. */
4217 /* Return the result of applying CODE to ARG0 and ARG1, but handle the case
4218 of ARG0 and/or ARG1 being omitted, meaning an unlimited range. UPPER0_P
4219 and UPPER1_P are nonzero if the respective argument is an upper bound
4220 and zero for a lower. TYPE, if nonzero, is the type of the result; it
4221 must be specified for a comparison. ARG1 will be converted to ARG0's
4222 type if both are specified. */
4225 range_binop (enum tree_code code
, tree type
, tree arg0
, int upper0_p
,
4226 tree arg1
, int upper1_p
)
4232 /* If neither arg represents infinity, do the normal operation.
4233 Else, if not a comparison, return infinity. Else handle the special
4234 comparison rules. Note that most of the cases below won't occur, but
4235 are handled for consistency. */
4237 if (arg0
!= 0 && arg1
!= 0)
4239 tem
= fold_build2 (code
, type
!= 0 ? type
: TREE_TYPE (arg0
),
4240 arg0
, fold_convert (TREE_TYPE (arg0
), arg1
));
4242 return TREE_CODE (tem
) == INTEGER_CST
? tem
: 0;
4245 if (TREE_CODE_CLASS (code
) != tcc_comparison
)
4248 /* Set SGN[01] to -1 if ARG[01] is a lower bound, 1 for upper, and 0
4249 for neither. In real maths, we cannot assume open ended ranges are
4250 the same. But, this is computer arithmetic, where numbers are finite.
4251 We can therefore make the transformation of any unbounded range with
4252 the value Z, Z being greater than any representable number. This permits
4253 us to treat unbounded ranges as equal. */
4254 sgn0
= arg0
!= 0 ? 0 : (upper0_p
? 1 : -1);
4255 sgn1
= arg1
!= 0 ? 0 : (upper1_p
? 1 : -1);
4259 result
= sgn0
== sgn1
;
4262 result
= sgn0
!= sgn1
;
4265 result
= sgn0
< sgn1
;
4268 result
= sgn0
<= sgn1
;
4271 result
= sgn0
> sgn1
;
4274 result
= sgn0
>= sgn1
;
4280 return constant_boolean_node (result
, type
);
4283 /* Helper routine for make_range. Perform one step for it, return
4284 new expression if the loop should continue or NULL_TREE if it should
4288 make_range_step (location_t loc
, enum tree_code code
, tree arg0
, tree arg1
,
4289 tree exp_type
, tree
*p_low
, tree
*p_high
, int *p_in_p
,
4290 bool *strict_overflow_p
)
4292 tree arg0_type
= TREE_TYPE (arg0
);
4293 tree n_low
, n_high
, low
= *p_low
, high
= *p_high
;
4294 int in_p
= *p_in_p
, n_in_p
;
4298 case TRUTH_NOT_EXPR
:
4299 /* We can only do something if the range is testing for zero. */
4300 if (low
== NULL_TREE
|| high
== NULL_TREE
4301 || ! integer_zerop (low
) || ! integer_zerop (high
))
4306 case EQ_EXPR
: case NE_EXPR
:
4307 case LT_EXPR
: case LE_EXPR
: case GE_EXPR
: case GT_EXPR
:
4308 /* We can only do something if the range is testing for zero
4309 and if the second operand is an integer constant. Note that
4310 saying something is "in" the range we make is done by
4311 complementing IN_P since it will set in the initial case of
4312 being not equal to zero; "out" is leaving it alone. */
4313 if (low
== NULL_TREE
|| high
== NULL_TREE
4314 || ! integer_zerop (low
) || ! integer_zerop (high
)
4315 || TREE_CODE (arg1
) != INTEGER_CST
)
4320 case NE_EXPR
: /* - [c, c] */
4323 case EQ_EXPR
: /* + [c, c] */
4324 in_p
= ! in_p
, low
= high
= arg1
;
4326 case GT_EXPR
: /* - [-, c] */
4327 low
= 0, high
= arg1
;
4329 case GE_EXPR
: /* + [c, -] */
4330 in_p
= ! in_p
, low
= arg1
, high
= 0;
4332 case LT_EXPR
: /* - [c, -] */
4333 low
= arg1
, high
= 0;
4335 case LE_EXPR
: /* + [-, c] */
4336 in_p
= ! in_p
, low
= 0, high
= arg1
;
4342 /* If this is an unsigned comparison, we also know that EXP is
4343 greater than or equal to zero. We base the range tests we make
4344 on that fact, so we record it here so we can parse existing
4345 range tests. We test arg0_type since often the return type
4346 of, e.g. EQ_EXPR, is boolean. */
4347 if (TYPE_UNSIGNED (arg0_type
) && (low
== 0 || high
== 0))
4349 if (! merge_ranges (&n_in_p
, &n_low
, &n_high
,
4351 build_int_cst (arg0_type
, 0),
4355 in_p
= n_in_p
, low
= n_low
, high
= n_high
;
4357 /* If the high bound is missing, but we have a nonzero low
4358 bound, reverse the range so it goes from zero to the low bound
4360 if (high
== 0 && low
&& ! integer_zerop (low
))
4363 high
= range_binop (MINUS_EXPR
, NULL_TREE
, low
, 0,
4364 build_int_cst (TREE_TYPE (low
), 1), 0);
4365 low
= build_int_cst (arg0_type
, 0);
4375 /* If flag_wrapv and ARG0_TYPE is signed, make sure
4376 low and high are non-NULL, then normalize will DTRT. */
4377 if (!TYPE_UNSIGNED (arg0_type
)
4378 && !TYPE_OVERFLOW_UNDEFINED (arg0_type
))
4380 if (low
== NULL_TREE
)
4381 low
= TYPE_MIN_VALUE (arg0_type
);
4382 if (high
== NULL_TREE
)
4383 high
= TYPE_MAX_VALUE (arg0_type
);
4386 /* (-x) IN [a,b] -> x in [-b, -a] */
4387 n_low
= range_binop (MINUS_EXPR
, exp_type
,
4388 build_int_cst (exp_type
, 0),
4390 n_high
= range_binop (MINUS_EXPR
, exp_type
,
4391 build_int_cst (exp_type
, 0),
4393 if (n_high
!= 0 && TREE_OVERFLOW (n_high
))
4399 return build2_loc (loc
, MINUS_EXPR
, exp_type
, negate_expr (arg0
),
4400 build_int_cst (exp_type
, 1));
4404 if (TREE_CODE (arg1
) != INTEGER_CST
)
4407 /* If flag_wrapv and ARG0_TYPE is signed, then we cannot
4408 move a constant to the other side. */
4409 if (!TYPE_UNSIGNED (arg0_type
)
4410 && !TYPE_OVERFLOW_UNDEFINED (arg0_type
))
4413 /* If EXP is signed, any overflow in the computation is undefined,
4414 so we don't worry about it so long as our computations on
4415 the bounds don't overflow. For unsigned, overflow is defined
4416 and this is exactly the right thing. */
4417 n_low
= range_binop (code
== MINUS_EXPR
? PLUS_EXPR
: MINUS_EXPR
,
4418 arg0_type
, low
, 0, arg1
, 0);
4419 n_high
= range_binop (code
== MINUS_EXPR
? PLUS_EXPR
: MINUS_EXPR
,
4420 arg0_type
, high
, 1, arg1
, 0);
4421 if ((n_low
!= 0 && TREE_OVERFLOW (n_low
))
4422 || (n_high
!= 0 && TREE_OVERFLOW (n_high
)))
4425 if (TYPE_OVERFLOW_UNDEFINED (arg0_type
))
4426 *strict_overflow_p
= true;
4429 /* Check for an unsigned range which has wrapped around the maximum
4430 value thus making n_high < n_low, and normalize it. */
4431 if (n_low
&& n_high
&& tree_int_cst_lt (n_high
, n_low
))
4433 low
= range_binop (PLUS_EXPR
, arg0_type
, n_high
, 0,
4434 build_int_cst (TREE_TYPE (n_high
), 1), 0);
4435 high
= range_binop (MINUS_EXPR
, arg0_type
, n_low
, 0,
4436 build_int_cst (TREE_TYPE (n_low
), 1), 0);
4438 /* If the range is of the form +/- [ x+1, x ], we won't
4439 be able to normalize it. But then, it represents the
4440 whole range or the empty set, so make it
4442 if (tree_int_cst_equal (n_low
, low
)
4443 && tree_int_cst_equal (n_high
, high
))
4449 low
= n_low
, high
= n_high
;
4457 case NON_LVALUE_EXPR
:
4458 if (TYPE_PRECISION (arg0_type
) > TYPE_PRECISION (exp_type
))
4461 if (! INTEGRAL_TYPE_P (arg0_type
)
4462 || (low
!= 0 && ! int_fits_type_p (low
, arg0_type
))
4463 || (high
!= 0 && ! int_fits_type_p (high
, arg0_type
)))
4466 n_low
= low
, n_high
= high
;
4469 n_low
= fold_convert_loc (loc
, arg0_type
, n_low
);
4472 n_high
= fold_convert_loc (loc
, arg0_type
, n_high
);
4474 /* If we're converting arg0 from an unsigned type, to exp,
4475 a signed type, we will be doing the comparison as unsigned.
4476 The tests above have already verified that LOW and HIGH
4479 So we have to ensure that we will handle large unsigned
4480 values the same way that the current signed bounds treat
4483 if (!TYPE_UNSIGNED (exp_type
) && TYPE_UNSIGNED (arg0_type
))
4487 /* For fixed-point modes, we need to pass the saturating flag
4488 as the 2nd parameter. */
4489 if (ALL_FIXED_POINT_MODE_P (TYPE_MODE (arg0_type
)))
4491 = lang_hooks
.types
.type_for_mode (TYPE_MODE (arg0_type
),
4492 TYPE_SATURATING (arg0_type
));
4495 = lang_hooks
.types
.type_for_mode (TYPE_MODE (arg0_type
), 1);
4497 /* A range without an upper bound is, naturally, unbounded.
4498 Since convert would have cropped a very large value, use
4499 the max value for the destination type. */
4501 = TYPE_MAX_VALUE (equiv_type
) ? TYPE_MAX_VALUE (equiv_type
)
4502 : TYPE_MAX_VALUE (arg0_type
);
4504 if (TYPE_PRECISION (exp_type
) == TYPE_PRECISION (arg0_type
))
4505 high_positive
= fold_build2_loc (loc
, RSHIFT_EXPR
, arg0_type
,
4506 fold_convert_loc (loc
, arg0_type
,
4508 build_int_cst (arg0_type
, 1));
4510 /* If the low bound is specified, "and" the range with the
4511 range for which the original unsigned value will be
4515 if (! merge_ranges (&n_in_p
, &n_low
, &n_high
, 1, n_low
, n_high
,
4516 1, fold_convert_loc (loc
, arg0_type
,
4521 in_p
= (n_in_p
== in_p
);
4525 /* Otherwise, "or" the range with the range of the input
4526 that will be interpreted as negative. */
4527 if (! merge_ranges (&n_in_p
, &n_low
, &n_high
, 0, n_low
, n_high
,
4528 1, fold_convert_loc (loc
, arg0_type
,
4533 in_p
= (in_p
!= n_in_p
);
4547 /* Given EXP, a logical expression, set the range it is testing into
4548 variables denoted by PIN_P, PLOW, and PHIGH. Return the expression
4549 actually being tested. *PLOW and *PHIGH will be made of the same
4550 type as the returned expression. If EXP is not a comparison, we
4551 will most likely not be returning a useful value and range. Set
4552 *STRICT_OVERFLOW_P to true if the return value is only valid
4553 because signed overflow is undefined; otherwise, do not change
4554 *STRICT_OVERFLOW_P. */
4557 make_range (tree exp
, int *pin_p
, tree
*plow
, tree
*phigh
,
4558 bool *strict_overflow_p
)
4560 enum tree_code code
;
4561 tree arg0
, arg1
= NULL_TREE
;
4562 tree exp_type
, nexp
;
4565 location_t loc
= EXPR_LOCATION (exp
);
4567 /* Start with simply saying "EXP != 0" and then look at the code of EXP
4568 and see if we can refine the range. Some of the cases below may not
4569 happen, but it doesn't seem worth worrying about this. We "continue"
4570 the outer loop when we've changed something; otherwise we "break"
4571 the switch, which will "break" the while. */
4574 low
= high
= build_int_cst (TREE_TYPE (exp
), 0);
4578 code
= TREE_CODE (exp
);
4579 exp_type
= TREE_TYPE (exp
);
4582 if (IS_EXPR_CODE_CLASS (TREE_CODE_CLASS (code
)))
4584 if (TREE_OPERAND_LENGTH (exp
) > 0)
4585 arg0
= TREE_OPERAND (exp
, 0);
4586 if (TREE_CODE_CLASS (code
) == tcc_binary
4587 || TREE_CODE_CLASS (code
) == tcc_comparison
4588 || (TREE_CODE_CLASS (code
) == tcc_expression
4589 && TREE_OPERAND_LENGTH (exp
) > 1))
4590 arg1
= TREE_OPERAND (exp
, 1);
4592 if (arg0
== NULL_TREE
)
4595 nexp
= make_range_step (loc
, code
, arg0
, arg1
, exp_type
, &low
,
4596 &high
, &in_p
, strict_overflow_p
);
4597 if (nexp
== NULL_TREE
)
4602 /* If EXP is a constant, we can evaluate whether this is true or false. */
4603 if (TREE_CODE (exp
) == INTEGER_CST
)
4605 in_p
= in_p
== (integer_onep (range_binop (GE_EXPR
, integer_type_node
,
4607 && integer_onep (range_binop (LE_EXPR
, integer_type_node
,
4613 *pin_p
= in_p
, *plow
= low
, *phigh
= high
;
4617 /* Given a range, LOW, HIGH, and IN_P, an expression, EXP, and a result
4618 type, TYPE, return an expression to test if EXP is in (or out of, depending
4619 on IN_P) the range. Return 0 if the test couldn't be created. */
4622 build_range_check (location_t loc
, tree type
, tree exp
, int in_p
,
4623 tree low
, tree high
)
4625 tree etype
= TREE_TYPE (exp
), value
;
4627 /* Disable this optimization for function pointer expressions
4628 on targets that require function pointer canonicalization. */
4629 if (targetm
.have_canonicalize_funcptr_for_compare ()
4630 && TREE_CODE (etype
) == POINTER_TYPE
4631 && TREE_CODE (TREE_TYPE (etype
)) == FUNCTION_TYPE
)
4636 value
= build_range_check (loc
, type
, exp
, 1, low
, high
);
4638 return invert_truthvalue_loc (loc
, value
);
4643 if (low
== 0 && high
== 0)
4644 return omit_one_operand_loc (loc
, type
, build_int_cst (type
, 1), exp
);
4647 return fold_build2_loc (loc
, LE_EXPR
, type
, exp
,
4648 fold_convert_loc (loc
, etype
, high
));
4651 return fold_build2_loc (loc
, GE_EXPR
, type
, exp
,
4652 fold_convert_loc (loc
, etype
, low
));
4654 if (operand_equal_p (low
, high
, 0))
4655 return fold_build2_loc (loc
, EQ_EXPR
, type
, exp
,
4656 fold_convert_loc (loc
, etype
, low
));
4658 if (integer_zerop (low
))
4660 if (! TYPE_UNSIGNED (etype
))
4662 etype
= unsigned_type_for (etype
);
4663 high
= fold_convert_loc (loc
, etype
, high
);
4664 exp
= fold_convert_loc (loc
, etype
, exp
);
4666 return build_range_check (loc
, type
, exp
, 1, 0, high
);
4669 /* Optimize (c>=1) && (c<=127) into (signed char)c > 0. */
4670 if (integer_onep (low
) && TREE_CODE (high
) == INTEGER_CST
)
4672 int prec
= TYPE_PRECISION (etype
);
4674 if (wi::mask (prec
- 1, false, prec
) == high
)
4676 if (TYPE_UNSIGNED (etype
))
4678 tree signed_etype
= signed_type_for (etype
);
4679 if (TYPE_PRECISION (signed_etype
) != TYPE_PRECISION (etype
))
4681 = build_nonstandard_integer_type (TYPE_PRECISION (etype
), 0);
4683 etype
= signed_etype
;
4684 exp
= fold_convert_loc (loc
, etype
, exp
);
4686 return fold_build2_loc (loc
, GT_EXPR
, type
, exp
,
4687 build_int_cst (etype
, 0));
4691 /* Optimize (c>=low) && (c<=high) into (c-low>=0) && (c-low<=high-low).
4692 This requires wrap-around arithmetics for the type of the expression.
4693 First make sure that arithmetics in this type is valid, then make sure
4694 that it wraps around. */
4695 if (TREE_CODE (etype
) == ENUMERAL_TYPE
|| TREE_CODE (etype
) == BOOLEAN_TYPE
)
4696 etype
= lang_hooks
.types
.type_for_size (TYPE_PRECISION (etype
),
4697 TYPE_UNSIGNED (etype
));
4699 if (TREE_CODE (etype
) == INTEGER_TYPE
&& !TYPE_OVERFLOW_WRAPS (etype
))
4701 tree utype
, minv
, maxv
;
4703 /* Check if (unsigned) INT_MAX + 1 == (unsigned) INT_MIN
4704 for the type in question, as we rely on this here. */
4705 utype
= unsigned_type_for (etype
);
4706 maxv
= fold_convert_loc (loc
, utype
, TYPE_MAX_VALUE (etype
));
4707 maxv
= range_binop (PLUS_EXPR
, NULL_TREE
, maxv
, 1,
4708 build_int_cst (TREE_TYPE (maxv
), 1), 1);
4709 minv
= fold_convert_loc (loc
, utype
, TYPE_MIN_VALUE (etype
));
4711 if (integer_zerop (range_binop (NE_EXPR
, integer_type_node
,
4718 high
= fold_convert_loc (loc
, etype
, high
);
4719 low
= fold_convert_loc (loc
, etype
, low
);
4720 exp
= fold_convert_loc (loc
, etype
, exp
);
4722 value
= const_binop (MINUS_EXPR
, high
, low
);
4725 if (POINTER_TYPE_P (etype
))
4727 if (value
!= 0 && !TREE_OVERFLOW (value
))
4729 low
= fold_build1_loc (loc
, NEGATE_EXPR
, TREE_TYPE (low
), low
);
4730 return build_range_check (loc
, type
,
4731 fold_build_pointer_plus_loc (loc
, exp
, low
),
4732 1, build_int_cst (etype
, 0), value
);
4737 if (value
!= 0 && !TREE_OVERFLOW (value
))
4738 return build_range_check (loc
, type
,
4739 fold_build2_loc (loc
, MINUS_EXPR
, etype
, exp
, low
),
4740 1, build_int_cst (etype
, 0), value
);
4745 /* Return the predecessor of VAL in its type, handling the infinite case. */
4748 range_predecessor (tree val
)
4750 tree type
= TREE_TYPE (val
);
4752 if (INTEGRAL_TYPE_P (type
)
4753 && operand_equal_p (val
, TYPE_MIN_VALUE (type
), 0))
4756 return range_binop (MINUS_EXPR
, NULL_TREE
, val
, 0,
4757 build_int_cst (TREE_TYPE (val
), 1), 0);
4760 /* Return the successor of VAL in its type, handling the infinite case. */
4763 range_successor (tree val
)
4765 tree type
= TREE_TYPE (val
);
4767 if (INTEGRAL_TYPE_P (type
)
4768 && operand_equal_p (val
, TYPE_MAX_VALUE (type
), 0))
4771 return range_binop (PLUS_EXPR
, NULL_TREE
, val
, 0,
4772 build_int_cst (TREE_TYPE (val
), 1), 0);
4775 /* Given two ranges, see if we can merge them into one. Return 1 if we
4776 can, 0 if we can't. Set the output range into the specified parameters. */
4779 merge_ranges (int *pin_p
, tree
*plow
, tree
*phigh
, int in0_p
, tree low0
,
4780 tree high0
, int in1_p
, tree low1
, tree high1
)
4788 int lowequal
= ((low0
== 0 && low1
== 0)
4789 || integer_onep (range_binop (EQ_EXPR
, integer_type_node
,
4790 low0
, 0, low1
, 0)));
4791 int highequal
= ((high0
== 0 && high1
== 0)
4792 || integer_onep (range_binop (EQ_EXPR
, integer_type_node
,
4793 high0
, 1, high1
, 1)));
4795 /* Make range 0 be the range that starts first, or ends last if they
4796 start at the same value. Swap them if it isn't. */
4797 if (integer_onep (range_binop (GT_EXPR
, integer_type_node
,
4800 && integer_onep (range_binop (GT_EXPR
, integer_type_node
,
4801 high1
, 1, high0
, 1))))
4803 temp
= in0_p
, in0_p
= in1_p
, in1_p
= temp
;
4804 tem
= low0
, low0
= low1
, low1
= tem
;
4805 tem
= high0
, high0
= high1
, high1
= tem
;
4808 /* Now flag two cases, whether the ranges are disjoint or whether the
4809 second range is totally subsumed in the first. Note that the tests
4810 below are simplified by the ones above. */
4811 no_overlap
= integer_onep (range_binop (LT_EXPR
, integer_type_node
,
4812 high0
, 1, low1
, 0));
4813 subset
= integer_onep (range_binop (LE_EXPR
, integer_type_node
,
4814 high1
, 1, high0
, 1));
4816 /* We now have four cases, depending on whether we are including or
4817 excluding the two ranges. */
4820 /* If they don't overlap, the result is false. If the second range
4821 is a subset it is the result. Otherwise, the range is from the start
4822 of the second to the end of the first. */
4824 in_p
= 0, low
= high
= 0;
4826 in_p
= 1, low
= low1
, high
= high1
;
4828 in_p
= 1, low
= low1
, high
= high0
;
4831 else if (in0_p
&& ! in1_p
)
4833 /* If they don't overlap, the result is the first range. If they are
4834 equal, the result is false. If the second range is a subset of the
4835 first, and the ranges begin at the same place, we go from just after
4836 the end of the second range to the end of the first. If the second
4837 range is not a subset of the first, or if it is a subset and both
4838 ranges end at the same place, the range starts at the start of the
4839 first range and ends just before the second range.
4840 Otherwise, we can't describe this as a single range. */
4842 in_p
= 1, low
= low0
, high
= high0
;
4843 else if (lowequal
&& highequal
)
4844 in_p
= 0, low
= high
= 0;
4845 else if (subset
&& lowequal
)
4847 low
= range_successor (high1
);
4852 /* We are in the weird situation where high0 > high1 but
4853 high1 has no successor. Punt. */
4857 else if (! subset
|| highequal
)
4860 high
= range_predecessor (low1
);
4864 /* low0 < low1 but low1 has no predecessor. Punt. */
4872 else if (! in0_p
&& in1_p
)
4874 /* If they don't overlap, the result is the second range. If the second
4875 is a subset of the first, the result is false. Otherwise,
4876 the range starts just after the first range and ends at the
4877 end of the second. */
4879 in_p
= 1, low
= low1
, high
= high1
;
4880 else if (subset
|| highequal
)
4881 in_p
= 0, low
= high
= 0;
4884 low
= range_successor (high0
);
4889 /* high1 > high0 but high0 has no successor. Punt. */
4897 /* The case where we are excluding both ranges. Here the complex case
4898 is if they don't overlap. In that case, the only time we have a
4899 range is if they are adjacent. If the second is a subset of the
4900 first, the result is the first. Otherwise, the range to exclude
4901 starts at the beginning of the first range and ends at the end of the
4905 if (integer_onep (range_binop (EQ_EXPR
, integer_type_node
,
4906 range_successor (high0
),
4908 in_p
= 0, low
= low0
, high
= high1
;
4911 /* Canonicalize - [min, x] into - [-, x]. */
4912 if (low0
&& TREE_CODE (low0
) == INTEGER_CST
)
4913 switch (TREE_CODE (TREE_TYPE (low0
)))
4916 if (TYPE_PRECISION (TREE_TYPE (low0
))
4917 != GET_MODE_BITSIZE (TYPE_MODE (TREE_TYPE (low0
))))
4921 if (tree_int_cst_equal (low0
,
4922 TYPE_MIN_VALUE (TREE_TYPE (low0
))))
4926 if (TYPE_UNSIGNED (TREE_TYPE (low0
))
4927 && integer_zerop (low0
))
4934 /* Canonicalize - [x, max] into - [x, -]. */
4935 if (high1
&& TREE_CODE (high1
) == INTEGER_CST
)
4936 switch (TREE_CODE (TREE_TYPE (high1
)))
4939 if (TYPE_PRECISION (TREE_TYPE (high1
))
4940 != GET_MODE_BITSIZE (TYPE_MODE (TREE_TYPE (high1
))))
4944 if (tree_int_cst_equal (high1
,
4945 TYPE_MAX_VALUE (TREE_TYPE (high1
))))
4949 if (TYPE_UNSIGNED (TREE_TYPE (high1
))
4950 && integer_zerop (range_binop (PLUS_EXPR
, NULL_TREE
,
4952 build_int_cst (TREE_TYPE (high1
), 1),
4960 /* The ranges might be also adjacent between the maximum and
4961 minimum values of the given type. For
4962 - [{min,-}, x] and - [y, {max,-}] ranges where x + 1 < y
4963 return + [x + 1, y - 1]. */
4964 if (low0
== 0 && high1
== 0)
4966 low
= range_successor (high0
);
4967 high
= range_predecessor (low1
);
4968 if (low
== 0 || high
== 0)
4978 in_p
= 0, low
= low0
, high
= high0
;
4980 in_p
= 0, low
= low0
, high
= high1
;
4983 *pin_p
= in_p
, *plow
= low
, *phigh
= high
;
4988 /* Subroutine of fold, looking inside expressions of the form
4989 A op B ? A : C, where ARG0, ARG1 and ARG2 are the three operands
4990 of the COND_EXPR. This function is being used also to optimize
4991 A op B ? C : A, by reversing the comparison first.
4993 Return a folded expression whose code is not a COND_EXPR
4994 anymore, or NULL_TREE if no folding opportunity is found. */
4997 fold_cond_expr_with_comparison (location_t loc
, tree type
,
4998 tree arg0
, tree arg1
, tree arg2
)
5000 enum tree_code comp_code
= TREE_CODE (arg0
);
5001 tree arg00
= TREE_OPERAND (arg0
, 0);
5002 tree arg01
= TREE_OPERAND (arg0
, 1);
5003 tree arg1_type
= TREE_TYPE (arg1
);
5009 /* If we have A op 0 ? A : -A, consider applying the following
5012 A == 0? A : -A same as -A
5013 A != 0? A : -A same as A
5014 A >= 0? A : -A same as abs (A)
5015 A > 0? A : -A same as abs (A)
5016 A <= 0? A : -A same as -abs (A)
5017 A < 0? A : -A same as -abs (A)
5019 None of these transformations work for modes with signed
5020 zeros. If A is +/-0, the first two transformations will
5021 change the sign of the result (from +0 to -0, or vice
5022 versa). The last four will fix the sign of the result,
5023 even though the original expressions could be positive or
5024 negative, depending on the sign of A.
5026 Note that all these transformations are correct if A is
5027 NaN, since the two alternatives (A and -A) are also NaNs. */
5028 if (!HONOR_SIGNED_ZEROS (element_mode (type
))
5029 && (FLOAT_TYPE_P (TREE_TYPE (arg01
))
5030 ? real_zerop (arg01
)
5031 : integer_zerop (arg01
))
5032 && ((TREE_CODE (arg2
) == NEGATE_EXPR
5033 && operand_equal_p (TREE_OPERAND (arg2
, 0), arg1
, 0))
5034 /* In the case that A is of the form X-Y, '-A' (arg2) may
5035 have already been folded to Y-X, check for that. */
5036 || (TREE_CODE (arg1
) == MINUS_EXPR
5037 && TREE_CODE (arg2
) == MINUS_EXPR
5038 && operand_equal_p (TREE_OPERAND (arg1
, 0),
5039 TREE_OPERAND (arg2
, 1), 0)
5040 && operand_equal_p (TREE_OPERAND (arg1
, 1),
5041 TREE_OPERAND (arg2
, 0), 0))))
5046 tem
= fold_convert_loc (loc
, arg1_type
, arg1
);
5047 return pedantic_non_lvalue_loc (loc
,
5048 fold_convert_loc (loc
, type
,
5049 negate_expr (tem
)));
5052 return pedantic_non_lvalue_loc (loc
, fold_convert_loc (loc
, type
, arg1
));
5055 if (flag_trapping_math
)
5060 if (TYPE_UNSIGNED (TREE_TYPE (arg1
)))
5062 tem
= fold_build1_loc (loc
, ABS_EXPR
, TREE_TYPE (arg1
), arg1
);
5063 return pedantic_non_lvalue_loc (loc
, fold_convert_loc (loc
, type
, tem
));
5066 if (flag_trapping_math
)
5070 if (TYPE_UNSIGNED (TREE_TYPE (arg1
)))
5072 tem
= fold_build1_loc (loc
, ABS_EXPR
, TREE_TYPE (arg1
), arg1
);
5073 return negate_expr (fold_convert_loc (loc
, type
, tem
));
5075 gcc_assert (TREE_CODE_CLASS (comp_code
) == tcc_comparison
);
5079 /* A != 0 ? A : 0 is simply A, unless A is -0. Likewise
5080 A == 0 ? A : 0 is always 0 unless A is -0. Note that
5081 both transformations are correct when A is NaN: A != 0
5082 is then true, and A == 0 is false. */
5084 if (!HONOR_SIGNED_ZEROS (element_mode (type
))
5085 && integer_zerop (arg01
) && integer_zerop (arg2
))
5087 if (comp_code
== NE_EXPR
)
5088 return pedantic_non_lvalue_loc (loc
, fold_convert_loc (loc
, type
, arg1
));
5089 else if (comp_code
== EQ_EXPR
)
5090 return build_zero_cst (type
);
5093 /* Try some transformations of A op B ? A : B.
5095 A == B? A : B same as B
5096 A != B? A : B same as A
5097 A >= B? A : B same as max (A, B)
5098 A > B? A : B same as max (B, A)
5099 A <= B? A : B same as min (A, B)
5100 A < B? A : B same as min (B, A)
5102 As above, these transformations don't work in the presence
5103 of signed zeros. For example, if A and B are zeros of
5104 opposite sign, the first two transformations will change
5105 the sign of the result. In the last four, the original
5106 expressions give different results for (A=+0, B=-0) and
5107 (A=-0, B=+0), but the transformed expressions do not.
5109 The first two transformations are correct if either A or B
5110 is a NaN. In the first transformation, the condition will
5111 be false, and B will indeed be chosen. In the case of the
5112 second transformation, the condition A != B will be true,
5113 and A will be chosen.
5115 The conversions to max() and min() are not correct if B is
5116 a number and A is not. The conditions in the original
5117 expressions will be false, so all four give B. The min()
5118 and max() versions would give a NaN instead. */
5119 if (!HONOR_SIGNED_ZEROS (element_mode (type
))
5120 && operand_equal_for_comparison_p (arg01
, arg2
, arg00
)
5121 /* Avoid these transformations if the COND_EXPR may be used
5122 as an lvalue in the C++ front-end. PR c++/19199. */
5124 || VECTOR_TYPE_P (type
)
5125 || (! lang_GNU_CXX ()
5126 && strcmp (lang_hooks
.name
, "GNU Objective-C++") != 0)
5127 || ! maybe_lvalue_p (arg1
)
5128 || ! maybe_lvalue_p (arg2
)))
5130 tree comp_op0
= arg00
;
5131 tree comp_op1
= arg01
;
5132 tree comp_type
= TREE_TYPE (comp_op0
);
5134 /* Avoid adding NOP_EXPRs in case this is an lvalue. */
5135 if (TYPE_MAIN_VARIANT (comp_type
) == TYPE_MAIN_VARIANT (type
))
5145 return pedantic_non_lvalue_loc (loc
, fold_convert_loc (loc
, type
, arg2
));
5147 return pedantic_non_lvalue_loc (loc
, fold_convert_loc (loc
, type
, arg1
));
5152 /* In C++ a ?: expression can be an lvalue, so put the
5153 operand which will be used if they are equal first
5154 so that we can convert this back to the
5155 corresponding COND_EXPR. */
5156 if (!HONOR_NANS (arg1
))
5158 comp_op0
= fold_convert_loc (loc
, comp_type
, comp_op0
);
5159 comp_op1
= fold_convert_loc (loc
, comp_type
, comp_op1
);
5160 tem
= (comp_code
== LE_EXPR
|| comp_code
== UNLE_EXPR
)
5161 ? fold_build2_loc (loc
, MIN_EXPR
, comp_type
, comp_op0
, comp_op1
)
5162 : fold_build2_loc (loc
, MIN_EXPR
, comp_type
,
5163 comp_op1
, comp_op0
);
5164 return pedantic_non_lvalue_loc (loc
,
5165 fold_convert_loc (loc
, type
, tem
));
5172 if (!HONOR_NANS (arg1
))
5174 comp_op0
= fold_convert_loc (loc
, comp_type
, comp_op0
);
5175 comp_op1
= fold_convert_loc (loc
, comp_type
, comp_op1
);
5176 tem
= (comp_code
== GE_EXPR
|| comp_code
== UNGE_EXPR
)
5177 ? fold_build2_loc (loc
, MAX_EXPR
, comp_type
, comp_op0
, comp_op1
)
5178 : fold_build2_loc (loc
, MAX_EXPR
, comp_type
,
5179 comp_op1
, comp_op0
);
5180 return pedantic_non_lvalue_loc (loc
,
5181 fold_convert_loc (loc
, type
, tem
));
5185 if (!HONOR_NANS (arg1
))
5186 return pedantic_non_lvalue_loc (loc
,
5187 fold_convert_loc (loc
, type
, arg2
));
5190 if (!HONOR_NANS (arg1
))
5191 return pedantic_non_lvalue_loc (loc
,
5192 fold_convert_loc (loc
, type
, arg1
));
5195 gcc_assert (TREE_CODE_CLASS (comp_code
) == tcc_comparison
);
5200 /* If this is A op C1 ? A : C2 with C1 and C2 constant integers,
5201 we might still be able to simplify this. For example,
5202 if C1 is one less or one more than C2, this might have started
5203 out as a MIN or MAX and been transformed by this function.
5204 Only good for INTEGER_TYPEs, because we need TYPE_MAX_VALUE. */
5206 if (INTEGRAL_TYPE_P (type
)
5207 && TREE_CODE (arg01
) == INTEGER_CST
5208 && TREE_CODE (arg2
) == INTEGER_CST
)
5212 if (TREE_CODE (arg1
) == INTEGER_CST
)
5214 /* We can replace A with C1 in this case. */
5215 arg1
= fold_convert_loc (loc
, type
, arg01
);
5216 return fold_build3_loc (loc
, COND_EXPR
, type
, arg0
, arg1
, arg2
);
5219 /* If C1 is C2 + 1, this is min(A, C2), but use ARG00's type for
5220 MIN_EXPR, to preserve the signedness of the comparison. */
5221 if (! operand_equal_p (arg2
, TYPE_MAX_VALUE (type
),
5223 && operand_equal_p (arg01
,
5224 const_binop (PLUS_EXPR
, arg2
,
5225 build_int_cst (type
, 1)),
5228 tem
= fold_build2_loc (loc
, MIN_EXPR
, TREE_TYPE (arg00
), arg00
,
5229 fold_convert_loc (loc
, TREE_TYPE (arg00
),
5231 return pedantic_non_lvalue_loc (loc
,
5232 fold_convert_loc (loc
, type
, tem
));
5237 /* If C1 is C2 - 1, this is min(A, C2), with the same care
5239 if (! operand_equal_p (arg2
, TYPE_MIN_VALUE (type
),
5241 && operand_equal_p (arg01
,
5242 const_binop (MINUS_EXPR
, arg2
,
5243 build_int_cst (type
, 1)),
5246 tem
= fold_build2_loc (loc
, MIN_EXPR
, TREE_TYPE (arg00
), arg00
,
5247 fold_convert_loc (loc
, TREE_TYPE (arg00
),
5249 return pedantic_non_lvalue_loc (loc
,
5250 fold_convert_loc (loc
, type
, tem
));
5255 /* If C1 is C2 - 1, this is max(A, C2), but use ARG00's type for
5256 MAX_EXPR, to preserve the signedness of the comparison. */
5257 if (! operand_equal_p (arg2
, TYPE_MIN_VALUE (type
),
5259 && operand_equal_p (arg01
,
5260 const_binop (MINUS_EXPR
, arg2
,
5261 build_int_cst (type
, 1)),
5264 tem
= fold_build2_loc (loc
, MAX_EXPR
, TREE_TYPE (arg00
), arg00
,
5265 fold_convert_loc (loc
, TREE_TYPE (arg00
),
5267 return pedantic_non_lvalue_loc (loc
, fold_convert_loc (loc
, type
, tem
));
5272 /* If C1 is C2 + 1, this is max(A, C2), with the same care as above. */
5273 if (! operand_equal_p (arg2
, TYPE_MAX_VALUE (type
),
5275 && operand_equal_p (arg01
,
5276 const_binop (PLUS_EXPR
, arg2
,
5277 build_int_cst (type
, 1)),
5280 tem
= fold_build2_loc (loc
, MAX_EXPR
, TREE_TYPE (arg00
), arg00
,
5281 fold_convert_loc (loc
, TREE_TYPE (arg00
),
5283 return pedantic_non_lvalue_loc (loc
, fold_convert_loc (loc
, type
, tem
));
5297 #ifndef LOGICAL_OP_NON_SHORT_CIRCUIT
5298 #define LOGICAL_OP_NON_SHORT_CIRCUIT \
5299 (BRANCH_COST (optimize_function_for_speed_p (cfun), \
5303 /* EXP is some logical combination of boolean tests. See if we can
5304 merge it into some range test. Return the new tree if so. */
5307 fold_range_test (location_t loc
, enum tree_code code
, tree type
,
5310 int or_op
= (code
== TRUTH_ORIF_EXPR
5311 || code
== TRUTH_OR_EXPR
);
5312 int in0_p
, in1_p
, in_p
;
5313 tree low0
, low1
, low
, high0
, high1
, high
;
5314 bool strict_overflow_p
= false;
5316 const char * const warnmsg
= G_("assuming signed overflow does not occur "
5317 "when simplifying range test");
5319 if (!INTEGRAL_TYPE_P (type
))
5322 lhs
= make_range (op0
, &in0_p
, &low0
, &high0
, &strict_overflow_p
);
5323 rhs
= make_range (op1
, &in1_p
, &low1
, &high1
, &strict_overflow_p
);
5325 /* If this is an OR operation, invert both sides; we will invert
5326 again at the end. */
5328 in0_p
= ! in0_p
, in1_p
= ! in1_p
;
5330 /* If both expressions are the same, if we can merge the ranges, and we
5331 can build the range test, return it or it inverted. If one of the
5332 ranges is always true or always false, consider it to be the same
5333 expression as the other. */
5334 if ((lhs
== 0 || rhs
== 0 || operand_equal_p (lhs
, rhs
, 0))
5335 && merge_ranges (&in_p
, &low
, &high
, in0_p
, low0
, high0
,
5337 && 0 != (tem
= (build_range_check (loc
, type
,
5339 : rhs
!= 0 ? rhs
: integer_zero_node
,
5342 if (strict_overflow_p
)
5343 fold_overflow_warning (warnmsg
, WARN_STRICT_OVERFLOW_COMPARISON
);
5344 return or_op
? invert_truthvalue_loc (loc
, tem
) : tem
;
5347 /* On machines where the branch cost is expensive, if this is a
5348 short-circuited branch and the underlying object on both sides
5349 is the same, make a non-short-circuit operation. */
5350 else if (LOGICAL_OP_NON_SHORT_CIRCUIT
5351 && lhs
!= 0 && rhs
!= 0
5352 && (code
== TRUTH_ANDIF_EXPR
5353 || code
== TRUTH_ORIF_EXPR
)
5354 && operand_equal_p (lhs
, rhs
, 0))
5356 /* If simple enough, just rewrite. Otherwise, make a SAVE_EXPR
5357 unless we are at top level or LHS contains a PLACEHOLDER_EXPR, in
5358 which cases we can't do this. */
5359 if (simple_operand_p (lhs
))
5360 return build2_loc (loc
, code
== TRUTH_ANDIF_EXPR
5361 ? TRUTH_AND_EXPR
: TRUTH_OR_EXPR
,
5364 else if (!lang_hooks
.decls
.global_bindings_p ()
5365 && !CONTAINS_PLACEHOLDER_P (lhs
))
5367 tree common
= save_expr (lhs
);
5369 if (0 != (lhs
= build_range_check (loc
, type
, common
,
5370 or_op
? ! in0_p
: in0_p
,
5372 && (0 != (rhs
= build_range_check (loc
, type
, common
,
5373 or_op
? ! in1_p
: in1_p
,
5376 if (strict_overflow_p
)
5377 fold_overflow_warning (warnmsg
,
5378 WARN_STRICT_OVERFLOW_COMPARISON
);
5379 return build2_loc (loc
, code
== TRUTH_ANDIF_EXPR
5380 ? TRUTH_AND_EXPR
: TRUTH_OR_EXPR
,
5389 /* Subroutine for fold_truth_andor_1: C is an INTEGER_CST interpreted as a P
5390 bit value. Arrange things so the extra bits will be set to zero if and
5391 only if C is signed-extended to its full width. If MASK is nonzero,
5392 it is an INTEGER_CST that should be AND'ed with the extra bits. */
5395 unextend (tree c
, int p
, int unsignedp
, tree mask
)
5397 tree type
= TREE_TYPE (c
);
5398 int modesize
= GET_MODE_BITSIZE (TYPE_MODE (type
));
5401 if (p
== modesize
|| unsignedp
)
5404 /* We work by getting just the sign bit into the low-order bit, then
5405 into the high-order bit, then sign-extend. We then XOR that value
5407 temp
= build_int_cst (TREE_TYPE (c
), wi::extract_uhwi (c
, p
- 1, 1));
5409 /* We must use a signed type in order to get an arithmetic right shift.
5410 However, we must also avoid introducing accidental overflows, so that
5411 a subsequent call to integer_zerop will work. Hence we must
5412 do the type conversion here. At this point, the constant is either
5413 zero or one, and the conversion to a signed type can never overflow.
5414 We could get an overflow if this conversion is done anywhere else. */
5415 if (TYPE_UNSIGNED (type
))
5416 temp
= fold_convert (signed_type_for (type
), temp
);
5418 temp
= const_binop (LSHIFT_EXPR
, temp
, size_int (modesize
- 1));
5419 temp
= const_binop (RSHIFT_EXPR
, temp
, size_int (modesize
- p
- 1));
5421 temp
= const_binop (BIT_AND_EXPR
, temp
,
5422 fold_convert (TREE_TYPE (c
), mask
));
5423 /* If necessary, convert the type back to match the type of C. */
5424 if (TYPE_UNSIGNED (type
))
5425 temp
= fold_convert (type
, temp
);
5427 return fold_convert (type
, const_binop (BIT_XOR_EXPR
, c
, temp
));
5430 /* For an expression that has the form
5434 we can drop one of the inner expressions and simplify to
5438 LOC is the location of the resulting expression. OP is the inner
5439 logical operation; the left-hand side in the examples above, while CMPOP
5440 is the right-hand side. RHS_ONLY is used to prevent us from accidentally
5441 removing a condition that guards another, as in
5442 (A != NULL && A->...) || A == NULL
5443 which we must not transform. If RHS_ONLY is true, only eliminate the
5444 right-most operand of the inner logical operation. */
5447 merge_truthop_with_opposite_arm (location_t loc
, tree op
, tree cmpop
,
5450 tree type
= TREE_TYPE (cmpop
);
5451 enum tree_code code
= TREE_CODE (cmpop
);
5452 enum tree_code truthop_code
= TREE_CODE (op
);
5453 tree lhs
= TREE_OPERAND (op
, 0);
5454 tree rhs
= TREE_OPERAND (op
, 1);
5455 tree orig_lhs
= lhs
, orig_rhs
= rhs
;
5456 enum tree_code rhs_code
= TREE_CODE (rhs
);
5457 enum tree_code lhs_code
= TREE_CODE (lhs
);
5458 enum tree_code inv_code
;
5460 if (TREE_SIDE_EFFECTS (op
) || TREE_SIDE_EFFECTS (cmpop
))
5463 if (TREE_CODE_CLASS (code
) != tcc_comparison
)
5466 if (rhs_code
== truthop_code
)
5468 tree newrhs
= merge_truthop_with_opposite_arm (loc
, rhs
, cmpop
, rhs_only
);
5469 if (newrhs
!= NULL_TREE
)
5472 rhs_code
= TREE_CODE (rhs
);
5475 if (lhs_code
== truthop_code
&& !rhs_only
)
5477 tree newlhs
= merge_truthop_with_opposite_arm (loc
, lhs
, cmpop
, false);
5478 if (newlhs
!= NULL_TREE
)
5481 lhs_code
= TREE_CODE (lhs
);
5485 inv_code
= invert_tree_comparison (code
, HONOR_NANS (type
));
5486 if (inv_code
== rhs_code
5487 && operand_equal_p (TREE_OPERAND (rhs
, 0), TREE_OPERAND (cmpop
, 0), 0)
5488 && operand_equal_p (TREE_OPERAND (rhs
, 1), TREE_OPERAND (cmpop
, 1), 0))
5490 if (!rhs_only
&& inv_code
== lhs_code
5491 && operand_equal_p (TREE_OPERAND (lhs
, 0), TREE_OPERAND (cmpop
, 0), 0)
5492 && operand_equal_p (TREE_OPERAND (lhs
, 1), TREE_OPERAND (cmpop
, 1), 0))
5494 if (rhs
!= orig_rhs
|| lhs
!= orig_lhs
)
5495 return fold_build2_loc (loc
, truthop_code
, TREE_TYPE (cmpop
),
5500 /* Find ways of folding logical expressions of LHS and RHS:
5501 Try to merge two comparisons to the same innermost item.
5502 Look for range tests like "ch >= '0' && ch <= '9'".
5503 Look for combinations of simple terms on machines with expensive branches
5504 and evaluate the RHS unconditionally.
5506 For example, if we have p->a == 2 && p->b == 4 and we can make an
5507 object large enough to span both A and B, we can do this with a comparison
5508 against the object ANDed with the a mask.
5510 If we have p->a == q->a && p->b == q->b, we may be able to use bit masking
5511 operations to do this with one comparison.
5513 We check for both normal comparisons and the BIT_AND_EXPRs made this by
5514 function and the one above.
5516 CODE is the logical operation being done. It can be TRUTH_ANDIF_EXPR,
5517 TRUTH_AND_EXPR, TRUTH_ORIF_EXPR, or TRUTH_OR_EXPR.
5519 TRUTH_TYPE is the type of the logical operand and LHS and RHS are its
5522 We return the simplified tree or 0 if no optimization is possible. */
5525 fold_truth_andor_1 (location_t loc
, enum tree_code code
, tree truth_type
,
5528 /* If this is the "or" of two comparisons, we can do something if
5529 the comparisons are NE_EXPR. If this is the "and", we can do something
5530 if the comparisons are EQ_EXPR. I.e.,
5531 (a->b == 2 && a->c == 4) can become (a->new == NEW).
5533 WANTED_CODE is this operation code. For single bit fields, we can
5534 convert EQ_EXPR to NE_EXPR so we need not reject the "wrong"
5535 comparison for one-bit fields. */
5537 enum tree_code wanted_code
;
5538 enum tree_code lcode
, rcode
;
5539 tree ll_arg
, lr_arg
, rl_arg
, rr_arg
;
5540 tree ll_inner
, lr_inner
, rl_inner
, rr_inner
;
5541 HOST_WIDE_INT ll_bitsize
, ll_bitpos
, lr_bitsize
, lr_bitpos
;
5542 HOST_WIDE_INT rl_bitsize
, rl_bitpos
, rr_bitsize
, rr_bitpos
;
5543 HOST_WIDE_INT xll_bitpos
, xlr_bitpos
, xrl_bitpos
, xrr_bitpos
;
5544 HOST_WIDE_INT lnbitsize
, lnbitpos
, rnbitsize
, rnbitpos
;
5545 int ll_unsignedp
, lr_unsignedp
, rl_unsignedp
, rr_unsignedp
;
5546 int ll_reversep
, lr_reversep
, rl_reversep
, rr_reversep
;
5547 machine_mode ll_mode
, lr_mode
, rl_mode
, rr_mode
;
5548 machine_mode lnmode
, rnmode
;
5549 tree ll_mask
, lr_mask
, rl_mask
, rr_mask
;
5550 tree ll_and_mask
, lr_and_mask
, rl_and_mask
, rr_and_mask
;
5551 tree l_const
, r_const
;
5552 tree lntype
, rntype
, result
;
5553 HOST_WIDE_INT first_bit
, end_bit
;
5556 /* Start by getting the comparison codes. Fail if anything is volatile.
5557 If one operand is a BIT_AND_EXPR with the constant one, treat it as if
5558 it were surrounded with a NE_EXPR. */
5560 if (TREE_SIDE_EFFECTS (lhs
) || TREE_SIDE_EFFECTS (rhs
))
5563 lcode
= TREE_CODE (lhs
);
5564 rcode
= TREE_CODE (rhs
);
5566 if (lcode
== BIT_AND_EXPR
&& integer_onep (TREE_OPERAND (lhs
, 1)))
5568 lhs
= build2 (NE_EXPR
, truth_type
, lhs
,
5569 build_int_cst (TREE_TYPE (lhs
), 0));
5573 if (rcode
== BIT_AND_EXPR
&& integer_onep (TREE_OPERAND (rhs
, 1)))
5575 rhs
= build2 (NE_EXPR
, truth_type
, rhs
,
5576 build_int_cst (TREE_TYPE (rhs
), 0));
5580 if (TREE_CODE_CLASS (lcode
) != tcc_comparison
5581 || TREE_CODE_CLASS (rcode
) != tcc_comparison
)
5584 ll_arg
= TREE_OPERAND (lhs
, 0);
5585 lr_arg
= TREE_OPERAND (lhs
, 1);
5586 rl_arg
= TREE_OPERAND (rhs
, 0);
5587 rr_arg
= TREE_OPERAND (rhs
, 1);
5589 /* Simplify (x<y) && (x==y) into (x<=y) and related optimizations. */
5590 if (simple_operand_p (ll_arg
)
5591 && simple_operand_p (lr_arg
))
5593 if (operand_equal_p (ll_arg
, rl_arg
, 0)
5594 && operand_equal_p (lr_arg
, rr_arg
, 0))
5596 result
= combine_comparisons (loc
, code
, lcode
, rcode
,
5597 truth_type
, ll_arg
, lr_arg
);
5601 else if (operand_equal_p (ll_arg
, rr_arg
, 0)
5602 && operand_equal_p (lr_arg
, rl_arg
, 0))
5604 result
= combine_comparisons (loc
, code
, lcode
,
5605 swap_tree_comparison (rcode
),
5606 truth_type
, ll_arg
, lr_arg
);
5612 code
= ((code
== TRUTH_AND_EXPR
|| code
== TRUTH_ANDIF_EXPR
)
5613 ? TRUTH_AND_EXPR
: TRUTH_OR_EXPR
);
5615 /* If the RHS can be evaluated unconditionally and its operands are
5616 simple, it wins to evaluate the RHS unconditionally on machines
5617 with expensive branches. In this case, this isn't a comparison
5618 that can be merged. */
5620 if (BRANCH_COST (optimize_function_for_speed_p (cfun
),
5622 && ! FLOAT_TYPE_P (TREE_TYPE (rl_arg
))
5623 && simple_operand_p (rl_arg
)
5624 && simple_operand_p (rr_arg
))
5626 /* Convert (a != 0) || (b != 0) into (a | b) != 0. */
5627 if (code
== TRUTH_OR_EXPR
5628 && lcode
== NE_EXPR
&& integer_zerop (lr_arg
)
5629 && rcode
== NE_EXPR
&& integer_zerop (rr_arg
)
5630 && TREE_TYPE (ll_arg
) == TREE_TYPE (rl_arg
)
5631 && INTEGRAL_TYPE_P (TREE_TYPE (ll_arg
)))
5632 return build2_loc (loc
, NE_EXPR
, truth_type
,
5633 build2 (BIT_IOR_EXPR
, TREE_TYPE (ll_arg
),
5635 build_int_cst (TREE_TYPE (ll_arg
), 0));
5637 /* Convert (a == 0) && (b == 0) into (a | b) == 0. */
5638 if (code
== TRUTH_AND_EXPR
5639 && lcode
== EQ_EXPR
&& integer_zerop (lr_arg
)
5640 && rcode
== EQ_EXPR
&& integer_zerop (rr_arg
)
5641 && TREE_TYPE (ll_arg
) == TREE_TYPE (rl_arg
)
5642 && INTEGRAL_TYPE_P (TREE_TYPE (ll_arg
)))
5643 return build2_loc (loc
, EQ_EXPR
, truth_type
,
5644 build2 (BIT_IOR_EXPR
, TREE_TYPE (ll_arg
),
5646 build_int_cst (TREE_TYPE (ll_arg
), 0));
5649 /* See if the comparisons can be merged. Then get all the parameters for
5652 if ((lcode
!= EQ_EXPR
&& lcode
!= NE_EXPR
)
5653 || (rcode
!= EQ_EXPR
&& rcode
!= NE_EXPR
))
5656 ll_reversep
= lr_reversep
= rl_reversep
= rr_reversep
= 0;
5658 ll_inner
= decode_field_reference (loc
, ll_arg
,
5659 &ll_bitsize
, &ll_bitpos
, &ll_mode
,
5660 &ll_unsignedp
, &ll_reversep
, &volatilep
,
5661 &ll_mask
, &ll_and_mask
);
5662 lr_inner
= decode_field_reference (loc
, lr_arg
,
5663 &lr_bitsize
, &lr_bitpos
, &lr_mode
,
5664 &lr_unsignedp
, &lr_reversep
, &volatilep
,
5665 &lr_mask
, &lr_and_mask
);
5666 rl_inner
= decode_field_reference (loc
, rl_arg
,
5667 &rl_bitsize
, &rl_bitpos
, &rl_mode
,
5668 &rl_unsignedp
, &rl_reversep
, &volatilep
,
5669 &rl_mask
, &rl_and_mask
);
5670 rr_inner
= decode_field_reference (loc
, rr_arg
,
5671 &rr_bitsize
, &rr_bitpos
, &rr_mode
,
5672 &rr_unsignedp
, &rr_reversep
, &volatilep
,
5673 &rr_mask
, &rr_and_mask
);
5675 /* It must be true that the inner operation on the lhs of each
5676 comparison must be the same if we are to be able to do anything.
5677 Then see if we have constants. If not, the same must be true for
5680 || ll_reversep
!= rl_reversep
5681 || ll_inner
== 0 || rl_inner
== 0
5682 || ! operand_equal_p (ll_inner
, rl_inner
, 0))
5685 if (TREE_CODE (lr_arg
) == INTEGER_CST
5686 && TREE_CODE (rr_arg
) == INTEGER_CST
)
5688 l_const
= lr_arg
, r_const
= rr_arg
;
5689 lr_reversep
= ll_reversep
;
5691 else if (lr_reversep
!= rr_reversep
5692 || lr_inner
== 0 || rr_inner
== 0
5693 || ! operand_equal_p (lr_inner
, rr_inner
, 0))
5696 l_const
= r_const
= 0;
5698 /* If either comparison code is not correct for our logical operation,
5699 fail. However, we can convert a one-bit comparison against zero into
5700 the opposite comparison against that bit being set in the field. */
5702 wanted_code
= (code
== TRUTH_AND_EXPR
? EQ_EXPR
: NE_EXPR
);
5703 if (lcode
!= wanted_code
)
5705 if (l_const
&& integer_zerop (l_const
) && integer_pow2p (ll_mask
))
5707 /* Make the left operand unsigned, since we are only interested
5708 in the value of one bit. Otherwise we are doing the wrong
5717 /* This is analogous to the code for l_const above. */
5718 if (rcode
!= wanted_code
)
5720 if (r_const
&& integer_zerop (r_const
) && integer_pow2p (rl_mask
))
5729 /* See if we can find a mode that contains both fields being compared on
5730 the left. If we can't, fail. Otherwise, update all constants and masks
5731 to be relative to a field of that size. */
5732 first_bit
= MIN (ll_bitpos
, rl_bitpos
);
5733 end_bit
= MAX (ll_bitpos
+ ll_bitsize
, rl_bitpos
+ rl_bitsize
);
5734 lnmode
= get_best_mode (end_bit
- first_bit
, first_bit
, 0, 0,
5735 TYPE_ALIGN (TREE_TYPE (ll_inner
)), word_mode
,
5737 if (lnmode
== VOIDmode
)
5740 lnbitsize
= GET_MODE_BITSIZE (lnmode
);
5741 lnbitpos
= first_bit
& ~ (lnbitsize
- 1);
5742 lntype
= lang_hooks
.types
.type_for_size (lnbitsize
, 1);
5743 xll_bitpos
= ll_bitpos
- lnbitpos
, xrl_bitpos
= rl_bitpos
- lnbitpos
;
5745 if (ll_reversep
? !BYTES_BIG_ENDIAN
: BYTES_BIG_ENDIAN
)
5747 xll_bitpos
= lnbitsize
- xll_bitpos
- ll_bitsize
;
5748 xrl_bitpos
= lnbitsize
- xrl_bitpos
- rl_bitsize
;
5751 ll_mask
= const_binop (LSHIFT_EXPR
, fold_convert_loc (loc
, lntype
, ll_mask
),
5752 size_int (xll_bitpos
));
5753 rl_mask
= const_binop (LSHIFT_EXPR
, fold_convert_loc (loc
, lntype
, rl_mask
),
5754 size_int (xrl_bitpos
));
5758 l_const
= fold_convert_loc (loc
, lntype
, l_const
);
5759 l_const
= unextend (l_const
, ll_bitsize
, ll_unsignedp
, ll_and_mask
);
5760 l_const
= const_binop (LSHIFT_EXPR
, l_const
, size_int (xll_bitpos
));
5761 if (! integer_zerop (const_binop (BIT_AND_EXPR
, l_const
,
5762 fold_build1_loc (loc
, BIT_NOT_EXPR
,
5765 warning (0, "comparison is always %d", wanted_code
== NE_EXPR
);
5767 return constant_boolean_node (wanted_code
== NE_EXPR
, truth_type
);
5772 r_const
= fold_convert_loc (loc
, lntype
, r_const
);
5773 r_const
= unextend (r_const
, rl_bitsize
, rl_unsignedp
, rl_and_mask
);
5774 r_const
= const_binop (LSHIFT_EXPR
, r_const
, size_int (xrl_bitpos
));
5775 if (! integer_zerop (const_binop (BIT_AND_EXPR
, r_const
,
5776 fold_build1_loc (loc
, BIT_NOT_EXPR
,
5779 warning (0, "comparison is always %d", wanted_code
== NE_EXPR
);
5781 return constant_boolean_node (wanted_code
== NE_EXPR
, truth_type
);
5785 /* If the right sides are not constant, do the same for it. Also,
5786 disallow this optimization if a size or signedness mismatch occurs
5787 between the left and right sides. */
5790 if (ll_bitsize
!= lr_bitsize
|| rl_bitsize
!= rr_bitsize
5791 || ll_unsignedp
!= lr_unsignedp
|| rl_unsignedp
!= rr_unsignedp
5792 /* Make sure the two fields on the right
5793 correspond to the left without being swapped. */
5794 || ll_bitpos
- rl_bitpos
!= lr_bitpos
- rr_bitpos
)
5797 first_bit
= MIN (lr_bitpos
, rr_bitpos
);
5798 end_bit
= MAX (lr_bitpos
+ lr_bitsize
, rr_bitpos
+ rr_bitsize
);
5799 rnmode
= get_best_mode (end_bit
- first_bit
, first_bit
, 0, 0,
5800 TYPE_ALIGN (TREE_TYPE (lr_inner
)), word_mode
,
5802 if (rnmode
== VOIDmode
)
5805 rnbitsize
= GET_MODE_BITSIZE (rnmode
);
5806 rnbitpos
= first_bit
& ~ (rnbitsize
- 1);
5807 rntype
= lang_hooks
.types
.type_for_size (rnbitsize
, 1);
5808 xlr_bitpos
= lr_bitpos
- rnbitpos
, xrr_bitpos
= rr_bitpos
- rnbitpos
;
5810 if (lr_reversep
? !BYTES_BIG_ENDIAN
: BYTES_BIG_ENDIAN
)
5812 xlr_bitpos
= rnbitsize
- xlr_bitpos
- lr_bitsize
;
5813 xrr_bitpos
= rnbitsize
- xrr_bitpos
- rr_bitsize
;
5816 lr_mask
= const_binop (LSHIFT_EXPR
, fold_convert_loc (loc
,
5818 size_int (xlr_bitpos
));
5819 rr_mask
= const_binop (LSHIFT_EXPR
, fold_convert_loc (loc
,
5821 size_int (xrr_bitpos
));
5823 /* Make a mask that corresponds to both fields being compared.
5824 Do this for both items being compared. If the operands are the
5825 same size and the bits being compared are in the same position
5826 then we can do this by masking both and comparing the masked
5828 ll_mask
= const_binop (BIT_IOR_EXPR
, ll_mask
, rl_mask
);
5829 lr_mask
= const_binop (BIT_IOR_EXPR
, lr_mask
, rr_mask
);
5830 if (lnbitsize
== rnbitsize
&& xll_bitpos
== xlr_bitpos
)
5832 lhs
= make_bit_field_ref (loc
, ll_inner
, lntype
, lnbitsize
, lnbitpos
,
5833 ll_unsignedp
|| rl_unsignedp
, ll_reversep
);
5834 if (! all_ones_mask_p (ll_mask
, lnbitsize
))
5835 lhs
= build2 (BIT_AND_EXPR
, lntype
, lhs
, ll_mask
);
5837 rhs
= make_bit_field_ref (loc
, lr_inner
, rntype
, rnbitsize
, rnbitpos
,
5838 lr_unsignedp
|| rr_unsignedp
, lr_reversep
);
5839 if (! all_ones_mask_p (lr_mask
, rnbitsize
))
5840 rhs
= build2 (BIT_AND_EXPR
, rntype
, rhs
, lr_mask
);
5842 return build2_loc (loc
, wanted_code
, truth_type
, lhs
, rhs
);
5845 /* There is still another way we can do something: If both pairs of
5846 fields being compared are adjacent, we may be able to make a wider
5847 field containing them both.
5849 Note that we still must mask the lhs/rhs expressions. Furthermore,
5850 the mask must be shifted to account for the shift done by
5851 make_bit_field_ref. */
5852 if ((ll_bitsize
+ ll_bitpos
== rl_bitpos
5853 && lr_bitsize
+ lr_bitpos
== rr_bitpos
)
5854 || (ll_bitpos
== rl_bitpos
+ rl_bitsize
5855 && lr_bitpos
== rr_bitpos
+ rr_bitsize
))
5859 lhs
= make_bit_field_ref (loc
, ll_inner
, lntype
,
5860 ll_bitsize
+ rl_bitsize
,
5861 MIN (ll_bitpos
, rl_bitpos
),
5862 ll_unsignedp
, ll_reversep
);
5863 rhs
= make_bit_field_ref (loc
, lr_inner
, rntype
,
5864 lr_bitsize
+ rr_bitsize
,
5865 MIN (lr_bitpos
, rr_bitpos
),
5866 lr_unsignedp
, lr_reversep
);
5868 ll_mask
= const_binop (RSHIFT_EXPR
, ll_mask
,
5869 size_int (MIN (xll_bitpos
, xrl_bitpos
)));
5870 lr_mask
= const_binop (RSHIFT_EXPR
, lr_mask
,
5871 size_int (MIN (xlr_bitpos
, xrr_bitpos
)));
5873 /* Convert to the smaller type before masking out unwanted bits. */
5875 if (lntype
!= rntype
)
5877 if (lnbitsize
> rnbitsize
)
5879 lhs
= fold_convert_loc (loc
, rntype
, lhs
);
5880 ll_mask
= fold_convert_loc (loc
, rntype
, ll_mask
);
5883 else if (lnbitsize
< rnbitsize
)
5885 rhs
= fold_convert_loc (loc
, lntype
, rhs
);
5886 lr_mask
= fold_convert_loc (loc
, lntype
, lr_mask
);
5891 if (! all_ones_mask_p (ll_mask
, ll_bitsize
+ rl_bitsize
))
5892 lhs
= build2 (BIT_AND_EXPR
, type
, lhs
, ll_mask
);
5894 if (! all_ones_mask_p (lr_mask
, lr_bitsize
+ rr_bitsize
))
5895 rhs
= build2 (BIT_AND_EXPR
, type
, rhs
, lr_mask
);
5897 return build2_loc (loc
, wanted_code
, truth_type
, lhs
, rhs
);
5903 /* Handle the case of comparisons with constants. If there is something in
5904 common between the masks, those bits of the constants must be the same.
5905 If not, the condition is always false. Test for this to avoid generating
5906 incorrect code below. */
5907 result
= const_binop (BIT_AND_EXPR
, ll_mask
, rl_mask
);
5908 if (! integer_zerop (result
)
5909 && simple_cst_equal (const_binop (BIT_AND_EXPR
, result
, l_const
),
5910 const_binop (BIT_AND_EXPR
, result
, r_const
)) != 1)
5912 if (wanted_code
== NE_EXPR
)
5914 warning (0, "%<or%> of unmatched not-equal tests is always 1");
5915 return constant_boolean_node (true, truth_type
);
5919 warning (0, "%<and%> of mutually exclusive equal-tests is always 0");
5920 return constant_boolean_node (false, truth_type
);
5924 /* Construct the expression we will return. First get the component
5925 reference we will make. Unless the mask is all ones the width of
5926 that field, perform the mask operation. Then compare with the
5928 result
= make_bit_field_ref (loc
, ll_inner
, lntype
, lnbitsize
, lnbitpos
,
5929 ll_unsignedp
|| rl_unsignedp
, ll_reversep
);
5931 ll_mask
= const_binop (BIT_IOR_EXPR
, ll_mask
, rl_mask
);
5932 if (! all_ones_mask_p (ll_mask
, lnbitsize
))
5933 result
= build2_loc (loc
, BIT_AND_EXPR
, lntype
, result
, ll_mask
);
5935 return build2_loc (loc
, wanted_code
, truth_type
, result
,
5936 const_binop (BIT_IOR_EXPR
, l_const
, r_const
));
5939 /* Optimize T, which is a comparison of a MIN_EXPR or MAX_EXPR with a
5943 optimize_minmax_comparison (location_t loc
, enum tree_code code
, tree type
,
5947 enum tree_code op_code
;
5950 int consts_equal
, consts_lt
;
5953 STRIP_SIGN_NOPS (arg0
);
5955 op_code
= TREE_CODE (arg0
);
5956 minmax_const
= TREE_OPERAND (arg0
, 1);
5957 comp_const
= fold_convert_loc (loc
, TREE_TYPE (arg0
), op1
);
5958 consts_equal
= tree_int_cst_equal (minmax_const
, comp_const
);
5959 consts_lt
= tree_int_cst_lt (minmax_const
, comp_const
);
5960 inner
= TREE_OPERAND (arg0
, 0);
5962 /* If something does not permit us to optimize, return the original tree. */
5963 if ((op_code
!= MIN_EXPR
&& op_code
!= MAX_EXPR
)
5964 || TREE_CODE (comp_const
) != INTEGER_CST
5965 || TREE_OVERFLOW (comp_const
)
5966 || TREE_CODE (minmax_const
) != INTEGER_CST
5967 || TREE_OVERFLOW (minmax_const
))
5970 /* Now handle all the various comparison codes. We only handle EQ_EXPR
5971 and GT_EXPR, doing the rest with recursive calls using logical
5975 case NE_EXPR
: case LT_EXPR
: case LE_EXPR
:
5978 = optimize_minmax_comparison (loc
,
5979 invert_tree_comparison (code
, false),
5982 return invert_truthvalue_loc (loc
, tem
);
5988 fold_build2_loc (loc
, TRUTH_ORIF_EXPR
, type
,
5989 optimize_minmax_comparison
5990 (loc
, EQ_EXPR
, type
, arg0
, comp_const
),
5991 optimize_minmax_comparison
5992 (loc
, GT_EXPR
, type
, arg0
, comp_const
));
5995 if (op_code
== MAX_EXPR
&& consts_equal
)
5996 /* MAX (X, 0) == 0 -> X <= 0 */
5997 return fold_build2_loc (loc
, LE_EXPR
, type
, inner
, comp_const
);
5999 else if (op_code
== MAX_EXPR
&& consts_lt
)
6000 /* MAX (X, 0) == 5 -> X == 5 */
6001 return fold_build2_loc (loc
, EQ_EXPR
, type
, inner
, comp_const
);
6003 else if (op_code
== MAX_EXPR
)
6004 /* MAX (X, 0) == -1 -> false */
6005 return omit_one_operand_loc (loc
, type
, integer_zero_node
, inner
);
6007 else if (consts_equal
)
6008 /* MIN (X, 0) == 0 -> X >= 0 */
6009 return fold_build2_loc (loc
, GE_EXPR
, type
, inner
, comp_const
);
6012 /* MIN (X, 0) == 5 -> false */
6013 return omit_one_operand_loc (loc
, type
, integer_zero_node
, inner
);
6016 /* MIN (X, 0) == -1 -> X == -1 */
6017 return fold_build2_loc (loc
, EQ_EXPR
, type
, inner
, comp_const
);
6020 if (op_code
== MAX_EXPR
&& (consts_equal
|| consts_lt
))
6021 /* MAX (X, 0) > 0 -> X > 0
6022 MAX (X, 0) > 5 -> X > 5 */
6023 return fold_build2_loc (loc
, GT_EXPR
, type
, inner
, comp_const
);
6025 else if (op_code
== MAX_EXPR
)
6026 /* MAX (X, 0) > -1 -> true */
6027 return omit_one_operand_loc (loc
, type
, integer_one_node
, inner
);
6029 else if (op_code
== MIN_EXPR
&& (consts_equal
|| consts_lt
))
6030 /* MIN (X, 0) > 0 -> false
6031 MIN (X, 0) > 5 -> false */
6032 return omit_one_operand_loc (loc
, type
, integer_zero_node
, inner
);
6035 /* MIN (X, 0) > -1 -> X > -1 */
6036 return fold_build2_loc (loc
, GT_EXPR
, type
, inner
, comp_const
);
6043 /* T is an integer expression that is being multiplied, divided, or taken a
6044 modulus (CODE says which and what kind of divide or modulus) by a
6045 constant C. See if we can eliminate that operation by folding it with
6046 other operations already in T. WIDE_TYPE, if non-null, is a type that
6047 should be used for the computation if wider than our type.
6049 For example, if we are dividing (X * 8) + (Y * 16) by 4, we can return
6050 (X * 2) + (Y * 4). We must, however, be assured that either the original
6051 expression would not overflow or that overflow is undefined for the type
6052 in the language in question.
6054 If we return a non-null expression, it is an equivalent form of the
6055 original computation, but need not be in the original type.
6057 We set *STRICT_OVERFLOW_P to true if the return values depends on
6058 signed overflow being undefined. Otherwise we do not change
6059 *STRICT_OVERFLOW_P. */
6062 extract_muldiv (tree t
, tree c
, enum tree_code code
, tree wide_type
,
6063 bool *strict_overflow_p
)
6065 /* To avoid exponential search depth, refuse to allow recursion past
6066 three levels. Beyond that (1) it's highly unlikely that we'll find
6067 something interesting and (2) we've probably processed it before
6068 when we built the inner expression. */
6077 ret
= extract_muldiv_1 (t
, c
, code
, wide_type
, strict_overflow_p
);
6084 extract_muldiv_1 (tree t
, tree c
, enum tree_code code
, tree wide_type
,
6085 bool *strict_overflow_p
)
6087 tree type
= TREE_TYPE (t
);
6088 enum tree_code tcode
= TREE_CODE (t
);
6089 tree ctype
= (wide_type
!= 0 && (GET_MODE_SIZE (TYPE_MODE (wide_type
))
6090 > GET_MODE_SIZE (TYPE_MODE (type
)))
6091 ? wide_type
: type
);
6093 int same_p
= tcode
== code
;
6094 tree op0
= NULL_TREE
, op1
= NULL_TREE
;
6095 bool sub_strict_overflow_p
;
6097 /* Don't deal with constants of zero here; they confuse the code below. */
6098 if (integer_zerop (c
))
6101 if (TREE_CODE_CLASS (tcode
) == tcc_unary
)
6102 op0
= TREE_OPERAND (t
, 0);
6104 if (TREE_CODE_CLASS (tcode
) == tcc_binary
)
6105 op0
= TREE_OPERAND (t
, 0), op1
= TREE_OPERAND (t
, 1);
6107 /* Note that we need not handle conditional operations here since fold
6108 already handles those cases. So just do arithmetic here. */
6112 /* For a constant, we can always simplify if we are a multiply
6113 or (for divide and modulus) if it is a multiple of our constant. */
6114 if (code
== MULT_EXPR
6115 || wi::multiple_of_p (t
, c
, TYPE_SIGN (type
)))
6117 tree tem
= const_binop (code
, fold_convert (ctype
, t
),
6118 fold_convert (ctype
, c
));
6119 /* If the multiplication overflowed, we lost information on it.
6120 See PR68142 and PR69845. */
6121 if (TREE_OVERFLOW (tem
))
6127 CASE_CONVERT
: case NON_LVALUE_EXPR
:
6128 /* If op0 is an expression ... */
6129 if ((COMPARISON_CLASS_P (op0
)
6130 || UNARY_CLASS_P (op0
)
6131 || BINARY_CLASS_P (op0
)
6132 || VL_EXP_CLASS_P (op0
)
6133 || EXPRESSION_CLASS_P (op0
))
6134 /* ... and has wrapping overflow, and its type is smaller
6135 than ctype, then we cannot pass through as widening. */
6136 && (((ANY_INTEGRAL_TYPE_P (TREE_TYPE (op0
))
6137 && TYPE_OVERFLOW_WRAPS (TREE_TYPE (op0
)))
6138 && (TYPE_PRECISION (ctype
)
6139 > TYPE_PRECISION (TREE_TYPE (op0
))))
6140 /* ... or this is a truncation (t is narrower than op0),
6141 then we cannot pass through this narrowing. */
6142 || (TYPE_PRECISION (type
)
6143 < TYPE_PRECISION (TREE_TYPE (op0
)))
6144 /* ... or signedness changes for division or modulus,
6145 then we cannot pass through this conversion. */
6146 || (code
!= MULT_EXPR
6147 && (TYPE_UNSIGNED (ctype
)
6148 != TYPE_UNSIGNED (TREE_TYPE (op0
))))
6149 /* ... or has undefined overflow while the converted to
6150 type has not, we cannot do the operation in the inner type
6151 as that would introduce undefined overflow. */
6152 || ((ANY_INTEGRAL_TYPE_P (TREE_TYPE (op0
))
6153 && TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (op0
)))
6154 && !TYPE_OVERFLOW_UNDEFINED (type
))))
6157 /* Pass the constant down and see if we can make a simplification. If
6158 we can, replace this expression with the inner simplification for
6159 possible later conversion to our or some other type. */
6160 if ((t2
= fold_convert (TREE_TYPE (op0
), c
)) != 0
6161 && TREE_CODE (t2
) == INTEGER_CST
6162 && !TREE_OVERFLOW (t2
)
6163 && (0 != (t1
= extract_muldiv (op0
, t2
, code
,
6165 ? ctype
: NULL_TREE
,
6166 strict_overflow_p
))))
6171 /* If widening the type changes it from signed to unsigned, then we
6172 must avoid building ABS_EXPR itself as unsigned. */
6173 if (TYPE_UNSIGNED (ctype
) && !TYPE_UNSIGNED (type
))
6175 tree cstype
= (*signed_type_for
) (ctype
);
6176 if ((t1
= extract_muldiv (op0
, c
, code
, cstype
, strict_overflow_p
))
6179 t1
= fold_build1 (tcode
, cstype
, fold_convert (cstype
, t1
));
6180 return fold_convert (ctype
, t1
);
6184 /* If the constant is negative, we cannot simplify this. */
6185 if (tree_int_cst_sgn (c
) == -1)
6189 /* For division and modulus, type can't be unsigned, as e.g.
6190 (-(x / 2U)) / 2U isn't equal to -((x / 2U) / 2U) for x >= 2.
6191 For signed types, even with wrapping overflow, this is fine. */
6192 if (code
!= MULT_EXPR
&& TYPE_UNSIGNED (type
))
6194 if ((t1
= extract_muldiv (op0
, c
, code
, wide_type
, strict_overflow_p
))
6196 return fold_build1 (tcode
, ctype
, fold_convert (ctype
, t1
));
6199 case MIN_EXPR
: case MAX_EXPR
:
6200 /* If widening the type changes the signedness, then we can't perform
6201 this optimization as that changes the result. */
6202 if (TYPE_UNSIGNED (ctype
) != TYPE_UNSIGNED (type
))
6205 /* MIN (a, b) / 5 -> MIN (a / 5, b / 5) */
6206 sub_strict_overflow_p
= false;
6207 if ((t1
= extract_muldiv (op0
, c
, code
, wide_type
,
6208 &sub_strict_overflow_p
)) != 0
6209 && (t2
= extract_muldiv (op1
, c
, code
, wide_type
,
6210 &sub_strict_overflow_p
)) != 0)
6212 if (tree_int_cst_sgn (c
) < 0)
6213 tcode
= (tcode
== MIN_EXPR
? MAX_EXPR
: MIN_EXPR
);
6214 if (sub_strict_overflow_p
)
6215 *strict_overflow_p
= true;
6216 return fold_build2 (tcode
, ctype
, fold_convert (ctype
, t1
),
6217 fold_convert (ctype
, t2
));
6221 case LSHIFT_EXPR
: case RSHIFT_EXPR
:
6222 /* If the second operand is constant, this is a multiplication
6223 or floor division, by a power of two, so we can treat it that
6224 way unless the multiplier or divisor overflows. Signed
6225 left-shift overflow is implementation-defined rather than
6226 undefined in C90, so do not convert signed left shift into
6228 if (TREE_CODE (op1
) == INTEGER_CST
6229 && (tcode
== RSHIFT_EXPR
|| TYPE_UNSIGNED (TREE_TYPE (op0
)))
6230 /* const_binop may not detect overflow correctly,
6231 so check for it explicitly here. */
6232 && wi::gtu_p (TYPE_PRECISION (TREE_TYPE (size_one_node
)), op1
)
6233 && 0 != (t1
= fold_convert (ctype
,
6234 const_binop (LSHIFT_EXPR
,
6237 && !TREE_OVERFLOW (t1
))
6238 return extract_muldiv (build2 (tcode
== LSHIFT_EXPR
6239 ? MULT_EXPR
: FLOOR_DIV_EXPR
,
6241 fold_convert (ctype
, op0
),
6243 c
, code
, wide_type
, strict_overflow_p
);
6246 case PLUS_EXPR
: case MINUS_EXPR
:
6247 /* See if we can eliminate the operation on both sides. If we can, we
6248 can return a new PLUS or MINUS. If we can't, the only remaining
6249 cases where we can do anything are if the second operand is a
6251 sub_strict_overflow_p
= false;
6252 t1
= extract_muldiv (op0
, c
, code
, wide_type
, &sub_strict_overflow_p
);
6253 t2
= extract_muldiv (op1
, c
, code
, wide_type
, &sub_strict_overflow_p
);
6254 if (t1
!= 0 && t2
!= 0
6255 && (code
== MULT_EXPR
6256 /* If not multiplication, we can only do this if both operands
6257 are divisible by c. */
6258 || (multiple_of_p (ctype
, op0
, c
)
6259 && multiple_of_p (ctype
, op1
, c
))))
6261 if (sub_strict_overflow_p
)
6262 *strict_overflow_p
= true;
6263 return fold_build2 (tcode
, ctype
, fold_convert (ctype
, t1
),
6264 fold_convert (ctype
, t2
));
6267 /* If this was a subtraction, negate OP1 and set it to be an addition.
6268 This simplifies the logic below. */
6269 if (tcode
== MINUS_EXPR
)
6271 tcode
= PLUS_EXPR
, op1
= negate_expr (op1
);
6272 /* If OP1 was not easily negatable, the constant may be OP0. */
6273 if (TREE_CODE (op0
) == INTEGER_CST
)
6275 std::swap (op0
, op1
);
6280 if (TREE_CODE (op1
) != INTEGER_CST
)
6283 /* If either OP1 or C are negative, this optimization is not safe for
6284 some of the division and remainder types while for others we need
6285 to change the code. */
6286 if (tree_int_cst_sgn (op1
) < 0 || tree_int_cst_sgn (c
) < 0)
6288 if (code
== CEIL_DIV_EXPR
)
6289 code
= FLOOR_DIV_EXPR
;
6290 else if (code
== FLOOR_DIV_EXPR
)
6291 code
= CEIL_DIV_EXPR
;
6292 else if (code
!= MULT_EXPR
6293 && code
!= CEIL_MOD_EXPR
&& code
!= FLOOR_MOD_EXPR
)
6297 /* If it's a multiply or a division/modulus operation of a multiple
6298 of our constant, do the operation and verify it doesn't overflow. */
6299 if (code
== MULT_EXPR
6300 || wi::multiple_of_p (op1
, c
, TYPE_SIGN (type
)))
6302 op1
= const_binop (code
, fold_convert (ctype
, op1
),
6303 fold_convert (ctype
, c
));
6304 /* We allow the constant to overflow with wrapping semantics. */
6306 || (TREE_OVERFLOW (op1
) && !TYPE_OVERFLOW_WRAPS (ctype
)))
6312 /* If we have an unsigned type, we cannot widen the operation since it
6313 will change the result if the original computation overflowed. */
6314 if (TYPE_UNSIGNED (ctype
) && ctype
!= type
)
6317 /* If we were able to eliminate our operation from the first side,
6318 apply our operation to the second side and reform the PLUS. */
6319 if (t1
!= 0 && (TREE_CODE (t1
) != code
|| code
== MULT_EXPR
))
6320 return fold_build2 (tcode
, ctype
, fold_convert (ctype
, t1
), op1
);
6322 /* The last case is if we are a multiply. In that case, we can
6323 apply the distributive law to commute the multiply and addition
6324 if the multiplication of the constants doesn't overflow
6325 and overflow is defined. With undefined overflow
6326 op0 * c might overflow, while (op0 + orig_op1) * c doesn't. */
6327 if (code
== MULT_EXPR
&& TYPE_OVERFLOW_WRAPS (ctype
))
6328 return fold_build2 (tcode
, ctype
,
6329 fold_build2 (code
, ctype
,
6330 fold_convert (ctype
, op0
),
6331 fold_convert (ctype
, c
)),
6337 /* We have a special case here if we are doing something like
6338 (C * 8) % 4 since we know that's zero. */
6339 if ((code
== TRUNC_MOD_EXPR
|| code
== CEIL_MOD_EXPR
6340 || code
== FLOOR_MOD_EXPR
|| code
== ROUND_MOD_EXPR
)
6341 /* If the multiplication can overflow we cannot optimize this. */
6342 && TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (t
))
6343 && TREE_CODE (TREE_OPERAND (t
, 1)) == INTEGER_CST
6344 && wi::multiple_of_p (op1
, c
, TYPE_SIGN (type
)))
6346 *strict_overflow_p
= true;
6347 return omit_one_operand (type
, integer_zero_node
, op0
);
6350 /* ... fall through ... */
6352 case TRUNC_DIV_EXPR
: case CEIL_DIV_EXPR
: case FLOOR_DIV_EXPR
:
6353 case ROUND_DIV_EXPR
: case EXACT_DIV_EXPR
:
6354 /* If we can extract our operation from the LHS, do so and return a
6355 new operation. Likewise for the RHS from a MULT_EXPR. Otherwise,
6356 do something only if the second operand is a constant. */
6358 && (t1
= extract_muldiv (op0
, c
, code
, wide_type
,
6359 strict_overflow_p
)) != 0)
6360 return fold_build2 (tcode
, ctype
, fold_convert (ctype
, t1
),
6361 fold_convert (ctype
, op1
));
6362 else if (tcode
== MULT_EXPR
&& code
== MULT_EXPR
6363 && (t1
= extract_muldiv (op1
, c
, code
, wide_type
,
6364 strict_overflow_p
)) != 0)
6365 return fold_build2 (tcode
, ctype
, fold_convert (ctype
, op0
),
6366 fold_convert (ctype
, t1
));
6367 else if (TREE_CODE (op1
) != INTEGER_CST
)
6370 /* If these are the same operation types, we can associate them
6371 assuming no overflow. */
6374 bool overflow_p
= false;
6375 bool overflow_mul_p
;
6376 signop sign
= TYPE_SIGN (ctype
);
6377 unsigned prec
= TYPE_PRECISION (ctype
);
6378 wide_int mul
= wi::mul (wide_int::from (op1
, prec
,
6379 TYPE_SIGN (TREE_TYPE (op1
))),
6380 wide_int::from (c
, prec
,
6381 TYPE_SIGN (TREE_TYPE (c
))),
6382 sign
, &overflow_mul_p
);
6383 overflow_p
= TREE_OVERFLOW (c
) | TREE_OVERFLOW (op1
);
6385 && ((sign
== UNSIGNED
&& tcode
!= MULT_EXPR
) || sign
== SIGNED
))
6388 return fold_build2 (tcode
, ctype
, fold_convert (ctype
, op0
),
6389 wide_int_to_tree (ctype
, mul
));
6392 /* If these operations "cancel" each other, we have the main
6393 optimizations of this pass, which occur when either constant is a
6394 multiple of the other, in which case we replace this with either an
6395 operation or CODE or TCODE.
6397 If we have an unsigned type, we cannot do this since it will change
6398 the result if the original computation overflowed. */
6399 if (TYPE_OVERFLOW_UNDEFINED (ctype
)
6400 && ((code
== MULT_EXPR
&& tcode
== EXACT_DIV_EXPR
)
6401 || (tcode
== MULT_EXPR
6402 && code
!= TRUNC_MOD_EXPR
&& code
!= CEIL_MOD_EXPR
6403 && code
!= FLOOR_MOD_EXPR
&& code
!= ROUND_MOD_EXPR
6404 && code
!= MULT_EXPR
)))
6406 if (wi::multiple_of_p (op1
, c
, TYPE_SIGN (type
)))
6408 if (TYPE_OVERFLOW_UNDEFINED (ctype
))
6409 *strict_overflow_p
= true;
6410 return fold_build2 (tcode
, ctype
, fold_convert (ctype
, op0
),
6411 fold_convert (ctype
,
6412 const_binop (TRUNC_DIV_EXPR
,
6415 else if (wi::multiple_of_p (c
, op1
, TYPE_SIGN (type
)))
6417 if (TYPE_OVERFLOW_UNDEFINED (ctype
))
6418 *strict_overflow_p
= true;
6419 return fold_build2 (code
, ctype
, fold_convert (ctype
, op0
),
6420 fold_convert (ctype
,
6421 const_binop (TRUNC_DIV_EXPR
,
6434 /* Return a node which has the indicated constant VALUE (either 0 or
6435 1 for scalars or {-1,-1,..} or {0,0,...} for vectors),
6436 and is of the indicated TYPE. */
6439 constant_boolean_node (bool value
, tree type
)
6441 if (type
== integer_type_node
)
6442 return value
? integer_one_node
: integer_zero_node
;
6443 else if (type
== boolean_type_node
)
6444 return value
? boolean_true_node
: boolean_false_node
;
6445 else if (TREE_CODE (type
) == VECTOR_TYPE
)
6446 return build_vector_from_val (type
,
6447 build_int_cst (TREE_TYPE (type
),
6450 return fold_convert (type
, value
? integer_one_node
: integer_zero_node
);
6454 /* Transform `a + (b ? x : y)' into `b ? (a + x) : (a + y)'.
6455 Transform, `a + (x < y)' into `(x < y) ? (a + 1) : (a + 0)'. Here
6456 CODE corresponds to the `+', COND to the `(b ? x : y)' or `(x < y)'
6457 expression, and ARG to `a'. If COND_FIRST_P is nonzero, then the
6458 COND is the first argument to CODE; otherwise (as in the example
6459 given here), it is the second argument. TYPE is the type of the
6460 original expression. Return NULL_TREE if no simplification is
6464 fold_binary_op_with_conditional_arg (location_t loc
,
6465 enum tree_code code
,
6466 tree type
, tree op0
, tree op1
,
6467 tree cond
, tree arg
, int cond_first_p
)
6469 tree cond_type
= cond_first_p
? TREE_TYPE (op0
) : TREE_TYPE (op1
);
6470 tree arg_type
= cond_first_p
? TREE_TYPE (op1
) : TREE_TYPE (op0
);
6471 tree test
, true_value
, false_value
;
6472 tree lhs
= NULL_TREE
;
6473 tree rhs
= NULL_TREE
;
6474 enum tree_code cond_code
= COND_EXPR
;
6476 if (TREE_CODE (cond
) == COND_EXPR
6477 || TREE_CODE (cond
) == VEC_COND_EXPR
)
6479 test
= TREE_OPERAND (cond
, 0);
6480 true_value
= TREE_OPERAND (cond
, 1);
6481 false_value
= TREE_OPERAND (cond
, 2);
6482 /* If this operand throws an expression, then it does not make
6483 sense to try to perform a logical or arithmetic operation
6485 if (VOID_TYPE_P (TREE_TYPE (true_value
)))
6487 if (VOID_TYPE_P (TREE_TYPE (false_value
)))
6490 else if (!(TREE_CODE (type
) != VECTOR_TYPE
6491 && TREE_CODE (TREE_TYPE (cond
)) == VECTOR_TYPE
))
6493 tree testtype
= TREE_TYPE (cond
);
6495 true_value
= constant_boolean_node (true, testtype
);
6496 false_value
= constant_boolean_node (false, testtype
);
6499 /* Detect the case of mixing vector and scalar types - bail out. */
6502 if (TREE_CODE (TREE_TYPE (test
)) == VECTOR_TYPE
)
6503 cond_code
= VEC_COND_EXPR
;
6505 /* This transformation is only worthwhile if we don't have to wrap ARG
6506 in a SAVE_EXPR and the operation can be simplified without recursing
6507 on at least one of the branches once its pushed inside the COND_EXPR. */
6508 if (!TREE_CONSTANT (arg
)
6509 && (TREE_SIDE_EFFECTS (arg
)
6510 || TREE_CODE (arg
) == COND_EXPR
|| TREE_CODE (arg
) == VEC_COND_EXPR
6511 || TREE_CONSTANT (true_value
) || TREE_CONSTANT (false_value
)))
6514 arg
= fold_convert_loc (loc
, arg_type
, arg
);
6517 true_value
= fold_convert_loc (loc
, cond_type
, true_value
);
6519 lhs
= fold_build2_loc (loc
, code
, type
, true_value
, arg
);
6521 lhs
= fold_build2_loc (loc
, code
, type
, arg
, true_value
);
6525 false_value
= fold_convert_loc (loc
, cond_type
, false_value
);
6527 rhs
= fold_build2_loc (loc
, code
, type
, false_value
, arg
);
6529 rhs
= fold_build2_loc (loc
, code
, type
, arg
, false_value
);
6532 /* Check that we have simplified at least one of the branches. */
6533 if (!TREE_CONSTANT (arg
) && !TREE_CONSTANT (lhs
) && !TREE_CONSTANT (rhs
))
6536 return fold_build3_loc (loc
, cond_code
, type
, test
, lhs
, rhs
);
6540 /* Subroutine of fold() that checks for the addition of +/- 0.0.
6542 If !NEGATE, return true if ADDEND is +/-0.0 and, for all X of type
6543 TYPE, X + ADDEND is the same as X. If NEGATE, return true if X -
6544 ADDEND is the same as X.
6546 X + 0 and X - 0 both give X when X is NaN, infinite, or nonzero
6547 and finite. The problematic cases are when X is zero, and its mode
6548 has signed zeros. In the case of rounding towards -infinity,
6549 X - 0 is not the same as X because 0 - 0 is -0. In other rounding
6550 modes, X + 0 is not the same as X because -0 + 0 is 0. */
6553 fold_real_zero_addition_p (const_tree type
, const_tree addend
, int negate
)
6555 if (!real_zerop (addend
))
6558 /* Don't allow the fold with -fsignaling-nans. */
6559 if (HONOR_SNANS (element_mode (type
)))
6562 /* Allow the fold if zeros aren't signed, or their sign isn't important. */
6563 if (!HONOR_SIGNED_ZEROS (element_mode (type
)))
6566 /* In a vector or complex, we would need to check the sign of all zeros. */
6567 if (TREE_CODE (addend
) != REAL_CST
)
6570 /* Treat x + -0 as x - 0 and x - -0 as x + 0. */
6571 if (REAL_VALUE_MINUS_ZERO (TREE_REAL_CST (addend
)))
6574 /* The mode has signed zeros, and we have to honor their sign.
6575 In this situation, there is only one case we can return true for.
6576 X - 0 is the same as X unless rounding towards -infinity is
6578 return negate
&& !HONOR_SIGN_DEPENDENT_ROUNDING (element_mode (type
));
6581 /* Subroutine of fold() that optimizes comparisons of a division by
6582 a nonzero integer constant against an integer constant, i.e.
6585 CODE is the comparison operator: EQ_EXPR, NE_EXPR, GT_EXPR, LT_EXPR,
6586 GE_EXPR or LE_EXPR. TYPE is the type of the result and ARG0 and ARG1
6587 are the operands of the comparison. ARG1 must be a TREE_REAL_CST.
6589 The function returns the constant folded tree if a simplification
6590 can be made, and NULL_TREE otherwise. */
6593 fold_div_compare (location_t loc
,
6594 enum tree_code code
, tree type
, tree arg0
, tree arg1
)
6596 tree prod
, tmp
, hi
, lo
;
6597 tree arg00
= TREE_OPERAND (arg0
, 0);
6598 tree arg01
= TREE_OPERAND (arg0
, 1);
6599 signop sign
= TYPE_SIGN (TREE_TYPE (arg0
));
6600 bool neg_overflow
= false;
6603 /* We have to do this the hard way to detect unsigned overflow.
6604 prod = int_const_binop (MULT_EXPR, arg01, arg1); */
6605 wide_int val
= wi::mul (arg01
, arg1
, sign
, &overflow
);
6606 prod
= force_fit_type (TREE_TYPE (arg00
), val
, -1, overflow
);
6607 neg_overflow
= false;
6609 if (sign
== UNSIGNED
)
6611 tmp
= int_const_binop (MINUS_EXPR
, arg01
,
6612 build_int_cst (TREE_TYPE (arg01
), 1));
6615 /* Likewise hi = int_const_binop (PLUS_EXPR, prod, tmp). */
6616 val
= wi::add (prod
, tmp
, sign
, &overflow
);
6617 hi
= force_fit_type (TREE_TYPE (arg00
), val
,
6618 -1, overflow
| TREE_OVERFLOW (prod
));
6620 else if (tree_int_cst_sgn (arg01
) >= 0)
6622 tmp
= int_const_binop (MINUS_EXPR
, arg01
,
6623 build_int_cst (TREE_TYPE (arg01
), 1));
6624 switch (tree_int_cst_sgn (arg1
))
6627 neg_overflow
= true;
6628 lo
= int_const_binop (MINUS_EXPR
, prod
, tmp
);
6633 lo
= fold_negate_const (tmp
, TREE_TYPE (arg0
));
6638 hi
= int_const_binop (PLUS_EXPR
, prod
, tmp
);
6648 /* A negative divisor reverses the relational operators. */
6649 code
= swap_tree_comparison (code
);
6651 tmp
= int_const_binop (PLUS_EXPR
, arg01
,
6652 build_int_cst (TREE_TYPE (arg01
), 1));
6653 switch (tree_int_cst_sgn (arg1
))
6656 hi
= int_const_binop (MINUS_EXPR
, prod
, tmp
);
6661 hi
= fold_negate_const (tmp
, TREE_TYPE (arg0
));
6666 neg_overflow
= true;
6667 lo
= int_const_binop (PLUS_EXPR
, prod
, tmp
);
6679 if (TREE_OVERFLOW (lo
) && TREE_OVERFLOW (hi
))
6680 return omit_one_operand_loc (loc
, type
, integer_zero_node
, arg00
);
6681 if (TREE_OVERFLOW (hi
))
6682 return fold_build2_loc (loc
, GE_EXPR
, type
, arg00
, lo
);
6683 if (TREE_OVERFLOW (lo
))
6684 return fold_build2_loc (loc
, LE_EXPR
, type
, arg00
, hi
);
6685 return build_range_check (loc
, type
, arg00
, 1, lo
, hi
);
6688 if (TREE_OVERFLOW (lo
) && TREE_OVERFLOW (hi
))
6689 return omit_one_operand_loc (loc
, type
, integer_one_node
, arg00
);
6690 if (TREE_OVERFLOW (hi
))
6691 return fold_build2_loc (loc
, LT_EXPR
, type
, arg00
, lo
);
6692 if (TREE_OVERFLOW (lo
))
6693 return fold_build2_loc (loc
, GT_EXPR
, type
, arg00
, hi
);
6694 return build_range_check (loc
, type
, arg00
, 0, lo
, hi
);
6697 if (TREE_OVERFLOW (lo
))
6699 tmp
= neg_overflow
? integer_zero_node
: integer_one_node
;
6700 return omit_one_operand_loc (loc
, type
, tmp
, arg00
);
6702 return fold_build2_loc (loc
, LT_EXPR
, type
, arg00
, lo
);
6705 if (TREE_OVERFLOW (hi
))
6707 tmp
= neg_overflow
? integer_zero_node
: integer_one_node
;
6708 return omit_one_operand_loc (loc
, type
, tmp
, arg00
);
6710 return fold_build2_loc (loc
, LE_EXPR
, type
, arg00
, hi
);
6713 if (TREE_OVERFLOW (hi
))
6715 tmp
= neg_overflow
? integer_one_node
: integer_zero_node
;
6716 return omit_one_operand_loc (loc
, type
, tmp
, arg00
);
6718 return fold_build2_loc (loc
, GT_EXPR
, type
, arg00
, hi
);
6721 if (TREE_OVERFLOW (lo
))
6723 tmp
= neg_overflow
? integer_one_node
: integer_zero_node
;
6724 return omit_one_operand_loc (loc
, type
, tmp
, arg00
);
6726 return fold_build2_loc (loc
, GE_EXPR
, type
, arg00
, lo
);
6736 /* If CODE with arguments ARG0 and ARG1 represents a single bit
6737 equality/inequality test, then return a simplified form of the test
6738 using a sign testing. Otherwise return NULL. TYPE is the desired
6742 fold_single_bit_test_into_sign_test (location_t loc
,
6743 enum tree_code code
, tree arg0
, tree arg1
,
6746 /* If this is testing a single bit, we can optimize the test. */
6747 if ((code
== NE_EXPR
|| code
== EQ_EXPR
)
6748 && TREE_CODE (arg0
) == BIT_AND_EXPR
&& integer_zerop (arg1
)
6749 && integer_pow2p (TREE_OPERAND (arg0
, 1)))
6751 /* If we have (A & C) != 0 where C is the sign bit of A, convert
6752 this into A < 0. Similarly for (A & C) == 0 into A >= 0. */
6753 tree arg00
= sign_bit_p (TREE_OPERAND (arg0
, 0), TREE_OPERAND (arg0
, 1));
6755 if (arg00
!= NULL_TREE
6756 /* This is only a win if casting to a signed type is cheap,
6757 i.e. when arg00's type is not a partial mode. */
6758 && TYPE_PRECISION (TREE_TYPE (arg00
))
6759 == GET_MODE_PRECISION (TYPE_MODE (TREE_TYPE (arg00
))))
6761 tree stype
= signed_type_for (TREE_TYPE (arg00
));
6762 return fold_build2_loc (loc
, code
== EQ_EXPR
? GE_EXPR
: LT_EXPR
,
6764 fold_convert_loc (loc
, stype
, arg00
),
6765 build_int_cst (stype
, 0));
6772 /* If CODE with arguments ARG0 and ARG1 represents a single bit
6773 equality/inequality test, then return a simplified form of
6774 the test using shifts and logical operations. Otherwise return
6775 NULL. TYPE is the desired result type. */
6778 fold_single_bit_test (location_t loc
, enum tree_code code
,
6779 tree arg0
, tree arg1
, tree result_type
)
6781 /* If this is testing a single bit, we can optimize the test. */
6782 if ((code
== NE_EXPR
|| code
== EQ_EXPR
)
6783 && TREE_CODE (arg0
) == BIT_AND_EXPR
&& integer_zerop (arg1
)
6784 && integer_pow2p (TREE_OPERAND (arg0
, 1)))
6786 tree inner
= TREE_OPERAND (arg0
, 0);
6787 tree type
= TREE_TYPE (arg0
);
6788 int bitnum
= tree_log2 (TREE_OPERAND (arg0
, 1));
6789 machine_mode operand_mode
= TYPE_MODE (type
);
6791 tree signed_type
, unsigned_type
, intermediate_type
;
6794 /* First, see if we can fold the single bit test into a sign-bit
6796 tem
= fold_single_bit_test_into_sign_test (loc
, code
, arg0
, arg1
,
6801 /* Otherwise we have (A & C) != 0 where C is a single bit,
6802 convert that into ((A >> C2) & 1). Where C2 = log2(C).
6803 Similarly for (A & C) == 0. */
6805 /* If INNER is a right shift of a constant and it plus BITNUM does
6806 not overflow, adjust BITNUM and INNER. */
6807 if (TREE_CODE (inner
) == RSHIFT_EXPR
6808 && TREE_CODE (TREE_OPERAND (inner
, 1)) == INTEGER_CST
6809 && bitnum
< TYPE_PRECISION (type
)
6810 && wi::ltu_p (TREE_OPERAND (inner
, 1),
6811 TYPE_PRECISION (type
) - bitnum
))
6813 bitnum
+= tree_to_uhwi (TREE_OPERAND (inner
, 1));
6814 inner
= TREE_OPERAND (inner
, 0);
6817 /* If we are going to be able to omit the AND below, we must do our
6818 operations as unsigned. If we must use the AND, we have a choice.
6819 Normally unsigned is faster, but for some machines signed is. */
6820 ops_unsigned
= (LOAD_EXTEND_OP (operand_mode
) == SIGN_EXTEND
6821 && !flag_syntax_only
) ? 0 : 1;
6823 signed_type
= lang_hooks
.types
.type_for_mode (operand_mode
, 0);
6824 unsigned_type
= lang_hooks
.types
.type_for_mode (operand_mode
, 1);
6825 intermediate_type
= ops_unsigned
? unsigned_type
: signed_type
;
6826 inner
= fold_convert_loc (loc
, intermediate_type
, inner
);
6829 inner
= build2 (RSHIFT_EXPR
, intermediate_type
,
6830 inner
, size_int (bitnum
));
6832 one
= build_int_cst (intermediate_type
, 1);
6834 if (code
== EQ_EXPR
)
6835 inner
= fold_build2_loc (loc
, BIT_XOR_EXPR
, intermediate_type
, inner
, one
);
6837 /* Put the AND last so it can combine with more things. */
6838 inner
= build2 (BIT_AND_EXPR
, intermediate_type
, inner
, one
);
6840 /* Make sure to return the proper type. */
6841 inner
= fold_convert_loc (loc
, result_type
, inner
);
6848 /* Check whether we are allowed to reorder operands arg0 and arg1,
6849 such that the evaluation of arg1 occurs before arg0. */
6852 reorder_operands_p (const_tree arg0
, const_tree arg1
)
6854 if (! flag_evaluation_order
)
6856 if (TREE_CONSTANT (arg0
) || TREE_CONSTANT (arg1
))
6858 return ! TREE_SIDE_EFFECTS (arg0
)
6859 && ! TREE_SIDE_EFFECTS (arg1
);
6862 /* Test whether it is preferable two swap two operands, ARG0 and
6863 ARG1, for example because ARG0 is an integer constant and ARG1
6864 isn't. If REORDER is true, only recommend swapping if we can
6865 evaluate the operands in reverse order. */
6868 tree_swap_operands_p (const_tree arg0
, const_tree arg1
, bool reorder
)
6870 if (CONSTANT_CLASS_P (arg1
))
6872 if (CONSTANT_CLASS_P (arg0
))
6878 if (TREE_CONSTANT (arg1
))
6880 if (TREE_CONSTANT (arg0
))
6883 if (reorder
&& flag_evaluation_order
6884 && (TREE_SIDE_EFFECTS (arg0
) || TREE_SIDE_EFFECTS (arg1
)))
6887 /* It is preferable to swap two SSA_NAME to ensure a canonical form
6888 for commutative and comparison operators. Ensuring a canonical
6889 form allows the optimizers to find additional redundancies without
6890 having to explicitly check for both orderings. */
6891 if (TREE_CODE (arg0
) == SSA_NAME
6892 && TREE_CODE (arg1
) == SSA_NAME
6893 && SSA_NAME_VERSION (arg0
) > SSA_NAME_VERSION (arg1
))
6896 /* Put SSA_NAMEs last. */
6897 if (TREE_CODE (arg1
) == SSA_NAME
)
6899 if (TREE_CODE (arg0
) == SSA_NAME
)
6902 /* Put variables last. */
6912 /* Fold A < X && A + 1 > Y to A < X && A >= Y. Normally A + 1 > Y
6913 means A >= Y && A != MAX, but in this case we know that
6914 A < X <= MAX. INEQ is A + 1 > Y, BOUND is A < X. */
6917 fold_to_nonsharp_ineq_using_bound (location_t loc
, tree ineq
, tree bound
)
6919 tree a
, typea
, type
= TREE_TYPE (ineq
), a1
, diff
, y
;
6921 if (TREE_CODE (bound
) == LT_EXPR
)
6922 a
= TREE_OPERAND (bound
, 0);
6923 else if (TREE_CODE (bound
) == GT_EXPR
)
6924 a
= TREE_OPERAND (bound
, 1);
6928 typea
= TREE_TYPE (a
);
6929 if (!INTEGRAL_TYPE_P (typea
)
6930 && !POINTER_TYPE_P (typea
))
6933 if (TREE_CODE (ineq
) == LT_EXPR
)
6935 a1
= TREE_OPERAND (ineq
, 1);
6936 y
= TREE_OPERAND (ineq
, 0);
6938 else if (TREE_CODE (ineq
) == GT_EXPR
)
6940 a1
= TREE_OPERAND (ineq
, 0);
6941 y
= TREE_OPERAND (ineq
, 1);
6946 if (TREE_TYPE (a1
) != typea
)
6949 if (POINTER_TYPE_P (typea
))
6951 /* Convert the pointer types into integer before taking the difference. */
6952 tree ta
= fold_convert_loc (loc
, ssizetype
, a
);
6953 tree ta1
= fold_convert_loc (loc
, ssizetype
, a1
);
6954 diff
= fold_binary_loc (loc
, MINUS_EXPR
, ssizetype
, ta1
, ta
);
6957 diff
= fold_binary_loc (loc
, MINUS_EXPR
, typea
, a1
, a
);
6959 if (!diff
|| !integer_onep (diff
))
6962 return fold_build2_loc (loc
, GE_EXPR
, type
, a
, y
);
6965 /* Fold a sum or difference of at least one multiplication.
6966 Returns the folded tree or NULL if no simplification could be made. */
6969 fold_plusminus_mult_expr (location_t loc
, enum tree_code code
, tree type
,
6970 tree arg0
, tree arg1
)
6972 tree arg00
, arg01
, arg10
, arg11
;
6973 tree alt0
= NULL_TREE
, alt1
= NULL_TREE
, same
;
6975 /* (A * C) +- (B * C) -> (A+-B) * C.
6976 (A * C) +- A -> A * (C+-1).
6977 We are most concerned about the case where C is a constant,
6978 but other combinations show up during loop reduction. Since
6979 it is not difficult, try all four possibilities. */
6981 if (TREE_CODE (arg0
) == MULT_EXPR
)
6983 arg00
= TREE_OPERAND (arg0
, 0);
6984 arg01
= TREE_OPERAND (arg0
, 1);
6986 else if (TREE_CODE (arg0
) == INTEGER_CST
)
6988 arg00
= build_one_cst (type
);
6993 /* We cannot generate constant 1 for fract. */
6994 if (ALL_FRACT_MODE_P (TYPE_MODE (type
)))
6997 arg01
= build_one_cst (type
);
6999 if (TREE_CODE (arg1
) == MULT_EXPR
)
7001 arg10
= TREE_OPERAND (arg1
, 0);
7002 arg11
= TREE_OPERAND (arg1
, 1);
7004 else if (TREE_CODE (arg1
) == INTEGER_CST
)
7006 arg10
= build_one_cst (type
);
7007 /* As we canonicalize A - 2 to A + -2 get rid of that sign for
7008 the purpose of this canonicalization. */
7009 if (wi::neg_p (arg1
, TYPE_SIGN (TREE_TYPE (arg1
)))
7010 && negate_expr_p (arg1
)
7011 && code
== PLUS_EXPR
)
7013 arg11
= negate_expr (arg1
);
7021 /* We cannot generate constant 1 for fract. */
7022 if (ALL_FRACT_MODE_P (TYPE_MODE (type
)))
7025 arg11
= build_one_cst (type
);
7029 if (operand_equal_p (arg01
, arg11
, 0))
7030 same
= arg01
, alt0
= arg00
, alt1
= arg10
;
7031 else if (operand_equal_p (arg00
, arg10
, 0))
7032 same
= arg00
, alt0
= arg01
, alt1
= arg11
;
7033 else if (operand_equal_p (arg00
, arg11
, 0))
7034 same
= arg00
, alt0
= arg01
, alt1
= arg10
;
7035 else if (operand_equal_p (arg01
, arg10
, 0))
7036 same
= arg01
, alt0
= arg00
, alt1
= arg11
;
7038 /* No identical multiplicands; see if we can find a common
7039 power-of-two factor in non-power-of-two multiplies. This
7040 can help in multi-dimensional array access. */
7041 else if (tree_fits_shwi_p (arg01
)
7042 && tree_fits_shwi_p (arg11
))
7044 HOST_WIDE_INT int01
, int11
, tmp
;
7047 int01
= tree_to_shwi (arg01
);
7048 int11
= tree_to_shwi (arg11
);
7050 /* Move min of absolute values to int11. */
7051 if (absu_hwi (int01
) < absu_hwi (int11
))
7053 tmp
= int01
, int01
= int11
, int11
= tmp
;
7054 alt0
= arg00
, arg00
= arg10
, arg10
= alt0
;
7061 if (exact_log2 (absu_hwi (int11
)) > 0 && int01
% int11
== 0
7062 /* The remainder should not be a constant, otherwise we
7063 end up folding i * 4 + 2 to (i * 2 + 1) * 2 which has
7064 increased the number of multiplications necessary. */
7065 && TREE_CODE (arg10
) != INTEGER_CST
)
7067 alt0
= fold_build2_loc (loc
, MULT_EXPR
, TREE_TYPE (arg00
), arg00
,
7068 build_int_cst (TREE_TYPE (arg00
),
7073 maybe_same
= alt0
, alt0
= alt1
, alt1
= maybe_same
;
7078 return fold_build2_loc (loc
, MULT_EXPR
, type
,
7079 fold_build2_loc (loc
, code
, type
,
7080 fold_convert_loc (loc
, type
, alt0
),
7081 fold_convert_loc (loc
, type
, alt1
)),
7082 fold_convert_loc (loc
, type
, same
));
7087 /* Subroutine of native_encode_expr. Encode the INTEGER_CST
7088 specified by EXPR into the buffer PTR of length LEN bytes.
7089 Return the number of bytes placed in the buffer, or zero
7093 native_encode_int (const_tree expr
, unsigned char *ptr
, int len
, int off
)
7095 tree type
= TREE_TYPE (expr
);
7096 int total_bytes
= GET_MODE_SIZE (TYPE_MODE (type
));
7097 int byte
, offset
, word
, words
;
7098 unsigned char value
;
7100 if ((off
== -1 && total_bytes
> len
)
7101 || off
>= total_bytes
)
7105 words
= total_bytes
/ UNITS_PER_WORD
;
7107 for (byte
= 0; byte
< total_bytes
; byte
++)
7109 int bitpos
= byte
* BITS_PER_UNIT
;
7110 /* Extend EXPR according to TYPE_SIGN if the precision isn't a whole
7112 value
= wi::extract_uhwi (wi::to_widest (expr
), bitpos
, BITS_PER_UNIT
);
7114 if (total_bytes
> UNITS_PER_WORD
)
7116 word
= byte
/ UNITS_PER_WORD
;
7117 if (WORDS_BIG_ENDIAN
)
7118 word
= (words
- 1) - word
;
7119 offset
= word
* UNITS_PER_WORD
;
7120 if (BYTES_BIG_ENDIAN
)
7121 offset
+= (UNITS_PER_WORD
- 1) - (byte
% UNITS_PER_WORD
);
7123 offset
+= byte
% UNITS_PER_WORD
;
7126 offset
= BYTES_BIG_ENDIAN
? (total_bytes
- 1) - byte
: byte
;
7128 && offset
- off
< len
)
7129 ptr
[offset
- off
] = value
;
7131 return MIN (len
, total_bytes
- off
);
7135 /* Subroutine of native_encode_expr. Encode the FIXED_CST
7136 specified by EXPR into the buffer PTR of length LEN bytes.
7137 Return the number of bytes placed in the buffer, or zero
7141 native_encode_fixed (const_tree expr
, unsigned char *ptr
, int len
, int off
)
7143 tree type
= TREE_TYPE (expr
);
7144 machine_mode mode
= TYPE_MODE (type
);
7145 int total_bytes
= GET_MODE_SIZE (mode
);
7146 FIXED_VALUE_TYPE value
;
7147 tree i_value
, i_type
;
7149 if (total_bytes
* BITS_PER_UNIT
> HOST_BITS_PER_DOUBLE_INT
)
7152 i_type
= lang_hooks
.types
.type_for_size (GET_MODE_BITSIZE (mode
), 1);
7154 if (NULL_TREE
== i_type
7155 || TYPE_PRECISION (i_type
) != total_bytes
)
7158 value
= TREE_FIXED_CST (expr
);
7159 i_value
= double_int_to_tree (i_type
, value
.data
);
7161 return native_encode_int (i_value
, ptr
, len
, off
);
7165 /* Subroutine of native_encode_expr. Encode the REAL_CST
7166 specified by EXPR into the buffer PTR of length LEN bytes.
7167 Return the number of bytes placed in the buffer, or zero
7171 native_encode_real (const_tree expr
, unsigned char *ptr
, int len
, int off
)
7173 tree type
= TREE_TYPE (expr
);
7174 int total_bytes
= GET_MODE_SIZE (TYPE_MODE (type
));
7175 int byte
, offset
, word
, words
, bitpos
;
7176 unsigned char value
;
7178 /* There are always 32 bits in each long, no matter the size of
7179 the hosts long. We handle floating point representations with
7183 if ((off
== -1 && total_bytes
> len
)
7184 || off
>= total_bytes
)
7188 words
= (32 / BITS_PER_UNIT
) / UNITS_PER_WORD
;
7190 real_to_target (tmp
, TREE_REAL_CST_PTR (expr
), TYPE_MODE (type
));
7192 for (bitpos
= 0; bitpos
< total_bytes
* BITS_PER_UNIT
;
7193 bitpos
+= BITS_PER_UNIT
)
7195 byte
= (bitpos
/ BITS_PER_UNIT
) & 3;
7196 value
= (unsigned char) (tmp
[bitpos
/ 32] >> (bitpos
& 31));
7198 if (UNITS_PER_WORD
< 4)
7200 word
= byte
/ UNITS_PER_WORD
;
7201 if (WORDS_BIG_ENDIAN
)
7202 word
= (words
- 1) - word
;
7203 offset
= word
* UNITS_PER_WORD
;
7204 if (BYTES_BIG_ENDIAN
)
7205 offset
+= (UNITS_PER_WORD
- 1) - (byte
% UNITS_PER_WORD
);
7207 offset
+= byte
% UNITS_PER_WORD
;
7210 offset
= BYTES_BIG_ENDIAN
? 3 - byte
: byte
;
7211 offset
= offset
+ ((bitpos
/ BITS_PER_UNIT
) & ~3);
7213 && offset
- off
< len
)
7214 ptr
[offset
- off
] = value
;
7216 return MIN (len
, total_bytes
- off
);
7219 /* Subroutine of native_encode_expr. Encode the COMPLEX_CST
7220 specified by EXPR into the buffer PTR of length LEN bytes.
7221 Return the number of bytes placed in the buffer, or zero
7225 native_encode_complex (const_tree expr
, unsigned char *ptr
, int len
, int off
)
7230 part
= TREE_REALPART (expr
);
7231 rsize
= native_encode_expr (part
, ptr
, len
, off
);
7235 part
= TREE_IMAGPART (expr
);
7237 off
= MAX (0, off
- GET_MODE_SIZE (TYPE_MODE (TREE_TYPE (part
))));
7238 isize
= native_encode_expr (part
, ptr
+rsize
, len
-rsize
, off
);
7242 return rsize
+ isize
;
7246 /* Subroutine of native_encode_expr. Encode the VECTOR_CST
7247 specified by EXPR into the buffer PTR of length LEN bytes.
7248 Return the number of bytes placed in the buffer, or zero
7252 native_encode_vector (const_tree expr
, unsigned char *ptr
, int len
, int off
)
7259 count
= VECTOR_CST_NELTS (expr
);
7260 itype
= TREE_TYPE (TREE_TYPE (expr
));
7261 size
= GET_MODE_SIZE (TYPE_MODE (itype
));
7262 for (i
= 0; i
< count
; i
++)
7269 elem
= VECTOR_CST_ELT (expr
, i
);
7270 int res
= native_encode_expr (elem
, ptr
+offset
, len
-offset
, off
);
7271 if ((off
== -1 && res
!= size
)
7284 /* Subroutine of native_encode_expr. Encode the STRING_CST
7285 specified by EXPR into the buffer PTR of length LEN bytes.
7286 Return the number of bytes placed in the buffer, or zero
7290 native_encode_string (const_tree expr
, unsigned char *ptr
, int len
, int off
)
7292 tree type
= TREE_TYPE (expr
);
7293 HOST_WIDE_INT total_bytes
;
7295 if (TREE_CODE (type
) != ARRAY_TYPE
7296 || TREE_CODE (TREE_TYPE (type
)) != INTEGER_TYPE
7297 || GET_MODE_BITSIZE (TYPE_MODE (TREE_TYPE (type
))) != BITS_PER_UNIT
7298 || !tree_fits_shwi_p (TYPE_SIZE_UNIT (type
)))
7300 total_bytes
= tree_to_shwi (TYPE_SIZE_UNIT (type
));
7301 if ((off
== -1 && total_bytes
> len
)
7302 || off
>= total_bytes
)
7306 if (TREE_STRING_LENGTH (expr
) - off
< MIN (total_bytes
, len
))
7309 if (off
< TREE_STRING_LENGTH (expr
))
7311 written
= MIN (len
, TREE_STRING_LENGTH (expr
) - off
);
7312 memcpy (ptr
, TREE_STRING_POINTER (expr
) + off
, written
);
7314 memset (ptr
+ written
, 0,
7315 MIN (total_bytes
- written
, len
- written
));
7318 memcpy (ptr
, TREE_STRING_POINTER (expr
) + off
, MIN (total_bytes
, len
));
7319 return MIN (total_bytes
- off
, len
);
7323 /* Subroutine of fold_view_convert_expr. Encode the INTEGER_CST,
7324 REAL_CST, COMPLEX_CST or VECTOR_CST specified by EXPR into the
7325 buffer PTR of length LEN bytes. If OFF is not -1 then start
7326 the encoding at byte offset OFF and encode at most LEN bytes.
7327 Return the number of bytes placed in the buffer, or zero upon failure. */
7330 native_encode_expr (const_tree expr
, unsigned char *ptr
, int len
, int off
)
7332 /* We don't support starting at negative offset and -1 is special. */
7336 switch (TREE_CODE (expr
))
7339 return native_encode_int (expr
, ptr
, len
, off
);
7342 return native_encode_real (expr
, ptr
, len
, off
);
7345 return native_encode_fixed (expr
, ptr
, len
, off
);
7348 return native_encode_complex (expr
, ptr
, len
, off
);
7351 return native_encode_vector (expr
, ptr
, len
, off
);
7354 return native_encode_string (expr
, ptr
, len
, off
);
7362 /* Subroutine of native_interpret_expr. Interpret the contents of
7363 the buffer PTR of length LEN as an INTEGER_CST of type TYPE.
7364 If the buffer cannot be interpreted, return NULL_TREE. */
7367 native_interpret_int (tree type
, const unsigned char *ptr
, int len
)
7369 int total_bytes
= GET_MODE_SIZE (TYPE_MODE (type
));
7371 if (total_bytes
> len
7372 || total_bytes
* BITS_PER_UNIT
> HOST_BITS_PER_DOUBLE_INT
)
7375 wide_int result
= wi::from_buffer (ptr
, total_bytes
);
7377 return wide_int_to_tree (type
, result
);
7381 /* Subroutine of native_interpret_expr. Interpret the contents of
7382 the buffer PTR of length LEN as a FIXED_CST of type TYPE.
7383 If the buffer cannot be interpreted, return NULL_TREE. */
7386 native_interpret_fixed (tree type
, const unsigned char *ptr
, int len
)
7388 int total_bytes
= GET_MODE_SIZE (TYPE_MODE (type
));
7390 FIXED_VALUE_TYPE fixed_value
;
7392 if (total_bytes
> len
7393 || total_bytes
* BITS_PER_UNIT
> HOST_BITS_PER_DOUBLE_INT
)
7396 result
= double_int::from_buffer (ptr
, total_bytes
);
7397 fixed_value
= fixed_from_double_int (result
, TYPE_MODE (type
));
7399 return build_fixed (type
, fixed_value
);
7403 /* Subroutine of native_interpret_expr. Interpret the contents of
7404 the buffer PTR of length LEN as a REAL_CST of type TYPE.
7405 If the buffer cannot be interpreted, return NULL_TREE. */
7408 native_interpret_real (tree type
, const unsigned char *ptr
, int len
)
7410 machine_mode mode
= TYPE_MODE (type
);
7411 int total_bytes
= GET_MODE_SIZE (mode
);
7412 unsigned char value
;
7413 /* There are always 32 bits in each long, no matter the size of
7414 the hosts long. We handle floating point representations with
7419 total_bytes
= GET_MODE_SIZE (TYPE_MODE (type
));
7420 if (total_bytes
> len
|| total_bytes
> 24)
7422 int words
= (32 / BITS_PER_UNIT
) / UNITS_PER_WORD
;
7424 memset (tmp
, 0, sizeof (tmp
));
7425 for (int bitpos
= 0; bitpos
< total_bytes
* BITS_PER_UNIT
;
7426 bitpos
+= BITS_PER_UNIT
)
7428 /* Both OFFSET and BYTE index within a long;
7429 bitpos indexes the whole float. */
7430 int offset
, byte
= (bitpos
/ BITS_PER_UNIT
) & 3;
7431 if (UNITS_PER_WORD
< 4)
7433 int word
= byte
/ UNITS_PER_WORD
;
7434 if (WORDS_BIG_ENDIAN
)
7435 word
= (words
- 1) - word
;
7436 offset
= word
* UNITS_PER_WORD
;
7437 if (BYTES_BIG_ENDIAN
)
7438 offset
+= (UNITS_PER_WORD
- 1) - (byte
% UNITS_PER_WORD
);
7440 offset
+= byte
% UNITS_PER_WORD
;
7445 if (BYTES_BIG_ENDIAN
)
7447 /* Reverse bytes within each long, or within the entire float
7448 if it's smaller than a long (for HFmode). */
7449 offset
= MIN (3, total_bytes
- 1) - offset
;
7450 gcc_assert (offset
>= 0);
7453 value
= ptr
[offset
+ ((bitpos
/ BITS_PER_UNIT
) & ~3)];
7455 tmp
[bitpos
/ 32] |= (unsigned long)value
<< (bitpos
& 31);
7458 real_from_target (&r
, tmp
, mode
);
7459 return build_real (type
, r
);
7463 /* Subroutine of native_interpret_expr. Interpret the contents of
7464 the buffer PTR of length LEN as a COMPLEX_CST of type TYPE.
7465 If the buffer cannot be interpreted, return NULL_TREE. */
7468 native_interpret_complex (tree type
, const unsigned char *ptr
, int len
)
7470 tree etype
, rpart
, ipart
;
7473 etype
= TREE_TYPE (type
);
7474 size
= GET_MODE_SIZE (TYPE_MODE (etype
));
7477 rpart
= native_interpret_expr (etype
, ptr
, size
);
7480 ipart
= native_interpret_expr (etype
, ptr
+size
, size
);
7483 return build_complex (type
, rpart
, ipart
);
7487 /* Subroutine of native_interpret_expr. Interpret the contents of
7488 the buffer PTR of length LEN as a VECTOR_CST of type TYPE.
7489 If the buffer cannot be interpreted, return NULL_TREE. */
7492 native_interpret_vector (tree type
, const unsigned char *ptr
, int len
)
7498 etype
= TREE_TYPE (type
);
7499 size
= GET_MODE_SIZE (TYPE_MODE (etype
));
7500 count
= TYPE_VECTOR_SUBPARTS (type
);
7501 if (size
* count
> len
)
7504 elements
= XALLOCAVEC (tree
, count
);
7505 for (i
= count
- 1; i
>= 0; i
--)
7507 elem
= native_interpret_expr (etype
, ptr
+(i
*size
), size
);
7512 return build_vector (type
, elements
);
7516 /* Subroutine of fold_view_convert_expr. Interpret the contents of
7517 the buffer PTR of length LEN as a constant of type TYPE. For
7518 INTEGRAL_TYPE_P we return an INTEGER_CST, for SCALAR_FLOAT_TYPE_P
7519 we return a REAL_CST, etc... If the buffer cannot be interpreted,
7520 return NULL_TREE. */
7523 native_interpret_expr (tree type
, const unsigned char *ptr
, int len
)
7525 switch (TREE_CODE (type
))
7531 case REFERENCE_TYPE
:
7532 return native_interpret_int (type
, ptr
, len
);
7535 return native_interpret_real (type
, ptr
, len
);
7537 case FIXED_POINT_TYPE
:
7538 return native_interpret_fixed (type
, ptr
, len
);
7541 return native_interpret_complex (type
, ptr
, len
);
7544 return native_interpret_vector (type
, ptr
, len
);
7551 /* Returns true if we can interpret the contents of a native encoding
7555 can_native_interpret_type_p (tree type
)
7557 switch (TREE_CODE (type
))
7563 case REFERENCE_TYPE
:
7564 case FIXED_POINT_TYPE
:
7574 /* Fold a VIEW_CONVERT_EXPR of a constant expression EXPR to type
7575 TYPE at compile-time. If we're unable to perform the conversion
7576 return NULL_TREE. */
7579 fold_view_convert_expr (tree type
, tree expr
)
7581 /* We support up to 512-bit values (for V8DFmode). */
7582 unsigned char buffer
[64];
7585 /* Check that the host and target are sane. */
7586 if (CHAR_BIT
!= 8 || BITS_PER_UNIT
!= 8)
7589 len
= native_encode_expr (expr
, buffer
, sizeof (buffer
));
7593 return native_interpret_expr (type
, buffer
, len
);
7596 /* Build an expression for the address of T. Folds away INDIRECT_REF
7597 to avoid confusing the gimplify process. */
7600 build_fold_addr_expr_with_type_loc (location_t loc
, tree t
, tree ptrtype
)
7602 /* The size of the object is not relevant when talking about its address. */
7603 if (TREE_CODE (t
) == WITH_SIZE_EXPR
)
7604 t
= TREE_OPERAND (t
, 0);
7606 if (TREE_CODE (t
) == INDIRECT_REF
)
7608 t
= TREE_OPERAND (t
, 0);
7610 if (TREE_TYPE (t
) != ptrtype
)
7611 t
= build1_loc (loc
, NOP_EXPR
, ptrtype
, t
);
7613 else if (TREE_CODE (t
) == MEM_REF
7614 && integer_zerop (TREE_OPERAND (t
, 1)))
7615 return TREE_OPERAND (t
, 0);
7616 else if (TREE_CODE (t
) == MEM_REF
7617 && TREE_CODE (TREE_OPERAND (t
, 0)) == INTEGER_CST
)
7618 return fold_binary (POINTER_PLUS_EXPR
, ptrtype
,
7619 TREE_OPERAND (t
, 0),
7620 convert_to_ptrofftype (TREE_OPERAND (t
, 1)));
7621 else if (TREE_CODE (t
) == VIEW_CONVERT_EXPR
)
7623 t
= build_fold_addr_expr_loc (loc
, TREE_OPERAND (t
, 0));
7625 if (TREE_TYPE (t
) != ptrtype
)
7626 t
= fold_convert_loc (loc
, ptrtype
, t
);
7629 t
= build1_loc (loc
, ADDR_EXPR
, ptrtype
, t
);
7634 /* Build an expression for the address of T. */
7637 build_fold_addr_expr_loc (location_t loc
, tree t
)
7639 tree ptrtype
= build_pointer_type (TREE_TYPE (t
));
7641 return build_fold_addr_expr_with_type_loc (loc
, t
, ptrtype
);
7644 /* Fold a unary expression of code CODE and type TYPE with operand
7645 OP0. Return the folded expression if folding is successful.
7646 Otherwise, return NULL_TREE. */
7649 fold_unary_loc (location_t loc
, enum tree_code code
, tree type
, tree op0
)
7653 enum tree_code_class kind
= TREE_CODE_CLASS (code
);
7655 gcc_assert (IS_EXPR_CODE_CLASS (kind
)
7656 && TREE_CODE_LENGTH (code
) == 1);
7661 if (CONVERT_EXPR_CODE_P (code
)
7662 || code
== FLOAT_EXPR
|| code
== ABS_EXPR
|| code
== NEGATE_EXPR
)
7664 /* Don't use STRIP_NOPS, because signedness of argument type
7666 STRIP_SIGN_NOPS (arg0
);
7670 /* Strip any conversions that don't change the mode. This
7671 is safe for every expression, except for a comparison
7672 expression because its signedness is derived from its
7675 Note that this is done as an internal manipulation within
7676 the constant folder, in order to find the simplest
7677 representation of the arguments so that their form can be
7678 studied. In any cases, the appropriate type conversions
7679 should be put back in the tree that will get out of the
7684 if (CONSTANT_CLASS_P (arg0
))
7686 tree tem
= const_unop (code
, type
, arg0
);
7689 if (TREE_TYPE (tem
) != type
)
7690 tem
= fold_convert_loc (loc
, type
, tem
);
7696 tem
= generic_simplify (loc
, code
, type
, op0
);
7700 if (TREE_CODE_CLASS (code
) == tcc_unary
)
7702 if (TREE_CODE (arg0
) == COMPOUND_EXPR
)
7703 return build2 (COMPOUND_EXPR
, type
, TREE_OPERAND (arg0
, 0),
7704 fold_build1_loc (loc
, code
, type
,
7705 fold_convert_loc (loc
, TREE_TYPE (op0
),
7706 TREE_OPERAND (arg0
, 1))));
7707 else if (TREE_CODE (arg0
) == COND_EXPR
)
7709 tree arg01
= TREE_OPERAND (arg0
, 1);
7710 tree arg02
= TREE_OPERAND (arg0
, 2);
7711 if (! VOID_TYPE_P (TREE_TYPE (arg01
)))
7712 arg01
= fold_build1_loc (loc
, code
, type
,
7713 fold_convert_loc (loc
,
7714 TREE_TYPE (op0
), arg01
));
7715 if (! VOID_TYPE_P (TREE_TYPE (arg02
)))
7716 arg02
= fold_build1_loc (loc
, code
, type
,
7717 fold_convert_loc (loc
,
7718 TREE_TYPE (op0
), arg02
));
7719 tem
= fold_build3_loc (loc
, COND_EXPR
, type
, TREE_OPERAND (arg0
, 0),
7722 /* If this was a conversion, and all we did was to move into
7723 inside the COND_EXPR, bring it back out. But leave it if
7724 it is a conversion from integer to integer and the
7725 result precision is no wider than a word since such a
7726 conversion is cheap and may be optimized away by combine,
7727 while it couldn't if it were outside the COND_EXPR. Then return
7728 so we don't get into an infinite recursion loop taking the
7729 conversion out and then back in. */
7731 if ((CONVERT_EXPR_CODE_P (code
)
7732 || code
== NON_LVALUE_EXPR
)
7733 && TREE_CODE (tem
) == COND_EXPR
7734 && TREE_CODE (TREE_OPERAND (tem
, 1)) == code
7735 && TREE_CODE (TREE_OPERAND (tem
, 2)) == code
7736 && ! VOID_TYPE_P (TREE_OPERAND (tem
, 1))
7737 && ! VOID_TYPE_P (TREE_OPERAND (tem
, 2))
7738 && (TREE_TYPE (TREE_OPERAND (TREE_OPERAND (tem
, 1), 0))
7739 == TREE_TYPE (TREE_OPERAND (TREE_OPERAND (tem
, 2), 0)))
7740 && (! (INTEGRAL_TYPE_P (TREE_TYPE (tem
))
7742 (TREE_TYPE (TREE_OPERAND (TREE_OPERAND (tem
, 1), 0))))
7743 && TYPE_PRECISION (TREE_TYPE (tem
)) <= BITS_PER_WORD
)
7744 || flag_syntax_only
))
7745 tem
= build1_loc (loc
, code
, type
,
7747 TREE_TYPE (TREE_OPERAND
7748 (TREE_OPERAND (tem
, 1), 0)),
7749 TREE_OPERAND (tem
, 0),
7750 TREE_OPERAND (TREE_OPERAND (tem
, 1), 0),
7751 TREE_OPERAND (TREE_OPERAND (tem
, 2),
7759 case NON_LVALUE_EXPR
:
7760 if (!maybe_lvalue_p (op0
))
7761 return fold_convert_loc (loc
, type
, op0
);
7766 case FIX_TRUNC_EXPR
:
7767 if (COMPARISON_CLASS_P (op0
))
7769 /* If we have (type) (a CMP b) and type is an integral type, return
7770 new expression involving the new type. Canonicalize
7771 (type) (a CMP b) to (a CMP b) ? (type) true : (type) false for
7773 Do not fold the result as that would not simplify further, also
7774 folding again results in recursions. */
7775 if (TREE_CODE (type
) == BOOLEAN_TYPE
)
7776 return build2_loc (loc
, TREE_CODE (op0
), type
,
7777 TREE_OPERAND (op0
, 0),
7778 TREE_OPERAND (op0
, 1));
7779 else if (!INTEGRAL_TYPE_P (type
) && !VOID_TYPE_P (type
)
7780 && TREE_CODE (type
) != VECTOR_TYPE
)
7781 return build3_loc (loc
, COND_EXPR
, type
, op0
,
7782 constant_boolean_node (true, type
),
7783 constant_boolean_node (false, type
));
7786 /* Handle (T *)&A.B.C for A being of type T and B and C
7787 living at offset zero. This occurs frequently in
7788 C++ upcasting and then accessing the base. */
7789 if (TREE_CODE (op0
) == ADDR_EXPR
7790 && POINTER_TYPE_P (type
)
7791 && handled_component_p (TREE_OPERAND (op0
, 0)))
7793 HOST_WIDE_INT bitsize
, bitpos
;
7796 int unsignedp
, reversep
, volatilep
;
7798 = get_inner_reference (TREE_OPERAND (op0
, 0), &bitsize
, &bitpos
,
7799 &offset
, &mode
, &unsignedp
, &reversep
,
7801 /* If the reference was to a (constant) zero offset, we can use
7802 the address of the base if it has the same base type
7803 as the result type and the pointer type is unqualified. */
7804 if (! offset
&& bitpos
== 0
7805 && (TYPE_MAIN_VARIANT (TREE_TYPE (type
))
7806 == TYPE_MAIN_VARIANT (TREE_TYPE (base
)))
7807 && TYPE_QUALS (type
) == TYPE_UNQUALIFIED
)
7808 return fold_convert_loc (loc
, type
,
7809 build_fold_addr_expr_loc (loc
, base
));
7812 if (TREE_CODE (op0
) == MODIFY_EXPR
7813 && TREE_CONSTANT (TREE_OPERAND (op0
, 1))
7814 /* Detect assigning a bitfield. */
7815 && !(TREE_CODE (TREE_OPERAND (op0
, 0)) == COMPONENT_REF
7817 (TREE_OPERAND (TREE_OPERAND (op0
, 0), 1))))
7819 /* Don't leave an assignment inside a conversion
7820 unless assigning a bitfield. */
7821 tem
= fold_build1_loc (loc
, code
, type
, TREE_OPERAND (op0
, 1));
7822 /* First do the assignment, then return converted constant. */
7823 tem
= build2_loc (loc
, COMPOUND_EXPR
, TREE_TYPE (tem
), op0
, tem
);
7824 TREE_NO_WARNING (tem
) = 1;
7825 TREE_USED (tem
) = 1;
7829 /* Convert (T)(x & c) into (T)x & (T)c, if c is an integer
7830 constants (if x has signed type, the sign bit cannot be set
7831 in c). This folds extension into the BIT_AND_EXPR.
7832 ??? We don't do it for BOOLEAN_TYPE or ENUMERAL_TYPE because they
7833 very likely don't have maximal range for their precision and this
7834 transformation effectively doesn't preserve non-maximal ranges. */
7835 if (TREE_CODE (type
) == INTEGER_TYPE
7836 && TREE_CODE (op0
) == BIT_AND_EXPR
7837 && TREE_CODE (TREE_OPERAND (op0
, 1)) == INTEGER_CST
)
7839 tree and_expr
= op0
;
7840 tree and0
= TREE_OPERAND (and_expr
, 0);
7841 tree and1
= TREE_OPERAND (and_expr
, 1);
7844 if (TYPE_UNSIGNED (TREE_TYPE (and_expr
))
7845 || (TYPE_PRECISION (type
)
7846 <= TYPE_PRECISION (TREE_TYPE (and_expr
))))
7848 else if (TYPE_PRECISION (TREE_TYPE (and1
))
7849 <= HOST_BITS_PER_WIDE_INT
7850 && tree_fits_uhwi_p (and1
))
7852 unsigned HOST_WIDE_INT cst
;
7854 cst
= tree_to_uhwi (and1
);
7855 cst
&= HOST_WIDE_INT_M1U
7856 << (TYPE_PRECISION (TREE_TYPE (and1
)) - 1);
7857 change
= (cst
== 0);
7859 && !flag_syntax_only
7860 && (LOAD_EXTEND_OP (TYPE_MODE (TREE_TYPE (and0
)))
7863 tree uns
= unsigned_type_for (TREE_TYPE (and0
));
7864 and0
= fold_convert_loc (loc
, uns
, and0
);
7865 and1
= fold_convert_loc (loc
, uns
, and1
);
7870 tem
= force_fit_type (type
, wi::to_widest (and1
), 0,
7871 TREE_OVERFLOW (and1
));
7872 return fold_build2_loc (loc
, BIT_AND_EXPR
, type
,
7873 fold_convert_loc (loc
, type
, and0
), tem
);
7877 /* Convert (T1)(X p+ Y) into ((T1)X p+ Y), for pointer type, when the new
7878 cast (T1)X will fold away. We assume that this happens when X itself
7880 if (POINTER_TYPE_P (type
)
7881 && TREE_CODE (arg0
) == POINTER_PLUS_EXPR
7882 && CONVERT_EXPR_P (TREE_OPERAND (arg0
, 0)))
7884 tree arg00
= TREE_OPERAND (arg0
, 0);
7885 tree arg01
= TREE_OPERAND (arg0
, 1);
7887 return fold_build_pointer_plus_loc
7888 (loc
, fold_convert_loc (loc
, type
, arg00
), arg01
);
7891 /* Convert (T1)(~(T2)X) into ~(T1)X if T1 and T2 are integral types
7892 of the same precision, and X is an integer type not narrower than
7893 types T1 or T2, i.e. the cast (T2)X isn't an extension. */
7894 if (INTEGRAL_TYPE_P (type
)
7895 && TREE_CODE (op0
) == BIT_NOT_EXPR
7896 && INTEGRAL_TYPE_P (TREE_TYPE (op0
))
7897 && CONVERT_EXPR_P (TREE_OPERAND (op0
, 0))
7898 && TYPE_PRECISION (type
) == TYPE_PRECISION (TREE_TYPE (op0
)))
7900 tem
= TREE_OPERAND (TREE_OPERAND (op0
, 0), 0);
7901 if (INTEGRAL_TYPE_P (TREE_TYPE (tem
))
7902 && TYPE_PRECISION (type
) <= TYPE_PRECISION (TREE_TYPE (tem
)))
7903 return fold_build1_loc (loc
, BIT_NOT_EXPR
, type
,
7904 fold_convert_loc (loc
, type
, tem
));
7907 /* Convert (T1)(X * Y) into (T1)X * (T1)Y if T1 is narrower than the
7908 type of X and Y (integer types only). */
7909 if (INTEGRAL_TYPE_P (type
)
7910 && TREE_CODE (op0
) == MULT_EXPR
7911 && INTEGRAL_TYPE_P (TREE_TYPE (op0
))
7912 && TYPE_PRECISION (type
) < TYPE_PRECISION (TREE_TYPE (op0
)))
7914 /* Be careful not to introduce new overflows. */
7916 if (TYPE_OVERFLOW_WRAPS (type
))
7919 mult_type
= unsigned_type_for (type
);
7921 if (TYPE_PRECISION (mult_type
) < TYPE_PRECISION (TREE_TYPE (op0
)))
7923 tem
= fold_build2_loc (loc
, MULT_EXPR
, mult_type
,
7924 fold_convert_loc (loc
, mult_type
,
7925 TREE_OPERAND (op0
, 0)),
7926 fold_convert_loc (loc
, mult_type
,
7927 TREE_OPERAND (op0
, 1)));
7928 return fold_convert_loc (loc
, type
, tem
);
7934 case VIEW_CONVERT_EXPR
:
7935 if (TREE_CODE (op0
) == MEM_REF
)
7937 tem
= fold_build2_loc (loc
, MEM_REF
, type
,
7938 TREE_OPERAND (op0
, 0), TREE_OPERAND (op0
, 1));
7939 REF_REVERSE_STORAGE_ORDER (tem
) = REF_REVERSE_STORAGE_ORDER (op0
);
7946 tem
= fold_negate_expr (loc
, arg0
);
7948 return fold_convert_loc (loc
, type
, tem
);
7952 /* Convert fabs((double)float) into (double)fabsf(float). */
7953 if (TREE_CODE (arg0
) == NOP_EXPR
7954 && TREE_CODE (type
) == REAL_TYPE
)
7956 tree targ0
= strip_float_extensions (arg0
);
7958 return fold_convert_loc (loc
, type
,
7959 fold_build1_loc (loc
, ABS_EXPR
,
7966 /* Convert ~(X ^ Y) to ~X ^ Y or X ^ ~Y if ~X or ~Y simplify. */
7967 if (TREE_CODE (arg0
) == BIT_XOR_EXPR
7968 && (tem
= fold_unary_loc (loc
, BIT_NOT_EXPR
, type
,
7969 fold_convert_loc (loc
, type
,
7970 TREE_OPERAND (arg0
, 0)))))
7971 return fold_build2_loc (loc
, BIT_XOR_EXPR
, type
, tem
,
7972 fold_convert_loc (loc
, type
,
7973 TREE_OPERAND (arg0
, 1)));
7974 else if (TREE_CODE (arg0
) == BIT_XOR_EXPR
7975 && (tem
= fold_unary_loc (loc
, BIT_NOT_EXPR
, type
,
7976 fold_convert_loc (loc
, type
,
7977 TREE_OPERAND (arg0
, 1)))))
7978 return fold_build2_loc (loc
, BIT_XOR_EXPR
, type
,
7979 fold_convert_loc (loc
, type
,
7980 TREE_OPERAND (arg0
, 0)), tem
);
7984 case TRUTH_NOT_EXPR
:
7985 /* Note that the operand of this must be an int
7986 and its values must be 0 or 1.
7987 ("true" is a fixed value perhaps depending on the language,
7988 but we don't handle values other than 1 correctly yet.) */
7989 tem
= fold_truth_not_expr (loc
, arg0
);
7992 return fold_convert_loc (loc
, type
, tem
);
7995 /* Fold *&X to X if X is an lvalue. */
7996 if (TREE_CODE (op0
) == ADDR_EXPR
)
7998 tree op00
= TREE_OPERAND (op0
, 0);
7999 if ((TREE_CODE (op00
) == VAR_DECL
8000 || TREE_CODE (op00
) == PARM_DECL
8001 || TREE_CODE (op00
) == RESULT_DECL
)
8002 && !TREE_READONLY (op00
))
8009 } /* switch (code) */
8013 /* If the operation was a conversion do _not_ mark a resulting constant
8014 with TREE_OVERFLOW if the original constant was not. These conversions
8015 have implementation defined behavior and retaining the TREE_OVERFLOW
8016 flag here would confuse later passes such as VRP. */
8018 fold_unary_ignore_overflow_loc (location_t loc
, enum tree_code code
,
8019 tree type
, tree op0
)
8021 tree res
= fold_unary_loc (loc
, code
, type
, op0
);
8023 && TREE_CODE (res
) == INTEGER_CST
8024 && TREE_CODE (op0
) == INTEGER_CST
8025 && CONVERT_EXPR_CODE_P (code
))
8026 TREE_OVERFLOW (res
) = TREE_OVERFLOW (op0
);
8031 /* Fold a binary bitwise/truth expression of code CODE and type TYPE with
8032 operands OP0 and OP1. LOC is the location of the resulting expression.
8033 ARG0 and ARG1 are the NOP_STRIPed results of OP0 and OP1.
8034 Return the folded expression if folding is successful. Otherwise,
8035 return NULL_TREE. */
8037 fold_truth_andor (location_t loc
, enum tree_code code
, tree type
,
8038 tree arg0
, tree arg1
, tree op0
, tree op1
)
8042 /* We only do these simplifications if we are optimizing. */
8046 /* Check for things like (A || B) && (A || C). We can convert this
8047 to A || (B && C). Note that either operator can be any of the four
8048 truth and/or operations and the transformation will still be
8049 valid. Also note that we only care about order for the
8050 ANDIF and ORIF operators. If B contains side effects, this
8051 might change the truth-value of A. */
8052 if (TREE_CODE (arg0
) == TREE_CODE (arg1
)
8053 && (TREE_CODE (arg0
) == TRUTH_ANDIF_EXPR
8054 || TREE_CODE (arg0
) == TRUTH_ORIF_EXPR
8055 || TREE_CODE (arg0
) == TRUTH_AND_EXPR
8056 || TREE_CODE (arg0
) == TRUTH_OR_EXPR
)
8057 && ! TREE_SIDE_EFFECTS (TREE_OPERAND (arg0
, 1)))
8059 tree a00
= TREE_OPERAND (arg0
, 0);
8060 tree a01
= TREE_OPERAND (arg0
, 1);
8061 tree a10
= TREE_OPERAND (arg1
, 0);
8062 tree a11
= TREE_OPERAND (arg1
, 1);
8063 int commutative
= ((TREE_CODE (arg0
) == TRUTH_OR_EXPR
8064 || TREE_CODE (arg0
) == TRUTH_AND_EXPR
)
8065 && (code
== TRUTH_AND_EXPR
8066 || code
== TRUTH_OR_EXPR
));
8068 if (operand_equal_p (a00
, a10
, 0))
8069 return fold_build2_loc (loc
, TREE_CODE (arg0
), type
, a00
,
8070 fold_build2_loc (loc
, code
, type
, a01
, a11
));
8071 else if (commutative
&& operand_equal_p (a00
, a11
, 0))
8072 return fold_build2_loc (loc
, TREE_CODE (arg0
), type
, a00
,
8073 fold_build2_loc (loc
, code
, type
, a01
, a10
));
8074 else if (commutative
&& operand_equal_p (a01
, a10
, 0))
8075 return fold_build2_loc (loc
, TREE_CODE (arg0
), type
, a01
,
8076 fold_build2_loc (loc
, code
, type
, a00
, a11
));
8078 /* This case if tricky because we must either have commutative
8079 operators or else A10 must not have side-effects. */
8081 else if ((commutative
|| ! TREE_SIDE_EFFECTS (a10
))
8082 && operand_equal_p (a01
, a11
, 0))
8083 return fold_build2_loc (loc
, TREE_CODE (arg0
), type
,
8084 fold_build2_loc (loc
, code
, type
, a00
, a10
),
8088 /* See if we can build a range comparison. */
8089 if (0 != (tem
= fold_range_test (loc
, code
, type
, op0
, op1
)))
8092 if ((code
== TRUTH_ANDIF_EXPR
&& TREE_CODE (arg0
) == TRUTH_ORIF_EXPR
)
8093 || (code
== TRUTH_ORIF_EXPR
&& TREE_CODE (arg0
) == TRUTH_ANDIF_EXPR
))
8095 tem
= merge_truthop_with_opposite_arm (loc
, arg0
, arg1
, true);
8097 return fold_build2_loc (loc
, code
, type
, tem
, arg1
);
8100 if ((code
== TRUTH_ANDIF_EXPR
&& TREE_CODE (arg1
) == TRUTH_ORIF_EXPR
)
8101 || (code
== TRUTH_ORIF_EXPR
&& TREE_CODE (arg1
) == TRUTH_ANDIF_EXPR
))
8103 tem
= merge_truthop_with_opposite_arm (loc
, arg1
, arg0
, false);
8105 return fold_build2_loc (loc
, code
, type
, arg0
, tem
);
8108 /* Check for the possibility of merging component references. If our
8109 lhs is another similar operation, try to merge its rhs with our
8110 rhs. Then try to merge our lhs and rhs. */
8111 if (TREE_CODE (arg0
) == code
8112 && 0 != (tem
= fold_truth_andor_1 (loc
, code
, type
,
8113 TREE_OPERAND (arg0
, 1), arg1
)))
8114 return fold_build2_loc (loc
, code
, type
, TREE_OPERAND (arg0
, 0), tem
);
8116 if ((tem
= fold_truth_andor_1 (loc
, code
, type
, arg0
, arg1
)) != 0)
8119 if (LOGICAL_OP_NON_SHORT_CIRCUIT
8120 && (code
== TRUTH_AND_EXPR
8121 || code
== TRUTH_ANDIF_EXPR
8122 || code
== TRUTH_OR_EXPR
8123 || code
== TRUTH_ORIF_EXPR
))
8125 enum tree_code ncode
, icode
;
8127 ncode
= (code
== TRUTH_ANDIF_EXPR
|| code
== TRUTH_AND_EXPR
)
8128 ? TRUTH_AND_EXPR
: TRUTH_OR_EXPR
;
8129 icode
= ncode
== TRUTH_AND_EXPR
? TRUTH_ANDIF_EXPR
: TRUTH_ORIF_EXPR
;
8131 /* Transform ((A AND-IF B) AND[-IF] C) into (A AND-IF (B AND C)),
8132 or ((A OR-IF B) OR[-IF] C) into (A OR-IF (B OR C))
8133 We don't want to pack more than two leafs to a non-IF AND/OR
8135 If tree-code of left-hand operand isn't an AND/OR-IF code and not
8136 equal to IF-CODE, then we don't want to add right-hand operand.
8137 If the inner right-hand side of left-hand operand has
8138 side-effects, or isn't simple, then we can't add to it,
8139 as otherwise we might destroy if-sequence. */
8140 if (TREE_CODE (arg0
) == icode
8141 && simple_operand_p_2 (arg1
)
8142 /* Needed for sequence points to handle trappings, and
8144 && simple_operand_p_2 (TREE_OPERAND (arg0
, 1)))
8146 tem
= fold_build2_loc (loc
, ncode
, type
, TREE_OPERAND (arg0
, 1),
8148 return fold_build2_loc (loc
, icode
, type
, TREE_OPERAND (arg0
, 0),
8151 /* Same as abouve but for (A AND[-IF] (B AND-IF C)) -> ((A AND B) AND-IF C),
8152 or (A OR[-IF] (B OR-IF C) -> ((A OR B) OR-IF C). */
8153 else if (TREE_CODE (arg1
) == icode
8154 && simple_operand_p_2 (arg0
)
8155 /* Needed for sequence points to handle trappings, and
8157 && simple_operand_p_2 (TREE_OPERAND (arg1
, 0)))
8159 tem
= fold_build2_loc (loc
, ncode
, type
,
8160 arg0
, TREE_OPERAND (arg1
, 0));
8161 return fold_build2_loc (loc
, icode
, type
, tem
,
8162 TREE_OPERAND (arg1
, 1));
8164 /* Transform (A AND-IF B) into (A AND B), or (A OR-IF B)
8166 For sequence point consistancy, we need to check for trapping,
8167 and side-effects. */
8168 else if (code
== icode
&& simple_operand_p_2 (arg0
)
8169 && simple_operand_p_2 (arg1
))
8170 return fold_build2_loc (loc
, ncode
, type
, arg0
, arg1
);
8176 /* Helper that tries to canonicalize the comparison ARG0 CODE ARG1
8177 by changing CODE to reduce the magnitude of constants involved in
8178 ARG0 of the comparison.
8179 Returns a canonicalized comparison tree if a simplification was
8180 possible, otherwise returns NULL_TREE.
8181 Set *STRICT_OVERFLOW_P to true if the canonicalization is only
8182 valid if signed overflow is undefined. */
8185 maybe_canonicalize_comparison_1 (location_t loc
, enum tree_code code
, tree type
,
8186 tree arg0
, tree arg1
,
8187 bool *strict_overflow_p
)
8189 enum tree_code code0
= TREE_CODE (arg0
);
8190 tree t
, cst0
= NULL_TREE
;
8193 /* Match A +- CST code arg1. We can change this only if overflow
8195 if (!((ANY_INTEGRAL_TYPE_P (TREE_TYPE (arg0
))
8196 && TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (arg0
)))
8197 /* In principle pointers also have undefined overflow behavior,
8198 but that causes problems elsewhere. */
8199 && !POINTER_TYPE_P (TREE_TYPE (arg0
))
8200 && (code0
== MINUS_EXPR
8201 || code0
== PLUS_EXPR
)
8202 && TREE_CODE (TREE_OPERAND (arg0
, 1)) == INTEGER_CST
))
8205 /* Identify the constant in arg0 and its sign. */
8206 cst0
= TREE_OPERAND (arg0
, 1);
8207 sgn0
= tree_int_cst_sgn (cst0
);
8209 /* Overflowed constants and zero will cause problems. */
8210 if (integer_zerop (cst0
)
8211 || TREE_OVERFLOW (cst0
))
8214 /* See if we can reduce the magnitude of the constant in
8215 arg0 by changing the comparison code. */
8216 /* A - CST < arg1 -> A - CST-1 <= arg1. */
8218 && code0
== ((sgn0
== -1) ? PLUS_EXPR
: MINUS_EXPR
))
8220 /* A + CST > arg1 -> A + CST-1 >= arg1. */
8221 else if (code
== GT_EXPR
8222 && code0
== ((sgn0
== -1) ? MINUS_EXPR
: PLUS_EXPR
))
8224 /* A + CST <= arg1 -> A + CST-1 < arg1. */
8225 else if (code
== LE_EXPR
8226 && code0
== ((sgn0
== -1) ? MINUS_EXPR
: PLUS_EXPR
))
8228 /* A - CST >= arg1 -> A - CST-1 > arg1. */
8229 else if (code
== GE_EXPR
8230 && code0
== ((sgn0
== -1) ? PLUS_EXPR
: MINUS_EXPR
))
8234 *strict_overflow_p
= true;
8236 /* Now build the constant reduced in magnitude. But not if that
8237 would produce one outside of its types range. */
8238 if (INTEGRAL_TYPE_P (TREE_TYPE (cst0
))
8240 && TYPE_MIN_VALUE (TREE_TYPE (cst0
))
8241 && tree_int_cst_equal (cst0
, TYPE_MIN_VALUE (TREE_TYPE (cst0
))))
8243 && TYPE_MAX_VALUE (TREE_TYPE (cst0
))
8244 && tree_int_cst_equal (cst0
, TYPE_MAX_VALUE (TREE_TYPE (cst0
))))))
8247 t
= int_const_binop (sgn0
== -1 ? PLUS_EXPR
: MINUS_EXPR
,
8248 cst0
, build_int_cst (TREE_TYPE (cst0
), 1));
8249 t
= fold_build2_loc (loc
, code0
, TREE_TYPE (arg0
), TREE_OPERAND (arg0
, 0), t
);
8250 t
= fold_convert (TREE_TYPE (arg1
), t
);
8252 return fold_build2_loc (loc
, code
, type
, t
, arg1
);
8255 /* Canonicalize the comparison ARG0 CODE ARG1 with type TYPE with undefined
8256 overflow further. Try to decrease the magnitude of constants involved
8257 by changing LE_EXPR and GE_EXPR to LT_EXPR and GT_EXPR or vice versa
8258 and put sole constants at the second argument position.
8259 Returns the canonicalized tree if changed, otherwise NULL_TREE. */
8262 maybe_canonicalize_comparison (location_t loc
, enum tree_code code
, tree type
,
8263 tree arg0
, tree arg1
)
8266 bool strict_overflow_p
;
8267 const char * const warnmsg
= G_("assuming signed overflow does not occur "
8268 "when reducing constant in comparison");
8270 /* Try canonicalization by simplifying arg0. */
8271 strict_overflow_p
= false;
8272 t
= maybe_canonicalize_comparison_1 (loc
, code
, type
, arg0
, arg1
,
8273 &strict_overflow_p
);
8276 if (strict_overflow_p
)
8277 fold_overflow_warning (warnmsg
, WARN_STRICT_OVERFLOW_MAGNITUDE
);
8281 /* Try canonicalization by simplifying arg1 using the swapped
8283 code
= swap_tree_comparison (code
);
8284 strict_overflow_p
= false;
8285 t
= maybe_canonicalize_comparison_1 (loc
, code
, type
, arg1
, arg0
,
8286 &strict_overflow_p
);
8287 if (t
&& strict_overflow_p
)
8288 fold_overflow_warning (warnmsg
, WARN_STRICT_OVERFLOW_MAGNITUDE
);
8292 /* Return whether BASE + OFFSET + BITPOS may wrap around the address
8293 space. This is used to avoid issuing overflow warnings for
8294 expressions like &p->x which can not wrap. */
8297 pointer_may_wrap_p (tree base
, tree offset
, HOST_WIDE_INT bitpos
)
8299 if (!POINTER_TYPE_P (TREE_TYPE (base
)))
8306 int precision
= TYPE_PRECISION (TREE_TYPE (base
));
8307 if (offset
== NULL_TREE
)
8308 wi_offset
= wi::zero (precision
);
8309 else if (TREE_CODE (offset
) != INTEGER_CST
|| TREE_OVERFLOW (offset
))
8315 wide_int units
= wi::shwi (bitpos
/ BITS_PER_UNIT
, precision
);
8316 wide_int total
= wi::add (wi_offset
, units
, UNSIGNED
, &overflow
);
8320 if (!wi::fits_uhwi_p (total
))
8323 HOST_WIDE_INT size
= int_size_in_bytes (TREE_TYPE (TREE_TYPE (base
)));
8327 /* We can do slightly better for SIZE if we have an ADDR_EXPR of an
8329 if (TREE_CODE (base
) == ADDR_EXPR
)
8331 HOST_WIDE_INT base_size
;
8333 base_size
= int_size_in_bytes (TREE_TYPE (TREE_OPERAND (base
, 0)));
8334 if (base_size
> 0 && size
< base_size
)
8338 return total
.to_uhwi () > (unsigned HOST_WIDE_INT
) size
;
8341 /* Return a positive integer when the symbol DECL is known to have
8342 a nonzero address, zero when it's known not to (e.g., it's a weak
8343 symbol), and a negative integer when the symbol is not yet in the
8344 symbol table and so whether or not its address is zero is unknown. */
8346 maybe_nonzero_address (tree decl
)
8348 if (DECL_P (decl
) && decl_in_symtab_p (decl
))
8349 if (struct symtab_node
*symbol
= symtab_node::get_create (decl
))
8350 return symbol
->nonzero_address ();
8355 /* Subroutine of fold_binary. This routine performs all of the
8356 transformations that are common to the equality/inequality
8357 operators (EQ_EXPR and NE_EXPR) and the ordering operators
8358 (LT_EXPR, LE_EXPR, GE_EXPR and GT_EXPR). Callers other than
8359 fold_binary should call fold_binary. Fold a comparison with
8360 tree code CODE and type TYPE with operands OP0 and OP1. Return
8361 the folded comparison or NULL_TREE. */
8364 fold_comparison (location_t loc
, enum tree_code code
, tree type
,
8367 const bool equality_code
= (code
== EQ_EXPR
|| code
== NE_EXPR
);
8368 tree arg0
, arg1
, tem
;
8373 STRIP_SIGN_NOPS (arg0
);
8374 STRIP_SIGN_NOPS (arg1
);
8376 /* Transform comparisons of the form X +- C1 CMP C2 to X CMP C2 -+ C1. */
8377 if ((TREE_CODE (arg0
) == PLUS_EXPR
|| TREE_CODE (arg0
) == MINUS_EXPR
)
8379 || (ANY_INTEGRAL_TYPE_P (TREE_TYPE (arg0
))
8380 && TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (arg0
))))
8381 && TREE_CODE (TREE_OPERAND (arg0
, 1)) == INTEGER_CST
8382 && !TREE_OVERFLOW (TREE_OPERAND (arg0
, 1))
8383 && TREE_CODE (arg1
) == INTEGER_CST
8384 && !TREE_OVERFLOW (arg1
))
8386 const enum tree_code
8387 reverse_op
= TREE_CODE (arg0
) == PLUS_EXPR
? MINUS_EXPR
: PLUS_EXPR
;
8388 tree const1
= TREE_OPERAND (arg0
, 1);
8389 tree const2
= fold_convert_loc (loc
, TREE_TYPE (const1
), arg1
);
8390 tree variable
= TREE_OPERAND (arg0
, 0);
8391 tree new_const
= int_const_binop (reverse_op
, const2
, const1
);
8393 /* If the constant operation overflowed this can be
8394 simplified as a comparison against INT_MAX/INT_MIN. */
8395 if (TREE_OVERFLOW (new_const
)
8396 && !TYPE_OVERFLOW_WRAPS (TREE_TYPE (arg0
)))
8398 int const1_sgn
= tree_int_cst_sgn (const1
);
8399 enum tree_code code2
= code
;
8401 /* Get the sign of the constant on the lhs if the
8402 operation were VARIABLE + CONST1. */
8403 if (TREE_CODE (arg0
) == MINUS_EXPR
)
8404 const1_sgn
= -const1_sgn
;
8406 /* The sign of the constant determines if we overflowed
8407 INT_MAX (const1_sgn == -1) or INT_MIN (const1_sgn == 1).
8408 Canonicalize to the INT_MIN overflow by swapping the comparison
8410 if (const1_sgn
== -1)
8411 code2
= swap_tree_comparison (code
);
8413 /* We now can look at the canonicalized case
8414 VARIABLE + 1 CODE2 INT_MIN
8415 and decide on the result. */
8422 omit_one_operand_loc (loc
, type
, boolean_false_node
, variable
);
8428 omit_one_operand_loc (loc
, type
, boolean_true_node
, variable
);
8437 fold_overflow_warning ("assuming signed overflow does not occur "
8438 "when changing X +- C1 cmp C2 to "
8440 WARN_STRICT_OVERFLOW_COMPARISON
);
8441 return fold_build2_loc (loc
, code
, type
, variable
, new_const
);
8445 /* For comparisons of pointers we can decompose it to a compile time
8446 comparison of the base objects and the offsets into the object.
8447 This requires at least one operand being an ADDR_EXPR or a
8448 POINTER_PLUS_EXPR to do more than the operand_equal_p test below. */
8449 if (POINTER_TYPE_P (TREE_TYPE (arg0
))
8450 && (TREE_CODE (arg0
) == ADDR_EXPR
8451 || TREE_CODE (arg1
) == ADDR_EXPR
8452 || TREE_CODE (arg0
) == POINTER_PLUS_EXPR
8453 || TREE_CODE (arg1
) == POINTER_PLUS_EXPR
))
8455 tree base0
, base1
, offset0
= NULL_TREE
, offset1
= NULL_TREE
;
8456 HOST_WIDE_INT bitsize
, bitpos0
= 0, bitpos1
= 0;
8458 int volatilep
, reversep
, unsignedp
;
8459 bool indirect_base0
= false, indirect_base1
= false;
8461 /* Get base and offset for the access. Strip ADDR_EXPR for
8462 get_inner_reference, but put it back by stripping INDIRECT_REF
8463 off the base object if possible. indirect_baseN will be true
8464 if baseN is not an address but refers to the object itself. */
8466 if (TREE_CODE (arg0
) == ADDR_EXPR
)
8469 = get_inner_reference (TREE_OPERAND (arg0
, 0),
8470 &bitsize
, &bitpos0
, &offset0
, &mode
,
8471 &unsignedp
, &reversep
, &volatilep
, false);
8472 if (TREE_CODE (base0
) == INDIRECT_REF
)
8473 base0
= TREE_OPERAND (base0
, 0);
8475 indirect_base0
= true;
8477 else if (TREE_CODE (arg0
) == POINTER_PLUS_EXPR
)
8479 base0
= TREE_OPERAND (arg0
, 0);
8480 STRIP_SIGN_NOPS (base0
);
8481 if (TREE_CODE (base0
) == ADDR_EXPR
)
8484 = get_inner_reference (TREE_OPERAND (base0
, 0),
8485 &bitsize
, &bitpos0
, &offset0
, &mode
,
8486 &unsignedp
, &reversep
, &volatilep
,
8488 if (TREE_CODE (base0
) == INDIRECT_REF
)
8489 base0
= TREE_OPERAND (base0
, 0);
8491 indirect_base0
= true;
8493 if (offset0
== NULL_TREE
|| integer_zerop (offset0
))
8494 offset0
= TREE_OPERAND (arg0
, 1);
8496 offset0
= size_binop (PLUS_EXPR
, offset0
,
8497 TREE_OPERAND (arg0
, 1));
8498 if (TREE_CODE (offset0
) == INTEGER_CST
)
8500 offset_int tem
= wi::sext (wi::to_offset (offset0
),
8501 TYPE_PRECISION (sizetype
));
8502 tem
= wi::lshift (tem
, LOG2_BITS_PER_UNIT
);
8504 if (wi::fits_shwi_p (tem
))
8506 bitpos0
= tem
.to_shwi ();
8507 offset0
= NULL_TREE
;
8513 if (TREE_CODE (arg1
) == ADDR_EXPR
)
8516 = get_inner_reference (TREE_OPERAND (arg1
, 0),
8517 &bitsize
, &bitpos1
, &offset1
, &mode
,
8518 &unsignedp
, &reversep
, &volatilep
, false);
8519 if (TREE_CODE (base1
) == INDIRECT_REF
)
8520 base1
= TREE_OPERAND (base1
, 0);
8522 indirect_base1
= true;
8524 else if (TREE_CODE (arg1
) == POINTER_PLUS_EXPR
)
8526 base1
= TREE_OPERAND (arg1
, 0);
8527 STRIP_SIGN_NOPS (base1
);
8528 if (TREE_CODE (base1
) == ADDR_EXPR
)
8531 = get_inner_reference (TREE_OPERAND (base1
, 0),
8532 &bitsize
, &bitpos1
, &offset1
, &mode
,
8533 &unsignedp
, &reversep
, &volatilep
,
8535 if (TREE_CODE (base1
) == INDIRECT_REF
)
8536 base1
= TREE_OPERAND (base1
, 0);
8538 indirect_base1
= true;
8540 if (offset1
== NULL_TREE
|| integer_zerop (offset1
))
8541 offset1
= TREE_OPERAND (arg1
, 1);
8543 offset1
= size_binop (PLUS_EXPR
, offset1
,
8544 TREE_OPERAND (arg1
, 1));
8545 if (TREE_CODE (offset1
) == INTEGER_CST
)
8547 offset_int tem
= wi::sext (wi::to_offset (offset1
),
8548 TYPE_PRECISION (sizetype
));
8549 tem
= wi::lshift (tem
, LOG2_BITS_PER_UNIT
);
8551 if (wi::fits_shwi_p (tem
))
8553 bitpos1
= tem
.to_shwi ();
8554 offset1
= NULL_TREE
;
8559 /* If we have equivalent bases we might be able to simplify. */
8560 if (indirect_base0
== indirect_base1
8561 && operand_equal_p (base0
, base1
,
8562 indirect_base0
? OEP_ADDRESS_OF
: 0))
8564 /* We can fold this expression to a constant if the non-constant
8565 offset parts are equal. */
8566 if ((offset0
== offset1
8567 || (offset0
&& offset1
8568 && operand_equal_p (offset0
, offset1
, 0)))
8571 || (indirect_base0
&& DECL_P (base0
))
8572 || POINTER_TYPE_OVERFLOW_UNDEFINED
))
8576 && bitpos0
!= bitpos1
8577 && (pointer_may_wrap_p (base0
, offset0
, bitpos0
)
8578 || pointer_may_wrap_p (base1
, offset1
, bitpos1
)))
8579 fold_overflow_warning (("assuming pointer wraparound does not "
8580 "occur when comparing P +- C1 with "
8582 WARN_STRICT_OVERFLOW_CONDITIONAL
);
8587 return constant_boolean_node (bitpos0
== bitpos1
, type
);
8589 return constant_boolean_node (bitpos0
!= bitpos1
, type
);
8591 return constant_boolean_node (bitpos0
< bitpos1
, type
);
8593 return constant_boolean_node (bitpos0
<= bitpos1
, type
);
8595 return constant_boolean_node (bitpos0
>= bitpos1
, type
);
8597 return constant_boolean_node (bitpos0
> bitpos1
, type
);
8601 /* We can simplify the comparison to a comparison of the variable
8602 offset parts if the constant offset parts are equal.
8603 Be careful to use signed sizetype here because otherwise we
8604 mess with array offsets in the wrong way. This is possible
8605 because pointer arithmetic is restricted to retain within an
8606 object and overflow on pointer differences is undefined as of
8607 6.5.6/8 and /9 with respect to the signed ptrdiff_t. */
8608 else if (bitpos0
== bitpos1
8610 || (indirect_base0
&& DECL_P (base0
))
8611 || POINTER_TYPE_OVERFLOW_UNDEFINED
))
8613 /* By converting to signed sizetype we cover middle-end pointer
8614 arithmetic which operates on unsigned pointer types of size
8615 type size and ARRAY_REF offsets which are properly sign or
8616 zero extended from their type in case it is narrower than
8618 if (offset0
== NULL_TREE
)
8619 offset0
= build_int_cst (ssizetype
, 0);
8621 offset0
= fold_convert_loc (loc
, ssizetype
, offset0
);
8622 if (offset1
== NULL_TREE
)
8623 offset1
= build_int_cst (ssizetype
, 0);
8625 offset1
= fold_convert_loc (loc
, ssizetype
, offset1
);
8628 && (pointer_may_wrap_p (base0
, offset0
, bitpos0
)
8629 || pointer_may_wrap_p (base1
, offset1
, bitpos1
)))
8630 fold_overflow_warning (("assuming pointer wraparound does not "
8631 "occur when comparing P +- C1 with "
8633 WARN_STRICT_OVERFLOW_COMPARISON
);
8635 return fold_build2_loc (loc
, code
, type
, offset0
, offset1
);
8638 /* For equal offsets we can simplify to a comparison of the
8640 else if (bitpos0
== bitpos1
8642 ? base0
!= TREE_OPERAND (arg0
, 0) : base0
!= arg0
)
8644 ? base1
!= TREE_OPERAND (arg1
, 0) : base1
!= arg1
)
8645 && ((offset0
== offset1
)
8646 || (offset0
&& offset1
8647 && operand_equal_p (offset0
, offset1
, 0))))
8650 base0
= build_fold_addr_expr_loc (loc
, base0
);
8652 base1
= build_fold_addr_expr_loc (loc
, base1
);
8653 return fold_build2_loc (loc
, code
, type
, base0
, base1
);
8655 /* Comparison between an ordinary (non-weak) symbol and a null
8656 pointer can be eliminated since such symbols must have a non
8657 null address. In C, relational expressions between pointers
8658 to objects and null pointers are undefined. The results
8659 below follow the C++ rules with the additional property that
8660 every object pointer compares greater than a null pointer.
8662 else if (DECL_P (base0
)
8663 && maybe_nonzero_address (base0
) > 0
8664 /* Avoid folding references to struct members at offset 0 to
8665 prevent tests like '&ptr->firstmember == 0' from getting
8666 eliminated. When ptr is null, although the -> expression
8667 is strictly speaking invalid, GCC retains it as a matter
8668 of QoI. See PR c/44555. */
8669 && (offset0
== NULL_TREE
&& bitpos0
!= 0)
8670 /* The caller guarantees that when one of the arguments is
8671 constant (i.e., null in this case) it is second. */
8672 && integer_zerop (arg1
))
8679 return constant_boolean_node (false, type
);
8683 return constant_boolean_node (true, type
);
8690 /* Transform comparisons of the form X +- C1 CMP Y +- C2 to
8691 X CMP Y +- C2 +- C1 for signed X, Y. This is valid if
8692 the resulting offset is smaller in absolute value than the
8693 original one and has the same sign. */
8694 if (ANY_INTEGRAL_TYPE_P (TREE_TYPE (arg0
))
8695 && TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (arg0
))
8696 && (TREE_CODE (arg0
) == PLUS_EXPR
|| TREE_CODE (arg0
) == MINUS_EXPR
)
8697 && (TREE_CODE (TREE_OPERAND (arg0
, 1)) == INTEGER_CST
8698 && !TREE_OVERFLOW (TREE_OPERAND (arg0
, 1)))
8699 && (TREE_CODE (arg1
) == PLUS_EXPR
|| TREE_CODE (arg1
) == MINUS_EXPR
)
8700 && (TREE_CODE (TREE_OPERAND (arg1
, 1)) == INTEGER_CST
8701 && !TREE_OVERFLOW (TREE_OPERAND (arg1
, 1))))
8703 tree const1
= TREE_OPERAND (arg0
, 1);
8704 tree const2
= TREE_OPERAND (arg1
, 1);
8705 tree variable1
= TREE_OPERAND (arg0
, 0);
8706 tree variable2
= TREE_OPERAND (arg1
, 0);
8708 const char * const warnmsg
= G_("assuming signed overflow does not "
8709 "occur when combining constants around "
8712 /* Put the constant on the side where it doesn't overflow and is
8713 of lower absolute value and of same sign than before. */
8714 cst
= int_const_binop (TREE_CODE (arg0
) == TREE_CODE (arg1
)
8715 ? MINUS_EXPR
: PLUS_EXPR
,
8717 if (!TREE_OVERFLOW (cst
)
8718 && tree_int_cst_compare (const2
, cst
) == tree_int_cst_sgn (const2
)
8719 && tree_int_cst_sgn (cst
) == tree_int_cst_sgn (const2
))
8721 fold_overflow_warning (warnmsg
, WARN_STRICT_OVERFLOW_COMPARISON
);
8722 return fold_build2_loc (loc
, code
, type
,
8724 fold_build2_loc (loc
, TREE_CODE (arg1
),
8729 cst
= int_const_binop (TREE_CODE (arg0
) == TREE_CODE (arg1
)
8730 ? MINUS_EXPR
: PLUS_EXPR
,
8732 if (!TREE_OVERFLOW (cst
)
8733 && tree_int_cst_compare (const1
, cst
) == tree_int_cst_sgn (const1
)
8734 && tree_int_cst_sgn (cst
) == tree_int_cst_sgn (const1
))
8736 fold_overflow_warning (warnmsg
, WARN_STRICT_OVERFLOW_COMPARISON
);
8737 return fold_build2_loc (loc
, code
, type
,
8738 fold_build2_loc (loc
, TREE_CODE (arg0
),
8745 tem
= maybe_canonicalize_comparison (loc
, code
, type
, arg0
, arg1
);
8749 /* If this is comparing a constant with a MIN_EXPR or a MAX_EXPR of a
8750 constant, we can simplify it. */
8751 if (TREE_CODE (arg1
) == INTEGER_CST
8752 && (TREE_CODE (arg0
) == MIN_EXPR
8753 || TREE_CODE (arg0
) == MAX_EXPR
)
8754 && TREE_CODE (TREE_OPERAND (arg0
, 1)) == INTEGER_CST
)
8756 tem
= optimize_minmax_comparison (loc
, code
, type
, op0
, op1
);
8761 /* If we are comparing an expression that just has comparisons
8762 of two integer values, arithmetic expressions of those comparisons,
8763 and constants, we can simplify it. There are only three cases
8764 to check: the two values can either be equal, the first can be
8765 greater, or the second can be greater. Fold the expression for
8766 those three values. Since each value must be 0 or 1, we have
8767 eight possibilities, each of which corresponds to the constant 0
8768 or 1 or one of the six possible comparisons.
8770 This handles common cases like (a > b) == 0 but also handles
8771 expressions like ((x > y) - (y > x)) > 0, which supposedly
8772 occur in macroized code. */
8774 if (TREE_CODE (arg1
) == INTEGER_CST
&& TREE_CODE (arg0
) != INTEGER_CST
)
8776 tree cval1
= 0, cval2
= 0;
8779 if (twoval_comparison_p (arg0
, &cval1
, &cval2
, &save_p
)
8780 /* Don't handle degenerate cases here; they should already
8781 have been handled anyway. */
8782 && cval1
!= 0 && cval2
!= 0
8783 && ! (TREE_CONSTANT (cval1
) && TREE_CONSTANT (cval2
))
8784 && TREE_TYPE (cval1
) == TREE_TYPE (cval2
)
8785 && INTEGRAL_TYPE_P (TREE_TYPE (cval1
))
8786 && TYPE_MAX_VALUE (TREE_TYPE (cval1
))
8787 && TYPE_MAX_VALUE (TREE_TYPE (cval2
))
8788 && ! operand_equal_p (TYPE_MIN_VALUE (TREE_TYPE (cval1
)),
8789 TYPE_MAX_VALUE (TREE_TYPE (cval2
)), 0))
8791 tree maxval
= TYPE_MAX_VALUE (TREE_TYPE (cval1
));
8792 tree minval
= TYPE_MIN_VALUE (TREE_TYPE (cval1
));
8794 /* We can't just pass T to eval_subst in case cval1 or cval2
8795 was the same as ARG1. */
8798 = fold_build2_loc (loc
, code
, type
,
8799 eval_subst (loc
, arg0
, cval1
, maxval
,
8803 = fold_build2_loc (loc
, code
, type
,
8804 eval_subst (loc
, arg0
, cval1
, maxval
,
8808 = fold_build2_loc (loc
, code
, type
,
8809 eval_subst (loc
, arg0
, cval1
, minval
,
8813 /* All three of these results should be 0 or 1. Confirm they are.
8814 Then use those values to select the proper code to use. */
8816 if (TREE_CODE (high_result
) == INTEGER_CST
8817 && TREE_CODE (equal_result
) == INTEGER_CST
8818 && TREE_CODE (low_result
) == INTEGER_CST
)
8820 /* Make a 3-bit mask with the high-order bit being the
8821 value for `>', the next for '=', and the low for '<'. */
8822 switch ((integer_onep (high_result
) * 4)
8823 + (integer_onep (equal_result
) * 2)
8824 + integer_onep (low_result
))
8828 return omit_one_operand_loc (loc
, type
, integer_zero_node
, arg0
);
8849 return omit_one_operand_loc (loc
, type
, integer_one_node
, arg0
);
8854 tem
= save_expr (build2 (code
, type
, cval1
, cval2
));
8855 SET_EXPR_LOCATION (tem
, loc
);
8858 return fold_build2_loc (loc
, code
, type
, cval1
, cval2
);
8863 /* We can fold X/C1 op C2 where C1 and C2 are integer constants
8864 into a single range test. */
8865 if ((TREE_CODE (arg0
) == TRUNC_DIV_EXPR
8866 || TREE_CODE (arg0
) == EXACT_DIV_EXPR
)
8867 && TREE_CODE (arg1
) == INTEGER_CST
8868 && TREE_CODE (TREE_OPERAND (arg0
, 1)) == INTEGER_CST
8869 && !integer_zerop (TREE_OPERAND (arg0
, 1))
8870 && !TREE_OVERFLOW (TREE_OPERAND (arg0
, 1))
8871 && !TREE_OVERFLOW (arg1
))
8873 tem
= fold_div_compare (loc
, code
, type
, arg0
, arg1
);
8874 if (tem
!= NULL_TREE
)
8882 /* Subroutine of fold_binary. Optimize complex multiplications of the
8883 form z * conj(z), as pow(realpart(z),2) + pow(imagpart(z),2). The
8884 argument EXPR represents the expression "z" of type TYPE. */
8887 fold_mult_zconjz (location_t loc
, tree type
, tree expr
)
8889 tree itype
= TREE_TYPE (type
);
8890 tree rpart
, ipart
, tem
;
8892 if (TREE_CODE (expr
) == COMPLEX_EXPR
)
8894 rpart
= TREE_OPERAND (expr
, 0);
8895 ipart
= TREE_OPERAND (expr
, 1);
8897 else if (TREE_CODE (expr
) == COMPLEX_CST
)
8899 rpart
= TREE_REALPART (expr
);
8900 ipart
= TREE_IMAGPART (expr
);
8904 expr
= save_expr (expr
);
8905 rpart
= fold_build1_loc (loc
, REALPART_EXPR
, itype
, expr
);
8906 ipart
= fold_build1_loc (loc
, IMAGPART_EXPR
, itype
, expr
);
8909 rpart
= save_expr (rpart
);
8910 ipart
= save_expr (ipart
);
8911 tem
= fold_build2_loc (loc
, PLUS_EXPR
, itype
,
8912 fold_build2_loc (loc
, MULT_EXPR
, itype
, rpart
, rpart
),
8913 fold_build2_loc (loc
, MULT_EXPR
, itype
, ipart
, ipart
));
8914 return fold_build2_loc (loc
, COMPLEX_EXPR
, type
, tem
,
8915 build_zero_cst (itype
));
8919 /* Helper function for fold_vec_perm. Store elements of VECTOR_CST or
8920 CONSTRUCTOR ARG into array ELTS and return true if successful. */
8923 vec_cst_ctor_to_array (tree arg
, tree
*elts
)
8925 unsigned int nelts
= TYPE_VECTOR_SUBPARTS (TREE_TYPE (arg
)), i
;
8927 if (TREE_CODE (arg
) == VECTOR_CST
)
8929 for (i
= 0; i
< VECTOR_CST_NELTS (arg
); ++i
)
8930 elts
[i
] = VECTOR_CST_ELT (arg
, i
);
8932 else if (TREE_CODE (arg
) == CONSTRUCTOR
)
8934 constructor_elt
*elt
;
8936 FOR_EACH_VEC_SAFE_ELT (CONSTRUCTOR_ELTS (arg
), i
, elt
)
8937 if (i
>= nelts
|| TREE_CODE (TREE_TYPE (elt
->value
)) == VECTOR_TYPE
)
8940 elts
[i
] = elt
->value
;
8944 for (; i
< nelts
; i
++)
8946 = fold_convert (TREE_TYPE (TREE_TYPE (arg
)), integer_zero_node
);
8950 /* Attempt to fold vector permutation of ARG0 and ARG1 vectors using SEL
8951 selector. Return the folded VECTOR_CST or CONSTRUCTOR if successful,
8952 NULL_TREE otherwise. */
8955 fold_vec_perm (tree type
, tree arg0
, tree arg1
, const unsigned char *sel
)
8957 unsigned int nelts
= TYPE_VECTOR_SUBPARTS (type
), i
;
8959 bool need_ctor
= false;
8961 gcc_assert (TYPE_VECTOR_SUBPARTS (TREE_TYPE (arg0
)) == nelts
8962 && TYPE_VECTOR_SUBPARTS (TREE_TYPE (arg1
)) == nelts
);
8963 if (TREE_TYPE (TREE_TYPE (arg0
)) != TREE_TYPE (type
)
8964 || TREE_TYPE (TREE_TYPE (arg1
)) != TREE_TYPE (type
))
8967 elts
= XALLOCAVEC (tree
, nelts
* 3);
8968 if (!vec_cst_ctor_to_array (arg0
, elts
)
8969 || !vec_cst_ctor_to_array (arg1
, elts
+ nelts
))
8972 for (i
= 0; i
< nelts
; i
++)
8974 if (!CONSTANT_CLASS_P (elts
[sel
[i
]]))
8976 elts
[i
+ 2 * nelts
] = unshare_expr (elts
[sel
[i
]]);
8981 vec
<constructor_elt
, va_gc
> *v
;
8982 vec_alloc (v
, nelts
);
8983 for (i
= 0; i
< nelts
; i
++)
8984 CONSTRUCTOR_APPEND_ELT (v
, NULL_TREE
, elts
[2 * nelts
+ i
]);
8985 return build_constructor (type
, v
);
8988 return build_vector (type
, &elts
[2 * nelts
]);
8991 /* Try to fold a pointer difference of type TYPE two address expressions of
8992 array references AREF0 and AREF1 using location LOC. Return a
8993 simplified expression for the difference or NULL_TREE. */
8996 fold_addr_of_array_ref_difference (location_t loc
, tree type
,
8997 tree aref0
, tree aref1
)
8999 tree base0
= TREE_OPERAND (aref0
, 0);
9000 tree base1
= TREE_OPERAND (aref1
, 0);
9001 tree base_offset
= build_int_cst (type
, 0);
9003 /* If the bases are array references as well, recurse. If the bases
9004 are pointer indirections compute the difference of the pointers.
9005 If the bases are equal, we are set. */
9006 if ((TREE_CODE (base0
) == ARRAY_REF
9007 && TREE_CODE (base1
) == ARRAY_REF
9009 = fold_addr_of_array_ref_difference (loc
, type
, base0
, base1
)))
9010 || (INDIRECT_REF_P (base0
)
9011 && INDIRECT_REF_P (base1
)
9013 = fold_binary_loc (loc
, MINUS_EXPR
, type
,
9014 fold_convert (type
, TREE_OPERAND (base0
, 0)),
9016 TREE_OPERAND (base1
, 0)))))
9017 || operand_equal_p (base0
, base1
, OEP_ADDRESS_OF
))
9019 tree op0
= fold_convert_loc (loc
, type
, TREE_OPERAND (aref0
, 1));
9020 tree op1
= fold_convert_loc (loc
, type
, TREE_OPERAND (aref1
, 1));
9021 tree esz
= fold_convert_loc (loc
, type
, array_ref_element_size (aref0
));
9022 tree diff
= build2 (MINUS_EXPR
, type
, op0
, op1
);
9023 return fold_build2_loc (loc
, PLUS_EXPR
, type
,
9025 fold_build2_loc (loc
, MULT_EXPR
, type
,
9031 /* If the real or vector real constant CST of type TYPE has an exact
9032 inverse, return it, else return NULL. */
9035 exact_inverse (tree type
, tree cst
)
9038 tree unit_type
, *elts
;
9040 unsigned vec_nelts
, i
;
9042 switch (TREE_CODE (cst
))
9045 r
= TREE_REAL_CST (cst
);
9047 if (exact_real_inverse (TYPE_MODE (type
), &r
))
9048 return build_real (type
, r
);
9053 vec_nelts
= VECTOR_CST_NELTS (cst
);
9054 elts
= XALLOCAVEC (tree
, vec_nelts
);
9055 unit_type
= TREE_TYPE (type
);
9056 mode
= TYPE_MODE (unit_type
);
9058 for (i
= 0; i
< vec_nelts
; i
++)
9060 r
= TREE_REAL_CST (VECTOR_CST_ELT (cst
, i
));
9061 if (!exact_real_inverse (mode
, &r
))
9063 elts
[i
] = build_real (unit_type
, r
);
9066 return build_vector (type
, elts
);
9073 /* Mask out the tz least significant bits of X of type TYPE where
9074 tz is the number of trailing zeroes in Y. */
9076 mask_with_tz (tree type
, const wide_int
&x
, const wide_int
&y
)
9078 int tz
= wi::ctz (y
);
9080 return wi::mask (tz
, true, TYPE_PRECISION (type
)) & x
;
9084 /* Return true when T is an address and is known to be nonzero.
9085 For floating point we further ensure that T is not denormal.
9086 Similar logic is present in nonzero_address in rtlanal.h.
9088 If the return value is based on the assumption that signed overflow
9089 is undefined, set *STRICT_OVERFLOW_P to true; otherwise, don't
9090 change *STRICT_OVERFLOW_P. */
9093 tree_expr_nonzero_warnv_p (tree t
, bool *strict_overflow_p
)
9095 tree type
= TREE_TYPE (t
);
9096 enum tree_code code
;
9098 /* Doing something useful for floating point would need more work. */
9099 if (!INTEGRAL_TYPE_P (type
) && !POINTER_TYPE_P (type
))
9102 code
= TREE_CODE (t
);
9103 switch (TREE_CODE_CLASS (code
))
9106 return tree_unary_nonzero_warnv_p (code
, type
, TREE_OPERAND (t
, 0),
9109 case tcc_comparison
:
9110 return tree_binary_nonzero_warnv_p (code
, type
,
9111 TREE_OPERAND (t
, 0),
9112 TREE_OPERAND (t
, 1),
9115 case tcc_declaration
:
9117 return tree_single_nonzero_warnv_p (t
, strict_overflow_p
);
9125 case TRUTH_NOT_EXPR
:
9126 return tree_unary_nonzero_warnv_p (code
, type
, TREE_OPERAND (t
, 0),
9129 case TRUTH_AND_EXPR
:
9131 case TRUTH_XOR_EXPR
:
9132 return tree_binary_nonzero_warnv_p (code
, type
,
9133 TREE_OPERAND (t
, 0),
9134 TREE_OPERAND (t
, 1),
9142 case WITH_SIZE_EXPR
:
9144 return tree_single_nonzero_warnv_p (t
, strict_overflow_p
);
9149 return tree_expr_nonzero_warnv_p (TREE_OPERAND (t
, 1),
9153 return tree_expr_nonzero_warnv_p (TREE_OPERAND (t
, 0),
9158 tree fndecl
= get_callee_fndecl (t
);
9159 if (!fndecl
) return false;
9160 if (flag_delete_null_pointer_checks
&& !flag_check_new
9161 && DECL_IS_OPERATOR_NEW (fndecl
)
9162 && !TREE_NOTHROW (fndecl
))
9164 if (flag_delete_null_pointer_checks
9165 && lookup_attribute ("returns_nonnull",
9166 TYPE_ATTRIBUTES (TREE_TYPE (fndecl
))))
9168 return alloca_call_p (t
);
9177 /* Return true when T is an address and is known to be nonzero.
9178 Handle warnings about undefined signed overflow. */
9181 tree_expr_nonzero_p (tree t
)
9183 bool ret
, strict_overflow_p
;
9185 strict_overflow_p
= false;
9186 ret
= tree_expr_nonzero_warnv_p (t
, &strict_overflow_p
);
9187 if (strict_overflow_p
)
9188 fold_overflow_warning (("assuming signed overflow does not occur when "
9189 "determining that expression is always "
9191 WARN_STRICT_OVERFLOW_MISC
);
9195 /* Return true if T is known not to be equal to an integer W. */
9198 expr_not_equal_to (tree t
, const wide_int
&w
)
9200 wide_int min
, max
, nz
;
9201 value_range_type rtype
;
9202 switch (TREE_CODE (t
))
9205 return wi::ne_p (t
, w
);
9208 if (!INTEGRAL_TYPE_P (TREE_TYPE (t
)))
9210 rtype
= get_range_info (t
, &min
, &max
);
9211 if (rtype
== VR_RANGE
)
9213 if (wi::lt_p (max
, w
, TYPE_SIGN (TREE_TYPE (t
))))
9215 if (wi::lt_p (w
, min
, TYPE_SIGN (TREE_TYPE (t
))))
9218 else if (rtype
== VR_ANTI_RANGE
9219 && wi::le_p (min
, w
, TYPE_SIGN (TREE_TYPE (t
)))
9220 && wi::le_p (w
, max
, TYPE_SIGN (TREE_TYPE (t
))))
9222 /* If T has some known zero bits and W has any of those bits set,
9223 then T is known not to be equal to W. */
9224 if (wi::ne_p (wi::zext (wi::bit_and_not (w
, get_nonzero_bits (t
)),
9225 TYPE_PRECISION (TREE_TYPE (t
))), 0))
9234 /* Fold a binary expression of code CODE and type TYPE with operands
9235 OP0 and OP1. LOC is the location of the resulting expression.
9236 Return the folded expression if folding is successful. Otherwise,
9237 return NULL_TREE. */
9240 fold_binary_loc (location_t loc
,
9241 enum tree_code code
, tree type
, tree op0
, tree op1
)
9243 enum tree_code_class kind
= TREE_CODE_CLASS (code
);
9244 tree arg0
, arg1
, tem
;
9245 tree t1
= NULL_TREE
;
9246 bool strict_overflow_p
;
9249 gcc_assert (IS_EXPR_CODE_CLASS (kind
)
9250 && TREE_CODE_LENGTH (code
) == 2
9252 && op1
!= NULL_TREE
);
9257 /* Strip any conversions that don't change the mode. This is
9258 safe for every expression, except for a comparison expression
9259 because its signedness is derived from its operands. So, in
9260 the latter case, only strip conversions that don't change the
9261 signedness. MIN_EXPR/MAX_EXPR also need signedness of arguments
9264 Note that this is done as an internal manipulation within the
9265 constant folder, in order to find the simplest representation
9266 of the arguments so that their form can be studied. In any
9267 cases, the appropriate type conversions should be put back in
9268 the tree that will get out of the constant folder. */
9270 if (kind
== tcc_comparison
|| code
== MIN_EXPR
|| code
== MAX_EXPR
)
9272 STRIP_SIGN_NOPS (arg0
);
9273 STRIP_SIGN_NOPS (arg1
);
9281 /* Note that TREE_CONSTANT isn't enough: static var addresses are
9282 constant but we can't do arithmetic on them. */
9283 if (CONSTANT_CLASS_P (arg0
) && CONSTANT_CLASS_P (arg1
))
9285 tem
= const_binop (code
, type
, arg0
, arg1
);
9286 if (tem
!= NULL_TREE
)
9288 if (TREE_TYPE (tem
) != type
)
9289 tem
= fold_convert_loc (loc
, type
, tem
);
9294 /* If this is a commutative operation, and ARG0 is a constant, move it
9295 to ARG1 to reduce the number of tests below. */
9296 if (commutative_tree_code (code
)
9297 && tree_swap_operands_p (arg0
, arg1
, true))
9298 return fold_build2_loc (loc
, code
, type
, op1
, op0
);
9300 /* Likewise if this is a comparison, and ARG0 is a constant, move it
9301 to ARG1 to reduce the number of tests below. */
9302 if (kind
== tcc_comparison
9303 && tree_swap_operands_p (arg0
, arg1
, true))
9304 return fold_build2_loc (loc
, swap_tree_comparison (code
), type
, op1
, op0
);
9306 tem
= generic_simplify (loc
, code
, type
, op0
, op1
);
9310 /* ARG0 is the first operand of EXPR, and ARG1 is the second operand.
9312 First check for cases where an arithmetic operation is applied to a
9313 compound, conditional, or comparison operation. Push the arithmetic
9314 operation inside the compound or conditional to see if any folding
9315 can then be done. Convert comparison to conditional for this purpose.
9316 The also optimizes non-constant cases that used to be done in
9319 Before we do that, see if this is a BIT_AND_EXPR or a BIT_IOR_EXPR,
9320 one of the operands is a comparison and the other is a comparison, a
9321 BIT_AND_EXPR with the constant 1, or a truth value. In that case, the
9322 code below would make the expression more complex. Change it to a
9323 TRUTH_{AND,OR}_EXPR. Likewise, convert a similar NE_EXPR to
9324 TRUTH_XOR_EXPR and an EQ_EXPR to the inversion of a TRUTH_XOR_EXPR. */
9326 if ((code
== BIT_AND_EXPR
|| code
== BIT_IOR_EXPR
9327 || code
== EQ_EXPR
|| code
== NE_EXPR
)
9328 && TREE_CODE (type
) != VECTOR_TYPE
9329 && ((truth_value_p (TREE_CODE (arg0
))
9330 && (truth_value_p (TREE_CODE (arg1
))
9331 || (TREE_CODE (arg1
) == BIT_AND_EXPR
9332 && integer_onep (TREE_OPERAND (arg1
, 1)))))
9333 || (truth_value_p (TREE_CODE (arg1
))
9334 && (truth_value_p (TREE_CODE (arg0
))
9335 || (TREE_CODE (arg0
) == BIT_AND_EXPR
9336 && integer_onep (TREE_OPERAND (arg0
, 1)))))))
9338 tem
= fold_build2_loc (loc
, code
== BIT_AND_EXPR
? TRUTH_AND_EXPR
9339 : code
== BIT_IOR_EXPR
? TRUTH_OR_EXPR
9342 fold_convert_loc (loc
, boolean_type_node
, arg0
),
9343 fold_convert_loc (loc
, boolean_type_node
, arg1
));
9345 if (code
== EQ_EXPR
)
9346 tem
= invert_truthvalue_loc (loc
, tem
);
9348 return fold_convert_loc (loc
, type
, tem
);
9351 if (TREE_CODE_CLASS (code
) == tcc_binary
9352 || TREE_CODE_CLASS (code
) == tcc_comparison
)
9354 if (TREE_CODE (arg0
) == COMPOUND_EXPR
)
9356 tem
= fold_build2_loc (loc
, code
, type
,
9357 fold_convert_loc (loc
, TREE_TYPE (op0
),
9358 TREE_OPERAND (arg0
, 1)), op1
);
9359 return build2_loc (loc
, COMPOUND_EXPR
, type
, TREE_OPERAND (arg0
, 0),
9362 if (TREE_CODE (arg1
) == COMPOUND_EXPR
9363 && reorder_operands_p (arg0
, TREE_OPERAND (arg1
, 0)))
9365 tem
= fold_build2_loc (loc
, code
, type
, op0
,
9366 fold_convert_loc (loc
, TREE_TYPE (op1
),
9367 TREE_OPERAND (arg1
, 1)));
9368 return build2_loc (loc
, COMPOUND_EXPR
, type
, TREE_OPERAND (arg1
, 0),
9372 if (TREE_CODE (arg0
) == COND_EXPR
9373 || TREE_CODE (arg0
) == VEC_COND_EXPR
9374 || COMPARISON_CLASS_P (arg0
))
9376 tem
= fold_binary_op_with_conditional_arg (loc
, code
, type
, op0
, op1
,
9378 /*cond_first_p=*/1);
9379 if (tem
!= NULL_TREE
)
9383 if (TREE_CODE (arg1
) == COND_EXPR
9384 || TREE_CODE (arg1
) == VEC_COND_EXPR
9385 || COMPARISON_CLASS_P (arg1
))
9387 tem
= fold_binary_op_with_conditional_arg (loc
, code
, type
, op0
, op1
,
9389 /*cond_first_p=*/0);
9390 if (tem
!= NULL_TREE
)
9398 /* MEM[&MEM[p, CST1], CST2] -> MEM[p, CST1 + CST2]. */
9399 if (TREE_CODE (arg0
) == ADDR_EXPR
9400 && TREE_CODE (TREE_OPERAND (arg0
, 0)) == MEM_REF
)
9402 tree iref
= TREE_OPERAND (arg0
, 0);
9403 return fold_build2 (MEM_REF
, type
,
9404 TREE_OPERAND (iref
, 0),
9405 int_const_binop (PLUS_EXPR
, arg1
,
9406 TREE_OPERAND (iref
, 1)));
9409 /* MEM[&a.b, CST2] -> MEM[&a, offsetof (a, b) + CST2]. */
9410 if (TREE_CODE (arg0
) == ADDR_EXPR
9411 && handled_component_p (TREE_OPERAND (arg0
, 0)))
9414 HOST_WIDE_INT coffset
;
9415 base
= get_addr_base_and_unit_offset (TREE_OPERAND (arg0
, 0),
9419 return fold_build2 (MEM_REF
, type
,
9420 build_fold_addr_expr (base
),
9421 int_const_binop (PLUS_EXPR
, arg1
,
9422 size_int (coffset
)));
9427 case POINTER_PLUS_EXPR
:
9428 /* INT +p INT -> (PTR)(INT + INT). Stripping types allows for this. */
9429 if (INTEGRAL_TYPE_P (TREE_TYPE (arg1
))
9430 && INTEGRAL_TYPE_P (TREE_TYPE (arg0
)))
9431 return fold_convert_loc (loc
, type
,
9432 fold_build2_loc (loc
, PLUS_EXPR
, sizetype
,
9433 fold_convert_loc (loc
, sizetype
,
9435 fold_convert_loc (loc
, sizetype
,
9441 if (INTEGRAL_TYPE_P (type
) || VECTOR_INTEGER_TYPE_P (type
))
9443 /* X + (X / CST) * -CST is X % CST. */
9444 if (TREE_CODE (arg1
) == MULT_EXPR
9445 && TREE_CODE (TREE_OPERAND (arg1
, 0)) == TRUNC_DIV_EXPR
9446 && operand_equal_p (arg0
,
9447 TREE_OPERAND (TREE_OPERAND (arg1
, 0), 0), 0))
9449 tree cst0
= TREE_OPERAND (TREE_OPERAND (arg1
, 0), 1);
9450 tree cst1
= TREE_OPERAND (arg1
, 1);
9451 tree sum
= fold_binary_loc (loc
, PLUS_EXPR
, TREE_TYPE (cst1
),
9453 if (sum
&& integer_zerop (sum
))
9454 return fold_convert_loc (loc
, type
,
9455 fold_build2_loc (loc
, TRUNC_MOD_EXPR
,
9456 TREE_TYPE (arg0
), arg0
,
9461 /* Handle (A1 * C1) + (A2 * C2) with A1, A2 or C1, C2 being the same or
9462 one. Make sure the type is not saturating and has the signedness of
9463 the stripped operands, as fold_plusminus_mult_expr will re-associate.
9464 ??? The latter condition should use TYPE_OVERFLOW_* flags instead. */
9465 if ((TREE_CODE (arg0
) == MULT_EXPR
9466 || TREE_CODE (arg1
) == MULT_EXPR
)
9467 && !TYPE_SATURATING (type
)
9468 && TYPE_UNSIGNED (type
) == TYPE_UNSIGNED (TREE_TYPE (arg0
))
9469 && TYPE_UNSIGNED (type
) == TYPE_UNSIGNED (TREE_TYPE (arg1
))
9470 && (!FLOAT_TYPE_P (type
) || flag_associative_math
))
9472 tree tem
= fold_plusminus_mult_expr (loc
, code
, type
, arg0
, arg1
);
9477 if (! FLOAT_TYPE_P (type
))
9479 /* Reassociate (plus (plus (mult) (foo)) (mult)) as
9480 (plus (plus (mult) (mult)) (foo)) so that we can
9481 take advantage of the factoring cases below. */
9482 if (ANY_INTEGRAL_TYPE_P (type
)
9483 && TYPE_OVERFLOW_WRAPS (type
)
9484 && (((TREE_CODE (arg0
) == PLUS_EXPR
9485 || TREE_CODE (arg0
) == MINUS_EXPR
)
9486 && TREE_CODE (arg1
) == MULT_EXPR
)
9487 || ((TREE_CODE (arg1
) == PLUS_EXPR
9488 || TREE_CODE (arg1
) == MINUS_EXPR
)
9489 && TREE_CODE (arg0
) == MULT_EXPR
)))
9491 tree parg0
, parg1
, parg
, marg
;
9492 enum tree_code pcode
;
9494 if (TREE_CODE (arg1
) == MULT_EXPR
)
9495 parg
= arg0
, marg
= arg1
;
9497 parg
= arg1
, marg
= arg0
;
9498 pcode
= TREE_CODE (parg
);
9499 parg0
= TREE_OPERAND (parg
, 0);
9500 parg1
= TREE_OPERAND (parg
, 1);
9504 if (TREE_CODE (parg0
) == MULT_EXPR
9505 && TREE_CODE (parg1
) != MULT_EXPR
)
9506 return fold_build2_loc (loc
, pcode
, type
,
9507 fold_build2_loc (loc
, PLUS_EXPR
, type
,
9508 fold_convert_loc (loc
, type
,
9510 fold_convert_loc (loc
, type
,
9512 fold_convert_loc (loc
, type
, parg1
));
9513 if (TREE_CODE (parg0
) != MULT_EXPR
9514 && TREE_CODE (parg1
) == MULT_EXPR
)
9516 fold_build2_loc (loc
, PLUS_EXPR
, type
,
9517 fold_convert_loc (loc
, type
, parg0
),
9518 fold_build2_loc (loc
, pcode
, type
,
9519 fold_convert_loc (loc
, type
, marg
),
9520 fold_convert_loc (loc
, type
,
9526 /* Fold __complex__ ( x, 0 ) + __complex__ ( 0, y )
9527 to __complex__ ( x, y ). This is not the same for SNaNs or
9528 if signed zeros are involved. */
9529 if (!HONOR_SNANS (element_mode (arg0
))
9530 && !HONOR_SIGNED_ZEROS (element_mode (arg0
))
9531 && COMPLEX_FLOAT_TYPE_P (TREE_TYPE (arg0
)))
9533 tree rtype
= TREE_TYPE (TREE_TYPE (arg0
));
9534 tree arg0r
= fold_unary_loc (loc
, REALPART_EXPR
, rtype
, arg0
);
9535 tree arg0i
= fold_unary_loc (loc
, IMAGPART_EXPR
, rtype
, arg0
);
9536 bool arg0rz
= false, arg0iz
= false;
9537 if ((arg0r
&& (arg0rz
= real_zerop (arg0r
)))
9538 || (arg0i
&& (arg0iz
= real_zerop (arg0i
))))
9540 tree arg1r
= fold_unary_loc (loc
, REALPART_EXPR
, rtype
, arg1
);
9541 tree arg1i
= fold_unary_loc (loc
, IMAGPART_EXPR
, rtype
, arg1
);
9542 if (arg0rz
&& arg1i
&& real_zerop (arg1i
))
9544 tree rp
= arg1r
? arg1r
9545 : build1 (REALPART_EXPR
, rtype
, arg1
);
9546 tree ip
= arg0i
? arg0i
9547 : build1 (IMAGPART_EXPR
, rtype
, arg0
);
9548 return fold_build2_loc (loc
, COMPLEX_EXPR
, type
, rp
, ip
);
9550 else if (arg0iz
&& arg1r
&& real_zerop (arg1r
))
9552 tree rp
= arg0r
? arg0r
9553 : build1 (REALPART_EXPR
, rtype
, arg0
);
9554 tree ip
= arg1i
? arg1i
9555 : build1 (IMAGPART_EXPR
, rtype
, arg1
);
9556 return fold_build2_loc (loc
, COMPLEX_EXPR
, type
, rp
, ip
);
9561 if (flag_unsafe_math_optimizations
9562 && (TREE_CODE (arg0
) == RDIV_EXPR
|| TREE_CODE (arg0
) == MULT_EXPR
)
9563 && (TREE_CODE (arg1
) == RDIV_EXPR
|| TREE_CODE (arg1
) == MULT_EXPR
)
9564 && (tem
= distribute_real_division (loc
, code
, type
, arg0
, arg1
)))
9567 /* Convert a + (b*c + d*e) into (a + b*c) + d*e.
9568 We associate floats only if the user has specified
9569 -fassociative-math. */
9570 if (flag_associative_math
9571 && TREE_CODE (arg1
) == PLUS_EXPR
9572 && TREE_CODE (arg0
) != MULT_EXPR
)
9574 tree tree10
= TREE_OPERAND (arg1
, 0);
9575 tree tree11
= TREE_OPERAND (arg1
, 1);
9576 if (TREE_CODE (tree11
) == MULT_EXPR
9577 && TREE_CODE (tree10
) == MULT_EXPR
)
9580 tree0
= fold_build2_loc (loc
, PLUS_EXPR
, type
, arg0
, tree10
);
9581 return fold_build2_loc (loc
, PLUS_EXPR
, type
, tree0
, tree11
);
9584 /* Convert (b*c + d*e) + a into b*c + (d*e +a).
9585 We associate floats only if the user has specified
9586 -fassociative-math. */
9587 if (flag_associative_math
9588 && TREE_CODE (arg0
) == PLUS_EXPR
9589 && TREE_CODE (arg1
) != MULT_EXPR
)
9591 tree tree00
= TREE_OPERAND (arg0
, 0);
9592 tree tree01
= TREE_OPERAND (arg0
, 1);
9593 if (TREE_CODE (tree01
) == MULT_EXPR
9594 && TREE_CODE (tree00
) == MULT_EXPR
)
9597 tree0
= fold_build2_loc (loc
, PLUS_EXPR
, type
, tree01
, arg1
);
9598 return fold_build2_loc (loc
, PLUS_EXPR
, type
, tree00
, tree0
);
9604 /* (A << C1) + (A >> C2) if A is unsigned and C1+C2 is the size of A
9605 is a rotate of A by C1 bits. */
9606 /* (A << B) + (A >> (Z - B)) if A is unsigned and Z is the size of A
9607 is a rotate of A by B bits. */
9609 enum tree_code code0
, code1
;
9611 code0
= TREE_CODE (arg0
);
9612 code1
= TREE_CODE (arg1
);
9613 if (((code0
== RSHIFT_EXPR
&& code1
== LSHIFT_EXPR
)
9614 || (code1
== RSHIFT_EXPR
&& code0
== LSHIFT_EXPR
))
9615 && operand_equal_p (TREE_OPERAND (arg0
, 0),
9616 TREE_OPERAND (arg1
, 0), 0)
9617 && (rtype
= TREE_TYPE (TREE_OPERAND (arg0
, 0)),
9618 TYPE_UNSIGNED (rtype
))
9619 /* Only create rotates in complete modes. Other cases are not
9620 expanded properly. */
9621 && (element_precision (rtype
)
9622 == GET_MODE_UNIT_PRECISION (TYPE_MODE (rtype
))))
9624 tree tree01
, tree11
;
9625 enum tree_code code01
, code11
;
9627 tree01
= TREE_OPERAND (arg0
, 1);
9628 tree11
= TREE_OPERAND (arg1
, 1);
9629 STRIP_NOPS (tree01
);
9630 STRIP_NOPS (tree11
);
9631 code01
= TREE_CODE (tree01
);
9632 code11
= TREE_CODE (tree11
);
9633 if (code01
== INTEGER_CST
9634 && code11
== INTEGER_CST
9635 && (wi::to_widest (tree01
) + wi::to_widest (tree11
)
9636 == element_precision (TREE_TYPE (TREE_OPERAND (arg0
, 0)))))
9638 tem
= build2_loc (loc
, LROTATE_EXPR
,
9639 TREE_TYPE (TREE_OPERAND (arg0
, 0)),
9640 TREE_OPERAND (arg0
, 0),
9641 code0
== LSHIFT_EXPR
9642 ? TREE_OPERAND (arg0
, 1)
9643 : TREE_OPERAND (arg1
, 1));
9644 return fold_convert_loc (loc
, type
, tem
);
9646 else if (code11
== MINUS_EXPR
)
9648 tree tree110
, tree111
;
9649 tree110
= TREE_OPERAND (tree11
, 0);
9650 tree111
= TREE_OPERAND (tree11
, 1);
9651 STRIP_NOPS (tree110
);
9652 STRIP_NOPS (tree111
);
9653 if (TREE_CODE (tree110
) == INTEGER_CST
9654 && 0 == compare_tree_int (tree110
,
9656 (TREE_TYPE (TREE_OPERAND
9658 && operand_equal_p (tree01
, tree111
, 0))
9660 fold_convert_loc (loc
, type
,
9661 build2 ((code0
== LSHIFT_EXPR
9664 TREE_TYPE (TREE_OPERAND (arg0
, 0)),
9665 TREE_OPERAND (arg0
, 0),
9666 TREE_OPERAND (arg0
, 1)));
9668 else if (code01
== MINUS_EXPR
)
9670 tree tree010
, tree011
;
9671 tree010
= TREE_OPERAND (tree01
, 0);
9672 tree011
= TREE_OPERAND (tree01
, 1);
9673 STRIP_NOPS (tree010
);
9674 STRIP_NOPS (tree011
);
9675 if (TREE_CODE (tree010
) == INTEGER_CST
9676 && 0 == compare_tree_int (tree010
,
9678 (TREE_TYPE (TREE_OPERAND
9680 && operand_equal_p (tree11
, tree011
, 0))
9681 return fold_convert_loc
9683 build2 ((code0
!= LSHIFT_EXPR
9686 TREE_TYPE (TREE_OPERAND (arg0
, 0)),
9687 TREE_OPERAND (arg0
, 0), TREE_OPERAND (arg1
, 1)));
9693 /* In most languages, can't associate operations on floats through
9694 parentheses. Rather than remember where the parentheses were, we
9695 don't associate floats at all, unless the user has specified
9697 And, we need to make sure type is not saturating. */
9699 if ((! FLOAT_TYPE_P (type
) || flag_associative_math
)
9700 && !TYPE_SATURATING (type
))
9702 tree var0
, con0
, lit0
, minus_lit0
;
9703 tree var1
, con1
, lit1
, minus_lit1
;
9707 /* Split both trees into variables, constants, and literals. Then
9708 associate each group together, the constants with literals,
9709 then the result with variables. This increases the chances of
9710 literals being recombined later and of generating relocatable
9711 expressions for the sum of a constant and literal. */
9712 var0
= split_tree (loc
, arg0
, type
, code
,
9713 &con0
, &lit0
, &minus_lit0
, 0);
9714 var1
= split_tree (loc
, arg1
, type
, code
,
9715 &con1
, &lit1
, &minus_lit1
, code
== MINUS_EXPR
);
9717 /* Recombine MINUS_EXPR operands by using PLUS_EXPR. */
9718 if (code
== MINUS_EXPR
)
9721 /* With undefined overflow prefer doing association in a type
9722 which wraps on overflow, if that is one of the operand types. */
9723 if ((POINTER_TYPE_P (type
) && POINTER_TYPE_OVERFLOW_UNDEFINED
)
9724 || (INTEGRAL_TYPE_P (type
) && !TYPE_OVERFLOW_WRAPS (type
)))
9726 if (INTEGRAL_TYPE_P (TREE_TYPE (arg0
))
9727 && TYPE_OVERFLOW_WRAPS (TREE_TYPE (arg0
)))
9728 atype
= TREE_TYPE (arg0
);
9729 else if (INTEGRAL_TYPE_P (TREE_TYPE (arg1
))
9730 && TYPE_OVERFLOW_WRAPS (TREE_TYPE (arg1
)))
9731 atype
= TREE_TYPE (arg1
);
9732 gcc_assert (TYPE_PRECISION (atype
) == TYPE_PRECISION (type
));
9735 /* With undefined overflow we can only associate constants with one
9736 variable, and constants whose association doesn't overflow. */
9737 if ((POINTER_TYPE_P (atype
) && POINTER_TYPE_OVERFLOW_UNDEFINED
)
9738 || (INTEGRAL_TYPE_P (atype
) && !TYPE_OVERFLOW_WRAPS (atype
)))
9744 bool one_neg
= false;
9746 if (TREE_CODE (tmp0
) == NEGATE_EXPR
)
9748 tmp0
= TREE_OPERAND (tmp0
, 0);
9751 if (CONVERT_EXPR_P (tmp0
)
9752 && INTEGRAL_TYPE_P (TREE_TYPE (TREE_OPERAND (tmp0
, 0)))
9753 && (TYPE_PRECISION (TREE_TYPE (TREE_OPERAND (tmp0
, 0)))
9754 <= TYPE_PRECISION (atype
)))
9755 tmp0
= TREE_OPERAND (tmp0
, 0);
9756 if (TREE_CODE (tmp1
) == NEGATE_EXPR
)
9758 tmp1
= TREE_OPERAND (tmp1
, 0);
9761 if (CONVERT_EXPR_P (tmp1
)
9762 && INTEGRAL_TYPE_P (TREE_TYPE (TREE_OPERAND (tmp1
, 0)))
9763 && (TYPE_PRECISION (TREE_TYPE (TREE_OPERAND (tmp1
, 0)))
9764 <= TYPE_PRECISION (atype
)))
9765 tmp1
= TREE_OPERAND (tmp1
, 0);
9766 /* The only case we can still associate with two variables
9767 is if they cancel out. */
9769 || !operand_equal_p (tmp0
, tmp1
, 0))
9774 /* Only do something if we found more than two objects. Otherwise,
9775 nothing has changed and we risk infinite recursion. */
9777 && (2 < ((var0
!= 0) + (var1
!= 0)
9778 + (con0
!= 0) + (con1
!= 0)
9779 + (lit0
!= 0) + (lit1
!= 0)
9780 + (minus_lit0
!= 0) + (minus_lit1
!= 0))))
9782 bool any_overflows
= false;
9783 if (lit0
) any_overflows
|= TREE_OVERFLOW (lit0
);
9784 if (lit1
) any_overflows
|= TREE_OVERFLOW (lit1
);
9785 if (minus_lit0
) any_overflows
|= TREE_OVERFLOW (minus_lit0
);
9786 if (minus_lit1
) any_overflows
|= TREE_OVERFLOW (minus_lit1
);
9787 var0
= associate_trees (loc
, var0
, var1
, code
, atype
);
9788 con0
= associate_trees (loc
, con0
, con1
, code
, atype
);
9789 lit0
= associate_trees (loc
, lit0
, lit1
, code
, atype
);
9790 minus_lit0
= associate_trees (loc
, minus_lit0
, minus_lit1
,
9793 /* Preserve the MINUS_EXPR if the negative part of the literal is
9794 greater than the positive part. Otherwise, the multiplicative
9795 folding code (i.e extract_muldiv) may be fooled in case
9796 unsigned constants are subtracted, like in the following
9797 example: ((X*2 + 4) - 8U)/2. */
9798 if (minus_lit0
&& lit0
)
9800 if (TREE_CODE (lit0
) == INTEGER_CST
9801 && TREE_CODE (minus_lit0
) == INTEGER_CST
9802 && tree_int_cst_lt (lit0
, minus_lit0
))
9804 minus_lit0
= associate_trees (loc
, minus_lit0
, lit0
,
9810 lit0
= associate_trees (loc
, lit0
, minus_lit0
,
9816 /* Don't introduce overflows through reassociation. */
9818 && ((lit0
&& TREE_OVERFLOW_P (lit0
))
9819 || (minus_lit0
&& TREE_OVERFLOW_P (minus_lit0
))))
9826 fold_convert_loc (loc
, type
,
9827 associate_trees (loc
, var0
, minus_lit0
,
9828 MINUS_EXPR
, atype
));
9831 con0
= associate_trees (loc
, con0
, minus_lit0
,
9834 fold_convert_loc (loc
, type
,
9835 associate_trees (loc
, var0
, con0
,
9840 con0
= associate_trees (loc
, con0
, lit0
, code
, atype
);
9842 fold_convert_loc (loc
, type
, associate_trees (loc
, var0
, con0
,
9850 /* (-A) - B -> (-B) - A where B is easily negated and we can swap. */
9851 if (TREE_CODE (arg0
) == NEGATE_EXPR
9852 && negate_expr_p (op1
)
9853 && reorder_operands_p (arg0
, arg1
))
9854 return fold_build2_loc (loc
, MINUS_EXPR
, type
,
9856 fold_convert_loc (loc
, type
,
9857 TREE_OPERAND (arg0
, 0)));
9859 /* Fold __complex__ ( x, 0 ) - __complex__ ( 0, y ) to
9860 __complex__ ( x, -y ). This is not the same for SNaNs or if
9861 signed zeros are involved. */
9862 if (!HONOR_SNANS (element_mode (arg0
))
9863 && !HONOR_SIGNED_ZEROS (element_mode (arg0
))
9864 && COMPLEX_FLOAT_TYPE_P (TREE_TYPE (arg0
)))
9866 tree rtype
= TREE_TYPE (TREE_TYPE (arg0
));
9867 tree arg0r
= fold_unary_loc (loc
, REALPART_EXPR
, rtype
, arg0
);
9868 tree arg0i
= fold_unary_loc (loc
, IMAGPART_EXPR
, rtype
, arg0
);
9869 bool arg0rz
= false, arg0iz
= false;
9870 if ((arg0r
&& (arg0rz
= real_zerop (arg0r
)))
9871 || (arg0i
&& (arg0iz
= real_zerop (arg0i
))))
9873 tree arg1r
= fold_unary_loc (loc
, REALPART_EXPR
, rtype
, arg1
);
9874 tree arg1i
= fold_unary_loc (loc
, IMAGPART_EXPR
, rtype
, arg1
);
9875 if (arg0rz
&& arg1i
&& real_zerop (arg1i
))
9877 tree rp
= fold_build1_loc (loc
, NEGATE_EXPR
, rtype
,
9879 : build1 (REALPART_EXPR
, rtype
, arg1
));
9880 tree ip
= arg0i
? arg0i
9881 : build1 (IMAGPART_EXPR
, rtype
, arg0
);
9882 return fold_build2_loc (loc
, COMPLEX_EXPR
, type
, rp
, ip
);
9884 else if (arg0iz
&& arg1r
&& real_zerop (arg1r
))
9886 tree rp
= arg0r
? arg0r
9887 : build1 (REALPART_EXPR
, rtype
, arg0
);
9888 tree ip
= fold_build1_loc (loc
, NEGATE_EXPR
, rtype
,
9890 : build1 (IMAGPART_EXPR
, rtype
, arg1
));
9891 return fold_build2_loc (loc
, COMPLEX_EXPR
, type
, rp
, ip
);
9896 /* A - B -> A + (-B) if B is easily negatable. */
9897 if (negate_expr_p (op1
)
9898 && ! TYPE_OVERFLOW_SANITIZED (type
)
9899 && ((FLOAT_TYPE_P (type
)
9900 /* Avoid this transformation if B is a positive REAL_CST. */
9901 && (TREE_CODE (op1
) != REAL_CST
9902 || REAL_VALUE_NEGATIVE (TREE_REAL_CST (op1
))))
9903 || INTEGRAL_TYPE_P (type
)))
9904 return fold_build2_loc (loc
, PLUS_EXPR
, type
,
9905 fold_convert_loc (loc
, type
, arg0
),
9908 /* Fold &a[i] - &a[j] to i-j. */
9909 if (TREE_CODE (arg0
) == ADDR_EXPR
9910 && TREE_CODE (TREE_OPERAND (arg0
, 0)) == ARRAY_REF
9911 && TREE_CODE (arg1
) == ADDR_EXPR
9912 && TREE_CODE (TREE_OPERAND (arg1
, 0)) == ARRAY_REF
)
9914 tree tem
= fold_addr_of_array_ref_difference (loc
, type
,
9915 TREE_OPERAND (arg0
, 0),
9916 TREE_OPERAND (arg1
, 0));
9921 if (FLOAT_TYPE_P (type
)
9922 && flag_unsafe_math_optimizations
9923 && (TREE_CODE (arg0
) == RDIV_EXPR
|| TREE_CODE (arg0
) == MULT_EXPR
)
9924 && (TREE_CODE (arg1
) == RDIV_EXPR
|| TREE_CODE (arg1
) == MULT_EXPR
)
9925 && (tem
= distribute_real_division (loc
, code
, type
, arg0
, arg1
)))
9928 /* Handle (A1 * C1) - (A2 * C2) with A1, A2 or C1, C2 being the same or
9929 one. Make sure the type is not saturating and has the signedness of
9930 the stripped operands, as fold_plusminus_mult_expr will re-associate.
9931 ??? The latter condition should use TYPE_OVERFLOW_* flags instead. */
9932 if ((TREE_CODE (arg0
) == MULT_EXPR
9933 || TREE_CODE (arg1
) == MULT_EXPR
)
9934 && !TYPE_SATURATING (type
)
9935 && TYPE_UNSIGNED (type
) == TYPE_UNSIGNED (TREE_TYPE (arg0
))
9936 && TYPE_UNSIGNED (type
) == TYPE_UNSIGNED (TREE_TYPE (arg1
))
9937 && (!FLOAT_TYPE_P (type
) || flag_associative_math
))
9939 tree tem
= fold_plusminus_mult_expr (loc
, code
, type
, arg0
, arg1
);
9947 if (! FLOAT_TYPE_P (type
))
9949 /* Transform x * -C into -x * C if x is easily negatable. */
9950 if (TREE_CODE (op1
) == INTEGER_CST
9951 && tree_int_cst_sgn (op1
) == -1
9952 && negate_expr_p (op0
)
9953 && (tem
= negate_expr (op1
)) != op1
9954 && ! TREE_OVERFLOW (tem
))
9955 return fold_build2_loc (loc
, MULT_EXPR
, type
,
9956 fold_convert_loc (loc
, type
,
9957 negate_expr (op0
)), tem
);
9959 /* (A + A) * C -> A * 2 * C */
9960 if (TREE_CODE (arg0
) == PLUS_EXPR
9961 && TREE_CODE (arg1
) == INTEGER_CST
9962 && operand_equal_p (TREE_OPERAND (arg0
, 0),
9963 TREE_OPERAND (arg0
, 1), 0))
9964 return fold_build2_loc (loc
, MULT_EXPR
, type
,
9965 omit_one_operand_loc (loc
, type
,
9966 TREE_OPERAND (arg0
, 0),
9967 TREE_OPERAND (arg0
, 1)),
9968 fold_build2_loc (loc
, MULT_EXPR
, type
,
9969 build_int_cst (type
, 2) , arg1
));
9971 /* ((T) (X /[ex] C)) * C cancels out if the conversion is
9972 sign-changing only. */
9973 if (TREE_CODE (arg1
) == INTEGER_CST
9974 && TREE_CODE (arg0
) == EXACT_DIV_EXPR
9975 && operand_equal_p (arg1
, TREE_OPERAND (arg0
, 1), 0))
9976 return fold_convert_loc (loc
, type
, TREE_OPERAND (arg0
, 0));
9978 strict_overflow_p
= false;
9979 if (TREE_CODE (arg1
) == INTEGER_CST
9980 && 0 != (tem
= extract_muldiv (op0
, arg1
, code
, NULL_TREE
,
9981 &strict_overflow_p
)))
9983 if (strict_overflow_p
)
9984 fold_overflow_warning (("assuming signed overflow does not "
9985 "occur when simplifying "
9987 WARN_STRICT_OVERFLOW_MISC
);
9988 return fold_convert_loc (loc
, type
, tem
);
9991 /* Optimize z * conj(z) for integer complex numbers. */
9992 if (TREE_CODE (arg0
) == CONJ_EXPR
9993 && operand_equal_p (TREE_OPERAND (arg0
, 0), arg1
, 0))
9994 return fold_mult_zconjz (loc
, type
, arg1
);
9995 if (TREE_CODE (arg1
) == CONJ_EXPR
9996 && operand_equal_p (arg0
, TREE_OPERAND (arg1
, 0), 0))
9997 return fold_mult_zconjz (loc
, type
, arg0
);
10001 /* Fold z * +-I to __complex__ (-+__imag z, +-__real z).
10002 This is not the same for NaNs or if signed zeros are
10004 if (!HONOR_NANS (arg0
)
10005 && !HONOR_SIGNED_ZEROS (element_mode (arg0
))
10006 && COMPLEX_FLOAT_TYPE_P (TREE_TYPE (arg0
))
10007 && TREE_CODE (arg1
) == COMPLEX_CST
10008 && real_zerop (TREE_REALPART (arg1
)))
10010 tree rtype
= TREE_TYPE (TREE_TYPE (arg0
));
10011 if (real_onep (TREE_IMAGPART (arg1
)))
10013 fold_build2_loc (loc
, COMPLEX_EXPR
, type
,
10014 negate_expr (fold_build1_loc (loc
, IMAGPART_EXPR
,
10016 fold_build1_loc (loc
, REALPART_EXPR
, rtype
, arg0
));
10017 else if (real_minus_onep (TREE_IMAGPART (arg1
)))
10019 fold_build2_loc (loc
, COMPLEX_EXPR
, type
,
10020 fold_build1_loc (loc
, IMAGPART_EXPR
, rtype
, arg0
),
10021 negate_expr (fold_build1_loc (loc
, REALPART_EXPR
,
10025 /* Optimize z * conj(z) for floating point complex numbers.
10026 Guarded by flag_unsafe_math_optimizations as non-finite
10027 imaginary components don't produce scalar results. */
10028 if (flag_unsafe_math_optimizations
10029 && TREE_CODE (arg0
) == CONJ_EXPR
10030 && operand_equal_p (TREE_OPERAND (arg0
, 0), arg1
, 0))
10031 return fold_mult_zconjz (loc
, type
, arg1
);
10032 if (flag_unsafe_math_optimizations
10033 && TREE_CODE (arg1
) == CONJ_EXPR
10034 && operand_equal_p (arg0
, TREE_OPERAND (arg1
, 0), 0))
10035 return fold_mult_zconjz (loc
, type
, arg0
);
10037 if (flag_unsafe_math_optimizations
)
10040 /* Canonicalize x*x as pow(x,2.0), which is expanded as x*x. */
10041 if (!in_gimple_form
10043 && operand_equal_p (arg0
, arg1
, 0))
10045 tree powfn
= mathfn_built_in (type
, BUILT_IN_POW
);
10049 tree arg
= build_real (type
, dconst2
);
10050 return build_call_expr_loc (loc
, powfn
, 2, arg0
, arg
);
10058 /* Canonicalize (X & C1) | C2. */
10059 if (TREE_CODE (arg0
) == BIT_AND_EXPR
10060 && TREE_CODE (arg1
) == INTEGER_CST
10061 && TREE_CODE (TREE_OPERAND (arg0
, 1)) == INTEGER_CST
)
10063 int width
= TYPE_PRECISION (type
), w
;
10064 wide_int c1
= TREE_OPERAND (arg0
, 1);
10065 wide_int c2
= arg1
;
10067 /* If (C1&C2) == C1, then (X&C1)|C2 becomes (X,C2). */
10068 if ((c1
& c2
) == c1
)
10069 return omit_one_operand_loc (loc
, type
, arg1
,
10070 TREE_OPERAND (arg0
, 0));
10072 wide_int msk
= wi::mask (width
, false,
10073 TYPE_PRECISION (TREE_TYPE (arg1
)));
10075 /* If (C1|C2) == ~0 then (X&C1)|C2 becomes X|C2. */
10076 if (msk
.and_not (c1
| c2
) == 0)
10077 return fold_build2_loc (loc
, BIT_IOR_EXPR
, type
,
10078 TREE_OPERAND (arg0
, 0), arg1
);
10080 /* Minimize the number of bits set in C1, i.e. C1 := C1 & ~C2,
10081 unless (C1 & ~C2) | (C2 & C3) for some C3 is a mask of some
10082 mode which allows further optimizations. */
10085 wide_int c3
= c1
.and_not (c2
);
10086 for (w
= BITS_PER_UNIT
; w
<= width
; w
<<= 1)
10088 wide_int mask
= wi::mask (w
, false,
10089 TYPE_PRECISION (type
));
10090 if (((c1
| c2
) & mask
) == mask
&& c1
.and_not (mask
) == 0)
10098 return fold_build2_loc (loc
, BIT_IOR_EXPR
, type
,
10099 fold_build2_loc (loc
, BIT_AND_EXPR
, type
,
10100 TREE_OPERAND (arg0
, 0),
10101 wide_int_to_tree (type
,
10106 /* See if this can be simplified into a rotate first. If that
10107 is unsuccessful continue in the association code. */
10111 /* Fold (X & 1) ^ 1 as (X & 1) == 0. */
10112 if (TREE_CODE (arg0
) == BIT_AND_EXPR
10113 && INTEGRAL_TYPE_P (type
)
10114 && integer_onep (TREE_OPERAND (arg0
, 1))
10115 && integer_onep (arg1
))
10116 return fold_build2_loc (loc
, EQ_EXPR
, type
, arg0
,
10117 build_zero_cst (TREE_TYPE (arg0
)));
10119 /* See if this can be simplified into a rotate first. If that
10120 is unsuccessful continue in the association code. */
10124 /* Fold (X ^ 1) & 1 as (X & 1) == 0. */
10125 if (TREE_CODE (arg0
) == BIT_XOR_EXPR
10126 && INTEGRAL_TYPE_P (type
)
10127 && integer_onep (TREE_OPERAND (arg0
, 1))
10128 && integer_onep (arg1
))
10131 tem
= TREE_OPERAND (arg0
, 0);
10132 tem2
= fold_convert_loc (loc
, TREE_TYPE (tem
), arg1
);
10133 tem2
= fold_build2_loc (loc
, BIT_AND_EXPR
, TREE_TYPE (tem
),
10135 return fold_build2_loc (loc
, EQ_EXPR
, type
, tem2
,
10136 build_zero_cst (TREE_TYPE (tem
)));
10138 /* Fold ~X & 1 as (X & 1) == 0. */
10139 if (TREE_CODE (arg0
) == BIT_NOT_EXPR
10140 && INTEGRAL_TYPE_P (type
)
10141 && integer_onep (arg1
))
10144 tem
= TREE_OPERAND (arg0
, 0);
10145 tem2
= fold_convert_loc (loc
, TREE_TYPE (tem
), arg1
);
10146 tem2
= fold_build2_loc (loc
, BIT_AND_EXPR
, TREE_TYPE (tem
),
10148 return fold_build2_loc (loc
, EQ_EXPR
, type
, tem2
,
10149 build_zero_cst (TREE_TYPE (tem
)));
10151 /* Fold !X & 1 as X == 0. */
10152 if (TREE_CODE (arg0
) == TRUTH_NOT_EXPR
10153 && integer_onep (arg1
))
10155 tem
= TREE_OPERAND (arg0
, 0);
10156 return fold_build2_loc (loc
, EQ_EXPR
, type
, tem
,
10157 build_zero_cst (TREE_TYPE (tem
)));
10160 /* Fold (X ^ Y) & Y as ~X & Y. */
10161 if (TREE_CODE (arg0
) == BIT_XOR_EXPR
10162 && operand_equal_p (TREE_OPERAND (arg0
, 1), arg1
, 0))
10164 tem
= fold_convert_loc (loc
, type
, TREE_OPERAND (arg0
, 0));
10165 return fold_build2_loc (loc
, BIT_AND_EXPR
, type
,
10166 fold_build1_loc (loc
, BIT_NOT_EXPR
, type
, tem
),
10167 fold_convert_loc (loc
, type
, arg1
));
10169 /* Fold (X ^ Y) & X as ~Y & X. */
10170 if (TREE_CODE (arg0
) == BIT_XOR_EXPR
10171 && operand_equal_p (TREE_OPERAND (arg0
, 0), arg1
, 0)
10172 && reorder_operands_p (TREE_OPERAND (arg0
, 1), arg1
))
10174 tem
= fold_convert_loc (loc
, type
, TREE_OPERAND (arg0
, 1));
10175 return fold_build2_loc (loc
, BIT_AND_EXPR
, type
,
10176 fold_build1_loc (loc
, BIT_NOT_EXPR
, type
, tem
),
10177 fold_convert_loc (loc
, type
, arg1
));
10179 /* Fold X & (X ^ Y) as X & ~Y. */
10180 if (TREE_CODE (arg1
) == BIT_XOR_EXPR
10181 && operand_equal_p (arg0
, TREE_OPERAND (arg1
, 0), 0))
10183 tem
= fold_convert_loc (loc
, type
, TREE_OPERAND (arg1
, 1));
10184 return fold_build2_loc (loc
, BIT_AND_EXPR
, type
,
10185 fold_convert_loc (loc
, type
, arg0
),
10186 fold_build1_loc (loc
, BIT_NOT_EXPR
, type
, tem
));
10188 /* Fold X & (Y ^ X) as ~Y & X. */
10189 if (TREE_CODE (arg1
) == BIT_XOR_EXPR
10190 && operand_equal_p (arg0
, TREE_OPERAND (arg1
, 1), 0)
10191 && reorder_operands_p (arg0
, TREE_OPERAND (arg1
, 0)))
10193 tem
= fold_convert_loc (loc
, type
, TREE_OPERAND (arg1
, 0));
10194 return fold_build2_loc (loc
, BIT_AND_EXPR
, type
,
10195 fold_build1_loc (loc
, BIT_NOT_EXPR
, type
, tem
),
10196 fold_convert_loc (loc
, type
, arg0
));
10199 /* Fold (X * Y) & -(1 << CST) to X * Y if Y is a constant
10200 multiple of 1 << CST. */
10201 if (TREE_CODE (arg1
) == INTEGER_CST
)
10203 wide_int cst1
= arg1
;
10204 wide_int ncst1
= -cst1
;
10205 if ((cst1
& ncst1
) == ncst1
10206 && multiple_of_p (type
, arg0
,
10207 wide_int_to_tree (TREE_TYPE (arg1
), ncst1
)))
10208 return fold_convert_loc (loc
, type
, arg0
);
10211 /* Fold (X * CST1) & CST2 to zero if we can, or drop known zero
10213 if (TREE_CODE (arg1
) == INTEGER_CST
10214 && TREE_CODE (arg0
) == MULT_EXPR
10215 && TREE_CODE (TREE_OPERAND (arg0
, 1)) == INTEGER_CST
)
10217 wide_int warg1
= arg1
;
10218 wide_int masked
= mask_with_tz (type
, warg1
, TREE_OPERAND (arg0
, 1));
10221 return omit_two_operands_loc (loc
, type
, build_zero_cst (type
),
10223 else if (masked
!= warg1
)
10225 /* Avoid the transform if arg1 is a mask of some
10226 mode which allows further optimizations. */
10227 int pop
= wi::popcount (warg1
);
10228 if (!(pop
>= BITS_PER_UNIT
10229 && exact_log2 (pop
) != -1
10230 && wi::mask (pop
, false, warg1
.get_precision ()) == warg1
))
10231 return fold_build2_loc (loc
, code
, type
, op0
,
10232 wide_int_to_tree (type
, masked
));
10236 /* For constants M and N, if M == (1LL << cst) - 1 && (N & M) == M,
10237 ((A & N) + B) & M -> (A + B) & M
10238 Similarly if (N & M) == 0,
10239 ((A | N) + B) & M -> (A + B) & M
10240 and for - instead of + (or unary - instead of +)
10241 and/or ^ instead of |.
10242 If B is constant and (B & M) == 0, fold into A & M. */
10243 if (TREE_CODE (arg1
) == INTEGER_CST
)
10245 wide_int cst1
= arg1
;
10246 if ((~cst1
!= 0) && (cst1
& (cst1
+ 1)) == 0
10247 && INTEGRAL_TYPE_P (TREE_TYPE (arg0
))
10248 && (TREE_CODE (arg0
) == PLUS_EXPR
10249 || TREE_CODE (arg0
) == MINUS_EXPR
10250 || TREE_CODE (arg0
) == NEGATE_EXPR
)
10251 && (TYPE_OVERFLOW_WRAPS (TREE_TYPE (arg0
))
10252 || TREE_CODE (TREE_TYPE (arg0
)) == INTEGER_TYPE
))
10258 /* Now we know that arg0 is (C + D) or (C - D) or
10259 -C and arg1 (M) is == (1LL << cst) - 1.
10260 Store C into PMOP[0] and D into PMOP[1]. */
10261 pmop
[0] = TREE_OPERAND (arg0
, 0);
10263 if (TREE_CODE (arg0
) != NEGATE_EXPR
)
10265 pmop
[1] = TREE_OPERAND (arg0
, 1);
10269 if ((wi::max_value (TREE_TYPE (arg0
)) & cst1
) != cst1
)
10272 for (; which
>= 0; which
--)
10273 switch (TREE_CODE (pmop
[which
]))
10278 if (TREE_CODE (TREE_OPERAND (pmop
[which
], 1))
10281 cst0
= TREE_OPERAND (pmop
[which
], 1);
10283 if (TREE_CODE (pmop
[which
]) == BIT_AND_EXPR
)
10288 else if (cst0
!= 0)
10290 /* If C or D is of the form (A & N) where
10291 (N & M) == M, or of the form (A | N) or
10292 (A ^ N) where (N & M) == 0, replace it with A. */
10293 pmop
[which
] = TREE_OPERAND (pmop
[which
], 0);
10296 /* If C or D is a N where (N & M) == 0, it can be
10297 omitted (assumed 0). */
10298 if ((TREE_CODE (arg0
) == PLUS_EXPR
10299 || (TREE_CODE (arg0
) == MINUS_EXPR
&& which
== 0))
10300 && (cst1
& pmop
[which
]) == 0)
10301 pmop
[which
] = NULL
;
10307 /* Only build anything new if we optimized one or both arguments
10309 if (pmop
[0] != TREE_OPERAND (arg0
, 0)
10310 || (TREE_CODE (arg0
) != NEGATE_EXPR
10311 && pmop
[1] != TREE_OPERAND (arg0
, 1)))
10313 tree utype
= TREE_TYPE (arg0
);
10314 if (! TYPE_OVERFLOW_WRAPS (TREE_TYPE (arg0
)))
10316 /* Perform the operations in a type that has defined
10317 overflow behavior. */
10318 utype
= unsigned_type_for (TREE_TYPE (arg0
));
10319 if (pmop
[0] != NULL
)
10320 pmop
[0] = fold_convert_loc (loc
, utype
, pmop
[0]);
10321 if (pmop
[1] != NULL
)
10322 pmop
[1] = fold_convert_loc (loc
, utype
, pmop
[1]);
10325 if (TREE_CODE (arg0
) == NEGATE_EXPR
)
10326 tem
= fold_build1_loc (loc
, NEGATE_EXPR
, utype
, pmop
[0]);
10327 else if (TREE_CODE (arg0
) == PLUS_EXPR
)
10329 if (pmop
[0] != NULL
&& pmop
[1] != NULL
)
10330 tem
= fold_build2_loc (loc
, PLUS_EXPR
, utype
,
10332 else if (pmop
[0] != NULL
)
10334 else if (pmop
[1] != NULL
)
10337 return build_int_cst (type
, 0);
10339 else if (pmop
[0] == NULL
)
10340 tem
= fold_build1_loc (loc
, NEGATE_EXPR
, utype
, pmop
[1]);
10342 tem
= fold_build2_loc (loc
, MINUS_EXPR
, utype
,
10344 /* TEM is now the new binary +, - or unary - replacement. */
10345 tem
= fold_build2_loc (loc
, BIT_AND_EXPR
, utype
, tem
,
10346 fold_convert_loc (loc
, utype
, arg1
));
10347 return fold_convert_loc (loc
, type
, tem
);
10352 /* Simplify ((int)c & 0377) into (int)c, if c is unsigned char. */
10353 if (TREE_CODE (arg1
) == INTEGER_CST
&& TREE_CODE (arg0
) == NOP_EXPR
10354 && TYPE_UNSIGNED (TREE_TYPE (TREE_OPERAND (arg0
, 0))))
10356 prec
= element_precision (TREE_TYPE (TREE_OPERAND (arg0
, 0)));
10358 wide_int mask
= wide_int::from (arg1
, prec
, UNSIGNED
);
10361 fold_convert_loc (loc
, type
, TREE_OPERAND (arg0
, 0));
10367 /* Don't touch a floating-point divide by zero unless the mode
10368 of the constant can represent infinity. */
10369 if (TREE_CODE (arg1
) == REAL_CST
10370 && !MODE_HAS_INFINITIES (TYPE_MODE (TREE_TYPE (arg1
)))
10371 && real_zerop (arg1
))
10374 /* (-A) / (-B) -> A / B */
10375 if (TREE_CODE (arg0
) == NEGATE_EXPR
&& negate_expr_p (arg1
))
10376 return fold_build2_loc (loc
, RDIV_EXPR
, type
,
10377 TREE_OPERAND (arg0
, 0),
10378 negate_expr (arg1
));
10379 if (TREE_CODE (arg1
) == NEGATE_EXPR
&& negate_expr_p (arg0
))
10380 return fold_build2_loc (loc
, RDIV_EXPR
, type
,
10381 negate_expr (arg0
),
10382 TREE_OPERAND (arg1
, 0));
10385 case TRUNC_DIV_EXPR
:
10388 case FLOOR_DIV_EXPR
:
10389 /* Simplify A / (B << N) where A and B are positive and B is
10390 a power of 2, to A >> (N + log2(B)). */
10391 strict_overflow_p
= false;
10392 if (TREE_CODE (arg1
) == LSHIFT_EXPR
10393 && (TYPE_UNSIGNED (type
)
10394 || tree_expr_nonnegative_warnv_p (op0
, &strict_overflow_p
)))
10396 tree sval
= TREE_OPERAND (arg1
, 0);
10397 if (integer_pow2p (sval
) && tree_int_cst_sgn (sval
) > 0)
10399 tree sh_cnt
= TREE_OPERAND (arg1
, 1);
10400 tree pow2
= build_int_cst (TREE_TYPE (sh_cnt
),
10401 wi::exact_log2 (sval
));
10403 if (strict_overflow_p
)
10404 fold_overflow_warning (("assuming signed overflow does not "
10405 "occur when simplifying A / (B << N)"),
10406 WARN_STRICT_OVERFLOW_MISC
);
10408 sh_cnt
= fold_build2_loc (loc
, PLUS_EXPR
, TREE_TYPE (sh_cnt
),
10410 return fold_build2_loc (loc
, RSHIFT_EXPR
, type
,
10411 fold_convert_loc (loc
, type
, arg0
), sh_cnt
);
10417 case ROUND_DIV_EXPR
:
10418 case CEIL_DIV_EXPR
:
10419 case EXACT_DIV_EXPR
:
10420 if (integer_zerop (arg1
))
10423 /* Convert -A / -B to A / B when the type is signed and overflow is
10425 if ((!INTEGRAL_TYPE_P (type
) || TYPE_OVERFLOW_UNDEFINED (type
))
10426 && TREE_CODE (arg0
) == NEGATE_EXPR
10427 && negate_expr_p (op1
))
10429 if (INTEGRAL_TYPE_P (type
))
10430 fold_overflow_warning (("assuming signed overflow does not occur "
10431 "when distributing negation across "
10433 WARN_STRICT_OVERFLOW_MISC
);
10434 return fold_build2_loc (loc
, code
, type
,
10435 fold_convert_loc (loc
, type
,
10436 TREE_OPERAND (arg0
, 0)),
10437 negate_expr (op1
));
10439 if ((!INTEGRAL_TYPE_P (type
) || TYPE_OVERFLOW_UNDEFINED (type
))
10440 && TREE_CODE (arg1
) == NEGATE_EXPR
10441 && negate_expr_p (op0
))
10443 if (INTEGRAL_TYPE_P (type
))
10444 fold_overflow_warning (("assuming signed overflow does not occur "
10445 "when distributing negation across "
10447 WARN_STRICT_OVERFLOW_MISC
);
10448 return fold_build2_loc (loc
, code
, type
,
10450 fold_convert_loc (loc
, type
,
10451 TREE_OPERAND (arg1
, 0)));
10454 /* If arg0 is a multiple of arg1, then rewrite to the fastest div
10455 operation, EXACT_DIV_EXPR.
10457 Note that only CEIL_DIV_EXPR and FLOOR_DIV_EXPR are rewritten now.
10458 At one time others generated faster code, it's not clear if they do
10459 after the last round to changes to the DIV code in expmed.c. */
10460 if ((code
== CEIL_DIV_EXPR
|| code
== FLOOR_DIV_EXPR
)
10461 && multiple_of_p (type
, arg0
, arg1
))
10462 return fold_build2_loc (loc
, EXACT_DIV_EXPR
, type
,
10463 fold_convert (type
, arg0
),
10464 fold_convert (type
, arg1
));
10466 strict_overflow_p
= false;
10467 if (TREE_CODE (arg1
) == INTEGER_CST
10468 && 0 != (tem
= extract_muldiv (op0
, arg1
, code
, NULL_TREE
,
10469 &strict_overflow_p
)))
10471 if (strict_overflow_p
)
10472 fold_overflow_warning (("assuming signed overflow does not occur "
10473 "when simplifying division"),
10474 WARN_STRICT_OVERFLOW_MISC
);
10475 return fold_convert_loc (loc
, type
, tem
);
10480 case CEIL_MOD_EXPR
:
10481 case FLOOR_MOD_EXPR
:
10482 case ROUND_MOD_EXPR
:
10483 case TRUNC_MOD_EXPR
:
10484 strict_overflow_p
= false;
10485 if (TREE_CODE (arg1
) == INTEGER_CST
10486 && 0 != (tem
= extract_muldiv (op0
, arg1
, code
, NULL_TREE
,
10487 &strict_overflow_p
)))
10489 if (strict_overflow_p
)
10490 fold_overflow_warning (("assuming signed overflow does not occur "
10491 "when simplifying modulus"),
10492 WARN_STRICT_OVERFLOW_MISC
);
10493 return fold_convert_loc (loc
, type
, tem
);
10502 /* Since negative shift count is not well-defined,
10503 don't try to compute it in the compiler. */
10504 if (TREE_CODE (arg1
) == INTEGER_CST
&& tree_int_cst_sgn (arg1
) < 0)
10507 prec
= element_precision (type
);
10509 /* If we have a rotate of a bit operation with the rotate count and
10510 the second operand of the bit operation both constant,
10511 permute the two operations. */
10512 if (code
== RROTATE_EXPR
&& TREE_CODE (arg1
) == INTEGER_CST
10513 && (TREE_CODE (arg0
) == BIT_AND_EXPR
10514 || TREE_CODE (arg0
) == BIT_IOR_EXPR
10515 || TREE_CODE (arg0
) == BIT_XOR_EXPR
)
10516 && TREE_CODE (TREE_OPERAND (arg0
, 1)) == INTEGER_CST
)
10517 return fold_build2_loc (loc
, TREE_CODE (arg0
), type
,
10518 fold_build2_loc (loc
, code
, type
,
10519 TREE_OPERAND (arg0
, 0), arg1
),
10520 fold_build2_loc (loc
, code
, type
,
10521 TREE_OPERAND (arg0
, 1), arg1
));
10523 /* Two consecutive rotates adding up to the some integer
10524 multiple of the precision of the type can be ignored. */
10525 if (code
== RROTATE_EXPR
&& TREE_CODE (arg1
) == INTEGER_CST
10526 && TREE_CODE (arg0
) == RROTATE_EXPR
10527 && TREE_CODE (TREE_OPERAND (arg0
, 1)) == INTEGER_CST
10528 && wi::umod_trunc (wi::add (arg1
, TREE_OPERAND (arg0
, 1)),
10530 return TREE_OPERAND (arg0
, 0);
10538 case TRUTH_ANDIF_EXPR
:
10539 /* Note that the operands of this must be ints
10540 and their values must be 0 or 1.
10541 ("true" is a fixed value perhaps depending on the language.) */
10542 /* If first arg is constant zero, return it. */
10543 if (integer_zerop (arg0
))
10544 return fold_convert_loc (loc
, type
, arg0
);
10545 case TRUTH_AND_EXPR
:
10546 /* If either arg is constant true, drop it. */
10547 if (TREE_CODE (arg0
) == INTEGER_CST
&& ! integer_zerop (arg0
))
10548 return non_lvalue_loc (loc
, fold_convert_loc (loc
, type
, arg1
));
10549 if (TREE_CODE (arg1
) == INTEGER_CST
&& ! integer_zerop (arg1
)
10550 /* Preserve sequence points. */
10551 && (code
!= TRUTH_ANDIF_EXPR
|| ! TREE_SIDE_EFFECTS (arg0
)))
10552 return non_lvalue_loc (loc
, fold_convert_loc (loc
, type
, arg0
));
10553 /* If second arg is constant zero, result is zero, but first arg
10554 must be evaluated. */
10555 if (integer_zerop (arg1
))
10556 return omit_one_operand_loc (loc
, type
, arg1
, arg0
);
10557 /* Likewise for first arg, but note that only the TRUTH_AND_EXPR
10558 case will be handled here. */
10559 if (integer_zerop (arg0
))
10560 return omit_one_operand_loc (loc
, type
, arg0
, arg1
);
10562 /* !X && X is always false. */
10563 if (TREE_CODE (arg0
) == TRUTH_NOT_EXPR
10564 && operand_equal_p (TREE_OPERAND (arg0
, 0), arg1
, 0))
10565 return omit_one_operand_loc (loc
, type
, integer_zero_node
, arg1
);
10566 /* X && !X is always false. */
10567 if (TREE_CODE (arg1
) == TRUTH_NOT_EXPR
10568 && operand_equal_p (arg0
, TREE_OPERAND (arg1
, 0), 0))
10569 return omit_one_operand_loc (loc
, type
, integer_zero_node
, arg0
);
10571 /* A < X && A + 1 > Y ==> A < X && A >= Y. Normally A + 1 > Y
10572 means A >= Y && A != MAX, but in this case we know that
10575 if (!TREE_SIDE_EFFECTS (arg0
)
10576 && !TREE_SIDE_EFFECTS (arg1
))
10578 tem
= fold_to_nonsharp_ineq_using_bound (loc
, arg0
, arg1
);
10579 if (tem
&& !operand_equal_p (tem
, arg0
, 0))
10580 return fold_build2_loc (loc
, code
, type
, tem
, arg1
);
10582 tem
= fold_to_nonsharp_ineq_using_bound (loc
, arg1
, arg0
);
10583 if (tem
&& !operand_equal_p (tem
, arg1
, 0))
10584 return fold_build2_loc (loc
, code
, type
, arg0
, tem
);
10587 if ((tem
= fold_truth_andor (loc
, code
, type
, arg0
, arg1
, op0
, op1
))
10593 case TRUTH_ORIF_EXPR
:
10594 /* Note that the operands of this must be ints
10595 and their values must be 0 or true.
10596 ("true" is a fixed value perhaps depending on the language.) */
10597 /* If first arg is constant true, return it. */
10598 if (TREE_CODE (arg0
) == INTEGER_CST
&& ! integer_zerop (arg0
))
10599 return fold_convert_loc (loc
, type
, arg0
);
10600 case TRUTH_OR_EXPR
:
10601 /* If either arg is constant zero, drop it. */
10602 if (TREE_CODE (arg0
) == INTEGER_CST
&& integer_zerop (arg0
))
10603 return non_lvalue_loc (loc
, fold_convert_loc (loc
, type
, arg1
));
10604 if (TREE_CODE (arg1
) == INTEGER_CST
&& integer_zerop (arg1
)
10605 /* Preserve sequence points. */
10606 && (code
!= TRUTH_ORIF_EXPR
|| ! TREE_SIDE_EFFECTS (arg0
)))
10607 return non_lvalue_loc (loc
, fold_convert_loc (loc
, type
, arg0
));
10608 /* If second arg is constant true, result is true, but we must
10609 evaluate first arg. */
10610 if (TREE_CODE (arg1
) == INTEGER_CST
&& ! integer_zerop (arg1
))
10611 return omit_one_operand_loc (loc
, type
, arg1
, arg0
);
10612 /* Likewise for first arg, but note this only occurs here for
10614 if (TREE_CODE (arg0
) == INTEGER_CST
&& ! integer_zerop (arg0
))
10615 return omit_one_operand_loc (loc
, type
, arg0
, arg1
);
10617 /* !X || X is always true. */
10618 if (TREE_CODE (arg0
) == TRUTH_NOT_EXPR
10619 && operand_equal_p (TREE_OPERAND (arg0
, 0), arg1
, 0))
10620 return omit_one_operand_loc (loc
, type
, integer_one_node
, arg1
);
10621 /* X || !X is always true. */
10622 if (TREE_CODE (arg1
) == TRUTH_NOT_EXPR
10623 && operand_equal_p (arg0
, TREE_OPERAND (arg1
, 0), 0))
10624 return omit_one_operand_loc (loc
, type
, integer_one_node
, arg0
);
10626 /* (X && !Y) || (!X && Y) is X ^ Y */
10627 if (TREE_CODE (arg0
) == TRUTH_AND_EXPR
10628 && TREE_CODE (arg1
) == TRUTH_AND_EXPR
)
10630 tree a0
, a1
, l0
, l1
, n0
, n1
;
10632 a0
= fold_convert_loc (loc
, type
, TREE_OPERAND (arg1
, 0));
10633 a1
= fold_convert_loc (loc
, type
, TREE_OPERAND (arg1
, 1));
10635 l0
= fold_convert_loc (loc
, type
, TREE_OPERAND (arg0
, 0));
10636 l1
= fold_convert_loc (loc
, type
, TREE_OPERAND (arg0
, 1));
10638 n0
= fold_build1_loc (loc
, TRUTH_NOT_EXPR
, type
, l0
);
10639 n1
= fold_build1_loc (loc
, TRUTH_NOT_EXPR
, type
, l1
);
10641 if ((operand_equal_p (n0
, a0
, 0)
10642 && operand_equal_p (n1
, a1
, 0))
10643 || (operand_equal_p (n0
, a1
, 0)
10644 && operand_equal_p (n1
, a0
, 0)))
10645 return fold_build2_loc (loc
, TRUTH_XOR_EXPR
, type
, l0
, n1
);
10648 if ((tem
= fold_truth_andor (loc
, code
, type
, arg0
, arg1
, op0
, op1
))
10654 case TRUTH_XOR_EXPR
:
10655 /* If the second arg is constant zero, drop it. */
10656 if (integer_zerop (arg1
))
10657 return non_lvalue_loc (loc
, fold_convert_loc (loc
, type
, arg0
));
10658 /* If the second arg is constant true, this is a logical inversion. */
10659 if (integer_onep (arg1
))
10661 tem
= invert_truthvalue_loc (loc
, arg0
);
10662 return non_lvalue_loc (loc
, fold_convert_loc (loc
, type
, tem
));
10664 /* Identical arguments cancel to zero. */
10665 if (operand_equal_p (arg0
, arg1
, 0))
10666 return omit_one_operand_loc (loc
, type
, integer_zero_node
, arg0
);
10668 /* !X ^ X is always true. */
10669 if (TREE_CODE (arg0
) == TRUTH_NOT_EXPR
10670 && operand_equal_p (TREE_OPERAND (arg0
, 0), arg1
, 0))
10671 return omit_one_operand_loc (loc
, type
, integer_one_node
, arg1
);
10673 /* X ^ !X is always true. */
10674 if (TREE_CODE (arg1
) == TRUTH_NOT_EXPR
10675 && operand_equal_p (arg0
, TREE_OPERAND (arg1
, 0), 0))
10676 return omit_one_operand_loc (loc
, type
, integer_one_node
, arg0
);
10685 tem
= fold_comparison (loc
, code
, type
, op0
, op1
);
10686 if (tem
!= NULL_TREE
)
10689 /* bool_var != 1 becomes !bool_var. */
10690 if (TREE_CODE (TREE_TYPE (arg0
)) == BOOLEAN_TYPE
&& integer_onep (arg1
)
10691 && code
== NE_EXPR
)
10692 return fold_convert_loc (loc
, type
,
10693 fold_build1_loc (loc
, TRUTH_NOT_EXPR
,
10694 TREE_TYPE (arg0
), arg0
));
10696 /* bool_var == 0 becomes !bool_var. */
10697 if (TREE_CODE (TREE_TYPE (arg0
)) == BOOLEAN_TYPE
&& integer_zerop (arg1
)
10698 && code
== EQ_EXPR
)
10699 return fold_convert_loc (loc
, type
,
10700 fold_build1_loc (loc
, TRUTH_NOT_EXPR
,
10701 TREE_TYPE (arg0
), arg0
));
10703 /* !exp != 0 becomes !exp */
10704 if (TREE_CODE (arg0
) == TRUTH_NOT_EXPR
&& integer_zerop (arg1
)
10705 && code
== NE_EXPR
)
10706 return non_lvalue_loc (loc
, fold_convert_loc (loc
, type
, arg0
));
10708 /* Transform comparisons of the form X +- Y CMP X to Y CMP 0. */
10709 if ((TREE_CODE (arg0
) == PLUS_EXPR
10710 || TREE_CODE (arg0
) == POINTER_PLUS_EXPR
10711 || TREE_CODE (arg0
) == MINUS_EXPR
)
10712 && operand_equal_p (tree_strip_nop_conversions (TREE_OPERAND (arg0
,
10715 && (INTEGRAL_TYPE_P (TREE_TYPE (arg0
))
10716 || POINTER_TYPE_P (TREE_TYPE (arg0
))))
10718 tree val
= TREE_OPERAND (arg0
, 1);
10719 val
= fold_build2_loc (loc
, code
, type
, val
,
10720 build_int_cst (TREE_TYPE (val
), 0));
10721 return omit_two_operands_loc (loc
, type
, val
,
10722 TREE_OPERAND (arg0
, 0), arg1
);
10725 /* Transform comparisons of the form X CMP X +- Y to Y CMP 0. */
10726 if ((TREE_CODE (arg1
) == PLUS_EXPR
10727 || TREE_CODE (arg1
) == POINTER_PLUS_EXPR
10728 || TREE_CODE (arg1
) == MINUS_EXPR
)
10729 && operand_equal_p (tree_strip_nop_conversions (TREE_OPERAND (arg1
,
10732 && (INTEGRAL_TYPE_P (TREE_TYPE (arg1
))
10733 || POINTER_TYPE_P (TREE_TYPE (arg1
))))
10735 tree val
= TREE_OPERAND (arg1
, 1);
10736 val
= fold_build2_loc (loc
, code
, type
, val
,
10737 build_int_cst (TREE_TYPE (val
), 0));
10738 return omit_two_operands_loc (loc
, type
, val
,
10739 TREE_OPERAND (arg1
, 0), arg0
);
10742 /* Transform comparisons of the form C - X CMP X if C % 2 == 1. */
10743 if (TREE_CODE (arg0
) == MINUS_EXPR
10744 && TREE_CODE (TREE_OPERAND (arg0
, 0)) == INTEGER_CST
10745 && operand_equal_p (tree_strip_nop_conversions (TREE_OPERAND (arg0
,
10748 && wi::extract_uhwi (TREE_OPERAND (arg0
, 0), 0, 1) == 1)
10749 return omit_two_operands_loc (loc
, type
,
10751 ? boolean_true_node
: boolean_false_node
,
10752 TREE_OPERAND (arg0
, 1), arg1
);
10754 /* Transform comparisons of the form X CMP C - X if C % 2 == 1. */
10755 if (TREE_CODE (arg1
) == MINUS_EXPR
10756 && TREE_CODE (TREE_OPERAND (arg1
, 0)) == INTEGER_CST
10757 && operand_equal_p (tree_strip_nop_conversions (TREE_OPERAND (arg1
,
10760 && wi::extract_uhwi (TREE_OPERAND (arg1
, 0), 0, 1) == 1)
10761 return omit_two_operands_loc (loc
, type
,
10763 ? boolean_true_node
: boolean_false_node
,
10764 TREE_OPERAND (arg1
, 1), arg0
);
10766 /* If this is an EQ or NE comparison with zero and ARG0 is
10767 (1 << foo) & bar, convert it to (bar >> foo) & 1. Both require
10768 two operations, but the latter can be done in one less insn
10769 on machines that have only two-operand insns or on which a
10770 constant cannot be the first operand. */
10771 if (TREE_CODE (arg0
) == BIT_AND_EXPR
10772 && integer_zerop (arg1
))
10774 tree arg00
= TREE_OPERAND (arg0
, 0);
10775 tree arg01
= TREE_OPERAND (arg0
, 1);
10776 if (TREE_CODE (arg00
) == LSHIFT_EXPR
10777 && integer_onep (TREE_OPERAND (arg00
, 0)))
10779 tree tem
= fold_build2_loc (loc
, RSHIFT_EXPR
, TREE_TYPE (arg00
),
10780 arg01
, TREE_OPERAND (arg00
, 1));
10781 tem
= fold_build2_loc (loc
, BIT_AND_EXPR
, TREE_TYPE (arg0
), tem
,
10782 build_int_cst (TREE_TYPE (arg0
), 1));
10783 return fold_build2_loc (loc
, code
, type
,
10784 fold_convert_loc (loc
, TREE_TYPE (arg1
), tem
),
10787 else if (TREE_CODE (arg01
) == LSHIFT_EXPR
10788 && integer_onep (TREE_OPERAND (arg01
, 0)))
10790 tree tem
= fold_build2_loc (loc
, RSHIFT_EXPR
, TREE_TYPE (arg01
),
10791 arg00
, TREE_OPERAND (arg01
, 1));
10792 tem
= fold_build2_loc (loc
, BIT_AND_EXPR
, TREE_TYPE (arg0
), tem
,
10793 build_int_cst (TREE_TYPE (arg0
), 1));
10794 return fold_build2_loc (loc
, code
, type
,
10795 fold_convert_loc (loc
, TREE_TYPE (arg1
), tem
),
10800 /* If this is an NE or EQ comparison of zero against the result of a
10801 signed MOD operation whose second operand is a power of 2, make
10802 the MOD operation unsigned since it is simpler and equivalent. */
10803 if (integer_zerop (arg1
)
10804 && !TYPE_UNSIGNED (TREE_TYPE (arg0
))
10805 && (TREE_CODE (arg0
) == TRUNC_MOD_EXPR
10806 || TREE_CODE (arg0
) == CEIL_MOD_EXPR
10807 || TREE_CODE (arg0
) == FLOOR_MOD_EXPR
10808 || TREE_CODE (arg0
) == ROUND_MOD_EXPR
)
10809 && integer_pow2p (TREE_OPERAND (arg0
, 1)))
10811 tree newtype
= unsigned_type_for (TREE_TYPE (arg0
));
10812 tree newmod
= fold_build2_loc (loc
, TREE_CODE (arg0
), newtype
,
10813 fold_convert_loc (loc
, newtype
,
10814 TREE_OPERAND (arg0
, 0)),
10815 fold_convert_loc (loc
, newtype
,
10816 TREE_OPERAND (arg0
, 1)));
10818 return fold_build2_loc (loc
, code
, type
, newmod
,
10819 fold_convert_loc (loc
, newtype
, arg1
));
10822 /* Fold ((X >> C1) & C2) == 0 and ((X >> C1) & C2) != 0 where
10823 C1 is a valid shift constant, and C2 is a power of two, i.e.
10825 if (TREE_CODE (arg0
) == BIT_AND_EXPR
10826 && TREE_CODE (TREE_OPERAND (arg0
, 0)) == RSHIFT_EXPR
10827 && TREE_CODE (TREE_OPERAND (TREE_OPERAND (arg0
, 0), 1))
10829 && integer_pow2p (TREE_OPERAND (arg0
, 1))
10830 && integer_zerop (arg1
))
10832 tree itype
= TREE_TYPE (arg0
);
10833 tree arg001
= TREE_OPERAND (TREE_OPERAND (arg0
, 0), 1);
10834 prec
= TYPE_PRECISION (itype
);
10836 /* Check for a valid shift count. */
10837 if (wi::ltu_p (arg001
, prec
))
10839 tree arg01
= TREE_OPERAND (arg0
, 1);
10840 tree arg000
= TREE_OPERAND (TREE_OPERAND (arg0
, 0), 0);
10841 unsigned HOST_WIDE_INT log2
= tree_log2 (arg01
);
10842 /* If (C2 << C1) doesn't overflow, then ((X >> C1) & C2) != 0
10843 can be rewritten as (X & (C2 << C1)) != 0. */
10844 if ((log2
+ TREE_INT_CST_LOW (arg001
)) < prec
)
10846 tem
= fold_build2_loc (loc
, LSHIFT_EXPR
, itype
, arg01
, arg001
);
10847 tem
= fold_build2_loc (loc
, BIT_AND_EXPR
, itype
, arg000
, tem
);
10848 return fold_build2_loc (loc
, code
, type
, tem
,
10849 fold_convert_loc (loc
, itype
, arg1
));
10851 /* Otherwise, for signed (arithmetic) shifts,
10852 ((X >> C1) & C2) != 0 is rewritten as X < 0, and
10853 ((X >> C1) & C2) == 0 is rewritten as X >= 0. */
10854 else if (!TYPE_UNSIGNED (itype
))
10855 return fold_build2_loc (loc
, code
== EQ_EXPR
? GE_EXPR
: LT_EXPR
, type
,
10856 arg000
, build_int_cst (itype
, 0));
10857 /* Otherwise, of unsigned (logical) shifts,
10858 ((X >> C1) & C2) != 0 is rewritten as (X,false), and
10859 ((X >> C1) & C2) == 0 is rewritten as (X,true). */
10861 return omit_one_operand_loc (loc
, type
,
10862 code
== EQ_EXPR
? integer_one_node
10863 : integer_zero_node
,
10868 /* If we have (A & C) == D where D & ~C != 0, convert this into 0.
10869 Similarly for NE_EXPR. */
10870 if (TREE_CODE (arg0
) == BIT_AND_EXPR
10871 && TREE_CODE (arg1
) == INTEGER_CST
10872 && TREE_CODE (TREE_OPERAND (arg0
, 1)) == INTEGER_CST
)
10874 tree notc
= fold_build1_loc (loc
, BIT_NOT_EXPR
,
10875 TREE_TYPE (TREE_OPERAND (arg0
, 1)),
10876 TREE_OPERAND (arg0
, 1));
10878 = fold_build2_loc (loc
, BIT_AND_EXPR
, TREE_TYPE (arg0
),
10879 fold_convert_loc (loc
, TREE_TYPE (arg0
), arg1
),
10881 tree rslt
= code
== EQ_EXPR
? integer_zero_node
: integer_one_node
;
10882 if (integer_nonzerop (dandnotc
))
10883 return omit_one_operand_loc (loc
, type
, rslt
, arg0
);
10886 /* If this is a comparison of a field, we may be able to simplify it. */
10887 if ((TREE_CODE (arg0
) == COMPONENT_REF
10888 || TREE_CODE (arg0
) == BIT_FIELD_REF
)
10889 /* Handle the constant case even without -O
10890 to make sure the warnings are given. */
10891 && (optimize
|| TREE_CODE (arg1
) == INTEGER_CST
))
10893 t1
= optimize_bit_field_compare (loc
, code
, type
, arg0
, arg1
);
10898 /* Optimize comparisons of strlen vs zero to a compare of the
10899 first character of the string vs zero. To wit,
10900 strlen(ptr) == 0 => *ptr == 0
10901 strlen(ptr) != 0 => *ptr != 0
10902 Other cases should reduce to one of these two (or a constant)
10903 due to the return value of strlen being unsigned. */
10904 if (TREE_CODE (arg0
) == CALL_EXPR
10905 && integer_zerop (arg1
))
10907 tree fndecl
= get_callee_fndecl (arg0
);
10910 && DECL_BUILT_IN_CLASS (fndecl
) == BUILT_IN_NORMAL
10911 && DECL_FUNCTION_CODE (fndecl
) == BUILT_IN_STRLEN
10912 && call_expr_nargs (arg0
) == 1
10913 && TREE_CODE (TREE_TYPE (CALL_EXPR_ARG (arg0
, 0))) == POINTER_TYPE
)
10915 tree iref
= build_fold_indirect_ref_loc (loc
,
10916 CALL_EXPR_ARG (arg0
, 0));
10917 return fold_build2_loc (loc
, code
, type
, iref
,
10918 build_int_cst (TREE_TYPE (iref
), 0));
10922 /* Fold (X >> C) != 0 into X < 0 if C is one less than the width
10923 of X. Similarly fold (X >> C) == 0 into X >= 0. */
10924 if (TREE_CODE (arg0
) == RSHIFT_EXPR
10925 && integer_zerop (arg1
)
10926 && TREE_CODE (TREE_OPERAND (arg0
, 1)) == INTEGER_CST
)
10928 tree arg00
= TREE_OPERAND (arg0
, 0);
10929 tree arg01
= TREE_OPERAND (arg0
, 1);
10930 tree itype
= TREE_TYPE (arg00
);
10931 if (wi::eq_p (arg01
, element_precision (itype
) - 1))
10933 if (TYPE_UNSIGNED (itype
))
10935 itype
= signed_type_for (itype
);
10936 arg00
= fold_convert_loc (loc
, itype
, arg00
);
10938 return fold_build2_loc (loc
, code
== EQ_EXPR
? GE_EXPR
: LT_EXPR
,
10939 type
, arg00
, build_zero_cst (itype
));
10943 /* Fold (~X & C) == 0 into (X & C) != 0 and (~X & C) != 0 into
10944 (X & C) == 0 when C is a single bit. */
10945 if (TREE_CODE (arg0
) == BIT_AND_EXPR
10946 && TREE_CODE (TREE_OPERAND (arg0
, 0)) == BIT_NOT_EXPR
10947 && integer_zerop (arg1
)
10948 && integer_pow2p (TREE_OPERAND (arg0
, 1)))
10950 tem
= fold_build2_loc (loc
, BIT_AND_EXPR
, TREE_TYPE (arg0
),
10951 TREE_OPERAND (TREE_OPERAND (arg0
, 0), 0),
10952 TREE_OPERAND (arg0
, 1));
10953 return fold_build2_loc (loc
, code
== EQ_EXPR
? NE_EXPR
: EQ_EXPR
,
10955 fold_convert_loc (loc
, TREE_TYPE (arg0
),
10959 /* Fold ((X & C) ^ C) eq/ne 0 into (X & C) ne/eq 0, when the
10960 constant C is a power of two, i.e. a single bit. */
10961 if (TREE_CODE (arg0
) == BIT_XOR_EXPR
10962 && TREE_CODE (TREE_OPERAND (arg0
, 0)) == BIT_AND_EXPR
10963 && integer_zerop (arg1
)
10964 && integer_pow2p (TREE_OPERAND (arg0
, 1))
10965 && operand_equal_p (TREE_OPERAND (TREE_OPERAND (arg0
, 0), 1),
10966 TREE_OPERAND (arg0
, 1), OEP_ONLY_CONST
))
10968 tree arg00
= TREE_OPERAND (arg0
, 0);
10969 return fold_build2_loc (loc
, code
== EQ_EXPR
? NE_EXPR
: EQ_EXPR
, type
,
10970 arg00
, build_int_cst (TREE_TYPE (arg00
), 0));
10973 /* Likewise, fold ((X ^ C) & C) eq/ne 0 into (X & C) ne/eq 0,
10974 when is C is a power of two, i.e. a single bit. */
10975 if (TREE_CODE (arg0
) == BIT_AND_EXPR
10976 && TREE_CODE (TREE_OPERAND (arg0
, 0)) == BIT_XOR_EXPR
10977 && integer_zerop (arg1
)
10978 && integer_pow2p (TREE_OPERAND (arg0
, 1))
10979 && operand_equal_p (TREE_OPERAND (TREE_OPERAND (arg0
, 0), 1),
10980 TREE_OPERAND (arg0
, 1), OEP_ONLY_CONST
))
10982 tree arg000
= TREE_OPERAND (TREE_OPERAND (arg0
, 0), 0);
10983 tem
= fold_build2_loc (loc
, BIT_AND_EXPR
, TREE_TYPE (arg000
),
10984 arg000
, TREE_OPERAND (arg0
, 1));
10985 return fold_build2_loc (loc
, code
== EQ_EXPR
? NE_EXPR
: EQ_EXPR
, type
,
10986 tem
, build_int_cst (TREE_TYPE (tem
), 0));
10989 if (integer_zerop (arg1
)
10990 && tree_expr_nonzero_p (arg0
))
10992 tree res
= constant_boolean_node (code
==NE_EXPR
, type
);
10993 return omit_one_operand_loc (loc
, type
, res
, arg0
);
10996 /* Fold (X & C) op (Y & C) as (X ^ Y) & C op 0", and symmetries. */
10997 if (TREE_CODE (arg0
) == BIT_AND_EXPR
10998 && TREE_CODE (arg1
) == BIT_AND_EXPR
)
11000 tree arg00
= TREE_OPERAND (arg0
, 0);
11001 tree arg01
= TREE_OPERAND (arg0
, 1);
11002 tree arg10
= TREE_OPERAND (arg1
, 0);
11003 tree arg11
= TREE_OPERAND (arg1
, 1);
11004 tree itype
= TREE_TYPE (arg0
);
11006 if (operand_equal_p (arg01
, arg11
, 0))
11007 return fold_build2_loc (loc
, code
, type
,
11008 fold_build2_loc (loc
, BIT_AND_EXPR
, itype
,
11009 fold_build2_loc (loc
,
11010 BIT_XOR_EXPR
, itype
,
11013 build_zero_cst (itype
));
11015 if (operand_equal_p (arg01
, arg10
, 0))
11016 return fold_build2_loc (loc
, code
, type
,
11017 fold_build2_loc (loc
, BIT_AND_EXPR
, itype
,
11018 fold_build2_loc (loc
,
11019 BIT_XOR_EXPR
, itype
,
11022 build_zero_cst (itype
));
11024 if (operand_equal_p (arg00
, arg11
, 0))
11025 return fold_build2_loc (loc
, code
, type
,
11026 fold_build2_loc (loc
, BIT_AND_EXPR
, itype
,
11027 fold_build2_loc (loc
,
11028 BIT_XOR_EXPR
, itype
,
11031 build_zero_cst (itype
));
11033 if (operand_equal_p (arg00
, arg10
, 0))
11034 return fold_build2_loc (loc
, code
, type
,
11035 fold_build2_loc (loc
, BIT_AND_EXPR
, itype
,
11036 fold_build2_loc (loc
,
11037 BIT_XOR_EXPR
, itype
,
11040 build_zero_cst (itype
));
11043 if (TREE_CODE (arg0
) == BIT_XOR_EXPR
11044 && TREE_CODE (arg1
) == BIT_XOR_EXPR
)
11046 tree arg00
= TREE_OPERAND (arg0
, 0);
11047 tree arg01
= TREE_OPERAND (arg0
, 1);
11048 tree arg10
= TREE_OPERAND (arg1
, 0);
11049 tree arg11
= TREE_OPERAND (arg1
, 1);
11050 tree itype
= TREE_TYPE (arg0
);
11052 /* Optimize (X ^ Z) op (Y ^ Z) as X op Y, and symmetries.
11053 operand_equal_p guarantees no side-effects so we don't need
11054 to use omit_one_operand on Z. */
11055 if (operand_equal_p (arg01
, arg11
, 0))
11056 return fold_build2_loc (loc
, code
, type
, arg00
,
11057 fold_convert_loc (loc
, TREE_TYPE (arg00
),
11059 if (operand_equal_p (arg01
, arg10
, 0))
11060 return fold_build2_loc (loc
, code
, type
, arg00
,
11061 fold_convert_loc (loc
, TREE_TYPE (arg00
),
11063 if (operand_equal_p (arg00
, arg11
, 0))
11064 return fold_build2_loc (loc
, code
, type
, arg01
,
11065 fold_convert_loc (loc
, TREE_TYPE (arg01
),
11067 if (operand_equal_p (arg00
, arg10
, 0))
11068 return fold_build2_loc (loc
, code
, type
, arg01
,
11069 fold_convert_loc (loc
, TREE_TYPE (arg01
),
11072 /* Optimize (X ^ C1) op (Y ^ C2) as (X ^ (C1 ^ C2)) op Y. */
11073 if (TREE_CODE (arg01
) == INTEGER_CST
11074 && TREE_CODE (arg11
) == INTEGER_CST
)
11076 tem
= fold_build2_loc (loc
, BIT_XOR_EXPR
, itype
, arg01
,
11077 fold_convert_loc (loc
, itype
, arg11
));
11078 tem
= fold_build2_loc (loc
, BIT_XOR_EXPR
, itype
, arg00
, tem
);
11079 return fold_build2_loc (loc
, code
, type
, tem
,
11080 fold_convert_loc (loc
, itype
, arg10
));
11084 /* Attempt to simplify equality/inequality comparisons of complex
11085 values. Only lower the comparison if the result is known or
11086 can be simplified to a single scalar comparison. */
11087 if ((TREE_CODE (arg0
) == COMPLEX_EXPR
11088 || TREE_CODE (arg0
) == COMPLEX_CST
)
11089 && (TREE_CODE (arg1
) == COMPLEX_EXPR
11090 || TREE_CODE (arg1
) == COMPLEX_CST
))
11092 tree real0
, imag0
, real1
, imag1
;
11095 if (TREE_CODE (arg0
) == COMPLEX_EXPR
)
11097 real0
= TREE_OPERAND (arg0
, 0);
11098 imag0
= TREE_OPERAND (arg0
, 1);
11102 real0
= TREE_REALPART (arg0
);
11103 imag0
= TREE_IMAGPART (arg0
);
11106 if (TREE_CODE (arg1
) == COMPLEX_EXPR
)
11108 real1
= TREE_OPERAND (arg1
, 0);
11109 imag1
= TREE_OPERAND (arg1
, 1);
11113 real1
= TREE_REALPART (arg1
);
11114 imag1
= TREE_IMAGPART (arg1
);
11117 rcond
= fold_binary_loc (loc
, code
, type
, real0
, real1
);
11118 if (rcond
&& TREE_CODE (rcond
) == INTEGER_CST
)
11120 if (integer_zerop (rcond
))
11122 if (code
== EQ_EXPR
)
11123 return omit_two_operands_loc (loc
, type
, boolean_false_node
,
11125 return fold_build2_loc (loc
, NE_EXPR
, type
, imag0
, imag1
);
11129 if (code
== NE_EXPR
)
11130 return omit_two_operands_loc (loc
, type
, boolean_true_node
,
11132 return fold_build2_loc (loc
, EQ_EXPR
, type
, imag0
, imag1
);
11136 icond
= fold_binary_loc (loc
, code
, type
, imag0
, imag1
);
11137 if (icond
&& TREE_CODE (icond
) == INTEGER_CST
)
11139 if (integer_zerop (icond
))
11141 if (code
== EQ_EXPR
)
11142 return omit_two_operands_loc (loc
, type
, boolean_false_node
,
11144 return fold_build2_loc (loc
, NE_EXPR
, type
, real0
, real1
);
11148 if (code
== NE_EXPR
)
11149 return omit_two_operands_loc (loc
, type
, boolean_true_node
,
11151 return fold_build2_loc (loc
, EQ_EXPR
, type
, real0
, real1
);
11162 tem
= fold_comparison (loc
, code
, type
, op0
, op1
);
11163 if (tem
!= NULL_TREE
)
11166 /* Transform comparisons of the form X +- C CMP X. */
11167 if ((TREE_CODE (arg0
) == PLUS_EXPR
|| TREE_CODE (arg0
) == MINUS_EXPR
)
11168 && operand_equal_p (TREE_OPERAND (arg0
, 0), arg1
, 0)
11169 && ((TREE_CODE (TREE_OPERAND (arg0
, 1)) == REAL_CST
11170 && !HONOR_SNANS (arg0
))
11171 || (TREE_CODE (TREE_OPERAND (arg0
, 1)) == INTEGER_CST
11172 && TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (arg1
)))))
11174 tree arg01
= TREE_OPERAND (arg0
, 1);
11175 enum tree_code code0
= TREE_CODE (arg0
);
11178 if (TREE_CODE (arg01
) == REAL_CST
)
11179 is_positive
= REAL_VALUE_NEGATIVE (TREE_REAL_CST (arg01
)) ? -1 : 1;
11181 is_positive
= tree_int_cst_sgn (arg01
);
11183 /* (X - c) > X becomes false. */
11184 if (code
== GT_EXPR
11185 && ((code0
== MINUS_EXPR
&& is_positive
>= 0)
11186 || (code0
== PLUS_EXPR
&& is_positive
<= 0)))
11188 if (TREE_CODE (arg01
) == INTEGER_CST
11189 && TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (arg1
)))
11190 fold_overflow_warning (("assuming signed overflow does not "
11191 "occur when assuming that (X - c) > X "
11192 "is always false"),
11193 WARN_STRICT_OVERFLOW_ALL
);
11194 return constant_boolean_node (0, type
);
11197 /* Likewise (X + c) < X becomes false. */
11198 if (code
== LT_EXPR
11199 && ((code0
== PLUS_EXPR
&& is_positive
>= 0)
11200 || (code0
== MINUS_EXPR
&& is_positive
<= 0)))
11202 if (TREE_CODE (arg01
) == INTEGER_CST
11203 && TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (arg1
)))
11204 fold_overflow_warning (("assuming signed overflow does not "
11205 "occur when assuming that "
11206 "(X + c) < X is always false"),
11207 WARN_STRICT_OVERFLOW_ALL
);
11208 return constant_boolean_node (0, type
);
11211 /* Convert (X - c) <= X to true. */
11212 if (!HONOR_NANS (arg1
)
11214 && ((code0
== MINUS_EXPR
&& is_positive
>= 0)
11215 || (code0
== PLUS_EXPR
&& is_positive
<= 0)))
11217 if (TREE_CODE (arg01
) == INTEGER_CST
11218 && TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (arg1
)))
11219 fold_overflow_warning (("assuming signed overflow does not "
11220 "occur when assuming that "
11221 "(X - c) <= X is always true"),
11222 WARN_STRICT_OVERFLOW_ALL
);
11223 return constant_boolean_node (1, type
);
11226 /* Convert (X + c) >= X to true. */
11227 if (!HONOR_NANS (arg1
)
11229 && ((code0
== PLUS_EXPR
&& is_positive
>= 0)
11230 || (code0
== MINUS_EXPR
&& is_positive
<= 0)))
11232 if (TREE_CODE (arg01
) == INTEGER_CST
11233 && TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (arg1
)))
11234 fold_overflow_warning (("assuming signed overflow does not "
11235 "occur when assuming that "
11236 "(X + c) >= X is always true"),
11237 WARN_STRICT_OVERFLOW_ALL
);
11238 return constant_boolean_node (1, type
);
11241 if (TREE_CODE (arg01
) == INTEGER_CST
)
11243 /* Convert X + c > X and X - c < X to true for integers. */
11244 if (code
== GT_EXPR
11245 && ((code0
== PLUS_EXPR
&& is_positive
> 0)
11246 || (code0
== MINUS_EXPR
&& is_positive
< 0)))
11248 if (TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (arg1
)))
11249 fold_overflow_warning (("assuming signed overflow does "
11250 "not occur when assuming that "
11251 "(X + c) > X is always true"),
11252 WARN_STRICT_OVERFLOW_ALL
);
11253 return constant_boolean_node (1, type
);
11256 if (code
== LT_EXPR
11257 && ((code0
== MINUS_EXPR
&& is_positive
> 0)
11258 || (code0
== PLUS_EXPR
&& is_positive
< 0)))
11260 if (TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (arg1
)))
11261 fold_overflow_warning (("assuming signed overflow does "
11262 "not occur when assuming that "
11263 "(X - c) < X is always true"),
11264 WARN_STRICT_OVERFLOW_ALL
);
11265 return constant_boolean_node (1, type
);
11268 /* Convert X + c <= X and X - c >= X to false for integers. */
11269 if (code
== LE_EXPR
11270 && ((code0
== PLUS_EXPR
&& is_positive
> 0)
11271 || (code0
== MINUS_EXPR
&& is_positive
< 0)))
11273 if (TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (arg1
)))
11274 fold_overflow_warning (("assuming signed overflow does "
11275 "not occur when assuming that "
11276 "(X + c) <= X is always false"),
11277 WARN_STRICT_OVERFLOW_ALL
);
11278 return constant_boolean_node (0, type
);
11281 if (code
== GE_EXPR
11282 && ((code0
== MINUS_EXPR
&& is_positive
> 0)
11283 || (code0
== PLUS_EXPR
&& is_positive
< 0)))
11285 if (TYPE_OVERFLOW_UNDEFINED (TREE_TYPE (arg1
)))
11286 fold_overflow_warning (("assuming signed overflow does "
11287 "not occur when assuming that "
11288 "(X - c) >= X is always false"),
11289 WARN_STRICT_OVERFLOW_ALL
);
11290 return constant_boolean_node (0, type
);
11295 /* If we are comparing an ABS_EXPR with a constant, we can
11296 convert all the cases into explicit comparisons, but they may
11297 well not be faster than doing the ABS and one comparison.
11298 But ABS (X) <= C is a range comparison, which becomes a subtraction
11299 and a comparison, and is probably faster. */
11300 if (code
== LE_EXPR
11301 && TREE_CODE (arg1
) == INTEGER_CST
11302 && TREE_CODE (arg0
) == ABS_EXPR
11303 && ! TREE_SIDE_EFFECTS (arg0
)
11304 && (0 != (tem
= negate_expr (arg1
)))
11305 && TREE_CODE (tem
) == INTEGER_CST
11306 && !TREE_OVERFLOW (tem
))
11307 return fold_build2_loc (loc
, TRUTH_ANDIF_EXPR
, type
,
11308 build2 (GE_EXPR
, type
,
11309 TREE_OPERAND (arg0
, 0), tem
),
11310 build2 (LE_EXPR
, type
,
11311 TREE_OPERAND (arg0
, 0), arg1
));
11313 /* Convert ABS_EXPR<x> >= 0 to true. */
11314 strict_overflow_p
= false;
11315 if (code
== GE_EXPR
11316 && (integer_zerop (arg1
)
11317 || (! HONOR_NANS (arg0
)
11318 && real_zerop (arg1
)))
11319 && tree_expr_nonnegative_warnv_p (arg0
, &strict_overflow_p
))
11321 if (strict_overflow_p
)
11322 fold_overflow_warning (("assuming signed overflow does not occur "
11323 "when simplifying comparison of "
11324 "absolute value and zero"),
11325 WARN_STRICT_OVERFLOW_CONDITIONAL
);
11326 return omit_one_operand_loc (loc
, type
,
11327 constant_boolean_node (true, type
),
11331 /* Convert ABS_EXPR<x> < 0 to false. */
11332 strict_overflow_p
= false;
11333 if (code
== LT_EXPR
11334 && (integer_zerop (arg1
) || real_zerop (arg1
))
11335 && tree_expr_nonnegative_warnv_p (arg0
, &strict_overflow_p
))
11337 if (strict_overflow_p
)
11338 fold_overflow_warning (("assuming signed overflow does not occur "
11339 "when simplifying comparison of "
11340 "absolute value and zero"),
11341 WARN_STRICT_OVERFLOW_CONDITIONAL
);
11342 return omit_one_operand_loc (loc
, type
,
11343 constant_boolean_node (false, type
),
11347 /* If X is unsigned, convert X < (1 << Y) into X >> Y == 0
11348 and similarly for >= into !=. */
11349 if ((code
== LT_EXPR
|| code
== GE_EXPR
)
11350 && TYPE_UNSIGNED (TREE_TYPE (arg0
))
11351 && TREE_CODE (arg1
) == LSHIFT_EXPR
11352 && integer_onep (TREE_OPERAND (arg1
, 0)))
11353 return build2_loc (loc
, code
== LT_EXPR
? EQ_EXPR
: NE_EXPR
, type
,
11354 build2 (RSHIFT_EXPR
, TREE_TYPE (arg0
), arg0
,
11355 TREE_OPERAND (arg1
, 1)),
11356 build_zero_cst (TREE_TYPE (arg0
)));
11358 /* Similarly for X < (cast) (1 << Y). But cast can't be narrowing,
11359 otherwise Y might be >= # of bits in X's type and thus e.g.
11360 (unsigned char) (1 << Y) for Y 15 might be 0.
11361 If the cast is widening, then 1 << Y should have unsigned type,
11362 otherwise if Y is number of bits in the signed shift type minus 1,
11363 we can't optimize this. E.g. (unsigned long long) (1 << Y) for Y
11364 31 might be 0xffffffff80000000. */
11365 if ((code
== LT_EXPR
|| code
== GE_EXPR
)
11366 && TYPE_UNSIGNED (TREE_TYPE (arg0
))
11367 && CONVERT_EXPR_P (arg1
)
11368 && TREE_CODE (TREE_OPERAND (arg1
, 0)) == LSHIFT_EXPR
11369 && (element_precision (TREE_TYPE (arg1
))
11370 >= element_precision (TREE_TYPE (TREE_OPERAND (arg1
, 0))))
11371 && (TYPE_UNSIGNED (TREE_TYPE (TREE_OPERAND (arg1
, 0)))
11372 || (element_precision (TREE_TYPE (arg1
))
11373 == element_precision (TREE_TYPE (TREE_OPERAND (arg1
, 0)))))
11374 && integer_onep (TREE_OPERAND (TREE_OPERAND (arg1
, 0), 0)))
11376 tem
= build2 (RSHIFT_EXPR
, TREE_TYPE (arg0
), arg0
,
11377 TREE_OPERAND (TREE_OPERAND (arg1
, 0), 1));
11378 return build2_loc (loc
, code
== LT_EXPR
? EQ_EXPR
: NE_EXPR
, type
,
11379 fold_convert_loc (loc
, TREE_TYPE (arg0
), tem
),
11380 build_zero_cst (TREE_TYPE (arg0
)));
11385 case UNORDERED_EXPR
:
11393 /* Fold (double)float1 CMP (double)float2 into float1 CMP float2. */
11395 tree targ0
= strip_float_extensions (arg0
);
11396 tree targ1
= strip_float_extensions (arg1
);
11397 tree newtype
= TREE_TYPE (targ0
);
11399 if (TYPE_PRECISION (TREE_TYPE (targ1
)) > TYPE_PRECISION (newtype
))
11400 newtype
= TREE_TYPE (targ1
);
11402 if (TYPE_PRECISION (newtype
) < TYPE_PRECISION (TREE_TYPE (arg0
)))
11403 return fold_build2_loc (loc
, code
, type
,
11404 fold_convert_loc (loc
, newtype
, targ0
),
11405 fold_convert_loc (loc
, newtype
, targ1
));
11410 case COMPOUND_EXPR
:
11411 /* When pedantic, a compound expression can be neither an lvalue
11412 nor an integer constant expression. */
11413 if (TREE_SIDE_EFFECTS (arg0
) || TREE_CONSTANT (arg1
))
11415 /* Don't let (0, 0) be null pointer constant. */
11416 tem
= integer_zerop (arg1
) ? build1 (NOP_EXPR
, type
, arg1
)
11417 : fold_convert_loc (loc
, type
, arg1
);
11418 return pedantic_non_lvalue_loc (loc
, tem
);
11421 /* An ASSERT_EXPR should never be passed to fold_binary. */
11422 gcc_unreachable ();
11426 } /* switch (code) */
11429 /* Callback for walk_tree, looking for LABEL_EXPR. Return *TP if it is
11430 a LABEL_EXPR; otherwise return NULL_TREE. Do not check the subtrees
11434 contains_label_1 (tree
*tp
, int *walk_subtrees
, void *data ATTRIBUTE_UNUSED
)
11436 switch (TREE_CODE (*tp
))
11442 *walk_subtrees
= 0;
11444 /* ... fall through ... */
11451 /* Return whether the sub-tree ST contains a label which is accessible from
11452 outside the sub-tree. */
11455 contains_label_p (tree st
)
11458 (walk_tree_without_duplicates (&st
, contains_label_1
, NULL
) != NULL_TREE
);
11461 /* Fold a ternary expression of code CODE and type TYPE with operands
11462 OP0, OP1, and OP2. Return the folded expression if folding is
11463 successful. Otherwise, return NULL_TREE. */
11466 fold_ternary_loc (location_t loc
, enum tree_code code
, tree type
,
11467 tree op0
, tree op1
, tree op2
)
11470 tree arg0
= NULL_TREE
, arg1
= NULL_TREE
, arg2
= NULL_TREE
;
11471 enum tree_code_class kind
= TREE_CODE_CLASS (code
);
11473 gcc_assert (IS_EXPR_CODE_CLASS (kind
)
11474 && TREE_CODE_LENGTH (code
) == 3);
11476 /* If this is a commutative operation, and OP0 is a constant, move it
11477 to OP1 to reduce the number of tests below. */
11478 if (commutative_ternary_tree_code (code
)
11479 && tree_swap_operands_p (op0
, op1
, true))
11480 return fold_build3_loc (loc
, code
, type
, op1
, op0
, op2
);
11482 tem
= generic_simplify (loc
, code
, type
, op0
, op1
, op2
);
11486 /* Strip any conversions that don't change the mode. This is safe
11487 for every expression, except for a comparison expression because
11488 its signedness is derived from its operands. So, in the latter
11489 case, only strip conversions that don't change the signedness.
11491 Note that this is done as an internal manipulation within the
11492 constant folder, in order to find the simplest representation of
11493 the arguments so that their form can be studied. In any cases,
11494 the appropriate type conversions should be put back in the tree
11495 that will get out of the constant folder. */
11516 case COMPONENT_REF
:
11517 if (TREE_CODE (arg0
) == CONSTRUCTOR
11518 && ! type_contains_placeholder_p (TREE_TYPE (arg0
)))
11520 unsigned HOST_WIDE_INT idx
;
11522 FOR_EACH_CONSTRUCTOR_ELT (CONSTRUCTOR_ELTS (arg0
), idx
, field
, value
)
11529 case VEC_COND_EXPR
:
11530 /* Pedantic ANSI C says that a conditional expression is never an lvalue,
11531 so all simple results must be passed through pedantic_non_lvalue. */
11532 if (TREE_CODE (arg0
) == INTEGER_CST
)
11534 tree unused_op
= integer_zerop (arg0
) ? op1
: op2
;
11535 tem
= integer_zerop (arg0
) ? op2
: op1
;
11536 /* Only optimize constant conditions when the selected branch
11537 has the same type as the COND_EXPR. This avoids optimizing
11538 away "c ? x : throw", where the throw has a void type.
11539 Avoid throwing away that operand which contains label. */
11540 if ((!TREE_SIDE_EFFECTS (unused_op
)
11541 || !contains_label_p (unused_op
))
11542 && (! VOID_TYPE_P (TREE_TYPE (tem
))
11543 || VOID_TYPE_P (type
)))
11544 return pedantic_non_lvalue_loc (loc
, tem
);
11547 else if (TREE_CODE (arg0
) == VECTOR_CST
)
11549 if ((TREE_CODE (arg1
) == VECTOR_CST
11550 || TREE_CODE (arg1
) == CONSTRUCTOR
)
11551 && (TREE_CODE (arg2
) == VECTOR_CST
11552 || TREE_CODE (arg2
) == CONSTRUCTOR
))
11554 unsigned int nelts
= TYPE_VECTOR_SUBPARTS (type
), i
;
11555 unsigned char *sel
= XALLOCAVEC (unsigned char, nelts
);
11556 gcc_assert (nelts
== VECTOR_CST_NELTS (arg0
));
11557 for (i
= 0; i
< nelts
; i
++)
11559 tree val
= VECTOR_CST_ELT (arg0
, i
);
11560 if (integer_all_onesp (val
))
11562 else if (integer_zerop (val
))
11563 sel
[i
] = nelts
+ i
;
11564 else /* Currently unreachable. */
11567 tree t
= fold_vec_perm (type
, arg1
, arg2
, sel
);
11568 if (t
!= NULL_TREE
)
11573 /* If we have A op B ? A : C, we may be able to convert this to a
11574 simpler expression, depending on the operation and the values
11575 of B and C. Signed zeros prevent all of these transformations,
11576 for reasons given above each one.
11578 Also try swapping the arguments and inverting the conditional. */
11579 if (COMPARISON_CLASS_P (arg0
)
11580 && operand_equal_for_comparison_p (TREE_OPERAND (arg0
, 0),
11581 arg1
, TREE_OPERAND (arg0
, 1))
11582 && !HONOR_SIGNED_ZEROS (element_mode (arg1
)))
11584 tem
= fold_cond_expr_with_comparison (loc
, type
, arg0
, op1
, op2
);
11589 if (COMPARISON_CLASS_P (arg0
)
11590 && operand_equal_for_comparison_p (TREE_OPERAND (arg0
, 0),
11592 TREE_OPERAND (arg0
, 1))
11593 && !HONOR_SIGNED_ZEROS (element_mode (op2
)))
11595 location_t loc0
= expr_location_or (arg0
, loc
);
11596 tem
= fold_invert_truthvalue (loc0
, arg0
);
11597 if (tem
&& COMPARISON_CLASS_P (tem
))
11599 tem
= fold_cond_expr_with_comparison (loc
, type
, tem
, op2
, op1
);
11605 /* If the second operand is simpler than the third, swap them
11606 since that produces better jump optimization results. */
11607 if (truth_value_p (TREE_CODE (arg0
))
11608 && tree_swap_operands_p (op1
, op2
, false))
11610 location_t loc0
= expr_location_or (arg0
, loc
);
11611 /* See if this can be inverted. If it can't, possibly because
11612 it was a floating-point inequality comparison, don't do
11614 tem
= fold_invert_truthvalue (loc0
, arg0
);
11616 return fold_build3_loc (loc
, code
, type
, tem
, op2
, op1
);
11619 /* Convert A ? 1 : 0 to simply A. */
11620 if ((code
== VEC_COND_EXPR
? integer_all_onesp (op1
)
11621 : (integer_onep (op1
)
11622 && !VECTOR_TYPE_P (type
)))
11623 && integer_zerop (op2
)
11624 /* If we try to convert OP0 to our type, the
11625 call to fold will try to move the conversion inside
11626 a COND, which will recurse. In that case, the COND_EXPR
11627 is probably the best choice, so leave it alone. */
11628 && type
== TREE_TYPE (arg0
))
11629 return pedantic_non_lvalue_loc (loc
, arg0
);
11631 /* Convert A ? 0 : 1 to !A. This prefers the use of NOT_EXPR
11632 over COND_EXPR in cases such as floating point comparisons. */
11633 if (integer_zerop (op1
)
11634 && (code
== VEC_COND_EXPR
? integer_all_onesp (op2
)
11635 : (integer_onep (op2
)
11636 && !VECTOR_TYPE_P (type
)))
11637 && truth_value_p (TREE_CODE (arg0
)))
11638 return pedantic_non_lvalue_loc (loc
,
11639 fold_convert_loc (loc
, type
,
11640 invert_truthvalue_loc (loc
,
11643 /* A < 0 ? <sign bit of A> : 0 is simply (A & <sign bit of A>). */
11644 if (TREE_CODE (arg0
) == LT_EXPR
11645 && integer_zerop (TREE_OPERAND (arg0
, 1))
11646 && integer_zerop (op2
)
11647 && (tem
= sign_bit_p (TREE_OPERAND (arg0
, 0), arg1
)))
11649 /* sign_bit_p looks through both zero and sign extensions,
11650 but for this optimization only sign extensions are
11652 tree tem2
= TREE_OPERAND (arg0
, 0);
11653 while (tem
!= tem2
)
11655 if (TREE_CODE (tem2
) != NOP_EXPR
11656 || TYPE_UNSIGNED (TREE_TYPE (TREE_OPERAND (tem2
, 0))))
11661 tem2
= TREE_OPERAND (tem2
, 0);
11663 /* sign_bit_p only checks ARG1 bits within A's precision.
11664 If <sign bit of A> has wider type than A, bits outside
11665 of A's precision in <sign bit of A> need to be checked.
11666 If they are all 0, this optimization needs to be done
11667 in unsigned A's type, if they are all 1 in signed A's type,
11668 otherwise this can't be done. */
11670 && TYPE_PRECISION (TREE_TYPE (tem
))
11671 < TYPE_PRECISION (TREE_TYPE (arg1
))
11672 && TYPE_PRECISION (TREE_TYPE (tem
))
11673 < TYPE_PRECISION (type
))
11675 int inner_width
, outer_width
;
11678 inner_width
= TYPE_PRECISION (TREE_TYPE (tem
));
11679 outer_width
= TYPE_PRECISION (TREE_TYPE (arg1
));
11680 if (outer_width
> TYPE_PRECISION (type
))
11681 outer_width
= TYPE_PRECISION (type
);
11683 wide_int mask
= wi::shifted_mask
11684 (inner_width
, outer_width
- inner_width
, false,
11685 TYPE_PRECISION (TREE_TYPE (arg1
)));
11687 wide_int common
= mask
& arg1
;
11688 if (common
== mask
)
11690 tem_type
= signed_type_for (TREE_TYPE (tem
));
11691 tem
= fold_convert_loc (loc
, tem_type
, tem
);
11693 else if (common
== 0)
11695 tem_type
= unsigned_type_for (TREE_TYPE (tem
));
11696 tem
= fold_convert_loc (loc
, tem_type
, tem
);
11704 fold_convert_loc (loc
, type
,
11705 fold_build2_loc (loc
, BIT_AND_EXPR
,
11706 TREE_TYPE (tem
), tem
,
11707 fold_convert_loc (loc
,
11712 /* (A >> N) & 1 ? (1 << N) : 0 is simply A & (1 << N). A & 1 was
11713 already handled above. */
11714 if (TREE_CODE (arg0
) == BIT_AND_EXPR
11715 && integer_onep (TREE_OPERAND (arg0
, 1))
11716 && integer_zerop (op2
)
11717 && integer_pow2p (arg1
))
11719 tree tem
= TREE_OPERAND (arg0
, 0);
11721 if (TREE_CODE (tem
) == RSHIFT_EXPR
11722 && tree_fits_uhwi_p (TREE_OPERAND (tem
, 1))
11723 && (unsigned HOST_WIDE_INT
) tree_log2 (arg1
) ==
11724 tree_to_uhwi (TREE_OPERAND (tem
, 1)))
11725 return fold_build2_loc (loc
, BIT_AND_EXPR
, type
,
11726 TREE_OPERAND (tem
, 0), arg1
);
11729 /* A & N ? N : 0 is simply A & N if N is a power of two. This
11730 is probably obsolete because the first operand should be a
11731 truth value (that's why we have the two cases above), but let's
11732 leave it in until we can confirm this for all front-ends. */
11733 if (integer_zerop (op2
)
11734 && TREE_CODE (arg0
) == NE_EXPR
11735 && integer_zerop (TREE_OPERAND (arg0
, 1))
11736 && integer_pow2p (arg1
)
11737 && TREE_CODE (TREE_OPERAND (arg0
, 0)) == BIT_AND_EXPR
11738 && operand_equal_p (TREE_OPERAND (TREE_OPERAND (arg0
, 0), 1),
11739 arg1
, OEP_ONLY_CONST
))
11740 return pedantic_non_lvalue_loc (loc
,
11741 fold_convert_loc (loc
, type
,
11742 TREE_OPERAND (arg0
, 0)));
11744 /* Disable the transformations below for vectors, since
11745 fold_binary_op_with_conditional_arg may undo them immediately,
11746 yielding an infinite loop. */
11747 if (code
== VEC_COND_EXPR
)
11750 /* Convert A ? B : 0 into A && B if A and B are truth values. */
11751 if (integer_zerop (op2
)
11752 && truth_value_p (TREE_CODE (arg0
))
11753 && truth_value_p (TREE_CODE (arg1
))
11754 && (code
== VEC_COND_EXPR
|| !VECTOR_TYPE_P (type
)))
11755 return fold_build2_loc (loc
, code
== VEC_COND_EXPR
? BIT_AND_EXPR
11756 : TRUTH_ANDIF_EXPR
,
11757 type
, fold_convert_loc (loc
, type
, arg0
), arg1
);
11759 /* Convert A ? B : 1 into !A || B if A and B are truth values. */
11760 if (code
== VEC_COND_EXPR
? integer_all_onesp (op2
) : integer_onep (op2
)
11761 && truth_value_p (TREE_CODE (arg0
))
11762 && truth_value_p (TREE_CODE (arg1
))
11763 && (code
== VEC_COND_EXPR
|| !VECTOR_TYPE_P (type
)))
11765 location_t loc0
= expr_location_or (arg0
, loc
);
11766 /* Only perform transformation if ARG0 is easily inverted. */
11767 tem
= fold_invert_truthvalue (loc0
, arg0
);
11769 return fold_build2_loc (loc
, code
== VEC_COND_EXPR
11772 type
, fold_convert_loc (loc
, type
, tem
),
11776 /* Convert A ? 0 : B into !A && B if A and B are truth values. */
11777 if (integer_zerop (arg1
)
11778 && truth_value_p (TREE_CODE (arg0
))
11779 && truth_value_p (TREE_CODE (op2
))
11780 && (code
== VEC_COND_EXPR
|| !VECTOR_TYPE_P (type
)))
11782 location_t loc0
= expr_location_or (arg0
, loc
);
11783 /* Only perform transformation if ARG0 is easily inverted. */
11784 tem
= fold_invert_truthvalue (loc0
, arg0
);
11786 return fold_build2_loc (loc
, code
== VEC_COND_EXPR
11787 ? BIT_AND_EXPR
: TRUTH_ANDIF_EXPR
,
11788 type
, fold_convert_loc (loc
, type
, tem
),
11792 /* Convert A ? 1 : B into A || B if A and B are truth values. */
11793 if (code
== VEC_COND_EXPR
? integer_all_onesp (arg1
) : integer_onep (arg1
)
11794 && truth_value_p (TREE_CODE (arg0
))
11795 && truth_value_p (TREE_CODE (op2
))
11796 && (code
== VEC_COND_EXPR
|| !VECTOR_TYPE_P (type
)))
11797 return fold_build2_loc (loc
, code
== VEC_COND_EXPR
11798 ? BIT_IOR_EXPR
: TRUTH_ORIF_EXPR
,
11799 type
, fold_convert_loc (loc
, type
, arg0
), op2
);
11804 /* CALL_EXPRs used to be ternary exprs. Catch any mistaken uses
11805 of fold_ternary on them. */
11806 gcc_unreachable ();
11808 case BIT_FIELD_REF
:
11809 if ((TREE_CODE (arg0
) == VECTOR_CST
11810 || (TREE_CODE (arg0
) == CONSTRUCTOR
11811 && TREE_CODE (TREE_TYPE (arg0
)) == VECTOR_TYPE
))
11812 && (type
== TREE_TYPE (TREE_TYPE (arg0
))
11813 || (TREE_CODE (type
) == VECTOR_TYPE
11814 && TREE_TYPE (type
) == TREE_TYPE (TREE_TYPE (arg0
)))))
11816 tree eltype
= TREE_TYPE (TREE_TYPE (arg0
));
11817 unsigned HOST_WIDE_INT width
= tree_to_uhwi (TYPE_SIZE (eltype
));
11818 unsigned HOST_WIDE_INT n
= tree_to_uhwi (arg1
);
11819 unsigned HOST_WIDE_INT idx
= tree_to_uhwi (op2
);
11822 && (idx
% width
) == 0
11823 && (n
% width
) == 0
11824 && ((idx
+ n
) / width
) <= TYPE_VECTOR_SUBPARTS (TREE_TYPE (arg0
)))
11829 if (TREE_CODE (arg0
) == VECTOR_CST
)
11832 return VECTOR_CST_ELT (arg0
, idx
);
11834 tree
*vals
= XALLOCAVEC (tree
, n
);
11835 for (unsigned i
= 0; i
< n
; ++i
)
11836 vals
[i
] = VECTOR_CST_ELT (arg0
, idx
+ i
);
11837 return build_vector (type
, vals
);
11840 /* Constructor elements can be subvectors. */
11841 unsigned HOST_WIDE_INT k
= 1;
11842 if (CONSTRUCTOR_NELTS (arg0
) != 0)
11844 tree cons_elem
= TREE_TYPE (CONSTRUCTOR_ELT (arg0
, 0)->value
);
11845 if (TREE_CODE (cons_elem
) == VECTOR_TYPE
)
11846 k
= TYPE_VECTOR_SUBPARTS (cons_elem
);
11849 /* We keep an exact subset of the constructor elements. */
11850 if ((idx
% k
) == 0 && (n
% k
) == 0)
11852 if (CONSTRUCTOR_NELTS (arg0
) == 0)
11853 return build_constructor (type
, NULL
);
11858 if (idx
< CONSTRUCTOR_NELTS (arg0
))
11859 return CONSTRUCTOR_ELT (arg0
, idx
)->value
;
11860 return build_zero_cst (type
);
11863 vec
<constructor_elt
, va_gc
> *vals
;
11864 vec_alloc (vals
, n
);
11865 for (unsigned i
= 0;
11866 i
< n
&& idx
+ i
< CONSTRUCTOR_NELTS (arg0
);
11868 CONSTRUCTOR_APPEND_ELT (vals
, NULL_TREE
,
11870 (arg0
, idx
+ i
)->value
);
11871 return build_constructor (type
, vals
);
11873 /* The bitfield references a single constructor element. */
11874 else if (idx
+ n
<= (idx
/ k
+ 1) * k
)
11876 if (CONSTRUCTOR_NELTS (arg0
) <= idx
/ k
)
11877 return build_zero_cst (type
);
11879 return CONSTRUCTOR_ELT (arg0
, idx
/ k
)->value
;
11881 return fold_build3_loc (loc
, code
, type
,
11882 CONSTRUCTOR_ELT (arg0
, idx
/ k
)->value
, op1
,
11883 build_int_cst (TREE_TYPE (op2
), (idx
% k
) * width
));
11888 /* A bit-field-ref that referenced the full argument can be stripped. */
11889 if (INTEGRAL_TYPE_P (TREE_TYPE (arg0
))
11890 && TYPE_PRECISION (TREE_TYPE (arg0
)) == tree_to_uhwi (arg1
)
11891 && integer_zerop (op2
))
11892 return fold_convert_loc (loc
, type
, arg0
);
11894 /* On constants we can use native encode/interpret to constant
11895 fold (nearly) all BIT_FIELD_REFs. */
11896 if (CONSTANT_CLASS_P (arg0
)
11897 && can_native_interpret_type_p (type
)
11898 && tree_fits_uhwi_p (TYPE_SIZE_UNIT (TREE_TYPE (arg0
)))
11899 /* This limitation should not be necessary, we just need to
11900 round this up to mode size. */
11901 && tree_to_uhwi (op1
) % BITS_PER_UNIT
== 0
11902 /* Need bit-shifting of the buffer to relax the following. */
11903 && tree_to_uhwi (op2
) % BITS_PER_UNIT
== 0)
11905 unsigned HOST_WIDE_INT bitpos
= tree_to_uhwi (op2
);
11906 unsigned HOST_WIDE_INT bitsize
= tree_to_uhwi (op1
);
11907 unsigned HOST_WIDE_INT clen
;
11908 clen
= tree_to_uhwi (TYPE_SIZE_UNIT (TREE_TYPE (arg0
)));
11909 /* ??? We cannot tell native_encode_expr to start at
11910 some random byte only. So limit us to a reasonable amount
11914 unsigned char *b
= XALLOCAVEC (unsigned char, clen
);
11915 unsigned HOST_WIDE_INT len
= native_encode_expr (arg0
, b
, clen
);
11917 && len
* BITS_PER_UNIT
>= bitpos
+ bitsize
)
11919 tree v
= native_interpret_expr (type
,
11920 b
+ bitpos
/ BITS_PER_UNIT
,
11921 bitsize
/ BITS_PER_UNIT
);
11931 /* For integers we can decompose the FMA if possible. */
11932 if (TREE_CODE (arg0
) == INTEGER_CST
11933 && TREE_CODE (arg1
) == INTEGER_CST
)
11934 return fold_build2_loc (loc
, PLUS_EXPR
, type
,
11935 const_binop (MULT_EXPR
, arg0
, arg1
), arg2
);
11936 if (integer_zerop (arg2
))
11937 return fold_build2_loc (loc
, MULT_EXPR
, type
, arg0
, arg1
);
11939 return fold_fma (loc
, type
, arg0
, arg1
, arg2
);
11941 case VEC_PERM_EXPR
:
11942 if (TREE_CODE (arg2
) == VECTOR_CST
)
11944 unsigned int nelts
= TYPE_VECTOR_SUBPARTS (type
), i
, mask
, mask2
;
11945 unsigned char *sel
= XALLOCAVEC (unsigned char, 2 * nelts
);
11946 unsigned char *sel2
= sel
+ nelts
;
11947 bool need_mask_canon
= false;
11948 bool need_mask_canon2
= false;
11949 bool all_in_vec0
= true;
11950 bool all_in_vec1
= true;
11951 bool maybe_identity
= true;
11952 bool single_arg
= (op0
== op1
);
11953 bool changed
= false;
11955 mask2
= 2 * nelts
- 1;
11956 mask
= single_arg
? (nelts
- 1) : mask2
;
11957 gcc_assert (nelts
== VECTOR_CST_NELTS (arg2
));
11958 for (i
= 0; i
< nelts
; i
++)
11960 tree val
= VECTOR_CST_ELT (arg2
, i
);
11961 if (TREE_CODE (val
) != INTEGER_CST
)
11964 /* Make sure that the perm value is in an acceptable
11967 need_mask_canon
|= wi::gtu_p (t
, mask
);
11968 need_mask_canon2
|= wi::gtu_p (t
, mask2
);
11969 sel
[i
] = t
.to_uhwi () & mask
;
11970 sel2
[i
] = t
.to_uhwi () & mask2
;
11972 if (sel
[i
] < nelts
)
11973 all_in_vec1
= false;
11975 all_in_vec0
= false;
11977 if ((sel
[i
] & (nelts
-1)) != i
)
11978 maybe_identity
= false;
11981 if (maybe_identity
)
11991 else if (all_in_vec1
)
11994 for (i
= 0; i
< nelts
; i
++)
11996 need_mask_canon
= true;
11999 if ((TREE_CODE (op0
) == VECTOR_CST
12000 || TREE_CODE (op0
) == CONSTRUCTOR
)
12001 && (TREE_CODE (op1
) == VECTOR_CST
12002 || TREE_CODE (op1
) == CONSTRUCTOR
))
12004 tree t
= fold_vec_perm (type
, op0
, op1
, sel
);
12005 if (t
!= NULL_TREE
)
12009 if (op0
== op1
&& !single_arg
)
12012 /* Some targets are deficient and fail to expand a single
12013 argument permutation while still allowing an equivalent
12014 2-argument version. */
12015 if (need_mask_canon
&& arg2
== op2
12016 && !can_vec_perm_p (TYPE_MODE (type
), false, sel
)
12017 && can_vec_perm_p (TYPE_MODE (type
), false, sel2
))
12019 need_mask_canon
= need_mask_canon2
;
12023 if (need_mask_canon
&& arg2
== op2
)
12025 tree
*tsel
= XALLOCAVEC (tree
, nelts
);
12026 tree eltype
= TREE_TYPE (TREE_TYPE (arg2
));
12027 for (i
= 0; i
< nelts
; i
++)
12028 tsel
[i
] = build_int_cst (eltype
, sel
[i
]);
12029 op2
= build_vector (TREE_TYPE (arg2
), tsel
);
12034 return build3_loc (loc
, VEC_PERM_EXPR
, type
, op0
, op1
, op2
);
12040 } /* switch (code) */
12043 /* Gets the element ACCESS_INDEX from CTOR, which must be a CONSTRUCTOR
12044 of an array (or vector). */
12047 get_array_ctor_element_at_index (tree ctor
, offset_int access_index
)
12049 tree index_type
= NULL_TREE
;
12050 offset_int low_bound
= 0;
12052 if (TREE_CODE (TREE_TYPE (ctor
)) == ARRAY_TYPE
)
12054 tree domain_type
= TYPE_DOMAIN (TREE_TYPE (ctor
));
12055 if (domain_type
&& TYPE_MIN_VALUE (domain_type
))
12057 /* Static constructors for variably sized objects makes no sense. */
12058 gcc_assert (TREE_CODE (TYPE_MIN_VALUE (domain_type
)) == INTEGER_CST
);
12059 index_type
= TREE_TYPE (TYPE_MIN_VALUE (domain_type
));
12060 low_bound
= wi::to_offset (TYPE_MIN_VALUE (domain_type
));
12065 access_index
= wi::ext (access_index
, TYPE_PRECISION (index_type
),
12066 TYPE_SIGN (index_type
));
12068 offset_int index
= low_bound
- 1;
12070 index
= wi::ext (index
, TYPE_PRECISION (index_type
),
12071 TYPE_SIGN (index_type
));
12073 offset_int max_index
;
12074 unsigned HOST_WIDE_INT cnt
;
12077 FOR_EACH_CONSTRUCTOR_ELT (CONSTRUCTOR_ELTS (ctor
), cnt
, cfield
, cval
)
12079 /* Array constructor might explicitly set index, or specify a range,
12080 or leave index NULL meaning that it is next index after previous
12084 if (TREE_CODE (cfield
) == INTEGER_CST
)
12085 max_index
= index
= wi::to_offset (cfield
);
12088 gcc_assert (TREE_CODE (cfield
) == RANGE_EXPR
);
12089 index
= wi::to_offset (TREE_OPERAND (cfield
, 0));
12090 max_index
= wi::to_offset (TREE_OPERAND (cfield
, 1));
12097 index
= wi::ext (index
, TYPE_PRECISION (index_type
),
12098 TYPE_SIGN (index_type
));
12102 /* Do we have match? */
12103 if (wi::cmpu (access_index
, index
) >= 0
12104 && wi::cmpu (access_index
, max_index
) <= 0)
12110 /* Perform constant folding and related simplification of EXPR.
12111 The related simplifications include x*1 => x, x*0 => 0, etc.,
12112 and application of the associative law.
12113 NOP_EXPR conversions may be removed freely (as long as we
12114 are careful not to change the type of the overall expression).
12115 We cannot simplify through a CONVERT_EXPR, FIX_EXPR or FLOAT_EXPR,
12116 but we can constant-fold them if they have constant operands. */
12118 #ifdef ENABLE_FOLD_CHECKING
12119 # define fold(x) fold_1 (x)
12120 static tree
fold_1 (tree
);
12126 const tree t
= expr
;
12127 enum tree_code code
= TREE_CODE (t
);
12128 enum tree_code_class kind
= TREE_CODE_CLASS (code
);
12130 location_t loc
= EXPR_LOCATION (expr
);
12132 /* Return right away if a constant. */
12133 if (kind
== tcc_constant
)
12136 /* CALL_EXPR-like objects with variable numbers of operands are
12137 treated specially. */
12138 if (kind
== tcc_vl_exp
)
12140 if (code
== CALL_EXPR
)
12142 tem
= fold_call_expr (loc
, expr
, false);
12143 return tem
? tem
: expr
;
12148 if (IS_EXPR_CODE_CLASS (kind
))
12150 tree type
= TREE_TYPE (t
);
12151 tree op0
, op1
, op2
;
12153 switch (TREE_CODE_LENGTH (code
))
12156 op0
= TREE_OPERAND (t
, 0);
12157 tem
= fold_unary_loc (loc
, code
, type
, op0
);
12158 return tem
? tem
: expr
;
12160 op0
= TREE_OPERAND (t
, 0);
12161 op1
= TREE_OPERAND (t
, 1);
12162 tem
= fold_binary_loc (loc
, code
, type
, op0
, op1
);
12163 return tem
? tem
: expr
;
12165 op0
= TREE_OPERAND (t
, 0);
12166 op1
= TREE_OPERAND (t
, 1);
12167 op2
= TREE_OPERAND (t
, 2);
12168 tem
= fold_ternary_loc (loc
, code
, type
, op0
, op1
, op2
);
12169 return tem
? tem
: expr
;
12179 tree op0
= TREE_OPERAND (t
, 0);
12180 tree op1
= TREE_OPERAND (t
, 1);
12182 if (TREE_CODE (op1
) == INTEGER_CST
12183 && TREE_CODE (op0
) == CONSTRUCTOR
12184 && ! type_contains_placeholder_p (TREE_TYPE (op0
)))
12186 tree val
= get_array_ctor_element_at_index (op0
,
12187 wi::to_offset (op1
));
12195 /* Return a VECTOR_CST if possible. */
12198 tree type
= TREE_TYPE (t
);
12199 if (TREE_CODE (type
) != VECTOR_TYPE
)
12204 FOR_EACH_CONSTRUCTOR_VALUE (CONSTRUCTOR_ELTS (t
), i
, val
)
12205 if (! CONSTANT_CLASS_P (val
))
12208 return build_vector_from_ctor (type
, CONSTRUCTOR_ELTS (t
));
12212 return fold (DECL_INITIAL (t
));
12216 } /* switch (code) */
12219 #ifdef ENABLE_FOLD_CHECKING
12222 static void fold_checksum_tree (const_tree
, struct md5_ctx
*,
12223 hash_table
<nofree_ptr_hash
<const tree_node
> > *);
12224 static void fold_check_failed (const_tree
, const_tree
);
12225 void print_fold_checksum (const_tree
);
12227 /* When --enable-checking=fold, compute a digest of expr before
12228 and after actual fold call to see if fold did not accidentally
12229 change original expr. */
12235 struct md5_ctx ctx
;
12236 unsigned char checksum_before
[16], checksum_after
[16];
12237 hash_table
<nofree_ptr_hash
<const tree_node
> > ht (32);
12239 md5_init_ctx (&ctx
);
12240 fold_checksum_tree (expr
, &ctx
, &ht
);
12241 md5_finish_ctx (&ctx
, checksum_before
);
12244 ret
= fold_1 (expr
);
12246 md5_init_ctx (&ctx
);
12247 fold_checksum_tree (expr
, &ctx
, &ht
);
12248 md5_finish_ctx (&ctx
, checksum_after
);
12250 if (memcmp (checksum_before
, checksum_after
, 16))
12251 fold_check_failed (expr
, ret
);
12257 print_fold_checksum (const_tree expr
)
12259 struct md5_ctx ctx
;
12260 unsigned char checksum
[16], cnt
;
12261 hash_table
<nofree_ptr_hash
<const tree_node
> > ht (32);
12263 md5_init_ctx (&ctx
);
12264 fold_checksum_tree (expr
, &ctx
, &ht
);
12265 md5_finish_ctx (&ctx
, checksum
);
12266 for (cnt
= 0; cnt
< 16; ++cnt
)
12267 fprintf (stderr
, "%02x", checksum
[cnt
]);
12268 putc ('\n', stderr
);
12272 fold_check_failed (const_tree expr ATTRIBUTE_UNUSED
, const_tree ret ATTRIBUTE_UNUSED
)
12274 internal_error ("fold check: original tree changed by fold");
12278 fold_checksum_tree (const_tree expr
, struct md5_ctx
*ctx
,
12279 hash_table
<nofree_ptr_hash
<const tree_node
> > *ht
)
12281 const tree_node
**slot
;
12282 enum tree_code code
;
12283 union tree_node buf
;
12289 slot
= ht
->find_slot (expr
, INSERT
);
12293 code
= TREE_CODE (expr
);
12294 if (TREE_CODE_CLASS (code
) == tcc_declaration
12295 && HAS_DECL_ASSEMBLER_NAME_P (expr
))
12297 /* Allow DECL_ASSEMBLER_NAME and symtab_node to be modified. */
12298 memcpy ((char *) &buf
, expr
, tree_size (expr
));
12299 SET_DECL_ASSEMBLER_NAME ((tree
)&buf
, NULL
);
12300 buf
.decl_with_vis
.symtab_node
= NULL
;
12301 expr
= (tree
) &buf
;
12303 else if (TREE_CODE_CLASS (code
) == tcc_type
12304 && (TYPE_POINTER_TO (expr
)
12305 || TYPE_REFERENCE_TO (expr
)
12306 || TYPE_CACHED_VALUES_P (expr
)
12307 || TYPE_CONTAINS_PLACEHOLDER_INTERNAL (expr
)
12308 || TYPE_NEXT_VARIANT (expr
)))
12310 /* Allow these fields to be modified. */
12312 memcpy ((char *) &buf
, expr
, tree_size (expr
));
12313 expr
= tmp
= (tree
) &buf
;
12314 TYPE_CONTAINS_PLACEHOLDER_INTERNAL (tmp
) = 0;
12315 TYPE_POINTER_TO (tmp
) = NULL
;
12316 TYPE_REFERENCE_TO (tmp
) = NULL
;
12317 TYPE_NEXT_VARIANT (tmp
) = NULL
;
12318 if (TYPE_CACHED_VALUES_P (tmp
))
12320 TYPE_CACHED_VALUES_P (tmp
) = 0;
12321 TYPE_CACHED_VALUES (tmp
) = NULL
;
12324 md5_process_bytes (expr
, tree_size (expr
), ctx
);
12325 if (CODE_CONTAINS_STRUCT (code
, TS_TYPED
))
12326 fold_checksum_tree (TREE_TYPE (expr
), ctx
, ht
);
12327 if (TREE_CODE_CLASS (code
) != tcc_type
12328 && TREE_CODE_CLASS (code
) != tcc_declaration
12329 && code
!= TREE_LIST
12330 && code
!= SSA_NAME
12331 && CODE_CONTAINS_STRUCT (code
, TS_COMMON
))
12332 fold_checksum_tree (TREE_CHAIN (expr
), ctx
, ht
);
12333 switch (TREE_CODE_CLASS (code
))
12339 md5_process_bytes (TREE_STRING_POINTER (expr
),
12340 TREE_STRING_LENGTH (expr
), ctx
);
12343 fold_checksum_tree (TREE_REALPART (expr
), ctx
, ht
);
12344 fold_checksum_tree (TREE_IMAGPART (expr
), ctx
, ht
);
12347 for (i
= 0; i
< (int) VECTOR_CST_NELTS (expr
); ++i
)
12348 fold_checksum_tree (VECTOR_CST_ELT (expr
, i
), ctx
, ht
);
12354 case tcc_exceptional
:
12358 fold_checksum_tree (TREE_PURPOSE (expr
), ctx
, ht
);
12359 fold_checksum_tree (TREE_VALUE (expr
), ctx
, ht
);
12360 expr
= TREE_CHAIN (expr
);
12361 goto recursive_label
;
12364 for (i
= 0; i
< TREE_VEC_LENGTH (expr
); ++i
)
12365 fold_checksum_tree (TREE_VEC_ELT (expr
, i
), ctx
, ht
);
12371 case tcc_expression
:
12372 case tcc_reference
:
12373 case tcc_comparison
:
12376 case tcc_statement
:
12378 len
= TREE_OPERAND_LENGTH (expr
);
12379 for (i
= 0; i
< len
; ++i
)
12380 fold_checksum_tree (TREE_OPERAND (expr
, i
), ctx
, ht
);
12382 case tcc_declaration
:
12383 fold_checksum_tree (DECL_NAME (expr
), ctx
, ht
);
12384 fold_checksum_tree (DECL_CONTEXT (expr
), ctx
, ht
);
12385 if (CODE_CONTAINS_STRUCT (TREE_CODE (expr
), TS_DECL_COMMON
))
12387 fold_checksum_tree (DECL_SIZE (expr
), ctx
, ht
);
12388 fold_checksum_tree (DECL_SIZE_UNIT (expr
), ctx
, ht
);
12389 fold_checksum_tree (DECL_INITIAL (expr
), ctx
, ht
);
12390 fold_checksum_tree (DECL_ABSTRACT_ORIGIN (expr
), ctx
, ht
);
12391 fold_checksum_tree (DECL_ATTRIBUTES (expr
), ctx
, ht
);
12394 if (CODE_CONTAINS_STRUCT (TREE_CODE (expr
), TS_DECL_NON_COMMON
))
12396 if (TREE_CODE (expr
) == FUNCTION_DECL
)
12398 fold_checksum_tree (DECL_VINDEX (expr
), ctx
, ht
);
12399 fold_checksum_tree (DECL_ARGUMENTS (expr
), ctx
, ht
);
12401 fold_checksum_tree (DECL_RESULT_FLD (expr
), ctx
, ht
);
12405 if (TREE_CODE (expr
) == ENUMERAL_TYPE
)
12406 fold_checksum_tree (TYPE_VALUES (expr
), ctx
, ht
);
12407 fold_checksum_tree (TYPE_SIZE (expr
), ctx
, ht
);
12408 fold_checksum_tree (TYPE_SIZE_UNIT (expr
), ctx
, ht
);
12409 fold_checksum_tree (TYPE_ATTRIBUTES (expr
), ctx
, ht
);
12410 fold_checksum_tree (TYPE_NAME (expr
), ctx
, ht
);
12411 if (INTEGRAL_TYPE_P (expr
)
12412 || SCALAR_FLOAT_TYPE_P (expr
))
12414 fold_checksum_tree (TYPE_MIN_VALUE (expr
), ctx
, ht
);
12415 fold_checksum_tree (TYPE_MAX_VALUE (expr
), ctx
, ht
);
12417 fold_checksum_tree (TYPE_MAIN_VARIANT (expr
), ctx
, ht
);
12418 if (TREE_CODE (expr
) == RECORD_TYPE
12419 || TREE_CODE (expr
) == UNION_TYPE
12420 || TREE_CODE (expr
) == QUAL_UNION_TYPE
)
12421 fold_checksum_tree (TYPE_BINFO (expr
), ctx
, ht
);
12422 fold_checksum_tree (TYPE_CONTEXT (expr
), ctx
, ht
);
12429 /* Helper function for outputting the checksum of a tree T. When
12430 debugging with gdb, you can "define mynext" to be "next" followed
12431 by "call debug_fold_checksum (op0)", then just trace down till the
12434 DEBUG_FUNCTION
void
12435 debug_fold_checksum (const_tree t
)
12438 unsigned char checksum
[16];
12439 struct md5_ctx ctx
;
12440 hash_table
<nofree_ptr_hash
<const tree_node
> > ht (32);
12442 md5_init_ctx (&ctx
);
12443 fold_checksum_tree (t
, &ctx
, &ht
);
12444 md5_finish_ctx (&ctx
, checksum
);
12447 for (i
= 0; i
< 16; i
++)
12448 fprintf (stderr
, "%d ", checksum
[i
]);
12450 fprintf (stderr
, "\n");
12455 /* Fold a unary tree expression with code CODE of type TYPE with an
12456 operand OP0. LOC is the location of the resulting expression.
12457 Return a folded expression if successful. Otherwise, return a tree
12458 expression with code CODE of type TYPE with an operand OP0. */
12461 fold_build1_stat_loc (location_t loc
,
12462 enum tree_code code
, tree type
, tree op0 MEM_STAT_DECL
)
12465 #ifdef ENABLE_FOLD_CHECKING
12466 unsigned char checksum_before
[16], checksum_after
[16];
12467 struct md5_ctx ctx
;
12468 hash_table
<nofree_ptr_hash
<const tree_node
> > ht (32);
12470 md5_init_ctx (&ctx
);
12471 fold_checksum_tree (op0
, &ctx
, &ht
);
12472 md5_finish_ctx (&ctx
, checksum_before
);
12476 tem
= fold_unary_loc (loc
, code
, type
, op0
);
12478 tem
= build1_stat_loc (loc
, code
, type
, op0 PASS_MEM_STAT
);
12480 #ifdef ENABLE_FOLD_CHECKING
12481 md5_init_ctx (&ctx
);
12482 fold_checksum_tree (op0
, &ctx
, &ht
);
12483 md5_finish_ctx (&ctx
, checksum_after
);
12485 if (memcmp (checksum_before
, checksum_after
, 16))
12486 fold_check_failed (op0
, tem
);
12491 /* Fold a binary tree expression with code CODE of type TYPE with
12492 operands OP0 and OP1. LOC is the location of the resulting
12493 expression. Return a folded expression if successful. Otherwise,
12494 return a tree expression with code CODE of type TYPE with operands
12498 fold_build2_stat_loc (location_t loc
,
12499 enum tree_code code
, tree type
, tree op0
, tree op1
12503 #ifdef ENABLE_FOLD_CHECKING
12504 unsigned char checksum_before_op0
[16],
12505 checksum_before_op1
[16],
12506 checksum_after_op0
[16],
12507 checksum_after_op1
[16];
12508 struct md5_ctx ctx
;
12509 hash_table
<nofree_ptr_hash
<const tree_node
> > ht (32);
12511 md5_init_ctx (&ctx
);
12512 fold_checksum_tree (op0
, &ctx
, &ht
);
12513 md5_finish_ctx (&ctx
, checksum_before_op0
);
12516 md5_init_ctx (&ctx
);
12517 fold_checksum_tree (op1
, &ctx
, &ht
);
12518 md5_finish_ctx (&ctx
, checksum_before_op1
);
12522 tem
= fold_binary_loc (loc
, code
, type
, op0
, op1
);
12524 tem
= build2_stat_loc (loc
, code
, type
, op0
, op1 PASS_MEM_STAT
);
12526 #ifdef ENABLE_FOLD_CHECKING
12527 md5_init_ctx (&ctx
);
12528 fold_checksum_tree (op0
, &ctx
, &ht
);
12529 md5_finish_ctx (&ctx
, checksum_after_op0
);
12532 if (memcmp (checksum_before_op0
, checksum_after_op0
, 16))
12533 fold_check_failed (op0
, tem
);
12535 md5_init_ctx (&ctx
);
12536 fold_checksum_tree (op1
, &ctx
, &ht
);
12537 md5_finish_ctx (&ctx
, checksum_after_op1
);
12539 if (memcmp (checksum_before_op1
, checksum_after_op1
, 16))
12540 fold_check_failed (op1
, tem
);
12545 /* Fold a ternary tree expression with code CODE of type TYPE with
12546 operands OP0, OP1, and OP2. Return a folded expression if
12547 successful. Otherwise, return a tree expression with code CODE of
12548 type TYPE with operands OP0, OP1, and OP2. */
12551 fold_build3_stat_loc (location_t loc
, enum tree_code code
, tree type
,
12552 tree op0
, tree op1
, tree op2 MEM_STAT_DECL
)
12555 #ifdef ENABLE_FOLD_CHECKING
12556 unsigned char checksum_before_op0
[16],
12557 checksum_before_op1
[16],
12558 checksum_before_op2
[16],
12559 checksum_after_op0
[16],
12560 checksum_after_op1
[16],
12561 checksum_after_op2
[16];
12562 struct md5_ctx ctx
;
12563 hash_table
<nofree_ptr_hash
<const tree_node
> > ht (32);
12565 md5_init_ctx (&ctx
);
12566 fold_checksum_tree (op0
, &ctx
, &ht
);
12567 md5_finish_ctx (&ctx
, checksum_before_op0
);
12570 md5_init_ctx (&ctx
);
12571 fold_checksum_tree (op1
, &ctx
, &ht
);
12572 md5_finish_ctx (&ctx
, checksum_before_op1
);
12575 md5_init_ctx (&ctx
);
12576 fold_checksum_tree (op2
, &ctx
, &ht
);
12577 md5_finish_ctx (&ctx
, checksum_before_op2
);
12581 gcc_assert (TREE_CODE_CLASS (code
) != tcc_vl_exp
);
12582 tem
= fold_ternary_loc (loc
, code
, type
, op0
, op1
, op2
);
12584 tem
= build3_stat_loc (loc
, code
, type
, op0
, op1
, op2 PASS_MEM_STAT
);
12586 #ifdef ENABLE_FOLD_CHECKING
12587 md5_init_ctx (&ctx
);
12588 fold_checksum_tree (op0
, &ctx
, &ht
);
12589 md5_finish_ctx (&ctx
, checksum_after_op0
);
12592 if (memcmp (checksum_before_op0
, checksum_after_op0
, 16))
12593 fold_check_failed (op0
, tem
);
12595 md5_init_ctx (&ctx
);
12596 fold_checksum_tree (op1
, &ctx
, &ht
);
12597 md5_finish_ctx (&ctx
, checksum_after_op1
);
12600 if (memcmp (checksum_before_op1
, checksum_after_op1
, 16))
12601 fold_check_failed (op1
, tem
);
12603 md5_init_ctx (&ctx
);
12604 fold_checksum_tree (op2
, &ctx
, &ht
);
12605 md5_finish_ctx (&ctx
, checksum_after_op2
);
12607 if (memcmp (checksum_before_op2
, checksum_after_op2
, 16))
12608 fold_check_failed (op2
, tem
);
12613 /* Fold a CALL_EXPR expression of type TYPE with operands FN and NARGS
12614 arguments in ARGARRAY, and a null static chain.
12615 Return a folded expression if successful. Otherwise, return a CALL_EXPR
12616 of type TYPE from the given operands as constructed by build_call_array. */
12619 fold_build_call_array_loc (location_t loc
, tree type
, tree fn
,
12620 int nargs
, tree
*argarray
)
12623 #ifdef ENABLE_FOLD_CHECKING
12624 unsigned char checksum_before_fn
[16],
12625 checksum_before_arglist
[16],
12626 checksum_after_fn
[16],
12627 checksum_after_arglist
[16];
12628 struct md5_ctx ctx
;
12629 hash_table
<nofree_ptr_hash
<const tree_node
> > ht (32);
12632 md5_init_ctx (&ctx
);
12633 fold_checksum_tree (fn
, &ctx
, &ht
);
12634 md5_finish_ctx (&ctx
, checksum_before_fn
);
12637 md5_init_ctx (&ctx
);
12638 for (i
= 0; i
< nargs
; i
++)
12639 fold_checksum_tree (argarray
[i
], &ctx
, &ht
);
12640 md5_finish_ctx (&ctx
, checksum_before_arglist
);
12644 tem
= fold_builtin_call_array (loc
, type
, fn
, nargs
, argarray
);
12646 tem
= build_call_array_loc (loc
, type
, fn
, nargs
, argarray
);
12648 #ifdef ENABLE_FOLD_CHECKING
12649 md5_init_ctx (&ctx
);
12650 fold_checksum_tree (fn
, &ctx
, &ht
);
12651 md5_finish_ctx (&ctx
, checksum_after_fn
);
12654 if (memcmp (checksum_before_fn
, checksum_after_fn
, 16))
12655 fold_check_failed (fn
, tem
);
12657 md5_init_ctx (&ctx
);
12658 for (i
= 0; i
< nargs
; i
++)
12659 fold_checksum_tree (argarray
[i
], &ctx
, &ht
);
12660 md5_finish_ctx (&ctx
, checksum_after_arglist
);
12662 if (memcmp (checksum_before_arglist
, checksum_after_arglist
, 16))
12663 fold_check_failed (NULL_TREE
, tem
);
12668 /* Perform constant folding and related simplification of initializer
12669 expression EXPR. These behave identically to "fold_buildN" but ignore
12670 potential run-time traps and exceptions that fold must preserve. */
12672 #define START_FOLD_INIT \
12673 int saved_signaling_nans = flag_signaling_nans;\
12674 int saved_trapping_math = flag_trapping_math;\
12675 int saved_rounding_math = flag_rounding_math;\
12676 int saved_trapv = flag_trapv;\
12677 int saved_folding_initializer = folding_initializer;\
12678 flag_signaling_nans = 0;\
12679 flag_trapping_math = 0;\
12680 flag_rounding_math = 0;\
12682 folding_initializer = 1;
12684 #define END_FOLD_INIT \
12685 flag_signaling_nans = saved_signaling_nans;\
12686 flag_trapping_math = saved_trapping_math;\
12687 flag_rounding_math = saved_rounding_math;\
12688 flag_trapv = saved_trapv;\
12689 folding_initializer = saved_folding_initializer;
12692 fold_build1_initializer_loc (location_t loc
, enum tree_code code
,
12693 tree type
, tree op
)
12698 result
= fold_build1_loc (loc
, code
, type
, op
);
12705 fold_build2_initializer_loc (location_t loc
, enum tree_code code
,
12706 tree type
, tree op0
, tree op1
)
12711 result
= fold_build2_loc (loc
, code
, type
, op0
, op1
);
12718 fold_build_call_array_initializer_loc (location_t loc
, tree type
, tree fn
,
12719 int nargs
, tree
*argarray
)
12724 result
= fold_build_call_array_loc (loc
, type
, fn
, nargs
, argarray
);
12730 #undef START_FOLD_INIT
12731 #undef END_FOLD_INIT
12733 /* Determine if first argument is a multiple of second argument. Return 0 if
12734 it is not, or we cannot easily determined it to be.
12736 An example of the sort of thing we care about (at this point; this routine
12737 could surely be made more general, and expanded to do what the *_DIV_EXPR's
12738 fold cases do now) is discovering that
12740 SAVE_EXPR (I) * SAVE_EXPR (J * 8)
12746 when we know that the two SAVE_EXPR (J * 8) nodes are the same node.
12748 This code also handles discovering that
12750 SAVE_EXPR (I) * SAVE_EXPR (J * 8)
12752 is a multiple of 8 so we don't have to worry about dealing with a
12753 possible remainder.
12755 Note that we *look* inside a SAVE_EXPR only to determine how it was
12756 calculated; it is not safe for fold to do much of anything else with the
12757 internals of a SAVE_EXPR, since it cannot know when it will be evaluated
12758 at run time. For example, the latter example above *cannot* be implemented
12759 as SAVE_EXPR (I) * J or any variant thereof, since the value of J at
12760 evaluation time of the original SAVE_EXPR is not necessarily the same at
12761 the time the new expression is evaluated. The only optimization of this
12762 sort that would be valid is changing
12764 SAVE_EXPR (I) * SAVE_EXPR (SAVE_EXPR (J) * 8)
12768 SAVE_EXPR (I) * SAVE_EXPR (J)
12770 (where the same SAVE_EXPR (J) is used in the original and the
12771 transformed version). */
12774 multiple_of_p (tree type
, const_tree top
, const_tree bottom
)
12776 if (operand_equal_p (top
, bottom
, 0))
12779 if (TREE_CODE (type
) != INTEGER_TYPE
)
12782 switch (TREE_CODE (top
))
12785 /* Bitwise and provides a power of two multiple. If the mask is
12786 a multiple of BOTTOM then TOP is a multiple of BOTTOM. */
12787 if (!integer_pow2p (bottom
))
12792 return (multiple_of_p (type
, TREE_OPERAND (top
, 0), bottom
)
12793 || multiple_of_p (type
, TREE_OPERAND (top
, 1), bottom
));
12797 return (multiple_of_p (type
, TREE_OPERAND (top
, 0), bottom
)
12798 && multiple_of_p (type
, TREE_OPERAND (top
, 1), bottom
));
12801 if (TREE_CODE (TREE_OPERAND (top
, 1)) == INTEGER_CST
)
12805 op1
= TREE_OPERAND (top
, 1);
12806 /* const_binop may not detect overflow correctly,
12807 so check for it explicitly here. */
12808 if (wi::gtu_p (TYPE_PRECISION (TREE_TYPE (size_one_node
)), op1
)
12809 && 0 != (t1
= fold_convert (type
,
12810 const_binop (LSHIFT_EXPR
,
12813 && !TREE_OVERFLOW (t1
))
12814 return multiple_of_p (type
, t1
, bottom
);
12819 /* Can't handle conversions from non-integral or wider integral type. */
12820 if ((TREE_CODE (TREE_TYPE (TREE_OPERAND (top
, 0))) != INTEGER_TYPE
)
12821 || (TYPE_PRECISION (type
)
12822 < TYPE_PRECISION (TREE_TYPE (TREE_OPERAND (top
, 0)))))
12825 /* .. fall through ... */
12828 return multiple_of_p (type
, TREE_OPERAND (top
, 0), bottom
);
12831 return (multiple_of_p (type
, TREE_OPERAND (top
, 1), bottom
)
12832 && multiple_of_p (type
, TREE_OPERAND (top
, 2), bottom
));
12835 if (TREE_CODE (bottom
) != INTEGER_CST
12836 || integer_zerop (bottom
)
12837 || (TYPE_UNSIGNED (type
)
12838 && (tree_int_cst_sgn (top
) < 0
12839 || tree_int_cst_sgn (bottom
) < 0)))
12841 return wi::multiple_of_p (wi::to_widest (top
), wi::to_widest (bottom
),
12849 #define tree_expr_nonnegative_warnv_p(X, Y) \
12850 _Pragma ("GCC error \"Use RECURSE for recursive calls\"") 0
12852 #define RECURSE(X) \
12853 ((tree_expr_nonnegative_warnv_p) (X, strict_overflow_p, depth + 1))
12855 /* Return true if CODE or TYPE is known to be non-negative. */
12858 tree_simple_nonnegative_warnv_p (enum tree_code code
, tree type
)
12860 if ((TYPE_PRECISION (type
) != 1 || TYPE_UNSIGNED (type
))
12861 && truth_value_p (code
))
12862 /* Truth values evaluate to 0 or 1, which is nonnegative unless we
12863 have a signed:1 type (where the value is -1 and 0). */
12868 /* Return true if (CODE OP0) is known to be non-negative. If the return
12869 value is based on the assumption that signed overflow is undefined,
12870 set *STRICT_OVERFLOW_P to true; otherwise, don't change
12871 *STRICT_OVERFLOW_P. DEPTH is the current nesting depth of the query. */
12874 tree_unary_nonnegative_warnv_p (enum tree_code code
, tree type
, tree op0
,
12875 bool *strict_overflow_p
, int depth
)
12877 if (TYPE_UNSIGNED (type
))
12883 /* We can't return 1 if flag_wrapv is set because
12884 ABS_EXPR<INT_MIN> = INT_MIN. */
12885 if (!ANY_INTEGRAL_TYPE_P (type
))
12887 if (TYPE_OVERFLOW_UNDEFINED (type
))
12889 *strict_overflow_p
= true;
12894 case NON_LVALUE_EXPR
:
12896 case FIX_TRUNC_EXPR
:
12897 return RECURSE (op0
);
12901 tree inner_type
= TREE_TYPE (op0
);
12902 tree outer_type
= type
;
12904 if (TREE_CODE (outer_type
) == REAL_TYPE
)
12906 if (TREE_CODE (inner_type
) == REAL_TYPE
)
12907 return RECURSE (op0
);
12908 if (INTEGRAL_TYPE_P (inner_type
))
12910 if (TYPE_UNSIGNED (inner_type
))
12912 return RECURSE (op0
);
12915 else if (INTEGRAL_TYPE_P (outer_type
))
12917 if (TREE_CODE (inner_type
) == REAL_TYPE
)
12918 return RECURSE (op0
);
12919 if (INTEGRAL_TYPE_P (inner_type
))
12920 return TYPE_PRECISION (inner_type
) < TYPE_PRECISION (outer_type
)
12921 && TYPE_UNSIGNED (inner_type
);
12927 return tree_simple_nonnegative_warnv_p (code
, type
);
12930 /* We don't know sign of `t', so be conservative and return false. */
12934 /* Return true if (CODE OP0 OP1) is known to be non-negative. If the return
12935 value is based on the assumption that signed overflow is undefined,
12936 set *STRICT_OVERFLOW_P to true; otherwise, don't change
12937 *STRICT_OVERFLOW_P. DEPTH is the current nesting depth of the query. */
12940 tree_binary_nonnegative_warnv_p (enum tree_code code
, tree type
, tree op0
,
12941 tree op1
, bool *strict_overflow_p
,
12944 if (TYPE_UNSIGNED (type
))
12949 case POINTER_PLUS_EXPR
:
12951 if (FLOAT_TYPE_P (type
))
12952 return RECURSE (op0
) && RECURSE (op1
);
12954 /* zero_extend(x) + zero_extend(y) is non-negative if x and y are
12955 both unsigned and at least 2 bits shorter than the result. */
12956 if (TREE_CODE (type
) == INTEGER_TYPE
12957 && TREE_CODE (op0
) == NOP_EXPR
12958 && TREE_CODE (op1
) == NOP_EXPR
)
12960 tree inner1
= TREE_TYPE (TREE_OPERAND (op0
, 0));
12961 tree inner2
= TREE_TYPE (TREE_OPERAND (op1
, 0));
12962 if (TREE_CODE (inner1
) == INTEGER_TYPE
&& TYPE_UNSIGNED (inner1
)
12963 && TREE_CODE (inner2
) == INTEGER_TYPE
&& TYPE_UNSIGNED (inner2
))
12965 unsigned int prec
= MAX (TYPE_PRECISION (inner1
),
12966 TYPE_PRECISION (inner2
)) + 1;
12967 return prec
< TYPE_PRECISION (type
);
12973 if (FLOAT_TYPE_P (type
) || TYPE_OVERFLOW_UNDEFINED (type
))
12975 /* x * x is always non-negative for floating point x
12976 or without overflow. */
12977 if (operand_equal_p (op0
, op1
, 0)
12978 || (RECURSE (op0
) && RECURSE (op1
)))
12980 if (ANY_INTEGRAL_TYPE_P (type
)
12981 && TYPE_OVERFLOW_UNDEFINED (type
))
12982 *strict_overflow_p
= true;
12987 /* zero_extend(x) * zero_extend(y) is non-negative if x and y are
12988 both unsigned and their total bits is shorter than the result. */
12989 if (TREE_CODE (type
) == INTEGER_TYPE
12990 && (TREE_CODE (op0
) == NOP_EXPR
|| TREE_CODE (op0
) == INTEGER_CST
)
12991 && (TREE_CODE (op1
) == NOP_EXPR
|| TREE_CODE (op1
) == INTEGER_CST
))
12993 tree inner0
= (TREE_CODE (op0
) == NOP_EXPR
)
12994 ? TREE_TYPE (TREE_OPERAND (op0
, 0))
12996 tree inner1
= (TREE_CODE (op1
) == NOP_EXPR
)
12997 ? TREE_TYPE (TREE_OPERAND (op1
, 0))
13000 bool unsigned0
= TYPE_UNSIGNED (inner0
);
13001 bool unsigned1
= TYPE_UNSIGNED (inner1
);
13003 if (TREE_CODE (op0
) == INTEGER_CST
)
13004 unsigned0
= unsigned0
|| tree_int_cst_sgn (op0
) >= 0;
13006 if (TREE_CODE (op1
) == INTEGER_CST
)
13007 unsigned1
= unsigned1
|| tree_int_cst_sgn (op1
) >= 0;
13009 if (TREE_CODE (inner0
) == INTEGER_TYPE
&& unsigned0
13010 && TREE_CODE (inner1
) == INTEGER_TYPE
&& unsigned1
)
13012 unsigned int precision0
= (TREE_CODE (op0
) == INTEGER_CST
)
13013 ? tree_int_cst_min_precision (op0
, UNSIGNED
)
13014 : TYPE_PRECISION (inner0
);
13016 unsigned int precision1
= (TREE_CODE (op1
) == INTEGER_CST
)
13017 ? tree_int_cst_min_precision (op1
, UNSIGNED
)
13018 : TYPE_PRECISION (inner1
);
13020 return precision0
+ precision1
< TYPE_PRECISION (type
);
13027 return RECURSE (op0
) || RECURSE (op1
);
13033 case TRUNC_DIV_EXPR
:
13034 case CEIL_DIV_EXPR
:
13035 case FLOOR_DIV_EXPR
:
13036 case ROUND_DIV_EXPR
:
13037 return RECURSE (op0
) && RECURSE (op1
);
13039 case TRUNC_MOD_EXPR
:
13040 return RECURSE (op0
);
13042 case FLOOR_MOD_EXPR
:
13043 return RECURSE (op1
);
13045 case CEIL_MOD_EXPR
:
13046 case ROUND_MOD_EXPR
:
13048 return tree_simple_nonnegative_warnv_p (code
, type
);
13051 /* We don't know sign of `t', so be conservative and return false. */
13055 /* Return true if T is known to be non-negative. If the return
13056 value is based on the assumption that signed overflow is undefined,
13057 set *STRICT_OVERFLOW_P to true; otherwise, don't change
13058 *STRICT_OVERFLOW_P. DEPTH is the current nesting depth of the query. */
13061 tree_single_nonnegative_warnv_p (tree t
, bool *strict_overflow_p
, int depth
)
13063 if (TYPE_UNSIGNED (TREE_TYPE (t
)))
13066 switch (TREE_CODE (t
))
13069 return tree_int_cst_sgn (t
) >= 0;
13072 return ! REAL_VALUE_NEGATIVE (TREE_REAL_CST (t
));
13075 return ! FIXED_VALUE_NEGATIVE (TREE_FIXED_CST (t
));
13078 return RECURSE (TREE_OPERAND (t
, 1)) && RECURSE (TREE_OPERAND (t
, 2));
13081 /* Limit the depth of recursion to avoid quadratic behavior.
13082 This is expected to catch almost all occurrences in practice.
13083 If this code misses important cases that unbounded recursion
13084 would not, passes that need this information could be revised
13085 to provide it through dataflow propagation. */
13086 return (!name_registered_for_update_p (t
)
13087 && depth
< PARAM_VALUE (PARAM_MAX_SSA_NAME_QUERY_DEPTH
)
13088 && gimple_stmt_nonnegative_warnv_p (SSA_NAME_DEF_STMT (t
),
13089 strict_overflow_p
, depth
));
13092 return tree_simple_nonnegative_warnv_p (TREE_CODE (t
), TREE_TYPE (t
));
13096 /* Return true if T is known to be non-negative. If the return
13097 value is based on the assumption that signed overflow is undefined,
13098 set *STRICT_OVERFLOW_P to true; otherwise, don't change
13099 *STRICT_OVERFLOW_P. DEPTH is the current nesting depth of the query. */
13102 tree_call_nonnegative_warnv_p (tree type
, combined_fn fn
, tree arg0
, tree arg1
,
13103 bool *strict_overflow_p
, int depth
)
13124 case CFN_BUILT_IN_BSWAP32
:
13125 case CFN_BUILT_IN_BSWAP64
:
13130 /* sqrt(-0.0) is -0.0. */
13131 if (!HONOR_SIGNED_ZEROS (element_mode (type
)))
13133 return RECURSE (arg0
);
13159 CASE_CFN_NEARBYINT
:
13166 CASE_CFN_SIGNIFICAND
:
13170 /* True if the 1st argument is nonnegative. */
13171 return RECURSE (arg0
);
13174 /* True if the 1st OR 2nd arguments are nonnegative. */
13175 return RECURSE (arg0
) || RECURSE (arg1
);
13178 /* True if the 1st AND 2nd arguments are nonnegative. */
13179 return RECURSE (arg0
) && RECURSE (arg1
);
13182 /* True if the 2nd argument is nonnegative. */
13183 return RECURSE (arg1
);
13186 /* True if the 1st argument is nonnegative or the second
13187 argument is an even integer. */
13188 if (TREE_CODE (arg1
) == INTEGER_CST
13189 && (TREE_INT_CST_LOW (arg1
) & 1) == 0)
13191 return RECURSE (arg0
);
13194 /* True if the 1st argument is nonnegative or the second
13195 argument is an even integer valued real. */
13196 if (TREE_CODE (arg1
) == REAL_CST
)
13201 c
= TREE_REAL_CST (arg1
);
13202 n
= real_to_integer (&c
);
13205 REAL_VALUE_TYPE cint
;
13206 real_from_integer (&cint
, VOIDmode
, n
, SIGNED
);
13207 if (real_identical (&c
, &cint
))
13211 return RECURSE (arg0
);
13216 return tree_simple_nonnegative_warnv_p (CALL_EXPR
, type
);
13219 /* Return true if T is known to be non-negative. If the return
13220 value is based on the assumption that signed overflow is undefined,
13221 set *STRICT_OVERFLOW_P to true; otherwise, don't change
13222 *STRICT_OVERFLOW_P. DEPTH is the current nesting depth of the query. */
13225 tree_invalid_nonnegative_warnv_p (tree t
, bool *strict_overflow_p
, int depth
)
13227 enum tree_code code
= TREE_CODE (t
);
13228 if (TYPE_UNSIGNED (TREE_TYPE (t
)))
13235 tree temp
= TARGET_EXPR_SLOT (t
);
13236 t
= TARGET_EXPR_INITIAL (t
);
13238 /* If the initializer is non-void, then it's a normal expression
13239 that will be assigned to the slot. */
13240 if (!VOID_TYPE_P (t
))
13241 return RECURSE (t
);
13243 /* Otherwise, the initializer sets the slot in some way. One common
13244 way is an assignment statement at the end of the initializer. */
13247 if (TREE_CODE (t
) == BIND_EXPR
)
13248 t
= expr_last (BIND_EXPR_BODY (t
));
13249 else if (TREE_CODE (t
) == TRY_FINALLY_EXPR
13250 || TREE_CODE (t
) == TRY_CATCH_EXPR
)
13251 t
= expr_last (TREE_OPERAND (t
, 0));
13252 else if (TREE_CODE (t
) == STATEMENT_LIST
)
13257 if (TREE_CODE (t
) == MODIFY_EXPR
13258 && TREE_OPERAND (t
, 0) == temp
)
13259 return RECURSE (TREE_OPERAND (t
, 1));
13266 tree arg0
= call_expr_nargs (t
) > 0 ? CALL_EXPR_ARG (t
, 0) : NULL_TREE
;
13267 tree arg1
= call_expr_nargs (t
) > 1 ? CALL_EXPR_ARG (t
, 1) : NULL_TREE
;
13269 return tree_call_nonnegative_warnv_p (TREE_TYPE (t
),
13270 get_call_combined_fn (t
),
13273 strict_overflow_p
, depth
);
13275 case COMPOUND_EXPR
:
13277 return RECURSE (TREE_OPERAND (t
, 1));
13280 return RECURSE (expr_last (TREE_OPERAND (t
, 1)));
13283 return RECURSE (TREE_OPERAND (t
, 0));
13286 return tree_simple_nonnegative_warnv_p (TREE_CODE (t
), TREE_TYPE (t
));
13291 #undef tree_expr_nonnegative_warnv_p
13293 /* Return true if T is known to be non-negative. If the return
13294 value is based on the assumption that signed overflow is undefined,
13295 set *STRICT_OVERFLOW_P to true; otherwise, don't change
13296 *STRICT_OVERFLOW_P. DEPTH is the current nesting depth of the query. */
13299 tree_expr_nonnegative_warnv_p (tree t
, bool *strict_overflow_p
, int depth
)
13301 enum tree_code code
;
13302 if (t
== error_mark_node
)
13305 code
= TREE_CODE (t
);
13306 switch (TREE_CODE_CLASS (code
))
13309 case tcc_comparison
:
13310 return tree_binary_nonnegative_warnv_p (TREE_CODE (t
),
13312 TREE_OPERAND (t
, 0),
13313 TREE_OPERAND (t
, 1),
13314 strict_overflow_p
, depth
);
13317 return tree_unary_nonnegative_warnv_p (TREE_CODE (t
),
13319 TREE_OPERAND (t
, 0),
13320 strict_overflow_p
, depth
);
13323 case tcc_declaration
:
13324 case tcc_reference
:
13325 return tree_single_nonnegative_warnv_p (t
, strict_overflow_p
, depth
);
13333 case TRUTH_AND_EXPR
:
13334 case TRUTH_OR_EXPR
:
13335 case TRUTH_XOR_EXPR
:
13336 return tree_binary_nonnegative_warnv_p (TREE_CODE (t
),
13338 TREE_OPERAND (t
, 0),
13339 TREE_OPERAND (t
, 1),
13340 strict_overflow_p
, depth
);
13341 case TRUTH_NOT_EXPR
:
13342 return tree_unary_nonnegative_warnv_p (TREE_CODE (t
),
13344 TREE_OPERAND (t
, 0),
13345 strict_overflow_p
, depth
);
13352 case WITH_SIZE_EXPR
:
13354 return tree_single_nonnegative_warnv_p (t
, strict_overflow_p
, depth
);
13357 return tree_invalid_nonnegative_warnv_p (t
, strict_overflow_p
, depth
);
13361 /* Return true if `t' is known to be non-negative. Handle warnings
13362 about undefined signed overflow. */
13365 tree_expr_nonnegative_p (tree t
)
13367 bool ret
, strict_overflow_p
;
13369 strict_overflow_p
= false;
13370 ret
= tree_expr_nonnegative_warnv_p (t
, &strict_overflow_p
);
13371 if (strict_overflow_p
)
13372 fold_overflow_warning (("assuming signed overflow does not occur when "
13373 "determining that expression is always "
13375 WARN_STRICT_OVERFLOW_MISC
);
13380 /* Return true when (CODE OP0) is an address and is known to be nonzero.
13381 For floating point we further ensure that T is not denormal.
13382 Similar logic is present in nonzero_address in rtlanal.h.
13384 If the return value is based on the assumption that signed overflow
13385 is undefined, set *STRICT_OVERFLOW_P to true; otherwise, don't
13386 change *STRICT_OVERFLOW_P. */
13389 tree_unary_nonzero_warnv_p (enum tree_code code
, tree type
, tree op0
,
13390 bool *strict_overflow_p
)
13395 return tree_expr_nonzero_warnv_p (op0
,
13396 strict_overflow_p
);
13400 tree inner_type
= TREE_TYPE (op0
);
13401 tree outer_type
= type
;
13403 return (TYPE_PRECISION (outer_type
) >= TYPE_PRECISION (inner_type
)
13404 && tree_expr_nonzero_warnv_p (op0
,
13405 strict_overflow_p
));
13409 case NON_LVALUE_EXPR
:
13410 return tree_expr_nonzero_warnv_p (op0
,
13411 strict_overflow_p
);
13420 /* Return true when (CODE OP0 OP1) is an address and is known to be nonzero.
13421 For floating point we further ensure that T is not denormal.
13422 Similar logic is present in nonzero_address in rtlanal.h.
13424 If the return value is based on the assumption that signed overflow
13425 is undefined, set *STRICT_OVERFLOW_P to true; otherwise, don't
13426 change *STRICT_OVERFLOW_P. */
13429 tree_binary_nonzero_warnv_p (enum tree_code code
,
13432 tree op1
, bool *strict_overflow_p
)
13434 bool sub_strict_overflow_p
;
13437 case POINTER_PLUS_EXPR
:
13439 if (ANY_INTEGRAL_TYPE_P (type
) && TYPE_OVERFLOW_UNDEFINED (type
))
13441 /* With the presence of negative values it is hard
13442 to say something. */
13443 sub_strict_overflow_p
= false;
13444 if (!tree_expr_nonnegative_warnv_p (op0
,
13445 &sub_strict_overflow_p
)
13446 || !tree_expr_nonnegative_warnv_p (op1
,
13447 &sub_strict_overflow_p
))
13449 /* One of operands must be positive and the other non-negative. */
13450 /* We don't set *STRICT_OVERFLOW_P here: even if this value
13451 overflows, on a twos-complement machine the sum of two
13452 nonnegative numbers can never be zero. */
13453 return (tree_expr_nonzero_warnv_p (op0
,
13455 || tree_expr_nonzero_warnv_p (op1
,
13456 strict_overflow_p
));
13461 if (TYPE_OVERFLOW_UNDEFINED (type
))
13463 if (tree_expr_nonzero_warnv_p (op0
,
13465 && tree_expr_nonzero_warnv_p (op1
,
13466 strict_overflow_p
))
13468 *strict_overflow_p
= true;
13475 sub_strict_overflow_p
= false;
13476 if (tree_expr_nonzero_warnv_p (op0
,
13477 &sub_strict_overflow_p
)
13478 && tree_expr_nonzero_warnv_p (op1
,
13479 &sub_strict_overflow_p
))
13481 if (sub_strict_overflow_p
)
13482 *strict_overflow_p
= true;
13487 sub_strict_overflow_p
= false;
13488 if (tree_expr_nonzero_warnv_p (op0
,
13489 &sub_strict_overflow_p
))
13491 if (sub_strict_overflow_p
)
13492 *strict_overflow_p
= true;
13494 /* When both operands are nonzero, then MAX must be too. */
13495 if (tree_expr_nonzero_warnv_p (op1
,
13496 strict_overflow_p
))
13499 /* MAX where operand 0 is positive is positive. */
13500 return tree_expr_nonnegative_warnv_p (op0
,
13501 strict_overflow_p
);
13503 /* MAX where operand 1 is positive is positive. */
13504 else if (tree_expr_nonzero_warnv_p (op1
,
13505 &sub_strict_overflow_p
)
13506 && tree_expr_nonnegative_warnv_p (op1
,
13507 &sub_strict_overflow_p
))
13509 if (sub_strict_overflow_p
)
13510 *strict_overflow_p
= true;
13516 return (tree_expr_nonzero_warnv_p (op1
,
13518 || tree_expr_nonzero_warnv_p (op0
,
13519 strict_overflow_p
));
13528 /* Return true when T is an address and is known to be nonzero.
13529 For floating point we further ensure that T is not denormal.
13530 Similar logic is present in nonzero_address in rtlanal.h.
13532 If the return value is based on the assumption that signed overflow
13533 is undefined, set *STRICT_OVERFLOW_P to true; otherwise, don't
13534 change *STRICT_OVERFLOW_P. */
13537 tree_single_nonzero_warnv_p (tree t
, bool *strict_overflow_p
)
13539 bool sub_strict_overflow_p
;
13540 switch (TREE_CODE (t
))
13543 return !integer_zerop (t
);
13547 tree base
= TREE_OPERAND (t
, 0);
13549 if (!DECL_P (base
))
13550 base
= get_base_address (base
);
13555 /* For objects in symbol table check if we know they are non-zero.
13556 Don't do anything for variables and functions before symtab is built;
13557 it is quite possible that they will be declared weak later. */
13558 int nonzero_addr
= maybe_nonzero_address (base
);
13559 if (nonzero_addr
>= 0)
13560 return nonzero_addr
;
13562 /* Function local objects are never NULL. */
13564 && (DECL_CONTEXT (base
)
13565 && TREE_CODE (DECL_CONTEXT (base
)) == FUNCTION_DECL
13566 && auto_var_in_fn_p (base
, DECL_CONTEXT (base
))))
13569 /* Constants are never weak. */
13570 if (CONSTANT_CLASS_P (base
))
13577 sub_strict_overflow_p
= false;
13578 if (tree_expr_nonzero_warnv_p (TREE_OPERAND (t
, 1),
13579 &sub_strict_overflow_p
)
13580 && tree_expr_nonzero_warnv_p (TREE_OPERAND (t
, 2),
13581 &sub_strict_overflow_p
))
13583 if (sub_strict_overflow_p
)
13584 *strict_overflow_p
= true;
13595 #define integer_valued_real_p(X) \
13596 _Pragma ("GCC error \"Use RECURSE for recursive calls\"") 0
13598 #define RECURSE(X) \
13599 ((integer_valued_real_p) (X, depth + 1))
13601 /* Return true if the floating point result of (CODE OP0) has an
13602 integer value. We also allow +Inf, -Inf and NaN to be considered
13603 integer values. Return false for signaling NaN.
13605 DEPTH is the current nesting depth of the query. */
13608 integer_valued_real_unary_p (tree_code code
, tree op0
, int depth
)
13616 return RECURSE (op0
);
13620 tree type
= TREE_TYPE (op0
);
13621 if (TREE_CODE (type
) == INTEGER_TYPE
)
13623 if (TREE_CODE (type
) == REAL_TYPE
)
13624 return RECURSE (op0
);
13634 /* Return true if the floating point result of (CODE OP0 OP1) has an
13635 integer value. We also allow +Inf, -Inf and NaN to be considered
13636 integer values. Return false for signaling NaN.
13638 DEPTH is the current nesting depth of the query. */
13641 integer_valued_real_binary_p (tree_code code
, tree op0
, tree op1
, int depth
)
13650 return RECURSE (op0
) && RECURSE (op1
);
13658 /* Return true if the floating point result of calling FNDECL with arguments
13659 ARG0 and ARG1 has an integer value. We also allow +Inf, -Inf and NaN to be
13660 considered integer values. Return false for signaling NaN. If FNDECL
13661 takes fewer than 2 arguments, the remaining ARGn are null.
13663 DEPTH is the current nesting depth of the query. */
13666 integer_valued_real_call_p (combined_fn fn
, tree arg0
, tree arg1
, int depth
)
13672 CASE_CFN_NEARBYINT
:
13680 return RECURSE (arg0
) && RECURSE (arg1
);
13688 /* Return true if the floating point expression T (a GIMPLE_SINGLE_RHS)
13689 has an integer value. We also allow +Inf, -Inf and NaN to be
13690 considered integer values. Return false for signaling NaN.
13692 DEPTH is the current nesting depth of the query. */
13695 integer_valued_real_single_p (tree t
, int depth
)
13697 switch (TREE_CODE (t
))
13700 return real_isinteger (TREE_REAL_CST_PTR (t
), TYPE_MODE (TREE_TYPE (t
)));
13703 return RECURSE (TREE_OPERAND (t
, 1)) && RECURSE (TREE_OPERAND (t
, 2));
13706 /* Limit the depth of recursion to avoid quadratic behavior.
13707 This is expected to catch almost all occurrences in practice.
13708 If this code misses important cases that unbounded recursion
13709 would not, passes that need this information could be revised
13710 to provide it through dataflow propagation. */
13711 return (!name_registered_for_update_p (t
)
13712 && depth
< PARAM_VALUE (PARAM_MAX_SSA_NAME_QUERY_DEPTH
)
13713 && gimple_stmt_integer_valued_real_p (SSA_NAME_DEF_STMT (t
),
13722 /* Return true if the floating point expression T (a GIMPLE_INVALID_RHS)
13723 has an integer value. We also allow +Inf, -Inf and NaN to be
13724 considered integer values. Return false for signaling NaN.
13726 DEPTH is the current nesting depth of the query. */
13729 integer_valued_real_invalid_p (tree t
, int depth
)
13731 switch (TREE_CODE (t
))
13733 case COMPOUND_EXPR
:
13736 return RECURSE (TREE_OPERAND (t
, 1));
13739 return RECURSE (TREE_OPERAND (t
, 0));
13748 #undef integer_valued_real_p
13750 /* Return true if the floating point expression T has an integer value.
13751 We also allow +Inf, -Inf and NaN to be considered integer values.
13752 Return false for signaling NaN.
13754 DEPTH is the current nesting depth of the query. */
13757 integer_valued_real_p (tree t
, int depth
)
13759 if (t
== error_mark_node
)
13762 tree_code code
= TREE_CODE (t
);
13763 switch (TREE_CODE_CLASS (code
))
13766 case tcc_comparison
:
13767 return integer_valued_real_binary_p (code
, TREE_OPERAND (t
, 0),
13768 TREE_OPERAND (t
, 1), depth
);
13771 return integer_valued_real_unary_p (code
, TREE_OPERAND (t
, 0), depth
);
13774 case tcc_declaration
:
13775 case tcc_reference
:
13776 return integer_valued_real_single_p (t
, depth
);
13786 return integer_valued_real_single_p (t
, depth
);
13790 tree arg0
= (call_expr_nargs (t
) > 0
13791 ? CALL_EXPR_ARG (t
, 0)
13793 tree arg1
= (call_expr_nargs (t
) > 1
13794 ? CALL_EXPR_ARG (t
, 1)
13796 return integer_valued_real_call_p (get_call_combined_fn (t
),
13797 arg0
, arg1
, depth
);
13801 return integer_valued_real_invalid_p (t
, depth
);
13805 /* Given the components of a binary expression CODE, TYPE, OP0 and OP1,
13806 attempt to fold the expression to a constant without modifying TYPE,
13809 If the expression could be simplified to a constant, then return
13810 the constant. If the expression would not be simplified to a
13811 constant, then return NULL_TREE. */
13814 fold_binary_to_constant (enum tree_code code
, tree type
, tree op0
, tree op1
)
13816 tree tem
= fold_binary (code
, type
, op0
, op1
);
13817 return (tem
&& TREE_CONSTANT (tem
)) ? tem
: NULL_TREE
;
13820 /* Given the components of a unary expression CODE, TYPE and OP0,
13821 attempt to fold the expression to a constant without modifying
13824 If the expression could be simplified to a constant, then return
13825 the constant. If the expression would not be simplified to a
13826 constant, then return NULL_TREE. */
13829 fold_unary_to_constant (enum tree_code code
, tree type
, tree op0
)
13831 tree tem
= fold_unary (code
, type
, op0
);
13832 return (tem
&& TREE_CONSTANT (tem
)) ? tem
: NULL_TREE
;
13835 /* If EXP represents referencing an element in a constant string
13836 (either via pointer arithmetic or array indexing), return the
13837 tree representing the value accessed, otherwise return NULL. */
13840 fold_read_from_constant_string (tree exp
)
13842 if ((TREE_CODE (exp
) == INDIRECT_REF
13843 || TREE_CODE (exp
) == ARRAY_REF
)
13844 && TREE_CODE (TREE_TYPE (exp
)) == INTEGER_TYPE
)
13846 tree exp1
= TREE_OPERAND (exp
, 0);
13849 location_t loc
= EXPR_LOCATION (exp
);
13851 if (TREE_CODE (exp
) == INDIRECT_REF
)
13852 string
= string_constant (exp1
, &index
);
13855 tree low_bound
= array_ref_low_bound (exp
);
13856 index
= fold_convert_loc (loc
, sizetype
, TREE_OPERAND (exp
, 1));
13858 /* Optimize the special-case of a zero lower bound.
13860 We convert the low_bound to sizetype to avoid some problems
13861 with constant folding. (E.g. suppose the lower bound is 1,
13862 and its mode is QI. Without the conversion,l (ARRAY
13863 +(INDEX-(unsigned char)1)) becomes ((ARRAY+(-(unsigned char)1))
13864 +INDEX), which becomes (ARRAY+255+INDEX). Oops!) */
13865 if (! integer_zerop (low_bound
))
13866 index
= size_diffop_loc (loc
, index
,
13867 fold_convert_loc (loc
, sizetype
, low_bound
));
13873 && TYPE_MODE (TREE_TYPE (exp
)) == TYPE_MODE (TREE_TYPE (TREE_TYPE (string
)))
13874 && TREE_CODE (string
) == STRING_CST
13875 && TREE_CODE (index
) == INTEGER_CST
13876 && compare_tree_int (index
, TREE_STRING_LENGTH (string
)) < 0
13877 && (GET_MODE_CLASS (TYPE_MODE (TREE_TYPE (TREE_TYPE (string
))))
13879 && (GET_MODE_SIZE (TYPE_MODE (TREE_TYPE (TREE_TYPE (string
)))) == 1))
13880 return build_int_cst_type (TREE_TYPE (exp
),
13881 (TREE_STRING_POINTER (string
)
13882 [TREE_INT_CST_LOW (index
)]));
13887 /* Return the tree for neg (ARG0) when ARG0 is known to be either
13888 an integer constant, real, or fixed-point constant.
13890 TYPE is the type of the result. */
13893 fold_negate_const (tree arg0
, tree type
)
13895 tree t
= NULL_TREE
;
13897 switch (TREE_CODE (arg0
))
13902 wide_int val
= wi::neg (arg0
, &overflow
);
13903 t
= force_fit_type (type
, val
, 1,
13904 (overflow
| TREE_OVERFLOW (arg0
))
13905 && !TYPE_UNSIGNED (type
));
13910 t
= build_real (type
, real_value_negate (&TREE_REAL_CST (arg0
)));
13915 FIXED_VALUE_TYPE f
;
13916 bool overflow_p
= fixed_arithmetic (&f
, NEGATE_EXPR
,
13917 &(TREE_FIXED_CST (arg0
)), NULL
,
13918 TYPE_SATURATING (type
));
13919 t
= build_fixed (type
, f
);
13920 /* Propagate overflow flags. */
13921 if (overflow_p
| TREE_OVERFLOW (arg0
))
13922 TREE_OVERFLOW (t
) = 1;
13927 gcc_unreachable ();
13933 /* Return the tree for abs (ARG0) when ARG0 is known to be either
13934 an integer constant or real constant.
13936 TYPE is the type of the result. */
13939 fold_abs_const (tree arg0
, tree type
)
13941 tree t
= NULL_TREE
;
13943 switch (TREE_CODE (arg0
))
13947 /* If the value is unsigned or non-negative, then the absolute value
13948 is the same as the ordinary value. */
13949 if (!wi::neg_p (arg0
, TYPE_SIGN (type
)))
13952 /* If the value is negative, then the absolute value is
13957 wide_int val
= wi::neg (arg0
, &overflow
);
13958 t
= force_fit_type (type
, val
, -1,
13959 overflow
| TREE_OVERFLOW (arg0
));
13965 if (REAL_VALUE_NEGATIVE (TREE_REAL_CST (arg0
)))
13966 t
= build_real (type
, real_value_negate (&TREE_REAL_CST (arg0
)));
13972 gcc_unreachable ();
13978 /* Return the tree for not (ARG0) when ARG0 is known to be an integer
13979 constant. TYPE is the type of the result. */
13982 fold_not_const (const_tree arg0
, tree type
)
13984 gcc_assert (TREE_CODE (arg0
) == INTEGER_CST
);
13986 return force_fit_type (type
, wi::bit_not (arg0
), 0, TREE_OVERFLOW (arg0
));
13989 /* Given CODE, a relational operator, the target type, TYPE and two
13990 constant operands OP0 and OP1, return the result of the
13991 relational operation. If the result is not a compile time
13992 constant, then return NULL_TREE. */
13995 fold_relational_const (enum tree_code code
, tree type
, tree op0
, tree op1
)
13997 int result
, invert
;
13999 /* From here on, the only cases we handle are when the result is
14000 known to be a constant. */
14002 if (TREE_CODE (op0
) == REAL_CST
&& TREE_CODE (op1
) == REAL_CST
)
14004 const REAL_VALUE_TYPE
*c0
= TREE_REAL_CST_PTR (op0
);
14005 const REAL_VALUE_TYPE
*c1
= TREE_REAL_CST_PTR (op1
);
14007 /* Handle the cases where either operand is a NaN. */
14008 if (real_isnan (c0
) || real_isnan (c1
))
14018 case UNORDERED_EXPR
:
14032 if (flag_trapping_math
)
14038 gcc_unreachable ();
14041 return constant_boolean_node (result
, type
);
14044 return constant_boolean_node (real_compare (code
, c0
, c1
), type
);
14047 if (TREE_CODE (op0
) == FIXED_CST
&& TREE_CODE (op1
) == FIXED_CST
)
14049 const FIXED_VALUE_TYPE
*c0
= TREE_FIXED_CST_PTR (op0
);
14050 const FIXED_VALUE_TYPE
*c1
= TREE_FIXED_CST_PTR (op1
);
14051 return constant_boolean_node (fixed_compare (code
, c0
, c1
), type
);
14054 /* Handle equality/inequality of complex constants. */
14055 if (TREE_CODE (op0
) == COMPLEX_CST
&& TREE_CODE (op1
) == COMPLEX_CST
)
14057 tree rcond
= fold_relational_const (code
, type
,
14058 TREE_REALPART (op0
),
14059 TREE_REALPART (op1
));
14060 tree icond
= fold_relational_const (code
, type
,
14061 TREE_IMAGPART (op0
),
14062 TREE_IMAGPART (op1
));
14063 if (code
== EQ_EXPR
)
14064 return fold_build2 (TRUTH_ANDIF_EXPR
, type
, rcond
, icond
);
14065 else if (code
== NE_EXPR
)
14066 return fold_build2 (TRUTH_ORIF_EXPR
, type
, rcond
, icond
);
14071 if (TREE_CODE (op0
) == VECTOR_CST
&& TREE_CODE (op1
) == VECTOR_CST
)
14073 if (!VECTOR_TYPE_P (type
))
14075 /* Have vector comparison with scalar boolean result. */
14076 bool result
= true;
14077 gcc_assert ((code
== EQ_EXPR
|| code
== NE_EXPR
)
14078 && VECTOR_CST_NELTS (op0
) == VECTOR_CST_NELTS (op1
));
14079 for (unsigned i
= 0; i
< VECTOR_CST_NELTS (op0
); i
++)
14081 tree elem0
= VECTOR_CST_ELT (op0
, i
);
14082 tree elem1
= VECTOR_CST_ELT (op1
, i
);
14083 tree tmp
= fold_relational_const (code
, type
, elem0
, elem1
);
14084 result
&= integer_onep (tmp
);
14086 if (code
== NE_EXPR
)
14088 return constant_boolean_node (result
, type
);
14090 unsigned count
= VECTOR_CST_NELTS (op0
);
14091 tree
*elts
= XALLOCAVEC (tree
, count
);
14092 gcc_assert (VECTOR_CST_NELTS (op1
) == count
14093 && TYPE_VECTOR_SUBPARTS (type
) == count
);
14095 for (unsigned i
= 0; i
< count
; i
++)
14097 tree elem_type
= TREE_TYPE (type
);
14098 tree elem0
= VECTOR_CST_ELT (op0
, i
);
14099 tree elem1
= VECTOR_CST_ELT (op1
, i
);
14101 tree tem
= fold_relational_const (code
, elem_type
,
14104 if (tem
== NULL_TREE
)
14107 elts
[i
] = build_int_cst (elem_type
, integer_zerop (tem
) ? 0 : -1);
14110 return build_vector (type
, elts
);
14113 /* From here on we only handle LT, LE, GT, GE, EQ and NE.
14115 To compute GT, swap the arguments and do LT.
14116 To compute GE, do LT and invert the result.
14117 To compute LE, swap the arguments, do LT and invert the result.
14118 To compute NE, do EQ and invert the result.
14120 Therefore, the code below must handle only EQ and LT. */
14122 if (code
== LE_EXPR
|| code
== GT_EXPR
)
14124 std::swap (op0
, op1
);
14125 code
= swap_tree_comparison (code
);
14128 /* Note that it is safe to invert for real values here because we
14129 have already handled the one case that it matters. */
14132 if (code
== NE_EXPR
|| code
== GE_EXPR
)
14135 code
= invert_tree_comparison (code
, false);
14138 /* Compute a result for LT or EQ if args permit;
14139 Otherwise return T. */
14140 if (TREE_CODE (op0
) == INTEGER_CST
&& TREE_CODE (op1
) == INTEGER_CST
)
14142 if (code
== EQ_EXPR
)
14143 result
= tree_int_cst_equal (op0
, op1
);
14145 result
= tree_int_cst_lt (op0
, op1
);
14152 return constant_boolean_node (result
, type
);
14155 /* If necessary, return a CLEANUP_POINT_EXPR for EXPR with the
14156 indicated TYPE. If no CLEANUP_POINT_EXPR is necessary, return EXPR
14160 fold_build_cleanup_point_expr (tree type
, tree expr
)
14162 /* If the expression does not have side effects then we don't have to wrap
14163 it with a cleanup point expression. */
14164 if (!TREE_SIDE_EFFECTS (expr
))
14167 /* If the expression is a return, check to see if the expression inside the
14168 return has no side effects or the right hand side of the modify expression
14169 inside the return. If either don't have side effects set we don't need to
14170 wrap the expression in a cleanup point expression. Note we don't check the
14171 left hand side of the modify because it should always be a return decl. */
14172 if (TREE_CODE (expr
) == RETURN_EXPR
)
14174 tree op
= TREE_OPERAND (expr
, 0);
14175 if (!op
|| !TREE_SIDE_EFFECTS (op
))
14177 op
= TREE_OPERAND (op
, 1);
14178 if (!TREE_SIDE_EFFECTS (op
))
14182 return build1 (CLEANUP_POINT_EXPR
, type
, expr
);
14185 /* Given a pointer value OP0 and a type TYPE, return a simplified version
14186 of an indirection through OP0, or NULL_TREE if no simplification is
14190 fold_indirect_ref_1 (location_t loc
, tree type
, tree op0
)
14196 subtype
= TREE_TYPE (sub
);
14197 if (!POINTER_TYPE_P (subtype
))
14200 if (TREE_CODE (sub
) == ADDR_EXPR
)
14202 tree op
= TREE_OPERAND (sub
, 0);
14203 tree optype
= TREE_TYPE (op
);
14204 /* *&CONST_DECL -> to the value of the const decl. */
14205 if (TREE_CODE (op
) == CONST_DECL
)
14206 return DECL_INITIAL (op
);
14207 /* *&p => p; make sure to handle *&"str"[cst] here. */
14208 if (type
== optype
)
14210 tree fop
= fold_read_from_constant_string (op
);
14216 /* *(foo *)&fooarray => fooarray[0] */
14217 else if (TREE_CODE (optype
) == ARRAY_TYPE
14218 && type
== TREE_TYPE (optype
)
14219 && (!in_gimple_form
14220 || TREE_CODE (TYPE_SIZE (type
)) == INTEGER_CST
))
14222 tree type_domain
= TYPE_DOMAIN (optype
);
14223 tree min_val
= size_zero_node
;
14224 if (type_domain
&& TYPE_MIN_VALUE (type_domain
))
14225 min_val
= TYPE_MIN_VALUE (type_domain
);
14227 && TREE_CODE (min_val
) != INTEGER_CST
)
14229 return build4_loc (loc
, ARRAY_REF
, type
, op
, min_val
,
14230 NULL_TREE
, NULL_TREE
);
14232 /* *(foo *)&complexfoo => __real__ complexfoo */
14233 else if (TREE_CODE (optype
) == COMPLEX_TYPE
14234 && type
== TREE_TYPE (optype
))
14235 return fold_build1_loc (loc
, REALPART_EXPR
, type
, op
);
14236 /* *(foo *)&vectorfoo => BIT_FIELD_REF<vectorfoo,...> */
14237 else if (TREE_CODE (optype
) == VECTOR_TYPE
14238 && type
== TREE_TYPE (optype
))
14240 tree part_width
= TYPE_SIZE (type
);
14241 tree index
= bitsize_int (0);
14242 return fold_build3_loc (loc
, BIT_FIELD_REF
, type
, op
, part_width
, index
);
14246 if (TREE_CODE (sub
) == POINTER_PLUS_EXPR
14247 && TREE_CODE (TREE_OPERAND (sub
, 1)) == INTEGER_CST
)
14249 tree op00
= TREE_OPERAND (sub
, 0);
14250 tree op01
= TREE_OPERAND (sub
, 1);
14253 if (TREE_CODE (op00
) == ADDR_EXPR
)
14256 op00
= TREE_OPERAND (op00
, 0);
14257 op00type
= TREE_TYPE (op00
);
14259 /* ((foo*)&vectorfoo)[1] => BIT_FIELD_REF<vectorfoo,...> */
14260 if (TREE_CODE (op00type
) == VECTOR_TYPE
14261 && type
== TREE_TYPE (op00type
))
14263 tree part_width
= TYPE_SIZE (type
);
14264 unsigned HOST_WIDE_INT max_offset
14265 = (tree_to_uhwi (part_width
) / BITS_PER_UNIT
14266 * TYPE_VECTOR_SUBPARTS (op00type
));
14267 if (tree_int_cst_sign_bit (op01
) == 0
14268 && compare_tree_int (op01
, max_offset
) == -1)
14270 unsigned HOST_WIDE_INT offset
= tree_to_uhwi (op01
);
14271 unsigned HOST_WIDE_INT indexi
= offset
* BITS_PER_UNIT
;
14272 tree index
= bitsize_int (indexi
);
14273 return fold_build3_loc (loc
,
14274 BIT_FIELD_REF
, type
, op00
,
14275 part_width
, index
);
14278 /* ((foo*)&complexfoo)[1] => __imag__ complexfoo */
14279 else if (TREE_CODE (op00type
) == COMPLEX_TYPE
14280 && type
== TREE_TYPE (op00type
))
14282 tree size
= TYPE_SIZE_UNIT (type
);
14283 if (tree_int_cst_equal (size
, op01
))
14284 return fold_build1_loc (loc
, IMAGPART_EXPR
, type
, op00
);
14286 /* ((foo *)&fooarray)[1] => fooarray[1] */
14287 else if (TREE_CODE (op00type
) == ARRAY_TYPE
14288 && type
== TREE_TYPE (op00type
))
14290 tree type_domain
= TYPE_DOMAIN (op00type
);
14291 tree min_val
= size_zero_node
;
14292 if (type_domain
&& TYPE_MIN_VALUE (type_domain
))
14293 min_val
= TYPE_MIN_VALUE (type_domain
);
14294 op01
= size_binop_loc (loc
, EXACT_DIV_EXPR
, op01
,
14295 TYPE_SIZE_UNIT (type
));
14296 op01
= size_binop_loc (loc
, PLUS_EXPR
, op01
, min_val
);
14297 return build4_loc (loc
, ARRAY_REF
, type
, op00
, op01
,
14298 NULL_TREE
, NULL_TREE
);
14303 /* *(foo *)fooarrptr => (*fooarrptr)[0] */
14304 if (TREE_CODE (TREE_TYPE (subtype
)) == ARRAY_TYPE
14305 && type
== TREE_TYPE (TREE_TYPE (subtype
))
14306 && (!in_gimple_form
14307 || TREE_CODE (TYPE_SIZE (type
)) == INTEGER_CST
))
14310 tree min_val
= size_zero_node
;
14311 sub
= build_fold_indirect_ref_loc (loc
, sub
);
14312 type_domain
= TYPE_DOMAIN (TREE_TYPE (sub
));
14313 if (type_domain
&& TYPE_MIN_VALUE (type_domain
))
14314 min_val
= TYPE_MIN_VALUE (type_domain
);
14316 && TREE_CODE (min_val
) != INTEGER_CST
)
14318 return build4_loc (loc
, ARRAY_REF
, type
, sub
, min_val
, NULL_TREE
,
14325 /* Builds an expression for an indirection through T, simplifying some
14329 build_fold_indirect_ref_loc (location_t loc
, tree t
)
14331 tree type
= TREE_TYPE (TREE_TYPE (t
));
14332 tree sub
= fold_indirect_ref_1 (loc
, type
, t
);
14337 return build1_loc (loc
, INDIRECT_REF
, type
, t
);
14340 /* Given an INDIRECT_REF T, return either T or a simplified version. */
14343 fold_indirect_ref_loc (location_t loc
, tree t
)
14345 tree sub
= fold_indirect_ref_1 (loc
, TREE_TYPE (t
), TREE_OPERAND (t
, 0));
14353 /* Strip non-trapping, non-side-effecting tree nodes from an expression
14354 whose result is ignored. The type of the returned tree need not be
14355 the same as the original expression. */
14358 fold_ignored_result (tree t
)
14360 if (!TREE_SIDE_EFFECTS (t
))
14361 return integer_zero_node
;
14364 switch (TREE_CODE_CLASS (TREE_CODE (t
)))
14367 t
= TREE_OPERAND (t
, 0);
14371 case tcc_comparison
:
14372 if (!TREE_SIDE_EFFECTS (TREE_OPERAND (t
, 1)))
14373 t
= TREE_OPERAND (t
, 0);
14374 else if (!TREE_SIDE_EFFECTS (TREE_OPERAND (t
, 0)))
14375 t
= TREE_OPERAND (t
, 1);
14380 case tcc_expression
:
14381 switch (TREE_CODE (t
))
14383 case COMPOUND_EXPR
:
14384 if (TREE_SIDE_EFFECTS (TREE_OPERAND (t
, 1)))
14386 t
= TREE_OPERAND (t
, 0);
14390 if (TREE_SIDE_EFFECTS (TREE_OPERAND (t
, 1))
14391 || TREE_SIDE_EFFECTS (TREE_OPERAND (t
, 2)))
14393 t
= TREE_OPERAND (t
, 0);
14406 /* Return the value of VALUE, rounded up to a multiple of DIVISOR. */
14409 round_up_loc (location_t loc
, tree value
, unsigned int divisor
)
14411 tree div
= NULL_TREE
;
14416 /* See if VALUE is already a multiple of DIVISOR. If so, we don't
14417 have to do anything. Only do this when we are not given a const,
14418 because in that case, this check is more expensive than just
14420 if (TREE_CODE (value
) != INTEGER_CST
)
14422 div
= build_int_cst (TREE_TYPE (value
), divisor
);
14424 if (multiple_of_p (TREE_TYPE (value
), value
, div
))
14428 /* If divisor is a power of two, simplify this to bit manipulation. */
14429 if (divisor
== (divisor
& -divisor
))
14431 if (TREE_CODE (value
) == INTEGER_CST
)
14433 wide_int val
= value
;
14436 if ((val
& (divisor
- 1)) == 0)
14439 overflow_p
= TREE_OVERFLOW (value
);
14440 val
+= divisor
- 1;
14441 val
&= - (int) divisor
;
14445 return force_fit_type (TREE_TYPE (value
), val
, -1, overflow_p
);
14451 t
= build_int_cst (TREE_TYPE (value
), divisor
- 1);
14452 value
= size_binop_loc (loc
, PLUS_EXPR
, value
, t
);
14453 t
= build_int_cst (TREE_TYPE (value
), - (int) divisor
);
14454 value
= size_binop_loc (loc
, BIT_AND_EXPR
, value
, t
);
14460 div
= build_int_cst (TREE_TYPE (value
), divisor
);
14461 value
= size_binop_loc (loc
, CEIL_DIV_EXPR
, value
, div
);
14462 value
= size_binop_loc (loc
, MULT_EXPR
, value
, div
);
14468 /* Likewise, but round down. */
14471 round_down_loc (location_t loc
, tree value
, int divisor
)
14473 tree div
= NULL_TREE
;
14475 gcc_assert (divisor
> 0);
14479 /* See if VALUE is already a multiple of DIVISOR. If so, we don't
14480 have to do anything. Only do this when we are not given a const,
14481 because in that case, this check is more expensive than just
14483 if (TREE_CODE (value
) != INTEGER_CST
)
14485 div
= build_int_cst (TREE_TYPE (value
), divisor
);
14487 if (multiple_of_p (TREE_TYPE (value
), value
, div
))
14491 /* If divisor is a power of two, simplify this to bit manipulation. */
14492 if (divisor
== (divisor
& -divisor
))
14496 t
= build_int_cst (TREE_TYPE (value
), -divisor
);
14497 value
= size_binop_loc (loc
, BIT_AND_EXPR
, value
, t
);
14502 div
= build_int_cst (TREE_TYPE (value
), divisor
);
14503 value
= size_binop_loc (loc
, FLOOR_DIV_EXPR
, value
, div
);
14504 value
= size_binop_loc (loc
, MULT_EXPR
, value
, div
);
14510 /* Returns the pointer to the base of the object addressed by EXP and
14511 extracts the information about the offset of the access, storing it
14512 to PBITPOS and POFFSET. */
14515 split_address_to_core_and_offset (tree exp
,
14516 HOST_WIDE_INT
*pbitpos
, tree
*poffset
)
14520 int unsignedp
, reversep
, volatilep
;
14521 HOST_WIDE_INT bitsize
;
14522 location_t loc
= EXPR_LOCATION (exp
);
14524 if (TREE_CODE (exp
) == ADDR_EXPR
)
14526 core
= get_inner_reference (TREE_OPERAND (exp
, 0), &bitsize
, pbitpos
,
14527 poffset
, &mode
, &unsignedp
, &reversep
,
14528 &volatilep
, false);
14529 core
= build_fold_addr_expr_loc (loc
, core
);
14535 *poffset
= NULL_TREE
;
14541 /* Returns true if addresses of E1 and E2 differ by a constant, false
14542 otherwise. If they do, E1 - E2 is stored in *DIFF. */
14545 ptr_difference_const (tree e1
, tree e2
, HOST_WIDE_INT
*diff
)
14548 HOST_WIDE_INT bitpos1
, bitpos2
;
14549 tree toffset1
, toffset2
, tdiff
, type
;
14551 core1
= split_address_to_core_and_offset (e1
, &bitpos1
, &toffset1
);
14552 core2
= split_address_to_core_and_offset (e2
, &bitpos2
, &toffset2
);
14554 if (bitpos1
% BITS_PER_UNIT
!= 0
14555 || bitpos2
% BITS_PER_UNIT
!= 0
14556 || !operand_equal_p (core1
, core2
, 0))
14559 if (toffset1
&& toffset2
)
14561 type
= TREE_TYPE (toffset1
);
14562 if (type
!= TREE_TYPE (toffset2
))
14563 toffset2
= fold_convert (type
, toffset2
);
14565 tdiff
= fold_build2 (MINUS_EXPR
, type
, toffset1
, toffset2
);
14566 if (!cst_and_fits_in_hwi (tdiff
))
14569 *diff
= int_cst_value (tdiff
);
14571 else if (toffset1
|| toffset2
)
14573 /* If only one of the offsets is non-constant, the difference cannot
14580 *diff
+= (bitpos1
- bitpos2
) / BITS_PER_UNIT
;
14584 /* Return OFF converted to a pointer offset type suitable as offset for
14585 POINTER_PLUS_EXPR. Use location LOC for this conversion. */
14587 convert_to_ptrofftype_loc (location_t loc
, tree off
)
14589 return fold_convert_loc (loc
, sizetype
, off
);
14592 /* Build and fold a POINTER_PLUS_EXPR at LOC offsetting PTR by OFF. */
14594 fold_build_pointer_plus_loc (location_t loc
, tree ptr
, tree off
)
14596 return fold_build2_loc (loc
, POINTER_PLUS_EXPR
, TREE_TYPE (ptr
),
14597 ptr
, convert_to_ptrofftype_loc (loc
, off
));
14600 /* Build and fold a POINTER_PLUS_EXPR at LOC offsetting PTR by OFF. */
14602 fold_build_pointer_plus_hwi_loc (location_t loc
, tree ptr
, HOST_WIDE_INT off
)
14604 return fold_build2_loc (loc
, POINTER_PLUS_EXPR
, TREE_TYPE (ptr
),
14605 ptr
, size_int (off
));
14608 /* Return a char pointer for a C string if it is a string constant
14609 or sum of string constant and integer constant. */
14612 c_getstr (tree src
)
14616 src
= string_constant (src
, &offset_node
);
14620 if (offset_node
== 0)
14621 return TREE_STRING_POINTER (src
);
14622 else if (!tree_fits_uhwi_p (offset_node
)
14623 || compare_tree_int (offset_node
, TREE_STRING_LENGTH (src
) - 1) > 0)
14626 return TREE_STRING_POINTER (src
) + tree_to_uhwi (offset_node
);