1 /* Constant folding for calls to built-in and internal functions.
2 Copyright (C) 1988-2021 Free Software Foundation, Inc.
4 This file is part of GCC.
6 GCC is free software; you can redistribute it and/or modify it under
7 the terms of the GNU General Public License as published by the Free
8 Software Foundation; either version 3, or (at your option) any later
11 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
12 WARRANTY; without even the implied warranty of MERCHANTABILITY or
13 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
16 You should have received a copy of the GNU General Public License
17 along with GCC; see the file COPYING3. If not see
18 <http://www.gnu.org/licenses/>. */
22 #include "coretypes.h"
25 #include "stor-layout.h"
27 #include "fold-const.h"
28 #include "fold-const-call.h"
29 #include "case-cfn-macros.h"
30 #include "tm.h" /* For C[LT]Z_DEFINED_AT_ZERO. */
32 #include "gimple-expr.h"
33 #include "tree-vector-builder.h"
35 /* Functions that test for certain constant types, abstracting away the
36 decision about whether to check for overflow. */
39 integer_cst_p (tree t
)
41 return TREE_CODE (t
) == INTEGER_CST
&& !TREE_OVERFLOW (t
);
47 return TREE_CODE (t
) == REAL_CST
&& !TREE_OVERFLOW (t
);
51 complex_cst_p (tree t
)
53 return TREE_CODE (t
) == COMPLEX_CST
;
56 /* Return true if ARG is a size_type_node constant.
57 Store it in *SIZE_OUT if so. */
60 size_t_cst_p (tree t
, unsigned HOST_WIDE_INT
*size_out
)
62 if (types_compatible_p (size_type_node
, TREE_TYPE (t
))
64 && tree_fits_uhwi_p (t
))
66 *size_out
= tree_to_uhwi (t
);
72 /* RES is the result of a comparison in which < 0 means "less", 0 means
73 "equal" and > 0 means "more". Canonicalize it to -1, 0 or 1 and
74 return it in type TYPE. */
77 build_cmp_result (tree type
, int res
)
79 return build_int_cst (type
, res
< 0 ? -1 : res
> 0 ? 1 : 0);
82 /* M is the result of trying to constant-fold an expression (starting
83 with clear MPFR flags) and INEXACT says whether the result in M is
84 exact or inexact. Return true if M can be used as a constant-folded
85 result in format FORMAT, storing the value in *RESULT if so. */
88 do_mpfr_ckconv (real_value
*result
, mpfr_srcptr m
, bool inexact
,
89 const real_format
*format
)
91 /* Proceed iff we get a normal number, i.e. not NaN or Inf and no
92 overflow/underflow occurred. If -frounding-math, proceed iff the
93 result of calling FUNC was exact. */
94 if (!mpfr_number_p (m
)
96 || mpfr_underflow_p ()
97 || (flag_rounding_math
&& inexact
))
101 real_from_mpfr (&tmp
, m
, format
, MPFR_RNDN
);
103 /* Proceed iff GCC's REAL_VALUE_TYPE can hold the MPFR values.
104 If the REAL_VALUE_TYPE is zero but the mpft_t is not, then we
105 underflowed in the conversion. */
106 if (!real_isfinite (&tmp
)
107 || ((tmp
.cl
== rvc_zero
) != (mpfr_zero_p (m
) != 0)))
110 real_convert (result
, format
, &tmp
);
111 return real_identical (result
, &tmp
);
118 in format FORMAT, given that FUNC is the MPFR implementation of f.
119 Return true on success. */
122 do_mpfr_arg1 (real_value
*result
,
123 int (*func
) (mpfr_ptr
, mpfr_srcptr
, mpfr_rnd_t
),
124 const real_value
*arg
, const real_format
*format
)
126 /* To proceed, MPFR must exactly represent the target floating point
127 format, which only happens when the target base equals two. */
128 if (format
->b
!= 2 || !real_isfinite (arg
))
131 int prec
= format
->p
;
132 mpfr_rnd_t rnd
= format
->round_towards_zero
? MPFR_RNDZ
: MPFR_RNDN
;
135 mpfr_init2 (m
, prec
);
136 mpfr_from_real (m
, arg
, MPFR_RNDN
);
138 bool inexact
= func (m
, m
, rnd
);
139 bool ok
= do_mpfr_ckconv (result
, m
, inexact
, format
);
147 *RESULT_SIN = sin (*ARG);
148 *RESULT_COS = cos (*ARG);
150 for format FORMAT. Return true on success. */
153 do_mpfr_sincos (real_value
*result_sin
, real_value
*result_cos
,
154 const real_value
*arg
, const real_format
*format
)
156 /* To proceed, MPFR must exactly represent the target floating point
157 format, which only happens when the target base equals two. */
158 if (format
->b
!= 2 || !real_isfinite (arg
))
161 int prec
= format
->p
;
162 mpfr_rnd_t rnd
= format
->round_towards_zero
? MPFR_RNDZ
: MPFR_RNDN
;
165 mpfr_inits2 (prec
, m
, ms
, mc
, NULL
);
166 mpfr_from_real (m
, arg
, MPFR_RNDN
);
168 bool inexact
= mpfr_sin_cos (ms
, mc
, m
, rnd
);
169 bool ok
= (do_mpfr_ckconv (result_sin
, ms
, inexact
, format
)
170 && do_mpfr_ckconv (result_cos
, mc
, inexact
, format
));
171 mpfr_clears (m
, ms
, mc
, NULL
);
178 *RESULT = f (*ARG0, *ARG1)
180 in format FORMAT, given that FUNC is the MPFR implementation of f.
181 Return true on success. */
184 do_mpfr_arg2 (real_value
*result
,
185 int (*func
) (mpfr_ptr
, mpfr_srcptr
, mpfr_srcptr
, mpfr_rnd_t
),
186 const real_value
*arg0
, const real_value
*arg1
,
187 const real_format
*format
)
189 /* To proceed, MPFR must exactly represent the target floating point
190 format, which only happens when the target base equals two. */
191 if (format
->b
!= 2 || !real_isfinite (arg0
) || !real_isfinite (arg1
))
194 int prec
= format
->p
;
195 mpfr_rnd_t rnd
= format
->round_towards_zero
? MPFR_RNDZ
: MPFR_RNDN
;
198 mpfr_inits2 (prec
, m0
, m1
, NULL
);
199 mpfr_from_real (m0
, arg0
, MPFR_RNDN
);
200 mpfr_from_real (m1
, arg1
, MPFR_RNDN
);
202 bool inexact
= func (m0
, m0
, m1
, rnd
);
203 bool ok
= do_mpfr_ckconv (result
, m0
, inexact
, format
);
204 mpfr_clears (m0
, m1
, NULL
);
211 *RESULT = f (ARG0, *ARG1)
213 in format FORMAT, given that FUNC is the MPFR implementation of f.
214 Return true on success. */
217 do_mpfr_arg2 (real_value
*result
,
218 int (*func
) (mpfr_ptr
, long, mpfr_srcptr
, mpfr_rnd_t
),
219 const wide_int_ref
&arg0
, const real_value
*arg1
,
220 const real_format
*format
)
222 if (format
->b
!= 2 || !real_isfinite (arg1
))
225 int prec
= format
->p
;
226 mpfr_rnd_t rnd
= format
->round_towards_zero
? MPFR_RNDZ
: MPFR_RNDN
;
229 mpfr_init2 (m
, prec
);
230 mpfr_from_real (m
, arg1
, MPFR_RNDN
);
232 bool inexact
= func (m
, arg0
.to_shwi (), m
, rnd
);
233 bool ok
= do_mpfr_ckconv (result
, m
, inexact
, format
);
241 *RESULT = f (*ARG0, *ARG1, *ARG2)
243 in format FORMAT, given that FUNC is the MPFR implementation of f.
244 Return true on success. */
247 do_mpfr_arg3 (real_value
*result
,
248 int (*func
) (mpfr_ptr
, mpfr_srcptr
, mpfr_srcptr
,
249 mpfr_srcptr
, mpfr_rnd_t
),
250 const real_value
*arg0
, const real_value
*arg1
,
251 const real_value
*arg2
, const real_format
*format
)
253 /* To proceed, MPFR must exactly represent the target floating point
254 format, which only happens when the target base equals two. */
256 || !real_isfinite (arg0
)
257 || !real_isfinite (arg1
)
258 || !real_isfinite (arg2
))
261 int prec
= format
->p
;
262 mpfr_rnd_t rnd
= format
->round_towards_zero
? MPFR_RNDZ
: MPFR_RNDN
;
265 mpfr_inits2 (prec
, m0
, m1
, m2
, NULL
);
266 mpfr_from_real (m0
, arg0
, MPFR_RNDN
);
267 mpfr_from_real (m1
, arg1
, MPFR_RNDN
);
268 mpfr_from_real (m2
, arg2
, MPFR_RNDN
);
270 bool inexact
= func (m0
, m0
, m1
, m2
, rnd
);
271 bool ok
= do_mpfr_ckconv (result
, m0
, inexact
, format
);
272 mpfr_clears (m0
, m1
, m2
, NULL
);
277 /* M is the result of trying to constant-fold an expression (starting
278 with clear MPFR flags) and INEXACT says whether the result in M is
279 exact or inexact. Return true if M can be used as a constant-folded
280 result in which the real and imaginary parts have format FORMAT.
281 Store those parts in *RESULT_REAL and *RESULT_IMAG if so. */
284 do_mpc_ckconv (real_value
*result_real
, real_value
*result_imag
,
285 mpc_srcptr m
, bool inexact
, const real_format
*format
)
287 /* Proceed iff we get a normal number, i.e. not NaN or Inf and no
288 overflow/underflow occurred. If -frounding-math, proceed iff the
289 result of calling FUNC was exact. */
290 if (!mpfr_number_p (mpc_realref (m
))
291 || !mpfr_number_p (mpc_imagref (m
))
292 || mpfr_overflow_p ()
293 || mpfr_underflow_p ()
294 || (flag_rounding_math
&& inexact
))
297 REAL_VALUE_TYPE tmp_real
, tmp_imag
;
298 real_from_mpfr (&tmp_real
, mpc_realref (m
), format
, MPFR_RNDN
);
299 real_from_mpfr (&tmp_imag
, mpc_imagref (m
), format
, MPFR_RNDN
);
301 /* Proceed iff GCC's REAL_VALUE_TYPE can hold the MPFR values.
302 If the REAL_VALUE_TYPE is zero but the mpft_t is not, then we
303 underflowed in the conversion. */
304 if (!real_isfinite (&tmp_real
)
305 || !real_isfinite (&tmp_imag
)
306 || (tmp_real
.cl
== rvc_zero
) != (mpfr_zero_p (mpc_realref (m
)) != 0)
307 || (tmp_imag
.cl
== rvc_zero
) != (mpfr_zero_p (mpc_imagref (m
)) != 0))
310 real_convert (result_real
, format
, &tmp_real
);
311 real_convert (result_imag
, format
, &tmp_imag
);
313 return (real_identical (result_real
, &tmp_real
)
314 && real_identical (result_imag
, &tmp_imag
));
321 in format FORMAT, given that FUNC is the mpc implementation of f.
322 Return true on success. Both RESULT and ARG are represented as
323 real and imaginary pairs. */
326 do_mpc_arg1 (real_value
*result_real
, real_value
*result_imag
,
327 int (*func
) (mpc_ptr
, mpc_srcptr
, mpc_rnd_t
),
328 const real_value
*arg_real
, const real_value
*arg_imag
,
329 const real_format
*format
)
331 /* To proceed, MPFR must exactly represent the target floating point
332 format, which only happens when the target base equals two. */
334 || !real_isfinite (arg_real
)
335 || !real_isfinite (arg_imag
))
338 int prec
= format
->p
;
339 mpc_rnd_t crnd
= format
->round_towards_zero
? MPC_RNDZZ
: MPC_RNDNN
;
343 mpfr_from_real (mpc_realref (m
), arg_real
, MPFR_RNDN
);
344 mpfr_from_real (mpc_imagref (m
), arg_imag
, MPFR_RNDN
);
346 bool inexact
= func (m
, m
, crnd
);
347 bool ok
= do_mpc_ckconv (result_real
, result_imag
, m
, inexact
, format
);
355 RESULT = f (ARG0, ARG1)
357 in format FORMAT, given that FUNC is the mpc implementation of f.
358 Return true on success. RESULT, ARG0 and ARG1 are represented as
359 real and imaginary pairs. */
362 do_mpc_arg2 (real_value
*result_real
, real_value
*result_imag
,
363 int (*func
)(mpc_ptr
, mpc_srcptr
, mpc_srcptr
, mpc_rnd_t
),
364 const real_value
*arg0_real
, const real_value
*arg0_imag
,
365 const real_value
*arg1_real
, const real_value
*arg1_imag
,
366 const real_format
*format
)
368 if (!real_isfinite (arg0_real
)
369 || !real_isfinite (arg0_imag
)
370 || !real_isfinite (arg1_real
)
371 || !real_isfinite (arg1_imag
))
374 int prec
= format
->p
;
375 mpc_rnd_t crnd
= format
->round_towards_zero
? MPC_RNDZZ
: MPC_RNDNN
;
378 mpc_init2 (m0
, prec
);
379 mpc_init2 (m1
, prec
);
380 mpfr_from_real (mpc_realref (m0
), arg0_real
, MPFR_RNDN
);
381 mpfr_from_real (mpc_imagref (m0
), arg0_imag
, MPFR_RNDN
);
382 mpfr_from_real (mpc_realref (m1
), arg1_real
, MPFR_RNDN
);
383 mpfr_from_real (mpc_imagref (m1
), arg1_imag
, MPFR_RNDN
);
385 bool inexact
= func (m0
, m0
, m1
, crnd
);
386 bool ok
= do_mpc_ckconv (result_real
, result_imag
, m0
, inexact
, format
);
395 *RESULT = logb (*ARG)
397 in format FORMAT. Return true on success. */
400 fold_const_logb (real_value
*result
, const real_value
*arg
,
401 const real_format
*format
)
406 /* If arg is +-NaN, then return it. */
411 /* If arg is +-Inf, then return +Inf. */
417 /* Zero may set errno and/or raise an exception. */
421 /* For normal numbers, proceed iff radix == 2. In GCC,
422 normalized significands are in the range [0.5, 1.0). We
423 want the exponent as if they were [1.0, 2.0) so get the
424 exponent and subtract 1. */
427 real_from_integer (result
, format
, REAL_EXP (arg
) - 1, SIGNED
);
437 *RESULT = significand (*ARG)
439 in format FORMAT. Return true on success. */
442 fold_const_significand (real_value
*result
, const real_value
*arg
,
443 const real_format
*format
)
450 /* If arg is +-0, +-Inf or +-NaN, then return it. */
455 /* For normal numbers, proceed iff radix == 2. */
459 /* In GCC, normalized significands are in the range [0.5, 1.0).
460 We want them to be [1.0, 2.0) so set the exponent to 1. */
461 SET_REAL_EXP (result
, 1);
473 where FORMAT is the format of *ARG and PRECISION is the number of
474 significant bits in the result. Return true on success. */
477 fold_const_conversion (wide_int
*result
,
478 void (*fn
) (real_value
*, format_helper
,
480 const real_value
*arg
, unsigned int precision
,
481 const real_format
*format
)
483 if (!real_isfinite (arg
))
487 fn (&rounded
, format
, arg
);
490 *result
= real_to_integer (&rounded
, &fail
, precision
);
496 *RESULT = pow (*ARG0, *ARG1)
498 in format FORMAT. Return true on success. */
501 fold_const_pow (real_value
*result
, const real_value
*arg0
,
502 const real_value
*arg1
, const real_format
*format
)
504 if (do_mpfr_arg2 (result
, mpfr_pow
, arg0
, arg1
, format
))
507 /* Check for an integer exponent. */
508 REAL_VALUE_TYPE cint1
;
509 HOST_WIDE_INT n1
= real_to_integer (arg1
);
510 real_from_integer (&cint1
, VOIDmode
, n1
, SIGNED
);
511 /* Attempt to evaluate pow at compile-time, unless this should
512 raise an exception. */
513 if (real_identical (arg1
, &cint1
)
515 || (!flag_trapping_math
&& !flag_errno_math
)
516 || !real_equal (arg0
, &dconst0
)))
518 bool inexact
= real_powi (result
, format
, arg0
, n1
);
519 /* Avoid the folding if flag_signaling_nans is on. */
520 if (flag_unsafe_math_optimizations
522 && !(flag_signaling_nans
523 && REAL_VALUE_ISSIGNALING_NAN (*arg0
))))
532 *RESULT = nextafter (*ARG0, *ARG1)
536 *RESULT = nexttoward (*ARG0, *ARG1)
538 in format FORMAT. Return true on success. */
541 fold_const_nextafter (real_value
*result
, const real_value
*arg0
,
542 const real_value
*arg1
, const real_format
*format
)
544 if (REAL_VALUE_ISSIGNALING_NAN (*arg0
)
545 || REAL_VALUE_ISSIGNALING_NAN (*arg1
))
548 /* Don't handle composite modes, nor decimal, nor modes without
549 inf or denorm at least for now. */
550 if (format
->pnan
< format
->p
553 || !format
->has_denorm
)
556 if (real_nextafter (result
, format
, arg0
, arg1
)
557 /* If raising underflow or overflow and setting errno to ERANGE,
558 fail if we care about those side-effects. */
559 && (flag_trapping_math
|| flag_errno_math
))
561 /* Similarly for nextafter (0, 1) raising underflow. */
562 else if (flag_trapping_math
563 && arg0
->cl
== rvc_zero
564 && result
->cl
!= rvc_zero
)
567 real_convert (result
, format
, result
);
574 *RESULT = ldexp (*ARG0, ARG1)
576 in format FORMAT. Return true on success. */
579 fold_const_builtin_load_exponent (real_value
*result
, const real_value
*arg0
,
580 const wide_int_ref
&arg1
,
581 const real_format
*format
)
583 /* Bound the maximum adjustment to twice the range of the
584 mode's valid exponents. Use abs to ensure the range is
585 positive as a sanity check. */
586 int max_exp_adj
= 2 * labs (format
->emax
- format
->emin
);
588 /* The requested adjustment must be inside this range. This
589 is a preliminary cap to avoid things like overflow, we
590 may still fail to compute the result for other reasons. */
591 if (wi::les_p (arg1
, -max_exp_adj
) || wi::ges_p (arg1
, max_exp_adj
))
594 /* Don't perform operation if we honor signaling NaNs and
595 operand is a signaling NaN. */
596 if (!flag_unsafe_math_optimizations
597 && flag_signaling_nans
598 && REAL_VALUE_ISSIGNALING_NAN (*arg0
))
601 REAL_VALUE_TYPE initial_result
;
602 real_ldexp (&initial_result
, arg0
, arg1
.to_shwi ());
604 /* Ensure we didn't overflow. */
605 if (real_isinf (&initial_result
))
608 /* Only proceed if the target mode can hold the
610 *result
= real_value_truncate (format
, initial_result
);
611 return real_equal (&initial_result
, result
);
614 /* Fold a call to __builtin_nan or __builtin_nans with argument ARG and
615 return type TYPE. QUIET is true if a quiet rather than signalling
619 fold_const_builtin_nan (tree type
, tree arg
, bool quiet
)
621 REAL_VALUE_TYPE real
;
622 const char *str
= c_getstr (arg
);
623 if (str
&& real_nan (&real
, str
, quiet
, TYPE_MODE (type
)))
624 return build_real (type
, real
);
628 /* Fold a call to IFN_REDUC_<CODE> (ARG), returning a value of type TYPE. */
631 fold_const_reduction (tree type
, tree arg
, tree_code code
)
633 unsigned HOST_WIDE_INT nelts
;
634 if (TREE_CODE (arg
) != VECTOR_CST
635 || !VECTOR_CST_NELTS (arg
).is_constant (&nelts
))
638 tree res
= VECTOR_CST_ELT (arg
, 0);
639 for (unsigned HOST_WIDE_INT i
= 1; i
< nelts
; i
++)
641 res
= const_binop (code
, type
, res
, VECTOR_CST_ELT (arg
, i
));
642 if (res
== NULL_TREE
|| !CONSTANT_CLASS_P (res
))
648 /* Fold a call to IFN_VEC_CONVERT (ARG) returning TYPE. */
651 fold_const_vec_convert (tree ret_type
, tree arg
)
653 enum tree_code code
= NOP_EXPR
;
654 tree arg_type
= TREE_TYPE (arg
);
655 if (TREE_CODE (arg
) != VECTOR_CST
)
658 gcc_checking_assert (VECTOR_TYPE_P (ret_type
) && VECTOR_TYPE_P (arg_type
));
660 if (INTEGRAL_TYPE_P (TREE_TYPE (ret_type
))
661 && SCALAR_FLOAT_TYPE_P (TREE_TYPE (arg_type
)))
662 code
= FIX_TRUNC_EXPR
;
663 else if (INTEGRAL_TYPE_P (TREE_TYPE (arg_type
))
664 && SCALAR_FLOAT_TYPE_P (TREE_TYPE (ret_type
)))
667 /* We can't handle steps directly when extending, since the
668 values need to wrap at the original precision first. */
670 = (INTEGRAL_TYPE_P (TREE_TYPE (ret_type
))
671 && INTEGRAL_TYPE_P (TREE_TYPE (arg_type
))
672 && (TYPE_PRECISION (TREE_TYPE (ret_type
))
673 <= TYPE_PRECISION (TREE_TYPE (arg_type
))));
674 tree_vector_builder elts
;
675 if (!elts
.new_unary_operation (ret_type
, arg
, step_ok_p
))
678 unsigned int count
= elts
.encoded_nelts ();
679 for (unsigned int i
= 0; i
< count
; ++i
)
681 tree elt
= fold_unary (code
, TREE_TYPE (ret_type
),
682 VECTOR_CST_ELT (arg
, i
));
683 if (elt
== NULL_TREE
|| !CONSTANT_CLASS_P (elt
))
685 elts
.quick_push (elt
);
688 return elts
.build ();
693 IFN_WHILE_ULT (ARG0, ARG1, (TYPE) { ... })
695 Return the value on success and null on failure. */
698 fold_while_ult (tree type
, poly_uint64 arg0
, poly_uint64 arg1
)
700 if (known_ge (arg0
, arg1
))
701 return build_zero_cst (type
);
703 if (maybe_ge (arg0
, arg1
))
706 poly_uint64 diff
= arg1
- arg0
;
707 poly_uint64 nelts
= TYPE_VECTOR_SUBPARTS (type
);
708 if (known_ge (diff
, nelts
))
709 return build_all_ones_cst (type
);
711 unsigned HOST_WIDE_INT const_diff
;
712 if (known_le (diff
, nelts
) && diff
.is_constant (&const_diff
))
714 tree minus_one
= build_minus_one_cst (TREE_TYPE (type
));
715 tree zero
= build_zero_cst (TREE_TYPE (type
));
716 return build_vector_a_then_b (type
, const_diff
, minus_one
, zero
);
725 in format FORMAT. Return true on success. */
728 fold_const_call_ss (real_value
*result
, combined_fn fn
,
729 const real_value
*arg
, const real_format
*format
)
735 return (real_compare (GE_EXPR
, arg
, &dconst0
)
736 && do_mpfr_arg1 (result
, mpfr_sqrt
, arg
, format
));
739 return do_mpfr_arg1 (result
, mpfr_cbrt
, arg
, format
);
742 return (real_compare (GE_EXPR
, arg
, &dconstm1
)
743 && real_compare (LE_EXPR
, arg
, &dconst1
)
744 && do_mpfr_arg1 (result
, mpfr_asin
, arg
, format
));
747 return (real_compare (GE_EXPR
, arg
, &dconstm1
)
748 && real_compare (LE_EXPR
, arg
, &dconst1
)
749 && do_mpfr_arg1 (result
, mpfr_acos
, arg
, format
));
752 return do_mpfr_arg1 (result
, mpfr_atan
, arg
, format
);
755 return do_mpfr_arg1 (result
, mpfr_asinh
, arg
, format
);
758 return (real_compare (GE_EXPR
, arg
, &dconst1
)
759 && do_mpfr_arg1 (result
, mpfr_acosh
, arg
, format
));
762 return (real_compare (GE_EXPR
, arg
, &dconstm1
)
763 && real_compare (LE_EXPR
, arg
, &dconst1
)
764 && do_mpfr_arg1 (result
, mpfr_atanh
, arg
, format
));
767 return do_mpfr_arg1 (result
, mpfr_sin
, arg
, format
);
770 return do_mpfr_arg1 (result
, mpfr_cos
, arg
, format
);
773 return do_mpfr_arg1 (result
, mpfr_tan
, arg
, format
);
776 return do_mpfr_arg1 (result
, mpfr_sinh
, arg
, format
);
779 return do_mpfr_arg1 (result
, mpfr_cosh
, arg
, format
);
782 return do_mpfr_arg1 (result
, mpfr_tanh
, arg
, format
);
785 return do_mpfr_arg1 (result
, mpfr_erf
, arg
, format
);
788 return do_mpfr_arg1 (result
, mpfr_erfc
, arg
, format
);
791 return do_mpfr_arg1 (result
, mpfr_gamma
, arg
, format
);
794 return do_mpfr_arg1 (result
, mpfr_exp
, arg
, format
);
797 return do_mpfr_arg1 (result
, mpfr_exp2
, arg
, format
);
801 return do_mpfr_arg1 (result
, mpfr_exp10
, arg
, format
);
804 return do_mpfr_arg1 (result
, mpfr_expm1
, arg
, format
);
807 return (real_compare (GT_EXPR
, arg
, &dconst0
)
808 && do_mpfr_arg1 (result
, mpfr_log
, arg
, format
));
811 return (real_compare (GT_EXPR
, arg
, &dconst0
)
812 && do_mpfr_arg1 (result
, mpfr_log2
, arg
, format
));
815 return (real_compare (GT_EXPR
, arg
, &dconst0
)
816 && do_mpfr_arg1 (result
, mpfr_log10
, arg
, format
));
819 return (real_compare (GT_EXPR
, arg
, &dconstm1
)
820 && do_mpfr_arg1 (result
, mpfr_log1p
, arg
, format
));
823 return do_mpfr_arg1 (result
, mpfr_j0
, arg
, format
);
826 return do_mpfr_arg1 (result
, mpfr_j1
, arg
, format
);
829 return (real_compare (GT_EXPR
, arg
, &dconst0
)
830 && do_mpfr_arg1 (result
, mpfr_y0
, arg
, format
));
833 return (real_compare (GT_EXPR
, arg
, &dconst0
)
834 && do_mpfr_arg1 (result
, mpfr_y1
, arg
, format
));
838 if (!REAL_VALUE_ISSIGNALING_NAN (*arg
))
840 real_floor (result
, format
, arg
);
847 if (!REAL_VALUE_ISSIGNALING_NAN (*arg
))
849 real_ceil (result
, format
, arg
);
856 if (!REAL_VALUE_ISSIGNALING_NAN (*arg
))
858 real_trunc (result
, format
, arg
);
865 if (!REAL_VALUE_ISSIGNALING_NAN (*arg
))
867 real_round (result
, format
, arg
);
873 CASE_CFN_ROUNDEVEN_FN
:
874 if (!REAL_VALUE_ISSIGNALING_NAN (*arg
))
876 real_roundeven (result
, format
, arg
);
882 return fold_const_logb (result
, arg
, format
);
884 CASE_CFN_SIGNIFICAND
:
885 return fold_const_significand (result
, arg
, format
);
896 where FORMAT is the format of ARG and PRECISION is the number of
897 significant bits in the result. Return true on success. */
900 fold_const_call_ss (wide_int
*result
, combined_fn fn
,
901 const real_value
*arg
, unsigned int precision
,
902 const real_format
*format
)
907 if (real_isneg (arg
))
908 *result
= wi::one (precision
);
910 *result
= wi::zero (precision
);
914 /* For ilogb we don't know FP_ILOGB0, so only handle normal values.
915 Proceed iff radix == 2. In GCC, normalized significands are in
916 the range [0.5, 1.0). We want the exponent as if they were
917 [1.0, 2.0) so get the exponent and subtract 1. */
918 if (arg
->cl
== rvc_normal
&& format
->b
== 2)
920 *result
= wi::shwi (REAL_EXP (arg
) - 1, precision
);
928 return fold_const_conversion (result
, real_ceil
, arg
,
934 return fold_const_conversion (result
, real_floor
, arg
,
940 return fold_const_conversion (result
, real_round
, arg
,
946 /* Not yet folded to a constant. */
950 case CFN_BUILT_IN_FINITED32
:
951 case CFN_BUILT_IN_FINITED64
:
952 case CFN_BUILT_IN_FINITED128
:
953 case CFN_BUILT_IN_ISFINITE
:
954 *result
= wi::shwi (real_isfinite (arg
) ? 1 : 0, precision
);
958 case CFN_BUILT_IN_ISINFD32
:
959 case CFN_BUILT_IN_ISINFD64
:
960 case CFN_BUILT_IN_ISINFD128
:
961 if (real_isinf (arg
))
962 *result
= wi::shwi (arg
->sign
? -1 : 1, precision
);
964 *result
= wi::shwi (0, precision
);
968 case CFN_BUILT_IN_ISNAND32
:
969 case CFN_BUILT_IN_ISNAND64
:
970 case CFN_BUILT_IN_ISNAND128
:
971 *result
= wi::shwi (real_isnan (arg
) ? 1 : 0, precision
);
983 where ARG_TYPE is the type of ARG and PRECISION is the number of bits
984 in the result. Return true on success. */
987 fold_const_call_ss (wide_int
*result
, combined_fn fn
, const wide_int_ref
&arg
,
988 unsigned int precision
, tree arg_type
)
993 *result
= wi::shwi (wi::ffs (arg
), precision
);
999 if (wi::ne_p (arg
, 0))
1000 tmp
= wi::clz (arg
);
1001 else if (!CLZ_DEFINED_VALUE_AT_ZERO (SCALAR_INT_TYPE_MODE (arg_type
),
1003 tmp
= TYPE_PRECISION (arg_type
);
1004 *result
= wi::shwi (tmp
, precision
);
1011 if (wi::ne_p (arg
, 0))
1012 tmp
= wi::ctz (arg
);
1013 else if (!CTZ_DEFINED_VALUE_AT_ZERO (SCALAR_INT_TYPE_MODE (arg_type
),
1015 tmp
= TYPE_PRECISION (arg_type
);
1016 *result
= wi::shwi (tmp
, precision
);
1021 *result
= wi::shwi (wi::clrsb (arg
), precision
);
1025 *result
= wi::shwi (wi::popcount (arg
), precision
);
1029 *result
= wi::shwi (wi::parity (arg
), precision
);
1032 case CFN_BUILT_IN_BSWAP16
:
1033 case CFN_BUILT_IN_BSWAP32
:
1034 case CFN_BUILT_IN_BSWAP64
:
1035 case CFN_BUILT_IN_BSWAP128
:
1036 *result
= wide_int::from (arg
, precision
, TYPE_SIGN (arg_type
)).bswap ();
1048 where FORMAT is the format of ARG and of the real and imaginary parts
1049 of RESULT, passed as RESULT_REAL and RESULT_IMAG respectively. Return
1053 fold_const_call_cs (real_value
*result_real
, real_value
*result_imag
,
1054 combined_fn fn
, const real_value
*arg
,
1055 const real_format
*format
)
1060 /* cexpi(x+yi) = cos(x)+sin(y)*i. */
1061 return do_mpfr_sincos (result_imag
, result_real
, arg
, format
);
1072 where FORMAT is the format of RESULT and of the real and imaginary parts
1073 of ARG, passed as ARG_REAL and ARG_IMAG respectively. Return true on
1077 fold_const_call_sc (real_value
*result
, combined_fn fn
,
1078 const real_value
*arg_real
, const real_value
*arg_imag
,
1079 const real_format
*format
)
1084 return do_mpfr_arg2 (result
, mpfr_hypot
, arg_real
, arg_imag
, format
);
1095 where FORMAT is the format of the real and imaginary parts of RESULT
1096 (RESULT_REAL and RESULT_IMAG) and of ARG (ARG_REAL and ARG_IMAG).
1097 Return true on success. */
1100 fold_const_call_cc (real_value
*result_real
, real_value
*result_imag
,
1101 combined_fn fn
, const real_value
*arg_real
,
1102 const real_value
*arg_imag
, const real_format
*format
)
1107 return do_mpc_arg1 (result_real
, result_imag
, mpc_cos
,
1108 arg_real
, arg_imag
, format
);
1111 return do_mpc_arg1 (result_real
, result_imag
, mpc_cosh
,
1112 arg_real
, arg_imag
, format
);
1115 if (real_isinf (arg_real
) || real_isinf (arg_imag
))
1117 real_inf (result_real
);
1118 *result_imag
= dconst0
;
1119 result_imag
->sign
= arg_imag
->sign
;
1123 *result_real
= *arg_real
;
1124 *result_imag
= *arg_imag
;
1129 return do_mpc_arg1 (result_real
, result_imag
, mpc_sin
,
1130 arg_real
, arg_imag
, format
);
1133 return do_mpc_arg1 (result_real
, result_imag
, mpc_sinh
,
1134 arg_real
, arg_imag
, format
);
1137 return do_mpc_arg1 (result_real
, result_imag
, mpc_tan
,
1138 arg_real
, arg_imag
, format
);
1141 return do_mpc_arg1 (result_real
, result_imag
, mpc_tanh
,
1142 arg_real
, arg_imag
, format
);
1145 return do_mpc_arg1 (result_real
, result_imag
, mpc_log
,
1146 arg_real
, arg_imag
, format
);
1149 return do_mpc_arg1 (result_real
, result_imag
, mpc_sqrt
,
1150 arg_real
, arg_imag
, format
);
1153 return do_mpc_arg1 (result_real
, result_imag
, mpc_asin
,
1154 arg_real
, arg_imag
, format
);
1157 return do_mpc_arg1 (result_real
, result_imag
, mpc_acos
,
1158 arg_real
, arg_imag
, format
);
1161 return do_mpc_arg1 (result_real
, result_imag
, mpc_atan
,
1162 arg_real
, arg_imag
, format
);
1165 return do_mpc_arg1 (result_real
, result_imag
, mpc_asinh
,
1166 arg_real
, arg_imag
, format
);
1169 return do_mpc_arg1 (result_real
, result_imag
, mpc_acosh
,
1170 arg_real
, arg_imag
, format
);
1173 return do_mpc_arg1 (result_real
, result_imag
, mpc_atanh
,
1174 arg_real
, arg_imag
, format
);
1177 return do_mpc_arg1 (result_real
, result_imag
, mpc_exp
,
1178 arg_real
, arg_imag
, format
);
1185 /* Subroutine of fold_const_call, with the same interface. Handle cases
1186 where the arguments and result are numerical. */
1189 fold_const_call_1 (combined_fn fn
, tree type
, tree arg
)
1191 machine_mode mode
= TYPE_MODE (type
);
1192 machine_mode arg_mode
= TYPE_MODE (TREE_TYPE (arg
));
1194 if (integer_cst_p (arg
))
1196 if (SCALAR_INT_MODE_P (mode
))
1199 if (fold_const_call_ss (&result
, fn
, wi::to_wide (arg
),
1200 TYPE_PRECISION (type
), TREE_TYPE (arg
)))
1201 return wide_int_to_tree (type
, result
);
1206 if (real_cst_p (arg
))
1208 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg_mode
));
1209 if (mode
== arg_mode
)
1212 REAL_VALUE_TYPE result
;
1213 if (fold_const_call_ss (&result
, fn
, TREE_REAL_CST_PTR (arg
),
1214 REAL_MODE_FORMAT (mode
)))
1215 return build_real (type
, result
);
1217 else if (COMPLEX_MODE_P (mode
)
1218 && GET_MODE_INNER (mode
) == arg_mode
)
1220 /* real -> complex real. */
1221 REAL_VALUE_TYPE result_real
, result_imag
;
1222 if (fold_const_call_cs (&result_real
, &result_imag
, fn
,
1223 TREE_REAL_CST_PTR (arg
),
1224 REAL_MODE_FORMAT (arg_mode
)))
1225 return build_complex (type
,
1226 build_real (TREE_TYPE (type
), result_real
),
1227 build_real (TREE_TYPE (type
), result_imag
));
1229 else if (INTEGRAL_TYPE_P (type
))
1233 if (fold_const_call_ss (&result
, fn
,
1234 TREE_REAL_CST_PTR (arg
),
1235 TYPE_PRECISION (type
),
1236 REAL_MODE_FORMAT (arg_mode
)))
1237 return wide_int_to_tree (type
, result
);
1242 if (complex_cst_p (arg
))
1244 gcc_checking_assert (COMPLEX_MODE_P (arg_mode
));
1245 machine_mode inner_mode
= GET_MODE_INNER (arg_mode
);
1246 tree argr
= TREE_REALPART (arg
);
1247 tree argi
= TREE_IMAGPART (arg
);
1248 if (mode
== arg_mode
1249 && real_cst_p (argr
)
1250 && real_cst_p (argi
))
1252 /* complex real -> complex real. */
1253 REAL_VALUE_TYPE result_real
, result_imag
;
1254 if (fold_const_call_cc (&result_real
, &result_imag
, fn
,
1255 TREE_REAL_CST_PTR (argr
),
1256 TREE_REAL_CST_PTR (argi
),
1257 REAL_MODE_FORMAT (inner_mode
)))
1258 return build_complex (type
,
1259 build_real (TREE_TYPE (type
), result_real
),
1260 build_real (TREE_TYPE (type
), result_imag
));
1262 if (mode
== inner_mode
1263 && real_cst_p (argr
)
1264 && real_cst_p (argi
))
1266 /* complex real -> real. */
1267 REAL_VALUE_TYPE result
;
1268 if (fold_const_call_sc (&result
, fn
,
1269 TREE_REAL_CST_PTR (argr
),
1270 TREE_REAL_CST_PTR (argi
),
1271 REAL_MODE_FORMAT (inner_mode
)))
1272 return build_real (type
, result
);
1280 /* Try to fold FN (ARG) to a constant. Return the constant on success,
1281 otherwise return null. TYPE is the type of the return value. */
1284 fold_const_call (combined_fn fn
, tree type
, tree arg
)
1288 case CFN_BUILT_IN_STRLEN
:
1289 if (const char *str
= c_getstr (arg
))
1290 return build_int_cst (type
, strlen (str
));
1294 CASE_FLT_FN_FLOATN_NX (CFN_BUILT_IN_NAN
):
1295 case CFN_BUILT_IN_NAND32
:
1296 case CFN_BUILT_IN_NAND64
:
1297 case CFN_BUILT_IN_NAND128
:
1298 return fold_const_builtin_nan (type
, arg
, true);
1301 CASE_FLT_FN_FLOATN_NX (CFN_BUILT_IN_NANS
):
1302 case CFN_BUILT_IN_NANSD32
:
1303 case CFN_BUILT_IN_NANSD64
:
1304 case CFN_BUILT_IN_NANSD128
:
1305 return fold_const_builtin_nan (type
, arg
, false);
1307 case CFN_REDUC_PLUS
:
1308 return fold_const_reduction (type
, arg
, PLUS_EXPR
);
1311 return fold_const_reduction (type
, arg
, MAX_EXPR
);
1314 return fold_const_reduction (type
, arg
, MIN_EXPR
);
1317 return fold_const_reduction (type
, arg
, BIT_AND_EXPR
);
1320 return fold_const_reduction (type
, arg
, BIT_IOR_EXPR
);
1323 return fold_const_reduction (type
, arg
, BIT_XOR_EXPR
);
1325 case CFN_VEC_CONVERT
:
1326 return fold_const_vec_convert (type
, arg
);
1329 return fold_const_call_1 (fn
, type
, arg
);
1333 /* Fold a call to IFN_FOLD_LEFT_<CODE> (ARG0, ARG1), returning a value
1337 fold_const_fold_left (tree type
, tree arg0
, tree arg1
, tree_code code
)
1339 if (TREE_CODE (arg1
) != VECTOR_CST
)
1342 unsigned HOST_WIDE_INT nelts
;
1343 if (!VECTOR_CST_NELTS (arg1
).is_constant (&nelts
))
1346 for (unsigned HOST_WIDE_INT i
= 0; i
< nelts
; i
++)
1348 arg0
= const_binop (code
, type
, arg0
, VECTOR_CST_ELT (arg1
, i
));
1349 if (arg0
== NULL_TREE
|| !CONSTANT_CLASS_P (arg0
))
1357 *RESULT = FN (*ARG0, *ARG1)
1359 in format FORMAT. Return true on success. */
1362 fold_const_call_sss (real_value
*result
, combined_fn fn
,
1363 const real_value
*arg0
, const real_value
*arg1
,
1364 const real_format
*format
)
1370 return do_mpfr_arg2 (result
, mpfr_remainder
, arg0
, arg1
, format
);
1373 return do_mpfr_arg2 (result
, mpfr_atan2
, arg0
, arg1
, format
);
1376 return do_mpfr_arg2 (result
, mpfr_dim
, arg0
, arg1
, format
);
1379 return do_mpfr_arg2 (result
, mpfr_fmod
, arg0
, arg1
, format
);
1382 return do_mpfr_arg2 (result
, mpfr_hypot
, arg0
, arg1
, format
);
1385 CASE_CFN_COPYSIGN_FN
:
1387 real_copysign (result
, arg1
);
1392 return do_mpfr_arg2 (result
, mpfr_min
, arg0
, arg1
, format
);
1396 return do_mpfr_arg2 (result
, mpfr_max
, arg0
, arg1
, format
);
1399 return fold_const_pow (result
, arg0
, arg1
, format
);
1402 CASE_CFN_NEXTTOWARD
:
1403 return fold_const_nextafter (result
, arg0
, arg1
, format
);
1412 *RESULT = FN (*ARG0, ARG1)
1414 where FORMAT is the format of *RESULT and *ARG0. Return true on
1418 fold_const_call_sss (real_value
*result
, combined_fn fn
,
1419 const real_value
*arg0
, const wide_int_ref
&arg1
,
1420 const real_format
*format
)
1425 return fold_const_builtin_load_exponent (result
, arg0
, arg1
, format
);
1429 return (format
->b
== 2
1430 && fold_const_builtin_load_exponent (result
, arg0
, arg1
,
1434 /* Avoid the folding if flag_signaling_nans is on and
1435 operand is a signaling NaN. */
1436 if (!flag_unsafe_math_optimizations
1437 && flag_signaling_nans
1438 && REAL_VALUE_ISSIGNALING_NAN (*arg0
))
1441 real_powi (result
, format
, arg0
, arg1
.to_shwi ());
1451 *RESULT = FN (ARG0, *ARG1)
1453 where FORMAT is the format of *RESULT and *ARG1. Return true on
1457 fold_const_call_sss (real_value
*result
, combined_fn fn
,
1458 const wide_int_ref
&arg0
, const real_value
*arg1
,
1459 const real_format
*format
)
1464 return do_mpfr_arg2 (result
, mpfr_jn
, arg0
, arg1
, format
);
1467 return (real_compare (GT_EXPR
, arg1
, &dconst0
)
1468 && do_mpfr_arg2 (result
, mpfr_yn
, arg0
, arg1
, format
));
1477 RESULT = fn (ARG0, ARG1)
1479 where FORMAT is the format of the real and imaginary parts of RESULT
1480 (RESULT_REAL and RESULT_IMAG), of ARG0 (ARG0_REAL and ARG0_IMAG)
1481 and of ARG1 (ARG1_REAL and ARG1_IMAG). Return true on success. */
1484 fold_const_call_ccc (real_value
*result_real
, real_value
*result_imag
,
1485 combined_fn fn
, const real_value
*arg0_real
,
1486 const real_value
*arg0_imag
, const real_value
*arg1_real
,
1487 const real_value
*arg1_imag
, const real_format
*format
)
1492 return do_mpc_arg2 (result_real
, result_imag
, mpc_pow
,
1493 arg0_real
, arg0_imag
, arg1_real
, arg1_imag
, format
);
1500 /* Subroutine of fold_const_call, with the same interface. Handle cases
1501 where the arguments and result are numerical. */
1504 fold_const_call_1 (combined_fn fn
, tree type
, tree arg0
, tree arg1
)
1506 machine_mode mode
= TYPE_MODE (type
);
1507 machine_mode arg0_mode
= TYPE_MODE (TREE_TYPE (arg0
));
1508 machine_mode arg1_mode
= TYPE_MODE (TREE_TYPE (arg1
));
1510 if (mode
== arg0_mode
1511 && real_cst_p (arg0
)
1512 && real_cst_p (arg1
))
1514 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg0_mode
));
1515 REAL_VALUE_TYPE result
;
1516 if (arg0_mode
== arg1_mode
)
1518 /* real, real -> real. */
1519 if (fold_const_call_sss (&result
, fn
, TREE_REAL_CST_PTR (arg0
),
1520 TREE_REAL_CST_PTR (arg1
),
1521 REAL_MODE_FORMAT (mode
)))
1522 return build_real (type
, result
);
1524 else if (arg1_mode
== TYPE_MODE (long_double_type_node
))
1527 CASE_CFN_NEXTTOWARD
:
1528 /* real, long double -> real. */
1529 if (fold_const_call_sss (&result
, fn
, TREE_REAL_CST_PTR (arg0
),
1530 TREE_REAL_CST_PTR (arg1
),
1531 REAL_MODE_FORMAT (mode
)))
1532 return build_real (type
, result
);
1540 if (real_cst_p (arg0
)
1541 && integer_cst_p (arg1
))
1543 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg0_mode
));
1544 if (mode
== arg0_mode
)
1546 /* real, int -> real. */
1547 REAL_VALUE_TYPE result
;
1548 if (fold_const_call_sss (&result
, fn
, TREE_REAL_CST_PTR (arg0
),
1550 REAL_MODE_FORMAT (mode
)))
1551 return build_real (type
, result
);
1556 if (integer_cst_p (arg0
)
1557 && real_cst_p (arg1
))
1559 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg1_mode
));
1560 if (mode
== arg1_mode
)
1562 /* int, real -> real. */
1563 REAL_VALUE_TYPE result
;
1564 if (fold_const_call_sss (&result
, fn
, wi::to_wide (arg0
),
1565 TREE_REAL_CST_PTR (arg1
),
1566 REAL_MODE_FORMAT (mode
)))
1567 return build_real (type
, result
);
1572 if (arg0_mode
== arg1_mode
1573 && complex_cst_p (arg0
)
1574 && complex_cst_p (arg1
))
1576 gcc_checking_assert (COMPLEX_MODE_P (arg0_mode
));
1577 machine_mode inner_mode
= GET_MODE_INNER (arg0_mode
);
1578 tree arg0r
= TREE_REALPART (arg0
);
1579 tree arg0i
= TREE_IMAGPART (arg0
);
1580 tree arg1r
= TREE_REALPART (arg1
);
1581 tree arg1i
= TREE_IMAGPART (arg1
);
1582 if (mode
== arg0_mode
1583 && real_cst_p (arg0r
)
1584 && real_cst_p (arg0i
)
1585 && real_cst_p (arg1r
)
1586 && real_cst_p (arg1i
))
1588 /* complex real, complex real -> complex real. */
1589 REAL_VALUE_TYPE result_real
, result_imag
;
1590 if (fold_const_call_ccc (&result_real
, &result_imag
, fn
,
1591 TREE_REAL_CST_PTR (arg0r
),
1592 TREE_REAL_CST_PTR (arg0i
),
1593 TREE_REAL_CST_PTR (arg1r
),
1594 TREE_REAL_CST_PTR (arg1i
),
1595 REAL_MODE_FORMAT (inner_mode
)))
1596 return build_complex (type
,
1597 build_real (TREE_TYPE (type
), result_real
),
1598 build_real (TREE_TYPE (type
), result_imag
));
1606 /* Try to fold FN (ARG0, ARG1) to a constant. Return the constant on success,
1607 otherwise return null. TYPE is the type of the return value. */
1610 fold_const_call (combined_fn fn
, tree type
, tree arg0
, tree arg1
)
1612 const char *p0
, *p1
;
1616 case CFN_BUILT_IN_STRSPN
:
1617 if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1618 return build_int_cst (type
, strspn (p0
, p1
));
1621 case CFN_BUILT_IN_STRCSPN
:
1622 if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1623 return build_int_cst (type
, strcspn (p0
, p1
));
1626 case CFN_BUILT_IN_STRCMP
:
1627 if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1628 return build_cmp_result (type
, strcmp (p0
, p1
));
1631 case CFN_BUILT_IN_STRCASECMP
:
1632 if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1634 int r
= strcmp (p0
, p1
);
1636 return build_cmp_result (type
, r
);
1640 case CFN_BUILT_IN_INDEX
:
1641 case CFN_BUILT_IN_STRCHR
:
1642 if ((p0
= c_getstr (arg0
)) && target_char_cst_p (arg1
, &c
))
1644 const char *r
= strchr (p0
, c
);
1646 return build_int_cst (type
, 0);
1647 return fold_convert (type
,
1648 fold_build_pointer_plus_hwi (arg0
, r
- p0
));
1652 case CFN_BUILT_IN_RINDEX
:
1653 case CFN_BUILT_IN_STRRCHR
:
1654 if ((p0
= c_getstr (arg0
)) && target_char_cst_p (arg1
, &c
))
1656 const char *r
= strrchr (p0
, c
);
1658 return build_int_cst (type
, 0);
1659 return fold_convert (type
,
1660 fold_build_pointer_plus_hwi (arg0
, r
- p0
));
1664 case CFN_BUILT_IN_STRSTR
:
1665 if ((p1
= c_getstr (arg1
)))
1667 if ((p0
= c_getstr (arg0
)))
1669 const char *r
= strstr (p0
, p1
);
1671 return build_int_cst (type
, 0);
1672 return fold_convert (type
,
1673 fold_build_pointer_plus_hwi (arg0
, r
- p0
));
1676 return fold_convert (type
, arg0
);
1680 case CFN_FOLD_LEFT_PLUS
:
1681 return fold_const_fold_left (type
, arg0
, arg1
, PLUS_EXPR
);
1684 return fold_const_call_1 (fn
, type
, arg0
, arg1
);
1690 *RESULT = FN (*ARG0, *ARG1, *ARG2)
1692 in format FORMAT. Return true on success. */
1695 fold_const_call_ssss (real_value
*result
, combined_fn fn
,
1696 const real_value
*arg0
, const real_value
*arg1
,
1697 const real_value
*arg2
, const real_format
*format
)
1703 return do_mpfr_arg3 (result
, mpfr_fma
, arg0
, arg1
, arg2
, format
);
1707 real_value new_arg2
= real_value_negate (arg2
);
1708 return do_mpfr_arg3 (result
, mpfr_fma
, arg0
, arg1
, &new_arg2
, format
);
1713 real_value new_arg0
= real_value_negate (arg0
);
1714 return do_mpfr_arg3 (result
, mpfr_fma
, &new_arg0
, arg1
, arg2
, format
);
1719 real_value new_arg0
= real_value_negate (arg0
);
1720 real_value new_arg2
= real_value_negate (arg2
);
1721 return do_mpfr_arg3 (result
, mpfr_fma
, &new_arg0
, arg1
,
1730 /* Subroutine of fold_const_call, with the same interface. Handle cases
1731 where the arguments and result are numerical. */
1734 fold_const_call_1 (combined_fn fn
, tree type
, tree arg0
, tree arg1
, tree arg2
)
1736 machine_mode mode
= TYPE_MODE (type
);
1737 machine_mode arg0_mode
= TYPE_MODE (TREE_TYPE (arg0
));
1738 machine_mode arg1_mode
= TYPE_MODE (TREE_TYPE (arg1
));
1739 machine_mode arg2_mode
= TYPE_MODE (TREE_TYPE (arg2
));
1741 if (arg0_mode
== arg1_mode
1742 && arg0_mode
== arg2_mode
1743 && real_cst_p (arg0
)
1744 && real_cst_p (arg1
)
1745 && real_cst_p (arg2
))
1747 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg0_mode
));
1748 if (mode
== arg0_mode
)
1750 /* real, real, real -> real. */
1751 REAL_VALUE_TYPE result
;
1752 if (fold_const_call_ssss (&result
, fn
, TREE_REAL_CST_PTR (arg0
),
1753 TREE_REAL_CST_PTR (arg1
),
1754 TREE_REAL_CST_PTR (arg2
),
1755 REAL_MODE_FORMAT (mode
)))
1756 return build_real (type
, result
);
1764 /* Try to fold FN (ARG0, ARG1, ARG2) to a constant. Return the constant on
1765 success, otherwise return null. TYPE is the type of the return value. */
1768 fold_const_call (combined_fn fn
, tree type
, tree arg0
, tree arg1
, tree arg2
)
1770 const char *p0
, *p1
;
1772 unsigned HOST_WIDE_INT s0
, s1
, s2
= 0;
1775 case CFN_BUILT_IN_STRNCMP
:
1776 if (!size_t_cst_p (arg2
, &s2
))
1779 && !TREE_SIDE_EFFECTS (arg0
)
1780 && !TREE_SIDE_EFFECTS (arg1
))
1781 return build_int_cst (type
, 0);
1782 else if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1783 return build_int_cst (type
, strncmp (p0
, p1
, MIN (s2
, SIZE_MAX
)));
1786 case CFN_BUILT_IN_STRNCASECMP
:
1787 if (!size_t_cst_p (arg2
, &s2
))
1790 && !TREE_SIDE_EFFECTS (arg0
)
1791 && !TREE_SIDE_EFFECTS (arg1
))
1792 return build_int_cst (type
, 0);
1793 else if ((p0
= c_getstr (arg0
))
1794 && (p1
= c_getstr (arg1
))
1795 && strncmp (p0
, p1
, MIN (s2
, SIZE_MAX
)) == 0)
1796 return build_int_cst (type
, 0);
1799 case CFN_BUILT_IN_BCMP
:
1800 case CFN_BUILT_IN_MEMCMP
:
1801 if (!size_t_cst_p (arg2
, &s2
))
1804 && !TREE_SIDE_EFFECTS (arg0
)
1805 && !TREE_SIDE_EFFECTS (arg1
))
1806 return build_int_cst (type
, 0);
1807 if ((p0
= getbyterep (arg0
, &s0
))
1808 && (p1
= getbyterep (arg1
, &s1
))
1811 return build_cmp_result (type
, memcmp (p0
, p1
, s2
));
1814 case CFN_BUILT_IN_MEMCHR
:
1815 if (!size_t_cst_p (arg2
, &s2
))
1818 && !TREE_SIDE_EFFECTS (arg0
)
1819 && !TREE_SIDE_EFFECTS (arg1
))
1820 return build_int_cst (type
, 0);
1821 if ((p0
= getbyterep (arg0
, &s0
))
1823 && target_char_cst_p (arg1
, &c
))
1825 const char *r
= (const char *) memchr (p0
, c
, s2
);
1827 return build_int_cst (type
, 0);
1828 return fold_convert (type
,
1829 fold_build_pointer_plus_hwi (arg0
, r
- p0
));
1835 poly_uint64 parg0
, parg1
;
1836 if (poly_int_tree_p (arg0
, &parg0
) && poly_int_tree_p (arg1
, &parg1
))
1837 return fold_while_ult (type
, parg0
, parg1
);
1842 return fold_const_call_1 (fn
, type
, arg0
, arg1
, arg2
);