1 /* Constant folding for calls to built-in and internal functions.
2 Copyright (C) 1988-2019 Free Software Foundation, Inc.
4 This file is part of GCC.
6 GCC is free software; you can redistribute it and/or modify it under
7 the terms of the GNU General Public License as published by the Free
8 Software Foundation; either version 3, or (at your option) any later
11 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
12 WARRANTY; without even the implied warranty of MERCHANTABILITY or
13 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
16 You should have received a copy of the GNU General Public License
17 along with GCC; see the file COPYING3. If not see
18 <http://www.gnu.org/licenses/>. */
22 #include "coretypes.h"
25 #include "stor-layout.h"
27 #include "fold-const.h"
28 #include "fold-const-call.h"
29 #include "case-cfn-macros.h"
30 #include "tm.h" /* For C[LT]Z_DEFINED_AT_ZERO. */
32 #include "gimple-expr.h"
33 #include "tree-vector-builder.h"
35 /* Functions that test for certain constant types, abstracting away the
36 decision about whether to check for overflow. */
39 integer_cst_p (tree t
)
41 return TREE_CODE (t
) == INTEGER_CST
&& !TREE_OVERFLOW (t
);
47 return TREE_CODE (t
) == REAL_CST
&& !TREE_OVERFLOW (t
);
51 complex_cst_p (tree t
)
53 return TREE_CODE (t
) == COMPLEX_CST
;
56 /* Return true if ARG is a constant in the range of the host size_t.
57 Store it in *SIZE_OUT if so. */
60 host_size_t_cst_p (tree t
, size_t *size_out
)
62 if (types_compatible_p (size_type_node
, TREE_TYPE (t
))
64 && (wi::min_precision (wi::to_wide (t
), UNSIGNED
)
65 <= sizeof (size_t) * CHAR_BIT
))
67 *size_out
= tree_to_uhwi (t
);
73 /* RES is the result of a comparison in which < 0 means "less", 0 means
74 "equal" and > 0 means "more". Canonicalize it to -1, 0 or 1 and
75 return it in type TYPE. */
78 build_cmp_result (tree type
, int res
)
80 return build_int_cst (type
, res
< 0 ? -1 : res
> 0 ? 1 : 0);
83 /* M is the result of trying to constant-fold an expression (starting
84 with clear MPFR flags) and INEXACT says whether the result in M is
85 exact or inexact. Return true if M can be used as a constant-folded
86 result in format FORMAT, storing the value in *RESULT if so. */
89 do_mpfr_ckconv (real_value
*result
, mpfr_srcptr m
, bool inexact
,
90 const real_format
*format
)
92 /* Proceed iff we get a normal number, i.e. not NaN or Inf and no
93 overflow/underflow occurred. If -frounding-math, proceed iff the
94 result of calling FUNC was exact. */
95 if (!mpfr_number_p (m
)
97 || mpfr_underflow_p ()
98 || (flag_rounding_math
&& inexact
))
102 real_from_mpfr (&tmp
, m
, format
, GMP_RNDN
);
104 /* Proceed iff GCC's REAL_VALUE_TYPE can hold the MPFR values.
105 If the REAL_VALUE_TYPE is zero but the mpft_t is not, then we
106 underflowed in the conversion. */
107 if (!real_isfinite (&tmp
)
108 || ((tmp
.cl
== rvc_zero
) != (mpfr_zero_p (m
) != 0)))
111 real_convert (result
, format
, &tmp
);
112 return real_identical (result
, &tmp
);
119 in format FORMAT, given that FUNC is the MPFR implementation of f.
120 Return true on success. */
123 do_mpfr_arg1 (real_value
*result
,
124 int (*func
) (mpfr_ptr
, mpfr_srcptr
, mpfr_rnd_t
),
125 const real_value
*arg
, const real_format
*format
)
127 /* To proceed, MPFR must exactly represent the target floating point
128 format, which only happens when the target base equals two. */
129 if (format
->b
!= 2 || !real_isfinite (arg
))
132 int prec
= format
->p
;
133 mp_rnd_t rnd
= format
->round_towards_zero
? GMP_RNDZ
: GMP_RNDN
;
136 mpfr_init2 (m
, prec
);
137 mpfr_from_real (m
, arg
, GMP_RNDN
);
139 bool inexact
= func (m
, m
, rnd
);
140 bool ok
= do_mpfr_ckconv (result
, m
, inexact
, format
);
148 *RESULT_SIN = sin (*ARG);
149 *RESULT_COS = cos (*ARG);
151 for format FORMAT. Return true on success. */
154 do_mpfr_sincos (real_value
*result_sin
, real_value
*result_cos
,
155 const real_value
*arg
, const real_format
*format
)
157 /* To proceed, MPFR must exactly represent the target floating point
158 format, which only happens when the target base equals two. */
159 if (format
->b
!= 2 || !real_isfinite (arg
))
162 int prec
= format
->p
;
163 mp_rnd_t rnd
= format
->round_towards_zero
? GMP_RNDZ
: GMP_RNDN
;
166 mpfr_inits2 (prec
, m
, ms
, mc
, NULL
);
167 mpfr_from_real (m
, arg
, GMP_RNDN
);
169 bool inexact
= mpfr_sin_cos (ms
, mc
, m
, rnd
);
170 bool ok
= (do_mpfr_ckconv (result_sin
, ms
, inexact
, format
)
171 && do_mpfr_ckconv (result_cos
, mc
, inexact
, format
));
172 mpfr_clears (m
, ms
, mc
, NULL
);
179 *RESULT = f (*ARG0, *ARG1)
181 in format FORMAT, given that FUNC is the MPFR implementation of f.
182 Return true on success. */
185 do_mpfr_arg2 (real_value
*result
,
186 int (*func
) (mpfr_ptr
, mpfr_srcptr
, mpfr_srcptr
, mpfr_rnd_t
),
187 const real_value
*arg0
, const real_value
*arg1
,
188 const real_format
*format
)
190 /* To proceed, MPFR must exactly represent the target floating point
191 format, which only happens when the target base equals two. */
192 if (format
->b
!= 2 || !real_isfinite (arg0
) || !real_isfinite (arg1
))
195 int prec
= format
->p
;
196 mp_rnd_t rnd
= format
->round_towards_zero
? GMP_RNDZ
: GMP_RNDN
;
199 mpfr_inits2 (prec
, m0
, m1
, NULL
);
200 mpfr_from_real (m0
, arg0
, GMP_RNDN
);
201 mpfr_from_real (m1
, arg1
, GMP_RNDN
);
203 bool inexact
= func (m0
, m0
, m1
, rnd
);
204 bool ok
= do_mpfr_ckconv (result
, m0
, inexact
, format
);
205 mpfr_clears (m0
, m1
, NULL
);
212 *RESULT = f (ARG0, *ARG1)
214 in format FORMAT, given that FUNC is the MPFR implementation of f.
215 Return true on success. */
218 do_mpfr_arg2 (real_value
*result
,
219 int (*func
) (mpfr_ptr
, long, mpfr_srcptr
, mp_rnd_t
),
220 const wide_int_ref
&arg0
, const real_value
*arg1
,
221 const real_format
*format
)
223 if (format
->b
!= 2 || !real_isfinite (arg1
))
226 int prec
= format
->p
;
227 mp_rnd_t rnd
= format
->round_towards_zero
? GMP_RNDZ
: GMP_RNDN
;
230 mpfr_init2 (m
, prec
);
231 mpfr_from_real (m
, arg1
, GMP_RNDN
);
233 bool inexact
= func (m
, arg0
.to_shwi (), m
, rnd
);
234 bool ok
= do_mpfr_ckconv (result
, m
, inexact
, format
);
242 *RESULT = f (*ARG0, *ARG1, *ARG2)
244 in format FORMAT, given that FUNC is the MPFR implementation of f.
245 Return true on success. */
248 do_mpfr_arg3 (real_value
*result
,
249 int (*func
) (mpfr_ptr
, mpfr_srcptr
, mpfr_srcptr
,
250 mpfr_srcptr
, mpfr_rnd_t
),
251 const real_value
*arg0
, const real_value
*arg1
,
252 const real_value
*arg2
, const real_format
*format
)
254 /* To proceed, MPFR must exactly represent the target floating point
255 format, which only happens when the target base equals two. */
257 || !real_isfinite (arg0
)
258 || !real_isfinite (arg1
)
259 || !real_isfinite (arg2
))
262 int prec
= format
->p
;
263 mp_rnd_t rnd
= format
->round_towards_zero
? GMP_RNDZ
: GMP_RNDN
;
266 mpfr_inits2 (prec
, m0
, m1
, m2
, NULL
);
267 mpfr_from_real (m0
, arg0
, GMP_RNDN
);
268 mpfr_from_real (m1
, arg1
, GMP_RNDN
);
269 mpfr_from_real (m2
, arg2
, GMP_RNDN
);
271 bool inexact
= func (m0
, m0
, m1
, m2
, rnd
);
272 bool ok
= do_mpfr_ckconv (result
, m0
, inexact
, format
);
273 mpfr_clears (m0
, m1
, m2
, NULL
);
278 /* M is the result of trying to constant-fold an expression (starting
279 with clear MPFR flags) and INEXACT says whether the result in M is
280 exact or inexact. Return true if M can be used as a constant-folded
281 result in which the real and imaginary parts have format FORMAT.
282 Store those parts in *RESULT_REAL and *RESULT_IMAG if so. */
285 do_mpc_ckconv (real_value
*result_real
, real_value
*result_imag
,
286 mpc_srcptr m
, bool inexact
, const real_format
*format
)
288 /* Proceed iff we get a normal number, i.e. not NaN or Inf and no
289 overflow/underflow occurred. If -frounding-math, proceed iff the
290 result of calling FUNC was exact. */
291 if (!mpfr_number_p (mpc_realref (m
))
292 || !mpfr_number_p (mpc_imagref (m
))
293 || mpfr_overflow_p ()
294 || mpfr_underflow_p ()
295 || (flag_rounding_math
&& inexact
))
298 REAL_VALUE_TYPE tmp_real
, tmp_imag
;
299 real_from_mpfr (&tmp_real
, mpc_realref (m
), format
, GMP_RNDN
);
300 real_from_mpfr (&tmp_imag
, mpc_imagref (m
), format
, GMP_RNDN
);
302 /* Proceed iff GCC's REAL_VALUE_TYPE can hold the MPFR values.
303 If the REAL_VALUE_TYPE is zero but the mpft_t is not, then we
304 underflowed in the conversion. */
305 if (!real_isfinite (&tmp_real
)
306 || !real_isfinite (&tmp_imag
)
307 || (tmp_real
.cl
== rvc_zero
) != (mpfr_zero_p (mpc_realref (m
)) != 0)
308 || (tmp_imag
.cl
== rvc_zero
) != (mpfr_zero_p (mpc_imagref (m
)) != 0))
311 real_convert (result_real
, format
, &tmp_real
);
312 real_convert (result_imag
, format
, &tmp_imag
);
314 return (real_identical (result_real
, &tmp_real
)
315 && real_identical (result_imag
, &tmp_imag
));
322 in format FORMAT, given that FUNC is the mpc implementation of f.
323 Return true on success. Both RESULT and ARG are represented as
324 real and imaginary pairs. */
327 do_mpc_arg1 (real_value
*result_real
, real_value
*result_imag
,
328 int (*func
) (mpc_ptr
, mpc_srcptr
, mpc_rnd_t
),
329 const real_value
*arg_real
, const real_value
*arg_imag
,
330 const real_format
*format
)
332 /* To proceed, MPFR must exactly represent the target floating point
333 format, which only happens when the target base equals two. */
335 || !real_isfinite (arg_real
)
336 || !real_isfinite (arg_imag
))
339 int prec
= format
->p
;
340 mpc_rnd_t crnd
= format
->round_towards_zero
? MPC_RNDZZ
: MPC_RNDNN
;
344 mpfr_from_real (mpc_realref (m
), arg_real
, GMP_RNDN
);
345 mpfr_from_real (mpc_imagref (m
), arg_imag
, GMP_RNDN
);
347 bool inexact
= func (m
, m
, crnd
);
348 bool ok
= do_mpc_ckconv (result_real
, result_imag
, m
, inexact
, format
);
356 RESULT = f (ARG0, ARG1)
358 in format FORMAT, given that FUNC is the mpc implementation of f.
359 Return true on success. RESULT, ARG0 and ARG1 are represented as
360 real and imaginary pairs. */
363 do_mpc_arg2 (real_value
*result_real
, real_value
*result_imag
,
364 int (*func
)(mpc_ptr
, mpc_srcptr
, mpc_srcptr
, mpc_rnd_t
),
365 const real_value
*arg0_real
, const real_value
*arg0_imag
,
366 const real_value
*arg1_real
, const real_value
*arg1_imag
,
367 const real_format
*format
)
369 if (!real_isfinite (arg0_real
)
370 || !real_isfinite (arg0_imag
)
371 || !real_isfinite (arg1_real
)
372 || !real_isfinite (arg1_imag
))
375 int prec
= format
->p
;
376 mpc_rnd_t crnd
= format
->round_towards_zero
? MPC_RNDZZ
: MPC_RNDNN
;
379 mpc_init2 (m0
, prec
);
380 mpc_init2 (m1
, prec
);
381 mpfr_from_real (mpc_realref (m0
), arg0_real
, GMP_RNDN
);
382 mpfr_from_real (mpc_imagref (m0
), arg0_imag
, GMP_RNDN
);
383 mpfr_from_real (mpc_realref (m1
), arg1_real
, GMP_RNDN
);
384 mpfr_from_real (mpc_imagref (m1
), arg1_imag
, GMP_RNDN
);
386 bool inexact
= func (m0
, m0
, m1
, crnd
);
387 bool ok
= do_mpc_ckconv (result_real
, result_imag
, m0
, inexact
, format
);
396 *RESULT = logb (*ARG)
398 in format FORMAT. Return true on success. */
401 fold_const_logb (real_value
*result
, const real_value
*arg
,
402 const real_format
*format
)
407 /* If arg is +-NaN, then return it. */
412 /* If arg is +-Inf, then return +Inf. */
418 /* Zero may set errno and/or raise an exception. */
422 /* For normal numbers, proceed iff radix == 2. In GCC,
423 normalized significands are in the range [0.5, 1.0). We
424 want the exponent as if they were [1.0, 2.0) so get the
425 exponent and subtract 1. */
428 real_from_integer (result
, format
, REAL_EXP (arg
) - 1, SIGNED
);
438 *RESULT = significand (*ARG)
440 in format FORMAT. Return true on success. */
443 fold_const_significand (real_value
*result
, const real_value
*arg
,
444 const real_format
*format
)
451 /* If arg is +-0, +-Inf or +-NaN, then return it. */
456 /* For normal numbers, proceed iff radix == 2. */
460 /* In GCC, normalized significands are in the range [0.5, 1.0).
461 We want them to be [1.0, 2.0) so set the exponent to 1. */
462 SET_REAL_EXP (result
, 1);
474 where FORMAT is the format of *ARG and PRECISION is the number of
475 significant bits in the result. Return true on success. */
478 fold_const_conversion (wide_int
*result
,
479 void (*fn
) (real_value
*, format_helper
,
481 const real_value
*arg
, unsigned int precision
,
482 const real_format
*format
)
484 if (!real_isfinite (arg
))
488 fn (&rounded
, format
, arg
);
491 *result
= real_to_integer (&rounded
, &fail
, precision
);
497 *RESULT = pow (*ARG0, *ARG1)
499 in format FORMAT. Return true on success. */
502 fold_const_pow (real_value
*result
, const real_value
*arg0
,
503 const real_value
*arg1
, const real_format
*format
)
505 if (do_mpfr_arg2 (result
, mpfr_pow
, arg0
, arg1
, format
))
508 /* Check for an integer exponent. */
509 REAL_VALUE_TYPE cint1
;
510 HOST_WIDE_INT n1
= real_to_integer (arg1
);
511 real_from_integer (&cint1
, VOIDmode
, n1
, SIGNED
);
512 /* Attempt to evaluate pow at compile-time, unless this should
513 raise an exception. */
514 if (real_identical (arg1
, &cint1
)
516 || (!flag_trapping_math
&& !flag_errno_math
)
517 || !real_equal (arg0
, &dconst0
)))
519 bool inexact
= real_powi (result
, format
, arg0
, n1
);
520 /* Avoid the folding if flag_signaling_nans is on. */
521 if (flag_unsafe_math_optimizations
523 && !(flag_signaling_nans
524 && REAL_VALUE_ISSIGNALING_NAN (*arg0
))))
533 *RESULT = nextafter (*ARG0, *ARG1)
537 *RESULT = nexttoward (*ARG0, *ARG1)
539 in format FORMAT. Return true on success. */
542 fold_const_nextafter (real_value
*result
, const real_value
*arg0
,
543 const real_value
*arg1
, const real_format
*format
)
545 if (REAL_VALUE_ISSIGNALING_NAN (*arg0
)
546 || REAL_VALUE_ISSIGNALING_NAN (*arg1
))
549 /* Don't handle composite modes, nor decimal, nor modes without
550 inf or denorm at least for now. */
551 if (format
->pnan
< format
->p
554 || !format
->has_denorm
)
557 if (real_nextafter (result
, format
, arg0
, arg1
)
558 /* If raising underflow or overflow and setting errno to ERANGE,
559 fail if we care about those side-effects. */
560 && (flag_trapping_math
|| flag_errno_math
))
562 /* Similarly for nextafter (0, 1) raising underflow. */
563 else if (flag_trapping_math
564 && arg0
->cl
== rvc_zero
565 && result
->cl
!= rvc_zero
)
568 real_convert (result
, format
, result
);
575 *RESULT = ldexp (*ARG0, ARG1)
577 in format FORMAT. Return true on success. */
580 fold_const_builtin_load_exponent (real_value
*result
, const real_value
*arg0
,
581 const wide_int_ref
&arg1
,
582 const real_format
*format
)
584 /* Bound the maximum adjustment to twice the range of the
585 mode's valid exponents. Use abs to ensure the range is
586 positive as a sanity check. */
587 int max_exp_adj
= 2 * labs (format
->emax
- format
->emin
);
589 /* The requested adjustment must be inside this range. This
590 is a preliminary cap to avoid things like overflow, we
591 may still fail to compute the result for other reasons. */
592 if (wi::les_p (arg1
, -max_exp_adj
) || wi::ges_p (arg1
, max_exp_adj
))
595 /* Don't perform operation if we honor signaling NaNs and
596 operand is a signaling NaN. */
597 if (!flag_unsafe_math_optimizations
598 && flag_signaling_nans
599 && REAL_VALUE_ISSIGNALING_NAN (*arg0
))
602 REAL_VALUE_TYPE initial_result
;
603 real_ldexp (&initial_result
, arg0
, arg1
.to_shwi ());
605 /* Ensure we didn't overflow. */
606 if (real_isinf (&initial_result
))
609 /* Only proceed if the target mode can hold the
611 *result
= real_value_truncate (format
, initial_result
);
612 return real_equal (&initial_result
, result
);
615 /* Fold a call to __builtin_nan or __builtin_nans with argument ARG and
616 return type TYPE. QUIET is true if a quiet rather than signalling
620 fold_const_builtin_nan (tree type
, tree arg
, bool quiet
)
622 REAL_VALUE_TYPE real
;
623 const char *str
= c_getstr (arg
);
624 if (str
&& real_nan (&real
, str
, quiet
, TYPE_MODE (type
)))
625 return build_real (type
, real
);
629 /* Fold a call to IFN_REDUC_<CODE> (ARG), returning a value of type TYPE. */
632 fold_const_reduction (tree type
, tree arg
, tree_code code
)
634 unsigned HOST_WIDE_INT nelts
;
635 if (TREE_CODE (arg
) != VECTOR_CST
636 || !VECTOR_CST_NELTS (arg
).is_constant (&nelts
))
639 tree res
= VECTOR_CST_ELT (arg
, 0);
640 for (unsigned HOST_WIDE_INT i
= 1; i
< nelts
; i
++)
642 res
= const_binop (code
, type
, res
, VECTOR_CST_ELT (arg
, i
));
643 if (res
== NULL_TREE
|| !CONSTANT_CLASS_P (res
))
649 /* Fold a call to IFN_VEC_CONVERT (ARG) returning TYPE. */
652 fold_const_vec_convert (tree ret_type
, tree arg
)
654 enum tree_code code
= NOP_EXPR
;
655 tree arg_type
= TREE_TYPE (arg
);
656 if (TREE_CODE (arg
) != VECTOR_CST
)
659 gcc_checking_assert (VECTOR_TYPE_P (ret_type
) && VECTOR_TYPE_P (arg_type
));
661 if (INTEGRAL_TYPE_P (TREE_TYPE (ret_type
))
662 && SCALAR_FLOAT_TYPE_P (TREE_TYPE (arg_type
)))
663 code
= FIX_TRUNC_EXPR
;
664 else if (INTEGRAL_TYPE_P (TREE_TYPE (arg_type
))
665 && SCALAR_FLOAT_TYPE_P (TREE_TYPE (ret_type
)))
668 /* We can't handle steps directly when extending, since the
669 values need to wrap at the original precision first. */
671 = (INTEGRAL_TYPE_P (TREE_TYPE (ret_type
))
672 && INTEGRAL_TYPE_P (TREE_TYPE (arg_type
))
673 && (TYPE_PRECISION (TREE_TYPE (ret_type
))
674 <= TYPE_PRECISION (TREE_TYPE (arg_type
))));
675 tree_vector_builder elts
;
676 if (!elts
.new_unary_operation (ret_type
, arg
, step_ok_p
))
679 unsigned int count
= elts
.encoded_nelts ();
680 for (unsigned int i
= 0; i
< count
; ++i
)
682 tree elt
= fold_unary (code
, TREE_TYPE (ret_type
),
683 VECTOR_CST_ELT (arg
, i
));
684 if (elt
== NULL_TREE
|| !CONSTANT_CLASS_P (elt
))
686 elts
.quick_push (elt
);
689 return elts
.build ();
694 IFN_WHILE_ULT (ARG0, ARG1, (TYPE) { ... })
696 Return the value on success and null on failure. */
699 fold_while_ult (tree type
, poly_uint64 arg0
, poly_uint64 arg1
)
701 if (known_ge (arg0
, arg1
))
702 return build_zero_cst (type
);
704 if (maybe_ge (arg0
, arg1
))
707 poly_uint64 diff
= arg1
- arg0
;
708 poly_uint64 nelts
= TYPE_VECTOR_SUBPARTS (type
);
709 if (known_ge (diff
, nelts
))
710 return build_all_ones_cst (type
);
712 unsigned HOST_WIDE_INT const_diff
;
713 if (known_le (diff
, nelts
) && diff
.is_constant (&const_diff
))
715 tree minus_one
= build_minus_one_cst (TREE_TYPE (type
));
716 tree zero
= build_zero_cst (TREE_TYPE (type
));
717 return build_vector_a_then_b (type
, const_diff
, minus_one
, zero
);
726 in format FORMAT. Return true on success. */
729 fold_const_call_ss (real_value
*result
, combined_fn fn
,
730 const real_value
*arg
, const real_format
*format
)
736 return (real_compare (GE_EXPR
, arg
, &dconst0
)
737 && do_mpfr_arg1 (result
, mpfr_sqrt
, arg
, format
));
740 return do_mpfr_arg1 (result
, mpfr_cbrt
, arg
, format
);
743 return (real_compare (GE_EXPR
, arg
, &dconstm1
)
744 && real_compare (LE_EXPR
, arg
, &dconst1
)
745 && do_mpfr_arg1 (result
, mpfr_asin
, arg
, format
));
748 return (real_compare (GE_EXPR
, arg
, &dconstm1
)
749 && real_compare (LE_EXPR
, arg
, &dconst1
)
750 && do_mpfr_arg1 (result
, mpfr_acos
, arg
, format
));
753 return do_mpfr_arg1 (result
, mpfr_atan
, arg
, format
);
756 return do_mpfr_arg1 (result
, mpfr_asinh
, arg
, format
);
759 return (real_compare (GE_EXPR
, arg
, &dconst1
)
760 && do_mpfr_arg1 (result
, mpfr_acosh
, arg
, format
));
763 return (real_compare (GE_EXPR
, arg
, &dconstm1
)
764 && real_compare (LE_EXPR
, arg
, &dconst1
)
765 && do_mpfr_arg1 (result
, mpfr_atanh
, arg
, format
));
768 return do_mpfr_arg1 (result
, mpfr_sin
, arg
, format
);
771 return do_mpfr_arg1 (result
, mpfr_cos
, arg
, format
);
774 return do_mpfr_arg1 (result
, mpfr_tan
, arg
, format
);
777 return do_mpfr_arg1 (result
, mpfr_sinh
, arg
, format
);
780 return do_mpfr_arg1 (result
, mpfr_cosh
, arg
, format
);
783 return do_mpfr_arg1 (result
, mpfr_tanh
, arg
, format
);
786 return do_mpfr_arg1 (result
, mpfr_erf
, arg
, format
);
789 return do_mpfr_arg1 (result
, mpfr_erfc
, arg
, format
);
792 return do_mpfr_arg1 (result
, mpfr_gamma
, arg
, format
);
795 return do_mpfr_arg1 (result
, mpfr_exp
, arg
, format
);
798 return do_mpfr_arg1 (result
, mpfr_exp2
, arg
, format
);
802 return do_mpfr_arg1 (result
, mpfr_exp10
, arg
, format
);
805 return do_mpfr_arg1 (result
, mpfr_expm1
, arg
, format
);
808 return (real_compare (GT_EXPR
, arg
, &dconst0
)
809 && do_mpfr_arg1 (result
, mpfr_log
, arg
, format
));
812 return (real_compare (GT_EXPR
, arg
, &dconst0
)
813 && do_mpfr_arg1 (result
, mpfr_log2
, arg
, format
));
816 return (real_compare (GT_EXPR
, arg
, &dconst0
)
817 && do_mpfr_arg1 (result
, mpfr_log10
, arg
, format
));
820 return (real_compare (GT_EXPR
, arg
, &dconstm1
)
821 && do_mpfr_arg1 (result
, mpfr_log1p
, arg
, format
));
824 return do_mpfr_arg1 (result
, mpfr_j0
, arg
, format
);
827 return do_mpfr_arg1 (result
, mpfr_j1
, arg
, format
);
830 return (real_compare (GT_EXPR
, arg
, &dconst0
)
831 && do_mpfr_arg1 (result
, mpfr_y0
, arg
, format
));
834 return (real_compare (GT_EXPR
, arg
, &dconst0
)
835 && do_mpfr_arg1 (result
, mpfr_y1
, arg
, format
));
839 if (!REAL_VALUE_ISNAN (*arg
) || !flag_errno_math
)
841 real_floor (result
, format
, arg
);
848 if (!REAL_VALUE_ISNAN (*arg
) || !flag_errno_math
)
850 real_ceil (result
, format
, arg
);
857 real_trunc (result
, format
, arg
);
862 if (!REAL_VALUE_ISNAN (*arg
) || !flag_errno_math
)
864 real_round (result
, format
, arg
);
870 return fold_const_logb (result
, arg
, format
);
872 CASE_CFN_SIGNIFICAND
:
873 return fold_const_significand (result
, arg
, format
);
884 where FORMAT is the format of ARG and PRECISION is the number of
885 significant bits in the result. Return true on success. */
888 fold_const_call_ss (wide_int
*result
, combined_fn fn
,
889 const real_value
*arg
, unsigned int precision
,
890 const real_format
*format
)
895 if (real_isneg (arg
))
896 *result
= wi::one (precision
);
898 *result
= wi::zero (precision
);
902 /* For ilogb we don't know FP_ILOGB0, so only handle normal values.
903 Proceed iff radix == 2. In GCC, normalized significands are in
904 the range [0.5, 1.0). We want the exponent as if they were
905 [1.0, 2.0) so get the exponent and subtract 1. */
906 if (arg
->cl
== rvc_normal
&& format
->b
== 2)
908 *result
= wi::shwi (REAL_EXP (arg
) - 1, precision
);
916 return fold_const_conversion (result
, real_ceil
, arg
,
922 return fold_const_conversion (result
, real_floor
, arg
,
928 return fold_const_conversion (result
, real_round
, arg
,
934 /* Not yet folded to a constant. */
938 case CFN_BUILT_IN_FINITED32
:
939 case CFN_BUILT_IN_FINITED64
:
940 case CFN_BUILT_IN_FINITED128
:
941 case CFN_BUILT_IN_ISFINITE
:
942 *result
= wi::shwi (real_isfinite (arg
) ? 1 : 0, precision
);
946 case CFN_BUILT_IN_ISINFD32
:
947 case CFN_BUILT_IN_ISINFD64
:
948 case CFN_BUILT_IN_ISINFD128
:
949 if (real_isinf (arg
))
950 *result
= wi::shwi (arg
->sign
? -1 : 1, precision
);
952 *result
= wi::shwi (0, precision
);
956 case CFN_BUILT_IN_ISNAND32
:
957 case CFN_BUILT_IN_ISNAND64
:
958 case CFN_BUILT_IN_ISNAND128
:
959 *result
= wi::shwi (real_isnan (arg
) ? 1 : 0, precision
);
971 where ARG_TYPE is the type of ARG and PRECISION is the number of bits
972 in the result. Return true on success. */
975 fold_const_call_ss (wide_int
*result
, combined_fn fn
, const wide_int_ref
&arg
,
976 unsigned int precision
, tree arg_type
)
981 *result
= wi::shwi (wi::ffs (arg
), precision
);
987 if (wi::ne_p (arg
, 0))
989 else if (!CLZ_DEFINED_VALUE_AT_ZERO (SCALAR_INT_TYPE_MODE (arg_type
),
991 tmp
= TYPE_PRECISION (arg_type
);
992 *result
= wi::shwi (tmp
, precision
);
999 if (wi::ne_p (arg
, 0))
1000 tmp
= wi::ctz (arg
);
1001 else if (!CTZ_DEFINED_VALUE_AT_ZERO (SCALAR_INT_TYPE_MODE (arg_type
),
1003 tmp
= TYPE_PRECISION (arg_type
);
1004 *result
= wi::shwi (tmp
, precision
);
1009 *result
= wi::shwi (wi::clrsb (arg
), precision
);
1013 *result
= wi::shwi (wi::popcount (arg
), precision
);
1017 *result
= wi::shwi (wi::parity (arg
), precision
);
1020 case CFN_BUILT_IN_BSWAP16
:
1021 case CFN_BUILT_IN_BSWAP32
:
1022 case CFN_BUILT_IN_BSWAP64
:
1023 *result
= wide_int::from (arg
, precision
, TYPE_SIGN (arg_type
)).bswap ();
1035 where FORMAT is the format of ARG and of the real and imaginary parts
1036 of RESULT, passed as RESULT_REAL and RESULT_IMAG respectively. Return
1040 fold_const_call_cs (real_value
*result_real
, real_value
*result_imag
,
1041 combined_fn fn
, const real_value
*arg
,
1042 const real_format
*format
)
1047 /* cexpi(x+yi) = cos(x)+sin(y)*i. */
1048 return do_mpfr_sincos (result_imag
, result_real
, arg
, format
);
1059 where FORMAT is the format of RESULT and of the real and imaginary parts
1060 of ARG, passed as ARG_REAL and ARG_IMAG respectively. Return true on
1064 fold_const_call_sc (real_value
*result
, combined_fn fn
,
1065 const real_value
*arg_real
, const real_value
*arg_imag
,
1066 const real_format
*format
)
1071 return do_mpfr_arg2 (result
, mpfr_hypot
, arg_real
, arg_imag
, format
);
1082 where FORMAT is the format of the real and imaginary parts of RESULT
1083 (RESULT_REAL and RESULT_IMAG) and of ARG (ARG_REAL and ARG_IMAG).
1084 Return true on success. */
1087 fold_const_call_cc (real_value
*result_real
, real_value
*result_imag
,
1088 combined_fn fn
, const real_value
*arg_real
,
1089 const real_value
*arg_imag
, const real_format
*format
)
1094 return do_mpc_arg1 (result_real
, result_imag
, mpc_cos
,
1095 arg_real
, arg_imag
, format
);
1098 return do_mpc_arg1 (result_real
, result_imag
, mpc_cosh
,
1099 arg_real
, arg_imag
, format
);
1102 if (real_isinf (arg_real
) || real_isinf (arg_imag
))
1104 real_inf (result_real
);
1105 *result_imag
= dconst0
;
1106 result_imag
->sign
= arg_imag
->sign
;
1110 *result_real
= *arg_real
;
1111 *result_imag
= *arg_imag
;
1116 return do_mpc_arg1 (result_real
, result_imag
, mpc_sin
,
1117 arg_real
, arg_imag
, format
);
1120 return do_mpc_arg1 (result_real
, result_imag
, mpc_sinh
,
1121 arg_real
, arg_imag
, format
);
1124 return do_mpc_arg1 (result_real
, result_imag
, mpc_tan
,
1125 arg_real
, arg_imag
, format
);
1128 return do_mpc_arg1 (result_real
, result_imag
, mpc_tanh
,
1129 arg_real
, arg_imag
, format
);
1132 return do_mpc_arg1 (result_real
, result_imag
, mpc_log
,
1133 arg_real
, arg_imag
, format
);
1136 return do_mpc_arg1 (result_real
, result_imag
, mpc_sqrt
,
1137 arg_real
, arg_imag
, format
);
1140 return do_mpc_arg1 (result_real
, result_imag
, mpc_asin
,
1141 arg_real
, arg_imag
, format
);
1144 return do_mpc_arg1 (result_real
, result_imag
, mpc_acos
,
1145 arg_real
, arg_imag
, format
);
1148 return do_mpc_arg1 (result_real
, result_imag
, mpc_atan
,
1149 arg_real
, arg_imag
, format
);
1152 return do_mpc_arg1 (result_real
, result_imag
, mpc_asinh
,
1153 arg_real
, arg_imag
, format
);
1156 return do_mpc_arg1 (result_real
, result_imag
, mpc_acosh
,
1157 arg_real
, arg_imag
, format
);
1160 return do_mpc_arg1 (result_real
, result_imag
, mpc_atanh
,
1161 arg_real
, arg_imag
, format
);
1164 return do_mpc_arg1 (result_real
, result_imag
, mpc_exp
,
1165 arg_real
, arg_imag
, format
);
1172 /* Subroutine of fold_const_call, with the same interface. Handle cases
1173 where the arguments and result are numerical. */
1176 fold_const_call_1 (combined_fn fn
, tree type
, tree arg
)
1178 machine_mode mode
= TYPE_MODE (type
);
1179 machine_mode arg_mode
= TYPE_MODE (TREE_TYPE (arg
));
1181 if (integer_cst_p (arg
))
1183 if (SCALAR_INT_MODE_P (mode
))
1186 if (fold_const_call_ss (&result
, fn
, wi::to_wide (arg
),
1187 TYPE_PRECISION (type
), TREE_TYPE (arg
)))
1188 return wide_int_to_tree (type
, result
);
1193 if (real_cst_p (arg
))
1195 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg_mode
));
1196 if (mode
== arg_mode
)
1199 REAL_VALUE_TYPE result
;
1200 if (fold_const_call_ss (&result
, fn
, TREE_REAL_CST_PTR (arg
),
1201 REAL_MODE_FORMAT (mode
)))
1202 return build_real (type
, result
);
1204 else if (COMPLEX_MODE_P (mode
)
1205 && GET_MODE_INNER (mode
) == arg_mode
)
1207 /* real -> complex real. */
1208 REAL_VALUE_TYPE result_real
, result_imag
;
1209 if (fold_const_call_cs (&result_real
, &result_imag
, fn
,
1210 TREE_REAL_CST_PTR (arg
),
1211 REAL_MODE_FORMAT (arg_mode
)))
1212 return build_complex (type
,
1213 build_real (TREE_TYPE (type
), result_real
),
1214 build_real (TREE_TYPE (type
), result_imag
));
1216 else if (INTEGRAL_TYPE_P (type
))
1220 if (fold_const_call_ss (&result
, fn
,
1221 TREE_REAL_CST_PTR (arg
),
1222 TYPE_PRECISION (type
),
1223 REAL_MODE_FORMAT (arg_mode
)))
1224 return wide_int_to_tree (type
, result
);
1229 if (complex_cst_p (arg
))
1231 gcc_checking_assert (COMPLEX_MODE_P (arg_mode
));
1232 machine_mode inner_mode
= GET_MODE_INNER (arg_mode
);
1233 tree argr
= TREE_REALPART (arg
);
1234 tree argi
= TREE_IMAGPART (arg
);
1235 if (mode
== arg_mode
1236 && real_cst_p (argr
)
1237 && real_cst_p (argi
))
1239 /* complex real -> complex real. */
1240 REAL_VALUE_TYPE result_real
, result_imag
;
1241 if (fold_const_call_cc (&result_real
, &result_imag
, fn
,
1242 TREE_REAL_CST_PTR (argr
),
1243 TREE_REAL_CST_PTR (argi
),
1244 REAL_MODE_FORMAT (inner_mode
)))
1245 return build_complex (type
,
1246 build_real (TREE_TYPE (type
), result_real
),
1247 build_real (TREE_TYPE (type
), result_imag
));
1249 if (mode
== inner_mode
1250 && real_cst_p (argr
)
1251 && real_cst_p (argi
))
1253 /* complex real -> real. */
1254 REAL_VALUE_TYPE result
;
1255 if (fold_const_call_sc (&result
, fn
,
1256 TREE_REAL_CST_PTR (argr
),
1257 TREE_REAL_CST_PTR (argi
),
1258 REAL_MODE_FORMAT (inner_mode
)))
1259 return build_real (type
, result
);
1267 /* Try to fold FN (ARG) to a constant. Return the constant on success,
1268 otherwise return null. TYPE is the type of the return value. */
1271 fold_const_call (combined_fn fn
, tree type
, tree arg
)
1275 case CFN_BUILT_IN_STRLEN
:
1276 if (const char *str
= c_getstr (arg
))
1277 return build_int_cst (type
, strlen (str
));
1281 CASE_FLT_FN_FLOATN_NX (CFN_BUILT_IN_NAN
):
1282 case CFN_BUILT_IN_NAND32
:
1283 case CFN_BUILT_IN_NAND64
:
1284 case CFN_BUILT_IN_NAND128
:
1285 return fold_const_builtin_nan (type
, arg
, true);
1288 CASE_FLT_FN_FLOATN_NX (CFN_BUILT_IN_NANS
):
1289 return fold_const_builtin_nan (type
, arg
, false);
1291 case CFN_REDUC_PLUS
:
1292 return fold_const_reduction (type
, arg
, PLUS_EXPR
);
1295 return fold_const_reduction (type
, arg
, MAX_EXPR
);
1298 return fold_const_reduction (type
, arg
, MIN_EXPR
);
1301 return fold_const_reduction (type
, arg
, BIT_AND_EXPR
);
1304 return fold_const_reduction (type
, arg
, BIT_IOR_EXPR
);
1307 return fold_const_reduction (type
, arg
, BIT_XOR_EXPR
);
1309 case CFN_VEC_CONVERT
:
1310 return fold_const_vec_convert (type
, arg
);
1313 return fold_const_call_1 (fn
, type
, arg
);
1317 /* Fold a call to IFN_FOLD_LEFT_<CODE> (ARG0, ARG1), returning a value
1321 fold_const_fold_left (tree type
, tree arg0
, tree arg1
, tree_code code
)
1323 if (TREE_CODE (arg1
) != VECTOR_CST
)
1326 unsigned HOST_WIDE_INT nelts
;
1327 if (!VECTOR_CST_NELTS (arg1
).is_constant (&nelts
))
1330 for (unsigned HOST_WIDE_INT i
= 0; i
< nelts
; i
++)
1332 arg0
= const_binop (code
, type
, arg0
, VECTOR_CST_ELT (arg1
, i
));
1333 if (arg0
== NULL_TREE
|| !CONSTANT_CLASS_P (arg0
))
1341 *RESULT = FN (*ARG0, *ARG1)
1343 in format FORMAT. Return true on success. */
1346 fold_const_call_sss (real_value
*result
, combined_fn fn
,
1347 const real_value
*arg0
, const real_value
*arg1
,
1348 const real_format
*format
)
1354 return do_mpfr_arg2 (result
, mpfr_remainder
, arg0
, arg1
, format
);
1357 return do_mpfr_arg2 (result
, mpfr_atan2
, arg0
, arg1
, format
);
1360 return do_mpfr_arg2 (result
, mpfr_dim
, arg0
, arg1
, format
);
1363 return do_mpfr_arg2 (result
, mpfr_hypot
, arg0
, arg1
, format
);
1366 CASE_CFN_COPYSIGN_FN
:
1368 real_copysign (result
, arg1
);
1373 return do_mpfr_arg2 (result
, mpfr_min
, arg0
, arg1
, format
);
1377 return do_mpfr_arg2 (result
, mpfr_max
, arg0
, arg1
, format
);
1380 return fold_const_pow (result
, arg0
, arg1
, format
);
1383 CASE_CFN_NEXTTOWARD
:
1384 return fold_const_nextafter (result
, arg0
, arg1
, format
);
1393 *RESULT = FN (*ARG0, ARG1)
1395 where FORMAT is the format of *RESULT and *ARG0. Return true on
1399 fold_const_call_sss (real_value
*result
, combined_fn fn
,
1400 const real_value
*arg0
, const wide_int_ref
&arg1
,
1401 const real_format
*format
)
1406 return fold_const_builtin_load_exponent (result
, arg0
, arg1
, format
);
1410 return (format
->b
== 2
1411 && fold_const_builtin_load_exponent (result
, arg0
, arg1
,
1415 /* Avoid the folding if flag_signaling_nans is on and
1416 operand is a signaling NaN. */
1417 if (!flag_unsafe_math_optimizations
1418 && flag_signaling_nans
1419 && REAL_VALUE_ISSIGNALING_NAN (*arg0
))
1422 real_powi (result
, format
, arg0
, arg1
.to_shwi ());
1432 *RESULT = FN (ARG0, *ARG1)
1434 where FORMAT is the format of *RESULT and *ARG1. Return true on
1438 fold_const_call_sss (real_value
*result
, combined_fn fn
,
1439 const wide_int_ref
&arg0
, const real_value
*arg1
,
1440 const real_format
*format
)
1445 return do_mpfr_arg2 (result
, mpfr_jn
, arg0
, arg1
, format
);
1448 return (real_compare (GT_EXPR
, arg1
, &dconst0
)
1449 && do_mpfr_arg2 (result
, mpfr_yn
, arg0
, arg1
, format
));
1458 RESULT = fn (ARG0, ARG1)
1460 where FORMAT is the format of the real and imaginary parts of RESULT
1461 (RESULT_REAL and RESULT_IMAG), of ARG0 (ARG0_REAL and ARG0_IMAG)
1462 and of ARG1 (ARG1_REAL and ARG1_IMAG). Return true on success. */
1465 fold_const_call_ccc (real_value
*result_real
, real_value
*result_imag
,
1466 combined_fn fn
, const real_value
*arg0_real
,
1467 const real_value
*arg0_imag
, const real_value
*arg1_real
,
1468 const real_value
*arg1_imag
, const real_format
*format
)
1473 return do_mpc_arg2 (result_real
, result_imag
, mpc_pow
,
1474 arg0_real
, arg0_imag
, arg1_real
, arg1_imag
, format
);
1481 /* Subroutine of fold_const_call, with the same interface. Handle cases
1482 where the arguments and result are numerical. */
1485 fold_const_call_1 (combined_fn fn
, tree type
, tree arg0
, tree arg1
)
1487 machine_mode mode
= TYPE_MODE (type
);
1488 machine_mode arg0_mode
= TYPE_MODE (TREE_TYPE (arg0
));
1489 machine_mode arg1_mode
= TYPE_MODE (TREE_TYPE (arg1
));
1491 if (mode
== arg0_mode
1492 && real_cst_p (arg0
)
1493 && real_cst_p (arg1
))
1495 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg0_mode
));
1496 REAL_VALUE_TYPE result
;
1497 if (arg0_mode
== arg1_mode
)
1499 /* real, real -> real. */
1500 if (fold_const_call_sss (&result
, fn
, TREE_REAL_CST_PTR (arg0
),
1501 TREE_REAL_CST_PTR (arg1
),
1502 REAL_MODE_FORMAT (mode
)))
1503 return build_real (type
, result
);
1505 else if (arg1_mode
== TYPE_MODE (long_double_type_node
))
1508 CASE_CFN_NEXTTOWARD
:
1509 /* real, long double -> real. */
1510 if (fold_const_call_sss (&result
, fn
, TREE_REAL_CST_PTR (arg0
),
1511 TREE_REAL_CST_PTR (arg1
),
1512 REAL_MODE_FORMAT (mode
)))
1513 return build_real (type
, result
);
1521 if (real_cst_p (arg0
)
1522 && integer_cst_p (arg1
))
1524 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg0_mode
));
1525 if (mode
== arg0_mode
)
1527 /* real, int -> real. */
1528 REAL_VALUE_TYPE result
;
1529 if (fold_const_call_sss (&result
, fn
, TREE_REAL_CST_PTR (arg0
),
1531 REAL_MODE_FORMAT (mode
)))
1532 return build_real (type
, result
);
1537 if (integer_cst_p (arg0
)
1538 && real_cst_p (arg1
))
1540 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg1_mode
));
1541 if (mode
== arg1_mode
)
1543 /* int, real -> real. */
1544 REAL_VALUE_TYPE result
;
1545 if (fold_const_call_sss (&result
, fn
, wi::to_wide (arg0
),
1546 TREE_REAL_CST_PTR (arg1
),
1547 REAL_MODE_FORMAT (mode
)))
1548 return build_real (type
, result
);
1553 if (arg0_mode
== arg1_mode
1554 && complex_cst_p (arg0
)
1555 && complex_cst_p (arg1
))
1557 gcc_checking_assert (COMPLEX_MODE_P (arg0_mode
));
1558 machine_mode inner_mode
= GET_MODE_INNER (arg0_mode
);
1559 tree arg0r
= TREE_REALPART (arg0
);
1560 tree arg0i
= TREE_IMAGPART (arg0
);
1561 tree arg1r
= TREE_REALPART (arg1
);
1562 tree arg1i
= TREE_IMAGPART (arg1
);
1563 if (mode
== arg0_mode
1564 && real_cst_p (arg0r
)
1565 && real_cst_p (arg0i
)
1566 && real_cst_p (arg1r
)
1567 && real_cst_p (arg1i
))
1569 /* complex real, complex real -> complex real. */
1570 REAL_VALUE_TYPE result_real
, result_imag
;
1571 if (fold_const_call_ccc (&result_real
, &result_imag
, fn
,
1572 TREE_REAL_CST_PTR (arg0r
),
1573 TREE_REAL_CST_PTR (arg0i
),
1574 TREE_REAL_CST_PTR (arg1r
),
1575 TREE_REAL_CST_PTR (arg1i
),
1576 REAL_MODE_FORMAT (inner_mode
)))
1577 return build_complex (type
,
1578 build_real (TREE_TYPE (type
), result_real
),
1579 build_real (TREE_TYPE (type
), result_imag
));
1587 /* Try to fold FN (ARG0, ARG1) to a constant. Return the constant on success,
1588 otherwise return null. TYPE is the type of the return value. */
1591 fold_const_call (combined_fn fn
, tree type
, tree arg0
, tree arg1
)
1593 const char *p0
, *p1
;
1597 case CFN_BUILT_IN_STRSPN
:
1598 if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1599 return build_int_cst (type
, strspn (p0
, p1
));
1602 case CFN_BUILT_IN_STRCSPN
:
1603 if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1604 return build_int_cst (type
, strcspn (p0
, p1
));
1607 case CFN_BUILT_IN_STRCMP
:
1608 if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1609 return build_cmp_result (type
, strcmp (p0
, p1
));
1612 case CFN_BUILT_IN_STRCASECMP
:
1613 if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1615 int r
= strcmp (p0
, p1
);
1617 return build_cmp_result (type
, r
);
1621 case CFN_BUILT_IN_INDEX
:
1622 case CFN_BUILT_IN_STRCHR
:
1623 if ((p0
= c_getstr (arg0
)) && target_char_cst_p (arg1
, &c
))
1625 const char *r
= strchr (p0
, c
);
1627 return build_int_cst (type
, 0);
1628 return fold_convert (type
,
1629 fold_build_pointer_plus_hwi (arg0
, r
- p0
));
1633 case CFN_BUILT_IN_RINDEX
:
1634 case CFN_BUILT_IN_STRRCHR
:
1635 if ((p0
= c_getstr (arg0
)) && target_char_cst_p (arg1
, &c
))
1637 const char *r
= strrchr (p0
, c
);
1639 return build_int_cst (type
, 0);
1640 return fold_convert (type
,
1641 fold_build_pointer_plus_hwi (arg0
, r
- p0
));
1645 case CFN_BUILT_IN_STRSTR
:
1646 if ((p1
= c_getstr (arg1
)))
1648 if ((p0
= c_getstr (arg0
)))
1650 const char *r
= strstr (p0
, p1
);
1652 return build_int_cst (type
, 0);
1653 return fold_convert (type
,
1654 fold_build_pointer_plus_hwi (arg0
, r
- p0
));
1657 return fold_convert (type
, arg0
);
1661 case CFN_FOLD_LEFT_PLUS
:
1662 return fold_const_fold_left (type
, arg0
, arg1
, PLUS_EXPR
);
1665 return fold_const_call_1 (fn
, type
, arg0
, arg1
);
1671 *RESULT = FN (*ARG0, *ARG1, *ARG2)
1673 in format FORMAT. Return true on success. */
1676 fold_const_call_ssss (real_value
*result
, combined_fn fn
,
1677 const real_value
*arg0
, const real_value
*arg1
,
1678 const real_value
*arg2
, const real_format
*format
)
1684 return do_mpfr_arg3 (result
, mpfr_fma
, arg0
, arg1
, arg2
, format
);
1688 real_value new_arg2
= real_value_negate (arg2
);
1689 return do_mpfr_arg3 (result
, mpfr_fma
, arg0
, arg1
, &new_arg2
, format
);
1694 real_value new_arg0
= real_value_negate (arg0
);
1695 return do_mpfr_arg3 (result
, mpfr_fma
, &new_arg0
, arg1
, arg2
, format
);
1700 real_value new_arg0
= real_value_negate (arg0
);
1701 real_value new_arg2
= real_value_negate (arg2
);
1702 return do_mpfr_arg3 (result
, mpfr_fma
, &new_arg0
, arg1
,
1711 /* Subroutine of fold_const_call, with the same interface. Handle cases
1712 where the arguments and result are numerical. */
1715 fold_const_call_1 (combined_fn fn
, tree type
, tree arg0
, tree arg1
, tree arg2
)
1717 machine_mode mode
= TYPE_MODE (type
);
1718 machine_mode arg0_mode
= TYPE_MODE (TREE_TYPE (arg0
));
1719 machine_mode arg1_mode
= TYPE_MODE (TREE_TYPE (arg1
));
1720 machine_mode arg2_mode
= TYPE_MODE (TREE_TYPE (arg2
));
1722 if (arg0_mode
== arg1_mode
1723 && arg0_mode
== arg2_mode
1724 && real_cst_p (arg0
)
1725 && real_cst_p (arg1
)
1726 && real_cst_p (arg2
))
1728 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg0_mode
));
1729 if (mode
== arg0_mode
)
1731 /* real, real, real -> real. */
1732 REAL_VALUE_TYPE result
;
1733 if (fold_const_call_ssss (&result
, fn
, TREE_REAL_CST_PTR (arg0
),
1734 TREE_REAL_CST_PTR (arg1
),
1735 TREE_REAL_CST_PTR (arg2
),
1736 REAL_MODE_FORMAT (mode
)))
1737 return build_real (type
, result
);
1745 /* Try to fold FN (ARG0, ARG1, ARG2) to a constant. Return the constant on
1746 success, otherwise return null. TYPE is the type of the return value. */
1749 fold_const_call (combined_fn fn
, tree type
, tree arg0
, tree arg1
, tree arg2
)
1751 const char *p0
, *p1
;
1753 unsigned HOST_WIDE_INT s0
, s1
;
1757 case CFN_BUILT_IN_STRNCMP
:
1758 if (!host_size_t_cst_p (arg2
, &s2
))
1761 && !TREE_SIDE_EFFECTS (arg0
)
1762 && !TREE_SIDE_EFFECTS (arg1
))
1763 return build_int_cst (type
, 0);
1764 else if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1765 return build_int_cst (type
, strncmp (p0
, p1
, s2
));
1768 case CFN_BUILT_IN_STRNCASECMP
:
1769 if (!host_size_t_cst_p (arg2
, &s2
))
1772 && !TREE_SIDE_EFFECTS (arg0
)
1773 && !TREE_SIDE_EFFECTS (arg1
))
1774 return build_int_cst (type
, 0);
1775 else if ((p0
= c_getstr (arg0
))
1776 && (p1
= c_getstr (arg1
))
1777 && strncmp (p0
, p1
, s2
) == 0)
1778 return build_int_cst (type
, 0);
1781 case CFN_BUILT_IN_BCMP
:
1782 case CFN_BUILT_IN_MEMCMP
:
1783 if (!host_size_t_cst_p (arg2
, &s2
))
1786 && !TREE_SIDE_EFFECTS (arg0
)
1787 && !TREE_SIDE_EFFECTS (arg1
))
1788 return build_int_cst (type
, 0);
1789 if ((p0
= c_getstr (arg0
, &s0
))
1790 && (p1
= c_getstr (arg1
, &s1
))
1793 return build_cmp_result (type
, memcmp (p0
, p1
, s2
));
1796 case CFN_BUILT_IN_MEMCHR
:
1797 if (!host_size_t_cst_p (arg2
, &s2
))
1800 && !TREE_SIDE_EFFECTS (arg0
)
1801 && !TREE_SIDE_EFFECTS (arg1
))
1802 return build_int_cst (type
, 0);
1803 if ((p0
= c_getstr (arg0
, &s0
))
1805 && target_char_cst_p (arg1
, &c
))
1807 const char *r
= (const char *) memchr (p0
, c
, s2
);
1809 return build_int_cst (type
, 0);
1810 return fold_convert (type
,
1811 fold_build_pointer_plus_hwi (arg0
, r
- p0
));
1817 poly_uint64 parg0
, parg1
;
1818 if (poly_int_tree_p (arg0
, &parg0
) && poly_int_tree_p (arg1
, &parg1
))
1819 return fold_while_ult (type
, parg0
, parg1
);
1824 return fold_const_call_1 (fn
, type
, arg0
, arg1
, arg2
);