1 /* Constant folding for calls to built-in and internal functions.
2 Copyright (C) 1988-2020 Free Software Foundation, Inc.
4 This file is part of GCC.
6 GCC is free software; you can redistribute it and/or modify it under
7 the terms of the GNU General Public License as published by the Free
8 Software Foundation; either version 3, or (at your option) any later
11 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
12 WARRANTY; without even the implied warranty of MERCHANTABILITY or
13 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
16 You should have received a copy of the GNU General Public License
17 along with GCC; see the file COPYING3. If not see
18 <http://www.gnu.org/licenses/>. */
22 #include "coretypes.h"
25 #include "stor-layout.h"
27 #include "fold-const.h"
28 #include "fold-const-call.h"
29 #include "case-cfn-macros.h"
30 #include "tm.h" /* For C[LT]Z_DEFINED_AT_ZERO. */
32 #include "gimple-expr.h"
33 #include "tree-vector-builder.h"
35 /* Functions that test for certain constant types, abstracting away the
36 decision about whether to check for overflow. */
39 integer_cst_p (tree t
)
41 return TREE_CODE (t
) == INTEGER_CST
&& !TREE_OVERFLOW (t
);
47 return TREE_CODE (t
) == REAL_CST
&& !TREE_OVERFLOW (t
);
51 complex_cst_p (tree t
)
53 return TREE_CODE (t
) == COMPLEX_CST
;
56 /* Return true if ARG is a constant in the range of the host size_t.
57 Store it in *SIZE_OUT if so. */
60 host_size_t_cst_p (tree t
, size_t *size_out
)
62 if (types_compatible_p (size_type_node
, TREE_TYPE (t
))
64 && (wi::min_precision (wi::to_wide (t
), UNSIGNED
)
65 <= sizeof (size_t) * CHAR_BIT
))
67 *size_out
= tree_to_uhwi (t
);
73 /* RES is the result of a comparison in which < 0 means "less", 0 means
74 "equal" and > 0 means "more". Canonicalize it to -1, 0 or 1 and
75 return it in type TYPE. */
78 build_cmp_result (tree type
, int res
)
80 return build_int_cst (type
, res
< 0 ? -1 : res
> 0 ? 1 : 0);
83 /* M is the result of trying to constant-fold an expression (starting
84 with clear MPFR flags) and INEXACT says whether the result in M is
85 exact or inexact. Return true if M can be used as a constant-folded
86 result in format FORMAT, storing the value in *RESULT if so. */
89 do_mpfr_ckconv (real_value
*result
, mpfr_srcptr m
, bool inexact
,
90 const real_format
*format
)
92 /* Proceed iff we get a normal number, i.e. not NaN or Inf and no
93 overflow/underflow occurred. If -frounding-math, proceed iff the
94 result of calling FUNC was exact. */
95 if (!mpfr_number_p (m
)
97 || mpfr_underflow_p ()
98 || (flag_rounding_math
&& inexact
))
102 real_from_mpfr (&tmp
, m
, format
, MPFR_RNDN
);
104 /* Proceed iff GCC's REAL_VALUE_TYPE can hold the MPFR values.
105 If the REAL_VALUE_TYPE is zero but the mpft_t is not, then we
106 underflowed in the conversion. */
107 if (!real_isfinite (&tmp
)
108 || ((tmp
.cl
== rvc_zero
) != (mpfr_zero_p (m
) != 0)))
111 real_convert (result
, format
, &tmp
);
112 return real_identical (result
, &tmp
);
119 in format FORMAT, given that FUNC is the MPFR implementation of f.
120 Return true on success. */
123 do_mpfr_arg1 (real_value
*result
,
124 int (*func
) (mpfr_ptr
, mpfr_srcptr
, mpfr_rnd_t
),
125 const real_value
*arg
, const real_format
*format
)
127 /* To proceed, MPFR must exactly represent the target floating point
128 format, which only happens when the target base equals two. */
129 if (format
->b
!= 2 || !real_isfinite (arg
))
132 int prec
= format
->p
;
133 mpfr_rnd_t rnd
= format
->round_towards_zero
? MPFR_RNDZ
: MPFR_RNDN
;
136 mpfr_init2 (m
, prec
);
137 mpfr_from_real (m
, arg
, MPFR_RNDN
);
139 bool inexact
= func (m
, m
, rnd
);
140 bool ok
= do_mpfr_ckconv (result
, m
, inexact
, format
);
148 *RESULT_SIN = sin (*ARG);
149 *RESULT_COS = cos (*ARG);
151 for format FORMAT. Return true on success. */
154 do_mpfr_sincos (real_value
*result_sin
, real_value
*result_cos
,
155 const real_value
*arg
, const real_format
*format
)
157 /* To proceed, MPFR must exactly represent the target floating point
158 format, which only happens when the target base equals two. */
159 if (format
->b
!= 2 || !real_isfinite (arg
))
162 int prec
= format
->p
;
163 mpfr_rnd_t rnd
= format
->round_towards_zero
? MPFR_RNDZ
: MPFR_RNDN
;
166 mpfr_inits2 (prec
, m
, ms
, mc
, NULL
);
167 mpfr_from_real (m
, arg
, MPFR_RNDN
);
169 bool inexact
= mpfr_sin_cos (ms
, mc
, m
, rnd
);
170 bool ok
= (do_mpfr_ckconv (result_sin
, ms
, inexact
, format
)
171 && do_mpfr_ckconv (result_cos
, mc
, inexact
, format
));
172 mpfr_clears (m
, ms
, mc
, NULL
);
179 *RESULT = f (*ARG0, *ARG1)
181 in format FORMAT, given that FUNC is the MPFR implementation of f.
182 Return true on success. */
185 do_mpfr_arg2 (real_value
*result
,
186 int (*func
) (mpfr_ptr
, mpfr_srcptr
, mpfr_srcptr
, mpfr_rnd_t
),
187 const real_value
*arg0
, const real_value
*arg1
,
188 const real_format
*format
)
190 /* To proceed, MPFR must exactly represent the target floating point
191 format, which only happens when the target base equals two. */
192 if (format
->b
!= 2 || !real_isfinite (arg0
) || !real_isfinite (arg1
))
195 int prec
= format
->p
;
196 mpfr_rnd_t rnd
= format
->round_towards_zero
? MPFR_RNDZ
: MPFR_RNDN
;
199 mpfr_inits2 (prec
, m0
, m1
, NULL
);
200 mpfr_from_real (m0
, arg0
, MPFR_RNDN
);
201 mpfr_from_real (m1
, arg1
, MPFR_RNDN
);
203 bool inexact
= func (m0
, m0
, m1
, rnd
);
204 bool ok
= do_mpfr_ckconv (result
, m0
, inexact
, format
);
205 mpfr_clears (m0
, m1
, NULL
);
212 *RESULT = f (ARG0, *ARG1)
214 in format FORMAT, given that FUNC is the MPFR implementation of f.
215 Return true on success. */
218 do_mpfr_arg2 (real_value
*result
,
219 int (*func
) (mpfr_ptr
, long, mpfr_srcptr
, mpfr_rnd_t
),
220 const wide_int_ref
&arg0
, const real_value
*arg1
,
221 const real_format
*format
)
223 if (format
->b
!= 2 || !real_isfinite (arg1
))
226 int prec
= format
->p
;
227 mpfr_rnd_t rnd
= format
->round_towards_zero
? MPFR_RNDZ
: MPFR_RNDN
;
230 mpfr_init2 (m
, prec
);
231 mpfr_from_real (m
, arg1
, MPFR_RNDN
);
233 bool inexact
= func (m
, arg0
.to_shwi (), m
, rnd
);
234 bool ok
= do_mpfr_ckconv (result
, m
, inexact
, format
);
242 *RESULT = f (*ARG0, *ARG1, *ARG2)
244 in format FORMAT, given that FUNC is the MPFR implementation of f.
245 Return true on success. */
248 do_mpfr_arg3 (real_value
*result
,
249 int (*func
) (mpfr_ptr
, mpfr_srcptr
, mpfr_srcptr
,
250 mpfr_srcptr
, mpfr_rnd_t
),
251 const real_value
*arg0
, const real_value
*arg1
,
252 const real_value
*arg2
, const real_format
*format
)
254 /* To proceed, MPFR must exactly represent the target floating point
255 format, which only happens when the target base equals two. */
257 || !real_isfinite (arg0
)
258 || !real_isfinite (arg1
)
259 || !real_isfinite (arg2
))
262 int prec
= format
->p
;
263 mpfr_rnd_t rnd
= format
->round_towards_zero
? MPFR_RNDZ
: MPFR_RNDN
;
266 mpfr_inits2 (prec
, m0
, m1
, m2
, NULL
);
267 mpfr_from_real (m0
, arg0
, MPFR_RNDN
);
268 mpfr_from_real (m1
, arg1
, MPFR_RNDN
);
269 mpfr_from_real (m2
, arg2
, MPFR_RNDN
);
271 bool inexact
= func (m0
, m0
, m1
, m2
, rnd
);
272 bool ok
= do_mpfr_ckconv (result
, m0
, inexact
, format
);
273 mpfr_clears (m0
, m1
, m2
, NULL
);
278 /* M is the result of trying to constant-fold an expression (starting
279 with clear MPFR flags) and INEXACT says whether the result in M is
280 exact or inexact. Return true if M can be used as a constant-folded
281 result in which the real and imaginary parts have format FORMAT.
282 Store those parts in *RESULT_REAL and *RESULT_IMAG if so. */
285 do_mpc_ckconv (real_value
*result_real
, real_value
*result_imag
,
286 mpc_srcptr m
, bool inexact
, const real_format
*format
)
288 /* Proceed iff we get a normal number, i.e. not NaN or Inf and no
289 overflow/underflow occurred. If -frounding-math, proceed iff the
290 result of calling FUNC was exact. */
291 if (!mpfr_number_p (mpc_realref (m
))
292 || !mpfr_number_p (mpc_imagref (m
))
293 || mpfr_overflow_p ()
294 || mpfr_underflow_p ()
295 || (flag_rounding_math
&& inexact
))
298 REAL_VALUE_TYPE tmp_real
, tmp_imag
;
299 real_from_mpfr (&tmp_real
, mpc_realref (m
), format
, MPFR_RNDN
);
300 real_from_mpfr (&tmp_imag
, mpc_imagref (m
), format
, MPFR_RNDN
);
302 /* Proceed iff GCC's REAL_VALUE_TYPE can hold the MPFR values.
303 If the REAL_VALUE_TYPE is zero but the mpft_t is not, then we
304 underflowed in the conversion. */
305 if (!real_isfinite (&tmp_real
)
306 || !real_isfinite (&tmp_imag
)
307 || (tmp_real
.cl
== rvc_zero
) != (mpfr_zero_p (mpc_realref (m
)) != 0)
308 || (tmp_imag
.cl
== rvc_zero
) != (mpfr_zero_p (mpc_imagref (m
)) != 0))
311 real_convert (result_real
, format
, &tmp_real
);
312 real_convert (result_imag
, format
, &tmp_imag
);
314 return (real_identical (result_real
, &tmp_real
)
315 && real_identical (result_imag
, &tmp_imag
));
322 in format FORMAT, given that FUNC is the mpc implementation of f.
323 Return true on success. Both RESULT and ARG are represented as
324 real and imaginary pairs. */
327 do_mpc_arg1 (real_value
*result_real
, real_value
*result_imag
,
328 int (*func
) (mpc_ptr
, mpc_srcptr
, mpc_rnd_t
),
329 const real_value
*arg_real
, const real_value
*arg_imag
,
330 const real_format
*format
)
332 /* To proceed, MPFR must exactly represent the target floating point
333 format, which only happens when the target base equals two. */
335 || !real_isfinite (arg_real
)
336 || !real_isfinite (arg_imag
))
339 int prec
= format
->p
;
340 mpc_rnd_t crnd
= format
->round_towards_zero
? MPC_RNDZZ
: MPC_RNDNN
;
344 mpfr_from_real (mpc_realref (m
), arg_real
, MPFR_RNDN
);
345 mpfr_from_real (mpc_imagref (m
), arg_imag
, MPFR_RNDN
);
347 bool inexact
= func (m
, m
, crnd
);
348 bool ok
= do_mpc_ckconv (result_real
, result_imag
, m
, inexact
, format
);
356 RESULT = f (ARG0, ARG1)
358 in format FORMAT, given that FUNC is the mpc implementation of f.
359 Return true on success. RESULT, ARG0 and ARG1 are represented as
360 real and imaginary pairs. */
363 do_mpc_arg2 (real_value
*result_real
, real_value
*result_imag
,
364 int (*func
)(mpc_ptr
, mpc_srcptr
, mpc_srcptr
, mpc_rnd_t
),
365 const real_value
*arg0_real
, const real_value
*arg0_imag
,
366 const real_value
*arg1_real
, const real_value
*arg1_imag
,
367 const real_format
*format
)
369 if (!real_isfinite (arg0_real
)
370 || !real_isfinite (arg0_imag
)
371 || !real_isfinite (arg1_real
)
372 || !real_isfinite (arg1_imag
))
375 int prec
= format
->p
;
376 mpc_rnd_t crnd
= format
->round_towards_zero
? MPC_RNDZZ
: MPC_RNDNN
;
379 mpc_init2 (m0
, prec
);
380 mpc_init2 (m1
, prec
);
381 mpfr_from_real (mpc_realref (m0
), arg0_real
, MPFR_RNDN
);
382 mpfr_from_real (mpc_imagref (m0
), arg0_imag
, MPFR_RNDN
);
383 mpfr_from_real (mpc_realref (m1
), arg1_real
, MPFR_RNDN
);
384 mpfr_from_real (mpc_imagref (m1
), arg1_imag
, MPFR_RNDN
);
386 bool inexact
= func (m0
, m0
, m1
, crnd
);
387 bool ok
= do_mpc_ckconv (result_real
, result_imag
, m0
, inexact
, format
);
396 *RESULT = logb (*ARG)
398 in format FORMAT. Return true on success. */
401 fold_const_logb (real_value
*result
, const real_value
*arg
,
402 const real_format
*format
)
407 /* If arg is +-NaN, then return it. */
412 /* If arg is +-Inf, then return +Inf. */
418 /* Zero may set errno and/or raise an exception. */
422 /* For normal numbers, proceed iff radix == 2. In GCC,
423 normalized significands are in the range [0.5, 1.0). We
424 want the exponent as if they were [1.0, 2.0) so get the
425 exponent and subtract 1. */
428 real_from_integer (result
, format
, REAL_EXP (arg
) - 1, SIGNED
);
438 *RESULT = significand (*ARG)
440 in format FORMAT. Return true on success. */
443 fold_const_significand (real_value
*result
, const real_value
*arg
,
444 const real_format
*format
)
451 /* If arg is +-0, +-Inf or +-NaN, then return it. */
456 /* For normal numbers, proceed iff radix == 2. */
460 /* In GCC, normalized significands are in the range [0.5, 1.0).
461 We want them to be [1.0, 2.0) so set the exponent to 1. */
462 SET_REAL_EXP (result
, 1);
474 where FORMAT is the format of *ARG and PRECISION is the number of
475 significant bits in the result. Return true on success. */
478 fold_const_conversion (wide_int
*result
,
479 void (*fn
) (real_value
*, format_helper
,
481 const real_value
*arg
, unsigned int precision
,
482 const real_format
*format
)
484 if (!real_isfinite (arg
))
488 fn (&rounded
, format
, arg
);
491 *result
= real_to_integer (&rounded
, &fail
, precision
);
497 *RESULT = pow (*ARG0, *ARG1)
499 in format FORMAT. Return true on success. */
502 fold_const_pow (real_value
*result
, const real_value
*arg0
,
503 const real_value
*arg1
, const real_format
*format
)
505 if (do_mpfr_arg2 (result
, mpfr_pow
, arg0
, arg1
, format
))
508 /* Check for an integer exponent. */
509 REAL_VALUE_TYPE cint1
;
510 HOST_WIDE_INT n1
= real_to_integer (arg1
);
511 real_from_integer (&cint1
, VOIDmode
, n1
, SIGNED
);
512 /* Attempt to evaluate pow at compile-time, unless this should
513 raise an exception. */
514 if (real_identical (arg1
, &cint1
)
516 || (!flag_trapping_math
&& !flag_errno_math
)
517 || !real_equal (arg0
, &dconst0
)))
519 bool inexact
= real_powi (result
, format
, arg0
, n1
);
520 /* Avoid the folding if flag_signaling_nans is on. */
521 if (flag_unsafe_math_optimizations
523 && !(flag_signaling_nans
524 && REAL_VALUE_ISSIGNALING_NAN (*arg0
))))
533 *RESULT = nextafter (*ARG0, *ARG1)
537 *RESULT = nexttoward (*ARG0, *ARG1)
539 in format FORMAT. Return true on success. */
542 fold_const_nextafter (real_value
*result
, const real_value
*arg0
,
543 const real_value
*arg1
, const real_format
*format
)
545 if (REAL_VALUE_ISSIGNALING_NAN (*arg0
)
546 || REAL_VALUE_ISSIGNALING_NAN (*arg1
))
549 /* Don't handle composite modes, nor decimal, nor modes without
550 inf or denorm at least for now. */
551 if (format
->pnan
< format
->p
554 || !format
->has_denorm
)
557 if (real_nextafter (result
, format
, arg0
, arg1
)
558 /* If raising underflow or overflow and setting errno to ERANGE,
559 fail if we care about those side-effects. */
560 && (flag_trapping_math
|| flag_errno_math
))
562 /* Similarly for nextafter (0, 1) raising underflow. */
563 else if (flag_trapping_math
564 && arg0
->cl
== rvc_zero
565 && result
->cl
!= rvc_zero
)
568 real_convert (result
, format
, result
);
575 *RESULT = ldexp (*ARG0, ARG1)
577 in format FORMAT. Return true on success. */
580 fold_const_builtin_load_exponent (real_value
*result
, const real_value
*arg0
,
581 const wide_int_ref
&arg1
,
582 const real_format
*format
)
584 /* Bound the maximum adjustment to twice the range of the
585 mode's valid exponents. Use abs to ensure the range is
586 positive as a sanity check. */
587 int max_exp_adj
= 2 * labs (format
->emax
- format
->emin
);
589 /* The requested adjustment must be inside this range. This
590 is a preliminary cap to avoid things like overflow, we
591 may still fail to compute the result for other reasons. */
592 if (wi::les_p (arg1
, -max_exp_adj
) || wi::ges_p (arg1
, max_exp_adj
))
595 /* Don't perform operation if we honor signaling NaNs and
596 operand is a signaling NaN. */
597 if (!flag_unsafe_math_optimizations
598 && flag_signaling_nans
599 && REAL_VALUE_ISSIGNALING_NAN (*arg0
))
602 REAL_VALUE_TYPE initial_result
;
603 real_ldexp (&initial_result
, arg0
, arg1
.to_shwi ());
605 /* Ensure we didn't overflow. */
606 if (real_isinf (&initial_result
))
609 /* Only proceed if the target mode can hold the
611 *result
= real_value_truncate (format
, initial_result
);
612 return real_equal (&initial_result
, result
);
615 /* Fold a call to __builtin_nan or __builtin_nans with argument ARG and
616 return type TYPE. QUIET is true if a quiet rather than signalling
620 fold_const_builtin_nan (tree type
, tree arg
, bool quiet
)
622 REAL_VALUE_TYPE real
;
623 const char *str
= c_getstr (arg
);
624 if (str
&& real_nan (&real
, str
, quiet
, TYPE_MODE (type
)))
625 return build_real (type
, real
);
629 /* Fold a call to IFN_REDUC_<CODE> (ARG), returning a value of type TYPE. */
632 fold_const_reduction (tree type
, tree arg
, tree_code code
)
634 unsigned HOST_WIDE_INT nelts
;
635 if (TREE_CODE (arg
) != VECTOR_CST
636 || !VECTOR_CST_NELTS (arg
).is_constant (&nelts
))
639 tree res
= VECTOR_CST_ELT (arg
, 0);
640 for (unsigned HOST_WIDE_INT i
= 1; i
< nelts
; i
++)
642 res
= const_binop (code
, type
, res
, VECTOR_CST_ELT (arg
, i
));
643 if (res
== NULL_TREE
|| !CONSTANT_CLASS_P (res
))
649 /* Fold a call to IFN_VEC_CONVERT (ARG) returning TYPE. */
652 fold_const_vec_convert (tree ret_type
, tree arg
)
654 enum tree_code code
= NOP_EXPR
;
655 tree arg_type
= TREE_TYPE (arg
);
656 if (TREE_CODE (arg
) != VECTOR_CST
)
659 gcc_checking_assert (VECTOR_TYPE_P (ret_type
) && VECTOR_TYPE_P (arg_type
));
661 if (INTEGRAL_TYPE_P (TREE_TYPE (ret_type
))
662 && SCALAR_FLOAT_TYPE_P (TREE_TYPE (arg_type
)))
663 code
= FIX_TRUNC_EXPR
;
664 else if (INTEGRAL_TYPE_P (TREE_TYPE (arg_type
))
665 && SCALAR_FLOAT_TYPE_P (TREE_TYPE (ret_type
)))
668 /* We can't handle steps directly when extending, since the
669 values need to wrap at the original precision first. */
671 = (INTEGRAL_TYPE_P (TREE_TYPE (ret_type
))
672 && INTEGRAL_TYPE_P (TREE_TYPE (arg_type
))
673 && (TYPE_PRECISION (TREE_TYPE (ret_type
))
674 <= TYPE_PRECISION (TREE_TYPE (arg_type
))));
675 tree_vector_builder elts
;
676 if (!elts
.new_unary_operation (ret_type
, arg
, step_ok_p
))
679 unsigned int count
= elts
.encoded_nelts ();
680 for (unsigned int i
= 0; i
< count
; ++i
)
682 tree elt
= fold_unary (code
, TREE_TYPE (ret_type
),
683 VECTOR_CST_ELT (arg
, i
));
684 if (elt
== NULL_TREE
|| !CONSTANT_CLASS_P (elt
))
686 elts
.quick_push (elt
);
689 return elts
.build ();
694 IFN_WHILE_ULT (ARG0, ARG1, (TYPE) { ... })
696 Return the value on success and null on failure. */
699 fold_while_ult (tree type
, poly_uint64 arg0
, poly_uint64 arg1
)
701 if (known_ge (arg0
, arg1
))
702 return build_zero_cst (type
);
704 if (maybe_ge (arg0
, arg1
))
707 poly_uint64 diff
= arg1
- arg0
;
708 poly_uint64 nelts
= TYPE_VECTOR_SUBPARTS (type
);
709 if (known_ge (diff
, nelts
))
710 return build_all_ones_cst (type
);
712 unsigned HOST_WIDE_INT const_diff
;
713 if (known_le (diff
, nelts
) && diff
.is_constant (&const_diff
))
715 tree minus_one
= build_minus_one_cst (TREE_TYPE (type
));
716 tree zero
= build_zero_cst (TREE_TYPE (type
));
717 return build_vector_a_then_b (type
, const_diff
, minus_one
, zero
);
726 in format FORMAT. Return true on success. */
729 fold_const_call_ss (real_value
*result
, combined_fn fn
,
730 const real_value
*arg
, const real_format
*format
)
736 return (real_compare (GE_EXPR
, arg
, &dconst0
)
737 && do_mpfr_arg1 (result
, mpfr_sqrt
, arg
, format
));
740 return do_mpfr_arg1 (result
, mpfr_cbrt
, arg
, format
);
743 return (real_compare (GE_EXPR
, arg
, &dconstm1
)
744 && real_compare (LE_EXPR
, arg
, &dconst1
)
745 && do_mpfr_arg1 (result
, mpfr_asin
, arg
, format
));
748 return (real_compare (GE_EXPR
, arg
, &dconstm1
)
749 && real_compare (LE_EXPR
, arg
, &dconst1
)
750 && do_mpfr_arg1 (result
, mpfr_acos
, arg
, format
));
753 return do_mpfr_arg1 (result
, mpfr_atan
, arg
, format
);
756 return do_mpfr_arg1 (result
, mpfr_asinh
, arg
, format
);
759 return (real_compare (GE_EXPR
, arg
, &dconst1
)
760 && do_mpfr_arg1 (result
, mpfr_acosh
, arg
, format
));
763 return (real_compare (GE_EXPR
, arg
, &dconstm1
)
764 && real_compare (LE_EXPR
, arg
, &dconst1
)
765 && do_mpfr_arg1 (result
, mpfr_atanh
, arg
, format
));
768 return do_mpfr_arg1 (result
, mpfr_sin
, arg
, format
);
771 return do_mpfr_arg1 (result
, mpfr_cos
, arg
, format
);
774 return do_mpfr_arg1 (result
, mpfr_tan
, arg
, format
);
777 return do_mpfr_arg1 (result
, mpfr_sinh
, arg
, format
);
780 return do_mpfr_arg1 (result
, mpfr_cosh
, arg
, format
);
783 return do_mpfr_arg1 (result
, mpfr_tanh
, arg
, format
);
786 return do_mpfr_arg1 (result
, mpfr_erf
, arg
, format
);
789 return do_mpfr_arg1 (result
, mpfr_erfc
, arg
, format
);
792 return do_mpfr_arg1 (result
, mpfr_gamma
, arg
, format
);
795 return do_mpfr_arg1 (result
, mpfr_exp
, arg
, format
);
798 return do_mpfr_arg1 (result
, mpfr_exp2
, arg
, format
);
802 return do_mpfr_arg1 (result
, mpfr_exp10
, arg
, format
);
805 return do_mpfr_arg1 (result
, mpfr_expm1
, arg
, format
);
808 return (real_compare (GT_EXPR
, arg
, &dconst0
)
809 && do_mpfr_arg1 (result
, mpfr_log
, arg
, format
));
812 return (real_compare (GT_EXPR
, arg
, &dconst0
)
813 && do_mpfr_arg1 (result
, mpfr_log2
, arg
, format
));
816 return (real_compare (GT_EXPR
, arg
, &dconst0
)
817 && do_mpfr_arg1 (result
, mpfr_log10
, arg
, format
));
820 return (real_compare (GT_EXPR
, arg
, &dconstm1
)
821 && do_mpfr_arg1 (result
, mpfr_log1p
, arg
, format
));
824 return do_mpfr_arg1 (result
, mpfr_j0
, arg
, format
);
827 return do_mpfr_arg1 (result
, mpfr_j1
, arg
, format
);
830 return (real_compare (GT_EXPR
, arg
, &dconst0
)
831 && do_mpfr_arg1 (result
, mpfr_y0
, arg
, format
));
834 return (real_compare (GT_EXPR
, arg
, &dconst0
)
835 && do_mpfr_arg1 (result
, mpfr_y1
, arg
, format
));
839 if (!REAL_VALUE_ISSIGNALING_NAN (*arg
))
841 real_floor (result
, format
, arg
);
848 if (!REAL_VALUE_ISSIGNALING_NAN (*arg
))
850 real_ceil (result
, format
, arg
);
857 if (!REAL_VALUE_ISSIGNALING_NAN (*arg
))
859 real_trunc (result
, format
, arg
);
866 if (!REAL_VALUE_ISSIGNALING_NAN (*arg
))
868 real_round (result
, format
, arg
);
874 CASE_CFN_ROUNDEVEN_FN
:
875 if (!REAL_VALUE_ISSIGNALING_NAN (*arg
))
877 real_roundeven (result
, format
, arg
);
883 return fold_const_logb (result
, arg
, format
);
885 CASE_CFN_SIGNIFICAND
:
886 return fold_const_significand (result
, arg
, format
);
897 where FORMAT is the format of ARG and PRECISION is the number of
898 significant bits in the result. Return true on success. */
901 fold_const_call_ss (wide_int
*result
, combined_fn fn
,
902 const real_value
*arg
, unsigned int precision
,
903 const real_format
*format
)
908 if (real_isneg (arg
))
909 *result
= wi::one (precision
);
911 *result
= wi::zero (precision
);
915 /* For ilogb we don't know FP_ILOGB0, so only handle normal values.
916 Proceed iff radix == 2. In GCC, normalized significands are in
917 the range [0.5, 1.0). We want the exponent as if they were
918 [1.0, 2.0) so get the exponent and subtract 1. */
919 if (arg
->cl
== rvc_normal
&& format
->b
== 2)
921 *result
= wi::shwi (REAL_EXP (arg
) - 1, precision
);
929 return fold_const_conversion (result
, real_ceil
, arg
,
935 return fold_const_conversion (result
, real_floor
, arg
,
941 return fold_const_conversion (result
, real_round
, arg
,
947 /* Not yet folded to a constant. */
951 case CFN_BUILT_IN_FINITED32
:
952 case CFN_BUILT_IN_FINITED64
:
953 case CFN_BUILT_IN_FINITED128
:
954 case CFN_BUILT_IN_ISFINITE
:
955 *result
= wi::shwi (real_isfinite (arg
) ? 1 : 0, precision
);
959 case CFN_BUILT_IN_ISINFD32
:
960 case CFN_BUILT_IN_ISINFD64
:
961 case CFN_BUILT_IN_ISINFD128
:
962 if (real_isinf (arg
))
963 *result
= wi::shwi (arg
->sign
? -1 : 1, precision
);
965 *result
= wi::shwi (0, precision
);
969 case CFN_BUILT_IN_ISNAND32
:
970 case CFN_BUILT_IN_ISNAND64
:
971 case CFN_BUILT_IN_ISNAND128
:
972 *result
= wi::shwi (real_isnan (arg
) ? 1 : 0, precision
);
984 where ARG_TYPE is the type of ARG and PRECISION is the number of bits
985 in the result. Return true on success. */
988 fold_const_call_ss (wide_int
*result
, combined_fn fn
, const wide_int_ref
&arg
,
989 unsigned int precision
, tree arg_type
)
994 *result
= wi::shwi (wi::ffs (arg
), precision
);
1000 if (wi::ne_p (arg
, 0))
1001 tmp
= wi::clz (arg
);
1002 else if (!CLZ_DEFINED_VALUE_AT_ZERO (SCALAR_INT_TYPE_MODE (arg_type
),
1004 tmp
= TYPE_PRECISION (arg_type
);
1005 *result
= wi::shwi (tmp
, precision
);
1012 if (wi::ne_p (arg
, 0))
1013 tmp
= wi::ctz (arg
);
1014 else if (!CTZ_DEFINED_VALUE_AT_ZERO (SCALAR_INT_TYPE_MODE (arg_type
),
1016 tmp
= TYPE_PRECISION (arg_type
);
1017 *result
= wi::shwi (tmp
, precision
);
1022 *result
= wi::shwi (wi::clrsb (arg
), precision
);
1026 *result
= wi::shwi (wi::popcount (arg
), precision
);
1030 *result
= wi::shwi (wi::parity (arg
), precision
);
1033 case CFN_BUILT_IN_BSWAP16
:
1034 case CFN_BUILT_IN_BSWAP32
:
1035 case CFN_BUILT_IN_BSWAP64
:
1036 case CFN_BUILT_IN_BSWAP128
:
1037 *result
= wide_int::from (arg
, precision
, TYPE_SIGN (arg_type
)).bswap ();
1049 where FORMAT is the format of ARG and of the real and imaginary parts
1050 of RESULT, passed as RESULT_REAL and RESULT_IMAG respectively. Return
1054 fold_const_call_cs (real_value
*result_real
, real_value
*result_imag
,
1055 combined_fn fn
, const real_value
*arg
,
1056 const real_format
*format
)
1061 /* cexpi(x+yi) = cos(x)+sin(y)*i. */
1062 return do_mpfr_sincos (result_imag
, result_real
, arg
, format
);
1073 where FORMAT is the format of RESULT and of the real and imaginary parts
1074 of ARG, passed as ARG_REAL and ARG_IMAG respectively. Return true on
1078 fold_const_call_sc (real_value
*result
, combined_fn fn
,
1079 const real_value
*arg_real
, const real_value
*arg_imag
,
1080 const real_format
*format
)
1085 return do_mpfr_arg2 (result
, mpfr_hypot
, arg_real
, arg_imag
, format
);
1096 where FORMAT is the format of the real and imaginary parts of RESULT
1097 (RESULT_REAL and RESULT_IMAG) and of ARG (ARG_REAL and ARG_IMAG).
1098 Return true on success. */
1101 fold_const_call_cc (real_value
*result_real
, real_value
*result_imag
,
1102 combined_fn fn
, const real_value
*arg_real
,
1103 const real_value
*arg_imag
, const real_format
*format
)
1108 return do_mpc_arg1 (result_real
, result_imag
, mpc_cos
,
1109 arg_real
, arg_imag
, format
);
1112 return do_mpc_arg1 (result_real
, result_imag
, mpc_cosh
,
1113 arg_real
, arg_imag
, format
);
1116 if (real_isinf (arg_real
) || real_isinf (arg_imag
))
1118 real_inf (result_real
);
1119 *result_imag
= dconst0
;
1120 result_imag
->sign
= arg_imag
->sign
;
1124 *result_real
= *arg_real
;
1125 *result_imag
= *arg_imag
;
1130 return do_mpc_arg1 (result_real
, result_imag
, mpc_sin
,
1131 arg_real
, arg_imag
, format
);
1134 return do_mpc_arg1 (result_real
, result_imag
, mpc_sinh
,
1135 arg_real
, arg_imag
, format
);
1138 return do_mpc_arg1 (result_real
, result_imag
, mpc_tan
,
1139 arg_real
, arg_imag
, format
);
1142 return do_mpc_arg1 (result_real
, result_imag
, mpc_tanh
,
1143 arg_real
, arg_imag
, format
);
1146 return do_mpc_arg1 (result_real
, result_imag
, mpc_log
,
1147 arg_real
, arg_imag
, format
);
1150 return do_mpc_arg1 (result_real
, result_imag
, mpc_sqrt
,
1151 arg_real
, arg_imag
, format
);
1154 return do_mpc_arg1 (result_real
, result_imag
, mpc_asin
,
1155 arg_real
, arg_imag
, format
);
1158 return do_mpc_arg1 (result_real
, result_imag
, mpc_acos
,
1159 arg_real
, arg_imag
, format
);
1162 return do_mpc_arg1 (result_real
, result_imag
, mpc_atan
,
1163 arg_real
, arg_imag
, format
);
1166 return do_mpc_arg1 (result_real
, result_imag
, mpc_asinh
,
1167 arg_real
, arg_imag
, format
);
1170 return do_mpc_arg1 (result_real
, result_imag
, mpc_acosh
,
1171 arg_real
, arg_imag
, format
);
1174 return do_mpc_arg1 (result_real
, result_imag
, mpc_atanh
,
1175 arg_real
, arg_imag
, format
);
1178 return do_mpc_arg1 (result_real
, result_imag
, mpc_exp
,
1179 arg_real
, arg_imag
, format
);
1186 /* Subroutine of fold_const_call, with the same interface. Handle cases
1187 where the arguments and result are numerical. */
1190 fold_const_call_1 (combined_fn fn
, tree type
, tree arg
)
1192 machine_mode mode
= TYPE_MODE (type
);
1193 machine_mode arg_mode
= TYPE_MODE (TREE_TYPE (arg
));
1195 if (integer_cst_p (arg
))
1197 if (SCALAR_INT_MODE_P (mode
))
1200 if (fold_const_call_ss (&result
, fn
, wi::to_wide (arg
),
1201 TYPE_PRECISION (type
), TREE_TYPE (arg
)))
1202 return wide_int_to_tree (type
, result
);
1207 if (real_cst_p (arg
))
1209 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg_mode
));
1210 if (mode
== arg_mode
)
1213 REAL_VALUE_TYPE result
;
1214 if (fold_const_call_ss (&result
, fn
, TREE_REAL_CST_PTR (arg
),
1215 REAL_MODE_FORMAT (mode
)))
1216 return build_real (type
, result
);
1218 else if (COMPLEX_MODE_P (mode
)
1219 && GET_MODE_INNER (mode
) == arg_mode
)
1221 /* real -> complex real. */
1222 REAL_VALUE_TYPE result_real
, result_imag
;
1223 if (fold_const_call_cs (&result_real
, &result_imag
, fn
,
1224 TREE_REAL_CST_PTR (arg
),
1225 REAL_MODE_FORMAT (arg_mode
)))
1226 return build_complex (type
,
1227 build_real (TREE_TYPE (type
), result_real
),
1228 build_real (TREE_TYPE (type
), result_imag
));
1230 else if (INTEGRAL_TYPE_P (type
))
1234 if (fold_const_call_ss (&result
, fn
,
1235 TREE_REAL_CST_PTR (arg
),
1236 TYPE_PRECISION (type
),
1237 REAL_MODE_FORMAT (arg_mode
)))
1238 return wide_int_to_tree (type
, result
);
1243 if (complex_cst_p (arg
))
1245 gcc_checking_assert (COMPLEX_MODE_P (arg_mode
));
1246 machine_mode inner_mode
= GET_MODE_INNER (arg_mode
);
1247 tree argr
= TREE_REALPART (arg
);
1248 tree argi
= TREE_IMAGPART (arg
);
1249 if (mode
== arg_mode
1250 && real_cst_p (argr
)
1251 && real_cst_p (argi
))
1253 /* complex real -> complex real. */
1254 REAL_VALUE_TYPE result_real
, result_imag
;
1255 if (fold_const_call_cc (&result_real
, &result_imag
, fn
,
1256 TREE_REAL_CST_PTR (argr
),
1257 TREE_REAL_CST_PTR (argi
),
1258 REAL_MODE_FORMAT (inner_mode
)))
1259 return build_complex (type
,
1260 build_real (TREE_TYPE (type
), result_real
),
1261 build_real (TREE_TYPE (type
), result_imag
));
1263 if (mode
== inner_mode
1264 && real_cst_p (argr
)
1265 && real_cst_p (argi
))
1267 /* complex real -> real. */
1268 REAL_VALUE_TYPE result
;
1269 if (fold_const_call_sc (&result
, fn
,
1270 TREE_REAL_CST_PTR (argr
),
1271 TREE_REAL_CST_PTR (argi
),
1272 REAL_MODE_FORMAT (inner_mode
)))
1273 return build_real (type
, result
);
1281 /* Try to fold FN (ARG) to a constant. Return the constant on success,
1282 otherwise return null. TYPE is the type of the return value. */
1285 fold_const_call (combined_fn fn
, tree type
, tree arg
)
1289 case CFN_BUILT_IN_STRLEN
:
1290 if (const char *str
= c_getstr (arg
))
1291 return build_int_cst (type
, strlen (str
));
1295 CASE_FLT_FN_FLOATN_NX (CFN_BUILT_IN_NAN
):
1296 case CFN_BUILT_IN_NAND32
:
1297 case CFN_BUILT_IN_NAND64
:
1298 case CFN_BUILT_IN_NAND128
:
1299 return fold_const_builtin_nan (type
, arg
, true);
1302 CASE_FLT_FN_FLOATN_NX (CFN_BUILT_IN_NANS
):
1303 case CFN_BUILT_IN_NANSD32
:
1304 case CFN_BUILT_IN_NANSD64
:
1305 case CFN_BUILT_IN_NANSD128
:
1306 return fold_const_builtin_nan (type
, arg
, false);
1308 case CFN_REDUC_PLUS
:
1309 return fold_const_reduction (type
, arg
, PLUS_EXPR
);
1312 return fold_const_reduction (type
, arg
, MAX_EXPR
);
1315 return fold_const_reduction (type
, arg
, MIN_EXPR
);
1318 return fold_const_reduction (type
, arg
, BIT_AND_EXPR
);
1321 return fold_const_reduction (type
, arg
, BIT_IOR_EXPR
);
1324 return fold_const_reduction (type
, arg
, BIT_XOR_EXPR
);
1326 case CFN_VEC_CONVERT
:
1327 return fold_const_vec_convert (type
, arg
);
1330 return fold_const_call_1 (fn
, type
, arg
);
1334 /* Fold a call to IFN_FOLD_LEFT_<CODE> (ARG0, ARG1), returning a value
1338 fold_const_fold_left (tree type
, tree arg0
, tree arg1
, tree_code code
)
1340 if (TREE_CODE (arg1
) != VECTOR_CST
)
1343 unsigned HOST_WIDE_INT nelts
;
1344 if (!VECTOR_CST_NELTS (arg1
).is_constant (&nelts
))
1347 for (unsigned HOST_WIDE_INT i
= 0; i
< nelts
; i
++)
1349 arg0
= const_binop (code
, type
, arg0
, VECTOR_CST_ELT (arg1
, i
));
1350 if (arg0
== NULL_TREE
|| !CONSTANT_CLASS_P (arg0
))
1358 *RESULT = FN (*ARG0, *ARG1)
1360 in format FORMAT. Return true on success. */
1363 fold_const_call_sss (real_value
*result
, combined_fn fn
,
1364 const real_value
*arg0
, const real_value
*arg1
,
1365 const real_format
*format
)
1371 return do_mpfr_arg2 (result
, mpfr_remainder
, arg0
, arg1
, format
);
1374 return do_mpfr_arg2 (result
, mpfr_atan2
, arg0
, arg1
, format
);
1377 return do_mpfr_arg2 (result
, mpfr_dim
, arg0
, arg1
, format
);
1380 return do_mpfr_arg2 (result
, mpfr_hypot
, arg0
, arg1
, format
);
1383 CASE_CFN_COPYSIGN_FN
:
1385 real_copysign (result
, arg1
);
1390 return do_mpfr_arg2 (result
, mpfr_min
, arg0
, arg1
, format
);
1394 return do_mpfr_arg2 (result
, mpfr_max
, arg0
, arg1
, format
);
1397 return fold_const_pow (result
, arg0
, arg1
, format
);
1400 CASE_CFN_NEXTTOWARD
:
1401 return fold_const_nextafter (result
, arg0
, arg1
, format
);
1410 *RESULT = FN (*ARG0, ARG1)
1412 where FORMAT is the format of *RESULT and *ARG0. Return true on
1416 fold_const_call_sss (real_value
*result
, combined_fn fn
,
1417 const real_value
*arg0
, const wide_int_ref
&arg1
,
1418 const real_format
*format
)
1423 return fold_const_builtin_load_exponent (result
, arg0
, arg1
, format
);
1427 return (format
->b
== 2
1428 && fold_const_builtin_load_exponent (result
, arg0
, arg1
,
1432 /* Avoid the folding if flag_signaling_nans is on and
1433 operand is a signaling NaN. */
1434 if (!flag_unsafe_math_optimizations
1435 && flag_signaling_nans
1436 && REAL_VALUE_ISSIGNALING_NAN (*arg0
))
1439 real_powi (result
, format
, arg0
, arg1
.to_shwi ());
1449 *RESULT = FN (ARG0, *ARG1)
1451 where FORMAT is the format of *RESULT and *ARG1. Return true on
1455 fold_const_call_sss (real_value
*result
, combined_fn fn
,
1456 const wide_int_ref
&arg0
, const real_value
*arg1
,
1457 const real_format
*format
)
1462 return do_mpfr_arg2 (result
, mpfr_jn
, arg0
, arg1
, format
);
1465 return (real_compare (GT_EXPR
, arg1
, &dconst0
)
1466 && do_mpfr_arg2 (result
, mpfr_yn
, arg0
, arg1
, format
));
1475 RESULT = fn (ARG0, ARG1)
1477 where FORMAT is the format of the real and imaginary parts of RESULT
1478 (RESULT_REAL and RESULT_IMAG), of ARG0 (ARG0_REAL and ARG0_IMAG)
1479 and of ARG1 (ARG1_REAL and ARG1_IMAG). Return true on success. */
1482 fold_const_call_ccc (real_value
*result_real
, real_value
*result_imag
,
1483 combined_fn fn
, const real_value
*arg0_real
,
1484 const real_value
*arg0_imag
, const real_value
*arg1_real
,
1485 const real_value
*arg1_imag
, const real_format
*format
)
1490 return do_mpc_arg2 (result_real
, result_imag
, mpc_pow
,
1491 arg0_real
, arg0_imag
, arg1_real
, arg1_imag
, format
);
1498 /* Subroutine of fold_const_call, with the same interface. Handle cases
1499 where the arguments and result are numerical. */
1502 fold_const_call_1 (combined_fn fn
, tree type
, tree arg0
, tree arg1
)
1504 machine_mode mode
= TYPE_MODE (type
);
1505 machine_mode arg0_mode
= TYPE_MODE (TREE_TYPE (arg0
));
1506 machine_mode arg1_mode
= TYPE_MODE (TREE_TYPE (arg1
));
1508 if (mode
== arg0_mode
1509 && real_cst_p (arg0
)
1510 && real_cst_p (arg1
))
1512 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg0_mode
));
1513 REAL_VALUE_TYPE result
;
1514 if (arg0_mode
== arg1_mode
)
1516 /* real, real -> real. */
1517 if (fold_const_call_sss (&result
, fn
, TREE_REAL_CST_PTR (arg0
),
1518 TREE_REAL_CST_PTR (arg1
),
1519 REAL_MODE_FORMAT (mode
)))
1520 return build_real (type
, result
);
1522 else if (arg1_mode
== TYPE_MODE (long_double_type_node
))
1525 CASE_CFN_NEXTTOWARD
:
1526 /* real, long double -> real. */
1527 if (fold_const_call_sss (&result
, fn
, TREE_REAL_CST_PTR (arg0
),
1528 TREE_REAL_CST_PTR (arg1
),
1529 REAL_MODE_FORMAT (mode
)))
1530 return build_real (type
, result
);
1538 if (real_cst_p (arg0
)
1539 && integer_cst_p (arg1
))
1541 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg0_mode
));
1542 if (mode
== arg0_mode
)
1544 /* real, int -> real. */
1545 REAL_VALUE_TYPE result
;
1546 if (fold_const_call_sss (&result
, fn
, TREE_REAL_CST_PTR (arg0
),
1548 REAL_MODE_FORMAT (mode
)))
1549 return build_real (type
, result
);
1554 if (integer_cst_p (arg0
)
1555 && real_cst_p (arg1
))
1557 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg1_mode
));
1558 if (mode
== arg1_mode
)
1560 /* int, real -> real. */
1561 REAL_VALUE_TYPE result
;
1562 if (fold_const_call_sss (&result
, fn
, wi::to_wide (arg0
),
1563 TREE_REAL_CST_PTR (arg1
),
1564 REAL_MODE_FORMAT (mode
)))
1565 return build_real (type
, result
);
1570 if (arg0_mode
== arg1_mode
1571 && complex_cst_p (arg0
)
1572 && complex_cst_p (arg1
))
1574 gcc_checking_assert (COMPLEX_MODE_P (arg0_mode
));
1575 machine_mode inner_mode
= GET_MODE_INNER (arg0_mode
);
1576 tree arg0r
= TREE_REALPART (arg0
);
1577 tree arg0i
= TREE_IMAGPART (arg0
);
1578 tree arg1r
= TREE_REALPART (arg1
);
1579 tree arg1i
= TREE_IMAGPART (arg1
);
1580 if (mode
== arg0_mode
1581 && real_cst_p (arg0r
)
1582 && real_cst_p (arg0i
)
1583 && real_cst_p (arg1r
)
1584 && real_cst_p (arg1i
))
1586 /* complex real, complex real -> complex real. */
1587 REAL_VALUE_TYPE result_real
, result_imag
;
1588 if (fold_const_call_ccc (&result_real
, &result_imag
, fn
,
1589 TREE_REAL_CST_PTR (arg0r
),
1590 TREE_REAL_CST_PTR (arg0i
),
1591 TREE_REAL_CST_PTR (arg1r
),
1592 TREE_REAL_CST_PTR (arg1i
),
1593 REAL_MODE_FORMAT (inner_mode
)))
1594 return build_complex (type
,
1595 build_real (TREE_TYPE (type
), result_real
),
1596 build_real (TREE_TYPE (type
), result_imag
));
1604 /* Try to fold FN (ARG0, ARG1) to a constant. Return the constant on success,
1605 otherwise return null. TYPE is the type of the return value. */
1608 fold_const_call (combined_fn fn
, tree type
, tree arg0
, tree arg1
)
1610 const char *p0
, *p1
;
1614 case CFN_BUILT_IN_STRSPN
:
1615 if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1616 return build_int_cst (type
, strspn (p0
, p1
));
1619 case CFN_BUILT_IN_STRCSPN
:
1620 if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1621 return build_int_cst (type
, strcspn (p0
, p1
));
1624 case CFN_BUILT_IN_STRCMP
:
1625 if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1626 return build_cmp_result (type
, strcmp (p0
, p1
));
1629 case CFN_BUILT_IN_STRCASECMP
:
1630 if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1632 int r
= strcmp (p0
, p1
);
1634 return build_cmp_result (type
, r
);
1638 case CFN_BUILT_IN_INDEX
:
1639 case CFN_BUILT_IN_STRCHR
:
1640 if ((p0
= c_getstr (arg0
)) && target_char_cst_p (arg1
, &c
))
1642 const char *r
= strchr (p0
, c
);
1644 return build_int_cst (type
, 0);
1645 return fold_convert (type
,
1646 fold_build_pointer_plus_hwi (arg0
, r
- p0
));
1650 case CFN_BUILT_IN_RINDEX
:
1651 case CFN_BUILT_IN_STRRCHR
:
1652 if ((p0
= c_getstr (arg0
)) && target_char_cst_p (arg1
, &c
))
1654 const char *r
= strrchr (p0
, c
);
1656 return build_int_cst (type
, 0);
1657 return fold_convert (type
,
1658 fold_build_pointer_plus_hwi (arg0
, r
- p0
));
1662 case CFN_BUILT_IN_STRSTR
:
1663 if ((p1
= c_getstr (arg1
)))
1665 if ((p0
= c_getstr (arg0
)))
1667 const char *r
= strstr (p0
, p1
);
1669 return build_int_cst (type
, 0);
1670 return fold_convert (type
,
1671 fold_build_pointer_plus_hwi (arg0
, r
- p0
));
1674 return fold_convert (type
, arg0
);
1678 case CFN_FOLD_LEFT_PLUS
:
1679 return fold_const_fold_left (type
, arg0
, arg1
, PLUS_EXPR
);
1682 return fold_const_call_1 (fn
, type
, arg0
, arg1
);
1688 *RESULT = FN (*ARG0, *ARG1, *ARG2)
1690 in format FORMAT. Return true on success. */
1693 fold_const_call_ssss (real_value
*result
, combined_fn fn
,
1694 const real_value
*arg0
, const real_value
*arg1
,
1695 const real_value
*arg2
, const real_format
*format
)
1701 return do_mpfr_arg3 (result
, mpfr_fma
, arg0
, arg1
, arg2
, format
);
1705 real_value new_arg2
= real_value_negate (arg2
);
1706 return do_mpfr_arg3 (result
, mpfr_fma
, arg0
, arg1
, &new_arg2
, format
);
1711 real_value new_arg0
= real_value_negate (arg0
);
1712 return do_mpfr_arg3 (result
, mpfr_fma
, &new_arg0
, arg1
, arg2
, format
);
1717 real_value new_arg0
= real_value_negate (arg0
);
1718 real_value new_arg2
= real_value_negate (arg2
);
1719 return do_mpfr_arg3 (result
, mpfr_fma
, &new_arg0
, arg1
,
1728 /* Subroutine of fold_const_call, with the same interface. Handle cases
1729 where the arguments and result are numerical. */
1732 fold_const_call_1 (combined_fn fn
, tree type
, tree arg0
, tree arg1
, tree arg2
)
1734 machine_mode mode
= TYPE_MODE (type
);
1735 machine_mode arg0_mode
= TYPE_MODE (TREE_TYPE (arg0
));
1736 machine_mode arg1_mode
= TYPE_MODE (TREE_TYPE (arg1
));
1737 machine_mode arg2_mode
= TYPE_MODE (TREE_TYPE (arg2
));
1739 if (arg0_mode
== arg1_mode
1740 && arg0_mode
== arg2_mode
1741 && real_cst_p (arg0
)
1742 && real_cst_p (arg1
)
1743 && real_cst_p (arg2
))
1745 gcc_checking_assert (SCALAR_FLOAT_MODE_P (arg0_mode
));
1746 if (mode
== arg0_mode
)
1748 /* real, real, real -> real. */
1749 REAL_VALUE_TYPE result
;
1750 if (fold_const_call_ssss (&result
, fn
, TREE_REAL_CST_PTR (arg0
),
1751 TREE_REAL_CST_PTR (arg1
),
1752 TREE_REAL_CST_PTR (arg2
),
1753 REAL_MODE_FORMAT (mode
)))
1754 return build_real (type
, result
);
1762 /* Try to fold FN (ARG0, ARG1, ARG2) to a constant. Return the constant on
1763 success, otherwise return null. TYPE is the type of the return value. */
1766 fold_const_call (combined_fn fn
, tree type
, tree arg0
, tree arg1
, tree arg2
)
1768 const char *p0
, *p1
;
1770 unsigned HOST_WIDE_INT s0
, s1
;
1774 case CFN_BUILT_IN_STRNCMP
:
1775 if (!host_size_t_cst_p (arg2
, &s2
))
1778 && !TREE_SIDE_EFFECTS (arg0
)
1779 && !TREE_SIDE_EFFECTS (arg1
))
1780 return build_int_cst (type
, 0);
1781 else if ((p0
= c_getstr (arg0
)) && (p1
= c_getstr (arg1
)))
1782 return build_int_cst (type
, strncmp (p0
, p1
, s2
));
1785 case CFN_BUILT_IN_STRNCASECMP
:
1786 if (!host_size_t_cst_p (arg2
, &s2
))
1789 && !TREE_SIDE_EFFECTS (arg0
)
1790 && !TREE_SIDE_EFFECTS (arg1
))
1791 return build_int_cst (type
, 0);
1792 else if ((p0
= c_getstr (arg0
))
1793 && (p1
= c_getstr (arg1
))
1794 && strncmp (p0
, p1
, s2
) == 0)
1795 return build_int_cst (type
, 0);
1798 case CFN_BUILT_IN_BCMP
:
1799 case CFN_BUILT_IN_MEMCMP
:
1800 if (!host_size_t_cst_p (arg2
, &s2
))
1803 && !TREE_SIDE_EFFECTS (arg0
)
1804 && !TREE_SIDE_EFFECTS (arg1
))
1805 return build_int_cst (type
, 0);
1806 if ((p0
= getbyterep (arg0
, &s0
))
1807 && (p1
= getbyterep (arg1
, &s1
))
1810 return build_cmp_result (type
, memcmp (p0
, p1
, s2
));
1813 case CFN_BUILT_IN_MEMCHR
:
1814 if (!host_size_t_cst_p (arg2
, &s2
))
1817 && !TREE_SIDE_EFFECTS (arg0
)
1818 && !TREE_SIDE_EFFECTS (arg1
))
1819 return build_int_cst (type
, 0);
1820 if ((p0
= getbyterep (arg0
, &s0
))
1822 && target_char_cst_p (arg1
, &c
))
1824 const char *r
= (const char *) memchr (p0
, c
, s2
);
1826 return build_int_cst (type
, 0);
1827 return fold_convert (type
,
1828 fold_build_pointer_plus_hwi (arg0
, r
- p0
));
1834 poly_uint64 parg0
, parg1
;
1835 if (poly_int_tree_p (arg0
, &parg0
) && poly_int_tree_p (arg1
, &parg1
))
1836 return fold_while_ult (type
, parg0
, parg1
);
1841 return fold_const_call_1 (fn
, type
, arg0
, arg1
, arg2
);