1 /* sqrmod_bnm1.c -- squaring mod B^n-1.
3 Contributed to the GNU project by Niels Möller, Torbjorn Granlund and
6 THE FUNCTIONS IN THIS FILE ARE INTERNAL WITH MUTABLE INTERFACES. IT IS ONLY
7 SAFE TO REACH THEM THROUGH DOCUMENTED INTERFACES. IN FACT, IT IS ALMOST
8 GUARANTEED THAT THEY WILL CHANGE OR DISAPPEAR IN A FUTURE GNU MP RELEASE.
10 Copyright 2009, 2010, 2012 Free Software Foundation, Inc.
12 This file is part of the GNU MP Library.
14 The GNU MP Library is free software; you can redistribute it and/or modify
15 it under the terms of either:
17 * the GNU Lesser General Public License as published by the Free
18 Software Foundation; either version 3 of the License, or (at your
19 option) any later version.
23 * the GNU General Public License as published by the Free Software
24 Foundation; either version 2 of the License, or (at your option) any
27 or both in parallel, as here.
29 The GNU MP Library is distributed in the hope that it will be useful, but
30 WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
31 or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
34 You should have received copies of the GNU General Public License and the
35 GNU Lesser General Public License along with the GNU MP Library. If not,
36 see https://www.gnu.org/licenses/. */
43 /* Input is {ap,rn}; output is {rp,rn}, computation is
44 mod B^rn - 1, and values are semi-normalised; zero is represented
45 as either 0 or B^n - 1. Needs a scratch of 2rn limbs at tp.
48 mpn_bc_sqrmod_bnm1 (mp_ptr rp
, mp_srcptr ap
, mp_size_t rn
, mp_ptr tp
)
55 cy
= mpn_add_n (rp
, tp
, tp
+ rn
, rn
);
56 /* If cy == 1, then the value of rp is at most B^rn - 2, so there can
57 * be no overflow when adding in the carry. */
58 MPN_INCR_U (rp
, rn
, cy
);
62 /* Input is {ap,rn+1}; output is {rp,rn+1}, in
63 semi-normalised representation, computation is mod B^rn + 1. Needs
64 a scratch area of 2rn + 2 limbs at tp; tp == rp is allowed.
65 Output is normalised. */
67 mpn_bc_sqrmod_bnp1 (mp_ptr rp
, mp_srcptr ap
, mp_size_t rn
, mp_ptr tp
)
73 mpn_sqr (tp
, ap
, rn
+ 1);
74 ASSERT (tp
[2*rn
+1] == 0);
75 ASSERT (tp
[2*rn
] < GMP_NUMB_MAX
);
76 cy
= tp
[2*rn
] + mpn_sub_n (rp
, tp
, tp
+rn
, rn
);
78 MPN_INCR_U (rp
, rn
+1, cy
);
82 /* Computes {rp,MIN(rn,2an)} <- {ap,an}^2 Mod(B^rn-1)
84 * The result is expected to be ZERO if and only if the operand
85 * already is. Otherwise the class [0] Mod(B^rn-1) is represented by
87 * It should not be a problem if sqrmod_bnm1 is used to
88 * compute the full square with an <= 2*rn, because this condition
89 * implies (B^an-1)^2 < (B^rn-1) .
91 * Requires rn/4 < an <= rn
92 * Scratch need: rn/2 + (need for recursive call OR rn + 3). This gives
94 * S(n) <= rn/2 + MAX (rn + 4, S(n/2)) <= 3/2 rn + 4
97 mpn_sqrmod_bnm1 (mp_ptr rp
, mp_size_t rn
, mp_srcptr ap
, mp_size_t an
, mp_ptr tp
)
102 if ((rn
& 1) != 0 || BELOW_THRESHOLD (rn
, SQRMOD_BNM1_THRESHOLD
))
104 if (UNLIKELY (an
< rn
))
106 if (UNLIKELY (2*an
<= rn
))
108 mpn_sqr (rp
, ap
, an
);
113 mpn_sqr (tp
, ap
, an
);
114 cy
= mpn_add (rp
, tp
, rn
, tp
+ rn
, 2*an
- rn
);
115 MPN_INCR_U (rp
, rn
, cy
);
119 mpn_bc_sqrmod_bnm1 (rp
, ap
, rn
, tp
);
131 /* Compute xm = a^2 mod (B^n - 1), xp = a^2 mod (B^n + 1)
134 x = -xp * B^n + (B^n + 1) * [ (xp + xm)/2 mod (B^n-1)]
140 #define xp tp /* 2n + 2 */
141 /* am1 maybe in {xp, n} */
142 #define sp1 (tp + 2*n + 2)
143 /* ap1 maybe in {sp1, n + 1} */
154 cy
= mpn_add (xp
, a0
, n
, a1
, an
- n
);
155 MPN_INCR_U (xp
, n
, cy
);
165 mpn_sqrmod_bnm1 (rp
, n
, am1
, anm
, so
);
173 if (LIKELY (an
> n
)) {
175 cy
= mpn_sub (sp1
, a0
, n
, a1
, an
- n
);
177 MPN_INCR_U (sp1
, n
+ 1, cy
);
184 if (BELOW_THRESHOLD (n
, MUL_FFT_MODF_THRESHOLD
))
189 k
= mpn_fft_best_k (n
, 1);
191 while (n
& mask
) {k
--; mask
>>=1;};
193 if (k
>= FFT_FIRST_K
)
194 xp
[n
] = mpn_mul_fft (xp
, n
, ap1
, anp
, ap1
, anp
, k
);
195 else if (UNLIKELY (ap1
== a0
))
199 mpn_sqr (xp
, a0
, an
);
201 cy
= mpn_sub (xp
, xp
, n
, xp
+ n
, anp
);
203 MPN_INCR_U (xp
, n
+1, cy
);
206 mpn_bc_sqrmod_bnp1 (xp
, ap1
, n
, xp
);
209 /* Here the CRT recomposition begins.
211 xm <- (xp + xm)/2 = (xp + xm)B^n/2 mod (B^n-1)
212 Division by 2 is a bitwise rotation.
214 Assumes xp normalised mod (B^n+1).
216 The residue class [0] is represented by [B^n-1]; except when
220 #if HAVE_NATIVE_mpn_rsh1add_n || HAVE_NATIVE_mpn_rsh1add_nc
221 #if HAVE_NATIVE_mpn_rsh1add_nc
222 cy
= mpn_rsh1add_nc(rp
, rp
, xp
, n
, xp
[n
]); /* B^n = 1 */
223 hi
= cy
<< (GMP_NUMB_BITS
- 1);
225 /* next update of rp[n-1] will set cy = 1 only if rp[n-1]+=hi
226 overflows, i.e. a further increment will not overflow again. */
228 cy
= xp
[n
] + mpn_rsh1add_n(rp
, rp
, xp
, n
); /* B^n = 1 */
229 hi
= (cy
<<(GMP_NUMB_BITS
-1))&GMP_NUMB_MASK
; /* (cy&1) << ... */
231 /* cy = 1 only if xp[n] = 1 i.e. {xp,n} = ZERO, this implies that
232 the rsh1add was a simple rshift: the top bit is 0. cy=1 => hi=0. */
234 #if GMP_NAIL_BITS == 0
235 add_ssaaaa(cy
, rp
[n
-1], cy
, rp
[n
-1], CNST_LIMB(0), hi
);
237 cy
+= (hi
& rp
[n
-1]) >> (GMP_NUMB_BITS
-1);
240 #else /* ! HAVE_NATIVE_mpn_rsh1add_n */
241 #if HAVE_NATIVE_mpn_add_nc
242 cy
= mpn_add_nc(rp
, rp
, xp
, n
, xp
[n
]);
244 cy
= xp
[n
] + mpn_add_n(rp
, rp
, xp
, n
); /* xp[n] == 1 implies {xp,n} == ZERO */
247 mpn_rshift(rp
, rp
, n
, 1);
249 hi
= (cy
<<(GMP_NUMB_BITS
-1))&GMP_NUMB_MASK
; /* (cy&1) << ... */
251 /* We can have cy != 0 only if hi = 0... */
252 ASSERT ((rp
[n
-1] & GMP_NUMB_HIGHBIT
) == 0);
254 /* ... rp[n-1] + cy can not overflow, the following INCR is correct. */
257 /* Next increment can not overflow, read the previous comments about cy. */
258 ASSERT ((cy
== 0) || ((rp
[n
-1] & GMP_NUMB_HIGHBIT
) == 0));
259 MPN_INCR_U(rp
, n
, cy
);
261 /* Compute the highest half:
262 ([(xp + xm)/2 mod (B^n-1)] - xp ) * B^n
264 if (UNLIKELY (2*an
< rn
))
266 /* Note that in this case, the only way the result can equal
267 zero mod B^{rn} - 1 is if the input is zero, and
268 then the output of both the recursive calls and this CRT
269 reconstruction is zero, not B^{rn} - 1. */
270 cy
= mpn_sub_n (rp
+ n
, rp
, xp
, 2*an
- n
);
272 /* FIXME: This subtraction of the high parts is not really
273 necessary, we do it to get the carry out, and for sanity
275 cy
= xp
[n
] + mpn_sub_nc (xp
+ 2*an
- n
, rp
+ 2*an
- n
,
276 xp
+ 2*an
- n
, rn
- 2*an
, cy
);
277 ASSERT (mpn_zero_p (xp
+ 2*an
- n
+1, rn
- 1 - 2*an
));
278 cy
= mpn_sub_1 (rp
, rp
, 2*an
, cy
);
279 ASSERT (cy
== (xp
+ 2*an
- n
)[0]);
283 cy
= xp
[n
] + mpn_sub_n (rp
+ n
, rp
, xp
, n
);
284 /* cy = 1 only if {xp,n+1} is not ZERO, i.e. {rp,n} is not ZERO.
285 DECR will affect _at most_ the lowest n limbs. */
286 MPN_DECR_U (rp
, 2*n
, cy
);
296 mpn_sqrmod_bnm1_next_size (mp_size_t n
)
300 if (BELOW_THRESHOLD (n
, SQRMOD_BNM1_THRESHOLD
))
302 if (BELOW_THRESHOLD (n
, 4 * (SQRMOD_BNM1_THRESHOLD
- 1) + 1))
303 return (n
+ (2-1)) & (-2);
304 if (BELOW_THRESHOLD (n
, 8 * (SQRMOD_BNM1_THRESHOLD
- 1) + 1))
305 return (n
+ (4-1)) & (-4);
309 if (BELOW_THRESHOLD (nh
, SQR_FFT_MODF_THRESHOLD
))
310 return (n
+ (8-1)) & (-8);
312 return 2 * mpn_fft_next_size (nh
, mpn_fft_best_k (nh
, 1));