Coverage Report

Created: 2026-02-14 06:49

next uncovered line (L), next uncovered region (R), next uncovered branch (B)
/src/gmp/mpz/powm.c
Line
Count
Source
1
/* mpz_powm(res,base,exp,mod) -- Set R to (U^E) mod M.
2
3
   Contributed to the GNU project by Torbjorn Granlund.
4
5
Copyright 1991, 1993, 1994, 1996, 1997, 2000-2002, 2005, 2008, 2009,
6
2011, 2012, 2015, 2019 Free Software Foundation, Inc.
7
8
This file is part of the GNU MP Library.
9
10
The GNU MP Library is free software; you can redistribute it and/or modify
11
it under the terms of either:
12
13
  * the GNU Lesser General Public License as published by the Free
14
    Software Foundation; either version 3 of the License, or (at your
15
    option) any later version.
16
17
or
18
19
  * the GNU General Public License as published by the Free Software
20
    Foundation; either version 2 of the License, or (at your option) any
21
    later version.
22
23
or both in parallel, as here.
24
25
The GNU MP Library is distributed in the hope that it will be useful, but
26
WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
27
or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
28
for more details.
29
30
You should have received copies of the GNU General Public License and the
31
GNU Lesser General Public License along with the GNU MP Library.  If not,
32
see https://www.gnu.org/licenses/.  */
33
34
35
#include "gmp-impl.h"
36
#include "longlong.h"
37
38
39
/* TODO
40
41
 * Improve handling of buffers.  It is pretty ugly now.
42
43
 * For even moduli, we compute a binvert of its odd part both here and in
44
   mpn_powm.  How can we avoid this recomputation?
45
*/
46
47
/*
48
  b ^ e mod m   res
49
  0   0     0    ?
50
  0   e     0    ?
51
  0   0     m    ?
52
  0   e     m    0
53
  b   0     0    ?
54
  b   e     0    ?
55
  b   0     m    1 mod m
56
  b   e     m    b^e mod m
57
*/
58
59
#define HANDLE_NEGATIVE_EXPONENT 1
60
61
void
62
mpz_powm (mpz_ptr r, mpz_srcptr b, mpz_srcptr e, mpz_srcptr m)
63
39.4k
{
64
39.4k
  mp_size_t n, nodd, ncnt;
65
39.4k
  int cnt;
66
39.4k
  mp_ptr rp, tp;
67
39.4k
  mp_srcptr bp, ep, mp;
68
39.4k
  mp_size_t rn, bn, es, en, itch;
69
39.4k
  mpz_t new_b;      /* note: value lives long via 'b' */
70
39.4k
  TMP_DECL;
71
72
39.4k
  n = ABSIZ(m);
73
39.4k
  if (UNLIKELY (n == 0))
74
0
    DIVIDE_BY_ZERO;
75
76
39.4k
  mp = PTR(m);
77
78
39.4k
  TMP_MARK;
79
80
39.4k
  es = SIZ(e);
81
39.4k
  if (UNLIKELY (es <= 0))
82
50
    {
83
50
      if (es == 0)
84
50
  {
85
    /* b^0 mod m,  b is anything and m is non-zero.
86
       Result is 1 mod m, i.e., 1 or 0 depending on if m = 1.  */
87
50
    SIZ(r) = n != 1 || mp[0] != 1;
88
50
    MPZ_NEWALLOC (r, 1)[0] = 1;
89
50
    TMP_FREE; /* we haven't really allocated anything here */
90
50
    return;
91
50
  }
92
0
#if HANDLE_NEGATIVE_EXPONENT
93
0
      MPZ_TMP_INIT (new_b, n + 1);
94
95
0
      if (UNLIKELY (! mpz_invert (new_b, b, m)))
96
0
  DIVIDE_BY_ZERO;
97
0
      b = new_b;
98
0
      es = -es;
99
#else
100
      DIVIDE_BY_ZERO;
101
#endif
102
0
    }
103
39.3k
  en = es;
104
105
39.3k
  bn = ABSIZ(b);
106
107
39.3k
  if (UNLIKELY (bn == 0))
108
2
    {
109
2
      SIZ(r) = 0;
110
2
      TMP_FREE;
111
2
      return;
112
2
    }
113
114
39.3k
  ep = PTR(e);
115
116
  /* Handle (b^1 mod m) early, since mpn_pow* do not handle that case.  */
117
39.3k
  if (UNLIKELY (en == 1 && ep[0] == 1))
118
163
    {
119
163
      rp = TMP_ALLOC_LIMBS (n);
120
163
      bp = PTR(b);
121
163
      if (bn >= n)
122
130
  {
123
130
    mp_ptr qp = TMP_ALLOC_LIMBS (bn - n + 1);
124
130
    mpn_tdiv_qr (qp, rp, 0L, bp, bn, mp, n);
125
130
    rn = n;
126
130
    MPN_NORMALIZE (rp, rn);
127
128
130
    if (rn != 0 && SIZ(b) < 0)
129
0
      {
130
0
        mpn_sub (rp, mp, n, rp, rn);
131
0
        rn = n;
132
0
        MPN_NORMALIZE_NOT_ZERO (rp, rn);
133
0
      }
134
130
  }
135
33
      else
136
33
  {
137
33
    if (SIZ(b) < 0)
138
0
      {
139
0
        mpn_sub (rp, mp, n, bp, bn);
140
0
        rn = n;
141
0
        MPN_NORMALIZE_NOT_ZERO (rp, rn);
142
0
      }
143
33
    else
144
33
      {
145
33
        MPN_COPY (rp, bp, bn);
146
33
        rn = bn;
147
33
      }
148
33
  }
149
163
      goto ret;
150
163
    }
151
152
  /* Remove low zero limbs from M.  This loop will terminate for correctly
153
     represented mpz numbers.  */
154
39.2k
  ncnt = 0;
155
39.2k
  while (UNLIKELY (mp[0] == 0))
156
152k
    {
157
152k
      mp++;
158
152k
      ncnt++;
159
152k
    }
160
39.2k
  nodd = n - ncnt;
161
39.2k
  cnt = 0;
162
39.2k
  if (mp[0] % 2 == 0)
163
7.77k
    {
164
7.77k
      mp_ptr newmp = TMP_ALLOC_LIMBS (nodd);
165
7.77k
      count_trailing_zeros (cnt, mp[0]);
166
7.77k
      mpn_rshift (newmp, mp, nodd, cnt);
167
7.77k
      nodd -= newmp[nodd - 1] == 0;
168
7.77k
      mp = newmp;
169
7.77k
      ncnt++;
170
7.77k
    }
171
172
39.2k
  if (ncnt != 0)
173
8.27k
    {
174
      /* We will call both mpn_powm and mpn_powlo.  */
175
      /* rp needs n, mpn_powlo needs 4n, the 2 mpn_binvert might need more */
176
8.27k
      mp_size_t n_largest_binvert = MAX (ncnt, nodd);
177
8.27k
      mp_size_t itch_binvert = mpn_binvert_itch (n_largest_binvert);
178
8.27k
      itch = 3 * n + MAX (itch_binvert, 2 * n);
179
8.27k
    }
180
30.9k
  else
181
30.9k
    {
182
      /* We will call just mpn_powm.  */
183
30.9k
      mp_size_t itch_binvert = mpn_binvert_itch (nodd);
184
30.9k
      itch = n + MAX (itch_binvert, 2 * n);
185
30.9k
    }
186
39.2k
  tp = TMP_ALLOC_LIMBS (itch);
187
188
39.2k
  rp = tp;  tp += n;
189
190
39.2k
  bp = PTR(b);
191
39.2k
  mpn_powm (rp, bp, bn, ep, en, mp, nodd, tp);
192
193
39.2k
  rn = n;
194
195
39.2k
  if (ncnt != 0)
196
8.27k
    {
197
8.27k
      mp_ptr r2, xp, yp, odd_inv_2exp;
198
8.27k
      unsigned long t;
199
8.27k
      int bcnt;
200
201
8.27k
      if (bn < ncnt)
202
4.49k
  {
203
4.49k
    mp_ptr newbp = TMP_ALLOC_LIMBS (ncnt);
204
4.49k
    MPN_COPY (newbp, bp, bn);
205
4.49k
    MPN_ZERO (newbp + bn, ncnt - bn);
206
4.49k
    bp = newbp;
207
4.49k
  }
208
209
8.27k
      r2 = tp;
210
211
8.27k
      if (bp[0] % 2 == 0)
212
5.85k
  {
213
5.85k
    if (en > 1)
214
4.93k
      {
215
4.93k
        MPN_ZERO (r2, ncnt);
216
4.93k
        goto zero;
217
4.93k
      }
218
219
924
    ASSERT (en == 1);
220
924
    t = (ncnt - (cnt != 0)) * GMP_NUMB_BITS + cnt;
221
222
    /* Count number of low zero bits in B, up to 3.  */
223
924
    bcnt = (0x1213 >> ((bp[0] & 7) << 1)) & 0x3;
224
    /* Note that ep[0] * bcnt might overflow, but that just results
225
       in a missed optimization.  */
226
924
    if (ep[0] * bcnt >= t)
227
902
      {
228
902
        MPN_ZERO (r2, ncnt);
229
902
        goto zero;
230
902
      }
231
924
  }
232
233
2.44k
      mpn_powlo (r2, bp, ep, en, ncnt, tp + ncnt);
234
235
8.27k
    zero:
236
8.27k
      if (nodd < ncnt)
237
2.47k
  {
238
2.47k
    mp_ptr newmp = TMP_ALLOC_LIMBS (ncnt);
239
2.47k
    MPN_COPY (newmp, mp, nodd);
240
2.47k
    MPN_ZERO (newmp + nodd, ncnt - nodd);
241
2.47k
    mp = newmp;
242
2.47k
  }
243
244
8.27k
      odd_inv_2exp = tp + n;
245
8.27k
      mpn_binvert (odd_inv_2exp, mp, ncnt, tp + 2 * n);
246
247
8.27k
      mpn_sub (r2, r2, ncnt, rp, nodd > ncnt ? ncnt : nodd);
248
249
8.27k
      xp = tp + 2 * n;
250
8.27k
      mpn_mullo_n (xp, odd_inv_2exp, r2, ncnt);
251
252
8.27k
      if (cnt != 0)
253
7.77k
  xp[ncnt - 1] &= (CNST_LIMB(1) << cnt) - 1;
254
255
8.27k
      yp = tp;
256
8.27k
      if (ncnt > nodd)
257
2.47k
  mpn_mul (yp, xp, ncnt, mp, nodd);
258
5.80k
      else
259
5.80k
  mpn_mul (yp, mp, nodd, xp, ncnt);
260
261
8.27k
      mpn_add (rp, yp, n, rp, nodd);
262
263
8.27k
      ASSERT (nodd + ncnt >= n);
264
8.27k
      ASSERT (nodd + ncnt <= n + 1);
265
8.27k
    }
266
267
39.2k
  MPN_NORMALIZE (rp, rn);
268
269
39.2k
  if ((ep[0] & 1) && SIZ(b) < 0 && rn != 0)
270
0
    {
271
0
      mpn_sub (rp, PTR(m), n, rp, rn);
272
0
      rn = n;
273
0
      MPN_NORMALIZE (rp, rn);
274
0
    }
275
276
39.3k
 ret:
277
39.3k
  MPZ_NEWALLOC (r, rn);
278
39.3k
  SIZ(r) = rn;
279
39.3k
  MPN_COPY (PTR(r), rp, rn);
280
281
39.3k
  TMP_FREE;
282
39.3k
}