Coverage Report

Created: 2026-03-31 06:37

next uncovered line (L), next uncovered region (R), next uncovered branch (B)
/src/gmp/mpz/powm.c
Line
Count
Source
1
/* mpz_powm(res,base,exp,mod) -- Set R to (U^E) mod M.
2
3
   Contributed to the GNU project by Torbjorn Granlund.
4
5
Copyright 1991, 1993, 1994, 1996, 1997, 2000-2002, 2005, 2008, 2009,
6
2011, 2012, 2015, 2019 Free Software Foundation, Inc.
7
8
This file is part of the GNU MP Library.
9
10
The GNU MP Library is free software; you can redistribute it and/or modify
11
it under the terms of either:
12
13
  * the GNU Lesser General Public License as published by the Free
14
    Software Foundation; either version 3 of the License, or (at your
15
    option) any later version.
16
17
or
18
19
  * the GNU General Public License as published by the Free Software
20
    Foundation; either version 2 of the License, or (at your option) any
21
    later version.
22
23
or both in parallel, as here.
24
25
The GNU MP Library is distributed in the hope that it will be useful, but
26
WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
27
or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
28
for more details.
29
30
You should have received copies of the GNU General Public License and the
31
GNU Lesser General Public License along with the GNU MP Library.  If not,
32
see https://www.gnu.org/licenses/.  */
33
34
35
#include "gmp-impl.h"
36
#include "longlong.h"
37
38
39
/* TODO
40
41
 * Improve handling of buffers.  It is pretty ugly now.
42
43
 * For even moduli, we compute a binvert of its odd part both here and in
44
   mpn_powm.  How can we avoid this recomputation?
45
*/
46
47
/*
48
  b ^ e mod m   res
49
  0   0     0    ?
50
  0   e     0    ?
51
  0   0     m    ?
52
  0   e     m    0
53
  b   0     0    ?
54
  b   e     0    ?
55
  b   0     m    1 mod m
56
  b   e     m    b^e mod m
57
*/
58
59
#define HANDLE_NEGATIVE_EXPONENT 1
60
61
void
62
mpz_powm (mpz_ptr r, mpz_srcptr b, mpz_srcptr e, mpz_srcptr m)
63
44.1k
{
64
44.1k
  mp_size_t n, nodd, ncnt;
65
44.1k
  int cnt;
66
44.1k
  mp_ptr rp, tp;
67
44.1k
  mp_srcptr bp, ep, mp;
68
44.1k
  mp_size_t rn, bn, es, en, itch;
69
44.1k
  mpz_t new_b;      /* note: value lives long via 'b' */
70
44.1k
  TMP_DECL;
71
72
44.1k
  n = ABSIZ(m);
73
44.1k
  if (UNLIKELY (n == 0))
74
0
    DIVIDE_BY_ZERO;
75
76
44.1k
  mp = PTR(m);
77
78
44.1k
  TMP_MARK;
79
80
44.1k
  es = SIZ(e);
81
44.1k
  if (UNLIKELY (es <= 0))
82
50
    {
83
50
      if (es == 0)
84
50
  {
85
    /* b^0 mod m,  b is anything and m is non-zero.
86
       Result is 1 mod m, i.e., 1 or 0 depending on if m = 1.  */
87
50
    SIZ(r) = n != 1 || mp[0] != 1;
88
50
    MPZ_NEWALLOC (r, 1)[0] = 1;
89
50
    TMP_FREE; /* we haven't really allocated anything here */
90
50
    return;
91
50
  }
92
0
#if HANDLE_NEGATIVE_EXPONENT
93
0
      MPZ_TMP_INIT (new_b, n + 1);
94
95
0
      if (UNLIKELY (! mpz_invert (new_b, b, m)))
96
0
  DIVIDE_BY_ZERO;
97
0
      b = new_b;
98
0
      es = -es;
99
#else
100
      DIVIDE_BY_ZERO;
101
#endif
102
0
    }
103
44.0k
  en = es;
104
105
44.0k
  bn = ABSIZ(b);
106
107
44.0k
  if (UNLIKELY (bn == 0))
108
2
    {
109
2
      SIZ(r) = 0;
110
2
      TMP_FREE;
111
2
      return;
112
2
    }
113
114
44.0k
  ep = PTR(e);
115
116
  /* Handle (b^1 mod m) early, since mpn_pow* do not handle that case.  */
117
44.0k
  if (UNLIKELY (en == 1 && ep[0] == 1))
118
179
    {
119
179
      rp = TMP_ALLOC_LIMBS (n);
120
179
      bp = PTR(b);
121
179
      if (bn >= n)
122
141
  {
123
141
    mp_ptr qp = TMP_ALLOC_LIMBS (bn - n + 1);
124
141
    mpn_tdiv_qr (qp, rp, 0L, bp, bn, mp, n);
125
141
    rn = n;
126
141
    MPN_NORMALIZE (rp, rn);
127
128
141
    if (rn != 0 && SIZ(b) < 0)
129
0
      {
130
0
        mpn_sub (rp, mp, n, rp, rn);
131
0
        rn = n;
132
0
        MPN_NORMALIZE_NOT_ZERO (rp, rn);
133
0
      }
134
141
  }
135
38
      else
136
38
  {
137
38
    if (SIZ(b) < 0)
138
0
      {
139
0
        mpn_sub (rp, mp, n, bp, bn);
140
0
        rn = n;
141
0
        MPN_NORMALIZE_NOT_ZERO (rp, rn);
142
0
      }
143
38
    else
144
38
      {
145
38
        MPN_COPY (rp, bp, bn);
146
38
        rn = bn;
147
38
      }
148
38
  }
149
179
      goto ret;
150
179
    }
151
152
  /* Remove low zero limbs from M.  This loop will terminate for correctly
153
     represented mpz numbers.  */
154
43.8k
  ncnt = 0;
155
43.8k
  while (UNLIKELY (mp[0] == 0))
156
166k
    {
157
166k
      mp++;
158
166k
      ncnt++;
159
166k
    }
160
43.8k
  nodd = n - ncnt;
161
43.8k
  cnt = 0;
162
43.8k
  if (mp[0] % 2 == 0)
163
8.20k
    {
164
8.20k
      mp_ptr newmp = TMP_ALLOC_LIMBS (nodd);
165
8.20k
      count_trailing_zeros (cnt, mp[0]);
166
8.20k
      mpn_rshift (newmp, mp, nodd, cnt);
167
8.20k
      nodd -= newmp[nodd - 1] == 0;
168
8.20k
      mp = newmp;
169
8.20k
      ncnt++;
170
8.20k
    }
171
172
43.8k
  if (ncnt != 0)
173
8.74k
    {
174
      /* We will call both mpn_powm and mpn_powlo.  */
175
      /* rp needs n, mpn_powlo needs 4n, the 2 mpn_binvert might need more */
176
8.74k
      mp_size_t n_largest_binvert = MAX (ncnt, nodd);
177
8.74k
      mp_size_t itch_binvert = mpn_binvert_itch (n_largest_binvert);
178
8.74k
      itch = 3 * n + MAX (itch_binvert, 2 * n);
179
8.74k
    }
180
35.1k
  else
181
35.1k
    {
182
      /* We will call just mpn_powm.  */
183
35.1k
      mp_size_t itch_binvert = mpn_binvert_itch (nodd);
184
35.1k
      itch = n + MAX (itch_binvert, 2 * n);
185
35.1k
    }
186
43.8k
  tp = TMP_ALLOC_LIMBS (itch);
187
188
43.8k
  rp = tp;  tp += n;
189
190
43.8k
  bp = PTR(b);
191
43.8k
  mpn_powm (rp, bp, bn, ep, en, mp, nodd, tp);
192
193
43.8k
  rn = n;
194
195
43.8k
  if (ncnt != 0)
196
8.74k
    {
197
8.74k
      mp_ptr r2, xp, yp, odd_inv_2exp;
198
8.74k
      unsigned long t;
199
8.74k
      int bcnt;
200
201
8.74k
      if (bn < ncnt)
202
4.59k
  {
203
4.59k
    mp_ptr newbp = TMP_ALLOC_LIMBS (ncnt);
204
4.59k
    MPN_COPY (newbp, bp, bn);
205
4.59k
    MPN_ZERO (newbp + bn, ncnt - bn);
206
4.59k
    bp = newbp;
207
4.59k
  }
208
209
8.74k
      r2 = tp;
210
211
8.74k
      if (bp[0] % 2 == 0)
212
6.28k
  {
213
6.28k
    if (en > 1)
214
5.21k
      {
215
5.21k
        MPN_ZERO (r2, ncnt);
216
5.21k
        goto zero;
217
5.21k
      }
218
219
1.07k
    ASSERT (en == 1);
220
1.07k
    t = (ncnt - (cnt != 0)) * GMP_NUMB_BITS + cnt;
221
222
    /* Count number of low zero bits in B, up to 3.  */
223
1.07k
    bcnt = (0x1213 >> ((bp[0] & 7) << 1)) & 0x3;
224
    /* Note that ep[0] * bcnt might overflow, but that just results
225
       in a missed optimization.  */
226
1.07k
    if (ep[0] * bcnt >= t)
227
1.05k
      {
228
1.05k
        MPN_ZERO (r2, ncnt);
229
1.05k
        goto zero;
230
1.05k
      }
231
1.07k
  }
232
233
2.47k
      mpn_powlo (r2, bp, ep, en, ncnt, tp + ncnt);
234
235
8.74k
    zero:
236
8.74k
      if (nodd < ncnt)
237
2.39k
  {
238
2.39k
    mp_ptr newmp = TMP_ALLOC_LIMBS (ncnt);
239
2.39k
    MPN_COPY (newmp, mp, nodd);
240
2.39k
    MPN_ZERO (newmp + nodd, ncnt - nodd);
241
2.39k
    mp = newmp;
242
2.39k
  }
243
244
8.74k
      odd_inv_2exp = tp + n;
245
8.74k
      mpn_binvert (odd_inv_2exp, mp, ncnt, tp + 2 * n);
246
247
8.74k
      mpn_sub (r2, r2, ncnt, rp, nodd > ncnt ? ncnt : nodd);
248
249
8.74k
      xp = tp + 2 * n;
250
8.74k
      mpn_mullo_n (xp, odd_inv_2exp, r2, ncnt);
251
252
8.74k
      if (cnt != 0)
253
8.20k
  xp[ncnt - 1] &= (CNST_LIMB(1) << cnt) - 1;
254
255
8.74k
      yp = tp;
256
8.74k
      if (ncnt > nodd)
257
2.39k
  mpn_mul (yp, xp, ncnt, mp, nodd);
258
6.34k
      else
259
6.34k
  mpn_mul (yp, mp, nodd, xp, ncnt);
260
261
8.74k
      mpn_add (rp, yp, n, rp, nodd);
262
263
8.74k
      ASSERT (nodd + ncnt >= n);
264
8.74k
      ASSERT (nodd + ncnt <= n + 1);
265
8.74k
    }
266
267
43.8k
  MPN_NORMALIZE (rp, rn);
268
269
43.8k
  if ((ep[0] & 1) && SIZ(b) < 0 && rn != 0)
270
0
    {
271
0
      mpn_sub (rp, PTR(m), n, rp, rn);
272
0
      rn = n;
273
0
      MPN_NORMALIZE (rp, rn);
274
0
    }
275
276
44.0k
 ret:
277
44.0k
  MPZ_NEWALLOC (r, rn);
278
44.0k
  SIZ(r) = rn;
279
44.0k
  MPN_COPY (PTR(r), rp, rn);
280
281
44.0k
  TMP_FREE;
282
44.0k
}