/src/gmp/mpn/toom53_mul.c
Line | Count | Source (jump to first uncovered line) |
1 | | /* mpn_toom53_mul -- Multiply {ap,an} and {bp,bn} where an is nominally 5/3 |
2 | | times as large as bn. Or more accurately, (4/3)bn < an < (5/2)bn. |
3 | | |
4 | | Contributed to the GNU project by Torbjorn Granlund and Marco Bodrato. |
5 | | |
6 | | The idea of applying toom to unbalanced multiplication is due to Marco |
7 | | Bodrato and Alberto Zanoni. |
8 | | |
9 | | THE FUNCTION IN THIS FILE IS INTERNAL WITH A MUTABLE INTERFACE. IT IS ONLY |
10 | | SAFE TO REACH IT THROUGH DOCUMENTED INTERFACES. IN FACT, IT IS ALMOST |
11 | | GUARANTEED THAT IT WILL CHANGE OR DISAPPEAR IN A FUTURE GNU MP RELEASE. |
12 | | |
13 | | Copyright 2006-2008, 2012, 2014, 2015 Free Software Foundation, Inc. |
14 | | |
15 | | This file is part of the GNU MP Library. |
16 | | |
17 | | The GNU MP Library is free software; you can redistribute it and/or modify |
18 | | it under the terms of either: |
19 | | |
20 | | * the GNU Lesser General Public License as published by the Free |
21 | | Software Foundation; either version 3 of the License, or (at your |
22 | | option) any later version. |
23 | | |
24 | | or |
25 | | |
26 | | * the GNU General Public License as published by the Free Software |
27 | | Foundation; either version 2 of the License, or (at your option) any |
28 | | later version. |
29 | | |
30 | | or both in parallel, as here. |
31 | | |
32 | | The GNU MP Library is distributed in the hope that it will be useful, but |
33 | | WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY |
34 | | or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
35 | | for more details. |
36 | | |
37 | | You should have received copies of the GNU General Public License and the |
38 | | GNU Lesser General Public License along with the GNU MP Library. If not, |
39 | | see https://www.gnu.org/licenses/. */ |
40 | | |
41 | | |
42 | | #include "gmp-impl.h" |
43 | | |
44 | | /* Evaluate in: 0, +1, -1, +2, -2, 1/2, +inf |
45 | | |
46 | | <-s-><--n--><--n--><--n--><--n--> |
47 | | ___ ______ ______ ______ ______ |
48 | | |a4_|___a3_|___a2_|___a1_|___a0_| |
49 | | |__b2|___b1_|___b0_| |
50 | | <-t--><--n--><--n--> |
51 | | |
52 | | v0 = a0 * b0 # A(0)*B(0) |
53 | | v1 = ( a0+ a1+ a2+ a3+ a4)*( b0+ b1+ b2) # A(1)*B(1) ah <= 4 bh <= 2 |
54 | | vm1 = ( a0- a1+ a2- a3+ a4)*( b0- b1+ b2) # A(-1)*B(-1) |ah| <= 2 bh <= 1 |
55 | | v2 = ( a0+2a1+4a2+8a3+16a4)*( b0+2b1+4b2) # A(2)*B(2) ah <= 30 bh <= 6 |
56 | | vm2 = ( a0-2a1+4a2-8a3+16a4)*( b0-2b1+4b2) # A(2)*B(2) -9<=ah<=20 -1<=bh<=4 |
57 | | vh = (16a0+8a1+4a2+2a3+ a4)*(4b0+2b1+ b2) # A(1/2)*B(1/2) ah <= 30 bh <= 6 |
58 | | vinf= a4 * b2 # A(inf)*B(inf) |
59 | | */ |
60 | | |
61 | | void |
62 | | mpn_toom53_mul (mp_ptr pp, |
63 | | mp_srcptr ap, mp_size_t an, |
64 | | mp_srcptr bp, mp_size_t bn, |
65 | | mp_ptr scratch) |
66 | 0 | { |
67 | 0 | mp_size_t n, s, t; |
68 | 0 | mp_limb_t cy; |
69 | 0 | mp_ptr gp; |
70 | 0 | mp_ptr as1, asm1, as2, asm2, ash; |
71 | 0 | mp_ptr bs1, bsm1, bs2, bsm2, bsh; |
72 | 0 | mp_ptr tmp; |
73 | 0 | enum toom7_flags flags; |
74 | 0 | TMP_DECL; |
75 | |
|
76 | 0 | #define a0 ap |
77 | 0 | #define a1 (ap + n) |
78 | 0 | #define a2 (ap + 2*n) |
79 | 0 | #define a3 (ap + 3*n) |
80 | 0 | #define a4 (ap + 4*n) |
81 | 0 | #define b0 bp |
82 | 0 | #define b1 (bp + n) |
83 | 0 | #define b2 (bp + 2*n) |
84 | |
|
85 | 0 | n = 1 + (3 * an >= 5 * bn ? (an - 1) / (size_t) 5 : (bn - 1) / (size_t) 3); |
86 | |
|
87 | 0 | s = an - 4 * n; |
88 | 0 | t = bn - 2 * n; |
89 | |
|
90 | 0 | ASSERT (0 < s && s <= n); |
91 | 0 | ASSERT (0 < t && t <= n); |
92 | |
|
93 | 0 | TMP_MARK; |
94 | |
|
95 | 0 | tmp = TMP_ALLOC_LIMBS (10 * (n + 1)); |
96 | 0 | as1 = tmp; tmp += n + 1; |
97 | 0 | asm1 = tmp; tmp += n + 1; |
98 | 0 | as2 = tmp; tmp += n + 1; |
99 | 0 | asm2 = tmp; tmp += n + 1; |
100 | 0 | ash = tmp; tmp += n + 1; |
101 | 0 | bs1 = tmp; tmp += n + 1; |
102 | 0 | bsm1 = tmp; tmp += n + 1; |
103 | 0 | bs2 = tmp; tmp += n + 1; |
104 | 0 | bsm2 = tmp; tmp += n + 1; |
105 | 0 | bsh = tmp; tmp += n + 1; |
106 | |
|
107 | 0 | gp = pp; |
108 | | |
109 | | /* Compute as1 and asm1. */ |
110 | 0 | flags = (enum toom7_flags) (toom7_w3_neg & mpn_toom_eval_pm1 (as1, asm1, 4, ap, n, s, gp)); |
111 | | |
112 | | /* Compute as2 and asm2. */ |
113 | 0 | flags = (enum toom7_flags) (flags | (toom7_w1_neg & mpn_toom_eval_pm2 (as2, asm2, 4, ap, n, s, gp))); |
114 | | |
115 | | /* Compute ash = 16 a0 + 8 a1 + 4 a2 + 2 a3 + a4 |
116 | | = 2*(2*(2*(2*a0 + a1) + a2) + a3) + a4 */ |
117 | 0 | #if HAVE_NATIVE_mpn_addlsh1_n |
118 | 0 | cy = mpn_addlsh1_n (ash, a1, a0, n); |
119 | 0 | cy = 2*cy + mpn_addlsh1_n (ash, a2, ash, n); |
120 | 0 | cy = 2*cy + mpn_addlsh1_n (ash, a3, ash, n); |
121 | 0 | if (s < n) |
122 | 0 | { |
123 | 0 | mp_limb_t cy2; |
124 | 0 | cy2 = mpn_addlsh1_n (ash, a4, ash, s); |
125 | 0 | ash[n] = 2*cy + mpn_lshift (ash + s, ash + s, n - s, 1); |
126 | 0 | MPN_INCR_U (ash + s, n+1-s, cy2); |
127 | 0 | } |
128 | 0 | else |
129 | 0 | ash[n] = 2*cy + mpn_addlsh1_n (ash, a4, ash, n); |
130 | | #else |
131 | | cy = mpn_lshift (ash, a0, n, 1); |
132 | | cy += mpn_add_n (ash, ash, a1, n); |
133 | | cy = 2*cy + mpn_lshift (ash, ash, n, 1); |
134 | | cy += mpn_add_n (ash, ash, a2, n); |
135 | | cy = 2*cy + mpn_lshift (ash, ash, n, 1); |
136 | | cy += mpn_add_n (ash, ash, a3, n); |
137 | | cy = 2*cy + mpn_lshift (ash, ash, n, 1); |
138 | | ash[n] = cy + mpn_add (ash, ash, n, a4, s); |
139 | | #endif |
140 | | |
141 | | /* Compute bs1 and bsm1. */ |
142 | 0 | bs1[n] = mpn_add (bs1, b0, n, b2, t); /* b0 + b2 */ |
143 | | #if HAVE_NATIVE_mpn_add_n_sub_n |
144 | | if (bs1[n] == 0 && mpn_cmp (bs1, b1, n) < 0) |
145 | | { |
146 | | bs1[n] = mpn_add_n_sub_n (bs1, bsm1, b1, bs1, n) >> 1; |
147 | | bsm1[n] = 0; |
148 | | flags = (enum toom7_flags) (flags ^ toom7_w3_neg); |
149 | | } |
150 | | else |
151 | | { |
152 | | cy = mpn_add_n_sub_n (bs1, bsm1, bs1, b1, n); |
153 | | bsm1[n] = bs1[n] - (cy & 1); |
154 | | bs1[n] += (cy >> 1); |
155 | | } |
156 | | #else |
157 | 0 | if (bs1[n] == 0 && mpn_cmp (bs1, b1, n) < 0) |
158 | 0 | { |
159 | 0 | mpn_sub_n (bsm1, b1, bs1, n); |
160 | 0 | bsm1[n] = 0; |
161 | 0 | flags = (enum toom7_flags) (flags ^ toom7_w3_neg); |
162 | 0 | } |
163 | 0 | else |
164 | 0 | { |
165 | 0 | bsm1[n] = bs1[n] - mpn_sub_n (bsm1, bs1, b1, n); |
166 | 0 | } |
167 | 0 | bs1[n] += mpn_add_n (bs1, bs1, b1, n); /* b0+b1+b2 */ |
168 | 0 | #endif |
169 | | |
170 | | /* Compute bs2 and bsm2. */ |
171 | 0 | #if HAVE_NATIVE_mpn_addlsh_n || HAVE_NATIVE_mpn_addlsh2_n |
172 | 0 | #if HAVE_NATIVE_mpn_addlsh2_n |
173 | 0 | cy = mpn_addlsh2_n (bs2, b0, b2, t); |
174 | | #else /* HAVE_NATIVE_mpn_addlsh_n */ |
175 | | cy = mpn_addlsh_n (bs2, b0, b2, t, 2); |
176 | | #endif |
177 | 0 | if (t < n) |
178 | 0 | cy = mpn_add_1 (bs2 + t, b0 + t, n - t, cy); |
179 | 0 | bs2[n] = cy; |
180 | | #else |
181 | | cy = mpn_lshift (gp, b2, t, 2); |
182 | | bs2[n] = mpn_add (bs2, b0, n, gp, t); |
183 | | MPN_INCR_U (bs2 + t, n+1-t, cy); |
184 | | #endif |
185 | |
|
186 | 0 | gp[n] = mpn_lshift (gp, b1, n, 1); |
187 | |
|
188 | | #if HAVE_NATIVE_mpn_add_n_sub_n |
189 | | if (mpn_cmp (bs2, gp, n+1) < 0) |
190 | | { |
191 | | ASSERT_NOCARRY (mpn_add_n_sub_n (bs2, bsm2, gp, bs2, n+1)); |
192 | | flags = (enum toom7_flags) (flags ^ toom7_w1_neg); |
193 | | } |
194 | | else |
195 | | { |
196 | | ASSERT_NOCARRY (mpn_add_n_sub_n (bs2, bsm2, bs2, gp, n+1)); |
197 | | } |
198 | | #else |
199 | 0 | if (mpn_cmp (bs2, gp, n+1) < 0) |
200 | 0 | { |
201 | 0 | ASSERT_NOCARRY (mpn_sub_n (bsm2, gp, bs2, n+1)); |
202 | 0 | flags = (enum toom7_flags) (flags ^ toom7_w1_neg); |
203 | 0 | } |
204 | 0 | else |
205 | 0 | { |
206 | 0 | ASSERT_NOCARRY (mpn_sub_n (bsm2, bs2, gp, n+1)); |
207 | 0 | } |
208 | 0 | mpn_add_n (bs2, bs2, gp, n+1); |
209 | 0 | #endif |
210 | | |
211 | | /* Compute bsh = 4 b0 + 2 b1 + b2 = 2*(2*b0 + b1)+b2. */ |
212 | 0 | #if HAVE_NATIVE_mpn_addlsh1_n |
213 | 0 | cy = mpn_addlsh1_n (bsh, b1, b0, n); |
214 | 0 | if (t < n) |
215 | 0 | { |
216 | 0 | mp_limb_t cy2; |
217 | 0 | cy2 = mpn_addlsh1_n (bsh, b2, bsh, t); |
218 | 0 | bsh[n] = 2*cy + mpn_lshift (bsh + t, bsh + t, n - t, 1); |
219 | 0 | MPN_INCR_U (bsh + t, n+1-t, cy2); |
220 | 0 | } |
221 | 0 | else |
222 | 0 | bsh[n] = 2*cy + mpn_addlsh1_n (bsh, b2, bsh, n); |
223 | | #else |
224 | | cy = mpn_lshift (bsh, b0, n, 1); |
225 | | cy += mpn_add_n (bsh, bsh, b1, n); |
226 | | cy = 2*cy + mpn_lshift (bsh, bsh, n, 1); |
227 | | bsh[n] = cy + mpn_add (bsh, bsh, n, b2, t); |
228 | | #endif |
229 | |
|
230 | 0 | ASSERT (as1[n] <= 4); |
231 | 0 | ASSERT (bs1[n] <= 2); |
232 | 0 | ASSERT (asm1[n] <= 2); |
233 | 0 | ASSERT (bsm1[n] <= 1); |
234 | 0 | ASSERT (as2[n] <= 30); |
235 | 0 | ASSERT (bs2[n] <= 6); |
236 | 0 | ASSERT (asm2[n] <= 20); |
237 | 0 | ASSERT (bsm2[n] <= 4); |
238 | 0 | ASSERT (ash[n] <= 30); |
239 | 0 | ASSERT (bsh[n] <= 6); |
240 | |
|
241 | 0 | #define v0 pp /* 2n */ |
242 | 0 | #define v1 (pp + 2 * n) /* 2n+1 */ |
243 | 0 | #define vinf (pp + 6 * n) /* s+t */ |
244 | 0 | #define v2 scratch /* 2n+1 */ |
245 | 0 | #define vm2 (scratch + 2 * n + 1) /* 2n+1 */ |
246 | 0 | #define vh (scratch + 4 * n + 2) /* 2n+1 */ |
247 | 0 | #define vm1 (scratch + 6 * n + 3) /* 2n+1 */ |
248 | 0 | #define scratch_out (scratch + 8 * n + 4) /* 2n+1 */ |
249 | | /* Total scratch need: 10*n+5 */ |
250 | | |
251 | | /* Must be in allocation order, as they overwrite one limb beyond |
252 | | * 2n+1. */ |
253 | 0 | mpn_mul_n (v2, as2, bs2, n + 1); /* v2, 2n+1 limbs */ |
254 | 0 | mpn_mul_n (vm2, asm2, bsm2, n + 1); /* vm2, 2n+1 limbs */ |
255 | 0 | mpn_mul_n (vh, ash, bsh, n + 1); /* vh, 2n+1 limbs */ |
256 | | |
257 | | /* vm1, 2n+1 limbs */ |
258 | | #ifdef SMALLER_RECURSION |
259 | | mpn_mul_n (vm1, asm1, bsm1, n); |
260 | | if (asm1[n] == 1) |
261 | | { |
262 | | cy = bsm1[n] + mpn_add_n (vm1 + n, vm1 + n, bsm1, n); |
263 | | } |
264 | | else if (asm1[n] == 2) |
265 | | { |
266 | | #if HAVE_NATIVE_mpn_addlsh1_n_ip1 |
267 | | cy = 2 * bsm1[n] + mpn_addlsh1_n_ip1 (vm1 + n, bsm1, n); |
268 | | #else |
269 | | cy = 2 * bsm1[n] + mpn_addmul_1 (vm1 + n, bsm1, n, CNST_LIMB(2)); |
270 | | #endif |
271 | | } |
272 | | else |
273 | | cy = 0; |
274 | | if (bsm1[n] != 0) |
275 | | cy += mpn_add_n (vm1 + n, vm1 + n, asm1, n); |
276 | | vm1[2 * n] = cy; |
277 | | #else /* SMALLER_RECURSION */ |
278 | 0 | vm1[2 * n] = 0; |
279 | 0 | mpn_mul_n (vm1, asm1, bsm1, n + ((asm1[n] | bsm1[n]) != 0)); |
280 | 0 | #endif /* SMALLER_RECURSION */ |
281 | | |
282 | | /* v1, 2n+1 limbs */ |
283 | | #ifdef SMALLER_RECURSION |
284 | | mpn_mul_n (v1, as1, bs1, n); |
285 | | if (as1[n] == 1) |
286 | | { |
287 | | cy = bs1[n] + mpn_add_n (v1 + n, v1 + n, bs1, n); |
288 | | } |
289 | | else if (as1[n] == 2) |
290 | | { |
291 | | #if HAVE_NATIVE_mpn_addlsh1_n_ip1 |
292 | | cy = 2 * bs1[n] + mpn_addlsh1_n_ip1 (v1 + n, bs1, n); |
293 | | #else |
294 | | cy = 2 * bs1[n] + mpn_addmul_1 (v1 + n, bs1, n, CNST_LIMB(2)); |
295 | | #endif |
296 | | } |
297 | | else if (as1[n] != 0) |
298 | | { |
299 | | cy = as1[n] * bs1[n] + mpn_addmul_1 (v1 + n, bs1, n, as1[n]); |
300 | | } |
301 | | else |
302 | | cy = 0; |
303 | | if (bs1[n] == 1) |
304 | | { |
305 | | cy += mpn_add_n (v1 + n, v1 + n, as1, n); |
306 | | } |
307 | | else if (bs1[n] == 2) |
308 | | { |
309 | | #if HAVE_NATIVE_mpn_addlsh1_n_ip1 |
310 | | cy += mpn_addlsh1_n_ip1 (v1 + n, as1, n); |
311 | | #else |
312 | | cy += mpn_addmul_1 (v1 + n, as1, n, CNST_LIMB(2)); |
313 | | #endif |
314 | | } |
315 | | v1[2 * n] = cy; |
316 | | #else /* SMALLER_RECURSION */ |
317 | 0 | v1[2 * n] = 0; |
318 | 0 | mpn_mul_n (v1, as1, bs1, n + ((as1[n] | bs1[n]) != 0)); |
319 | 0 | #endif /* SMALLER_RECURSION */ |
320 | |
|
321 | 0 | mpn_mul_n (v0, a0, b0, n); /* v0, 2n limbs */ |
322 | | |
323 | | /* vinf, s+t limbs */ |
324 | 0 | if (s > t) mpn_mul (vinf, a4, s, b2, t); |
325 | 0 | else mpn_mul (vinf, b2, t, a4, s); |
326 | |
|
327 | 0 | mpn_toom_interpolate_7pts (pp, n, flags, vm2, vm1, v2, vh, s + t, |
328 | 0 | scratch_out); |
329 | |
|
330 | 0 | TMP_FREE; |
331 | 0 | } |