/* -*-c-*-
*
- * $Id: mpmont.c,v 1.10 2000/07/29 17:05:43 mdw Exp $
+ * $Id: mpmont.c,v 1.16 2002/01/13 13:40:31 mdw Exp $
*
* Montgomery reduction
*
/*----- Revision history --------------------------------------------------*
*
* $Log: mpmont.c,v $
+ * Revision 1.16 2002/01/13 13:40:31 mdw
+ * Avoid trashing arguments before we've used them.
+ *
+ * Revision 1.15 2001/06/16 13:00:20 mdw
+ * Use the generic exponentiation functions.
+ *
+ * Revision 1.14 2001/02/22 09:04:26 mdw
+ * Cosmetic fix.
+ *
+ * Revision 1.13 2001/02/03 12:00:29 mdw
+ * Now @mp_drop@ checks its argument is non-NULL before attempting to free
+ * it. Note that the macro version @MP_DROP@ doesn't do this.
+ *
+ * Revision 1.12 2000/10/08 15:48:35 mdw
+ * Rename Karatsuba constants now that we have @gfx_kmul@ too.
+ *
+ * Revision 1.11 2000/10/08 12:04:27 mdw
+ * (mpmont_reduce, mpmont_mul): Cope with negative numbers.
+ *
* Revision 1.10 2000/07/29 17:05:43 mdw
* (mpmont_expr): Use sliding window exponentiation, with a drop-through
* for small exponents to use a simple left-to-right bitwise routine. This
#include "mp.h"
#include "mpmont.h"
+#include "mpmont-exp.h"
/*----- Tweakables --------------------------------------------------------*/
/* #define MPMONT_DISABLE */
-/*----- Main code ---------------------------------------------------------*/
+/*----- Reduction and multiplication --------------------------------------*/
/* --- @mpmont_create@ --- *
*
/* --- Check for serious Karatsuba reduction --- */
- if (n > KARATSUBA_CUTOFF * 3) {
+ if (n > MPK_THRESH * 3) {
mp al;
mpw *vl;
mp *u;
memmove(d->v, d->v + n, MPWS(MP_LEN(d) - n));
d->vl -= n;
- if (MP_CMP(d, >=, mm->m))
- d = mp_sub(d, d, mm->m);
+ if (MPX_UCMP(d->v, d->vl, >=, mm->m->v, mm->m->vl))
+ mpx_usub(d->v, d->vl, d->v, d->vl, mm->m->v, mm->m->vl);
+ if (d->f & MP_NEG) {
+ mpx_usub(d->v, d->vl, mm->m->v, mm->m->vl, d->v, d->vl);
+ d->f &= ~MP_NEG;
+ }
MP_SHRINK(d);
return (d);
}
mp *mpmont_mul(mpmont *mm, mp *d, mp *a, mp *b)
{
- if (mm->n > KARATSUBA_CUTOFF * 3) {
+ if (mm->n > MPK_THRESH * 3) {
d = mp_mul(d, a, b);
d = mpmont_reduce(mm, d, d);
} else {
memmove(d->v, dv, MPWS(dvl - dv));
d->vl -= dv - d->v;
+ if (MPX_UCMP(d->v, d->vl, >=, mm->m->v, mm->m->vl))
+ mpx_usub(d->v, d->vl, d->v, d->vl, mm->m->v, mm->m->vl);
+ if ((a->f ^ b->f) & MP_NEG)
+ mpx_usub(d->v, d->vl, mm->m->v, mm->m->vl, d->v, d->vl);
MP_SHRINK(d);
d->f = (a->f | b->f) & MP_BURN;
- if (MP_CMP(d, >=, mm->m))
- d = mp_sub(d, d, mm->m);
MP_DROP(a);
MP_DROP(b);
}
#endif
+/*----- Exponentiation ----------------------------------------------------*/
+
/* --- @mpmont_expr@ --- *
*
* Arguments: @mpmont *mm@ = pointer to Montgomery reduction context
* @mp *a@ = base
* @mp *e@ = exponent
*
- * Returns: Result, %$a^e R \bmod m$%.
+ * Returns: Result, %$(a R^{-1})^e R \bmod m$%.
*/
-#define WINSZ 5
-#define TABSZ (1 << (WINSZ - 1))
-
-#define THRESH (((MPW_BITS / WINSZ) << 2) + 1)
-
-static mp *exp_simple(mpmont *mm, mp *d, mp *a, mp *e)
-{
- mpscan sc;
- mp *ar;
- mp *x = MP_COPY(mm->r);
- mp *spare = (e->f & MP_BURN) ? MP_NEWSEC : MP_NEW;
- unsigned sq = 0;
-
- mp_rscan(&sc, e);
- if (!MP_RSTEP(&sc))
- goto exit;
- while (!MP_RBIT(&sc))
- MP_RSTEP(&sc);
-
- /* --- Do the main body of the work --- */
-
- ar = mpmont_mul(mm, MP_NEW, a, mm->r2);
- for (;;) {
- sq++;
- while (sq) {
- mp *y;
- y = mp_sqr(spare, x);
- y = mpmont_reduce(mm, y, y);
- spare = x; x = y;
- sq--;
- }
- { mp *y = mpmont_mul(mm, spare, x, ar); spare = x; x = y; }
- sq = 0;
- for (;;) {
- if (!MP_RSTEP(&sc))
- goto done;
- if (MP_RBIT(&sc))
- break;
- sq++;
- }
- }
-
- /* --- Do a final round of squaring --- */
-
-done:
- while (sq) {
- mp *y;
- y = mp_sqr(spare, x);
- y = mpmont_reduce(mm, y, y);
- spare = x; x = y;
- sq--;
- }
-
- /* --- Done --- */
-
- MP_DROP(ar);
-exit:
- if (spare != MP_NEW)
- MP_DROP(spare);
- if (d != MP_NEW)
- MP_DROP(d);
- return (x);
-}
-
mp *mpmont_expr(mpmont *mm, mp *d, mp *a, mp *e)
{
- mp **tab;
- mp *ar, *a2;
- mp *spare = (e->f & MP_BURN) ? MP_NEWSEC : MP_NEW;
mp *x = MP_COPY(mm->r);
- unsigned i, sq = 0;
- mpscan sc;
-
- /* --- Do we bother? --- */
+ mp *spare = (e->f & MP_BURN) ? MP_NEWSEC : MP_NEW;
MP_SHRINK(e);
if (MP_LEN(e) == 0)
- goto exit;
- if (MP_LEN(e) < THRESH) {
- x->ref--;
- return (exp_simple(mm, d, a, e));
- }
-
- /* --- Do the precomputation --- */
-
- ar = mpmont_mul(mm, MP_NEW, a, mm->r2);
- a2 = mp_sqr(MP_NEW, ar);
- a2 = mpmont_reduce(mm, a2, a2);
- tab = xmalloc(TABSZ * sizeof(mp *));
- tab[0] = ar;
- for (i = 1; i < TABSZ; i++)
- tab[i] = mpmont_mul(mm, MP_NEW, tab[i - 1], a2);
- mp_drop(a2);
- mp_rscan(&sc, e);
-
- /* --- Skip top-end zero bits --- *
- *
- * If the initial step worked, there must be a set bit somewhere, so keep
- * stepping until I find it.
- */
-
- MP_RSTEP(&sc);
- while (!MP_RBIT(&sc)) {
- MP_RSTEP(&sc);
- }
-
- /* --- Now for the main work --- */
-
- for (;;) {
- unsigned l = 0;
- unsigned z = 0;
-
- /* --- The next bit is set, so read a window index --- *
- *
- * Reset @i@ to zero and increment @sq@. Then, until either I read
- * @WINSZ@ bits or I run out of bits, scan in a bit: if it's clear, bump
- * the @z@ counter; if it's set, push a set bit into @i@, shift it over
- * by @z@ bits, bump @sq@ by @z + 1@ and clear @z@. By the end of this
- * palaver, @i@ is an index to the precomputed value in @tab@.
- */
-
- i = 0;
- sq++;
- for (;;) {
- l++;
- if (l >= WINSZ || !MP_RSTEP(&sc))
- break;
- if (!MP_RBIT(&sc))
- z++;
- else {
- i = ((i << 1) | 1) << z;
- sq += z + 1;
- z = 0;
- }
- }
-
- /* --- Do the squaring --- *
- *
- * Remember that @sq@ carries over from the zero-skipping stuff below.
- */
-
- while (sq) {
- mp *y;
- y = mp_sqr(spare, x);
- y = mpmont_reduce(mm, y, y);
- spare = x; x = y;
- sq--;
- }
-
- /* --- Do the multiply --- */
-
- { mp *y = mpmont_mul(mm, spare, x, tab[i]); spare = x; x = y; }
-
- /* --- Now grind along through the rest of the bits --- */
-
- sq = z;
- for (;;) {
- if (!MP_RSTEP(&sc))
- goto done;
- if (MP_RBIT(&sc))
- break;
- sq++;
- }
- }
-
- /* --- Do a final round of squaring --- */
-
-done:
- while (sq) {
- mp *y;
- y = mp_sqr(spare, x);
- y = mpmont_reduce(mm, y, y);
- spare = x; x = y;
- sq--;
- }
-
- /* --- Done --- */
-
- for (i = 0; i < TABSZ; i++)
- mp_drop(tab[i]);
- xfree(tab);
-exit:
- if (d != MP_NEW)
- mp_drop(d);
- if (spare)
- mp_drop(spare);
+ ;
+ else if (MP_LEN(e) < EXP_THRESH)
+ EXP_SIMPLE(x, a, e);
+ else
+ EXP_WINDOW(x, a, e);
+ mp_drop(d);
+ mp_drop(spare);
return (x);
}
mp *mpmont_exp(mpmont *mm, mp *d, mp *a, mp *e)
{
- d = mpmont_expr(mm, d, a, e);
+ e = MP_COPY(e);
+ d = mpmont_mul(mm, d, a, mm->r2);
+ d = mpmont_expr(mm, d, d, e);
d = mpmont_reduce(mm, d, d);
+ MP_DROP(e);
return (d);
}
ok = 0;
}
- if (MP_CMP(mm.r, !=, r)) {
+ if (!MP_EQ(mm.r, r)) {
fputs("\n*** bad r", stderr);
fputs("\nm = ", stderr); mp_writefile(m, stderr, 10);
fputs("\nexpected ", stderr); mp_writefile(r, stderr, 10);
ok = 0;
}
- if (MP_CMP(mm.r2, !=, r2)) {
+ if (!MP_EQ(mm.r2, r2)) {
fputs("\n*** bad r2", stderr);
fputs("\nm = ", stderr); mp_writefile(m, stderr, 10);
fputs("\nexpected ", stderr); mp_writefile(r2, stderr, 10);
mp *qr = mp_mul(MP_NEW, a, b);
mp_div(0, &qr, qr, m);
- if (MP_CMP(qr, !=, r)) {
+ if (!MP_EQ(qr, r)) {
fputs("\n*** classical modmul failed", stderr);
fputs("\n m = ", stderr); mp_writefile(m, stderr, 10);
fputs("\n a = ", stderr); mp_writefile(a, stderr, 10);
mp *br = mpmont_mul(&mm, MP_NEW, b, mm.r2);
mp *mr = mpmont_mul(&mm, MP_NEW, ar, br);
mr = mpmont_reduce(&mm, mr, mr);
- if (MP_CMP(mr, !=, r)) {
+ if (!MP_EQ(mr, r)) {
fputs("\n*** montgomery modmul failed", stderr);
fputs("\n m = ", stderr); mp_writefile(m, stderr, 10);
fputs("\n a = ", stderr); mp_writefile(a, stderr, 10);
mr = mpmont_exp(&mm, MP_NEW, a, b);
- if (MP_CMP(mr, !=, r)) {
+ if (!MP_EQ(mr, r)) {
fputs("\n*** montgomery modexp failed", stderr);
fputs("\n m = ", stderr); mp_writefile(m, stderr, 10);
fputs("\n a = ", stderr); mp_writefile(a, stderr, 10);