X-Git-Url: https://git.distorted.org.uk/u/mdw/catacomb/blobdiff_plain/01f6ed1aafd567672bfb07d98c3bfa2ec19f8b3e..ab0350a65e5ac7f1f96e0e2df18c504dd8a441ce:/mpmont.c diff --git a/mpmont.c b/mpmont.c index cabb2cb..d051778 100644 --- a/mpmont.c +++ b/mpmont.c @@ -1,6 +1,6 @@ /* -*-c-*- * - * $Id: mpmont.c,v 1.8 1999/12/22 15:55:00 mdw Exp $ + * $Id: mpmont.c,v 1.14 2001/02/22 09:04:26 mdw Exp $ * * Montgomery reduction * @@ -30,6 +30,29 @@ /*----- Revision history --------------------------------------------------* * * $Log: mpmont.c,v $ + * Revision 1.14 2001/02/22 09:04:26 mdw + * Cosmetic fix. + * + * Revision 1.13 2001/02/03 12:00:29 mdw + * Now @mp_drop@ checks its argument is non-NULL before attempting to free + * it. Note that the macro version @MP_DROP@ doesn't do this. + * + * Revision 1.12 2000/10/08 15:48:35 mdw + * Rename Karatsuba constants now that we have @gfx_kmul@ too. + * + * Revision 1.11 2000/10/08 12:04:27 mdw + * (mpmont_reduce, mpmont_mul): Cope with negative numbers. + * + * Revision 1.10 2000/07/29 17:05:43 mdw + * (mpmont_expr): Use sliding window exponentiation, with a drop-through + * for small exponents to use a simple left-to-right bitwise routine. This + * can reduce modexp times by up to a quarter. + * + * Revision 1.9 2000/06/17 11:45:09 mdw + * Major memory management overhaul. Added arena support. Use the secure + * arena for secret integers. Replace and improve the MP management macros + * (e.g., replace MP_MODIFY by MP_DEST). + * * Revision 1.8 1999/12/22 15:55:00 mdw * Adjust Karatsuba parameters. * @@ -105,7 +128,7 @@ void mpmont_create(mpmont *mm, mp *m) void mpmont_create(mpmont *mm, mp *m) { size_t n = MP_LEN(m); - mp *r2 = mp_create(2 * n + 1); + mp *r2 = mp_new(2 * n + 1, 0); mp r; /* --- Validate the arguments --- */ @@ -185,7 +208,7 @@ mp *mpmont_reduce(mpmont *mm, mp *d, mp *a) /* --- Check for serious Karatsuba reduction --- */ - if (n > KARATSUBA_CUTOFF * 3) { + if (n > MPK_THRESH * 3) { mp al; mpw *vl; mp *u; @@ -213,13 +236,12 @@ mp *mpmont_reduce(mpmont *mm, mp *d, mp *a) /* --- Initial conditioning of the arguments --- */ - if (d == a) - MP_MODIFY(d, 2 * n + 1); - else { - MP_MODIFY(d, 2 * n + 1); - MPX_COPY(d->v, d->vl, a->v, a->vl); - } - + a = MP_COPY(a); + if (d) + MP_DROP(d); + d = a; + MP_DEST(d, 2 * n + 1, a->f); + dv = d->v; dvl = d->vl; mv = mm->m->v; mvl = mm->m->vl; @@ -235,11 +257,14 @@ mp *mpmont_reduce(mpmont *mm, mp *d, mp *a) /* --- Wrap everything up --- */ - d->f = a->f & MP_BURN; memmove(d->v, d->v + n, MPWS(MP_LEN(d) - n)); d->vl -= n; - if (MP_CMP(d, >=, mm->m)) - d = mp_sub(d, d, mm->m); + if (MPX_UCMP(d->v, d->vl, >=, mm->m->v, mm->m->vl)) + mpx_usub(d->v, d->vl, d->v, d->vl, mm->m->v, mm->m->vl); + if (d->f & MP_NEG) { + mpx_usub(d->v, d->vl, mm->m->v, mm->m->vl, d->v, d->vl); + d->f &= ~MP_NEG; + } MP_SHRINK(d); return (d); } @@ -268,7 +293,7 @@ mp *mpmont_mul(mpmont *mm, mp *d, mp *a, mp *b) mp *mpmont_mul(mpmont *mm, mp *d, mp *a, mp *b) { - if (mm->n > KARATSUBA_CUTOFF * 3) { + if (mm->n > MPK_THRESH * 3) { d = mp_mul(d, a, b); d = mpmont_reduce(mm, d, d); } else { @@ -289,7 +314,7 @@ mp *mpmont_mul(mpmont *mm, mp *d, mp *a, mp *b) a = MP_COPY(a); b = MP_COPY(b); - MP_MODIFY(d, 2 * n + 1); + MP_DEST(d, 2 * n + 1, a->f | b->f | MP_UNDEF); dv = d->v; dvl = d->vl; MPX_ZERO(dv, dvl); av = a->v; avl = a->vl; @@ -323,10 +348,12 @@ mp *mpmont_mul(mpmont *mm, mp *d, mp *a, mp *b) memmove(d->v, dv, MPWS(dvl - dv)); d->vl -= dv - d->v; + if (MPX_UCMP(d->v, d->vl, >=, mm->m->v, mm->m->vl)) + mpx_usub(d->v, d->vl, d->v, d->vl, mm->m->v, mm->m->vl); + if ((a->f ^ b->f) & MP_NEG) + mpx_usub(d->v, d->vl, mm->m->v, mm->m->vl, d->v, d->vl); MP_SHRINK(d); d->f = (a->f | b->f) & MP_BURN; - if (MP_CMP(d, >=, mm->m)) - d = mp_sub(d, d, mm->m); MP_DROP(a); MP_DROP(b); } @@ -346,35 +373,63 @@ mp *mpmont_mul(mpmont *mm, mp *d, mp *a, mp *b) * Returns: Result, %$a^e R \bmod m$%. */ -mp *mpmont_expr(mpmont *mm, mp *d, mp *a, mp *e) +#define WINSZ 5 +#define TABSZ (1 << (WINSZ - 1)) + +#define THRESH (((MPW_BITS / WINSZ) << 2) + 1) + +static mp *exp_simple(mpmont *mm, mp *d, mp *a, mp *e) { mpscan sc; - mp *ar = mpmont_mul(mm, MP_NEW, a, mm->r2); + mp *ar; mp *x = MP_COPY(mm->r); - mp *spare = MP_NEW; - - mp_scan(&sc, e); - - if (MP_STEP(&sc)) { - size_t sq = 0; + mp *spare = (e->f & MP_BURN) ? MP_NEWSEC : MP_NEW; + unsigned sq = 0; + + mp_rscan(&sc, e); + if (!MP_RSTEP(&sc)) + goto exit; + while (!MP_RBIT(&sc)) + MP_RSTEP(&sc); + + /* --- Do the main body of the work --- */ + + ar = mpmont_mul(mm, MP_NEW, a, mm->r2); + for (;;) { + sq++; + while (sq) { + mp *y; + y = mp_sqr(spare, x); + y = mpmont_reduce(mm, y, y); + spare = x; x = y; + sq--; + } + { mp *y = mpmont_mul(mm, spare, x, ar); spare = x; x = y; } + sq = 0; for (;;) { - mp *dd; - if (MP_BIT(&sc)) { - while (sq) { - dd = mp_sqr(spare, ar); - dd = mpmont_reduce(mm, dd, dd); - spare = ar; ar = dd; - sq--; - } - dd = mpmont_mul(mm, spare, x, ar); - spare = x; x = dd; - } - sq++; - if (!MP_STEP(&sc)) + if (!MP_RSTEP(&sc)) + goto done; + if (MP_RBIT(&sc)) break; + sq++; } } + + /* --- Do a final round of squaring --- */ + +done: + while (sq) { + mp *y; + y = mp_sqr(spare, x); + y = mpmont_reduce(mm, y, y); + spare = x; x = y; + sq--; + } + + /* --- Done --- */ + MP_DROP(ar); +exit: if (spare != MP_NEW) MP_DROP(spare); if (d != MP_NEW) @@ -382,6 +437,128 @@ mp *mpmont_expr(mpmont *mm, mp *d, mp *a, mp *e) return (x); } +mp *mpmont_expr(mpmont *mm, mp *d, mp *a, mp *e) +{ + mp **tab; + mp *ar, *a2; + mp *spare = (e->f & MP_BURN) ? MP_NEWSEC : MP_NEW; + mp *x = MP_COPY(mm->r); + unsigned i, sq = 0; + mpscan sc; + + /* --- Do we bother? --- */ + + MP_SHRINK(e); + if (MP_LEN(e) == 0) + goto exit; + if (MP_LEN(e) < THRESH) { + x->ref--; + return (exp_simple(mm, d, a, e)); + } + + /* --- Do the precomputation --- */ + + ar = mpmont_mul(mm, MP_NEW, a, mm->r2); + a2 = mp_sqr(MP_NEW, ar); + a2 = mpmont_reduce(mm, a2, a2); + tab = xmalloc(TABSZ * sizeof(mp *)); + tab[0] = ar; + for (i = 1; i < TABSZ; i++) + tab[i] = mpmont_mul(mm, MP_NEW, tab[i - 1], a2); + mp_drop(a2); + mp_rscan(&sc, e); + + /* --- Skip top-end zero bits --- * + * + * If the initial step worked, there must be a set bit somewhere, so keep + * stepping until I find it. + */ + + MP_RSTEP(&sc); + while (!MP_RBIT(&sc)) + MP_RSTEP(&sc); + + /* --- Now for the main work --- */ + + for (;;) { + unsigned l = 0; + unsigned z = 0; + + /* --- The next bit is set, so read a window index --- * + * + * Reset @i@ to zero and increment @sq@. Then, until either I read + * @WINSZ@ bits or I run out of bits, scan in a bit: if it's clear, bump + * the @z@ counter; if it's set, push a set bit into @i@, shift it over + * by @z@ bits, bump @sq@ by @z + 1@ and clear @z@. By the end of this + * palaver, @i@ is an index to the precomputed value in @tab@. + */ + + i = 0; + sq++; + for (;;) { + l++; + if (l >= WINSZ || !MP_RSTEP(&sc)) + break; + if (!MP_RBIT(&sc)) + z++; + else { + i = ((i << 1) | 1) << z; + sq += z + 1; + z = 0; + } + } + + /* --- Do the squaring --- * + * + * Remember that @sq@ carries over from the zero-skipping stuff below. + */ + + while (sq) { + mp *y; + y = mp_sqr(spare, x); + y = mpmont_reduce(mm, y, y); + spare = x; x = y; + sq--; + } + + /* --- Do the multiply --- */ + + { mp *y = mpmont_mul(mm, spare, x, tab[i]); spare = x; x = y; } + + /* --- Now grind along through the rest of the bits --- */ + + sq = z; + for (;;) { + if (!MP_RSTEP(&sc)) + goto done; + if (MP_RBIT(&sc)) + break; + sq++; + } + } + + /* --- Do a final round of squaring --- */ + +done: + while (sq) { + mp *y; + y = mp_sqr(spare, x); + y = mpmont_reduce(mm, y, y); + spare = x; x = y; + sq--; + } + + /* --- Done --- */ + + for (i = 0; i < TABSZ; i++) + mp_drop(tab[i]); + xfree(tab); +exit: + mp_drop(d); + mp_drop(spare); + return (x); +} + /* --- @mpmont_exp@ --- * * * Arguments: @mpmont *mm@ = pointer to Montgomery reduction context @@ -423,7 +600,7 @@ static int tcreate(dstr *v) ok = 0; } - if (MP_CMP(mm.r, !=, r)) { + if (!MP_EQ(mm.r, r)) { fputs("\n*** bad r", stderr); fputs("\nm = ", stderr); mp_writefile(m, stderr, 10); fputs("\nexpected ", stderr); mp_writefile(r, stderr, 10); @@ -432,7 +609,7 @@ static int tcreate(dstr *v) ok = 0; } - if (MP_CMP(mm.r2, !=, r2)) { + if (!MP_EQ(mm.r2, r2)) { fputs("\n*** bad r2", stderr); fputs("\nm = ", stderr); mp_writefile(m, stderr, 10); fputs("\nexpected ", stderr); mp_writefile(r2, stderr, 10); @@ -465,7 +642,7 @@ static int tmul(dstr *v) mp *qr = mp_mul(MP_NEW, a, b); mp_div(0, &qr, qr, m); - if (MP_CMP(qr, !=, r)) { + if (!MP_EQ(qr, r)) { fputs("\n*** classical modmul failed", stderr); fputs("\n m = ", stderr); mp_writefile(m, stderr, 10); fputs("\n a = ", stderr); mp_writefile(a, stderr, 10); @@ -484,7 +661,7 @@ static int tmul(dstr *v) mp *br = mpmont_mul(&mm, MP_NEW, b, mm.r2); mp *mr = mpmont_mul(&mm, MP_NEW, ar, br); mr = mpmont_reduce(&mm, mr, mr); - if (MP_CMP(mr, !=, r)) { + if (!MP_EQ(mr, r)) { fputs("\n*** montgomery modmul failed", stderr); fputs("\n m = ", stderr); mp_writefile(m, stderr, 10); fputs("\n a = ", stderr); mp_writefile(a, stderr, 10); @@ -522,7 +699,7 @@ static int texp(dstr *v) mr = mpmont_exp(&mm, MP_NEW, a, b); - if (MP_CMP(mr, !=, r)) { + if (!MP_EQ(mr, r)) { fputs("\n*** montgomery modexp failed", stderr); fputs("\n m = ", stderr); mp_writefile(m, stderr, 10); fputs("\n a = ", stderr); mp_writefile(a, stderr, 10);