7#ifndef SECP256K1_FIELD_REPR_IMPL_H
8#define SECP256K1_FIELD_REPR_IMPL_H
14#if defined(USE_ASM_X86_64)
37 const uint64_t *d = a->
n;
38 int m = a->normalized ? 1 : 2 * a->magnitude, r = 1;
40 r &= (d[0] <= 0xFFFFFFFFFFFFFULL * m);
41 r &= (d[1] <= 0xFFFFFFFFFFFFFULL * m);
42 r &= (d[2] <= 0xFFFFFFFFFFFFFULL * m);
43 r &= (d[3] <= 0xFFFFFFFFFFFFFULL * m);
44 r &= (d[4] <= 0x0FFFFFFFFFFFFULL * m);
45 r &= (a->magnitude >= 0);
46 r &= (a->magnitude <= 2048);
48 r &= (a->magnitude <= 1);
49 if (r && (d[4] == 0x0FFFFFFFFFFFFULL) && ((d[3] & d[2] & d[1]) == 0xFFFFFFFFFFFFFULL)) {
50 r &= (d[0] < 0xFFFFEFFFFFC2FULL);
60 r->
n[0] = 0xFFFFFFFFFFFFFULL * 2 * m;
61 r->
n[1] = 0xFFFFFFFFFFFFFULL * 2 * m;
62 r->
n[2] = 0xFFFFFFFFFFFFFULL * 2 * m;
63 r->
n[3] = 0xFFFFFFFFFFFFFULL * 2 * m;
64 r->
n[4] = 0x0FFFFFFFFFFFFULL * 2 * m;
67 r->normalized = (m == 0);
68 secp256k1_fe_verify(r);
73 uint64_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4];
77 uint64_t x = t4 >> 48; t4 &= 0x0FFFFFFFFFFFFULL;
80 t0 += x * 0x1000003D1ULL;
81 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
82 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL; m = t1;
83 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL; m &= t2;
84 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL; m &= t3;
90 x = (t4 >> 48) | ((t4 == 0x0FFFFFFFFFFFFULL) & (m == 0xFFFFFFFFFFFFFULL)
91 & (t0 >= 0xFFFFEFFFFFC2FULL));
94 t0 += x * 0x1000003D1ULL;
95 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
96 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL;
97 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL;
98 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL;
104 t4 &= 0x0FFFFFFFFFFFFULL;
106 r->
n[0] = t0; r->
n[1] = t1; r->
n[2] = t2; r->
n[3] = t3; r->
n[4] = t4;
111 secp256k1_fe_verify(r);
116 uint64_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4];
119 uint64_t x = t4 >> 48; t4 &= 0x0FFFFFFFFFFFFULL;
122 t0 += x * 0x1000003D1ULL;
123 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
124 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL;
125 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL;
126 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL;
131 r->
n[0] = t0; r->
n[1] = t1; r->
n[2] = t2; r->
n[3] = t3; r->
n[4] = t4;
135 secp256k1_fe_verify(r);
140 uint64_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4];
144 uint64_t x = t4 >> 48; t4 &= 0x0FFFFFFFFFFFFULL;
147 t0 += x * 0x1000003D1ULL;
148 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
149 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL; m = t1;
150 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL; m &= t2;
151 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL; m &= t3;
157 x = (t4 >> 48) | ((t4 == 0x0FFFFFFFFFFFFULL) & (m == 0xFFFFFFFFFFFFFULL)
158 & (t0 >= 0xFFFFEFFFFFC2FULL));
161 t0 += 0x1000003D1ULL;
162 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
163 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL;
164 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL;
165 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL;
171 t4 &= 0x0FFFFFFFFFFFFULL;
174 r->
n[0] = t0; r->
n[1] = t1; r->
n[2] = t2; r->
n[3] = t3; r->
n[4] = t4;
179 secp256k1_fe_verify(r);
184 uint64_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4];
190 uint64_t x = t4 >> 48; t4 &= 0x0FFFFFFFFFFFFULL;
193 t0 += x * 0x1000003D1ULL;
194 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL; z0 = t0; z1 = t0 ^ 0x1000003D0ULL;
195 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL; z0 |= t1; z1 &= t1;
196 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL; z0 |= t2; z1 &= t2;
197 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL; z0 |= t3; z1 &= t3;
198 z0 |= t4; z1 &= t4 ^ 0xF000000000000ULL;
203 return (z0 == 0) | (z1 == 0xFFFFFFFFFFFFFULL);
207 uint64_t t0, t1, t2, t3, t4;
218 t0 += x * 0x1000003D1ULL;
221 z0 = t0 & 0xFFFFFFFFFFFFFULL;
222 z1 = z0 ^ 0x1000003D0ULL;
225 if ((z0 != 0ULL) & (z1 != 0xFFFFFFFFFFFFFULL)) {
233 t4 &= 0x0FFFFFFFFFFFFULL;
236 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL; z0 |= t1; z1 &= t1;
237 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL; z0 |= t2; z1 &= t2;
238 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL; z0 |= t3; z1 &= t3;
239 z0 |= t4; z1 &= t4 ^ 0xF000000000000ULL;
244 return (z0 == 0) | (z1 == 0xFFFFFFFFFFFFFULL);
250 r->
n[1] = r->
n[2] = r->
n[3] = r->
n[4] = 0;
252 r->magnitude = (a != 0);
254 secp256k1_fe_verify(r);
259 const uint64_t *t = a->
n;
262 secp256k1_fe_verify(a);
264 return (t[0] | t[1] | t[2] | t[3] | t[4]) == 0;
270 secp256k1_fe_verify(a);
281 for (i=0; i<5; i++) {
291 secp256k1_fe_verify(a);
292 secp256k1_fe_verify(b);
294 for (i = 4; i >= 0; i--) {
295 if (a->
n[i] > b->
n[i]) {
298 if (a->
n[i] < b->
n[i]) {
307 r->
n[0] = (uint64_t)a[31]
308 | ((uint64_t)a[30] << 8)
309 | ((uint64_t)a[29] << 16)
310 | ((uint64_t)a[28] << 24)
311 | ((uint64_t)a[27] << 32)
312 | ((uint64_t)a[26] << 40)
313 | ((uint64_t)(a[25] & 0xF) << 48);
314 r->
n[1] = (uint64_t)((a[25] >> 4) & 0xF)
315 | ((uint64_t)a[24] << 4)
316 | ((uint64_t)a[23] << 12)
317 | ((uint64_t)a[22] << 20)
318 | ((uint64_t)a[21] << 28)
319 | ((uint64_t)a[20] << 36)
320 | ((uint64_t)a[19] << 44);
321 r->
n[2] = (uint64_t)a[18]
322 | ((uint64_t)a[17] << 8)
323 | ((uint64_t)a[16] << 16)
324 | ((uint64_t)a[15] << 24)
325 | ((uint64_t)a[14] << 32)
326 | ((uint64_t)a[13] << 40)
327 | ((uint64_t)(a[12] & 0xF) << 48);
328 r->
n[3] = (uint64_t)((a[12] >> 4) & 0xF)
329 | ((uint64_t)a[11] << 4)
330 | ((uint64_t)a[10] << 12)
331 | ((uint64_t)a[9] << 20)
332 | ((uint64_t)a[8] << 28)
333 | ((uint64_t)a[7] << 36)
334 | ((uint64_t)a[6] << 44);
335 r->
n[4] = (uint64_t)a[5]
336 | ((uint64_t)a[4] << 8)
337 | ((uint64_t)a[3] << 16)
338 | ((uint64_t)a[2] << 24)
339 | ((uint64_t)a[1] << 32)
340 | ((uint64_t)a[0] << 40);
341 ret = !((r->
n[4] == 0x0FFFFFFFFFFFFULL) & ((r->
n[3] & r->
n[2] & r->
n[1]) == 0xFFFFFFFFFFFFFULL) & (r->
n[0] >= 0xFFFFEFFFFFC2FULL));
346 secp256k1_fe_verify(r);
358 secp256k1_fe_verify(a);
360 r[0] = (a->
n[4] >> 40) & 0xFF;
361 r[1] = (a->
n[4] >> 32) & 0xFF;
362 r[2] = (a->
n[4] >> 24) & 0xFF;
363 r[3] = (a->
n[4] >> 16) & 0xFF;
364 r[4] = (a->
n[4] >> 8) & 0xFF;
365 r[5] = a->
n[4] & 0xFF;
366 r[6] = (a->
n[3] >> 44) & 0xFF;
367 r[7] = (a->
n[3] >> 36) & 0xFF;
368 r[8] = (a->
n[3] >> 28) & 0xFF;
369 r[9] = (a->
n[3] >> 20) & 0xFF;
370 r[10] = (a->
n[3] >> 12) & 0xFF;
371 r[11] = (a->
n[3] >> 4) & 0xFF;
372 r[12] = ((a->
n[2] >> 48) & 0xF) | ((a->
n[3] & 0xF) << 4);
373 r[13] = (a->
n[2] >> 40) & 0xFF;
374 r[14] = (a->
n[2] >> 32) & 0xFF;
375 r[15] = (a->
n[2] >> 24) & 0xFF;
376 r[16] = (a->
n[2] >> 16) & 0xFF;
377 r[17] = (a->
n[2] >> 8) & 0xFF;
378 r[18] = a->
n[2] & 0xFF;
379 r[19] = (a->
n[1] >> 44) & 0xFF;
380 r[20] = (a->
n[1] >> 36) & 0xFF;
381 r[21] = (a->
n[1] >> 28) & 0xFF;
382 r[22] = (a->
n[1] >> 20) & 0xFF;
383 r[23] = (a->
n[1] >> 12) & 0xFF;
384 r[24] = (a->
n[1] >> 4) & 0xFF;
385 r[25] = ((a->
n[0] >> 48) & 0xF) | ((a->
n[1] & 0xF) << 4);
386 r[26] = (a->
n[0] >> 40) & 0xFF;
387 r[27] = (a->
n[0] >> 32) & 0xFF;
388 r[28] = (a->
n[0] >> 24) & 0xFF;
389 r[29] = (a->
n[0] >> 16) & 0xFF;
390 r[30] = (a->
n[0] >> 8) & 0xFF;
391 r[31] = a->
n[0] & 0xFF;
397 secp256k1_fe_verify(a);
398 VERIFY_CHECK(0xFFFFEFFFFFC2FULL * 2 * (m + 1) >= 0xFFFFFFFFFFFFFULL * 2 * m);
399 VERIFY_CHECK(0xFFFFFFFFFFFFFULL * 2 * (m + 1) >= 0xFFFFFFFFFFFFFULL * 2 * m);
400 VERIFY_CHECK(0x0FFFFFFFFFFFFULL * 2 * (m + 1) >= 0x0FFFFFFFFFFFFULL * 2 * m);
402 r->
n[0] = 0xFFFFEFFFFFC2FULL * 2 * (m + 1) - a->
n[0];
403 r->
n[1] = 0xFFFFFFFFFFFFFULL * 2 * (m + 1) - a->
n[1];
404 r->
n[2] = 0xFFFFFFFFFFFFFULL * 2 * (m + 1) - a->
n[2];
405 r->
n[3] = 0xFFFFFFFFFFFFFULL * 2 * (m + 1) - a->
n[3];
406 r->
n[4] = 0x0FFFFFFFFFFFFULL * 2 * (m + 1) - a->
n[4];
408 r->magnitude = m + 1;
410 secp256k1_fe_verify(r);
423 secp256k1_fe_verify(r);
429 secp256k1_fe_verify(a);
437 r->magnitude += a->magnitude;
439 secp256k1_fe_verify(r);
447 secp256k1_fe_verify(a);
448 secp256k1_fe_verify(b);
456 secp256k1_fe_verify(r);
463 secp256k1_fe_verify(a);
469 secp256k1_fe_verify(r);
474 uint64_t mask0, mask1;
475 volatile int vflag = flag;
477 mask0 = vflag + ~((uint64_t)0);
479 r->
n[0] = (r->
n[0] & mask0) | (a->
n[0] & mask1);
480 r->
n[1] = (r->
n[1] & mask0) | (a->
n[1] & mask1);
481 r->
n[2] = (r->
n[2] & mask0) | (a->
n[2] & mask1);
482 r->
n[3] = (r->
n[3] & mask0) | (a->
n[3] & mask1);
483 r->
n[4] = (r->
n[4] & mask0) | (a->
n[4] & mask1);
486 r->magnitude = a->magnitude;
487 r->normalized = a->normalized;
493 uint64_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4];
494 uint64_t one = (uint64_t)1;
495 uint64_t mask = -(t0 & one) >> 12;
498 secp256k1_fe_verify(r);
512 t0 += 0xFFFFEFFFFFC2FULL & mask;
527 r->
n[0] = (t0 >> 1) + ((t1 & one) << 51);
528 r->
n[1] = (t1 >> 1) + ((t2 & one) << 51);
529 r->
n[2] = (t2 >> 1) + ((t3 & one) << 51);
530 r->
n[3] = (t3 >> 1) + ((t4 & one) << 51);
551 r->magnitude = (r->magnitude >> 1) + 1;
553 secp256k1_fe_verify(r);
558 uint64_t mask0, mask1;
559 volatile int vflag = flag;
561 mask0 = vflag + ~((uint64_t)0);
563 r->
n[0] = (r->
n[0] & mask0) | (a->
n[0] & mask1);
564 r->
n[1] = (r->
n[1] & mask0) | (a->
n[1] & mask1);
565 r->
n[2] = (r->
n[2] & mask0) | (a->
n[2] & mask1);
566 r->
n[3] = (r->
n[3] & mask0) | (a->
n[3] & mask1);
573 r->
n[0] = a->
n[0] | a->
n[1] << 52;
574 r->
n[1] = a->
n[1] >> 12 | a->
n[2] << 40;
575 r->
n[2] = a->
n[2] >> 24 | a->
n[3] << 28;
576 r->
n[3] = a->
n[3] >> 36 | a->
n[4] << 16;
580 r->
n[0] = a->
n[0] & 0xFFFFFFFFFFFFFULL;
581 r->
n[1] = a->
n[0] >> 52 | ((a->
n[1] << 12) & 0xFFFFFFFFFFFFFULL);
582 r->
n[2] = a->
n[1] >> 40 | ((a->
n[2] << 24) & 0xFFFFFFFFFFFFFULL);
583 r->
n[3] = a->
n[2] >> 28 | ((a->
n[3] << 36) & 0xFFFFFFFFFFFFFULL);
584 r->
n[4] = a->
n[3] >> 16;
588 secp256k1_fe_verify(r);
593 const uint64_t M52 = UINT64_MAX >> 12;
594 const uint64_t a0 = a->
v[0], a1 = a->
v[1], a2 = a->
v[2], a3 = a->
v[3], a4 = a->
v[4];
606 r->
n[1] = (a0 >> 52 | a1 << 10) & M52;
607 r->
n[2] = (a1 >> 42 | a2 << 20) & M52;
608 r->
n[3] = (a2 >> 32 | a3 << 30) & M52;
609 r->
n[4] = (a3 >> 22 | a4 << 40);
614 secp256k1_fe_verify(r);
619 const uint64_t M62 = UINT64_MAX >> 2;
620 const uint64_t a0 = a->
n[0], a1 = a->
n[1], a2 = a->
n[2], a3 = a->
n[3], a4 = a->
n[4];
626 r->
v[0] = (a0 | a1 << 52) & M62;
627 r->
v[1] = (a1 >> 10 | a2 << 42) & M62;
628 r->
v[2] = (a2 >> 20 | a3 << 32) & M62;
629 r->
v[3] = (a3 >> 30 | a4 << 22) & M62;
634 {{-0x1000003D1LL, 0, 0, 0, 256}},
static SECP256K1_INLINE void secp256k1_fe_sqr_inner(uint32_t *r, const uint32_t *a)
static SECP256K1_INLINE void secp256k1_fe_mul_inner(uint32_t *r, const uint32_t *a, const uint32_t *SECP256K1_RESTRICT b)
static int secp256k1_fe_normalizes_to_zero_var(const secp256k1_fe *r)
static SECP256K1_INLINE void secp256k1_fe_set_int(secp256k1_fe *r, int a)
static void secp256k1_fe_normalize_weak(secp256k1_fe *r)
static SECP256K1_INLINE int secp256k1_fe_is_zero(const secp256k1_fe *a)
static void secp256k1_fe_normalize_var(secp256k1_fe *r)
static SECP256K1_INLINE void secp256k1_fe_mul_int(secp256k1_fe *r, int a)
static void secp256k1_fe_get_bounds(secp256k1_fe *r, int m)
Implements arithmetic modulo FFFFFFFF FFFFFFFF FFFFFFFF FFFFFFFF FFFFFFFF FFFFFFFF FFFFFFFE FFFFFC2F,...
static void secp256k1_fe_mul(secp256k1_fe *r, const secp256k1_fe *a, const secp256k1_fe *SECP256K1_RESTRICT b)
static int secp256k1_fe_set_b32(secp256k1_fe *r, const unsigned char *a)
static SECP256K1_INLINE void secp256k1_fe_storage_cmov(secp256k1_fe_storage *r, const secp256k1_fe_storage *a, int flag)
static void secp256k1_fe_sqr(secp256k1_fe *r, const secp256k1_fe *a)
static SECP256K1_INLINE void secp256k1_fe_cmov(secp256k1_fe *r, const secp256k1_fe *a, int flag)
static int secp256k1_fe_normalizes_to_zero(const secp256k1_fe *r)
static SECP256K1_INLINE void secp256k1_fe_negate(secp256k1_fe *r, const secp256k1_fe *a, int m)
static void secp256k1_fe_to_signed62(secp256k1_modinv64_signed62 *r, const secp256k1_fe *a)
static void secp256k1_fe_normalize(secp256k1_fe *r)
static SECP256K1_INLINE int secp256k1_fe_is_odd(const secp256k1_fe *a)
static SECP256K1_INLINE void secp256k1_fe_clear(secp256k1_fe *a)
static void secp256k1_fe_to_storage(secp256k1_fe_storage *r, const secp256k1_fe *a)
static SECP256K1_INLINE void secp256k1_fe_from_storage(secp256k1_fe *r, const secp256k1_fe_storage *a)
static void secp256k1_fe_get_b32(unsigned char *r, const secp256k1_fe *a)
Convert a field element to a 32-byte big endian value.
static SECP256K1_INLINE void secp256k1_fe_half(secp256k1_fe *r)
static SECP256K1_INLINE void secp256k1_fe_add(secp256k1_fe *r, const secp256k1_fe *a)
static void secp256k1_fe_from_signed62(secp256k1_fe *r, const secp256k1_modinv64_signed62 *a)
static const secp256k1_modinv64_modinfo secp256k1_const_modinfo_fe
static int secp256k1_fe_cmp_var(const secp256k1_fe *a, const secp256k1_fe *b)
static void secp256k1_fe_inv(secp256k1_fe *r, const secp256k1_fe *x)
static void secp256k1_fe_inv_var(secp256k1_fe *r, const secp256k1_fe *x)
static void secp256k1_modinv64(secp256k1_modinv64_signed62 *x, const secp256k1_modinv64_modinfo *modinfo)
static void secp256k1_modinv64_var(secp256k1_modinv64_signed62 *x, const secp256k1_modinv64_modinfo *modinfo)
#define VG_CHECK_VERIFY(x, y)
#define VERIFY_CHECK(cond)
#define SECP256K1_RESTRICT