Lines Matching refs:u64

39                 u64 T = U64(0xe100000000000000) & (0-(V.lo&1)); \
46 V.hi = (V.hi>>1 )^((u64)T<<32); \
88 static void gcm_init_4bit(u128 Htable[16], const u64 H[2]) in gcm_init_4bit()
167 static void gcm_gmult_4bit(u64 Xi[2], const u128 Htable[16]) in gcm_gmult_4bit()
188 Z.hi ^= (u64)rem_4bit[rem] << 32; in gcm_gmult_4bit()
206 Z.hi ^= (u64)rem_4bit[rem] << 32; in gcm_gmult_4bit()
245 static void gcm_ghash_4bit(u64 Xi[2], const u128 Htable[16], in gcm_ghash_4bit()
270 Z.hi ^= (u64)rem_4bit[rem] << 32; in gcm_ghash_4bit()
289 Z.hi ^= (u64)rem_4bit[rem] << 32; in gcm_ghash_4bit()
323 void gcm_gmult_4bit(u64 Xi[2], const u128 Htable[16]);
324 void gcm_ghash_4bit(u64 Xi[2], const u128 Htable[16], const u8 *inp,
346 void gcm_init_clmul(u128 Htable[16], const u64 Xi[2]);
347 void gcm_gmult_clmul(u64 Xi[2], const u128 Htable[16]);
348 void gcm_ghash_clmul(u64 Xi[2], const u128 Htable[16], const u8 *inp,
356 void gcm_init_avx(u128 Htable[16], const u64 Xi[2]);
357 void gcm_gmult_avx(u64 Xi[2], const u128 Htable[16]);
358 void gcm_ghash_avx(u64 Xi[2], const u128 Htable[16], const u8 *inp,
364 void gcm_gmult_4bit_mmx(u64 Xi[2], const u128 Htable[16]);
365 void gcm_ghash_4bit_mmx(u64 Xi[2], const u128 Htable[16], const u8 *inp,
368 void gcm_gmult_4bit_x86(u64 Xi[2], const u128 Htable[16]);
369 void gcm_ghash_4bit_x86(u64 Xi[2], const u128 Htable[16], const u8 *inp,
380 void gcm_init_neon(u128 Htable[16], const u64 Xi[2]);
381 void gcm_gmult_neon(u64 Xi[2], const u128 Htable[16]);
382 void gcm_ghash_neon(u64 Xi[2], const u128 Htable[16], const u8 *inp,
384 void gcm_init_v8(u128 Htable[16], const u64 Xi[2]);
385 void gcm_gmult_v8(u64 Xi[2], const u128 Htable[16]);
386 void gcm_ghash_v8(u64 Xi[2], const u128 Htable[16], const u8 *inp,
392 void gcm_init_vis3(u128 Htable[16], const u64 Xi[2]);
393 void gcm_gmult_vis3(u64 Xi[2], const u128 Htable[16]);
394 void gcm_ghash_vis3(u64 Xi[2], const u128 Htable[16], const u8 *inp,
399 void gcm_init_p8(u128 Htable[16], const u64 Xi[2]);
400 void gcm_gmult_p8(u64 Xi[2], const u128 Htable[16]);
401 void gcm_ghash_p8(u64 Xi[2], const u128 Htable[16], const u8 *inp,
407 void gcm_init_rv64i_zbc(u128 Htable[16], const u64 Xi[2]);
408 void gcm_init_rv64i_zbc__zbb(u128 Htable[16], const u64 Xi[2]);
409 void gcm_init_rv64i_zbc__zbkb(u128 Htable[16], const u64 Xi[2]);
410 void gcm_gmult_rv64i_zbc(u64 Xi[2], const u128 Htable[16]);
411 void gcm_gmult_rv64i_zbc__zbkb(u64 Xi[2], const u128 Htable[16]);
412 void gcm_ghash_rv64i_zbc(u64 Xi[2], const u128 Htable[16],
414 void gcm_ghash_rv64i_zbc__zbkb(u64 Xi[2], const u128 Htable[16],
417 void gcm_init_rv64i_zvkb_zvbc(u128 Htable[16], const u64 Xi[2]);
418 void gcm_gmult_rv64i_zvkb_zvbc(u64 Xi[2], const u128 Htable[16]);
419 void gcm_ghash_rv64i_zvkb_zvbc(u64 Xi[2], const u128 Htable[16],
422 void gcm_init_rv64i_zvkg(u128 Htable[16], const u64 Xi[2]);
423 void gcm_init_rv64i_zvkg_zvkb(u128 Htable[16], const u64 Xi[2]);
424 void gcm_gmult_rv64i_zvkg(u64 Xi[2], const u128 Htable[16]);
425 void gcm_ghash_rv64i_zvkg(u64 Xi[2], const u128 Htable[16],
565 void ossl_gcm_init_4bit(u128 Htable[16], const u64 H[2]) in ossl_gcm_init_4bit()
573 void ossl_gcm_gmult_4bit(u64 Xi[2], const u128 Htable[16]) in ossl_gcm_gmult_4bit()
581 void ossl_gcm_ghash_4bit(u64 Xi[2], const u128 Htable[16], in ossl_gcm_ghash_4bit()
585 u64 tmp[2]; in ossl_gcm_ghash_4bit()
619 u64 hi, lo; in CRYPTO_gcm128_init()
620 hi = (u64)GETU32(p) << 32 | GETU32(p + 4); in CRYPTO_gcm128_init()
621 lo = (u64)GETU32(p + 8) << 32 | GETU32(p + 12); in CRYPTO_gcm128_init()
651 u64 len0 = len; in CRYPTO_gcm128_setiv()
723 u64 alen = ctx->len.u[0]; in CRYPTO_gcm128_aad()
779 u64 mlen = ctx->len.u[1]; in CRYPTO_gcm128_encrypt()
1004 u64 mlen = ctx->len.u[1]; in CRYPTO_gcm128_decrypt()
1240 u64 mlen = ctx->len.u[1]; in CRYPTO_gcm128_encrypt_ctr32()
1394 u64 mlen = ctx->len.u[1]; in CRYPTO_gcm128_decrypt_ctr32()
1549 u64 alen = ctx->len.u[0] << 3; in CRYPTO_gcm128_finish()
1550 u64 clen = ctx->len.u[1] << 3; in CRYPTO_gcm128_finish()
1583 alen = (u64)GETU32(p) << 32 | GETU32(p + 4); in CRYPTO_gcm128_finish()
1584 clen = (u64)GETU32(p + 8) << 32 | GETU32(p + 12); in CRYPTO_gcm128_finish()