Lines Matching refs:xxh_u64

1564 typedef XXH64_hash_t xxh_u64;  typedef
1567 # define U64 xxh_u64
1607 static xxh_u64 XXH_read64(const void* memPtr) { return *(const xxh_u64*) memPtr; } in XXH_read64()
1618 typedef union { xxh_u32 u32; xxh_u64 u64; } __attribute__((packed)) unalign64;
1620 static xxh_u64 XXH_read64(const void* ptr) in XXH_read64()
1622 typedef union { xxh_u32 u32; xxh_u64 u64; } __attribute__((packed)) xxh_unalign64; in XXH_read64()
1632 static xxh_u64 XXH_read64(const void* memPtr) in XXH_read64()
1634 xxh_u64 val; in XXH_read64()
1646 static xxh_u64 XXH_swap64 (xxh_u64 x) in XXH_swap64()
1663 XXH_FORCE_INLINE xxh_u64 XXH_readLE64(const void* memPtr) in XXH_readLE64()
1667 | ((xxh_u64)bytePtr[1] << 8) in XXH_readLE64()
1668 | ((xxh_u64)bytePtr[2] << 16) in XXH_readLE64()
1669 | ((xxh_u64)bytePtr[3] << 24) in XXH_readLE64()
1670 | ((xxh_u64)bytePtr[4] << 32) in XXH_readLE64()
1671 | ((xxh_u64)bytePtr[5] << 40) in XXH_readLE64()
1672 | ((xxh_u64)bytePtr[6] << 48) in XXH_readLE64()
1673 | ((xxh_u64)bytePtr[7] << 56); in XXH_readLE64()
1676 XXH_FORCE_INLINE xxh_u64 XXH_readBE64(const void* memPtr) in XXH_readBE64()
1680 | ((xxh_u64)bytePtr[6] << 8) in XXH_readBE64()
1681 | ((xxh_u64)bytePtr[5] << 16) in XXH_readBE64()
1682 | ((xxh_u64)bytePtr[4] << 24) in XXH_readBE64()
1683 | ((xxh_u64)bytePtr[3] << 32) in XXH_readBE64()
1684 | ((xxh_u64)bytePtr[2] << 40) in XXH_readBE64()
1685 | ((xxh_u64)bytePtr[1] << 48) in XXH_readBE64()
1686 | ((xxh_u64)bytePtr[0] << 56); in XXH_readBE64()
1690 XXH_FORCE_INLINE xxh_u64 XXH_readLE64(const void* ptr) in XXH_readLE64()
1695 static xxh_u64 XXH_readBE64(const void* ptr) in XXH_readBE64()
1701 XXH_FORCE_INLINE xxh_u64
1707 return XXH_CPU_LITTLE_ENDIAN ? *(const xxh_u64*)ptr : XXH_swap64(*(const xxh_u64*)ptr); in XXH_readLE64_align()
1713 static const xxh_u64 XXH_PRIME64_1 = 0x9E3779B185EBCA87ULL; /* 0b10011110001101110111100110110001…
1714 static const xxh_u64 XXH_PRIME64_2 = 0xC2B2AE3D27D4EB4FULL; /* 0b11000010101100101010111000111101…
1715 static const xxh_u64 XXH_PRIME64_3 = 0x165667B19E3779F9ULL; /* 0b00010110010101100110011110110001…
1716 static const xxh_u64 XXH_PRIME64_4 = 0x85EBCA77C2B2AE63ULL; /* 0b10000101111010111100101001110111…
1717 static const xxh_u64 XXH_PRIME64_5 = 0x27D4EB2F165667C5ULL; /* 0b00100111110101001110101100101111…
1727 static xxh_u64 XXH64_round(xxh_u64 acc, xxh_u64 input) in XXH64_round()
1735 static xxh_u64 XXH64_mergeRound(xxh_u64 acc, xxh_u64 val) in XXH64_mergeRound()
1743 static xxh_u64 XXH64_avalanche(xxh_u64 h64) in XXH64_avalanche()
1756 static xxh_u64
1757 XXH64_finalize(xxh_u64 h64, const xxh_u8* ptr, size_t len, XXH_alignment align) in XXH64_finalize()
1765 h64 ^= (xxh_u64)(XXH_get32bits(ptr)) * XXH_PRIME64_1; \ in XXH64_finalize()
1771 xxh_u64 const k1 = XXH64_round(0, XXH_get64bits(ptr)); \ in XXH64_finalize()
1891 XXH_FORCE_INLINE xxh_u64
1892 XXH64_endian_align(const xxh_u8* input, size_t len, xxh_u64 seed, XXH_alignment align) in XXH64_endian_align()
1895 xxh_u64 h64; in XXH64_endian_align()
1906 xxh_u64 v1 = seed + XXH_PRIME64_1 + XXH_PRIME64_2; in XXH64_endian_align()
1907 xxh_u64 v2 = seed + XXH_PRIME64_2; in XXH64_endian_align()
1908 xxh_u64 v3 = seed + 0; in XXH64_endian_align()
1909 xxh_u64 v4 = seed - XXH_PRIME64_1; in XXH64_endian_align()
1928 h64 += (xxh_u64) len; in XXH64_endian_align()
2018 xxh_u64 v1 = state->v1; in XXH64_update()
2019 xxh_u64 v2 = state->v2; in XXH64_update()
2020 xxh_u64 v3 = state->v3; in XXH64_update()
2021 xxh_u64 v4 = state->v4; in XXH64_update()
2048 xxh_u64 h64; in XXH64_digest()
2051 xxh_u64 const v1 = state->v1; in XXH64_digest()
2052 xxh_u64 const v2 = state->v2; in XXH64_digest()
2053 xxh_u64 const v3 = state->v3; in XXH64_digest()
2054 xxh_u64 const v4 = state->v4; in XXH64_digest()
2065 h64 += (xxh_u64) state->total_len; in XXH64_digest()
2565 # define XXH_mult32to64(x, y) ((xxh_u64)(xxh_u32)(x) * (xxh_u64)(xxh_u32)(y))
2574 XXH_mult64to128(xxh_u64 lhs, xxh_u64 rhs) in XXH_mult64to128()
2597 r128.low64 = (xxh_u64)(product); in XXH_mult64to128()
2598 r128.high64 = (xxh_u64)(product >> 64); in XXH_mult64to128()
2613 xxh_u64 product_high; in XXH_mult64to128()
2614 xxh_u64 const product_low = _umul128(lhs, rhs, &product_high); in XXH_mult64to128()
2665 xxh_u64 const lo_lo = XXH_mult32to64(lhs & 0xFFFFFFFF, rhs & 0xFFFFFFFF); in XXH_mult64to128()
2666 xxh_u64 const hi_lo = XXH_mult32to64(lhs >> 32, rhs & 0xFFFFFFFF); in XXH_mult64to128()
2667 xxh_u64 const lo_hi = XXH_mult32to64(lhs & 0xFFFFFFFF, rhs >> 32); in XXH_mult64to128()
2668 xxh_u64 const hi_hi = XXH_mult32to64(lhs >> 32, rhs >> 32); in XXH_mult64to128()
2671 xxh_u64 const cross = (lo_lo >> 32) + (hi_lo & 0xFFFFFFFF) + lo_hi; in XXH_mult64to128()
2672 xxh_u64 const upper = (hi_lo >> 32) + (cross >> 32) + hi_hi; in XXH_mult64to128()
2673 xxh_u64 const lower = (cross << 32) | (lo_lo & 0xFFFFFFFF); in XXH_mult64to128()
2688 static xxh_u64
2689 XXH3_mul128_fold64(xxh_u64 lhs, xxh_u64 rhs) in XXH3_mul128_fold64()
2696 XXH_FORCE_INLINE xxh_u64 XXH_xorshift64(xxh_u64 v64, int shift) in XXH_xorshift64()
2706 static XXH64_hash_t XXH3_avalanche(xxh_u64 h64) in XXH3_avalanche()
2719 static XXH64_hash_t XXH3_rrmxmx(xxh_u64 h64, xxh_u64 len) in XXH3_rrmxmx()
2779 xxh_u64 const bitflip = (XXH_readLE32(secret) ^ XXH_readLE32(secret+4)) + seed; in XXH3_len_1to3_64b()
2780 xxh_u64 const keyed = (xxh_u64)combined ^ bitflip; in XXH3_len_1to3_64b()
2791 seed ^= (xxh_u64)XXH_swap32((xxh_u32)seed) << 32; in XXH3_len_4to8_64b()
2794 xxh_u64 const bitflip = (XXH_readLE64(secret+8) ^ XXH_readLE64(secret+16)) - seed; in XXH3_len_4to8_64b()
2795 xxh_u64 const input64 = input2 + (((xxh_u64)input1) << 32); in XXH3_len_4to8_64b()
2796 xxh_u64 const keyed = input64 ^ bitflip; in XXH3_len_4to8_64b()
2807 { xxh_u64 const bitflip1 = (XXH_readLE64(secret+24) ^ XXH_readLE64(secret+32)) + seed; in XXH3_len_9to16_64b()
2808 xxh_u64 const bitflip2 = (XXH_readLE64(secret+40) ^ XXH_readLE64(secret+48)) - seed; in XXH3_len_9to16_64b()
2809 xxh_u64 const input_lo = XXH_readLE64(input) ^ bitflip1; in XXH3_len_9to16_64b()
2810 xxh_u64 const input_hi = XXH_readLE64(input + len - 8) ^ bitflip2; in XXH3_len_9to16_64b()
2811 xxh_u64 const acc = len in XXH3_len_9to16_64b()
2855 XXH_FORCE_INLINE xxh_u64 XXH3_mix16B(const xxh_u8* XXH_RESTRICT input, in XXH3_mix16B()
2856 const xxh_u8* XXH_RESTRICT secret, xxh_u64 seed64) in XXH3_mix16B()
2878 { xxh_u64 const input_lo = XXH_readLE64(input); in XXH3_mix16B()
2879 xxh_u64 const input_hi = XXH_readLE64(input+8); in XXH3_mix16B()
2896 { xxh_u64 acc = len * XXH_PRIME64_1; in XXH3_len_17to128_64b()
2929 { xxh_u64 acc = len * XXH_PRIME64_1; in XXH3_len_129to240_64b()
2976 #define XXH_ACC_NB (XXH_STRIPE_LEN / sizeof(xxh_u64))
2983 XXH_FORCE_INLINE void XXH_writeLE64(void* dst, xxh_u64 v64) in XXH_writeLE64()
3106 XXH3_initCustomSecret_avx512(void* XXH_RESTRICT customSecret, xxh_u64 seed64) in XXH3_initCustomSecret_avx512()
3201 …NE XXH_TARGET_AVX2 void XXH3_initCustomSecret_avx2(void* XXH_RESTRICT customSecret, xxh_u64 seed64) in XXH3_initCustomSecret_avx2()
3308 …NE XXH_TARGET_SSE2 void XXH3_initCustomSecret_sse2(void* XXH_RESTRICT customSecret, xxh_u64 seed64) in XXH3_initCustomSecret_sse2()
3507 XXH_ALIGN(XXH_ACC_ALIGN) xxh_u64* const xacc = (xxh_u64*) acc; /* presumed aligned */ in XXH3_accumulate_512_scalar()
3513 xxh_u64 const data_val = XXH_readLE64(xinput + 8*i); in XXH3_accumulate_512_scalar()
3514 xxh_u64 const data_key = data_val ^ XXH_readLE64(xsecret + i*8); in XXH3_accumulate_512_scalar()
3523 XXH_ALIGN(XXH_ACC_ALIGN) xxh_u64* const xacc = (xxh_u64*) acc; /* presumed aligned */ in XXH3_scrambleAcc_scalar()
3528 xxh_u64 const key64 = XXH_readLE64(xsecret + 8*i); in XXH3_scrambleAcc_scalar()
3529 xxh_u64 acc64 = xacc[i]; in XXH3_scrambleAcc_scalar()
3538 XXH3_initCustomSecret_scalar(void* XXH_RESTRICT customSecret, xxh_u64 seed64) in XXH3_initCustomSecret_scalar()
3594 xxh_u64 lo = XXH_readLE64(kSecretPtr + 16*i) + seed64; in XXH3_initCustomSecret_scalar()
3595 xxh_u64 hi = XXH_readLE64(kSecretPtr + 16*i + 8) - seed64; in XXH3_initCustomSecret_scalar()
3604 typedef void (*XXH3_f_initCustomSecret)(void* XXH_RESTRICT, xxh_u64);
3665 XXH3_accumulate( xxh_u64* XXH_RESTRICT acc, in XXH3_accumulate()
3682 XXH3_hashLong_internal_loop(xxh_u64* XXH_RESTRICT acc, in XXH3_hashLong_internal_loop()
3714 XXH_FORCE_INLINE xxh_u64
3715 XXH3_mix2Accs(const xxh_u64* XXH_RESTRICT acc, const xxh_u8* XXH_RESTRICT secret) in XXH3_mix2Accs()
3723 XXH3_mergeAccs(const xxh_u64* XXH_RESTRICT acc, const xxh_u8* XXH_RESTRICT secret, xxh_u64 start) in XXH3_mergeAccs()
3725 xxh_u64 result64 = start; in XXH3_mergeAccs()
3758 XXH_ALIGN(XXH_ACC_ALIGN) xxh_u64 acc[XXH_ACC_NB] = XXH3_INIT_ACC; in XXH3_hashLong_64b_internal()
3767 …return XXH3_mergeAccs(acc, (const xxh_u8*)secret + XXH_SECRET_MERGEACCS_START, (xxh_u64)len * XXH_… in XXH3_hashLong_64b_internal()
4029 XXH3_consumeStripes(xxh_u64* XXH_RESTRICT acc, in XXH3_consumeStripes()
4174 (xxh_u64)state->totalLen * XXH_PRIME64_1); in XXH3_64bits_digest()
4264 xxh_u64 const bitflipl = (XXH_readLE32(secret) ^ XXH_readLE32(secret+4)) + seed; in XXH3_len_1to3_128b()
4265 xxh_u64 const bitfliph = (XXH_readLE32(secret+8) ^ XXH_readLE32(secret+12)) - seed; in XXH3_len_1to3_128b()
4266 xxh_u64 const keyed_lo = (xxh_u64)combinedl ^ bitflipl; in XXH3_len_1to3_128b()
4267 xxh_u64 const keyed_hi = (xxh_u64)combinedh ^ bitfliph; in XXH3_len_1to3_128b()
4281 seed ^= (xxh_u64)XXH_swap32((xxh_u32)seed) << 32; in XXH3_len_4to8_128b()
4284 xxh_u64 const input_64 = input_lo + ((xxh_u64)input_hi << 32); in XXH3_len_4to8_128b()
4285 xxh_u64 const bitflip = (XXH_readLE64(secret+16) ^ XXH_readLE64(secret+24)) + seed; in XXH3_len_4to8_128b()
4286 xxh_u64 const keyed = input_64 ^ bitflip; in XXH3_len_4to8_128b()
4308 { xxh_u64 const bitflipl = (XXH_readLE64(secret+32) ^ XXH_readLE64(secret+40)) - seed; in XXH3_len_9to16_128b()
4309 xxh_u64 const bitfliph = (XXH_readLE64(secret+48) ^ XXH_readLE64(secret+56)) + seed; in XXH3_len_9to16_128b()
4310 xxh_u64 const input_lo = XXH_readLE64(input); in XXH3_len_9to16_128b()
4311 xxh_u64 input_hi = XXH_readLE64(input + len - 8); in XXH3_len_9to16_128b()
4317 m128.low64 += (xxh_u64)(len - 1) << 54; in XXH3_len_9to16_128b()
4326 if (sizeof(void *) < sizeof(xxh_u64)) { /* 32-bit */ in XXH3_len_9to16_128b()
4385 xxh_u64 const bitflipl = XXH_readLE64(secret+64) ^ XXH_readLE64(secret+72); in XXH3_len_0to16_128b()
4386 xxh_u64 const bitfliph = XXH_readLE64(secret+80) ^ XXH_readLE64(secret+88); in XXH3_len_0to16_128b()
4496 XXH_ALIGN(XXH_ACC_ALIGN) xxh_u64 acc[XXH_ACC_NB] = XXH3_INIT_ACC; in XXH3_hashLong_128b_internal()
4506 (xxh_u64)len * XXH_PRIME64_1); in XXH3_hashLong_128b_internal()
4510 ~((xxh_u64)len * XXH_PRIME64_2)); in XXH3_hashLong_128b_internal()
4688 (xxh_u64)state->totalLen * XXH_PRIME64_1); in XXH3_128bits_digest()
4692 ~((xxh_u64)state->totalLen * XXH_PRIME64_2)); in XXH3_128bits_digest()