Lines Matching refs:h64

1743 static xxh_u64 XXH64_avalanche(xxh_u64 h64)  in XXH64_avalanche()  argument
1745 h64 ^= h64 >> 33; in XXH64_avalanche()
1746 h64 *= XXH_PRIME64_2; in XXH64_avalanche()
1747 h64 ^= h64 >> 29; in XXH64_avalanche()
1748 h64 *= XXH_PRIME64_3; in XXH64_avalanche()
1749 h64 ^= h64 >> 32; in XXH64_avalanche()
1750 return h64; in XXH64_avalanche()
1757 XXH64_finalize(xxh_u64 h64, const xxh_u8* ptr, size_t len, XXH_alignment align) in XXH64_finalize() argument
1760 h64 ^= (*ptr++) * XXH_PRIME64_5; \ in XXH64_finalize()
1761 h64 = XXH_rotl64(h64, 11) * XXH_PRIME64_1; \ in XXH64_finalize()
1765 h64 ^= (xxh_u64)(XXH_get32bits(ptr)) * XXH_PRIME64_1; \ in XXH64_finalize()
1767 h64 = XXH_rotl64(h64, 23) * XXH_PRIME64_2 + XXH_PRIME64_3; \ in XXH64_finalize()
1773 h64 ^= k1; \ in XXH64_finalize()
1774 h64 = XXH_rotl64(h64,27) * XXH_PRIME64_1 + XXH_PRIME64_4; \ in XXH64_finalize()
1792 return XXH64_avalanche(h64); in XXH64_finalize()
1800 return XXH64_avalanche(h64); in XXH64_finalize()
1809 return XXH64_avalanche(h64); in XXH64_finalize()
1817 return XXH64_avalanche(h64); in XXH64_finalize()
1827 return XXH64_avalanche(h64); in XXH64_finalize()
1836 return XXH64_avalanche(h64); in XXH64_finalize()
1847 return XXH64_avalanche(h64); in XXH64_finalize()
1857 return XXH64_avalanche(h64); in XXH64_finalize()
1873 case 0: return XXH64_avalanche(h64); in XXH64_finalize()
1895 xxh_u64 h64; in XXH64_endian_align() local
1918 h64 = XXH_rotl64(v1, 1) + XXH_rotl64(v2, 7) + XXH_rotl64(v3, 12) + XXH_rotl64(v4, 18); in XXH64_endian_align()
1919 h64 = XXH64_mergeRound(h64, v1); in XXH64_endian_align()
1920 h64 = XXH64_mergeRound(h64, v2); in XXH64_endian_align()
1921 h64 = XXH64_mergeRound(h64, v3); in XXH64_endian_align()
1922 h64 = XXH64_mergeRound(h64, v4); in XXH64_endian_align()
1925 h64 = seed + XXH_PRIME64_5; in XXH64_endian_align()
1928 h64 += (xxh_u64) len; in XXH64_endian_align()
1930 return XXH64_finalize(h64, input, len, align); in XXH64_endian_align()
2048 xxh_u64 h64; in XXH64_digest() local
2056 h64 = XXH_rotl64(v1, 1) + XXH_rotl64(v2, 7) + XXH_rotl64(v3, 12) + XXH_rotl64(v4, 18); in XXH64_digest()
2057 h64 = XXH64_mergeRound(h64, v1); in XXH64_digest()
2058 h64 = XXH64_mergeRound(h64, v2); in XXH64_digest()
2059 h64 = XXH64_mergeRound(h64, v3); in XXH64_digest()
2060 h64 = XXH64_mergeRound(h64, v4); in XXH64_digest()
2062 h64 = state->v3 /*seed*/ + XXH_PRIME64_5; in XXH64_digest()
2065 h64 += (xxh_u64) state->total_len; in XXH64_digest()
2067 return XXH64_finalize(h64, (const xxh_u8*)state->mem64, (size_t)state->total_len, XXH_aligned); in XXH64_digest()
2706 static XXH64_hash_t XXH3_avalanche(xxh_u64 h64) in XXH3_avalanche() argument
2708 h64 = XXH_xorshift64(h64, 37); in XXH3_avalanche()
2709 h64 *= 0x165667919E3779F9ULL; in XXH3_avalanche()
2710 h64 = XXH_xorshift64(h64, 32); in XXH3_avalanche()
2711 return h64; in XXH3_avalanche()
2719 static XXH64_hash_t XXH3_rrmxmx(xxh_u64 h64, xxh_u64 len) in XXH3_rrmxmx() argument
2722 h64 ^= XXH_rotl64(h64, 49) ^ XXH_rotl64(h64, 24); in XXH3_rrmxmx()
2723 h64 *= 0x9FB21C651E98DF25ULL; in XXH3_rrmxmx()
2724 h64 ^= (h64 >> 35) + len ; in XXH3_rrmxmx()
2725 h64 *= 0x9FB21C651E98DF25ULL; in XXH3_rrmxmx()
2726 return XXH_xorshift64(h64, 28); in XXH3_rrmxmx()