Lines Matching refs:xxh_u64

2319 typedef XXH64_hash_t xxh_u64;  typedef
2322 # define U64 xxh_u64
2333 static xxh_u64 XXH_read64(const void* memPtr) in XXH_read64()
2335 return *(const xxh_u64*) memPtr; in XXH_read64()
2347 typedef union { xxh_u32 u32; xxh_u64 u64; } __attribute__((packed)) unalign64;
2349 static xxh_u64 XXH_read64(const void* ptr) in XXH_read64()
2351 typedef union { xxh_u32 u32; xxh_u64 u64; } __attribute__((packed)) xxh_unalign64; in XXH_read64()
2361 static xxh_u64 XXH_read64(const void* memPtr) in XXH_read64()
2363 xxh_u64 val; in XXH_read64()
2375 static xxh_u64 XXH_swap64(xxh_u64 x) in XXH_swap64()
2392 XXH_FORCE_INLINE xxh_u64 XXH_readLE64(const void* memPtr) in XXH_readLE64()
2396 | ((xxh_u64)bytePtr[1] << 8) in XXH_readLE64()
2397 | ((xxh_u64)bytePtr[2] << 16) in XXH_readLE64()
2398 | ((xxh_u64)bytePtr[3] << 24) in XXH_readLE64()
2399 | ((xxh_u64)bytePtr[4] << 32) in XXH_readLE64()
2400 | ((xxh_u64)bytePtr[5] << 40) in XXH_readLE64()
2401 | ((xxh_u64)bytePtr[6] << 48) in XXH_readLE64()
2402 | ((xxh_u64)bytePtr[7] << 56); in XXH_readLE64()
2405 XXH_FORCE_INLINE xxh_u64 XXH_readBE64(const void* memPtr) in XXH_readBE64()
2409 | ((xxh_u64)bytePtr[6] << 8) in XXH_readBE64()
2410 | ((xxh_u64)bytePtr[5] << 16) in XXH_readBE64()
2411 | ((xxh_u64)bytePtr[4] << 24) in XXH_readBE64()
2412 | ((xxh_u64)bytePtr[3] << 32) in XXH_readBE64()
2413 | ((xxh_u64)bytePtr[2] << 40) in XXH_readBE64()
2414 | ((xxh_u64)bytePtr[1] << 48) in XXH_readBE64()
2415 | ((xxh_u64)bytePtr[0] << 56); in XXH_readBE64()
2419 XXH_FORCE_INLINE xxh_u64 XXH_readLE64(const void* ptr) in XXH_readLE64()
2424 static xxh_u64 XXH_readBE64(const void* ptr) in XXH_readBE64()
2430 XXH_FORCE_INLINE xxh_u64
2436 return XXH_CPU_LITTLE_ENDIAN ? *(const xxh_u64*)ptr : XXH_swap64(*(const xxh_u64*)ptr); in XXH_readLE64_align()
2462 static xxh_u64 XXH64_round(xxh_u64 acc, xxh_u64 input) in XXH64_round()
2470 static xxh_u64 XXH64_mergeRound(xxh_u64 acc, xxh_u64 val) in XXH64_mergeRound()
2478 static xxh_u64 XXH64_avalanche(xxh_u64 h64) in XXH64_avalanche()
2491 static xxh_u64
2492 XXH64_finalize(xxh_u64 h64, const xxh_u8* ptr, size_t len, XXH_alignment align) in XXH64_finalize()
2497 xxh_u64 const k1 = XXH64_round(0, XXH_get64bits(ptr)); in XXH64_finalize()
2504 h64 ^= (xxh_u64)(XXH_get32bits(ptr)) * XXH_PRIME64_1; in XXH64_finalize()
2527 XXH_FORCE_INLINE xxh_u64
2528 XXH64_endian_align(const xxh_u8* input, size_t len, xxh_u64 seed, XXH_alignment align) in XXH64_endian_align()
2530 xxh_u64 h64; in XXH64_endian_align()
2536 xxh_u64 v1 = seed + XXH_PRIME64_1 + XXH_PRIME64_2; in XXH64_endian_align()
2537 xxh_u64 v2 = seed + XXH_PRIME64_2; in XXH64_endian_align()
2538 xxh_u64 v3 = seed + 0; in XXH64_endian_align()
2539 xxh_u64 v4 = seed - XXH_PRIME64_1; in XXH64_endian_align()
2558 h64 += (xxh_u64) len; in XXH64_endian_align()
2673 xxh_u64 h64; in XXH64_digest()
2685 h64 += (xxh_u64) state->total_len; in XXH64_digest()
3240 XXH_FORCE_INLINE xxh_u64
3241 XXH_mult32to64(xxh_u64 x, xxh_u64 y) in XXH_mult32to64()
3256 # define XXH_mult32to64(x, y) ((xxh_u64)(xxh_u32)(x) * (xxh_u64)(xxh_u32)(y))
3269 XXH_mult64to128(xxh_u64 lhs, xxh_u64 rhs) in XXH_mult64to128()
3292 r128.low64 = (xxh_u64)(product); in XXH_mult64to128()
3293 r128.high64 = (xxh_u64)(product >> 64); in XXH_mult64to128()
3308 xxh_u64 product_high; in XXH_mult64to128()
3309 xxh_u64 const product_low = _umul128(lhs, rhs, &product_high); in XXH_mult64to128()
3375 xxh_u64 const lo_lo = XXH_mult32to64(lhs & 0xFFFFFFFF, rhs & 0xFFFFFFFF); in XXH_mult64to128()
3376 xxh_u64 const hi_lo = XXH_mult32to64(lhs >> 32, rhs & 0xFFFFFFFF); in XXH_mult64to128()
3377 xxh_u64 const lo_hi = XXH_mult32to64(lhs & 0xFFFFFFFF, rhs >> 32); in XXH_mult64to128()
3378 xxh_u64 const hi_hi = XXH_mult32to64(lhs >> 32, rhs >> 32); in XXH_mult64to128()
3381 xxh_u64 const cross = (lo_lo >> 32) + (hi_lo & 0xFFFFFFFF) + lo_hi; in XXH_mult64to128()
3382 xxh_u64 const upper = (hi_lo >> 32) + (cross >> 32) + hi_hi; in XXH_mult64to128()
3383 xxh_u64 const lower = (cross << 32) | (lo_lo & 0xFFFFFFFF); in XXH_mult64to128()
3402 static xxh_u64
3403 XXH3_mul128_fold64(xxh_u64 lhs, xxh_u64 rhs) in XXH3_mul128_fold64()
3410 XXH_FORCE_INLINE xxh_u64 XXH_xorshift64(xxh_u64 v64, int shift) in XXH_xorshift64()
3420 static XXH64_hash_t XXH3_avalanche(xxh_u64 h64) in XXH3_avalanche()
3433 static XXH64_hash_t XXH3_rrmxmx(xxh_u64 h64, xxh_u64 len) in XXH3_rrmxmx()
3493 xxh_u64 const bitflip = (XXH_readLE32(secret) ^ XXH_readLE32(secret+4)) + seed; in XXH3_len_1to3_64b()
3494 xxh_u64 const keyed = (xxh_u64)combined ^ bitflip; in XXH3_len_1to3_64b()
3505 seed ^= (xxh_u64)XXH_swap32((xxh_u32)seed) << 32; in XXH3_len_4to8_64b()
3508 xxh_u64 const bitflip = (XXH_readLE64(secret+8) ^ XXH_readLE64(secret+16)) - seed; in XXH3_len_4to8_64b()
3509 xxh_u64 const input64 = input2 + (((xxh_u64)input1) << 32); in XXH3_len_4to8_64b()
3510 xxh_u64 const keyed = input64 ^ bitflip; in XXH3_len_4to8_64b()
3521 { xxh_u64 const bitflip1 = (XXH_readLE64(secret+24) ^ XXH_readLE64(secret+32)) + seed; in XXH3_len_9to16_64b()
3522 xxh_u64 const bitflip2 = (XXH_readLE64(secret+40) ^ XXH_readLE64(secret+48)) - seed; in XXH3_len_9to16_64b()
3523 xxh_u64 const input_lo = XXH_readLE64(input) ^ bitflip1; in XXH3_len_9to16_64b()
3524 xxh_u64 const input_hi = XXH_readLE64(input + len - 8) ^ bitflip2; in XXH3_len_9to16_64b()
3525 xxh_u64 const acc = len in XXH3_len_9to16_64b()
3569 XXH_FORCE_INLINE xxh_u64 XXH3_mix16B(const xxh_u8* XXH_RESTRICT input, in XXH3_mix16B()
3570 const xxh_u8* XXH_RESTRICT secret, xxh_u64 seed64) in XXH3_mix16B()
3592 { xxh_u64 const input_lo = XXH_readLE64(input); in XXH3_mix16B()
3593 xxh_u64 const input_hi = XXH_readLE64(input+8); in XXH3_mix16B()
3610 { xxh_u64 acc = len * XXH_PRIME64_1; in XXH3_len_17to128_64b()
3643 { xxh_u64 acc = len * XXH_PRIME64_1; in XXH3_len_129to240_64b()
3690 #define XXH_ACC_NB (XXH_STRIPE_LEN / sizeof(xxh_u64))
3697 XXH_FORCE_INLINE void XXH_writeLE64(void* dst, xxh_u64 v64) in XXH_writeLE64()
3821 XXH3_initCustomSecret_avx512(void* XXH_RESTRICT customSecret, xxh_u64 seed64) in XXH3_initCustomSecret_avx512()
3919 …NE XXH_TARGET_AVX2 void XXH3_initCustomSecret_avx2(void* XXH_RESTRICT customSecret, xxh_u64 seed64) in XXH3_initCustomSecret_avx2()
4025 …NE XXH_TARGET_SSE2 void XXH3_initCustomSecret_sse2(void* XXH_RESTRICT customSecret, xxh_u64 seed64) in XXH3_initCustomSecret_sse2()
4231 xxh_u64* const xacc = (xxh_u64*) acc; /* presumed aligned */ in XXH3_accumulate_512_scalar()
4237 xxh_u64 const data_val = XXH_readLE64(xinput + 8*i); in XXH3_accumulate_512_scalar()
4238 xxh_u64 const data_key = data_val ^ XXH_readLE64(xsecret + i*8); in XXH3_accumulate_512_scalar()
4247 xxh_u64* const xacc = (xxh_u64*) acc; /* presumed aligned */ in XXH3_scrambleAcc_scalar()
4252 xxh_u64 const key64 = XXH_readLE64(xsecret + 8*i); in XXH3_scrambleAcc_scalar()
4253 xxh_u64 acc64 = xacc[i]; in XXH3_scrambleAcc_scalar()
4262 XXH3_initCustomSecret_scalar(void* XXH_RESTRICT customSecret, xxh_u64 seed64) in XXH3_initCustomSecret_scalar()
4318 xxh_u64 lo = XXH_readLE64(kSecretPtr + 16*i) + seed64; in XXH3_initCustomSecret_scalar()
4319 xxh_u64 hi = XXH_readLE64(kSecretPtr + 16*i + 8) - seed64; in XXH3_initCustomSecret_scalar()
4328 typedef void (*XXH3_f_initCustomSecret)(void* XXH_RESTRICT, xxh_u64);
4389 XXH3_accumulate( xxh_u64* XXH_RESTRICT acc, in XXH3_accumulate()
4406 XXH3_hashLong_internal_loop(xxh_u64* XXH_RESTRICT acc, in XXH3_hashLong_internal_loop()
4438 XXH_FORCE_INLINE xxh_u64
4439 XXH3_mix2Accs(const xxh_u64* XXH_RESTRICT acc, const xxh_u8* XXH_RESTRICT secret) in XXH3_mix2Accs()
4447 XXH3_mergeAccs(const xxh_u64* XXH_RESTRICT acc, const xxh_u8* XXH_RESTRICT secret, xxh_u64 start) in XXH3_mergeAccs()
4449 xxh_u64 result64 = start; in XXH3_mergeAccs()
4482 XXH_ALIGN(XXH_ACC_ALIGN) xxh_u64 acc[XXH_ACC_NB] = XXH3_INIT_ACC; in XXH3_hashLong_64b_internal()
4491 …return XXH3_mergeAccs(acc, (const xxh_u8*)secret + XXH_SECRET_MERGEACCS_START, (xxh_u64)len * XXH_… in XXH3_hashLong_64b_internal()
4785 XXH3_consumeStripes(xxh_u64* XXH_RESTRICT acc, in XXH3_consumeStripes()
4835 XXH_ALIGN(XXH_ACC_ALIGN) xxh_u64 acc[8]; memcpy(acc, state->acc, sizeof(acc)); in XXH3_update()
4837 xxh_u64* XXH_RESTRICT const acc = state->acc; in XXH3_update()
4983 (xxh_u64)state->totalLen * XXH_PRIME64_1); in XXH3_64bits_digest()
5029 xxh_u64 const bitflipl = (XXH_readLE32(secret) ^ XXH_readLE32(secret+4)) + seed; in XXH3_len_1to3_128b()
5030 xxh_u64 const bitfliph = (XXH_readLE32(secret+8) ^ XXH_readLE32(secret+12)) - seed; in XXH3_len_1to3_128b()
5031 xxh_u64 const keyed_lo = (xxh_u64)combinedl ^ bitflipl; in XXH3_len_1to3_128b()
5032 xxh_u64 const keyed_hi = (xxh_u64)combinedh ^ bitfliph; in XXH3_len_1to3_128b()
5046 seed ^= (xxh_u64)XXH_swap32((xxh_u32)seed) << 32; in XXH3_len_4to8_128b()
5049 xxh_u64 const input_64 = input_lo + ((xxh_u64)input_hi << 32); in XXH3_len_4to8_128b()
5050 xxh_u64 const bitflip = (XXH_readLE64(secret+16) ^ XXH_readLE64(secret+24)) + seed; in XXH3_len_4to8_128b()
5051 xxh_u64 const keyed = input_64 ^ bitflip; in XXH3_len_4to8_128b()
5073 { xxh_u64 const bitflipl = (XXH_readLE64(secret+32) ^ XXH_readLE64(secret+40)) - seed; in XXH3_len_9to16_128b()
5074 xxh_u64 const bitfliph = (XXH_readLE64(secret+48) ^ XXH_readLE64(secret+56)) + seed; in XXH3_len_9to16_128b()
5075 xxh_u64 const input_lo = XXH_readLE64(input); in XXH3_len_9to16_128b()
5076 xxh_u64 input_hi = XXH_readLE64(input + len - 8); in XXH3_len_9to16_128b()
5082 m128.low64 += (xxh_u64)(len - 1) << 54; in XXH3_len_9to16_128b()
5091 if (sizeof(void *) < sizeof(xxh_u64)) { /* 32-bit */ in XXH3_len_9to16_128b()
5150 xxh_u64 const bitflipl = XXH_readLE64(secret+64) ^ XXH_readLE64(secret+72); in XXH3_len_0to16_128b()
5151 xxh_u64 const bitfliph = XXH_readLE64(secret+80) ^ XXH_readLE64(secret+88); in XXH3_len_0to16_128b()
5261 XXH_ALIGN(XXH_ACC_ALIGN) xxh_u64 acc[XXH_ACC_NB] = XXH3_INIT_ACC; in XXH3_hashLong_128b_internal()
5271 (xxh_u64)len * XXH_PRIME64_1); in XXH3_hashLong_128b_internal()
5275 ~((xxh_u64)len * XXH_PRIME64_2)); in XXH3_hashLong_128b_internal()
5461 (xxh_u64)state->totalLen * XXH_PRIME64_1); in XXH3_128bits_digest()
5465 ~((xxh_u64)state->totalLen * XXH_PRIME64_2)); in XXH3_128bits_digest()