Lines Matching refs:xxh_u64

2288 typedef XXH64_hash_t xxh_u64;  typedef
2291 # define U64 xxh_u64
2302 static xxh_u64 XXH_read64(const void* memPtr) in XXH_read64()
2304 return *(const xxh_u64*) memPtr; in XXH_read64()
2316 typedef union { xxh_u32 u32; xxh_u64 u64; } __attribute__((packed)) unalign64;
2318 static xxh_u64 XXH_read64(const void* ptr) in XXH_read64()
2320 typedef union { xxh_u32 u32; xxh_u64 u64; } __attribute__((packed)) xxh_unalign64; in XXH_read64()
2330 static xxh_u64 XXH_read64(const void* memPtr) in XXH_read64()
2332 xxh_u64 val; in XXH_read64()
2344 static xxh_u64 XXH_swap64(xxh_u64 x) in XXH_swap64()
2361 XXH_FORCE_INLINE xxh_u64 XXH_readLE64(const void* memPtr) in XXH_readLE64()
2365 | ((xxh_u64)bytePtr[1] << 8) in XXH_readLE64()
2366 | ((xxh_u64)bytePtr[2] << 16) in XXH_readLE64()
2367 | ((xxh_u64)bytePtr[3] << 24) in XXH_readLE64()
2368 | ((xxh_u64)bytePtr[4] << 32) in XXH_readLE64()
2369 | ((xxh_u64)bytePtr[5] << 40) in XXH_readLE64()
2370 | ((xxh_u64)bytePtr[6] << 48) in XXH_readLE64()
2371 | ((xxh_u64)bytePtr[7] << 56); in XXH_readLE64()
2374 XXH_FORCE_INLINE xxh_u64 XXH_readBE64(const void* memPtr) in XXH_readBE64()
2378 | ((xxh_u64)bytePtr[6] << 8) in XXH_readBE64()
2379 | ((xxh_u64)bytePtr[5] << 16) in XXH_readBE64()
2380 | ((xxh_u64)bytePtr[4] << 24) in XXH_readBE64()
2381 | ((xxh_u64)bytePtr[3] << 32) in XXH_readBE64()
2382 | ((xxh_u64)bytePtr[2] << 40) in XXH_readBE64()
2383 | ((xxh_u64)bytePtr[1] << 48) in XXH_readBE64()
2384 | ((xxh_u64)bytePtr[0] << 56); in XXH_readBE64()
2388 XXH_FORCE_INLINE xxh_u64 XXH_readLE64(const void* ptr) in XXH_readLE64()
2393 static xxh_u64 XXH_readBE64(const void* ptr) in XXH_readBE64()
2399 XXH_FORCE_INLINE xxh_u64
2405 return XXH_CPU_LITTLE_ENDIAN ? *(const xxh_u64*)ptr : XXH_swap64(*(const xxh_u64*)ptr); in XXH_readLE64_align()
2431 static xxh_u64 XXH64_round(xxh_u64 acc, xxh_u64 input) in XXH64_round()
2439 static xxh_u64 XXH64_mergeRound(xxh_u64 acc, xxh_u64 val) in XXH64_mergeRound()
2447 static xxh_u64 XXH64_avalanche(xxh_u64 h64) in XXH64_avalanche()
2460 static xxh_u64
2461 XXH64_finalize(xxh_u64 h64, const xxh_u8* ptr, size_t len, XXH_alignment align) in XXH64_finalize()
2466 xxh_u64 const k1 = XXH64_round(0, XXH_get64bits(ptr)); in XXH64_finalize()
2473 h64 ^= (xxh_u64)(XXH_get32bits(ptr)) * XXH_PRIME64_1; in XXH64_finalize()
2496 XXH_FORCE_INLINE xxh_u64
2497 XXH64_endian_align(const xxh_u8* input, size_t len, xxh_u64 seed, XXH_alignment align) in XXH64_endian_align()
2499 xxh_u64 h64; in XXH64_endian_align()
2505 xxh_u64 v1 = seed + XXH_PRIME64_1 + XXH_PRIME64_2; in XXH64_endian_align()
2506 xxh_u64 v2 = seed + XXH_PRIME64_2; in XXH64_endian_align()
2507 xxh_u64 v3 = seed + 0; in XXH64_endian_align()
2508 xxh_u64 v4 = seed - XXH_PRIME64_1; in XXH64_endian_align()
2527 h64 += (xxh_u64) len; in XXH64_endian_align()
2642 xxh_u64 h64; in XXH64_digest()
2654 h64 += (xxh_u64) state->total_len; in XXH64_digest()
3209 XXH_FORCE_INLINE xxh_u64
3210 XXH_mult32to64(xxh_u64 x, xxh_u64 y) in XXH_mult32to64()
3225 # define XXH_mult32to64(x, y) ((xxh_u64)(xxh_u32)(x) * (xxh_u64)(xxh_u32)(y))
3238 XXH_mult64to128(xxh_u64 lhs, xxh_u64 rhs) in XXH_mult64to128()
3261 r128.low64 = (xxh_u64)(product); in XXH_mult64to128()
3262 r128.high64 = (xxh_u64)(product >> 64); in XXH_mult64to128()
3277 xxh_u64 product_high; in XXH_mult64to128()
3278 xxh_u64 const product_low = _umul128(lhs, rhs, &product_high); in XXH_mult64to128()
3344 xxh_u64 const lo_lo = XXH_mult32to64(lhs & 0xFFFFFFFF, rhs & 0xFFFFFFFF); in XXH_mult64to128()
3345 xxh_u64 const hi_lo = XXH_mult32to64(lhs >> 32, rhs & 0xFFFFFFFF); in XXH_mult64to128()
3346 xxh_u64 const lo_hi = XXH_mult32to64(lhs & 0xFFFFFFFF, rhs >> 32); in XXH_mult64to128()
3347 xxh_u64 const hi_hi = XXH_mult32to64(lhs >> 32, rhs >> 32); in XXH_mult64to128()
3350 xxh_u64 const cross = (lo_lo >> 32) + (hi_lo & 0xFFFFFFFF) + lo_hi; in XXH_mult64to128()
3351 xxh_u64 const upper = (hi_lo >> 32) + (cross >> 32) + hi_hi; in XXH_mult64to128()
3352 xxh_u64 const lower = (cross << 32) | (lo_lo & 0xFFFFFFFF); in XXH_mult64to128()
3371 static xxh_u64
3372 XXH3_mul128_fold64(xxh_u64 lhs, xxh_u64 rhs) in XXH3_mul128_fold64()
3379 XXH_FORCE_INLINE xxh_u64 XXH_xorshift64(xxh_u64 v64, int shift) in XXH_xorshift64()
3389 static XXH64_hash_t XXH3_avalanche(xxh_u64 h64) in XXH3_avalanche()
3402 static XXH64_hash_t XXH3_rrmxmx(xxh_u64 h64, xxh_u64 len) in XXH3_rrmxmx()
3462 xxh_u64 const bitflip = (XXH_readLE32(secret) ^ XXH_readLE32(secret+4)) + seed; in XXH3_len_1to3_64b()
3463 xxh_u64 const keyed = (xxh_u64)combined ^ bitflip; in XXH3_len_1to3_64b()
3474 seed ^= (xxh_u64)XXH_swap32((xxh_u32)seed) << 32; in XXH3_len_4to8_64b()
3477 xxh_u64 const bitflip = (XXH_readLE64(secret+8) ^ XXH_readLE64(secret+16)) - seed; in XXH3_len_4to8_64b()
3478 xxh_u64 const input64 = input2 + (((xxh_u64)input1) << 32); in XXH3_len_4to8_64b()
3479 xxh_u64 const keyed = input64 ^ bitflip; in XXH3_len_4to8_64b()
3490 { xxh_u64 const bitflip1 = (XXH_readLE64(secret+24) ^ XXH_readLE64(secret+32)) + seed; in XXH3_len_9to16_64b()
3491 xxh_u64 const bitflip2 = (XXH_readLE64(secret+40) ^ XXH_readLE64(secret+48)) - seed; in XXH3_len_9to16_64b()
3492 xxh_u64 const input_lo = XXH_readLE64(input) ^ bitflip1; in XXH3_len_9to16_64b()
3493 xxh_u64 const input_hi = XXH_readLE64(input + len - 8) ^ bitflip2; in XXH3_len_9to16_64b()
3494 xxh_u64 const acc = len in XXH3_len_9to16_64b()
3538 XXH_FORCE_INLINE xxh_u64 XXH3_mix16B(const xxh_u8* XXH_RESTRICT input, in XXH3_mix16B()
3539 const xxh_u8* XXH_RESTRICT secret, xxh_u64 seed64) in XXH3_mix16B()
3561 { xxh_u64 const input_lo = XXH_readLE64(input); in XXH3_mix16B()
3562 xxh_u64 const input_hi = XXH_readLE64(input+8); in XXH3_mix16B()
3579 { xxh_u64 acc = len * XXH_PRIME64_1; in XXH3_len_17to128_64b()
3612 { xxh_u64 acc = len * XXH_PRIME64_1; in XXH3_len_129to240_64b()
3659 #define XXH_ACC_NB (XXH_STRIPE_LEN / sizeof(xxh_u64))
3666 XXH_FORCE_INLINE void XXH_writeLE64(void* dst, xxh_u64 v64) in XXH_writeLE64()
3790 XXH3_initCustomSecret_avx512(void* XXH_RESTRICT customSecret, xxh_u64 seed64) in XXH3_initCustomSecret_avx512()
3888 …NE XXH_TARGET_AVX2 void XXH3_initCustomSecret_avx2(void* XXH_RESTRICT customSecret, xxh_u64 seed64) in XXH3_initCustomSecret_avx2()
3994 …NE XXH_TARGET_SSE2 void XXH3_initCustomSecret_sse2(void* XXH_RESTRICT customSecret, xxh_u64 seed64) in XXH3_initCustomSecret_sse2()
4200 xxh_u64* const xacc = (xxh_u64*) acc; /* presumed aligned */ in XXH3_accumulate_512_scalar()
4206 xxh_u64 const data_val = XXH_readLE64(xinput + 8*i); in XXH3_accumulate_512_scalar()
4207 xxh_u64 const data_key = data_val ^ XXH_readLE64(xsecret + i*8); in XXH3_accumulate_512_scalar()
4216 xxh_u64* const xacc = (xxh_u64*) acc; /* presumed aligned */ in XXH3_scrambleAcc_scalar()
4221 xxh_u64 const key64 = XXH_readLE64(xsecret + 8*i); in XXH3_scrambleAcc_scalar()
4222 xxh_u64 acc64 = xacc[i]; in XXH3_scrambleAcc_scalar()
4231 XXH3_initCustomSecret_scalar(void* XXH_RESTRICT customSecret, xxh_u64 seed64) in XXH3_initCustomSecret_scalar()
4287 xxh_u64 lo = XXH_readLE64(kSecretPtr + 16*i) + seed64; in XXH3_initCustomSecret_scalar()
4288 xxh_u64 hi = XXH_readLE64(kSecretPtr + 16*i + 8) - seed64; in XXH3_initCustomSecret_scalar()
4297 typedef void (*XXH3_f_initCustomSecret)(void* XXH_RESTRICT, xxh_u64);
4358 XXH3_accumulate( xxh_u64* XXH_RESTRICT acc, in XXH3_accumulate()
4375 XXH3_hashLong_internal_loop(xxh_u64* XXH_RESTRICT acc, in XXH3_hashLong_internal_loop()
4407 XXH_FORCE_INLINE xxh_u64
4408 XXH3_mix2Accs(const xxh_u64* XXH_RESTRICT acc, const xxh_u8* XXH_RESTRICT secret) in XXH3_mix2Accs()
4416 XXH3_mergeAccs(const xxh_u64* XXH_RESTRICT acc, const xxh_u8* XXH_RESTRICT secret, xxh_u64 start) in XXH3_mergeAccs()
4418 xxh_u64 result64 = start; in XXH3_mergeAccs()
4451 XXH_ALIGN(XXH_ACC_ALIGN) xxh_u64 acc[XXH_ACC_NB] = XXH3_INIT_ACC; in XXH3_hashLong_64b_internal()
4460 …return XXH3_mergeAccs(acc, (const xxh_u8*)secret + XXH_SECRET_MERGEACCS_START, (xxh_u64)len * XXH_… in XXH3_hashLong_64b_internal()
4754 XXH3_consumeStripes(xxh_u64* XXH_RESTRICT acc, in XXH3_consumeStripes()
4804 XXH_ALIGN(XXH_ACC_ALIGN) xxh_u64 acc[8]; memcpy(acc, state->acc, sizeof(acc)); in XXH3_update()
4806 xxh_u64* XXH_RESTRICT const acc = state->acc; in XXH3_update()
4952 (xxh_u64)state->totalLen * XXH_PRIME64_1); in XXH3_64bits_digest()
4998 xxh_u64 const bitflipl = (XXH_readLE32(secret) ^ XXH_readLE32(secret+4)) + seed; in XXH3_len_1to3_128b()
4999 xxh_u64 const bitfliph = (XXH_readLE32(secret+8) ^ XXH_readLE32(secret+12)) - seed; in XXH3_len_1to3_128b()
5000 xxh_u64 const keyed_lo = (xxh_u64)combinedl ^ bitflipl; in XXH3_len_1to3_128b()
5001 xxh_u64 const keyed_hi = (xxh_u64)combinedh ^ bitfliph; in XXH3_len_1to3_128b()
5015 seed ^= (xxh_u64)XXH_swap32((xxh_u32)seed) << 32; in XXH3_len_4to8_128b()
5018 xxh_u64 const input_64 = input_lo + ((xxh_u64)input_hi << 32); in XXH3_len_4to8_128b()
5019 xxh_u64 const bitflip = (XXH_readLE64(secret+16) ^ XXH_readLE64(secret+24)) + seed; in XXH3_len_4to8_128b()
5020 xxh_u64 const keyed = input_64 ^ bitflip; in XXH3_len_4to8_128b()
5042 { xxh_u64 const bitflipl = (XXH_readLE64(secret+32) ^ XXH_readLE64(secret+40)) - seed; in XXH3_len_9to16_128b()
5043 xxh_u64 const bitfliph = (XXH_readLE64(secret+48) ^ XXH_readLE64(secret+56)) + seed; in XXH3_len_9to16_128b()
5044 xxh_u64 const input_lo = XXH_readLE64(input); in XXH3_len_9to16_128b()
5045 xxh_u64 input_hi = XXH_readLE64(input + len - 8); in XXH3_len_9to16_128b()
5051 m128.low64 += (xxh_u64)(len - 1) << 54; in XXH3_len_9to16_128b()
5060 if (sizeof(void *) < sizeof(xxh_u64)) { /* 32-bit */ in XXH3_len_9to16_128b()
5119 xxh_u64 const bitflipl = XXH_readLE64(secret+64) ^ XXH_readLE64(secret+72); in XXH3_len_0to16_128b()
5120 xxh_u64 const bitfliph = XXH_readLE64(secret+80) ^ XXH_readLE64(secret+88); in XXH3_len_0to16_128b()
5230 XXH_ALIGN(XXH_ACC_ALIGN) xxh_u64 acc[XXH_ACC_NB] = XXH3_INIT_ACC; in XXH3_hashLong_128b_internal()
5240 (xxh_u64)len * XXH_PRIME64_1); in XXH3_hashLong_128b_internal()
5244 ~((xxh_u64)len * XXH_PRIME64_2)); in XXH3_hashLong_128b_internal()
5430 (xxh_u64)state->totalLen * XXH_PRIME64_1); in XXH3_128bits_digest()
5434 ~((xxh_u64)state->totalLen * XXH_PRIME64_2)); in XXH3_128bits_digest()