Lines Matching refs:x3
90 __m128i x3, x4; in crc32_pclmul_batch() local
96 x3 = _mm_loadu_si128((__m128i *)(p + 0x20)); in crc32_pclmul_batch()
97 x3 = _mm_shuffle_epi8(x3, shuf_mask); /* endianness swap */ in crc32_pclmul_batch()
108 x7 = _mm_clmulepi64_si128(x3, k, 0x00); in crc32_pclmul_batch()
112 x3 = _mm_clmulepi64_si128(x3, k, 0x11); in crc32_pclmul_batch()
124 x3 = _mm_xor_si128(x3, x7); in crc32_pclmul_batch()
128 x3 = _mm_xor_si128(x3, x11); in crc32_pclmul_batch()
146 x0 = _mm_xor_si128(x0, x3); in crc32_pclmul_batch()
202 __m128i x3, x4; in crc32_pclmul_reflected_batch() local
206 x3 = _mm_loadu_si128((__m128i *)(p + 0x20)); in crc32_pclmul_reflected_batch()
217 x7 = _mm_clmulepi64_si128(x3, k, 0x00); in crc32_pclmul_reflected_batch()
221 x3 = _mm_clmulepi64_si128(x3, k, 0x11); in crc32_pclmul_reflected_batch()
229 x3 = _mm_xor_si128(x3, x7); in crc32_pclmul_reflected_batch()
233 x3 = _mm_xor_si128(x3, x11); in crc32_pclmul_reflected_batch()
251 x0 = _mm_xor_si128(x0, x3); in crc32_pclmul_reflected_batch()