Home
last modified time | relevance | path

Searched refs:Xi (Results 1 – 25 of 33) sorted by relevance

12

/openssl/crypto/modes/asm/
H A Dghash-x86_64.pl1094 vpxor $T2,$Xi,$Xi #
1110 vpxor $T1,$Xi,$Xi #
1115 vpxor $T2,$Xi,$Xi #
1117 vpxor $T2,$Xi,$Xi #
1357 vpxor $T2, $Xi, $Xi
1567 vpxor $T2, $Xi, $Xi
1571 vpalignr \$8,$Xi,$Xi,$Xi
1572 vpxor $T2,$Xi,$Xi
1575 vpalignr \$8,$Xi,$Xi,$Xi
1576 vpxor $Xo,$Xi,$Xi
[all …]
H A Dghash-x86.pl909 &psllq ($Xi,5);
911 &psllq ($Xi,1);
915 &pslldq ($Xi,8);
917 &pxor ($Xi,$T2);
922 &psrlq ($Xi,1);
924 &pxor ($T2,$Xi);
925 &psrlq ($Xi,5);
1077 &psllq ($Xi,5);
1152 my ($Xhi,$Xi)=@_;
1157 &pslld ($Xi,1);
[all …]
H A Daesni-gcm-x86_64.pl152 vpxor 16+8(%rsp),$Xi,$Xi # modulo-scheduled [vpxor $Z3,$Xi,$Xi]
240 vpclmulqdq \$0x11,$Hkey,$Xi,$Xi
465 vpshufb $Ii,$Xi,$Xi
724 vpshufb $Ii,$Xi,$Xi
796 vpunpckhqdq $Xi,$Xi,$T3
807 vpclmulqdq \$0x11,$Ii,$Xi,$Xi
852 vxorps $inout5,$Xi,$Xi
868 vxorps $inout5,$Xi,$Xi
873 vpunpckhqdq $Xi,$Xi,$T3
897 vpxor $T2,$Xi,$Xi
[all …]
H A Dghash-riscv64.pl156 ld $x0, 0($Xi)
157 ld $x1, 8($Xi)
196 sd $x0, 0($Xi)
197 sd $x1, 8($Xi)
213 ld $x0, 0($Xi)
214 ld $x1, 8($Xi)
253 sd $x0, 0($Xi)
254 sd $x1, 8($Xi)
281 ld $x0, 0($Xi)
406 sd $x0, 0($Xi)
[all …]
H A Dghash-s390x.pl96 aghi $Xi,-1
101 lg $Zlo,8+1($Xi) # Xi
121 lmg %r0,%r1,0($Xi)
130 stmg %r0,%r1,0($Xi)
141 aghi $Xi,-1
147 lg $Zhi,0+1($Xi)
153 stg $Zlo,8+1($Xi)
154 stg $Zhi,0+1($Xi)
190 llgc $xi,0($cnt,$Xi)
236 stg $Zlo,8+1($Xi)
[all …]
H A Dghash-armv4.pl126 str $_,[$Xi,#$i]
128 str $_,[$Xi,#$i]
131 strb $_,[$Xi,#$i+3]
133 strb $Tlh,[$Xi,#$i+2]
135 strb $Thl,[$Xi,#$i+1]
136 strb $Thh,[$Xi,#$i]
197 ldrb $nhi,[$Xi,#15]
215 ldrb $nhi,[$Xi,#14]
301 ldrb $nlo,[$Xi,#15]
310 ldrb $nlo,[$Xi,#14]
[all …]
H A Dghash-riscv64-zvkb-zvbc.pl124 my ($Xi,$Htable,$TMP0,$TMP1,$TMP2,$TMP3,$TMP4) = ("a0","a1","t0","t1","t2","t3","t4");
141 add $Xi, $Xi, 8
146 @{[vlse64_v $V5, $Xi, $TMP4]} # vlse64.v v5, (a0), t4
229 @{[vsse64_v $V2, $Xi, $TMP4]} # vsse64.v v2, (a0), t4
245 my ($Xi,$Htable,$inp,$len,$TMP0,$TMP1,$TMP2,$TMP3,$M8,$TMP5,$TMP6) = ("a0","a1","a2","a3","t0","t1"…
261 add $Xi, $Xi, 8
267 @{[vlse64_v $V5, $Xi, $M8]} # vlse64.v v5, (a0), t4
362 @{[vsse64_v $V5, $Xi, $M8]} # vsse64.v v2, (a0), t4
H A Dghash-riscv64-zvkg.pl113 my ($Xi,$Htable) = ("a0","a1");
123 @{[vle32_v $VD, $Xi]}
125 @{[vse32_v $VD, $Xi]}
141 my ($Xi,$Htable,$inp,$len) = ("a0","a1","a2","a3");
151 @{[vle32_v $vXi, $Xi]}
160 @{[vse32_v $vXi, $Xi]}
H A Dghash-parisc.pl134 ldb 15($Xi),$nlo
146 ldb 14($Xi),$nlo
215 std $Zll,8($Xi)
216 std $Zhh,0($Xi)
224 ldb 15($Xi),$nlo
237 ldb 14($Xi),$nlo
325 stw $Zlh,8($Xi)
327 stw $Zhl,4($Xi)
328 stw $Zhh,0($Xi)
476 std $Zll,8($Xi)
[all …]
H A Dghash-sparcv9.pl109 ldub [$Xi+15],$xi0
110 ldub [$Xi+14],$xi1
159 ldub [$Xi+$cnt],$xi1
211 stx $Zlo,[$Xi+8]
213 stx $Zhi,[$Xi]
231 stx $Zlo,[$Xi+8]
233 stx $Zhi,[$Xi]
249 ldub [$Xi+15],$nlo
261 ldub [$Xi+14],$nlo
339 stx $Zlo,[$Xi+8]
[all …]
H A Dghash-alpha.pl41 $Xi="a0"; # $16, input argument block
273 ldq $Xlo,8($Xi)
274 ldq $Xhi,0($Xi)
321 stq $Xlo,8($Xi)
322 stq $Xhi,0($Xi)
348 ldq $Xhi,0($Xi)
349 ldq $Xlo,8($Xi)
434 stq $Xlo,8($Xi)
435 stq $Xhi,0($Xi)
H A Dghash-ia64.pl116 $ADDP Xi=15,in0 // &Xi[15]
270 ld1 $xi[0]=[$Xip],-1 //(p16) *Xi--
276 { .mmi; ld1 $xi[0]=[$Xip],-1 //(p16) *Xi--
304 ld1 $xi[0]=[$Xip],-1 //(p16) *Xi--
334 ld1 $xi[0]=[$Xip],-1 //(p16) *Xi--
395 add $Xip=9,$Xip };; // &Xi.lo
/openssl/crypto/sha/asm/
H A Dsha1-mb-x86_64.pl817 vpunpckldq @Xi[2],@Xi[0],@Xi[0]
845 vinserti128 @Xi[2],@Xi[0],@Xi[0]
911 vpxor @Xi[-2],@Xi[1],@Xi[1] # "X[13]"
925 vpxor @Xi[3],@Xi[1],@Xi[1]
933 vpaddd @Xi[1],@Xi[1],@Xi[1]
949 vpxor @Xi[-2],@Xi[1],@Xi[1] # "X[13]"
964 vpxor @Xi[3],@Xi[1],@Xi[1]
970 vpaddd @Xi[1],@Xi[1],@Xi[1]
1002 vpxor @Xi[-2],@Xi[1],@Xi[1] # "X[13]"
1013 vpxor @Xi[3],@Xi[1],@Xi[1]
[all …]
H A Dsha256-mb-x86_64.pl243 ($Xi,$Xn)=($Xn,$Xi);
809 vpunpckldq $t1,$Xi,$Xi
810 vpshufb $Xn,$Xi,$Xi
821 vpunpckldq $t1,$Xi,$Xi
822 vpshufb $Xn,$Xi,$Xi
832 vpunpckldq $t2,$Xi,$Xi
835 vinserti128 $t1,$Xi,$Xi
836 vpshufb $Xn,$Xi,$Xi
853 vpunpckldq $t2,$Xi,$Xi
858 vpshufb $Xn,$Xi,$Xi
[all …]
H A Dsha1-armv4-large.pl108 $Xi="r14";
226 mov $Xi,sp
239 teq $Xi,$t3
241 teq $Xi,sp
263 teq $Xi,$t3
282 teq $Xi,sp
334 my $Xi=4;
513 $Xi=0;
522 &vrev32_8 (@X[($Xi-3)&7],@X[($Xi-3)&7]);
525 &vadd_i32 (@X[$Xi&7],@X[($Xi-4)&7],$K);
[all …]
H A Dsha1-thumb.pl46 $Xi="r12";
139 mov $Xi,sp
176 mov $t0,$Xi
181 mov $Xi,$t1
187 cmp $Xi,$t0
193 mov $Xi,$t1
199 cmp $Xi,$t0
207 mov $Xi,$t1
213 cmp $Xi,$t0
217 mov $Xi,sp
H A Dsha1-sparcv9.pl40 $Xi="%g4";
104 or $tmp1,$Xi,$Xi
107 srlx @X[$j%8],31,$Xi
109 and $Xi,$rot1m,$Xi
112 or $Xi,@X[$j%8],@X[$j%8]
124 $xi=$Xi;
148 $xi=$Xi;
171 $xi=$Xi;
330 srlx @X[$i+1],$tmp2,$Xi
332 or $Xi,@X[$i],@X[$i]
[all …]
H A Dsha1-x86_64.pl479 my $Xi=4;
685 if ($Xi%5) {
769 $Xi=0;
799 $Xi++;
1211 $Xi=0;
1222 &vpshufb(@X[($Xi-3)&7],@X[($Xi-3)&7],@X[2]);
1225 &vpaddd (@X[$Xi&7],@X[($Xi-4)&7],$Kx);
1235 $Xi++;
1432 for (;$Xi<8;$Xi++) { # Xupdate_avx2_16_31
1594 $Xi++;
[all …]
H A Dsha1-586.pl781 if ($Xi%5) {
872 $Xi=0;
893 &paddd (@X[($Xi-4)&7],@X[3]);
898 &movdqa (&QWP(0+16*$Xi,"esp"),@X[($Xi-4)&7]); # X[]+K xfer to IALU
906 $Xi++;
1296 if ($Xi%5) {
1376 $Xi=0;
1387 &vpshufb (@X[($Xi-3)&7],@X[($Xi-3)&7],@X[2]);
1390 &vpaddd (@X[$Xi&7],@X[($Xi-4)&7],@X[3]);
1395 &vmovdqa (&QWP(0+16*$Xi,"esp"),@X[$Xi&7]); # X[]+K xfer to IALU
[all …]
H A Dsha1-sparcv9a.pl64 $Xi="%o7";
166 add $Xi,$e,$e
186 add $Xi,$e,$e
223 add $Xi,$e,$e
243 add $Xi,$e,$e
260 add $Xi,$e,$e
274 add $Xi,$e,$e
301 add $Xi,$e,$e
323 add $Xi,$e,$e
353 add $Xi,$e,$e
[all …]
/openssl/crypto/modes/
H A Dgcm128.c217 u8 *p = (u8 *)Xi; in gcm_gmult_4bit()
229 Xi[0] = Z.hi; in gcm_gmult_4bit()
230 Xi[1] = Z.lo; in gcm_gmult_4bit()
312 Xi[0] = Z.hi; in gcm_ghash_4bit()
703 ctx->Xi.u[0] = 0; in CRYPTO_gcm128_setiv()
704 ctx->Xi.u[1] = 0; in CRYPTO_gcm128_setiv()
798 memcpy(ctx->Xn, ctx->Xi.c, sizeof(ctx->Xi)); in CRYPTO_gcm128_encrypt()
1023 memcpy(ctx->Xn, ctx->Xi.c, sizeof(ctx->Xi)); in CRYPTO_gcm128_decrypt()
1258 memcpy(ctx->Xn, ctx->Xi.c, sizeof(ctx->Xi)); in CRYPTO_gcm128_encrypt_ctr32()
1412 memcpy(ctx->Xn, ctx->Xi.c, sizeof(ctx->Xi)); in CRYPTO_gcm128_decrypt_ctr32()
[all …]
/openssl/providers/implementations/ciphers/
H A Dcipher_aes_gcm_hw_armv8.inc16 const void *key, unsigned char ivec[16], u64 *Xi)
26 … unroll8_eor3_aes_gcm_enc_128_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
28 aes_gcm_enc_128_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
33 … unroll8_eor3_aes_gcm_enc_192_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
35 aes_gcm_enc_192_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
40 … unroll8_eor3_aes_gcm_enc_256_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
42 aes_gcm_enc_256_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
50 const void *key, unsigned char ivec[16], u64 *Xi)
62 aes_gcm_dec_128_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
69 aes_gcm_dec_192_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
[all …]
H A Dcipher_aes_gcm_hw_vaes_avx512.inc42 void ossl_gcm_gmult_avx512(u64 Xi[2], const void *gcm128ctx);
68 gcmctx->Xi.u[0] = 0; /* AAD hash */
69 gcmctx->Xi.u[1] = 0;
112 gcmctx->Xi.c[15 - ares] ^= *(aad++);
118 ossl_gcm_gmult_avx512(gcmctx->Xi.u, gcmctx);
137 gcmctx->Xi.c[15 - i] ^= aad[i];
159 ossl_gcm_gmult_avx512(gcmctx->Xi.u, gcmctx);
184 memcpy(tag, gcmctx->Xi.c,
185 ctx->taglen <= sizeof(gcmctx->Xi.c) ? ctx->taglen :
186 sizeof(gcmctx->Xi.c));
[all …]
/openssl/include/crypto/
H A Daes_platform.h85 u64 *Xi);
88 u64 *Xi);
91 void gcm_ghash_p8(u64 Xi[2],const u128 Htable[16],const u8 *inp, size_t len);
155 unsigned char ivec[16], u64 *Xi);
157 unsigned char ivec[16], u64 *Xi);
158 void gcm_ghash_v8(u64 Xi[2],const u128 Htable[16],const u8 *inp, size_t len);
262 const void *key, unsigned char ivec[16], u64 *Xi);
264 const void *key, unsigned char ivec[16], u64 *Xi);
265 void gcm_ghash_avx(u64 Xi[2], const u128 Htable[16], const u8 *in, size_t len);
487 unsigned char ivec[16], u64 *Xi);
[all …]
/openssl/crypto/aes/asm/
H A Daesni-sha1-x86_64.pl163 my $Xi=4;
429 if ($Xi%5) {
513 $Xi=0;
543 $Xi++;
766 $Xi=4;
1057 my $Xi=4;
1294 &vmovdqa ($Kx,eval(16*($Xi/5))."($K_XX_XX)") if ($Xi%5==0);
1365 $Xi=0;
1376 &vpshufb(@X[($Xi-3)&7],@X[($Xi-3)&7],@Tx[1]);
1389 $Xi++;
[all …]

Completed in 92 milliseconds

12