/openssl/crypto/modes/asm/ |
H A D | ghash-x86_64.pl | 1094 vpxor $T2,$Xi,$Xi # 1110 vpxor $T1,$Xi,$Xi # 1115 vpxor $T2,$Xi,$Xi # 1117 vpxor $T2,$Xi,$Xi # 1357 vpxor $T2, $Xi, $Xi 1567 vpxor $T2, $Xi, $Xi 1571 vpalignr \$8,$Xi,$Xi,$Xi 1572 vpxor $T2,$Xi,$Xi 1575 vpalignr \$8,$Xi,$Xi,$Xi 1576 vpxor $Xo,$Xi,$Xi [all …]
|
H A D | ghash-x86.pl | 909 &psllq ($Xi,5); 911 &psllq ($Xi,1); 915 &pslldq ($Xi,8); 917 &pxor ($Xi,$T2); 922 &psrlq ($Xi,1); 924 &pxor ($T2,$Xi); 925 &psrlq ($Xi,5); 1077 &psllq ($Xi,5); 1152 my ($Xhi,$Xi)=@_; 1157 &pslld ($Xi,1); [all …]
|
H A D | aesni-gcm-x86_64.pl | 152 vpxor 16+8(%rsp),$Xi,$Xi # modulo-scheduled [vpxor $Z3,$Xi,$Xi] 240 vpclmulqdq \$0x11,$Hkey,$Xi,$Xi 465 vpshufb $Ii,$Xi,$Xi 724 vpshufb $Ii,$Xi,$Xi 796 vpunpckhqdq $Xi,$Xi,$T3 807 vpclmulqdq \$0x11,$Ii,$Xi,$Xi 852 vxorps $inout5,$Xi,$Xi 868 vxorps $inout5,$Xi,$Xi 873 vpunpckhqdq $Xi,$Xi,$T3 897 vpxor $T2,$Xi,$Xi [all …]
|
H A D | ghash-riscv64.pl | 156 ld $x0, 0($Xi) 157 ld $x1, 8($Xi) 196 sd $x0, 0($Xi) 197 sd $x1, 8($Xi) 213 ld $x0, 0($Xi) 214 ld $x1, 8($Xi) 253 sd $x0, 0($Xi) 254 sd $x1, 8($Xi) 281 ld $x0, 0($Xi) 406 sd $x0, 0($Xi) [all …]
|
H A D | ghash-s390x.pl | 96 aghi $Xi,-1 101 lg $Zlo,8+1($Xi) # Xi 121 lmg %r0,%r1,0($Xi) 130 stmg %r0,%r1,0($Xi) 141 aghi $Xi,-1 147 lg $Zhi,0+1($Xi) 153 stg $Zlo,8+1($Xi) 154 stg $Zhi,0+1($Xi) 190 llgc $xi,0($cnt,$Xi) 236 stg $Zlo,8+1($Xi) [all …]
|
H A D | ghash-armv4.pl | 126 str $_,[$Xi,#$i] 128 str $_,[$Xi,#$i] 131 strb $_,[$Xi,#$i+3] 133 strb $Tlh,[$Xi,#$i+2] 135 strb $Thl,[$Xi,#$i+1] 136 strb $Thh,[$Xi,#$i] 197 ldrb $nhi,[$Xi,#15] 215 ldrb $nhi,[$Xi,#14] 301 ldrb $nlo,[$Xi,#15] 310 ldrb $nlo,[$Xi,#14] [all …]
|
H A D | ghash-riscv64-zvkb-zvbc.pl | 124 my ($Xi,$Htable,$TMP0,$TMP1,$TMP2,$TMP3,$TMP4) = ("a0","a1","t0","t1","t2","t3","t4"); 141 add $Xi, $Xi, 8 146 @{[vlse64_v $V5, $Xi, $TMP4]} # vlse64.v v5, (a0), t4 229 @{[vsse64_v $V2, $Xi, $TMP4]} # vsse64.v v2, (a0), t4 245 my ($Xi,$Htable,$inp,$len,$TMP0,$TMP1,$TMP2,$TMP3,$M8,$TMP5,$TMP6) = ("a0","a1","a2","a3","t0","t1"… 261 add $Xi, $Xi, 8 267 @{[vlse64_v $V5, $Xi, $M8]} # vlse64.v v5, (a0), t4 362 @{[vsse64_v $V5, $Xi, $M8]} # vsse64.v v2, (a0), t4
|
H A D | ghash-riscv64-zvkg.pl | 113 my ($Xi,$Htable) = ("a0","a1"); 123 @{[vle32_v $VD, $Xi]} 125 @{[vse32_v $VD, $Xi]} 141 my ($Xi,$Htable,$inp,$len) = ("a0","a1","a2","a3"); 151 @{[vle32_v $vXi, $Xi]} 160 @{[vse32_v $vXi, $Xi]}
|
H A D | ghash-parisc.pl | 134 ldb 15($Xi),$nlo 146 ldb 14($Xi),$nlo 215 std $Zll,8($Xi) 216 std $Zhh,0($Xi) 224 ldb 15($Xi),$nlo 237 ldb 14($Xi),$nlo 325 stw $Zlh,8($Xi) 327 stw $Zhl,4($Xi) 328 stw $Zhh,0($Xi) 476 std $Zll,8($Xi) [all …]
|
H A D | ghash-sparcv9.pl | 109 ldub [$Xi+15],$xi0 110 ldub [$Xi+14],$xi1 159 ldub [$Xi+$cnt],$xi1 211 stx $Zlo,[$Xi+8] 213 stx $Zhi,[$Xi] 231 stx $Zlo,[$Xi+8] 233 stx $Zhi,[$Xi] 249 ldub [$Xi+15],$nlo 261 ldub [$Xi+14],$nlo 339 stx $Zlo,[$Xi+8] [all …]
|
H A D | ghash-alpha.pl | 41 $Xi="a0"; # $16, input argument block 273 ldq $Xlo,8($Xi) 274 ldq $Xhi,0($Xi) 321 stq $Xlo,8($Xi) 322 stq $Xhi,0($Xi) 348 ldq $Xhi,0($Xi) 349 ldq $Xlo,8($Xi) 434 stq $Xlo,8($Xi) 435 stq $Xhi,0($Xi)
|
H A D | ghash-ia64.pl | 116 $ADDP Xi=15,in0 // &Xi[15] 270 ld1 $xi[0]=[$Xip],-1 //(p16) *Xi-- 276 { .mmi; ld1 $xi[0]=[$Xip],-1 //(p16) *Xi-- 304 ld1 $xi[0]=[$Xip],-1 //(p16) *Xi-- 334 ld1 $xi[0]=[$Xip],-1 //(p16) *Xi-- 395 add $Xip=9,$Xip };; // &Xi.lo
|
/openssl/crypto/sha/asm/ |
H A D | sha1-mb-x86_64.pl | 817 vpunpckldq @Xi[2],@Xi[0],@Xi[0] 845 vinserti128 @Xi[2],@Xi[0],@Xi[0] 911 vpxor @Xi[-2],@Xi[1],@Xi[1] # "X[13]" 925 vpxor @Xi[3],@Xi[1],@Xi[1] 933 vpaddd @Xi[1],@Xi[1],@Xi[1] 949 vpxor @Xi[-2],@Xi[1],@Xi[1] # "X[13]" 964 vpxor @Xi[3],@Xi[1],@Xi[1] 970 vpaddd @Xi[1],@Xi[1],@Xi[1] 1002 vpxor @Xi[-2],@Xi[1],@Xi[1] # "X[13]" 1013 vpxor @Xi[3],@Xi[1],@Xi[1] [all …]
|
H A D | sha256-mb-x86_64.pl | 243 ($Xi,$Xn)=($Xn,$Xi); 809 vpunpckldq $t1,$Xi,$Xi 810 vpshufb $Xn,$Xi,$Xi 821 vpunpckldq $t1,$Xi,$Xi 822 vpshufb $Xn,$Xi,$Xi 832 vpunpckldq $t2,$Xi,$Xi 835 vinserti128 $t1,$Xi,$Xi 836 vpshufb $Xn,$Xi,$Xi 853 vpunpckldq $t2,$Xi,$Xi 858 vpshufb $Xn,$Xi,$Xi [all …]
|
H A D | sha1-armv4-large.pl | 108 $Xi="r14"; 226 mov $Xi,sp 239 teq $Xi,$t3 241 teq $Xi,sp 263 teq $Xi,$t3 282 teq $Xi,sp 334 my $Xi=4; 513 $Xi=0; 522 &vrev32_8 (@X[($Xi-3)&7],@X[($Xi-3)&7]); 525 &vadd_i32 (@X[$Xi&7],@X[($Xi-4)&7],$K); [all …]
|
H A D | sha1-thumb.pl | 46 $Xi="r12"; 139 mov $Xi,sp 176 mov $t0,$Xi 181 mov $Xi,$t1 187 cmp $Xi,$t0 193 mov $Xi,$t1 199 cmp $Xi,$t0 207 mov $Xi,$t1 213 cmp $Xi,$t0 217 mov $Xi,sp
|
H A D | sha1-sparcv9.pl | 40 $Xi="%g4"; 104 or $tmp1,$Xi,$Xi 107 srlx @X[$j%8],31,$Xi 109 and $Xi,$rot1m,$Xi 112 or $Xi,@X[$j%8],@X[$j%8] 124 $xi=$Xi; 148 $xi=$Xi; 171 $xi=$Xi; 330 srlx @X[$i+1],$tmp2,$Xi 332 or $Xi,@X[$i],@X[$i] [all …]
|
H A D | sha1-x86_64.pl | 479 my $Xi=4; 685 if ($Xi%5) { 769 $Xi=0; 799 $Xi++; 1211 $Xi=0; 1222 &vpshufb(@X[($Xi-3)&7],@X[($Xi-3)&7],@X[2]); 1225 &vpaddd (@X[$Xi&7],@X[($Xi-4)&7],$Kx); 1235 $Xi++; 1432 for (;$Xi<8;$Xi++) { # Xupdate_avx2_16_31 1594 $Xi++; [all …]
|
H A D | sha1-586.pl | 781 if ($Xi%5) { 872 $Xi=0; 893 &paddd (@X[($Xi-4)&7],@X[3]); 898 &movdqa (&QWP(0+16*$Xi,"esp"),@X[($Xi-4)&7]); # X[]+K xfer to IALU 906 $Xi++; 1296 if ($Xi%5) { 1376 $Xi=0; 1387 &vpshufb (@X[($Xi-3)&7],@X[($Xi-3)&7],@X[2]); 1390 &vpaddd (@X[$Xi&7],@X[($Xi-4)&7],@X[3]); 1395 &vmovdqa (&QWP(0+16*$Xi,"esp"),@X[$Xi&7]); # X[]+K xfer to IALU [all …]
|
H A D | sha1-sparcv9a.pl | 64 $Xi="%o7"; 166 add $Xi,$e,$e 186 add $Xi,$e,$e 223 add $Xi,$e,$e 243 add $Xi,$e,$e 260 add $Xi,$e,$e 274 add $Xi,$e,$e 301 add $Xi,$e,$e 323 add $Xi,$e,$e 353 add $Xi,$e,$e [all …]
|
/openssl/crypto/modes/ |
H A D | gcm128.c | 217 u8 *p = (u8 *)Xi; in gcm_gmult_4bit() 229 Xi[0] = Z.hi; in gcm_gmult_4bit() 230 Xi[1] = Z.lo; in gcm_gmult_4bit() 312 Xi[0] = Z.hi; in gcm_ghash_4bit() 703 ctx->Xi.u[0] = 0; in CRYPTO_gcm128_setiv() 704 ctx->Xi.u[1] = 0; in CRYPTO_gcm128_setiv() 798 memcpy(ctx->Xn, ctx->Xi.c, sizeof(ctx->Xi)); in CRYPTO_gcm128_encrypt() 1023 memcpy(ctx->Xn, ctx->Xi.c, sizeof(ctx->Xi)); in CRYPTO_gcm128_decrypt() 1258 memcpy(ctx->Xn, ctx->Xi.c, sizeof(ctx->Xi)); in CRYPTO_gcm128_encrypt_ctr32() 1412 memcpy(ctx->Xn, ctx->Xi.c, sizeof(ctx->Xi)); in CRYPTO_gcm128_decrypt_ctr32() [all …]
|
/openssl/providers/implementations/ciphers/ |
H A D | cipher_aes_gcm_hw_armv8.inc | 16 const void *key, unsigned char ivec[16], u64 *Xi) 26 … unroll8_eor3_aes_gcm_enc_128_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key); 28 aes_gcm_enc_128_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key); 33 … unroll8_eor3_aes_gcm_enc_192_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key); 35 aes_gcm_enc_192_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key); 40 … unroll8_eor3_aes_gcm_enc_256_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key); 42 aes_gcm_enc_256_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key); 50 const void *key, unsigned char ivec[16], u64 *Xi) 62 aes_gcm_dec_128_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key); 69 aes_gcm_dec_192_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key); [all …]
|
H A D | cipher_aes_gcm_hw_vaes_avx512.inc | 42 void ossl_gcm_gmult_avx512(u64 Xi[2], const void *gcm128ctx); 68 gcmctx->Xi.u[0] = 0; /* AAD hash */ 69 gcmctx->Xi.u[1] = 0; 112 gcmctx->Xi.c[15 - ares] ^= *(aad++); 118 ossl_gcm_gmult_avx512(gcmctx->Xi.u, gcmctx); 137 gcmctx->Xi.c[15 - i] ^= aad[i]; 159 ossl_gcm_gmult_avx512(gcmctx->Xi.u, gcmctx); 184 memcpy(tag, gcmctx->Xi.c, 185 ctx->taglen <= sizeof(gcmctx->Xi.c) ? ctx->taglen : 186 sizeof(gcmctx->Xi.c)); [all …]
|
/openssl/include/crypto/ |
H A D | aes_platform.h | 85 u64 *Xi); 88 u64 *Xi); 91 void gcm_ghash_p8(u64 Xi[2],const u128 Htable[16],const u8 *inp, size_t len); 155 unsigned char ivec[16], u64 *Xi); 157 unsigned char ivec[16], u64 *Xi); 158 void gcm_ghash_v8(u64 Xi[2],const u128 Htable[16],const u8 *inp, size_t len); 262 const void *key, unsigned char ivec[16], u64 *Xi); 264 const void *key, unsigned char ivec[16], u64 *Xi); 265 void gcm_ghash_avx(u64 Xi[2], const u128 Htable[16], const u8 *in, size_t len); 487 unsigned char ivec[16], u64 *Xi); [all …]
|
/openssl/crypto/aes/asm/ |
H A D | aesni-sha1-x86_64.pl | 163 my $Xi=4; 429 if ($Xi%5) { 513 $Xi=0; 543 $Xi++; 766 $Xi=4; 1057 my $Xi=4; 1294 &vmovdqa ($Kx,eval(16*($Xi/5))."($K_XX_XX)") if ($Xi%5==0); 1365 $Xi=0; 1376 &vpshufb(@X[($Xi-3)&7],@X[($Xi-3)&7],@Tx[1]); 1389 $Xi++; [all …]
|