/openssl/crypto/aes/asm/ |
H A D | aes-riscv32-zkn.pl | 271 xor $Q3,$Q3,$T3 532 srli $T4,$T3,8 533 slli $T5,$T3,24 544 xor $T3,$T3,$T2 598 xor $T3,$T3,$T2 671 xor $T3,$T3,$T2 823 srli $T4,$T3,8 824 slli $T5,$T3,24 832 xor $T3,$T3,$T2 911 xor $T3,$T3,$T2 [all …]
|
H A D | aes-riscv64.pl | 323 xor $T3,$T3,$T7 410 slli $T3,$T3,24+32 553 xor $T3,$T3,$T7 590 [$T0,$T1,$T2,$T3] 596 [$T0,$T1,$T2,$T3] 602 [$T0,$T1,$T2,$T3] 630 slli $T3,$T3,24+32 853 xor $T3,$T3,$T2 885 xor $T3,$T3,$T2 921 xor $T3,$T3,$T2 [all …]
|
H A D | aes-riscv64-zkn.pl | 356 ld $T3,24($UKEY) 360 sd $T3,24($KEYP) 377 @{[aes64ks2 $T3,$T2,$T3]} 495 @{[aes64im $T3,$T2]} 496 sd $T3,16($KEYP) 507 @{[aes64im $T3,$T0]} 508 sd $T3,0($KEYP) 509 @{[aes64im $T3,$T1]} 534 ld $T3,24($UKEY) 539 @{[aes64im $T4,$T3]} [all …]
|
H A D | aes-riscv64-zvkb-zvkned.pl | 69 my ($T0, $T1, $T2, $T3) = ("t0", "t1", "t2", "t3"); 125 li $T3, 10 131 beq $T0, $T3, ctr32_encrypt_blocks_128
|
/openssl/crypto/poly1305/asm/ |
H A D | poly1305-x86_64.pl | 829 vpsrlq \$30,$T3,$T3 1086 vpsrlq \$30,$T3,$T3 1139 vpaddq $H3,$T3,$T3 1768 vpermd $T3,$T0,$T3 1948 vpsrlq \$30,$T3,$T3 2995 vpsllq \$20,$T3,$T3 3152 vpxorq $T3,$T3,$T3 3165 vpsllq \$20,$T3,$T3 3228 vpsllq \$20,$T3,$T3 3531 vpsllq \$20,$T3,$T3 [all …]
|
H A D | poly1305-s390x.pl | 517 vperm ($I0,$T3,$T4,$bswaplo); 518 vperm ($I2,$T3,$T4,$bswapmi); 519 vperm ($T3,$T3,$T4,$bswaphi); 531 vperm ($T3,$T1,$T2,$bswaplo); 535 verimg ($I0,$T3,$mask26,0); 583 vperm ($T3,$T3,$T4,$bswaphi); 625 verimg ($I0,$T3,$mask26,0); 799 vperm ($I0,$T3,$T4,$bswaplo); 800 vperm ($I2,$T3,$T4,$bswapmi); 801 vperm ($T3,$T3,$T4,$bswaphi); [all …]
|
H A D | poly1305-ppc.pl | 1348 be?vperm $T3,$T3,$T3,$_4 1368 vpermdi $T3,$T3,$T4,0b11 1372 vsrd $T4,$T3,$_40 1373 vsrd $T3,$T3,$_14 1377 vand $T3,$T3,$mask26 1484 be?vperm $T3,$T3,$T3,$_4 1517 vpermdi $T3,$T3,$T4,0b11 1569 vsrd $T3,$T3,$_14 1585 vand $T3,$T3,$mask26 1816 be?vperm $T3,$T3,$T3,$_4 [all …]
|
/openssl/crypto/modes/asm/ |
H A D | ghash-x86.pl | 882 &pxor ($T3,$Hkey); 947 &pxor ($T3,$T3); # 986 &pshufb ($Xi,$T3); 992 &pshufb ($Xi,$T3); 1012 &pshufb ($Xi,$T3); 1024 &pshufb ($T1,$T3); 1025 &pshufb ($Xn,$T3); 1071 &pshufb ($Xn,$T3); 1138 &pshufb ($T1,$T3); 1145 &pshufb ($Xi,$T3); [all …]
|
H A D | ghash-x86_64.pl | 554 pxor $T3,$T3 # 585 movdqa $Xi,$T3 626 pshufb $T3,$Xi 637 movq %r11,$T3 # borrow $T3 638 pand $Xi,$T3 640 movq %rax,$T3 647 movdqa .Lbswap_mask(%rip),$T3 # reload $T3 654 pshufb $T3,$Xi 1053 vpxor $T3,$T3,$T3 # 1054 vpcmpgtd $T2,$T3,$T3 # broadcast carry bit [all …]
|
H A D | aesni-gcm-x86_64.pl | 762 vpunpckhqdq $T3,$T3,$Z1 775 vpclmulqdq \$0x11,$Hkey,$T3,$T3 777 vpxor $Z2,$T3,$T3 788 vpxor $T3,$T1,$T1 799 vpxor $Xi,$T3,$T3 810 vpclmulqdq \$0x10,$HK,$T3,$T3 813 vpxor $Z0,$T3,$Z2 833 vpxor $inout3,$T3,$T3 848 vpclmulqdq \$0x00,$HK,$T3,$T3 849 vpxor $T2,$T3,$T3 [all …]
|
H A D | aes-gcm-riscv64-zvkb-zvkg-zvkned.pl | 80 my ($T0, $T1, $T2, $T3) = ("t0", "t1", "t2", "t3"); 626 li $T3, 10 630 beq $T0, $T3, aes_gcm_enc_blocks_128 811 li $T3, 10 815 beq $T0, $T3, aes_gcm_dec_blocks_128
|
H A D | aes-gcm-avx512.pl | 1244 vpclmulqdq \$0x01,$HK,$GH,$T3 # ; $T3 = a1*b0 1246 vpxorq $T3,$GH,$GH 1250 vpxorq $T3,$T1,$T1 1255 vmovdqu64 POLY2(%rip),$T3 1257 vpclmulqdq \$0x01,$GH,$T3,$T2 1263 vpclmulqdq \$0x00,$GH,$T3,$T2 1265 vpclmulqdq \$0x10,$GH,$T3,$GH 1287 my $ZT3 = &ZWORD($T3); 1294 my $YT3 = &YWORD($T3); 1592 lea ($T3,$T2,8),$T3 [all …]
|
/openssl/crypto/aria/ |
H A D | aria.c | 345 #define ARIA_ADD_ROUND_KEY(RK, T0, T1, T2, T3) \ argument 350 (T3) ^= (RK)->u[3]; \ 354 #define ARIA_SBOX_LAYER1_WITH_PRE_DIFF(T0, T1, T2, T3) \ argument 371 (T3) = \ 372 S1[GET_U8_BE(T3, 0)] ^ \ 373 S2[GET_U8_BE(T3, 1)] ^ \ 374 X1[GET_U8_BE(T3, 2)] ^ \ 404 #define ARIA_DIFF_WORD(T0,T1,T2,T3) \ argument 407 (T2) ^= (T3); \ 410 (T3) ^= (T1); \ [all …]
|
/openssl/crypto/chacha/asm/ |
H A D | chacha-riscv64-v-zbb.pl | 90 my ( $T3 ) = ( "t6" ); 382 li $T3, 0x6b206574 383 @{[vadd_vx $V3, $V3, $T3]} 384 add $STATE3, $STATE3, $T3 388 lw $T3, 12($KEY) 393 @{[vadd_vx $V7, $V7, $T3]} 394 add $STATE7, $STATE7, $T3 418 li $T3, 64 424 lw $T3, 28($KEY) 431 @{[vadd_vx $V11, $V11, $T3]} [all …]
|
H A D | chacha-armv8.pl | 844 my ($T0,$T1,$T2,$T3,$T4,$T5)=@K; 985 my @thread3=&NEONROUND($A3,$B3,$C3,$D3,$T3,0); 1004 @thread3=&NEONROUND($A3,$B3,$C3,$D3,$T3,1); 1099 @thread3=&NEONROUND($A3,$B3,$C3,$D3,$T3,0); 1116 @thread3=&NEONROUND($A3,$B3,$C3,$D3,$T3,1); 1202 ld1.8 {$T0-$T3},[$inp],#64 1214 eor $D0,$D0,$T3 1215 ld1.8 {$T0-$T3},[$inp],#64 1229 eor $D1,$D1,$T3
|
/openssl/crypto/sha/asm/ |
H A D | sha512-armv8.pl | 478 &ext_8 ($T3,@X[2],@X[3],4); # X[9..12] 488 &add_32 (@X[0],@X[0],$T3); # X[0..3] += X[9..12] 493 &ushr_32 ($T3,$T0,$sigma0[1]); 499 &sli_32 ($T3,$T0,32-$sigma0[1]); 505 &eor_8 ($T1,$T1,$T3); # sigma0(X[1..4]) 514 &ushr_32 ($T3,$T7,$sigma1[1]); 520 &sli_u32 ($T3,$T7,32-$sigma1[1]); 527 &eor_8 ($T5,$T5,$T3); # sigma1(X[14..15]) 654 ld1.32 {$T3},[$Ktbl], #16 664 add.32 $T3,$T3,@X[3] [all …]
|
H A D | sha256-armv4.pl | 312 my ($T0,$T1,$T2,$T3,$T4,$T5)=("q8","q9","q10","q11","d24","d25"); 352 &vshr_u32 ($T3,$T0,$sigma0[1]); 358 &vsli_32 ($T3,$T0,32-$sigma0[1]); 364 &veor ($T1,$T1,$T3); # sigma0(X[1..4]) 507 vld1.32 {$T3},[$Ktbl,:128]! 522 vadd.i32 $T3,$T3,@X[3] 524 vst1.32 {$T3},[$Xfer,:128]!
|
/openssl/crypto/bn/asm/ |
H A D | rsaz-3k-avx512.pl | 121 my ($T0,$T0h,$T1,$T1h,$T2,$T2h,$T3,$T3h) = ("$zero", "$Bi", "$Yi", map("%ymm$_", (19..23))); 224 vpsrlq \$52, $_R3, $T3 228 valignq \$3, $T3, $T3h, $T3h 229 valignq \$3, $T2h, $T3, $T3 254 vpaddq $T3, $_R3, $_R3
|
H A D | rsaz-4k-avx512.pl | 118 my ($T0,$T0h,$T1,$T1h,$T2,$T2h,$T3,$T3h,$T4,$T4h) = ("$zero", "$Bi", "$Yi", map("%ymm$_", (23..29))… 231 vpsrlq \$52, $_R3, $T3 239 valignq \$3, $T3, $T3h, $T3h 240 valignq \$3, $T2h, $T3, $T3 267 vpaddq $T3, $_R3, $_R3
|
/openssl/test/ |
H A D | modes_internal_test.c | 328 static const u8 T3[] = { variable
|