/openssl/crypto/poly1305/asm/ |
H A D | poly1305-mips.pl | 138 and $in0,$tmp0 152 li $tmp0,1 153 dsll $tmp0,32 155 dsll $tmp0,28 279 sltu $tmp0,$h1,$tmp0 294 sltu $tmp0,$d0,$tmp0 310 sltu $tmp0,$d1,$tmp0 320 and $tmp0,$d2 325 sltu $tmp0,$h0,$tmp0 327 sltu $tmp0,$h1,$tmp0 [all …]
|
/openssl/crypto/sha/asm/ |
H A D | sha1-sparcv9a.pl | 55 $tmp0="%i3"; 161 add $tmp0,$e,$e 180 add $tmp0,$e,$e 213 add $tmp0,$e,$e 215 xor $c,$b,$tmp0 231 add $tmp0,$e,$e 233 xor $c,$b,$tmp0 251 add $tmp0,$e,$e 396 and $tmp0,-8,$tmp0 469 mov 5,$tmp0 [all …]
|
H A D | sha512-mips.pl | 162 or @X[0],$tmp0 178 and $tmp2,$tmp0 184 and @X[0],$tmp0 200 xor $tmp0,$tmp1 210 $ADDU $T1,$tmp0 211 and $tmp0,$b,$c 223 xor $h,$tmp0 227 xor $h,$tmp0 239 xor $h,$tmp0 243 xor $h,$tmp0 [all …]
|
H A D | sha512-sparcv9.pl | 119 $tmp0="%g3"; 241 xor $tmp0,$h,$h 264 or $a,$b,$tmp0 266 and $c,$tmp0,$tmp0 352 or %l3,$tmp0,$tmp0 358 srlx $tmp0,@sigma0[1],$tmp0 362 srlx $tmp0,`@sigma0[2]-@sigma0[1]`,$tmp0 374 sllx $tmp0,`@sigma1[2]-@sigma1[1]`,$tmp0 381 or %l5,$tmp0,$tmp0 749 or %l1,$tmp0,$tmp0 [all …]
|
H A D | sha1-sparcv9.pl | 56 $tmp0="%i3"; 68 add $tmp0,$e,$e 69 and $c,$b,$tmp0 128 add $tmp0,$e,$e !! 129 and $c,$b,$tmp0 152 add $tmp0,$e,$e !! 153 xor $c,$b,$tmp0 156 xor $d,$tmp0,$tmp1 176 and $c,$b,$tmp0 300 andn $inp,7,$tmp0 [all …]
|
/openssl/crypto/ec/ |
H A D | ecp_sm2p256.c | 203 ecp_sm2p256_sqr(tmp0, P->Z); in ecp_sm2p256_point_double() 205 ecp_sm2p256_add(tmp0, P->X, tmp0); in ecp_sm2p256_point_double() 218 ecp_sm2p256_mul(tmp0, tmp0, tmp1); in ecp_sm2p256_point_double() 245 ecp_sm2p256_sqr(tmp0, P->Z); in ecp_sm2p256_point_add_affine() 247 ecp_sm2p256_mul(tmp0, tmp0, Q->X); in ecp_sm2p256_point_add_affine() 249 ecp_sm2p256_sub(tmp0, tmp0, P->X); in ecp_sm2p256_point_add_affine() 253 if (is_zeros(tmp0)) { in ecp_sm2p256_point_add_affine() 320 ecp_sm2p256_mul(tmp0, tmp0, Q->X); in ecp_sm2p256_point_add() 327 ecp_sm2p256_sub(tmp0, tmp0, R->X); in ecp_sm2p256_point_add() 331 ecp_sm2p256_mul(tmp0, tmp0, tmp2); in ecp_sm2p256_point_add() [all …]
|
H A D | curve25519.c | 236 fe64_sub(tmp0, x3, z3); in x25519_scalar_mulx() 240 fe64_mul(z3, x2, tmp0); in x25519_scalar_mulx() 242 fe64_sqr(tmp0, tmp1); in x25519_scalar_mulx() 246 fe64_mul(x2, tmp1, tmp0); in x25519_scalar_mulx() 251 fe64_add(tmp0, tmp0, z3); in x25519_scalar_mulx() 728 fe51_sub(tmp0, x3, z3); in x25519_scalar_mult() 732 fe51_mul(z3, tmp0, x2); in x25519_scalar_mult() 734 fe51_sq(tmp0, tmp1); in x25519_scalar_mult() 743 fe51_add(tmp0, tmp0, z3); in x25519_scalar_mult() 4402 fe_sq(tmp0, tmp1); in x25519_scalar_mult_generic() [all …]
|
H A D | ecp_nistp224.c | 494 limb tmp0, tmp1, tmp2; in felem_square() local 495 tmp0 = 2 * in[0]; in felem_square() 501 out[3] = ((widelimb) in[3]) * tmp0 + ((widelimb) in[1]) * tmp2; in felem_square()
|
H A D | ecp_nistz256.c | 262 BN_ULONG tmp0[P256_LIMBS]; in ecp_nistz256_point_double() local 291 ecp_nistz256_mul_by_2(tmp0, S); in ecp_nistz256_point_double() 295 ecp_nistz256_sub(res_x, res_x, tmp0); in ecp_nistz256_point_double()
|
/openssl/crypto/bn/asm/ |
H A D | sparcv9-mont.pl | 71 $tmp0="%g4"; 141 mulx $apj,$mul0,$tmp0 151 mov $tmp0,$acc0 170 add $tmp0,$car0,$car0 212 mulx $apj,$mul0,$tmp0 224 mov $tmp0,$acc0 345 mov $tmp0,$acc0 390 mulx $n0,$tmp0,$mul1 498 srlx $car1,32,$tmp0 566 add $i,4,$tmp0 [all …]
|
/openssl/crypto/aes/asm/ |
H A D | aesv8-armx.pl | 1554 veor $tmp0,$tmp0,$dat0 1669 veor $tmp0,$tmp0,$dat0 2357 aesmc $tmp0,$tmp0 2366 aesmc $tmp0,$tmp0 2375 aesmc $tmp0,$tmp0 2387 aesmc $tmp0,$tmp0 2802 aesmc $tmp0,$tmp0 2816 aesmc $tmp0,$tmp0 2826 aesmc $tmp0,$tmp0 2846 aesmc $tmp0,$tmp0 [all …]
|
H A D | aes-x86_64.pl | 252 my $tmp0=$acc0; 258 $tmp0=$s[1]; 288 my $tmp0=$acc0; 294 $tmp0=$s[1]; 310 $code.=" mov 0($sbox,$tmp0,8),$tmp0\n"; 318 $code.=" xor $tmp0,$out\n"; 845 my $tmp0=$acc0; 860 $tmp0=$s[1] if ($i==3); 877 my $tmp0=$acc0; 897 $code.=" movzb 2048($sbox,$tmp0,1),$tmp0\n"; [all …]
|
/openssl/crypto/ec/asm/ |
H A D | ecp_nistz256-x86.pl | 1234 { my ($S,$M,$Zsqr,$in_x,$tmp0)=map(32*$_,(0..4)); 1289 &lea ("edi",&DWP($tmp0,"esp")); 1298 &lea ("esi",&DWP($tmp0,"esp")); 1299 &lea ("ebp",&DWP($tmp0,"esp")); 1311 &lea ("edi",&DWP($tmp0,"esp")); 1321 &lea ("esi",&DWP($tmp0,"esp")); 1327 &lea ("edi",&DWP($tmp0,"esp")); 1336 &lea ("esi",&DWP($tmp0,"esp")); 1343 &lea ("edi",&DWP($tmp0,"esp")); 1353 &lea ("ebp",&DWP($tmp0,"esp"));
|
H A D | ecp_nistz256-armv4.pl | 1276 my ($S,$M,$Zsqr,$in_x,$tmp0)=map(32*$_,(0..4)); 1311 add $r_ptr,sp,#$tmp0 1330 add $r_ptr,sp,#$tmp0 1339 add $a_ptr,sp,#$tmp0 1352 add $r_ptr,sp,#$tmp0 1360 add $b_ptr,sp,#$tmp0
|
H A D | ecp_nistz256-armv8.pl | 717 my ($S,$M,$Zsqr,$tmp0)=map(32*$_,(0..3)); 779 add $rp,sp,#$tmp0 791 add $rp,sp,#$tmp0 833 add $rp,sp,#$tmp0 839 add $bp,sp,#$tmp0
|
H A D | ecp_nistz256-sparcv9.pl | 749 my ($S,$M,$Zsqr,$tmp0)=map(32*$_,(0..3)); 814 add %sp,LOCALS+$tmp0,$rp 830 add %sp,LOCALS+$tmp0,$rp 841 add %sp,LOCALS+$tmp0,$rp 848 add %sp,LOCALS+$tmp0,$bp 2032 $S,$M,$Zsqr,$tmp0)=map(32*$_,(0..9)); 2155 add %sp,LOCALS64+$tmp0,$rp 2188 add %sp,LOCALS64+$tmp0,$rp 2209 add %sp,LOCALS64+$tmp0,$rp 2214 add %sp,LOCALS64+$tmp0,$bp
|
H A D | ecp_nistz256-ppc64.pl | 848 my ($S,$M,$Zsqr,$tmp0)=map(64+32*$_,(0..3)); 929 addi $rp,$sp,$tmp0 943 addi $rp,$sp,$tmp0 991 addi $rp,$sp,$tmp0 997 addi $bp,$sp,$tmp0
|
H A D | ecp_nistz256-x86_64.pl | 3226 my ($S,$M,$Zsqr,$in_x,$tmp0)=map(32*$_,(0..4)); 3393 lea $tmp0(%rsp), $r_ptr 3404 lea $tmp0(%rsp), $r_ptr 3411 lea $tmp0(%rsp), $b_ptr
|
/openssl/crypto/chacha/asm/ |
H A D | chacha-armv8-sve.pl | 712 my ($tmpw0,$tmp0,$tmpw1,$tmp1) = ("w9","x9", "w10","x10");
|