/openssl/crypto/ec/asm/ |
H A D | x25519-ppc64.pl | 233 mulhdu $a0,$a0,$a0 341 addc $a0,$a0,$ap 350 add $a0,$a0,$t0 373 addc $a0,$a0,$t0 382 addc $a0,$a0,$t1 391 add $a0,$a0,$t0 414 subfc $a0,$t0,$a0 424 subfc $a0,$t1,$a0 433 subf $a0,$t0,$a0 459 addc $a0,$a0,$t0 [all …]
|
H A D | ecp_nistz256-armv8.pl | 131 ldp $a0,$a1,[$ap] 156 ldp $a0,$a1,[$ap] 182 ldp $a0,$a1,[$ap] 305 mov $a0,$acc0 312 mov $t0,$a0 484 // to $a0-$a3 521 mul $acc0,$a0,$a0 // a[0]*a[0] 525 umulh $a0,$a0,$a0 1182 mov $a0,$acc0 1546 mul $acc0,$a0,$a0 // a[0]*a[0] [all …]
|
H A D | ecp_nistz256-armv4.pl | 219 adds $a0,$a0,$t0 242 subs $a0,$a0,#-1 304 adds $a0,$a0,$a0 @ a[0:7]+=a[0:7] 469 subs $a0,$a0,$t0 535 subs $a0,$ff,$a0 1215 adds $a0,$a0,$a0 @ a[0:7]+=a[0:7] 1491 orr $a0,$a0,$a2 1493 orr $a0,$a0,$a7 1495 orr $a0,$a0,$a4 1514 orr $a0,$a0,$a2 [all …]
|
H A D | ecp_nistz256-ppc64.pl | 121 ld $a0,0($ap) 562 # to $a0-$a3 616 mulld $acc0,$a0,$a0 # a[0]*a[0] 617 mulhdu $a0,$a0,$a0 1189 xor $a0,$a0,$a0 1322 and $a0,$a0,$in1infty 1366 and $a0,$a0,$in1infty 1629 and $a0,$a0,$in1infty 1682 and $a0,$a0,$in1infty 2008 mulld $acc0,$a0,$a0 # a[0]*a[0] [all …]
|
H A D | ecp_nistz256-sparcv9.pl | 1942 umulxhi $a0,$a0,$a0 2060 or $a0,$t0,$a0 2100 or $a0,$t0,$a0 2328 or $a0,$t0,$a0 2397 or $a0,$t0,$a0 2738 or $a0,$t0,$a0 2768 or $a1,$a0,$a0 2772 or $a2,$a0,$a0 2774 or $acc0,$a0,$a0 2789 or $a0,$t0,$a0 [all …]
|
H A D | ecp_nistz256-x86_64.pl | 138 add $a0, $a0 # a0:a3+a0:a3 143 mov $a0, $t0 193 mov $a0, $t0 216 shr \$1, $a0 220 or $t0, $a0 263 add $a0, $a0 # a0:a3+a0:a3 267 mov $a0, $t0 273 sub \$-1, $a0 448 xor $a0, $a0 3193 add $a0, $a0 # a0:a3+a0:a3 [all …]
|
/openssl/crypto/sha/asm/ |
H A D | sha512-ppc.pl | 140 xor $a0,$a0,$a1 146 xor $a0,$a0,$a1 ; Sigma1(e) 154 xor $a0,$a0,$a1 158 xor $a0,$a0,$a1 ; Sigma0(a) 180 xor $a0,$a0,$a1 185 xor $a0,$a0,$a1 ; sigma0(X[(i+1)&0x0f]) 485 and $a0,$a0,$elo 494 xor $a0,$a0,$glo ; Ch(e,f,g) 586 xor $a0,$a0,$s0 642 ($a0,$a1,$a2,$a3) = ($a2,$a3,$a0,$a1); [all …]
|
H A D | sha512-parisc.pl | 98 $a0 ="%r26"; 117 xor $a1,$a0,$a0 129 xor $a1,$a0,$a0 152 xor $a1,$a0,$a0 156 xor $a1,$a0,$a0 ; sigma0(X[(i+1)&0x0f]) 377 $a0 ="%r17"; 418 xor $a2,$a0,$a0 442 xor $a2,$a0,$a0 450 xor $a2,$a0,$a0 494 xor $t2,$a0,$a0 [all …]
|
H A D | sha512-x86_64.pl | 196 xor $e,$a0 209 xor $e,$a0 243 mov $a0,$T1 249 xor $T1,$a0 251 ror \$$sigma0[0],$a0 261 mov $e,$a0 867 mov $E,$a0 1231 mov $E,$a0 1393 mov $E,$a0 1627 mov $E,$a0 [all …]
|
/openssl/crypto/bn/asm/ |
H A D | mips.pl | 145 move $a0,$v0 293 move $a0,$v0 304 move $a0,$v0 421 move $a0,$v0 432 move $a0,$v0 537 move $a0,$v0 549 move $a0,$v0 668 move $a0,$v0 800 move $a0,$v0 929 and $t2,$a0 [all …]
|
H A D | ppc-mont.pl | 396 mr $a0,$sp 843 andc $a0,$a0,$carry 854 or $acc0,$t0,$a0 875 andc $a0,$a0,$carry 886 or $acc0,$t0,$a0 912 addc $a0,$a0,$m0 987 mr $a0,$sp 1240 mr $n0,$a0 1845 andc $a0,$a0,$carry 1880 andc $a0,$a0,$carry [all …]
|
H A D | armv8-mont.pl | 762 mul $t1,$a2,$a0 763 mul $t2,$a3,$a0 764 mul $t3,$a4,$a0 766 mul $t0,$a5,$a0 768 mul $t1,$a6,$a0 770 mul $t2,$a7,$a0 895 mov $n0,$a0 940 mul $t0,$a0,$n0 1392 subs $a0,$acc0,$a0 1412 csel $a0,$acc0,$a0,lo [all …]
|
H A D | armv4-gf2m.pl | 77 ($a0,$a1,$a2,$a12,$a4,$a14)= 86 mov $a0,#0 88 str $a0,[sp,#0] @ tab[0]=0 97 eor $a0,$a2,$a4 @ a2^a4 100 str $a0,[sp,#24] @ tab[6]=a2^a4
|
H A D | c64xplus-gf2m.pl | 28 ($rp,$a1,$a0,$b1,$b0)=("A4","B4","A6","B6","A8"); # argument vector 129 &mul_1x1_upper($a0,$b0); # a0·b0 137 XOR $a0,$a1,$A
|
/openssl/crypto/chacha/asm/ |
H A D | chacha-armv4.pl | 99 "&add (@x[$a0],@x[$a0],@x[$b0])", 113 "&add (@x[$a0],@x[$a0],@x[$b0])", 704 vmov $a1,$a0 706 vmov $a2,$a0 716 vmov $a1,$a0 718 vmov $a2,$a0 783 vadd.i32 $a0,$a0,$t0 @ accumulate key material 809 veor $a0,$a0,$t0 @ xor with input 1010 veor $a0,$a0,$t0 1030 veor $a0,$a0,$t0 [all …]
|
H A D | chacha-s390x.pl | 160 vaf (@x[$a0],@x[$a0],@x[$b0]); # Q1 186 vaf (@x[$a0],@x[$a0],@x[$b0]); 771 vlr ($a0,@K[0]); 819 vaf ($a0,$a0,@K[0]); 826 vperm ($a0,$a0,$a0,$beperm); 838 vx ($a0,$a0,$t0); 865 vx ($a0,$a0,$a1); 891 vx ($a0,$a0,$a1); 919 vx ($a0,$a0,$a1); 948 vx ($a0,$a0,$a1); [all …]
|
H A D | chacha-c64xplus.pl | 129 || ADD @X[$b0],@X[$a0],@X[$a0] 164 ADD @Y[$b0],@Y[$a0],@Y[$a0] 173 || ADD @X[$b0],@X[$a0],@X[$a0] 219 || ADD @X[$b0],@X[$a0],@X[$a0] 230 || ADD @Y[$b0],@Y[$a0],@Y[$a0] 276 || ADD @Y[$b0],@Y[$a0],@Y[$a0] 281 || ADD @X[$b0],@X[$a0],@X[$a0] 598 ADD @X[$b0],@X[$a0],@X[$a0] 624 ADD @X[$b0],@X[$a0],@X[$a0] 654 ADD @X[$b0],@X[$a0],@X[$a0] [all …]
|
H A D | chacha-ia64.pl | 101 my ($a0,$b0,$c0,$d0)=@_; 102 my ($a1,$b1,$c1,$d1)=map(($_&~3)+(($_+1)&3),($a0,$b0,$c0,$d0)); 107 { .mmi; add @x[$a0]=@x[$a0],@x[$b0] 111 xor @x[$d0]=@x[$d0],@x[$a0] 136 { .mii; add @x[$a0]=@x[$a0],@x[$b0] 147 xor @x[$d0]=@x[$d0],@x[$a0] };;
|
H A D | chacha-loongarch64.pl | 67 # $a0 = arg #1 (out pointer) 108 my ($a0,$b0,$c0,$d0) = @_; 114 add.w @x[$a0],@x[$a0],@x[$b0] 115 xor @x[$d0],@x[$d0],@x[$a0] 128 add.w @x[$a0],@x[$a0],@x[$b0] 397 my ($a0,$b0,$c0,$d0) = @_; 403 vadd.w @x[$a0],@x[$a0],@x[$b0] 417 vadd.w @x[$a0],@x[$a0],@x[$b0] 821 my ($a0,$b0,$c0,$d0) = @_; 827 xvadd.w @x[$a0],@x[$a0],@x[$b0] [all …]
|
H A D | chacha-armv8.pl | 74 my ($a0,$b0,$c0,$d0)=@_; 75 my ($a1,$b1,$c1,$d1)=map(($_&~3)+(($_+1)&3),($a0,$b0,$c0,$d0)); 80 "&add_32 (@x[$a0],@x[$a0],@x[$b0])", 84 "&eor_32 (@x[$d0],@x[$d0],@x[$a0])", 106 "&add_32 (@x[$a0],@x[$a0],@x[$b0])", 110 "&eor_32 (@x[$d0],@x[$d0],@x[$a0])", 394 my ($a0,$b0,$c0,$d0)=@_; 401 "&add (@x[$a0],@x[$a0],@x[$b0])", # Q1 405 "&eor (@x[$d0],@x[$d0],@x[$a0])", 431 "&add (@x[$a0],@x[$a0],@x[$b0])", [all …]
|
H A D | chacha-armv8-sve.pl | 177 my ($a0,$b0,$c0,$d0,$a1,$b1,$c1,$d1,$a2,$b2,$c2,$d2,$a3,$b3,$c3,$d3) = @_; 179 &SVE_ADD($a0,$b0,$a1,$b1,$a2,$b2,$a3,$b3); 180 &SVE2_XAR(16,$d0,$a0,$d1,$a1,$d2,$a2,$d3,$a3); 185 &SVE_ADD($a0,$b0,$a1,$b1,$a2,$b2,$a3,$b3); 186 &SVE2_XAR(8,$d0,$a0,$d1,$a1,$d2,$a2,$d3,$a3); 193 my ($a0,$b0,$c0,$d0,$a1,$b1,$c1,$d1,$a2,$b2,$c2,$d2,$a3,$b3,$c3,$d3) = @_; 195 &SVE_ADD($a0,$b0,$a1,$b1,$a2,$b2,$a3,$b3); 196 &SVE_EOR($d0,$a0,$d1,$a1,$d2,$a2,$d3,$a3); 205 &SVE_ADD($a0,$b0,$a1,$b1,$a2,$b2,$a3,$b3); 206 &SVE_EOR($d0,$a0,$d1,$a1,$d2,$a2,$d3,$a3);
|
/openssl/crypto/ |
H A D | loongarch64cpuid.pl | 13 ($a0,$a1,$a2,$a3,$a4,$a5,$a6,$a7)=map("\$r$_",(4..11)); 26 my ($in_a,$in_b,$len,$m,$temp1,$temp2) = ($a0,$a1,$a2,$t0,$t1,$t2); 48 move $a0,$m 53 my ($ptr,$len,$temp1,$temp2) = ($a0,$a1,$t0,$t1); 104 rdtimel.w $a0,$zero
|
/openssl/crypto/md5/asm/ |
H A D | md5-loongarch64.pl | 18 my ($a0,$a1,$a2,$a3,$a4,$a5,$a6,$a7)=map("\$r$_",(4..11)); 176 # $a0 = arg #1 (ctx, MD5_CTX pointer) 190 ld.w $a4,$a0,0 # a4 = ctx->A 191 ld.w $a5,$a0,4 # a5 = ctx->B 192 ld.w $a6,$a0,8 # a6 = ctx->C 193 ld.w $a7,$a0,12 # a7 = ctx->D 283 st.w $a4,$a0,0 # ctx->A = A 284 st.w $a5,$a0,4 # ctx->B = B 285 st.w $a6,$a0,8 # ctx->C = C 286 st.w $a7,$a0,12 # ctx->D = D
|
/openssl/crypto/aes/asm/ |
H A D | vpaes-loongarch64.pl | 26 ($a0,$a1,$a2,$a3,$a4,$a5,$a6,$a7)=map("\$r$_",(4..11)); 276 vld $vr0,$a0,0 # load key (unaligned) 340 vld $vr0,$a0,8 #load key part 2 375 vld $vr0,$a0,16 # load key part 2 (unaligned) 706 xor $a0,$a0,$a0 745 xor $a0,$a0,$a0 764 vld $vr0,$a0,0x0 788 vld $vr0,$a0,0x0 802 my ($inp,$out,$len,$key,$ivp,$enc)=("$a0","$a1","$a2","$a3","$a4","$a5");
|
/openssl/test/recipes/30-test_evp_pkey_provided/ |
H A D | DSA.pub.txt | 13 6e:0a:1b:73:a0:4b:5f:4d:ab:35:28:fa:da:3a:0c: 16 21:a3:48:e9:f0:97:47:a0:27:ba:87:b8:15:5f:ff: 22 00:a0:b7:02:c4:ac:a6:42:ab:f2:34:0b:22:47:1f: 30 a8:40:48:42:7b:60:a0:c4:f2:3b:2d:0a:0c:b8:19:
|