Lines Matching refs:Xi

138 $Xi="%rdi";
261 movzb 15($Xi),$Zlo
264 &loop ($Xi);
266 mov $Zlo,8($Xi)
267 mov $Zhi,($Xi)
343 &mov ($Zlo,"8($Xi)");
344 &mov ($Zhi,"0($Xi)");
354 &mov ("($Xi)",$Zhi);
355 &mov ("8($Xi)","%rdx");
390 &mov ($dat,"$j($Xi)") if (--$j%4==0);
431 mov $Zlo,8($Xi)
432 mov $Zhi,($Xi)
462 ($Xi,$Xhi)=("%xmm0","%xmm1"); $Hkey="%xmm2";
466 my ($Xhi,$Xi,$Hkey,$HK)=@_;
470 movdqa $Xi,$Xhi #
471 pshufd \$0b01001110,$Xi,$T1
473 pxor $Xi,$T1 #
478 movdqa $Xi,$Xhi #
479 pshufd \$0b01001110,$Xi,$T1
480 pxor $Xi,$T1 #
484 pclmulqdq \$0x00,$Hkey,$Xi #######
487 pxor $Xi,$T1 #
494 pxor $T2,$Xi #
499 my ($Xhi,$Xi) = @_;
503 movdqa $Xi,$T2 #
504 movdqa $Xi,$T1
505 psllq \$5,$Xi
506 pxor $Xi,$T1 #
507 psllq \$1,$Xi
508 pxor $T1,$Xi #
509 psllq \$57,$Xi #
510 movdqa $Xi,$T1 #
511 pslldq \$8,$Xi
513 pxor $T2,$Xi
517 movdqa $Xi,$T2
518 psrlq \$1,$Xi
520 pxor $Xi,$T2
521 psrlq \$5,$Xi
522 pxor $T2,$Xi #
523 psrlq \$1,$Xi #
524 pxor $Xhi,$Xi #
566 movdqa $Hkey,$Xi
569 &clmul64x64_T2 ($Xhi,$Xi,$Hkey,$HK);
570 &reduction_alg9 ($Xhi,$Xi);
573 pshufd \$0b01001110,$Xi,$T2
576 pxor $Xi,$T2 # Karatsuba pre-processing
577 movdqu $Xi,0x10($Htbl) # save H^2
582 &clmul64x64_T2 ($Xhi,$Xi,$Hkey,$HK); # H^3
583 &reduction_alg9 ($Xhi,$Xi);
585 movdqa $Xi,$T3
587 &clmul64x64_T2 ($Xhi,$Xi,$Hkey,$HK); # H^4
588 &reduction_alg9 ($Xhi,$Xi);
591 pshufd \$0b01001110,$Xi,$T2
594 pxor $Xi,$T2 # Karatsuba pre-processing
595 movdqu $Xi,0x40($Htbl) # save H^4
622 movdqu ($Xip),$Xi
626 pshufb $T3,$Xi
628 &clmul64x64_T2 ($Xhi,$Xi,$Hkey,$T2);
629 $code.=<<___ if (0 || (&reduction_alg9($Xhi,$Xi)&&0));
638 pand $Xi,$T3
639 pshufb $T3,$T2 # ($Xi&7)·0xE0
641 pclmulqdq \$0x00,$Xi,$T1 # ·(0xE1<<1)
642 pxor $Xi,$T2
645 pxor $T2,$Xi
646 pclmulqdq \$0x01,$T3,$Xi
650 pslldq \$7,$Xi
651 pxor $Xhi,$Xi
654 pshufb $T3,$Xi
655 movdqu $Xi,($Xip)
694 movdqu ($Xip),$Xi
697 pshufb $T3,$Xi
752 pxor $T1,$Xi
755 movdqa $Xi,$Xhi
756 pshufd \$0b01001110,$Xi,$T1
757 pxor $Xi,$T1
770 pclmulqdq \$0x00,$Hkey4,$Xi
775 xorps $Xln,$Xi
788 pxor $Xi,$T1 # aggregated Karatsuba post-processing
796 pxor $T1,$Xi
801 pand $Xi,$T1 # 1st phase
803 pxor $Xi,$T2 #
810 pxor $T2,$Xi
814 movdqa $Xi,$T2 # 2nd phase
815 psrlq \$1,$Xi
825 pxor $Xi,$T2
826 psrlq \$5,$Xi
831 pxor $T2,$Xi #
835 psrlq \$1,$Xi #
836 pxor $Xhi,$Xi #
837 movdqa $Xi,$Xhi
840 pshufd \$0b01001110,$Xi,$T1
841 pxor $Xi,$T1
851 pclmulqdq \$0x00,$Hkey4,$Xi
855 xorps $Xln,$Xi
857 pxor $Xi,$Xhi # aggregated Karatsuba post-processing
861 pxor $Xi,$Xhi
867 pxor $T2,$Xi #
869 &reduction_alg9($Xhi,$Xi);
882 # [(H*Ii+1) + (H*Xi+1)] mod P =
889 pxor $T1,$Xi # Ii+Xi
907 movdqa $Xi,$Xhi
909 pshufd \$0b01001110,$Xi,$Xmn #
910 pxor $Xi,$Xmn #
912 pclmulqdq \$0x00,$Hkey2,$Xi
916 pxor $Xln,$Xi # (H*Ii+1) + H^2*(Ii+Xi)
919 pxor $Xi,$T1 # aggregated Karatsuba post-processing
931 pxor $Xmn,$Xi #
935 movdqa $Xi,$T2 # 1st phase
936 movdqa $Xi,$T1
937 psllq \$5,$Xi
938 pxor $Xi,$T1 #
940 psllq \$1,$Xi
941 pxor $T1,$Xi #
942 psllq \$57,$Xi #
943 movdqa $Xi,$T1 #
944 pslldq \$8,$Xi
946 pxor $T2,$Xi
951 movdqa $Xi,$T2 # 2nd phase
952 psrlq \$1,$Xi
955 pxor $Xi,$T2
956 psrlq \$5,$Xi
957 pxor $T2,$Xi #
959 psrlq \$1,$Xi #
961 pxor $Xhi,$Xi #
967 movdqa $Xi,$Xhi
969 pshufd \$0b01001110,$Xi,$Xmn #
970 pxor $Xi,$Xmn #
972 pclmulqdq \$0x00,$Hkey2,$Xi
976 pxor $Xln,$Xi # (H*Ii+1) + H^2*(Ii+Xi)
978 pxor $Xi,$T1
985 pxor $Xmn,$Xi #
987 &reduction_alg9 ($Xhi,$Xi);
995 pxor $T1,$Xi # Ii+Xi
997 &clmul64x64_T2 ($Xhi,$Xi,$Hkey,$HK); # H*(Ii+Xi)
998 &reduction_alg9 ($Xhi,$Xi);
1001 pshufb $T3,$Xi
1002 movdqu $Xi,($Xip)
1063 vmovdqa $Hkey,$Xi
1070 my ($Xhi,$Xi,$Hkey,$HK)=@_;
1074 vpunpckhqdq $Xi,$Xi,$T1
1076 vpxor $Xi,$T1,$T1 #
1081 vpunpckhqdq $Xi,$Xi,$T1
1082 vpxor $Xi,$T1,$T1 #
1086 vpclmulqdq \$0x11,$Hkey,$Xi,$Xhi #######
1087 vpclmulqdq \$0x00,$Hkey,$Xi,$Xi #######
1089 vpxor $Xi,$Xhi,$T2 #
1094 vpxor $T2,$Xi,$Xi #
1100 my ($Xhi,$Xi) = @_;
1103 vpsllq \$57,$Xi,$T1 # 1st phase
1104 vpsllq \$62,$Xi,$T2
1106 vpsllq \$63,$Xi,$T1
1110 vpxor $T1,$Xi,$Xi #
1113 vpsrlq \$1,$Xi,$T2 # 2nd phase
1114 vpxor $Xi,$Xhi,$Xhi
1115 vpxor $T2,$Xi,$Xi #
1117 vpxor $T2,$Xi,$Xi #
1118 vpsrlq \$1,$Xi,$Xi #
1119 vpxor $Xhi,$Xi,$Xi #
1129 &clmul64x64_avx ($Xhi,$Xi,$Hkey,$HK); # calculate H^3,5,7
1130 &reduction_avx ($Xhi,$Xi);
1133 vmovdqa $Xi,$T3
1135 &clmul64x64_avx ($Xhi,$Xi,$Hkey,$HK); # calculate H^2,4,6,8
1136 &reduction_avx ($Xhi,$Xi);
1139 vpshufd \$0b01001110,$Xi,$T2
1142 vpxor $Xi,$T2,$T2 # Karatsuba pre-processing
1143 vmovdqu $Xi,0x10($Htbl) # save H^2,4,6,8
1196 $Xi,$Xo,$Tred,$bswap,$Ii,$Ij) = map("%xmm$_",(0..15));
1217 vmovdqu ($Xip),$Xi # load $Xi
1221 vpshufb $bswap,$Xi,$Xi
1316 vpxor $Xi,$Ij,$Ij # accumulate $Xi
1326 vpclmulqdq \$0x00,$Hkey,$Ij,$Xi
1339 vpxor $Zlo,$Xi,$Xi # collect result
1350 vpxor $Xi,$Tred,$Tred # aggregated Karatsuba post-processing
1357 vpxor $T2, $Xi, $Xi
1369 vpalignr \$8,$Xi,$Xi,$Tred # 1st phase
1382 vpclmulqdq \$0x10,(%r10),$Xi,$Xi
1406 vxorps $Tred,$Xi,$Xi
1409 vpalignr \$8,$Xi,$Xi,$Tred # 2nd phase
1415 vpclmulqdq \$0x10,(%r10),$Xi,$Xi
1431 vpxor $Xi,$Ij,$Ij # accumulate $Xi
1546 vpxor $Xi,$Ij,$Ij # accumulate $Xi
1559 vpxor $Xlo,$Zlo,$Xi
1563 vpxor $Xi, $Zmi,$Zmi # aggregated Karatsuba post-processing
1567 vpxor $T2, $Xi, $Xi
1570 vpclmulqdq \$0x10,$Tred,$Xi,$T2 # 1st phase
1571 vpalignr \$8,$Xi,$Xi,$Xi
1572 vpxor $T2,$Xi,$Xi
1574 vpclmulqdq \$0x10,$Tred,$Xi,$T2 # 2nd phase
1575 vpalignr \$8,$Xi,$Xi,$Xi
1576 vpxor $Xo,$Xi,$Xi
1577 vpxor $T2,$Xi,$Xi
1582 vpshufb $bswap,$Xi,$Xi
1583 vmovdqu $Xi,($Xip)