Lines Matching refs:rsp

217 	"&mov	(\"4*$c0(%rsp)\",$xc)",	# reload pair of 'c's
218 "&mov (\"4*$c1(%rsp)\",$xc_)",
219 "&mov ($xc,\"4*$c2(%rsp)\")",
220 "&mov ($xc_,\"4*$c3(%rsp)\")",
286 sub \$64+24,%rsp
296 #movdqa %xmm0,4*0(%rsp) # key[0]
297 movdqa %xmm1,4*4(%rsp) # key[1]
298 movdqa %xmm2,4*8(%rsp) # key[2]
299 movdqa %xmm3,4*12(%rsp) # key[3]
309 mov 4*4(%rsp),@x[4]
310 mov 4*5(%rsp),@x[5]
311 mov 4*6(%rsp),@x[6]
312 mov 4*7(%rsp),@x[7]
314 mov 4*13(%rsp),@x[13]
315 mov 4*14(%rsp),@x[14]
316 mov 4*15(%rsp),@x[15]
318 mov %rbp,64+0(%rsp) # save len
320 mov $inp,64+8(%rsp) # save inp
322 mov $out,64+16(%rsp) # save out
336 mov @t[1],4*9(%rsp) # modulo-scheduled
337 mov @t[0],4*8(%rsp)
338 mov 64(%rsp),%rbp # load len
340 mov 64+8(%rsp),$inp # load inp
342 mov 64+16(%rsp),$out # load out
348 add 4*4(%rsp),@x[4]
349 add 4*5(%rsp),@x[5]
350 add 4*6(%rsp),@x[6]
351 add 4*7(%rsp),@x[7]
352 add 4*12(%rsp),@x[12]
353 add 4*13(%rsp),@x[13]
354 add 4*14(%rsp),@x[14]
355 add 4*15(%rsp),@x[15]
356 paddd 4*8(%rsp),%xmm1
377 movdqa %xmm2,4*8(%rsp)
378 movd %xmm3,4*12(%rsp)
402 mov @x[0],4*0(%rsp)
403 mov @x[1],4*1(%rsp)
405 mov @x[2],4*2(%rsp)
406 mov @x[3],4*3(%rsp)
407 mov @x[4],4*4(%rsp)
408 mov @x[5],4*5(%rsp)
409 mov @x[6],4*6(%rsp)
410 mov @x[7],4*7(%rsp)
411 movdqa %xmm1,4*8(%rsp)
412 mov @x[12],4*12(%rsp)
413 mov @x[13],4*13(%rsp)
414 mov @x[14],4*14(%rsp)
415 mov @x[15],4*15(%rsp)
419 movzb (%rsp,%rbx),%edx
427 lea 64+24+48(%rsp),%rsi
441 lea (%rsi),%rsp
442 .cfi_def_cfa_register %rsp
486 mov %rsp,%r9 # frame pointer
499 sub \$64+$xframe,%rsp
514 movdqa $a,0x00(%rsp)
515 movdqa $b,0x10(%rsp)
516 movdqa $c,0x20(%rsp)
517 movdqa $d,0x30(%rsp)
524 movdqa 0x00(%rsp),$a
525 movdqa 0x10(%rsp),$b
526 movdqa 0x20(%rsp),$c
527 paddd 0x30(%rsp),$d
529 movdqa $d,0x30(%rsp)
550 paddd 0x00(%rsp),$a
551 paddd 0x10(%rsp),$b
552 paddd 0x20(%rsp),$c
553 paddd 0x30(%rsp),$d
581 movdqa $a,0x00(%rsp)
582 movdqa $b,0x10(%rsp)
583 movdqa $c,0x20(%rsp)
584 movdqa $d,0x30(%rsp)
589 movzb (%rsp,$counter),%ecx
603 lea (%r9),%rsp
604 .cfi_def_cfa_register %rsp
668 mov %rsp,%r9 # frame pointer
670 sub \$64+$xframe,%rsp
691 movdqa $a,0x00(%rsp)
693 movdqa $b,0x10(%rsp)
695 movdqa $c,0x20(%rsp)
697 movdqa $d,0x30(%rsp)
724 paddd 0x00(%rsp),$a
725 paddd 0x10(%rsp),$b
726 paddd 0x20(%rsp),$c
727 paddd 0x30(%rsp),$d
729 paddd 0x00(%rsp),$a1
730 paddd 0x10(%rsp),$b1
731 paddd 0x20(%rsp),$c1
732 paddd 0x30(%rsp),$d1
769 lea (%r9),%rsp
770 .cfi_def_cfa_register %rsp
860 "&movdqa (\"`16*($c0-8)`(%rsp)\",$xc)", # reload pair of 'c's
861 "&movdqa (\"`16*($c1-8)`(%rsp)\",$xc_)",
862 "&movdqa ($xc,\"`16*($c2-8)`(%rsp)\")",
863 "&movdqa ($xc_,\"`16*($c3-8)`(%rsp)\")",
917 mov %rsp,%r9 # frame pointer
935 sub \$0x140+$xframe,%rsp
963 lea 0x100(%rsp),%rcx # size optimization
969 movdqa $xa0,0x40(%rsp) # ... and offload
971 movdqa $xa1,0x50(%rsp)
973 movdqa $xa2,0x60(%rsp)
974 movdqa $xa3,0x70(%rsp)
1007 movdqa 0x40(%rsp),$xa0 # re-load smashed key
1008 movdqa 0x50(%rsp),$xa1
1009 movdqa 0x60(%rsp),$xa2
1010 movdqa 0x70(%rsp),$xa3
1026 movdqa $xt2,0x20(%rsp) # SIMD equivalent of "@x[10]"
1027 movdqa $xt3,0x30(%rsp) # SIMD equivalent of "@x[11]"
1042 paddd 0x40(%rsp),$xa0 # accumulate key material
1043 paddd 0x50(%rsp),$xa1
1044 paddd 0x60(%rsp),$xa2
1045 paddd 0x70(%rsp),$xa3
1067 movdqa $xa0,0x00(%rsp) # offload $xaN
1068 movdqa $xa1,0x10(%rsp)
1069 movdqa 0x20(%rsp),$xa0 # "xc2"
1070 movdqa 0x30(%rsp),$xa1 # "xc3"
1093 movdqa $xa2,0x20(%rsp) # keep offloading $xaN
1094 movdqa $xa3,0x30(%rsp)
1139 pxor 0x00(%rsp),$xt0 # $xaN is offloaded, remember?
1153 pxor 0x10(%rsp),$xt0
1167 pxor 0x20(%rsp),$xt0
1181 pxor 0x30(%rsp),$xt0
1204 #movdqa 0x00(%rsp),$xt0 # $xaN is offloaded, remember?
1206 #movdqa $xt0,0x00(%rsp)
1207 movdqa $xb0,0x10(%rsp)
1208 movdqa $xc0,0x20(%rsp)
1209 movdqa $xd0,0x30(%rsp)
1218 pxor 0x00(%rsp),$xt0 # $xaxN is offloaded, remember?
1228 movdqa 0x10(%rsp),$xt0 # $xaN is offloaded, remember?
1231 movdqa $xt0,0x00(%rsp)
1232 movdqa $xb1,0x10(%rsp)
1234 movdqa $xc1,0x20(%rsp)
1236 movdqa $xd1,0x30(%rsp)
1245 pxor 0x00(%rsp),$xt0 # $xaN is offloaded, remember?
1258 pxor 0x10(%rsp),$xt0
1268 movdqa 0x20(%rsp),$xt0 # $xaN is offloaded, remember?
1271 movdqa $xt0,0x00(%rsp)
1272 movdqa $xb2,0x10(%rsp)
1274 movdqa $xc2,0x20(%rsp)
1276 movdqa $xd2,0x30(%rsp)
1285 pxor 0x00(%rsp),$xt0 # $xaN is offloaded, remember?
1299 pxor 0x10(%rsp),$xt0
1313 pxor 0x20(%rsp),$xt0
1323 movdqa 0x30(%rsp),$xt0 # $xaN is offloaded, remember?
1326 movdqa $xt0,0x00(%rsp)
1327 movdqa $xb3,0x10(%rsp)
1329 movdqa $xc3,0x20(%rsp)
1331 movdqa $xd3,0x30(%rsp)
1335 movzb (%rsp,%r10),%ecx
1357 lea (%r9),%rsp
1358 .cfi_def_cfa_register %rsp
1450 mov %rsp,%r9 # frame pointer
1452 sub \$0x140+$xframe,%rsp
1482 lea 0x100(%rsp),%rcx # size optimization
1486 vmovdqa $xa0,0x40(%rsp) # ... and offload
1488 vmovdqa $xa1,0x50(%rsp)
1490 vmovdqa $xa2,0x60(%rsp)
1491 vmovdqa $xa3,0x70(%rsp)
1524 vmovdqa 0x40(%rsp),$xa0 # re-load smashed key
1525 vmovdqa 0x50(%rsp),$xa1
1526 vmovdqa 0x60(%rsp),$xa2
1527 vmovdqa 0x70(%rsp),$xa3
1556 vpaddd 0x40(%rsp),$xa0,$xa0 # accumulate key material
1557 vpaddd 0x50(%rsp),$xa1,$xa1
1558 vpaddd 0x60(%rsp),$xa2,$xa2
1559 vpaddd 0x70(%rsp),$xa3,$xa3
1561 vmovdqa $xt2,0x20(%rsp) # offload $xc2,3
1562 vmovdqa $xt3,0x30(%rsp)
1580 vmovdqa $xa0,0x00(%rsp) # offload $xa0,1
1581 vmovdqa $xa1,0x10(%rsp)
1582 vmovdqa 0x20(%rsp),$xa0 # "xc2"
1583 vmovdqa 0x30(%rsp),$xa1 # "xc3"
1630 vmovdqa 0x00(%rsp),$xa0 # restore $xa0,1
1631 vmovdqa 0x10(%rsp),$xa1
1689 vmovdqa $xa0,0x00(%rsp)
1690 vmovdqa $xb0,0x10(%rsp)
1691 vmovdqa $xc0,0x20(%rsp)
1692 vmovdqa $xd0,0x30(%rsp)
1708 vmovdqa $xa1,0x00(%rsp)
1710 vmovdqa $xb1,0x10(%rsp)
1712 vmovdqa $xc1,0x20(%rsp)
1714 vmovdqa $xd1,0x30(%rsp)
1739 vmovdqa $xa2,0x00(%rsp)
1741 vmovdqa $xb2,0x10(%rsp)
1743 vmovdqa $xc2,0x20(%rsp)
1745 vmovdqa $xd2,0x30(%rsp)
1780 vmovdqa $xa3,0x00(%rsp)
1782 vmovdqa $xb3,0x10(%rsp)
1784 vmovdqa $xc3,0x20(%rsp)
1786 vmovdqa $xd3,0x30(%rsp)
1790 movzb (%rsp,%r10),%ecx
1813 lea (%r9),%rsp
1814 .cfi_def_cfa_register %rsp
1899 "&vmovdqa (\"`32*($c0-8)`(%rsp)\",$xc)", # reload pair of 'c's
1900 "&vmovdqa (\"`32*($c1-8)`(%rsp)\",$xc_)",
1901 "&vmovdqa ($xc,\"`32*($c2-8)`(%rsp)\")",
1902 "&vmovdqa ($xc_,\"`32*($c3-8)`(%rsp)\")",
1952 mov %rsp,%r9 # frame register
1954 sub \$0x280+$xframe,%rsp
1955 and \$-32,%rsp
1986 lea 0x100(%rsp),%rcx # size optimization
1987 lea 0x200(%rsp),%rax # size optimization
2050 vmovdqa $xt2,0x40(%rsp) # SIMD equivalent of "@x[10]"
2051 vmovdqa $xt3,0x60(%rsp) # SIMD equivalent of "@x[11]"
2066 lea 0x200(%rsp),%rax # size optimization
2111 vmovdqa $xa0,0x00(%rsp) # offload $xaN
2112 vmovdqa $xa1,0x20(%rsp)
2113 vmovdqa 0x40(%rsp),$xc2 # $xa0
2114 vmovdqa 0x60(%rsp),$xc3 # $xa1
2162 vmovdqa 0x00(%rsp),$xa0 # $xaN was offloaded, remember?
2163 vmovdqa 0x20(%rsp),$xa1
2234 vmovdqa $xa0,0x00(%rsp)
2235 vmovdqa $xb0,0x20(%rsp)
2248 vmovdqa $xc0,0x00(%rsp)
2251 vmovdqa $xd0,0x20(%rsp)
2268 vmovdqa $xa1,0x00(%rsp)
2271 vmovdqa $xb1,0x20(%rsp)
2292 vmovdqa $xc1,0x00(%rsp)
2295 vmovdqa $xd1,0x20(%rsp)
2320 vmovdqa $xa2,0x00(%rsp)
2323 vmovdqa $xb2,0x20(%rsp)
2352 vmovdqa $xc2,0x00(%rsp)
2355 vmovdqa $xd2,0x20(%rsp)
2388 vmovdqa $xa3,0x00(%rsp)
2391 vmovdqa $xb3,0x20(%rsp)
2428 vmovdqa $xc3,0x00(%rsp)
2431 vmovdqa $xd3,0x20(%rsp)
2435 movzb (%rsp,%r10),%ecx
2458 lea (%r9),%rsp
2459 .cfi_def_cfa_register %rsp
2514 mov %rsp,%r9 # frame pointer
2519 sub \$64+$xframe,%rsp
2668 vmovdqa %x#$a,0x00(%rsp)
2669 vmovdqa %x#$b,0x10(%rsp)
2670 vmovdqa %x#$c,0x20(%rsp)
2671 vmovdqa %x#$d,0x30(%rsp)
2677 vmovdqa $t0,0x00(%rsp)
2678 vmovdqa $t1,0x10(%rsp)
2679 vmovdqa $t2,0x20(%rsp)
2680 vmovdqa $t3,0x30(%rsp)
2685 movzb (%rsp,$counter),%ecx
2692 vmovdqu32 $a_,0x00(%rsp)
2710 lea (%r9),%rsp
2711 .cfi_def_cfa_register %rsp
2726 mov %rsp,%r9 # frame pointer
2731 sub \$64+$xframe,%rsp
2836 vmovdqa %x#$a,0x00(%rsp)
2837 vmovdqa %x#$b,0x10(%rsp)
2838 vmovdqa %x#$c,0x20(%rsp)
2839 vmovdqa %x#$d,0x30(%rsp)
2845 vmovdqa $t0,0x00(%rsp)
2846 vmovdqa $t1,0x10(%rsp)
2847 vmovdqa $t2,0x20(%rsp)
2848 vmovdqa $t3,0x30(%rsp)
2853 movzb (%rsp,$counter),%ecx
2860 vmovdqu32 $a_,0x00(%rsp)
2861 vmovdqu32 $a_,0x20(%rsp)
2879 lea (%r9),%rsp
2880 .cfi_def_cfa_register %rsp
2967 mov %rsp,%r9 # frame register
2969 sub \$64+$xframe,%rsp
2970 and \$-64,%rsp
3353 vmovdqa32 $xa0,0x00(%rsp)
3359 movzb (%rsp,%r10),%ecx
3367 vmovdqa32 $xa0,0(%rsp)
3385 lea (%r9),%rsp
3386 .cfi_def_cfa_register %rsp
3407 mov %rsp,%r9 # frame register
3409 sub \$64+$xframe,%rsp
3410 and \$-64,%rsp
3741 vmovdqa $xa0,0x00(%rsp)
3742 vmovdqa $xb0,0x20(%rsp)
3748 movzb (%rsp,%r10),%ecx
3756 vmovdqa $xa0,0x00(%rsp)
3757 vmovdqa $xa0,0x20(%rsp)
3775 lea (%r9),%rsp
3776 .cfi_def_cfa_register %rsp
3806 sub \$64,%rsp
3859 mov %r10,32(%rsp) # arg5
3860 mov %r11,40(%rsp) # arg6
3861 mov %r12,48(%rsp) # arg7
3862 mov %rcx,56(%rsp) # arg8, (NULL)
3866 add \$64,%rsp
3891 sub \$64,%rsp