Lines Matching refs:rsp
354 mov 0(%rsp),$ctx
361 mov 8(%rsp),%r15
363 mov 16(%rsp),%r14
365 mov 24(%rsp),%r13
367 mov 32(%rsp),%r12
369 mov 40(%rsp),%rbx
371 lea 48(%rsp),%rsp
438 mov %rsp,%rbp
623 mov %rsp,%rbp
757 mov %rsp,%rbp
853 lea 8(%rsp),%r10
855 and \$-32,%rsp
856 sub \$-8,%rsp
857 lea -0x58(%rsp),%r11
858 sub \$0x178,%rsp
861 lea -0xf8(%rsp),%r11
862 sub \$0x218,%rsp
914 vmovdqa $D0,0x00(%rsp)
919 vmovdqa $D1,0x10(%rsp)
924 vmovdqa $D2,0x20(%rsp)
929 vmovdqa $D0,0x30(%rsp)
934 vmovdqa $D1,0x40(%rsp)
939 vmovdqa $D2,0x50(%rsp)
944 vmovdqa $D0,0x60(%rsp)
948 vmovdqa $D1,0x70(%rsp)
950 vmovdqa 0x00(%rsp),$D4 # preload r0^2
953 vmovdqa $D2,0x80(%rsp)
983 vmovdqa 0x10(%rsp),$H2 # r1^2
988 vpmuludq 0x20(%rsp),$T4,$H0 # h4*s1
997 vmovdqa 0x30(%rsp),$H3 # r2^2
1004 vmovdqa 0x40(%rsp),$H4 # s2^2
1009 vmovdqa 0x50(%rsp),$H2 # r3^2
1014 vmovdqa 0x60(%rsp),$H3 # s3^2
1017 vmovdqa 0x80(%rsp),$H4 # s4^2
1039 vpmuludq 0x70(%rsp),$T0,$T4 # h0*r4
1153 vmovdqa 0x00(%rsp),$D4 # preload r0^2
1445 lea 0xf8(%r11),%rsp
1449 lea -8(%r10),%rsp
1450 .cfi_def_cfa_register %rsp
1545 mov %rsp,%rbp
1685 mov %rsp,%rbp
1809 lea 8(%rsp),%r10
1811 sub \$0x128,%rsp
1814 lea 8(%rsp),%r10
1815 sub \$0x1c8,%rsp
1835 and \$-512,%rsp
1841 lea 0x90(%rsp),%rax # size optimization
1848 vmovdqa $T2,0x00(%rsp)
1905 vmovdqa `32*0`(%rsp),$T0 # r0^4
1907 vmovdqa `32*1`(%rsp),$T1 # r1^4
1909 vmovdqa `32*3`(%rsp),$T2 # r2^4
1940 vpmuludq `32*2`(%rsp),$H4,$H2 # h4*s1
2055 vmovdqu `32*0+4`(%rsp),$T0 # r0^4
2057 vmovdqu `32*1+4`(%rsp),$T1 # r1^4
2059 vmovdqu `32*3+4`(%rsp),$T2 # r2^4
2075 vpmuludq `32*2+4`(%rsp),$H4,$H2 # h4*s1
2194 lea -8(%r10),%rsp
2198 lea -8(%r10),%rsp
2199 .cfi_def_cfa_register %rsp
2223 lea 8(%rsp),%r10
2225 sub \$0x128,%rsp
2228 lea 8(%rsp),%r10
2229 sub \$0x1c8,%rsp
2249 and \$-512,%rsp
2264 vmovdqa64 $R0,0x00(%rsp){%k2} # save in case $len%128 != 0
2267 vmovdqu64 $R1,0x00(%rsp,%rax){%k2}
2270 vmovdqa64 $S1,0x40(%rsp){%k2}
2273 vmovdqu64 $R2,0x40(%rsp,%rax){%k2}
2275 vmovdqa64 $S2,0x80(%rsp){%k2}
2276 vmovdqu64 $R3,0x80(%rsp,%rax){%k2}
2277 vmovdqa64 $S3,0xc0(%rsp){%k2}
2278 vmovdqu64 $R4,0xc0(%rsp,%rax){%k2}
2279 vmovdqa64 $S4,0x100(%rsp){%k2}
2763 lea 0x90(%rsp),%rax # size optimization for .Ltail_avx2
2786 lea -8(%r10),%rsp
2790 lea -8(%r10),%rsp
2791 .cfi_def_cfa_register %rsp
3996 sub \$64,%rsp
4046 sub \$64,%rsp
4094 mov %r10,32(%rsp) # arg5
4095 mov %r11,40(%rsp) # arg6
4096 mov %r12,48(%rsp) # arg7
4097 mov %rcx,56(%rsp) # arg8, (NULL)
4101 add \$64,%rsp