/external/libvpx/libvpx/vp8/common/x86/ |
D | recon_mmx.asm | 26 push rsi 30 mov rsi, arg(0) ;src; 31 movq mm0, [rsi] 36 movq mm1, [rsi+rax] 37 movq mm2, [rsi+rax*2] 40 lea rsi, [rsi+rax*2] 43 add rsi, rax 50 movq mm3, [rsi] 53 movq mm4, [rsi+rax] 55 movq mm5, [rsi+rax*2] [all …]
|
D | recon_sse2.asm | 25 push rsi 29 mov rsi, arg(0) ;src; 30 movdqu xmm0, [rsi] 35 movdqu xmm1, [rsi+rax] 36 movdqu xmm2, [rsi+rax*2] 39 lea rsi, [rsi+rax*2] 42 add rsi, rax 48 movdqu xmm3, [rsi] 51 movdqu xmm4, [rsi+rax] 53 movdqu xmm5, [rsi+rax*2] [all …]
|
D | subpixel_ssse3.asm | 44 push rsi 49 xor rsi, rsi 65 mov rsi, arg(0) ;src_ptr 74 movq xmm0, MMWORD PTR [rsi - 2] ; -2 -1 0 1 2 3 4 5 76 movq xmm2, MMWORD PTR [rsi + 3] ; 3 4 5 6 7 8 9 10 92 lea rsi, [rsi + rax] 109 pop rsi 123 mov rsi, arg(0) ;src_ptr 133 movq xmm0, MMWORD PTR [rsi - 2] ; -2 -1 0 1 2 3 4 5 135 movq xmm1, MMWORD PTR [rsi + 3] ; 3 4 5 6 7 8 9 10 [all …]
|
D | copy_sse2.asm | 27 push rsi 31 mov rsi, arg(0) ;src_ptr 39 movdqu xmm0, XMMWORD PTR [rsi] 40 movdqu xmm1, XMMWORD PTR [rsi + 16] 41 movdqu xmm2, XMMWORD PTR [rsi + rax] 42 movdqu xmm3, XMMWORD PTR [rsi + rax + 16] 44 lea rsi, [rsi+rax*2] 46 movdqu xmm4, XMMWORD PTR [rsi] 47 movdqu xmm5, XMMWORD PTR [rsi + 16] 48 movdqu xmm6, XMMWORD PTR [rsi + rax] [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | vpx_high_subpixel_8t_sse2.asm | 93 mov rsi, arg(0) ;src_ptr 137 movdqu xmm0, [rsi + %1] ;0 138 movdqu xmm1, [rsi + rax + %1] ;1 139 movdqu xmm6, [rsi + rdx * 2 + %1] ;6 140 lea rsi, [rsi + rax] 141 movdqu xmm7, [rsi + rdx * 2 + %1] ;7 142 movdqu xmm2, [rsi + rax + %1] ;2 143 movdqu xmm3, [rsi + rax * 2 + %1] ;3 144 movdqu xmm4, [rsi + rdx + %1] ;4 145 movdqu xmm5, [rsi + rax * 4 + %1] ;5 [all …]
|
D | vpx_subpixel_bilinear_ssse3.asm | 15 mov rsi, arg(0) ;src_ptr 45 lea rsi, [rsi + rax] 52 mov rsi, arg(0) ;src_ptr 84 lea rsi, [rsi + rax] 107 lea rsi, [rsi + rax] 117 push rsi 123 movd xmm0, [rsi] ;load src 124 movd xmm1, [rsi + rax] 131 pop rsi 142 push rsi [all …]
|
D | vpx_high_subpixel_bilinear_sse2.asm | 15 mov rsi, arg(0) ;src_ptr 62 lea rsi, [rsi + 2*rax] 70 mov rsi, arg(0) ;src_ptr 122 lea rsi, [rsi + 2*rax] 168 lea rsi, [rsi + 2*rax] 179 push rsi 185 movq xmm0, [rsi] ;load src 186 movq xmm1, [rsi + 2*rax] 193 pop rsi 205 push rsi [all …]
|
D | vpx_subpixel_bilinear_sse2.asm | 15 mov rsi, arg(0) ;src_ptr 55 lea rsi, [rsi + rax] 62 mov rsi, arg(0) ;src_ptr 99 lea rsi, [rsi + rax] 129 lea rsi, [rsi + rax] 139 push rsi 145 movd xmm0, [rsi] ;load src 146 movd xmm1, [rsi + rax] 153 pop rsi 164 push rsi [all …]
|
D | sad_mmx.asm | 29 push rsi 33 mov rsi, arg(0) ;src_ptr 39 lea rcx, [rsi+rax*8] 48 movq mm0, QWORD PTR [rsi] 49 movq mm2, QWORD PTR [rsi+8] 79 lea rsi, [rsi+rax] 85 cmp rsi, rcx 104 pop rsi 121 push rsi 125 mov rsi, arg(0) ;src_ptr [all …]
|
/external/boringssl/linux-x86_64/crypto/rc4/ |
D | rc4-x86_64.S | 11 orq %rsi,%rsi 19 movq %rsi,%r11 41 leaq 1(%r10),%rsi 59 leaq 1(%r10),%rsi 66 movl 0(%rdi,%rsi,4),%ebx 74 movl 4(%rdi,%rsi,4),%eax 82 movl 8(%rdi,%rsi,4),%ebx 90 movl 12(%rdi,%rsi,4),%eax 98 movl 16(%rdi,%rsi,4),%ebx 106 movl 20(%rdi,%rsi,4),%eax [all …]
|
/external/boringssl/src/crypto/curve25519/asm/ |
D | x25519-asm-x86_64.S | 71 movq 0(%rdi),%rsi 81 mov %rsi,%r12 83 and %rax,%rsi 101 add %r12,%rsi 105 cmp %r10,%rsi 118 sub %r10,%rsi 123 movq %rsi,0(%rdi) 137 mov %rsi,%rdx 157 movq 24(%rsi),%rdx 163 movq 32(%rsi),%rdx [all …]
|
/external/compiler-rt/lib/tsan/rtl/ |
D | tsan_rtl_amd64.S | 22 push %rsi 24 CFI_REL_OFFSET(%rsi, 0) 67 pop %rsi 79 CFI_RESTORE(%rsi) 102 push %rsi 104 CFI_REL_OFFSET(%rsi, 0) 147 pop %rsi 159 CFI_RESTORE(%rsi) 183 mov %rdi, %rsi 186 mov %rdi, %rsi [all …]
|
/external/boringssl/linux-x86_64/crypto/bn/ |
D | rsaz-x86_64.S | 22 movq (%rsi),%rdx 23 movq 8(%rsi),%rax 34 movq 16(%rsi),%rax 39 movq 24(%rsi),%rax 45 movq 32(%rsi),%rax 51 movq 40(%rsi),%rax 57 movq 48(%rsi),%rax 63 movq 56(%rsi),%rax 86 movq 8(%rsi),%r8 87 movq 16(%rsi),%rax [all …]
|
/external/boringssl/mac-x86_64/crypto/bn/ |
D | rsaz-x86_64.S | 21 movq (%rsi),%rdx 22 movq 8(%rsi),%rax 33 movq 16(%rsi),%rax 38 movq 24(%rsi),%rax 44 movq 32(%rsi),%rax 50 movq 40(%rsi),%rax 56 movq 48(%rsi),%rax 62 movq 56(%rsi),%rax 85 movq 8(%rsi),%r8 86 movq 16(%rsi),%rax [all …]
|
/external/boringssl/mac-x86_64/crypto/rc4/ |
D | rc4-x86_64.S | 10 orq %rsi,%rsi 18 movq %rsi,%r11 40 leaq 1(%r10),%rsi 58 leaq 1(%r10),%rsi 65 movl 0(%rdi,%rsi,4),%ebx 73 movl 4(%rdi,%rsi,4),%eax 81 movl 8(%rdi,%rsi,4),%ebx 89 movl 12(%rdi,%rsi,4),%eax 97 movl 16(%rdi,%rsi,4),%ebx 105 movl 20(%rdi,%rsi,4),%eax [all …]
|
/external/boringssl/linux-x86_64/crypto/aes/ |
D | aesni-x86_64.S | 25 movups %xmm2,(%rsi) 50 movups %xmm2,(%rsi) 527 movups %xmm2,(%rsi) 531 movups %xmm3,16(%rsi) 533 movups %xmm4,32(%rsi) 535 movups %xmm5,48(%rsi) 537 movups %xmm6,64(%rsi) 539 movups %xmm7,80(%rsi) 541 movups %xmm8,96(%rsi) 543 movups %xmm9,112(%rsi) [all …]
|
/external/boringssl/win-x86_64/crypto/rc4/ |
D | rc4-x86_64.asm | 14 mov QWORD[16+rsp],rsi 18 mov rsi,rdx 23 or rsi,rsi 26 mov rsi,QWORD[16+rsp] 33 mov r11,rsi 55 lea rsi,[1+r10] 73 lea rsi,[1+r10] 80 mov ebx,DWORD[rsi*4+rdi] 88 mov eax,DWORD[4+rsi*4+rdi] 96 mov ebx,DWORD[8+rsi*4+rdi] [all …]
|
/external/libvpx/libvpx/vp8/encoder/x86/ |
D | encodeopt.asm | 20 push rsi 24 mov rsi, arg(0) ;coeff_ptr 27 movdqa xmm0, [rsi] 30 movdqa xmm2, [rsi+16] 56 pop rsi 68 push rsi 73 mov rsi, arg(0) ;coeff_ptr 77 movq mm3, [rsi] 80 movq mm5, [rsi+8] 98 movq mm3, [rsi+16] [all …]
|
/external/boringssl/mac-x86_64/crypto/ec/ |
D | p256-x86_64-asm.S | 25 movq 0(%rsi),%r8 26 movq 8(%rsi),%r9 28 movq 16(%rsi),%r10 30 movq 24(%rsi),%r11 31 leaq L$poly(%rip),%rsi 38 subq 0(%rsi),%r8 40 sbbq 8(%rsi),%r9 41 sbbq 16(%rsi),%r10 43 sbbq 24(%rsi),%r11 76 subq 0(%rsi),%r8 [all …]
|
/external/boringssl/linux-x86_64/crypto/ec/ |
D | p256-x86_64-asm.S | 26 movq 0(%rsi),%r8 27 movq 8(%rsi),%r9 29 movq 16(%rsi),%r10 31 movq 24(%rsi),%r11 32 leaq .Lpoly(%rip),%rsi 39 subq 0(%rsi),%r8 41 sbbq 8(%rsi),%r9 42 sbbq 16(%rsi),%r10 44 sbbq 24(%rsi),%r11 77 subq 0(%rsi),%r8 [all …]
|
/external/boringssl/win-x86_64/crypto/md5/ |
D | md5-x86_64.asm | 13 mov QWORD[16+rsp],rsi 17 mov rsi,rdx 33 lea rdi,[rdx*1+rsi] 45 cmp rsi,rdi 54 mov r10d,DWORD[rsi] 60 mov r10d,DWORD[4+rsi] 69 mov r10d,DWORD[8+rsi] 78 mov r10d,DWORD[12+rsi] 87 mov r10d,DWORD[16+rsi] 96 mov r10d,DWORD[20+rsi] [all …]
|
/external/boringssl/mac-x86_64/crypto/aes/ |
D | aesni-x86_64.S | 24 movups %xmm2,(%rsi) 49 movups %xmm2,(%rsi) 526 movups %xmm2,(%rsi) 530 movups %xmm3,16(%rsi) 532 movups %xmm4,32(%rsi) 534 movups %xmm5,48(%rsi) 536 movups %xmm6,64(%rsi) 538 movups %xmm7,80(%rsi) 540 movups %xmm8,96(%rsi) 542 movups %xmm9,112(%rsi) [all …]
|
/external/boringssl/src/crypto/md5/asm/ |
D | md5-x86_64.pl | 22 $code .= " mov 0*4(%rsi), %r10d /* (NEXT STEP) X[0] */\n" if ($pos == -1); 29 mov $k_next*4(%rsi),%r10d /* (NEXT STEP) X[$k_next] */ 46 $code .= " mov 1*4(%rsi), %r10d /* (NEXT STEP) X[1] */\n" if ($pos == -1); 54 mov $k_next*4(%rsi),%r10d /* (NEXT STEP) X[$k_next] */ 72 $code .= " mov 5*4(%rsi), %r10d /* (NEXT STEP) X[5] */\n" if ($pos == -1); 76 mov $k_next*4(%rsi),%r10d /* (NEXT STEP) X[$k_next] */ 94 $code .= " mov 0*4(%rsi), %r10d /* (NEXT STEP) X[0] */\n" if ($pos == -1); 103 mov $k_next*4(%rsi),%r10d /* (NEXT STEP) X[$k_next] */ 145 lea (%rsi,%rdx), %rdi # rdi = end 157 cmp %rdi, %rsi # cmp end with ptr [all …]
|
/external/boringssl/win-x86_64/crypto/bn/ |
D | rsaz-x86_64.asm | 15 mov QWORD[16+rsp],rsi 19 mov rsi,rdx 35 mov rdx,QWORD[rsi] 36 mov rax,QWORD[8+rsi] 47 mov rax,QWORD[16+rsi] 52 mov rax,QWORD[24+rsi] 58 mov rax,QWORD[32+rsi] 64 mov rax,QWORD[40+rsi] 70 mov rax,QWORD[48+rsi] 76 mov rax,QWORD[56+rsi] [all …]
|
/external/boringssl/mac-x86_64/crypto/md5/ |
D | md5-x86_64.S | 21 leaq (%rsi,%rdx,1),%rdi 33 cmpq %rdi,%rsi 42 movl 0(%rsi),%r10d 48 movl 4(%rsi),%r10d 57 movl 8(%rsi),%r10d 66 movl 12(%rsi),%r10d 75 movl 16(%rsi),%r10d 84 movl 20(%rsi),%r10d 93 movl 24(%rsi),%r10d 102 movl 28(%rsi),%r10d [all …]
|