/external/libvpx/libvpx/vp8/common/x86/ |
D | recon_sse2.asm | 27 push rsi 31 mov rsi, arg(0) ;src; 32 movdqu xmm0, [rsi] 37 movdqu xmm1, [rsi+rax] 38 movdqu xmm2, [rsi+rax*2] 41 lea rsi, [rsi+rax*2] 44 add rsi, rax 50 movdqu xmm3, [rsi] 53 movdqu xmm4, [rsi+rax] 55 movdqu xmm5, [rsi+rax*2] [all …]
|
D | recon_mmx.asm | 27 push rsi 31 mov rsi, arg(0) ;src; 32 movq mm0, [rsi] 37 movq mm1, [rsi+rax] 38 movq mm2, [rsi+rax*2] 41 lea rsi, [rsi+rax*2] 44 add rsi, rax 51 movq mm3, [rsi] 54 movq mm4, [rsi+rax] 56 movq mm5, [rsi+rax*2] [all …]
|
D | subpixel_ssse3.asm | 45 push rsi 50 xor rsi, rsi 66 mov rsi, arg(0) ;src_ptr 75 movq xmm0, MMWORD PTR [rsi - 2] ; -2 -1 0 1 2 3 4 5 77 movq xmm2, MMWORD PTR [rsi + 3] ; 3 4 5 6 7 8 9 10 93 lea rsi, [rsi + rax] 110 pop rsi 124 mov rsi, arg(0) ;src_ptr 134 movq xmm0, MMWORD PTR [rsi - 2] ; -2 -1 0 1 2 3 4 5 136 movq xmm1, MMWORD PTR [rsi + 3] ; 3 4 5 6 7 8 9 10 [all …]
|
D | copy_sse2.asm | 28 push rsi 32 mov rsi, arg(0) ;src_ptr 40 movdqu xmm0, XMMWORD PTR [rsi] 41 movdqu xmm1, XMMWORD PTR [rsi + 16] 42 movdqu xmm2, XMMWORD PTR [rsi + rax] 43 movdqu xmm3, XMMWORD PTR [rsi + rax + 16] 45 lea rsi, [rsi+rax*2] 47 movdqu xmm4, XMMWORD PTR [rsi] 48 movdqu xmm5, XMMWORD PTR [rsi + 16] 49 movdqu xmm6, XMMWORD PTR [rsi + rax] [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | vpx_high_subpixel_8t_sse2.asm | 93 mov rsi, arg(0) ;src_ptr 137 movdqu xmm0, [rsi + %1] ;0 138 movdqu xmm1, [rsi + rax + %1] ;1 139 movdqu xmm6, [rsi + rdx * 2 + %1] ;6 140 lea rsi, [rsi + rax] 141 movdqu xmm7, [rsi + rdx * 2 + %1] ;7 142 movdqu xmm2, [rsi + rax + %1] ;2 143 movdqu xmm3, [rsi + rax * 2 + %1] ;3 144 movdqu xmm4, [rsi + rdx + %1] ;4 145 movdqu xmm5, [rsi + rax * 4 + %1] ;5 [all …]
|
D | vpx_subpixel_bilinear_ssse3.asm | 15 mov rsi, arg(0) ;src_ptr 44 lea rsi, [rsi + rax] 51 mov rsi, arg(0) ;src_ptr 82 lea rsi, [rsi + rax] 103 lea rsi, [rsi + rax] 115 push rsi 121 movd xmm0, [rsi] ;load src 122 movd xmm1, [rsi + rax] 129 pop rsi 140 push rsi [all …]
|
D | vpx_high_subpixel_bilinear_sse2.asm | 15 mov rsi, arg(0) ;src_ptr 62 lea rsi, [rsi + 2*rax] 70 mov rsi, arg(0) ;src_ptr 122 lea rsi, [rsi + 2*rax] 168 lea rsi, [rsi + 2*rax] 181 push rsi 187 movq xmm0, [rsi] ;load src 188 movq xmm1, [rsi + 2*rax] 195 pop rsi 207 push rsi [all …]
|
D | vpx_subpixel_bilinear_sse2.asm | 15 mov rsi, arg(0) ;src_ptr 55 lea rsi, [rsi + rax] 62 mov rsi, arg(0) ;src_ptr 99 lea rsi, [rsi + rax] 129 lea rsi, [rsi + rax] 141 push rsi 147 movd xmm0, [rsi] ;load src 148 movd xmm1, [rsi + rax] 155 pop rsi 166 push rsi [all …]
|
D | vpx_subpixel_8t_sse2.asm | 90 mov rsi, arg(0) ;src_ptr 131 movq xmm0, [rsi + %1] ;0 132 movq xmm1, [rsi + rax + %1] ;1 133 movq xmm6, [rsi + rdx * 2 + %1] ;6 134 lea rsi, [rsi + rax] 135 movq xmm7, [rsi + rdx * 2 + %1] ;7 136 movq xmm2, [rsi + rax + %1] ;2 137 movq xmm3, [rsi + rax * 2 + %1] ;3 138 movq xmm4, [rsi + rdx + %1] ;4 139 movq xmm5, [rsi + rax * 4 + %1] ;5 [all …]
|
/external/boringssl/linux-x86_64/crypto/fipsmodule/ |
D | p256-x86_64-asm.S | 60 subq 0(%rsi),%r8 61 sbbq 8(%rsi),%r9 62 sbbq 16(%rsi),%r10 64 sbbq 24(%rsi),%r11 65 leaq .Lpoly(%rip),%rsi 69 addq 0(%rsi),%r8 71 adcq 8(%rsi),%r9 72 adcq 16(%rsi),%r10 74 adcq 24(%rsi),%r11 140 mulq 0(%rsi) [all …]
|
D | aesni-x86_64.S | 45 movups %xmm2,(%rsi) 72 movups %xmm2,(%rsi) 571 movups %xmm2,(%rsi) 575 movups %xmm3,16(%rsi) 577 movups %xmm4,32(%rsi) 579 movups %xmm5,48(%rsi) 581 movups %xmm6,64(%rsi) 583 movups %xmm7,80(%rsi) 585 movups %xmm8,96(%rsi) 587 movups %xmm9,112(%rsi) [all …]
|
D | md5-x86_64.S | 44 leaq (%rsi,%rdx,1),%rdi 56 cmpq %rdi,%rsi 65 movl 0(%rsi),%r10d 71 movl 4(%rsi),%r10d 80 movl 8(%rsi),%r10d 89 movl 12(%rsi),%r10d 98 movl 16(%rsi),%r10d 107 movl 20(%rsi),%r10d 116 movl 24(%rsi),%r10d 125 movl 28(%rsi),%r10d [all …]
|
/external/boringssl/mac-x86_64/crypto/fipsmodule/ |
D | p256-x86_64-asm.S | 57 subq 0(%rsi),%r8 58 sbbq 8(%rsi),%r9 59 sbbq 16(%rsi),%r10 61 sbbq 24(%rsi),%r11 62 leaq L$poly(%rip),%rsi 66 addq 0(%rsi),%r8 68 adcq 8(%rsi),%r9 69 adcq 16(%rsi),%r10 71 adcq 24(%rsi),%r11 131 mulq 0(%rsi) [all …]
|
D | aesni-x86_64.S | 43 movups %xmm2,(%rsi) 70 movups %xmm2,(%rsi) 569 movups %xmm2,(%rsi) 573 movups %xmm3,16(%rsi) 575 movups %xmm4,32(%rsi) 577 movups %xmm5,48(%rsi) 579 movups %xmm6,64(%rsi) 581 movups %xmm7,80(%rsi) 583 movups %xmm8,96(%rsi) 585 movups %xmm9,112(%rsi) [all …]
|
D | md5-x86_64.S | 39 leaq (%rsi,%rdx,1),%rdi 51 cmpq %rdi,%rsi 60 movl 0(%rsi),%r10d 66 movl 4(%rsi),%r10d 75 movl 8(%rsi),%r10d 84 movl 12(%rsi),%r10d 93 movl 16(%rsi),%r10d 102 movl 20(%rsi),%r10d 111 movl 24(%rsi),%r10d 120 movl 28(%rsi),%r10d [all …]
|
/external/boringssl/linux-x86_64/crypto/chacha/ |
D | chacha-x86_64.S | 112 movq %rsi,64+8(%rsp) 115 movq %rsi,%rdi 231 movq 64+8(%rsp),%rsi 252 xorl 0(%rsi),%eax 253 xorl 4(%rsi),%ebx 254 xorl 8(%rsi),%ecx 255 xorl 12(%rsi),%edx 256 xorl 16(%rsi),%r8d 257 xorl 20(%rsi),%r9d 258 xorl 24(%rsi),%r10d [all …]
|
/external/boringssl/mac-x86_64/crypto/chacha/ |
D | chacha-x86_64.S | 105 movq %rsi,64+8(%rsp) 108 movq %rsi,%rdi 224 movq 64+8(%rsp),%rsi 245 xorl 0(%rsi),%eax 246 xorl 4(%rsi),%ebx 247 xorl 8(%rsi),%ecx 248 xorl 12(%rsi),%edx 249 xorl 16(%rsi),%r8d 250 xorl 20(%rsi),%r9d 251 xorl 24(%rsi),%r10d [all …]
|
/external/libaom/libaom/aom_dsp/x86/ |
D | aom_high_subpixel_8t_sse2.asm | 96 mov rsi, arg(0) ;src_ptr 140 movdqu xmm0, [rsi + %1] ;0 141 movdqu xmm1, [rsi + rax + %1] ;1 142 movdqu xmm6, [rsi + rdx * 2 + %1] ;6 143 lea rsi, [rsi + rax] 144 movdqu xmm7, [rsi + rdx * 2 + %1] ;7 145 movdqu xmm2, [rsi + rax + %1] ;2 146 movdqu xmm3, [rsi + rax * 2 + %1] ;3 147 movdqu xmm4, [rsi + rdx + %1] ;4 148 movdqu xmm5, [rsi + rax * 4 + %1] ;5 [all …]
|
D | aom_subpixel_bilinear_ssse3.asm | 18 mov rsi, arg(0) ;src_ptr 47 lea rsi, [rsi + rax] 54 mov rsi, arg(0) ;src_ptr 85 lea rsi, [rsi + rax] 106 lea rsi, [rsi + rax] 118 push rsi 124 movd xmm0, [rsi] ;load src 125 movd xmm1, [rsi + rax] 132 pop rsi 143 push rsi [all …]
|
/external/compiler-rt/lib/tsan/rtl/ |
D | tsan_rtl_amd64.S | 22 push %rsi 24 CFI_REL_OFFSET(%rsi, 0) 67 pop %rsi 79 CFI_RESTORE(%rsi) 102 push %rsi 104 CFI_REL_OFFSET(%rsi, 0) 147 pop %rsi 159 CFI_RESTORE(%rsi) 183 mov %rdi, %rsi 186 mov %rdi, %rsi [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/X86/ |
D | I386-64.s | 49 cmpsb %es:(%rdi), %gs:(%rsi) 53 cmpsl %es:(%rdi), %gs:(%rsi) 57 cmpsq %es:(%rdi), %gs:(%rsi) 61 cmpsw %es:(%rdi), %gs:(%rsi) 97 lodsl %gs:(%rsi), %eax 105 movsb %gs:(%rsi), %es:(%rdi) 141 movsl %gs:(%rsi), %es:(%rdi) 145 movsq %gs:(%rsi), %es:(%rdi) 149 movsw %gs:(%rsi), %es:(%rdi) 249 outsb %gs:(%rsi), %dx [all …]
|
/external/boringssl/win-x86_64/crypto/fipsmodule/ |
D | md5-x86_64.asm | 20 mov QWORD[16+rsp],rsi 24 mov rsi,rdx 46 lea rdi,[rdx*1+rsi] 58 cmp rsi,rdi 67 mov r10d,DWORD[rsi] 73 mov r10d,DWORD[4+rsi] 82 mov r10d,DWORD[8+rsi] 91 mov r10d,DWORD[12+rsi] 100 mov r10d,DWORD[16+rsi] 109 mov r10d,DWORD[20+rsi] [all …]
|
D | p256-x86_64-asm.asm | 43 mov QWORD[16+rsp],rsi 47 mov rsi,rdx 63 sub r8,QWORD[rsi] 64 sbb r9,QWORD[8+rsi] 65 sbb r10,QWORD[16+rsi] 67 sbb r11,QWORD[24+rsi] 68 lea rsi,[$L$poly] 72 add r8,QWORD[rsi] 74 adc r9,QWORD[8+rsi] 75 adc r10,QWORD[16+rsi] [all …]
|
/external/boringssl/win-x86_64/crypto/chacha/ |
D | chacha-x86_64.asm | 55 mov QWORD[16+rsp],rsi 59 mov rsi,rdx 118 mov QWORD[((64+8))+rsp],rsi 121 mov rdi,rsi 237 mov rsi,QWORD[((64+8))+rsp] 258 xor eax,DWORD[rsi] 259 xor ebx,DWORD[4+rsi] 260 xor ecx,DWORD[8+rsi] 261 xor edx,DWORD[12+rsi] 262 xor r8d,DWORD[16+rsi] [all …]
|
/external/boringssl/src/crypto/fipsmodule/md5/asm/ |
D | md5-x86_64.pl | 22 $code .= " mov 0*4(%rsi), %r10d /* (NEXT STEP) X[0] */\n" if ($pos == -1); 29 mov $k_next*4(%rsi),%r10d /* (NEXT STEP) X[$k_next] */ 46 $code .= " mov 1*4(%rsi), %r10d /* (NEXT STEP) X[1] */\n" if ($pos == -1); 54 mov $k_next*4(%rsi),%r10d /* (NEXT STEP) X[$k_next] */ 72 $code .= " mov 5*4(%rsi), %r10d /* (NEXT STEP) X[5] */\n" if ($pos == -1); 76 mov $k_next*4(%rsi),%r10d /* (NEXT STEP) X[$k_next] */ 94 $code .= " mov 0*4(%rsi), %r10d /* (NEXT STEP) X[0] */\n" if ($pos == -1); 103 mov $k_next*4(%rsi),%r10d /* (NEXT STEP) X[$k_next] */ 151 lea (%rsi,%rdx), %rdi # rdi = end 163 cmp %rdi, %rsi # cmp end with ptr [all …]
|