/external/libvpx/libvpx/vp8/common/x86/ |
D | loopfilter_block_sse2_x86_64.asm | 18 movdqa scratch1, %2 ; v2 31 movdqa scratch2, %3 ; save p1 68 movdqa scratch2, %6 ; save hev 72 movdqa scratch1, %1 78 movdqa scratch1, %3 85 movdqa scratch2, %5 90 movdqa scratch1, zero 101 movdqa scratch1, zero 113 movdqa scratch1, zero 199 movdqa xmm0, i0 [all …]
|
D | idctllm_sse2.asm | 128 movdqa xmm0, [rax] 129 movdqa xmm2, [rax+16] 130 movdqa xmm1, [rax+32] 131 movdqa xmm3, [rax+48] 134 movdqa [rax], xmm7 135 movdqa [rax+16], xmm7 136 movdqa [rax+32], xmm7 137 movdqa [rax+48], xmm7 147 movdqa xmm4, xmm0 154 movdqa xmm4, xmm2 [all …]
|
D | loopfilter_sse2.asm | 30 movdqa xmm2, [rdi+2*rax] ; q3 31 movdqa xmm1, [rsi+2*rax] ; q2 32 movdqa xmm4, [rsi+rax] ; q1 33 movdqa xmm5, [rsi] ; q0 49 movdqa [rsp+_q2], xmm1 ; store q2 50 movdqa [rsp+_q1], xmm4 ; store q1 52 movdqa xmm7, [rdx] ;limit 54 movdqa xmm6, xmm1 ; q2 55 movdqa xmm3, xmm4 ; q1 66 movdqa xmm0, xmm5 ; q0 [all …]
|
D | recon_sse2.asm | 41 movdqa [rdi], xmm0 44 movdqa [rdi+rcx], xmm1 45 movdqa [rdi+rcx*2],xmm2 56 movdqa [rdi], xmm3 59 movdqa [rdi+rcx], xmm4 60 movdqa [rdi+rcx*2],xmm5 71 movdqa [rdi], xmm0 74 movdqa [rdi+rcx], xmm1 76 movdqa [rdi+rcx*2], xmm2 86 movdqa [rdi], xmm3 [all …]
|
/external/boringssl/src/crypto/aes/asm/ |
D | vpaes-x86.pl | 162 &movdqa ("xmm7",&QWP($k_inv,$const)); 163 &movdqa ("xmm6",&QWP($k_s0F,$const)); 184 &movdqa ("xmm1","xmm6") 185 &movdqa ("xmm2",&QWP($k_ipt,$const)); 190 &movdqa ("xmm0",&QWP($k_ipt+16,$const)); 202 &movdqa ("xmm4",&QWP($k_sb1,$const)); # 4 : sb1u 203 &movdqa ("xmm0",&QWP($k_sb1+16,$const));# 0 : sb1t 207 &movdqa ("xmm5",&QWP($k_sb2,$const)); # 4 : sb2u 209 &movdqa ("xmm1",&QWP(-0x40,$base,$magic));# .Lk_mc_forward[] 211 &movdqa ("xmm2",&QWP($k_sb2+16,$const));# 2 : sb2t [all …]
|
/external/boringssl/linux-x86_64/crypto/aes/ |
D | vpaes-x86_64.S | 25 movdqa %xmm9,%xmm1 26 movdqa .Lk_ipt(%rip),%xmm2 32 movdqa .Lk_ipt+16(%rip),%xmm0 43 movdqa %xmm13,%xmm4 44 movdqa %xmm12,%xmm0 48 movdqa %xmm15,%xmm5 50 movdqa -64(%r11,%r10,1),%xmm1 52 movdqa (%r11,%r10,1),%xmm4 53 movdqa %xmm14,%xmm2 55 movdqa %xmm0,%xmm3 [all …]
|
D | bsaes-x86_64.S | 14 movdqa (%rax),%xmm8 16 movdqa 80(%r11),%xmm7 34 movdqa 0(%r11),%xmm7 35 movdqa 16(%r11),%xmm8 36 movdqa %xmm5,%xmm9 38 movdqa %xmm3,%xmm10 50 movdqa %xmm1,%xmm9 52 movdqa %xmm15,%xmm10 64 movdqa 32(%r11),%xmm7 65 movdqa %xmm4,%xmm9 [all …]
|
/external/boringssl/mac-x86_64/crypto/aes/ |
D | vpaes-x86_64.S | 25 movdqa %xmm9,%xmm1 26 movdqa L$k_ipt(%rip),%xmm2 32 movdqa L$k_ipt+16(%rip),%xmm0 43 movdqa %xmm13,%xmm4 44 movdqa %xmm12,%xmm0 48 movdqa %xmm15,%xmm5 50 movdqa -64(%r11,%r10,1),%xmm1 52 movdqa (%r11,%r10,1),%xmm4 53 movdqa %xmm14,%xmm2 55 movdqa %xmm0,%xmm3 [all …]
|
D | bsaes-x86_64.S | 12 movdqa (%rax),%xmm8 14 movdqa 80(%r11),%xmm7 32 movdqa 0(%r11),%xmm7 33 movdqa 16(%r11),%xmm8 34 movdqa %xmm5,%xmm9 36 movdqa %xmm3,%xmm10 48 movdqa %xmm1,%xmm9 50 movdqa %xmm15,%xmm10 62 movdqa 32(%r11),%xmm7 63 movdqa %xmm4,%xmm9 [all …]
|
/external/boringssl/mac-x86/crypto/aes/ |
D | vpaes-x86.S | 65 movdqa -48(%ebp),%xmm7 66 movdqa -16(%ebp),%xmm6 73 movdqa %xmm6,%xmm1 74 movdqa (%ebp),%xmm2 79 movdqa 16(%ebp),%xmm0 89 movdqa 32(%ebp),%xmm4 90 movdqa 48(%ebp),%xmm0 94 movdqa 64(%ebp),%xmm5 96 movdqa -64(%ebx,%ecx,1),%xmm1 98 movdqa 80(%ebp),%xmm2 [all …]
|
/external/boringssl/linux-x86/crypto/aes/ |
D | vpaes-x86.S | 66 movdqa -48(%ebp),%xmm7 67 movdqa -16(%ebp),%xmm6 76 movdqa %xmm6,%xmm1 77 movdqa (%ebp),%xmm2 82 movdqa 16(%ebp),%xmm0 92 movdqa 32(%ebp),%xmm4 93 movdqa 48(%ebp),%xmm0 97 movdqa 64(%ebp),%xmm5 99 movdqa -64(%ebx,%ecx,1),%xmm1 101 movdqa 80(%ebp),%xmm2 [all …]
|
/external/boringssl/win-x86/crypto/aes/ |
D | vpaes-x86.asm | 77 movdqa xmm7,[ebp-48] 78 movdqa xmm6,[ebp-16] 84 movdqa xmm1,xmm6 85 movdqa xmm2,[ebp] 90 movdqa xmm0,[16+ebp] 100 movdqa xmm4,[32+ebp] 101 movdqa xmm0,[48+ebp] 105 movdqa xmm5,[64+ebp] 107 movdqa xmm1,[ecx*1+ebx-64] 109 movdqa xmm2,[80+ebp] [all …]
|
/external/llvm/test/MC/X86/ |
D | cfi_def_cfa-crash.s | 22 movdqa %xmm0, -0x80(%rbp) 24 movdqa %xmm1, -0x70(%rbp) 26 movdqa %xmm2, -0x60(%rbp) 28 movdqa %xmm3, -0x50(%rbp) 30 movdqa %xmm4, -0x40(%rbp) 32 movdqa %xmm5, -0x30(%rbp) 34 movdqa %xmm6, -0x20(%rbp) 36 movdqa %xmm7, -0x10(%rbp) 41 movdqa -0x80(%rbp), %xmm0 43 movdqa -0x70(%rbp), %xmm1 [all …]
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | rotate_win.cc | 46 movdqa xmm1, xmm0 in TransposeWx8_SSSE3() 51 movdqa xmm3, xmm2 in TransposeWx8_SSSE3() 57 movdqa xmm5, xmm4 in TransposeWx8_SSSE3() 63 movdqa xmm7, xmm6 in TransposeWx8_SSSE3() 68 movdqa xmm2, xmm0 in TransposeWx8_SSSE3() 69 movdqa xmm3, xmm1 in TransposeWx8_SSSE3() 74 movdqa xmm6, xmm4 in TransposeWx8_SSSE3() 75 movdqa xmm7, xmm5 in TransposeWx8_SSSE3() 82 movdqa xmm4, xmm0 in TransposeWx8_SSSE3() 87 movdqa xmm6, xmm2 in TransposeWx8_SSSE3() [all …]
|
/external/libyuv/files/source/ |
D | rotate_win.cc | 45 movdqa xmm1, xmm0 in TransposeWx8_SSSE3() 50 movdqa xmm3, xmm2 in TransposeWx8_SSSE3() 56 movdqa xmm5, xmm4 in TransposeWx8_SSSE3() 62 movdqa xmm7, xmm6 in TransposeWx8_SSSE3() 67 movdqa xmm2, xmm0 in TransposeWx8_SSSE3() 68 movdqa xmm3, xmm1 in TransposeWx8_SSSE3() 73 movdqa xmm6, xmm4 in TransposeWx8_SSSE3() 74 movdqa xmm7, xmm5 in TransposeWx8_SSSE3() 81 movdqa xmm4, xmm0 in TransposeWx8_SSSE3() 86 movdqa xmm6, xmm2 in TransposeWx8_SSSE3() [all …]
|
/external/boringssl/win-x86_64/crypto/aes/ |
D | bsaes-x86_64.asm | 16 movdqa xmm8,XMMWORD[rax] 18 movdqa xmm7,XMMWORD[80+r11] 36 movdqa xmm7,XMMWORD[r11] 37 movdqa xmm8,XMMWORD[16+r11] 38 movdqa xmm9,xmm5 40 movdqa xmm10,xmm3 52 movdqa xmm9,xmm1 54 movdqa xmm10,xmm15 66 movdqa xmm7,XMMWORD[32+r11] 67 movdqa xmm9,xmm4 [all …]
|
D | vpaes-x86_64.asm | 29 movdqa xmm1,xmm9 30 movdqa xmm2,XMMWORD[$L$k_ipt] 36 movdqa xmm0,XMMWORD[(($L$k_ipt+16))] 47 movdqa xmm4,xmm13 48 movdqa xmm0,xmm12 52 movdqa xmm5,xmm15 54 movdqa xmm1,XMMWORD[((-64))+r10*1+r11] 56 movdqa xmm4,XMMWORD[r10*1+r11] 57 movdqa xmm2,xmm14 59 movdqa xmm3,xmm0 [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | vector-popcnt-128.ll | 12 ; SSE2-NEXT: movdqa %xmm0, %xmm1 16 ; SSE2-NEXT: movdqa {{.*#+}} xmm1 = [3689348814741910323,3689348814741910323] 17 ; SSE2-NEXT: movdqa %xmm0, %xmm2 22 ; SSE2-NEXT: movdqa %xmm0, %xmm1 28 ; SSE2-NEXT: movdqa %xmm1, %xmm0 33 ; SSE3-NEXT: movdqa %xmm0, %xmm1 37 ; SSE3-NEXT: movdqa {{.*#+}} xmm1 = [3689348814741910323,3689348814741910323] 38 ; SSE3-NEXT: movdqa %xmm0, %xmm2 43 ; SSE3-NEXT: movdqa %xmm0, %xmm1 49 ; SSE3-NEXT: movdqa %xmm1, %xmm0 [all …]
|
D | vselect-minmax.ll | 11 ; SSE2-NEXT: movdqa %xmm1, %xmm2 36 ; SSE2-NEXT: movdqa %xmm0, %xmm2 43 ; SSE2-NEXT: movdqa %xmm2, %xmm0 64 ; SSE2-NEXT: movdqa %xmm0, %xmm2 69 ; SSE2-NEXT: movdqa %xmm2, %xmm0 90 ; SSE2-NEXT: movdqa %xmm1, %xmm3 97 ; SSE2-NEXT: movdqa %xmm2, %xmm0 246 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [32768,32768,32768,32768,32768,32768,32768,32768] 247 ; SSE2-NEXT: movdqa %xmm0, %xmm3 274 ; SSE2-NEXT: movdqa %xmm0, %xmm2 [all …]
|
D | vector-rotate-128.ll | 19 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [64,64] 22 ; SSE2-NEXT: movdqa %xmm0, %xmm4 24 ; SSE2-NEXT: movdqa %xmm0, %xmm3 28 ; SSE2-NEXT: movdqa %xmm0, %xmm1 38 ; SSE41-NEXT: movdqa {{.*#+}} xmm2 = [64,64] 40 ; SSE41-NEXT: movdqa %xmm0, %xmm3 43 ; SSE41-NEXT: movdqa %xmm0, %xmm4 46 ; SSE41-NEXT: movdqa %xmm0, %xmm1 85 ; X32-SSE-NEXT: movdqa {{.*#+}} xmm2 = [64,0,64,0] 88 ; X32-SSE-NEXT: movdqa %xmm0, %xmm4 [all …]
|
D | vector-tzcnt-128.ll | 23 ; SSE2-NEXT: movdqa %xmm1, %xmm0 39 ; SSE3-NEXT: movdqa %xmm1, %xmm0 55 ; SSSE3-NEXT: movdqa %xmm1, %xmm0 100 ; SSE2-NEXT: movdqa %xmm1, %xmm0 113 ; SSE3-NEXT: movdqa %xmm1, %xmm0 126 ; SSSE3-NEXT: movdqa %xmm1, %xmm0 162 ; SSE2-NEXT: movdqa %xmm2, %xmm0 166 ; SSE2-NEXT: movdqa {{.*#+}} xmm0 = [858993459,858993459,858993459,858993459] 167 ; SSE2-NEXT: movdqa %xmm2, %xmm3 172 ; SSE2-NEXT: movdqa %xmm2, %xmm0 [all …]
|
/external/libvpx/libvpx/vp8/encoder/x86/ |
D | dct_sse2.asm | 78 movdqa xmm2, xmm0 81 movdqa xmm1, xmm0 87 movdqa xmm3, xmm0 93 movdqa xmm1, xmm0 96 movdqa xmm4, xmm3 111 movdqa xmm2, xmm0 115 movdqa xmm3, xmm0 118 movdqa xmm2, xmm0 122 movdqa xmm5, XMMWORD PTR[GLOBAL(_7)] 124 movdqa xmm3, xmm0 [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | vpx_subpixel_8t_sse2.asm | 21 movdqa xmm7, [rdx] ;load filters 37 movdqa k0k1, xmm0 38 movdqa k2k3, xmm2 39 movdqa k5k4, xmm5 40 movdqa k6k7, xmm6 44 movdqa krd, xmm6 47 movdqa zero, xmm7 67 movdqa xmm1, xmm0 94 movdqa xmm7, [rdx] ;load filters 113 movdqa k0, xmm0 ;store filter factors on stack [all …]
|
/external/libjpeg-turbo/simd/ |
D | jccolext-sse2-64.asm | 126 movdqa xmmF,xmmA 134 movdqa xmmB,xmmA 149 movdqa xmmG,xmmA 159 movdqa xmmD,xmmA 169 movdqa xmmE,xmmA 181 movdqa xmmC,xmmA 185 movdqa xmmB,xmmE 189 movdqa xmmF,xmmD 211 movdqa xmmE,xmmA 217 movdqa xmmF,xmmA [all …]
|
D | jdsample-sse2-64.asm | 258 movdqa xmm4,xmm0 261 movdqa xmm5,xmm1 264 movdqa xmm6,xmm2 279 movdqa XMMWORD [rdx+0*SIZEOF_XMMWORD], xmm1 ; temporarily save 280 movdqa XMMWORD [rdx+1*SIZEOF_XMMWORD], xmm5 ; the intermediate data 281 movdqa XMMWORD [rdi+0*SIZEOF_XMMWORD], xmm2 282 movdqa XMMWORD [rdi+1*SIZEOF_XMMWORD], xmm6 287 movdqa XMMWORD [wk(0)], xmm1 288 movdqa XMMWORD [wk(1)], xmm2 300 movdqa xmm2,xmm1 [all …]
|