/external/boringssl/mac-x86_64/crypto/fipsmodule/ |
D | bsaes-x86_64.S | 988 movdqu (%rcx),%xmm7 997 movdqu (%rcx),%xmm6 1050 movdqu (%rcx),%xmm6 1108 movdqu (%rbx),%xmm14 1111 movdqu 0(%r12),%xmm15 1112 movdqu 16(%r12),%xmm0 1113 movdqu 32(%r12),%xmm1 1114 movdqu 48(%r12),%xmm2 1115 movdqu 64(%r12),%xmm3 1116 movdqu 80(%r12),%xmm4 [all …]
|
D | aesni-x86_64.S | 513 movdqu (%rdi),%xmm2 514 movdqu 16(%rdi),%xmm3 515 movdqu 32(%rdi),%xmm4 516 movdqu 48(%rdi),%xmm5 517 movdqu 64(%rdi),%xmm6 518 movdqu 80(%rdi),%xmm7 519 movdqu 96(%rdi),%xmm8 520 movdqu 112(%rdi),%xmm9 528 movdqu (%rdi),%xmm2 531 movdqu 16(%rdi),%xmm3 [all …]
|
/external/boringssl/linux-x86_64/crypto/fipsmodule/ |
D | bsaes-x86_64.S | 990 movdqu (%rcx),%xmm7 999 movdqu (%rcx),%xmm6 1052 movdqu (%rcx),%xmm6 1111 movdqu (%rbx),%xmm14 1114 movdqu 0(%r12),%xmm15 1115 movdqu 16(%r12),%xmm0 1116 movdqu 32(%r12),%xmm1 1117 movdqu 48(%r12),%xmm2 1118 movdqu 64(%r12),%xmm3 1119 movdqu 80(%r12),%xmm4 [all …]
|
D | aesni-x86_64.S | 514 movdqu (%rdi),%xmm2 515 movdqu 16(%rdi),%xmm3 516 movdqu 32(%rdi),%xmm4 517 movdqu 48(%rdi),%xmm5 518 movdqu 64(%rdi),%xmm6 519 movdqu 80(%rdi),%xmm7 520 movdqu 96(%rdi),%xmm8 521 movdqu 112(%rdi),%xmm9 529 movdqu (%rdi),%xmm2 532 movdqu 16(%rdi),%xmm3 [all …]
|
/external/boringssl/win-x86_64/crypto/fipsmodule/ |
D | bsaes-x86_64.asm | 992 movdqu xmm7,XMMWORD[rcx] 1001 movdqu xmm6,XMMWORD[rcx] 1054 movdqu xmm6,XMMWORD[rcx] 1125 movdqu xmm14,XMMWORD[rbx] 1128 movdqu xmm15,XMMWORD[r12] 1129 movdqu xmm0,XMMWORD[16+r12] 1130 movdqu xmm1,XMMWORD[32+r12] 1131 movdqu xmm2,XMMWORD[48+r12] 1132 movdqu xmm3,XMMWORD[64+r12] 1133 movdqu xmm4,XMMWORD[80+r12] [all …]
|
D | aesni-x86_64.asm | 531 movdqu xmm2,XMMWORD[rdi] 532 movdqu xmm3,XMMWORD[16+rdi] 533 movdqu xmm4,XMMWORD[32+rdi] 534 movdqu xmm5,XMMWORD[48+rdi] 535 movdqu xmm6,XMMWORD[64+rdi] 536 movdqu xmm7,XMMWORD[80+rdi] 537 movdqu xmm8,XMMWORD[96+rdi] 538 movdqu xmm9,XMMWORD[112+rdi] 546 movdqu xmm2,XMMWORD[rdi] 549 movdqu xmm3,XMMWORD[16+rdi] [all …]
|
/external/libvpx/libvpx/vp8/common/x86/ |
D | recon_sse2.asm | 32 movdqu xmm0, [rsi] 37 movdqu xmm1, [rsi+rax] 38 movdqu xmm2, [rsi+rax*2] 50 movdqu xmm3, [rsi] 53 movdqu xmm4, [rsi+rax] 55 movdqu xmm5, [rsi+rax*2] 65 movdqu xmm0, [rsi] 68 movdqu xmm1, [rsi+rax] 70 movdqu xmm2, [rsi+rax*2] 79 movdqu xmm3, [rsi] [all …]
|
D | copy_sse2.asm | 40 movdqu xmm0, XMMWORD PTR [rsi] 41 movdqu xmm1, XMMWORD PTR [rsi + 16] 42 movdqu xmm2, XMMWORD PTR [rsi + rax] 43 movdqu xmm3, XMMWORD PTR [rsi + rax + 16] 47 movdqu xmm4, XMMWORD PTR [rsi] 48 movdqu xmm5, XMMWORD PTR [rsi + 16] 49 movdqu xmm6, XMMWORD PTR [rsi + rax] 50 movdqu xmm7, XMMWORD PTR [rsi + rax + 16] 76 movdqu xmm0, XMMWORD PTR [rsi] 77 movdqu xmm1, XMMWORD PTR [rsi + 16]
|
D | copy_sse3.asm | 102 movdqu xmm0, XMMWORD PTR [src_ptr] 103 movdqu xmm1, XMMWORD PTR [src_ptr + 16] 104 movdqu xmm2, XMMWORD PTR [src_ptr + src_stride] 105 movdqu xmm3, XMMWORD PTR [src_ptr + src_stride + 16] 106 movdqu xmm4, XMMWORD PTR [end_ptr] 107 movdqu xmm5, XMMWORD PTR [end_ptr + 16] 108 movdqu xmm6, XMMWORD PTR [end_ptr + src_stride] 109 movdqu xmm7, XMMWORD PTR [end_ptr + src_stride + 16] 135 movdqu xmm0, XMMWORD PTR [src_ptr] 136 movdqu xmm1, XMMWORD PTR [src_ptr + 16]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | vpx_high_subpixel_8t_sse2.asm | 137 movdqu xmm0, [rsi + %1] ;0 138 movdqu xmm1, [rsi + rax + %1] ;1 139 movdqu xmm6, [rsi + rdx * 2 + %1] ;6 141 movdqu xmm7, [rsi + rdx * 2 + %1] ;7 142 movdqu xmm2, [rsi + rax + %1] ;2 143 movdqu xmm3, [rsi + rax * 2 + %1] ;3 144 movdqu xmm4, [rsi + rdx + %1] ;4 145 movdqu xmm5, [rsi + rax * 4 + %1] ;5 149 movdqu temp, xmm4 160 movdqu xmm5, temp [all …]
|
D | vpx_high_subpixel_bilinear_sse2.asm | 117 movdqu xmm1, [rdi] 120 movdqu [rdi], xmm0 ;store the result 160 movdqu xmm1, [rdi] 161 movdqu xmm3, [rdi + 16] 165 movdqu [rdi], xmm0 ;store the result 166 movdqu [rdi + 16], xmm2 ;store the result 213 movdqu xmm0, [rsi] ;0 214 movdqu xmm1, [rsi + 2*rax] ;1 239 movdqu xmm0, [rsi] ;0 240 movdqu xmm2, [rsi + 16] [all …]
|
D | highbd_variance_impl_sse2.asm | 72 movdqu xmm1, XMMWORD PTR [rsi] 73 movdqu xmm2, XMMWORD PTR [rdi] 89 movdqu xmm3, XMMWORD PTR [rsi+16] 92 movdqu xmm2, XMMWORD PTR [rdi+16] 96 movdqu xmm1, XMMWORD PTR [rsi+rax] 99 movdqu xmm2, XMMWORD PTR [rdi+rdx] 103 movdqu xmm3, XMMWORD PTR [rsi+rax+16] 106 movdqu xmm2, XMMWORD PTR [rdi+rdx+16] 217 movdqu xmm1, XMMWORD PTR [rsi] 218 movdqu xmm2, XMMWORD PTR [rdi] [all …]
|
/external/libyuv/files/source/ |
D | row_win.cc | 277 movdqu [edx], xmm0 278 movdqu [edx + 16], xmm1 331 movdqu xmm0, [eax] 332 movdqu xmm1, [eax + 16] 333 movdqu xmm3, [eax + 32] 341 movdqu [edx + 32], xmm2 344 movdqu [edx], xmm0 348 movdqu [edx + 16], xmm1 350 movdqu [edx + 48], xmm3 370 movdqu xmm0, [eax] [all …]
|
D | scale_win.cc | 102 movdqu xmm0, [eax] in ScaleRowDown2_SSSE3() 103 movdqu xmm1, [eax + 16] in ScaleRowDown2_SSSE3() 108 movdqu [edx], xmm0 in ScaleRowDown2_SSSE3() 134 movdqu xmm0, [eax] in ScaleRowDown2Linear_SSSE3() 135 movdqu xmm1, [eax + 16] in ScaleRowDown2Linear_SSSE3() 142 movdqu [edx], xmm0 in ScaleRowDown2Linear_SSSE3() 169 movdqu xmm0, [eax] in ScaleRowDown2Box_SSSE3() 170 movdqu xmm1, [eax + 16] in ScaleRowDown2Box_SSSE3() 171 movdqu xmm2, [eax + esi] in ScaleRowDown2Box_SSSE3() 172 movdqu xmm3, [eax + esi + 16] in ScaleRowDown2Box_SSSE3() [all …]
|
D | scale_gcc.cc | 166 MEMOPREG(movdqu,0x00,0,3,1,xmm2) // movdqu (%0,%3,1),%%xmm2 in ScaleRowDown2Box_SSSE3() 167 MEMOPREG(movdqu,0x10,0,3,1,xmm3) // movdqu 0x10(%0,%3,1),%%xmm3 in ScaleRowDown2Box_SSSE3() 347 MEMOPREG(movdqu,0x00,0,4,1,xmm2) // movdqu (%0,%4,1),%%xmm2 in ScaleRowDown4Box_SSSE3() 348 MEMOPREG(movdqu,0x10,0,4,1,xmm3) // movdqu 0x10(%0,%4,1),%%xmm3 in ScaleRowDown4Box_SSSE3() 355 MEMOPREG(movdqu,0x00,0,4,2,xmm2) // movdqu (%0,%4,2),%%xmm2 in ScaleRowDown4Box_SSSE3() 356 MEMOPREG(movdqu,0x10,0,4,2,xmm3) // movdqu 0x10(%0,%4,2),%%xmm3 in ScaleRowDown4Box_SSSE3() 361 MEMOPREG(movdqu,0x00,0,3,1,xmm2) // movdqu (%0,%3,1),%%xmm2 in ScaleRowDown4Box_SSSE3() 362 MEMOPREG(movdqu,0x10,0,3,1,xmm3) // movdqu 0x10(%0,%3,1),%%xmm3 in ScaleRowDown4Box_SSSE3() 541 MEMOPREG(movdqu,0x00,0,3,1,xmm7) // movdqu (%0,%3),%%xmm7 in ScaleRowDown34_1_Box_SSSE3() 550 MEMOPREG(movdqu,0x8,0,3,1,xmm7) // movdqu 0x8(%0,%3),%%xmm7 in ScaleRowDown34_1_Box_SSSE3() [all …]
|
/external/boringssl/mac-x86_64/crypto/chacha/ |
D | chacha-x86_64.S | 56 movdqu (%rcx),%xmm1 57 movdqu 16(%rcx),%xmm2 58 movdqu (%r8),%xmm3 233 movdqu 32(%rsi),%xmm0 252 movdqu %xmm0,32(%rdi) 313 movdqu (%rcx),%xmm1 314 movdqu 16(%rcx),%xmm2 315 movdqu (%r8),%xmm3 392 movdqu 0(%rsi),%xmm4 393 movdqu 16(%rsi),%xmm5 [all …]
|
/external/boringssl/linux-x86_64/crypto/chacha/ |
D | chacha-x86_64.S | 57 movdqu (%rcx),%xmm1 58 movdqu 16(%rcx),%xmm2 59 movdqu (%r8),%xmm3 234 movdqu 32(%rsi),%xmm0 253 movdqu %xmm0,32(%rdi) 314 movdqu (%rcx),%xmm1 315 movdqu 16(%rcx),%xmm2 316 movdqu (%r8),%xmm3 393 movdqu 0(%rsi),%xmm4 394 movdqu 16(%rsi),%xmm5 [all …]
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | row_win.cc | 310 movdqu [edx], xmm0 311 movdqu [edx + 16], xmm1 362 movdqu xmm0, [eax] 363 movdqu xmm1, [eax + 16] 364 movdqu xmm3, [eax + 32] 372 movdqu [edx + 32], xmm2 375 movdqu [edx], xmm0 379 movdqu [edx + 16], xmm1 381 movdqu [edx + 48], xmm3 401 movdqu xmm0, [eax] [all …]
|
D | scale_win.cc | 107 movdqu xmm0, [eax] in ScaleRowDown2_SSSE3() 108 movdqu xmm1, [eax + 16] in ScaleRowDown2_SSSE3() 113 movdqu [edx], xmm0 in ScaleRowDown2_SSSE3() 138 movdqu xmm0, [eax] in ScaleRowDown2Linear_SSSE3() 139 movdqu xmm1, [eax + 16] in ScaleRowDown2Linear_SSSE3() 146 movdqu [edx], xmm0 in ScaleRowDown2Linear_SSSE3() 172 movdqu xmm0, [eax] in ScaleRowDown2Box_SSSE3() 173 movdqu xmm1, [eax + 16] in ScaleRowDown2Box_SSSE3() 174 movdqu xmm2, [eax + esi] in ScaleRowDown2Box_SSSE3() 175 movdqu xmm3, [eax + esi + 16] in ScaleRowDown2Box_SSSE3() [all …]
|
D | scale_gcc.cc | 164 MEMOPREG(movdqu,0x00,0,3,1,xmm2) // movdqu (%0,%3,1),%%xmm2 in ScaleRowDown2Box_SSSE3() 165 MEMOPREG(movdqu,0x10,0,3,1,xmm3) // movdqu 0x10(%0,%3,1),%%xmm3 in ScaleRowDown2Box_SSSE3() 332 MEMOPREG(movdqu,0x00,0,4,1,xmm2) // movdqu (%0,%4,1),%%xmm2 in ScaleRowDown4Box_SSSE3() 333 MEMOPREG(movdqu,0x10,0,4,1,xmm3) // movdqu 0x10(%0,%4,1),%%xmm3 in ScaleRowDown4Box_SSSE3() 340 MEMOPREG(movdqu,0x00,0,4,2,xmm2) // movdqu (%0,%4,2),%%xmm2 in ScaleRowDown4Box_SSSE3() 341 MEMOPREG(movdqu,0x10,0,4,2,xmm3) // movdqu 0x10(%0,%4,2),%%xmm3 in ScaleRowDown4Box_SSSE3() 346 MEMOPREG(movdqu,0x00,0,3,1,xmm2) // movdqu (%0,%3,1),%%xmm2 in ScaleRowDown4Box_SSSE3() 347 MEMOPREG(movdqu,0x10,0,3,1,xmm3) // movdqu 0x10(%0,%3,1),%%xmm3 in ScaleRowDown4Box_SSSE3() 518 MEMOPREG(movdqu,0x00,0,3,1,xmm7) // movdqu (%0,%3),%%xmm7 in ScaleRowDown34_1_Box_SSSE3() 527 MEMOPREG(movdqu,0x8,0,3,1,xmm7) // movdqu 0x8(%0,%3),%%xmm7 in ScaleRowDown34_1_Box_SSSE3() [all …]
|
/external/boringssl/win-x86_64/crypto/chacha/ |
D | chacha-x86_64.asm | 74 movdqu xmm1,XMMWORD[rcx] 75 movdqu xmm2,XMMWORD[16+rcx] 76 movdqu xmm3,XMMWORD[r8] 251 movdqu xmm0,XMMWORD[32+rsi] 270 movdqu XMMWORD[32+rdi],xmm0 347 movdqu xmm1,XMMWORD[rcx] 348 movdqu xmm2,XMMWORD[16+rcx] 349 movdqu xmm3,XMMWORD[r8] 426 movdqu xmm4,XMMWORD[rsi] 427 movdqu xmm5,XMMWORD[16+rsi] [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | psubus.ll | 9 ; SSE-NEXT: movdqu (%rdi), %xmm0 11 ; SSE-NEXT: movdqu %xmm0, (%rdi) 34 ; SSE-NEXT: movdqu (%rdi), %xmm0 36 ; SSE-NEXT: movdqu %xmm0, (%rdi) 62 ; SSE-NEXT: movdqu (%rdi), %xmm1 64 ; SSE-NEXT: movdqu %xmm1, (%rdi) 101 ; SSE-NEXT: movdqu (%rdi), %xmm0 103 ; SSE-NEXT: movdqu %xmm0, (%rdi) 126 ; SSE-NEXT: movdqu (%rdi), %xmm0 128 ; SSE-NEXT: movdqu %xmm0, (%rdi) [all …]
|
/external/boringssl/mac-x86/crypto/chacha/ |
D | chacha-x86.S | 395 movdqu (%ebx),%xmm3 402 movdqu (%edx),%xmm7 417 movdqu 16(%edx),%xmm3 704 movdqu -128(%esi),%xmm4 705 movdqu -64(%esi),%xmm5 706 movdqu (%esi),%xmm2 707 movdqu 64(%esi),%xmm7 717 movdqu %xmm4,-128(%edi) 718 movdqu %xmm5,-64(%edi) 719 movdqu %xmm6,(%edi) [all …]
|
/external/boringssl/linux-x86/crypto/chacha/ |
D | chacha-x86.S | 398 movdqu (%ebx),%xmm3 405 movdqu (%edx),%xmm7 420 movdqu 16(%edx),%xmm3 707 movdqu -128(%esi),%xmm4 708 movdqu -64(%esi),%xmm5 709 movdqu (%esi),%xmm2 710 movdqu 64(%esi),%xmm7 720 movdqu %xmm4,-128(%edi) 721 movdqu %xmm5,-64(%edi) 722 movdqu %xmm6,(%edi) [all …]
|
/external/boringssl/src/crypto/fipsmodule/aes/asm/ |
D | vpaes-x86.pl | 192 &movdqu ("xmm5",&QWP(0,$key)); 251 &movdqu ("xmm5",&QWP(0,$key)); 280 &movdqu ("xmm5",&QWP(0,$key)); 353 &movdqu ("xmm0",&QWP(0,$key)); 376 &movdqu ("xmm0",&QWP(0,$inp)); # load key (unaligned) 390 &movdqu (&QWP(0,$key),"xmm0"); 397 &movdqu (&QWP(0,$key),"xmm3"); 440 &movdqu ("xmm0",&QWP(8,$inp)); # load key part 2 (very unaligned) 471 &movdqu ("xmm0",&QWP(16,$inp)); # load key part 2 (unaligned) 520 &movdqu (&QWP(0,$key),"xmm0"); # save last key [all …]
|