/external/libjpeg-turbo/simd/x86_64/ |
D | jcsample-sse2.asm | 100 psrlw xmm6, BYTE_BIT ; xmm6={0xFF 0x00 0xFF 0x00 ..} 130 psrlw xmm2, BYTE_BIT 132 psrlw xmm3, BYTE_BIT 138 psrlw xmm0, 1 139 psrlw xmm1, 1 245 psrlw xmm6, BYTE_BIT ; xmm6={0xFF 0x00 0xFF 0x00 ..} 279 psrlw xmm4, BYTE_BIT 281 psrlw xmm5, BYTE_BIT 288 psrlw xmm4, BYTE_BIT 290 psrlw xmm5, BYTE_BIT [all …]
|
D | jcphuff-sse2.asm | 311 psrlw X0, AL 312 psrlw X1, AL 338 psrlw X0, AL 339 psrlw X1, AL 353 psrlw X0, AL 364 psrlw X0, AL 469 psrlw ONE, 15 484 psrlw X0, AL 485 psrlw X1, AL 522 psrlw X0, AL [all …]
|
/external/libaom/aom_dsp/x86/ |
D | highbd_subpel_variance_impl_sse2.asm | 323 psrlw m1, 4 324 psrlw m0, 4 351 psrlw m1, 4 352 psrlw m0, 4 553 psrlw m1, 4 556 psrlw m0, 4 592 psrlw m4, 4 595 psrlw m0, 4 667 psrlw m1, 4 668 psrlw m0, 4 [all …]
|
D | quantize_ssse3_x86_64.asm | 38 psrlw m5, 15 41 psrlw m0, 1 ; m0 = (m0 + 1) / 2 42 psrlw m1, 1 ; m1 = (m1 + 1) / 2 84 psrlw m5, 15 94 psrlw m5, 15 129 psrlw m8, 1 130 psrlw m13, 1 192 psrlw m5, 15 199 psrlw m5, 15 234 psrlw m14, 1 [all …]
|
/external/libvpx/vpx_dsp/x86/ |
D | highbd_subpel_variance_impl_sse2.asm | 320 psrlw m1, 4 321 psrlw m0, 4 348 psrlw m1, 4 349 psrlw m0, 4 550 psrlw m1, 4 553 psrlw m0, 4 589 psrlw m4, 4 592 psrlw m0, 4 664 psrlw m1, 4 665 psrlw m0, 4 [all …]
|
/external/libjpeg-turbo/simd/i386/ |
D | jcsample-mmx.asm | 103 psrlw mm6, BYTE_BIT ; mm6={0xFF 0x00 0xFF 0x00 ..} 124 psrlw mm2, BYTE_BIT 126 psrlw mm3, BYTE_BIT 132 psrlw mm0, 1 133 psrlw mm1, 1 246 psrlw mm6, BYTE_BIT ; mm6={0xFF 0x00 0xFF 0x00 ..} 270 psrlw mm4, BYTE_BIT 272 psrlw mm5, BYTE_BIT 279 psrlw mm4, BYTE_BIT 281 psrlw mm5, BYTE_BIT [all …]
|
D | jcsample-sse2.asm | 103 psrlw xmm6, BYTE_BIT ; xmm6={0xFF 0x00 0xFF 0x00 ..} 136 psrlw xmm2, BYTE_BIT 138 psrlw xmm3, BYTE_BIT 144 psrlw xmm0, 1 145 psrlw xmm1, 1 259 psrlw xmm6, BYTE_BIT ; xmm6={0xFF 0x00 0xFF 0x00 ..} 296 psrlw xmm4, BYTE_BIT 298 psrlw xmm5, BYTE_BIT 305 psrlw xmm4, BYTE_BIT 307 psrlw xmm5, BYTE_BIT [all …]
|
D | jcphuff-sse2.asm | 318 psrlw X0, AL 319 psrlw X1, AL 348 psrlw X0, AL 349 psrlw X1, AL 363 psrlw X0, AL 374 psrlw X0, AL 478 psrlw ONE, 15 503 psrlw X0, AL 504 psrlw X1, AL 544 psrlw X0, AL [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | vector-popcnt-128.ll | 60 ; SSSE3-NEXT: psrlw $4, %xmm0 76 ; SSE41-NEXT: psrlw $4, %xmm0 162 ; SSSE3-NEXT: psrlw $4, %xmm0 184 ; SSE41-NEXT: psrlw $4, %xmm0 223 ; SSE2-NEXT: psrlw $1, %xmm1 229 ; SSE2-NEXT: psrlw $2, %xmm0 233 ; SSE2-NEXT: psrlw $4, %xmm1 239 ; SSE2-NEXT: psrlw $8, %xmm0 245 ; SSE3-NEXT: psrlw $1, %xmm1 251 ; SSE3-NEXT: psrlw $2, %xmm0 [all …]
|
D | vector-shift-lshr-128.ll | 200 ; SSE2-NEXT: psrlw $8, %xmm0 208 ; SSE2-NEXT: psrlw $4, %xmm0 216 ; SSE2-NEXT: psrlw $2, %xmm0 223 ; SSE2-NEXT: psrlw $1, %xmm0 238 ; SSE41-NEXT: psrlw $8, %xmm4 242 ; SSE41-NEXT: psrlw $4, %xmm1 246 ; SSE41-NEXT: psrlw $2, %xmm1 251 ; SSE41-NEXT: psrlw $1, %xmm1 309 ; X32-SSE-NEXT: psrlw $8, %xmm0 317 ; X32-SSE-NEXT: psrlw $4, %xmm0 [all …]
|
D | vector-idiv-udiv-128.ll | 155 ; SSE-NEXT: psrlw $1, %xmm0 157 ; SSE-NEXT: psrlw $2, %xmm0 176 ; SSE2-NEXT: psrlw $8, %xmm1 179 ; SSE2-NEXT: psrlw $8, %xmm2 181 ; SSE2-NEXT: psrlw $8, %xmm2 184 ; SSE2-NEXT: psrlw $8, %xmm3 186 ; SSE2-NEXT: psrlw $8, %xmm3 189 ; SSE2-NEXT: psrlw $1, %xmm0 192 ; SSE2-NEXT: psrlw $2, %xmm0 201 ; SSE41-NEXT: psrlw $8, %xmm1 [all …]
|
D | vector-bitreverse.ll | 1025 ; SSE2-NEXT: psrlw $7, %xmm2 1048 ; SSE2-NEXT: psrlw $1, %xmm4 1053 ; SSE2-NEXT: psrlw $3, %xmm3 1057 ; SSE2-NEXT: psrlw $5, %xmm0 1072 ; SSSE3-NEXT: psrlw $4, %xmm0 1132 ; SSE2-NEXT: psrlw $1, %xmm3 1137 ; SSE2-NEXT: psrlw $3, %xmm0 1142 ; SSE2-NEXT: psrlw $5, %xmm3 1146 ; SSE2-NEXT: psrlw $7, %xmm1 1162 ; SSSE3-NEXT: psrlw $4, %xmm0 [all …]
|
D | vector-tzcnt-128.ll | 309 ; SSSE3-NEXT: psrlw $4, %xmm2 334 ; SSE41-NEXT: psrlw $4, %xmm2 445 ; X32-SSE-NEXT: psrlw $4, %xmm2 532 ; SSSE3-NEXT: psrlw $4, %xmm2 557 ; SSE41-NEXT: psrlw $4, %xmm2 645 ; X32-SSE-NEXT: psrlw $4, %xmm2 668 ; SSE2-NEXT: psrlw $1, %xmm0 674 ; SSE2-NEXT: psrlw $2, %xmm1 678 ; SSE2-NEXT: psrlw $4, %xmm2 684 ; SSE2-NEXT: psrlw $8, %xmm0 [all …]
|
D | 2012-02-23-mmx-inlineasm.ll | 5 ; CHECK: psrlw %mm0, %mm1 7 call void asm sideeffect "psrlw $0, %mm1", "y,~{dirflag},~{fpsr},~{flags}"(i32 8) nounwind
|
D | 2007-03-24-InlineAsmXConstraint.ll | 8 ; CHECK: psrlw $8, %xmm0 11 tail call void asm sideeffect "psrlw $0, %xmm0", "X,~{dirflag},~{fpsr},~{flags}"( i32 8 )
|
D | vector-rotate-128.ll | 296 ; SSE2-NEXT: psrlw $8, %xmm0 304 ; SSE2-NEXT: psrlw $4, %xmm0 312 ; SSE2-NEXT: psrlw $2, %xmm0 319 ; SSE2-NEXT: psrlw $1, %xmm0 363 ; SSE41-NEXT: psrlw $8, %xmm4 367 ; SSE41-NEXT: psrlw $4, %xmm2 371 ; SSE41-NEXT: psrlw $2, %xmm2 376 ; SSE41-NEXT: psrlw $1, %xmm2 481 ; X32-SSE-NEXT: psrlw $8, %xmm0 489 ; X32-SSE-NEXT: psrlw $4, %xmm0 [all …]
|
D | pr16807.ll | 11 ; CHECK: psrlw 15 ; CHECK: psrlw
|
D | vector-idiv-sdiv-128.ll | 159 ; SSE-NEXT: psrlw $15, %xmm1 184 ; SSE2-NEXT: psrlw $8, %xmm3 189 ; SSE2-NEXT: psrlw $8, %xmm1 193 ; SSE2-NEXT: psrlw $2, %xmm0 198 ; SSE2-NEXT: psrlw $7, %xmm1 209 ; SSE41-NEXT: psrlw $8, %xmm1 213 ; SSE41-NEXT: psrlw $8, %xmm3 217 ; SSE41-NEXT: psrlw $2, %xmm0 222 ; SSE41-NEXT: psrlw $7, %xmm1 463 ; SSE-NEXT: psrlw $15, %xmm2 [all …]
|
/external/libaom/third_party/libyuv/source/ |
D | row_x86.asm | 28 psrlw m2, m2, 8 40 psrlw m0, m0, 8 ; UYVY odd bytes are Y 41 psrlw m1, m1, 8 74 psrlw m4, m4, 8 82 psrlw m2, m0, 8 ; odd bytes 83 psrlw m3, m1, 8
|
D | scale_win.cc | 106 psrlw xmm0, 8 // isolate odd pixels. in ScaleRowDown2_SSSE3() 107 psrlw xmm1, 8 in ScaleRowDown2_SSSE3() 130 psrlw xmm4, 15 in ScaleRowDown2Linear_SSSE3() 165 psrlw xmm4, 15 in ScaleRowDown2Box_SSSE3() 181 psrlw xmm0, 1 in ScaleRowDown2Box_SSSE3() 182 psrlw xmm1, 1 in ScaleRowDown2Box_SSSE3() 332 psrlw xmm0, 8 in ScaleRowDown4_SSSE3() 357 psrlw xmm4, 15 in ScaleRowDown4Box_SSSE3() 388 psrlw xmm0, 4 // /16 for average of 4 * 4 in ScaleRowDown4Box_SSSE3() 574 psrlw xmm0, 2 in ScaleRowDown34_1_Box_SSSE3() [all …]
|
/external/libaom/av1/encoder/x86/ |
D | av1_quantize_ssse3_x86_64.asm | 41 psrlw m5, 15 43 psrlw m1, 1 ; m1 = (m1 + 1) / 2 85 psrlw m8, 1 86 psrlw m13, 1 89 psrlw m0, m3, 2 91 psrlw m0, m3, 1 138 psrlw m14, 1 139 psrlw m13, 1
|
/external/libvpx/third_party/libyuv/source/ |
D | scale_win.cc | 106 psrlw xmm0, 8 // isolate odd pixels. in ScaleRowDown2_SSSE3() 107 psrlw xmm1, 8 in ScaleRowDown2_SSSE3() 130 psrlw xmm4, 15 in ScaleRowDown2Linear_SSSE3() 165 psrlw xmm4, 15 in ScaleRowDown2Box_SSSE3() 181 psrlw xmm0, 1 in ScaleRowDown2Box_SSSE3() 182 psrlw xmm1, 1 in ScaleRowDown2Box_SSSE3() 332 psrlw xmm0, 8 in ScaleRowDown4_SSSE3() 357 psrlw xmm4, 15 in ScaleRowDown4Box_SSSE3() 388 psrlw xmm0, 4 // /16 for average of 4 * 4 in ScaleRowDown4Box_SSSE3() 574 psrlw xmm0, 2 in ScaleRowDown34_1_Box_SSSE3() [all …]
|
/external/libyuv/files/source/ |
D | scale_win.cc | 107 psrlw xmm0, 8 // isolate odd pixels. in ScaleRowDown2_SSSE3() 108 psrlw xmm1, 8 in ScaleRowDown2_SSSE3() 131 psrlw xmm4, 15 in ScaleRowDown2Linear_SSSE3() 166 psrlw xmm4, 15 in ScaleRowDown2Box_SSSE3() 182 psrlw xmm0, 1 in ScaleRowDown2Box_SSSE3() 183 psrlw xmm1, 1 in ScaleRowDown2Box_SSSE3() 333 psrlw xmm0, 8 in ScaleRowDown4_SSSE3() 358 psrlw xmm4, 15 in ScaleRowDown4Box_SSSE3() 389 psrlw xmm0, 4 // /16 for average of 4 * 4 in ScaleRowDown4Box_SSSE3() 575 psrlw xmm0, 2 in ScaleRowDown34_1_Box_SSSE3() [all …]
|
/external/llvm/test/Analysis/CostModel/X86/ |
D | testshiftlshr.ll | 34 ; SSE2-CODEGEN: psrlw 46 ; SSE2-CODEGEN: psrlw 58 ; SSE2-CODEGEN: psrlw 214 ; SSE2-CODEGEN: psrlw 226 ; SSE2-CODEGEN: psrlw 238 ; SSE2-CODEGEN: psrlw 276 ; SSE2-CODEGEN: psrlw $3 290 ; SSE2-CODEGEN: psrlw $3 306 ; SSE2-CODEGEN: psrlw $3 490 ; SSE2-CODEGEN: psrlw $3 [all …]
|
/external/mesa3d/src/mesa/x86/ |
D | read_rgba_span_x86.S | 565 psrlw $SCALE_ADJUST, %mm0 566 psrlw $SCALE_ADJUST, %mm2 600 psrlw $SCALE_ADJUST, %mm0 601 psrlw $SCALE_ADJUST, %mm2 638 psrlw $SCALE_ADJUST, %mm0 639 psrlw $SCALE_ADJUST, %mm2 668 psrlw $SCALE_ADJUST, %mm0
|