/external/skia/src/opts/ |
D | SkBitmapFilter_opts_SSE2.cpp | 90 __m128i src16 = _mm_unpacklo_epi8(src8, zero); in convolveHorizontally_SSE2() local 91 __m128i mul_hi = _mm_mulhi_epi16(src16, coeff16); in convolveHorizontally_SSE2() 92 __m128i mul_lo = _mm_mullo_epi16(src16, coeff16); in convolveHorizontally_SSE2() 108 src16 = _mm_unpackhi_epi8(src8, zero); in convolveHorizontally_SSE2() 109 mul_hi = _mm_mulhi_epi16(src16, coeff16); in convolveHorizontally_SSE2() 110 mul_lo = _mm_mullo_epi16(src16, coeff16); in convolveHorizontally_SSE2() 140 __m128i src16 = _mm_unpacklo_epi8(src8, zero); in convolveHorizontally_SSE2() local 141 __m128i mul_hi = _mm_mulhi_epi16(src16, coeff16); in convolveHorizontally_SSE2() 142 __m128i mul_lo = _mm_mullo_epi16(src16, coeff16); in convolveHorizontally_SSE2() 148 src16 = _mm_unpackhi_epi8(src8, zero); in convolveHorizontally_SSE2() [all …]
|
/external/libvpx/libvpx/test/ |
D | fdct8x8_test.cc | 224 DECLARE_ALIGNED(16, uint16_t, src16[64]); in RunRoundTripErrorCheck() 236 src16[j] = rnd.Rand16() & mask_; in RunRoundTripErrorCheck() 238 test_input_block[j] = src16[j] - dst16[j]; in RunRoundTripErrorCheck() 269 bit_depth_ == VPX_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; in RunRoundTripErrorCheck() 302 DECLARE_ALIGNED(16, uint16_t, src16[64]); in RunExtremalCheck() 323 src16[j] = mask_; in RunExtremalCheck() 326 src16[j] = 0; in RunExtremalCheck() 329 src16[j] = rnd.Rand8() % 2 ? mask_ : 0; in RunExtremalCheck() 332 test_input_block[j] = src16[j] - dst16[j]; in RunExtremalCheck() 354 bit_depth_ == VPX_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; in RunExtremalCheck() [all …]
|
D | dct32x32_test.cc | 130 DECLARE_ALIGNED(16, uint16_t, src16[kNumCoeffs]); in TEST_P() 142 src16[j] = rnd.Rand16() & mask_; in TEST_P() 144 test_input_block[j] = src16[j] - dst16[j]; in TEST_P() 162 bit_depth_ == VPX_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; in TEST_P() 266 DECLARE_ALIGNED(16, uint16_t, src16[kNumCoeffs]); in TEST_P() 280 src16[j] = rnd.Rand16() & mask_; in TEST_P() 282 in[j] = src16[j] - dst16[j]; in TEST_P() 300 bit_depth_ == VPX_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; in TEST_P()
|
D | fdct4x4_test.cc | 113 DECLARE_ALIGNED(16, uint16_t, src16[kNumCoeffs]); in RunAccuracyCheck() 124 src16[j] = rnd.Rand16() & mask_; in RunAccuracyCheck() 126 test_input_block[j] = src16[j] - dst16[j]; in RunAccuracyCheck() 145 bit_depth_ == VPX_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; in RunAccuracyCheck() 229 DECLARE_ALIGNED(16, uint16_t, src16[kNumCoeffs]); in RunInvAccuracyCheck() 241 src16[j] = rnd.Rand16() & mask_; in RunInvAccuracyCheck() 243 in[j] = src16[j] - dst16[j]; in RunInvAccuracyCheck() 262 bit_depth_ == VPX_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; in RunInvAccuracyCheck()
|
D | dct16x16_test.cc | 344 DECLARE_ALIGNED(16, uint16_t, src16[kNumCoeffs]); in RunAccuracyCheck() 355 src16[j] = rnd.Rand16() & mask_; in RunAccuracyCheck() 357 test_input_block[j] = src16[j] - dst16[j]; in RunAccuracyCheck() 377 bit_depth_ == VPX_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; in RunAccuracyCheck() 520 DECLARE_ALIGNED(16, uint16_t, src16[kNumCoeffs]); in RunInvAccuracyCheck() 534 src16[j] = rnd.Rand16() & mask_; in RunInvAccuracyCheck() 536 in[j] = src16[j] - dst16[j]; in RunInvAccuracyCheck() 557 bit_depth_ == VPX_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; in RunInvAccuracyCheck()
|
D | variance_test.cc | 147 uint16_t *src16 = CONVERT_TO_SHORTPTR(src); in subpel_variance_ref() local 155 const int diff = r - src16[w * y + x]; in subpel_variance_ref() 518 uint16_t *src16 = CONVERT_TO_SHORTPTR(src); in subpel_avg_variance_ref() local 527 const int diff = ((r + sec16[w * y + x] + 1) >> 1) - src16[w * y + x]; in subpel_avg_variance_ref()
|
/external/icu/icu4c/source/test/cintltst/ |
D | custrtrn.c | 101 static const UChar src16[] = { variable 150 u_strToUTF32(u32Target, 0, &u32DestLen, src16, UPRV_LENGTHOF(src16),&err); in Test_strToUTF32() 159 u_strToUTF32(u32Target, UPRV_LENGTHOF(src32)+1, &u32DestLen, src16, UPRV_LENGTHOF(src16),&err); in Test_strToUTF32() 183 u_strToUTF32(NULL,0, &u32DestLen, src16, -1,&err); in Test_strToUTF32() 192 u_strToUTF32(u32Target, UPRV_LENGTHOF(src32), &u32DestLen, src16, -1,&err); in Test_strToUTF32() 330 if(err != U_BUFFER_OVERFLOW_ERROR || uDestLen != UPRV_LENGTHOF(src16)) { in Test_strFromUTF32() 333 (long)uDestLen, (long)UPRV_LENGTHOF(src16), u_errorName(err)); in Test_strFromUTF32() 338 u_strFromUTF32(uTarget, UPRV_LENGTHOF(src16)+1,&uDestLen,src32,UPRV_LENGTHOF(src32),&err); in Test_strFromUTF32() 339 if(err != U_ZERO_ERROR || uDestLen != UPRV_LENGTHOF(src16)) { in Test_strFromUTF32() 342 (long)uDestLen, (long)UPRV_LENGTHOF(src16), u_errorName(err)); in Test_strFromUTF32() [all …]
|
/external/mesa3d/src/gallium/drivers/llvmpipe/ |
D | lp_tile_image.c | 171 const uint16_t *src16 = (const uint16_t *) src; in lp_tiled_to_linear() local 184 untile_4_4_uint16(src16 + src_offset, in lp_tiled_to_linear() 261 const uint16_t *src16 = (const uint16_t *) src; in lp_linear_to_tiled() local 274 tile_4_4_uint16(src16 + src_offset, in lp_linear_to_tiled()
|
/external/libvpx/libvpx/vpx_scale/generic/ |
D | yv12extend.c | 308 const uint16_t *src16 = CONVERT_TO_SHORTPTR(src); in vpx_yv12_copy_y_c() local 311 memcpy(dst16, src16, src_ybc->y_width * sizeof(uint16_t)); in vpx_yv12_copy_y_c() 312 src16 += src_ybc->y_stride; in vpx_yv12_copy_y_c()
|
/external/skia/src/codec/ |
D | SkSwizzler.cpp | 632 const uint16_t* src16 = (const uint16_t*) (src + offset); in SkipLeadingGrayAlphaZerosThen() local 637 while (width > 0 && *src16 == 0x0000) { in SkipLeadingGrayAlphaZerosThen() 640 src16 += deltaSrc / 2; in SkipLeadingGrayAlphaZerosThen() 642 proc(dst32, (const uint8_t*)src16, width, bpp, deltaSrc, 0, ctable); in SkipLeadingGrayAlphaZerosThen()
|
/external/libvpx/libvpx/vp9/encoder/ |
D | vp9_encoder.c | 2480 uint16_t *src16 = CONVERT_TO_SHORTPTR(s->y_buffer); in vp9_write_yuv_rec_frame() local 2483 fwrite(src16, s->y_width, 2, yuv_rec_file); in vp9_write_yuv_rec_frame() 2484 src16 += s->y_stride; in vp9_write_yuv_rec_frame() 2487 src16 = CONVERT_TO_SHORTPTR(s->u_buffer); in vp9_write_yuv_rec_frame() 2491 fwrite(src16, s->uv_width, 2, yuv_rec_file); in vp9_write_yuv_rec_frame() 2492 src16 += s->uv_stride; in vp9_write_yuv_rec_frame() 2495 src16 = CONVERT_TO_SHORTPTR(s->v_buffer); in vp9_write_yuv_rec_frame() 2499 fwrite(src16, s->uv_width, 2, yuv_rec_file); in vp9_write_yuv_rec_frame() 2500 src16 += s->uv_stride; in vp9_write_yuv_rec_frame()
|
/external/valgrind/VEX/priv/ |
D | guest_s390_toIR.c | 7239 IRTemp src16 = newTemp(Ity_D128); in s390_irgen_PFPO() local 7406 assign(src16, get_dpr_pair(4)); /* get source from FPR 4,6 */ in s390_irgen_PFPO() 7407 assign(dst16, binop(Iop_D128toF32, irrm, mkexpr(src16))); in s390_irgen_PFPO() 7410 s390_cc_thunk_put1d128Z(S390_CC_OP_PFPO_128, src16, gr0); in s390_irgen_PFPO()
|