/third_party/ffmpeg/tests/checkasm/ |
D | sbrdsp.c | 39 LOCAL_ALIGNED_16(INTFLOAT, dst1, [64 + 256]); in test_sum64x5() 44 memcpy(dst1, dst0, (64 + 256) * sizeof(INTFLOAT)); in test_sum64x5() 46 call_new(dst1); in test_sum64x5() 47 if (!float_near_abs_eps_array(dst0, dst1, EPS, 64 + 256)) in test_sum64x5() 49 bench_new(dst1); in test_sum64x5() 72 LOCAL_ALIGNED_16(INTFLOAT, dst1, [64]); in test_neg_odd_64() 77 memcpy(dst1, dst0, (64) * sizeof(INTFLOAT)); in test_neg_odd_64() 79 call_new(dst1); in test_neg_odd_64() 80 if (!float_near_abs_eps_array(dst0, dst1, EPS, 64)) in test_neg_odd_64() 82 bench_new(dst1); in test_neg_odd_64() [all …]
|
D | llviddsp.c | 48 uint8_t *dst1 = av_mallocz(width); in check_add_bytes() local 55 if (!dst0 || !dst1) in check_add_bytes() 61 call_new(dst1, src1, width); in check_add_bytes() 62 if (memcmp(dst0, dst1, width)) in check_add_bytes() 64 bench_new(dst1, src1, width); in check_add_bytes() 70 av_free(dst1); in check_add_bytes() 76 uint8_t *dst1 = av_mallocz(width); in check_add_median_pred() local 96 call_new(dst1, src1, diff1, width, &A1, &B1); in check_add_median_pred() 97 if (memcmp(dst0, dst1, width) || (A0 != A1) || (B0 != B1)) in check_add_median_pred() 99 bench_new(dst1, src1, diff1, width, &A1, &B1); in check_add_median_pred() [all …]
|
D | fmtconvert.c | 50 LOCAL_ALIGNED(32, float, dst1, [BUF_SIZE]); in checkasm_check_fmtconvert() 62 memset(dst1, 0, sizeof(*dst1) * BUF_SIZE); in checkasm_check_fmtconvert() 73 call_new(dst1, in, scale_arr[i], length[j]); in checkasm_check_fmtconvert() 75 if (!float_near_ulp_array(dst0, dst1, 3, length[j])) { in checkasm_check_fmtconvert() 80 bench_new(dst1, in, scale_arr[i], length[j]); in checkasm_check_fmtconvert() 94 call_new(&c, dst1, in, scale_arr, length[j]); in checkasm_check_fmtconvert() 96 if (!float_near_ulp_array(dst0, dst1, 3, length[j])) { in checkasm_check_fmtconvert() 102 bench_new(&c, dst1, in, scale_arr, length[j]); in checkasm_check_fmtconvert()
|
D | hevc_pel.c | 49 AV_WN32A(dst1 + k, r); \ 73 LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); in checkasm_check_hevc_qpel() 95 int16_t *dstw0 = (int16_t *) dst0, *dstw1 = (int16_t *) dst1; in checkasm_check_hevc_qpel() 117 LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); in checkasm_check_hevc_qpel_uni() 141 …call_new(dst1, sizes[size] * SIZEOF_PIXEL, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], i, j, si… in checkasm_check_hevc_qpel_uni() 142 if (memcmp(dst0, dst1, sizes[size] * sizes[size] * SIZEOF_PIXEL)) in checkasm_check_hevc_qpel_uni() 144 …bench_new(dst1, sizes[size] * SIZEOF_PIXEL, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], i, j, s… in checkasm_check_hevc_qpel_uni() 158 LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); in checkasm_check_hevc_qpel_uni_w() 186 …call_new(dst1, sizes[size] * SIZEOF_PIXEL, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], *denom, … in checkasm_check_hevc_qpel_uni_w() 187 … if (memcmp(dst0, dst1, sizes[size] * sizes[size] * SIZEOF_PIXEL)) in checkasm_check_hevc_qpel_uni_w() [all …]
|
D | llviddspenc.c | 50 LOCAL_ALIGNED_32(uint8_t, dst1, [MAX_STRIDE]); in check_diff_bytes() 60 memset(dst1, 0, MAX_STRIDE); in check_diff_bytes() 69 call_new(dst1, src1, src3, planes[i].w); in check_diff_bytes() 70 if (memcmp(dst0, dst1, planes[i].w)) in check_diff_bytes() 73 bench_new(dst1, src0, src2, planes[4].w); in check_diff_bytes() 81 LOCAL_ALIGNED_32(uint8_t, dst1, [MAX_STRIDE * MAX_HEIGHT]); in check_sub_left_pred() 89 memset(dst1, 0, MAX_STRIDE * MAX_HEIGHT); in check_sub_left_pred() 96 call_new(dst1, src1, planes[i].s, planes[i].w, planes[i].h); in check_sub_left_pred() 97 if (memcmp(dst0, dst1, planes[i].w * planes[i].h)) in check_sub_left_pred() 101 bench_new(dst1, src0, planes[4].s, planes[4].w, planes[4].h); in check_sub_left_pred()
|
D | aacpsdsp.c | 52 LOCAL_ALIGNED_16(INTFLOAT, dst1, [BUF_SIZE]); in test_add_squares() 60 memcpy(dst1, dst0, BUF_SIZE * sizeof(INTFLOAT)); in test_add_squares() 62 call_new(dst1, src, BUF_SIZE); in test_add_squares() 63 if (!float_near_abs_eps_array(dst0, dst1, EPS, BUF_SIZE)) in test_add_squares() 65 bench_new(dst1, src, BUF_SIZE); in test_add_squares() 71 LOCAL_ALIGNED_16(INTFLOAT, dst1, [BUF_SIZE], [2]); in test_mul_pair_single() 81 call_new(dst1, src0, src1, BUF_SIZE); in test_mul_pair_single() 82 if (!float_near_abs_eps_array((float *)dst0, (float *)dst1, EPS, BUF_SIZE * 2)) in test_mul_pair_single() 84 bench_new(dst1, src0, src1, BUF_SIZE); in test_mul_pair_single() 90 LOCAL_ALIGNED_16(INTFLOAT, dst1, [BUF_SIZE], [2]); in test_hybrid_analysis() [all …]
|
D | vf_colorspace.c | 72 uint8_t *dst0[3] = { dst0_y, dst0_u, dst0_v }, *dst1[3] = { dst1_y, dst1_u, dst1_v }; in check_yuv2yuv() local 106 call_new(dst1, (ptrdiff_t[3]) { y_dst_stride, uv_dst_stride, uv_dst_stride }, in check_yuv2yuv() 109 if (memcmp(dst0[0], dst1[0], y_dst_stride * H) || in check_yuv2yuv() 110 memcmp(dst0[1], dst1[1], uv_dst_stride * H >> ss_h) || in check_yuv2yuv() 111 memcmp(dst0[2], dst1[2], uv_dst_stride * H >> ss_h)) { in check_yuv2yuv() 140 int16_t *dst0[3] = { dst0_y, dst0_u, dst0_v }, *dst1[3] = { dst1_y, dst1_u, dst1_v }; in check_yuv2rgb() local 169 call_new(dst1, W, src, in check_yuv2rgb() 172 if (memcmp(dst0[0], dst1[0], W * H * sizeof(int16_t)) || in check_yuv2rgb() 173 memcmp(dst0[1], dst1[1], W * H * sizeof(int16_t)) || in check_yuv2rgb() 174 memcmp(dst0[2], dst1[2], W * H * sizeof(int16_t))) { in check_yuv2rgb() [all …]
|
D | hevc_sao.c | 73 LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); in check_sao_band() 88 memset(dst1, 0, BUF_SIZE); in check_sao_band() 92 call_new(dst1, src1, stride, stride, offset_val, left_class, block_size, block_size); in check_sao_band() 93 if (memcmp(dst0, dst1, BUF_SIZE)) in check_sao_band() 95 bench_new(dst1, src1, stride, stride, offset_val, left_class, block_size, block_size); in check_sao_band() 104 LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); in check_sao_edge() 120 memset(dst1, 0, BUF_SIZE); in check_sao_edge() 124 call_new(dst1, src1 + offset, stride, offset_val, eo, block_size, block_size); in check_sao_edge() 125 if (memcmp(dst0, dst1, BUF_SIZE)) in check_sao_edge() 127 bench_new(dst1, src1 + offset, stride, offset_val, eo, block_size, block_size); in check_sao_edge()
|
D | huffyuvdsp.c | 43 uint16_t *dst1 = av_mallocz(width * sizeof(uint16_t)); in check_add_int16() local 47 if (!src0 || !src1 || !dst0 || !dst1) in check_add_int16() 55 call_new(dst1, src1, mask, width); in check_add_int16() 56 if (memcmp(dst0, dst1, width * sizeof(uint16_t))) in check_add_int16() 58 bench_new(dst1, src1, mask, width); in check_add_int16() 64 av_free(dst1); in check_add_int16()
|
D | sw_scale.c | 82 LOCAL_ALIGNED_8(uint8_t, dst1, [LARGEST_INPUT_SIZE]); in check_yuv2yuvX() 112 memset(dst1, 0, LARGEST_INPUT_SIZE * sizeof(dst1[0])); in check_yuv2yuvX() 121 … call_new((const int16_t*)vFilterData, filter_sizes[fsi], src, dst1, dstW - osi, dither, osi); in check_yuv2yuvX() 122 if (memcmp(dst0, dst1, LARGEST_INPUT_SIZE * sizeof(dst0[0]))) in check_yuv2yuvX() 125 … bench_new((const int16_t*)vFilterData, filter_sizes[fsi], src, dst1, dstW - osi, dither, osi); in check_yuv2yuvX() 158 LOCAL_ALIGNED_32(uint32_t, dst1, [SRC_PIXELS]); in check_hscale() 217 memset(dst1, 0, SRC_PIXELS * sizeof(dst1[0])); in check_hscale() 220 call_new(NULL, dst1, SRC_PIXELS, src, filter, filterPos, width); in check_hscale() 221 if (memcmp(dst0, dst1, SRC_PIXELS * sizeof(dst0[0]))) in check_hscale()
|
D | h264dsp.c | 208 uint8_t *dst1 = dst1_base + align; in check_idct() local 216 memcpy(dst1, dst, sz * PIXEL_STRIDE); in check_idct() 219 call_new(dst1, subcoef1, PIXEL_STRIDE); in check_idct() 220 if (memcmp(dst0, dst1, sz * PIXEL_STRIDE) || in check_idct() 223 bench_new(dst1, subcoef1, sz * SIZEOF_PIXEL); in check_idct() 236 LOCAL_ALIGNED_16(uint8_t, dst1, [16 * 16 * 2]); in check_idct_multiple() 305 memcpy(dst1, dst_full, 16 * 16 * SIZEOF_PIXEL); in check_idct_multiple() 307 call_new(dst1, block_offset, coef1, 16 * SIZEOF_PIXEL, nnzc); in check_idct_multiple() 308 if (memcmp(dst0, dst1, 16 * 16 * SIZEOF_PIXEL) || in check_idct_multiple() 311 bench_new(dst1, block_offset, coef1, 16 * SIZEOF_PIXEL, nnzc); in check_idct_multiple() [all …]
|
D | pixblockdsp.c | 58 call_new(dst1 + dst_offset, src11 + src_offset, 8); \ 59 if (memcmp(src10, src11, BUF_SIZE)|| memcmp(dst0, dst1, BUF_SIZE)) \ 61 bench_new(dst1 + dst_offset, src11 + src_offset, 8); \ 75 call_new(dst1 + dst_offset, src11 + src_offset, src21 + src_offset, 8); \ 76 …memcmp(src10, src11, BUF_SIZE) || memcmp(src20, src21, BUF_SIZE) || memcmp(dst0, dst1, BUF_SIZE)) \ 78 bench_new(dst1 + dst_offset, src11 + src_offset, src21 + src_offset, 8); \ 91 uint16_t *dst1 = (uint16_t *)dst1_; in checkasm_check_pixblockdsp() local
|
D | audiodsp.c | 89 LOCAL_ALIGNED(32, int32_t, dst1, [MAX_SIZE]); in checkasm_check_audiodsp() 110 call_new(dst1, src, min, max, len); in checkasm_check_audiodsp() 111 if (memcmp(dst0, dst1, len * sizeof(*dst0))) in checkasm_check_audiodsp() 113 bench_new(dst1, src, min, max, MAX_SIZE); in checkasm_check_audiodsp() 119 LOCAL_ALIGNED(32, float, dst1, [MAX_SIZE]); in checkasm_check_audiodsp() 138 call_new(dst1, src, len, min, max); in checkasm_check_audiodsp() 140 if (!float_near_ulp_array(dst0, dst1, 3, len)) in checkasm_check_audiodsp() 143 bench_new(dst1, src, MAX_SIZE, min, max); in checkasm_check_audiodsp()
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/intrinsics/sse/ |
D | MatMul_Sse.c | 38 __m128 dst1 = _mm_setzero_ps(); in MatrixMultiplyWinograd() local 54 dst1 = _mm_add_ps(dst1, tmp1); in MatrixMultiplyWinograd() 61 _mm_storeu_ps(matrix_c, dst1); in MatrixMultiplyWinograd() 71 __m128 dst1 = _mm_setzero_ps(); in MatrixMultiplyWinograd() local 81 dst1 = _mm_add_ps(dst1, tmp1); in MatrixMultiplyWinograd() 86 _mm_storeu_ps(matrix_c, dst1); in MatrixMultiplyWinograd() 117 …__m128 dst1 = _mm_setzero_ps(), dst2 = _mm_setzero_ps(), dst3 = _mm_setzero_ps(), dst4 = _mm_setze… in MatmulFloatSse64Opt() local 125 dst1 = _mm_add_ps(dst1, tmp1), dst2 = _mm_add_ps(dst2, tmp2); in MatmulFloatSse64Opt() 136 DoBiasBlock8(bias_d, &dst1, &dst2, &dst3, &dst4, &dst5, &dst6, &dst7, &dst8); in MatmulFloatSse64Opt() 140 ActBlock8(&dst1, &dst2, &dst3, &dst4, &dst5, &dst6, &dst7, &dst8, act_type); in MatmulFloatSse64Opt() [all …]
|
D | ConvDwFp32Row_sse.c | 31 __m128 dst1 = _mm_loadu_ps(output_ptr); in ConvDwFp32Row() local 43 dst1 = MS_MLAQ_F32(dst1, w1, in1); in ConvDwFp32Row() 47 _mm_storeu_ps(output_ptr, dst1); in ConvDwFp32Row() 56 __m128 dst1 = _mm_loadu_ps(output_ptr); in ConvDwFp32Row() local 62 dst1 = MS_MLAQ_F32(dst1, w1, in1); in ConvDwFp32Row() 64 _mm_storeu_ps(output_ptr, dst1); in ConvDwFp32Row() 71 __m128 dst1 = _mm_loadu_ps(output_ptr); in ConvDwFp32Row() local 74 dst1 = MS_MLAQ_F32(dst1, w1, in1); in ConvDwFp32Row() 75 _mm_storeu_ps(output_ptr, dst1); in ConvDwFp32Row()
|
D | sse_common.h | 88 static inline void WriteCol1(float **dst, __m128 *dst1, __m128 *dst2, __m128 *dst3, __m128 *dst4, _… in WriteCol1() argument 90 _mm_store_ss(*dst, *dst1); in WriteCol1() 107 static inline void WriteCol2(float **dst, __m128 *dst1, __m128 *dst2, __m128 *dst3, __m128 *dst4, _… in WriteCol2() argument 109 _mm_store_ss(*dst, *dst1); in WriteCol2() 110 *dst1 = _mm_shuffle_ps(*dst1, *dst1, _MM_SHUFFLE(0, 3, 2, 1)); in WriteCol2() 111 _mm_store_ss(*dst, *dst1); in WriteCol2() 132 static inline void WriteCol2Opt(float **dst, __m128 *dst1, __m128 *dst2, __m128 *dst3, __m128 *dst4… in WriteCol2Opt() argument 134 _mm_store_ss(*dst, *dst1); in WriteCol2Opt() 135 *dst1 = _mm_shuffle_ps(*dst1, *dst1, _MM_SHUFFLE(0, 3, 2, 1)); in WriteCol2Opt() 136 _mm_store_ss(*dst + 1, *dst1); in WriteCol2Opt() [all …]
|
D | TiledC4MatMulFp32.c | 20 static inline void TiledC4MatmulFp32_Transfer(__m128 *dst1, __m128 *dst2, __m128 *dst3, __m128 *dst… in TiledC4MatmulFp32_Transfer() argument 23 *dst1 = _mm_add_ps(*dst1, _mm_mul_ps(weight, _mm_set_ps1(v1))); in TiledC4MatmulFp32_Transfer() 54 __m128 dst1 = _mm_mul_ps(weight_data[0], _mm_set_ps1(MS_F32X4_GETI(src1, 0))); in TiledC4MatmulFp32() local 59 TiledC4MatmulFp32_Transfer(&dst1, &dst2, &dst3, &dst4, weight_data[j], MS_F32X4_GETI(src1, j), in TiledC4MatmulFp32() 80 dst1 = _mm_add_ps(dst1, _mm_mul_ps(weight_data[0], _mm_set_ps1(MS_F32X4_GETI(src1, 0)))); in TiledC4MatmulFp32() 86 … TiledC4MatmulFp32_Transfer(&dst1, &dst2, &dst3, &dst4, weight_data[1], MS_F32X4_GETI(src1, 1), in TiledC4MatmulFp32() 93 … TiledC4MatmulFp32_Transfer(&dst1, &dst2, &dst3, &dst4, weight_data[2], MS_F32X4_GETI(src1, 2), in TiledC4MatmulFp32() 96 dst1 = _mm_add_ps(dst1, _mm_mul_ps(weight_data[3], _mm_set_ps1(MS_F32X4_GETI(src1, 3)))); in TiledC4MatmulFp32() 124 dst1 = _mm_add_ps(dst1, _mm_mul_ps(weight_data[0], _mm_set_ps1(MS_F32X4_GETI(src1, 0)))); in TiledC4MatmulFp32() 130 TiledC4MatmulFp32_Transfer(&dst1, &dst2, &dst3, &dst4, weight_data[1], MS_F32X4_GETI(src1, 1), in TiledC4MatmulFp32() [all …]
|
/third_party/flutter/skia/third_party/externals/sdl/src/video/ |
D | SDL_blit_A.c | 185 __m64 src1, src2, dst1, dst2, lmask, hmask, dsta; in BlitRGBtoRGBSurfaceAlpha128MMX() local 202 dst1 = *(__m64 *) dstp; /* 2 x dst -> dst1(ARGBARGB) */ in BlitRGBtoRGBSurfaceAlpha128MMX() 203 dst2 = dst1; /* 2 x dst -> dst2(ARGBARGB) */ in BlitRGBtoRGBSurfaceAlpha128MMX() 213 dst1 = _mm_and_si64(dst1, src1); /* src & dst -> dst1 */ in BlitRGBtoRGBSurfaceAlpha128MMX() 214 dst1 = _mm_and_si64(dst1, lmask); /* dst1 & !mask -> dst1 */ in BlitRGBtoRGBSurfaceAlpha128MMX() 215 dst1 = _mm_add_pi32(dst1, src2); /* src2 + dst1 -> dst1 */ in BlitRGBtoRGBSurfaceAlpha128MMX() 216 dst1 = _mm_or_si64(dst1, dsta); /* dsta(full alpha) | dst1 -> dst1 */ in BlitRGBtoRGBSurfaceAlpha128MMX() 218 *(__m64 *) dstp = dst1; /* dst1 -> 2 x dst pixels */ in BlitRGBtoRGBSurfaceAlpha128MMX() 250 __m64 src1, src2, dst1, dst2, mm_alpha, mm_zero, dsta; in BlitRGBtoRGBSurfaceAlphaMMX() local 271 dst1 = _mm_cvtsi32_si64(*dstp); /* dst(ARGB) -> dst1 (0000ARGB) */ in BlitRGBtoRGBSurfaceAlphaMMX() [all …]
|
/third_party/boost/libs/multi_index/test/ |
D | test_node_handling.cpp | 297 Src& src,Dst& dst0,Dst& dst1,Dst& /* dst2 */,Dst& /* dst3 */, in test_transfer() argument 302 test_transfer(src,src.begin(),dst1,dst1.find(*src.begin())); in test_transfer() 303 test_transfer(src,src.begin(),dst1,dst1.find(*src.begin())); in test_transfer() 309 Src& src,Dst& dst0,Dst& dst1,Dst& /* dst2 */,Dst& /* dst3 */, in test_transfer() argument 314 test_transfer(src,*src.begin(),dst1,dst1.begin()); in test_transfer() 315 test_transfer(src,*src.begin(),dst1,dst1.begin()); in test_transfer() 316 test_transfer(src, -1,dst1,dst1.begin()); in test_transfer() 322 Src& src,Dst& dst0,Dst& dst1,Dst& dst2,Dst& dst3, in test_transfer() argument 327 test_transfer(src,*src.begin(),dst1); in test_transfer() 328 test_transfer(src,*src.begin(),dst1); in test_transfer() [all …]
|
/third_party/flutter/skia/third_party/externals/libwebp/src/dsp/ |
D | lossless_msa.c | 25 v16u8 src0, src1, src2, src3, dst0, dst1, dst2; \ 27 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \ 29 ST_UB2(dst0, dst1, pdst, 16); \ 35 v16u8 src0, src1, src2, dst0, dst1, dst2; \ 37 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \ 39 ST_UB2(dst0, dst1, pdst, 16); \ 46 v16u8 src0, src1, src2 = { 0 }, dst0, dst1; \ 48 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \ 50 pix_d = __msa_copy_s_d((v2i64)dst1, 0); \ 81 #define TRANSFORM_COLOR_INVERSE_8(src0, src1, dst0, dst1, \ argument [all …]
|
/third_party/skia/third_party/externals/libwebp/src/dsp/ |
D | lossless_msa.c | 25 v16u8 src0, src1, src2, src3, dst0, dst1, dst2; \ 27 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \ 29 ST_UB2(dst0, dst1, pdst, 16); \ 35 v16u8 src0, src1, src2, dst0, dst1, dst2; \ 37 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \ 39 ST_UB2(dst0, dst1, pdst, 16); \ 46 v16u8 src0, src1, src2 = { 0 }, dst0, dst1; \ 48 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \ 50 pix_d = __msa_copy_s_d((v2i64)dst1, 0); \ 81 #define TRANSFORM_COLOR_INVERSE_8(src0, src1, dst0, dst1, \ argument [all …]
|
/third_party/mesa3d/src/intel/compiler/ |
D | test_fs_saturate_propagation.cpp | 117 fs_reg dst1 = v->vgrf(glsl_type::float_type); in TEST_F() local 121 set_saturate(true, bld.MOV(dst1, dst0)); in TEST_F() 152 fs_reg dst1 = v->vgrf(glsl_type::float_type); in TEST_F() local 157 set_saturate(true, bld.MOV(dst1, dst0)); in TEST_F() 190 fs_reg dst1 = v->vgrf(glsl_type::float_type); in TEST_F() local 195 set_saturate(true, bld.MOV(dst1, dst0)); in TEST_F() 225 fs_reg dst1 = v->vgrf(glsl_type::float_type); in TEST_F() local 229 set_saturate(true, bld.MOV(dst1, dst0)); in TEST_F() 259 fs_reg dst1 = v->vgrf(glsl_type::float_type); in TEST_F() local 264 set_saturate(true, bld.MOV(dst1, dst0)); in TEST_F() [all …]
|
/third_party/ffmpeg/libavcodec/mips/ |
D | hevc_mc_bi_msa.c | 74 v8i16 dst0, dst1, dst2, dst3; in hevc_bi_copy_4w_msa() local 96 ILVRL_B2_SH(zero, src0, dst0, dst1); in hevc_bi_copy_4w_msa() 97 SLLI_2V(dst0, dst1, 6); in hevc_bi_copy_4w_msa() 98 HEVC_BI_RND_CLIP2_MAX_SATU(in0, in1, dst0, dst1, 7, dst0, dst1); in hevc_bi_copy_4w_msa() 99 dst0 = (v8i16) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_bi_copy_4w_msa() 117 ILVRL_B2_SH(zero, src0, dst0, dst1); in hevc_bi_copy_4w_msa() 119 SLLI_4V(dst0, dst1, dst2, dst3, 6); in hevc_bi_copy_4w_msa() 120 HEVC_BI_RND_CLIP4_MAX_SATU(in0, in1, in2, in3, dst0, dst1, dst2, in hevc_bi_copy_4w_msa() 121 dst3, 7, dst0, dst1, dst2, dst3); in hevc_bi_copy_4w_msa() 122 PCKEV_B2_SH(dst1, dst0, dst3, dst2, dst0, dst1); in hevc_bi_copy_4w_msa() [all …]
|
D | hevc_mc_uniw_msa.c | 70 v8i16 dst0, dst1, dst2, dst3, offset_vec; in hevc_uniwgt_copy_4w_msa() local 97 ILVRL_B2_SH(zero, src0, dst0, dst1); in hevc_uniwgt_copy_4w_msa() 98 SLLI_2V(dst0, dst1, 6); in hevc_uniwgt_copy_4w_msa() 99 HEVC_UNIW_RND_CLIP2_MAX_SATU_H(dst0, dst1, weight_vec, offset_vec, in hevc_uniwgt_copy_4w_msa() 100 rnd_vec, dst0, dst1); in hevc_uniwgt_copy_4w_msa() 101 out0 = (v16u8) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_uniwgt_copy_4w_msa() 111 ILVRL_B2_SH(zero, src0, dst0, dst1); in hevc_uniwgt_copy_4w_msa() 113 SLLI_4V(dst0, dst1, dst2, dst3, 6); in hevc_uniwgt_copy_4w_msa() 114 HEVC_UNIW_RND_CLIP4_MAX_SATU_H(dst0, dst1, dst2, dst3, weight_vec, in hevc_uniwgt_copy_4w_msa() 115 offset_vec, rnd_vec, dst0, dst1, in hevc_uniwgt_copy_4w_msa() [all …]
|
/third_party/ffmpeg/libavfilter/x86/ |
D | vf_spp.c | 38 #define REQUANT_CORE(dst0, dst1, dst2, dst3, src0, src1, src2, src3) \ in hardthresh_mmx() argument 77 "movq %%mm7, " #dst1 " \n" \ in hardthresh_mmx() 109 #define REQUANT_CORE(dst0, dst1, dst2, dst3, src0, src1, src2, src3) \ in softthresh_mmx() argument 157 "movq %%mm7, " #dst1 " \n" \ in softthresh_mmx() 186 uint8_t *dst1 = dst; in store_slice_mmx() local 211 : "+r" (src1), "+r"(dst1) in store_slice_mmx()
|