Home
last modified time | relevance | path

Searched refs:dst1 (Results 1 – 25 of 174) sorted by relevance

1234567

/third_party/ffmpeg/tests/checkasm/
Dsbrdsp.c39 LOCAL_ALIGNED_16(INTFLOAT, dst1, [64 + 256]); in test_sum64x5()
44 memcpy(dst1, dst0, (64 + 256) * sizeof(INTFLOAT)); in test_sum64x5()
46 call_new(dst1); in test_sum64x5()
47 if (!float_near_abs_eps_array(dst0, dst1, EPS, 64 + 256)) in test_sum64x5()
49 bench_new(dst1); in test_sum64x5()
72 LOCAL_ALIGNED_16(INTFLOAT, dst1, [64]); in test_neg_odd_64()
77 memcpy(dst1, dst0, (64) * sizeof(INTFLOAT)); in test_neg_odd_64()
79 call_new(dst1); in test_neg_odd_64()
80 if (!float_near_abs_eps_array(dst0, dst1, EPS, 64)) in test_neg_odd_64()
82 bench_new(dst1); in test_neg_odd_64()
[all …]
Dllviddsp.c48 uint8_t *dst1 = av_mallocz(width); in check_add_bytes() local
55 if (!dst0 || !dst1) in check_add_bytes()
61 call_new(dst1, src1, width); in check_add_bytes()
62 if (memcmp(dst0, dst1, width)) in check_add_bytes()
64 bench_new(dst1, src1, width); in check_add_bytes()
70 av_free(dst1); in check_add_bytes()
76 uint8_t *dst1 = av_mallocz(width); in check_add_median_pred() local
96 call_new(dst1, src1, diff1, width, &A1, &B1); in check_add_median_pred()
97 if (memcmp(dst0, dst1, width) || (A0 != A1) || (B0 != B1)) in check_add_median_pred()
99 bench_new(dst1, src1, diff1, width, &A1, &B1); in check_add_median_pred()
[all …]
Dfmtconvert.c50 LOCAL_ALIGNED(32, float, dst1, [BUF_SIZE]); in checkasm_check_fmtconvert()
62 memset(dst1, 0, sizeof(*dst1) * BUF_SIZE); in checkasm_check_fmtconvert()
73 call_new(dst1, in, scale_arr[i], length[j]); in checkasm_check_fmtconvert()
75 if (!float_near_ulp_array(dst0, dst1, 3, length[j])) { in checkasm_check_fmtconvert()
80 bench_new(dst1, in, scale_arr[i], length[j]); in checkasm_check_fmtconvert()
94 call_new(&c, dst1, in, scale_arr, length[j]); in checkasm_check_fmtconvert()
96 if (!float_near_ulp_array(dst0, dst1, 3, length[j])) { in checkasm_check_fmtconvert()
102 bench_new(&c, dst1, in, scale_arr, length[j]); in checkasm_check_fmtconvert()
Dhevc_pel.c49 AV_WN32A(dst1 + k, r); \
73 LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); in checkasm_check_hevc_qpel()
95 int16_t *dstw0 = (int16_t *) dst0, *dstw1 = (int16_t *) dst1; in checkasm_check_hevc_qpel()
117 LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); in checkasm_check_hevc_qpel_uni()
141 …call_new(dst1, sizes[size] * SIZEOF_PIXEL, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], i, j, si… in checkasm_check_hevc_qpel_uni()
142 if (memcmp(dst0, dst1, sizes[size] * sizes[size] * SIZEOF_PIXEL)) in checkasm_check_hevc_qpel_uni()
144 …bench_new(dst1, sizes[size] * SIZEOF_PIXEL, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], i, j, s… in checkasm_check_hevc_qpel_uni()
158 LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); in checkasm_check_hevc_qpel_uni_w()
186 …call_new(dst1, sizes[size] * SIZEOF_PIXEL, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], *denom, … in checkasm_check_hevc_qpel_uni_w()
187 … if (memcmp(dst0, dst1, sizes[size] * sizes[size] * SIZEOF_PIXEL)) in checkasm_check_hevc_qpel_uni_w()
[all …]
Dllviddspenc.c50 LOCAL_ALIGNED_32(uint8_t, dst1, [MAX_STRIDE]); in check_diff_bytes()
60 memset(dst1, 0, MAX_STRIDE); in check_diff_bytes()
69 call_new(dst1, src1, src3, planes[i].w); in check_diff_bytes()
70 if (memcmp(dst0, dst1, planes[i].w)) in check_diff_bytes()
73 bench_new(dst1, src0, src2, planes[4].w); in check_diff_bytes()
81 LOCAL_ALIGNED_32(uint8_t, dst1, [MAX_STRIDE * MAX_HEIGHT]); in check_sub_left_pred()
89 memset(dst1, 0, MAX_STRIDE * MAX_HEIGHT); in check_sub_left_pred()
96 call_new(dst1, src1, planes[i].s, planes[i].w, planes[i].h); in check_sub_left_pred()
97 if (memcmp(dst0, dst1, planes[i].w * planes[i].h)) in check_sub_left_pred()
101 bench_new(dst1, src0, planes[4].s, planes[4].w, planes[4].h); in check_sub_left_pred()
Daacpsdsp.c52 LOCAL_ALIGNED_16(INTFLOAT, dst1, [BUF_SIZE]); in test_add_squares()
60 memcpy(dst1, dst0, BUF_SIZE * sizeof(INTFLOAT)); in test_add_squares()
62 call_new(dst1, src, BUF_SIZE); in test_add_squares()
63 if (!float_near_abs_eps_array(dst0, dst1, EPS, BUF_SIZE)) in test_add_squares()
65 bench_new(dst1, src, BUF_SIZE); in test_add_squares()
71 LOCAL_ALIGNED_16(INTFLOAT, dst1, [BUF_SIZE], [2]); in test_mul_pair_single()
81 call_new(dst1, src0, src1, BUF_SIZE); in test_mul_pair_single()
82 if (!float_near_abs_eps_array((float *)dst0, (float *)dst1, EPS, BUF_SIZE * 2)) in test_mul_pair_single()
84 bench_new(dst1, src0, src1, BUF_SIZE); in test_mul_pair_single()
90 LOCAL_ALIGNED_16(INTFLOAT, dst1, [BUF_SIZE], [2]); in test_hybrid_analysis()
[all …]
Dvf_colorspace.c72 uint8_t *dst0[3] = { dst0_y, dst0_u, dst0_v }, *dst1[3] = { dst1_y, dst1_u, dst1_v }; in check_yuv2yuv() local
106 call_new(dst1, (ptrdiff_t[3]) { y_dst_stride, uv_dst_stride, uv_dst_stride }, in check_yuv2yuv()
109 if (memcmp(dst0[0], dst1[0], y_dst_stride * H) || in check_yuv2yuv()
110 memcmp(dst0[1], dst1[1], uv_dst_stride * H >> ss_h) || in check_yuv2yuv()
111 memcmp(dst0[2], dst1[2], uv_dst_stride * H >> ss_h)) { in check_yuv2yuv()
140 int16_t *dst0[3] = { dst0_y, dst0_u, dst0_v }, *dst1[3] = { dst1_y, dst1_u, dst1_v }; in check_yuv2rgb() local
169 call_new(dst1, W, src, in check_yuv2rgb()
172 if (memcmp(dst0[0], dst1[0], W * H * sizeof(int16_t)) || in check_yuv2rgb()
173 memcmp(dst0[1], dst1[1], W * H * sizeof(int16_t)) || in check_yuv2rgb()
174 memcmp(dst0[2], dst1[2], W * H * sizeof(int16_t))) { in check_yuv2rgb()
[all …]
Dhevc_sao.c73 LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); in check_sao_band()
88 memset(dst1, 0, BUF_SIZE); in check_sao_band()
92 call_new(dst1, src1, stride, stride, offset_val, left_class, block_size, block_size); in check_sao_band()
93 if (memcmp(dst0, dst1, BUF_SIZE)) in check_sao_band()
95 bench_new(dst1, src1, stride, stride, offset_val, left_class, block_size, block_size); in check_sao_band()
104 LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); in check_sao_edge()
120 memset(dst1, 0, BUF_SIZE); in check_sao_edge()
124 call_new(dst1, src1 + offset, stride, offset_val, eo, block_size, block_size); in check_sao_edge()
125 if (memcmp(dst0, dst1, BUF_SIZE)) in check_sao_edge()
127 bench_new(dst1, src1 + offset, stride, offset_val, eo, block_size, block_size); in check_sao_edge()
Dhuffyuvdsp.c43 uint16_t *dst1 = av_mallocz(width * sizeof(uint16_t)); in check_add_int16() local
47 if (!src0 || !src1 || !dst0 || !dst1) in check_add_int16()
55 call_new(dst1, src1, mask, width); in check_add_int16()
56 if (memcmp(dst0, dst1, width * sizeof(uint16_t))) in check_add_int16()
58 bench_new(dst1, src1, mask, width); in check_add_int16()
64 av_free(dst1); in check_add_int16()
Dsw_scale.c82 LOCAL_ALIGNED_8(uint8_t, dst1, [LARGEST_INPUT_SIZE]); in check_yuv2yuvX()
112 memset(dst1, 0, LARGEST_INPUT_SIZE * sizeof(dst1[0])); in check_yuv2yuvX()
121 … call_new((const int16_t*)vFilterData, filter_sizes[fsi], src, dst1, dstW - osi, dither, osi); in check_yuv2yuvX()
122 if (memcmp(dst0, dst1, LARGEST_INPUT_SIZE * sizeof(dst0[0]))) in check_yuv2yuvX()
125 … bench_new((const int16_t*)vFilterData, filter_sizes[fsi], src, dst1, dstW - osi, dither, osi); in check_yuv2yuvX()
158 LOCAL_ALIGNED_32(uint32_t, dst1, [SRC_PIXELS]); in check_hscale()
217 memset(dst1, 0, SRC_PIXELS * sizeof(dst1[0])); in check_hscale()
220 call_new(NULL, dst1, SRC_PIXELS, src, filter, filterPos, width); in check_hscale()
221 if (memcmp(dst0, dst1, SRC_PIXELS * sizeof(dst0[0]))) in check_hscale()
Dh264dsp.c208 uint8_t *dst1 = dst1_base + align; in check_idct() local
216 memcpy(dst1, dst, sz * PIXEL_STRIDE); in check_idct()
219 call_new(dst1, subcoef1, PIXEL_STRIDE); in check_idct()
220 if (memcmp(dst0, dst1, sz * PIXEL_STRIDE) || in check_idct()
223 bench_new(dst1, subcoef1, sz * SIZEOF_PIXEL); in check_idct()
236 LOCAL_ALIGNED_16(uint8_t, dst1, [16 * 16 * 2]); in check_idct_multiple()
305 memcpy(dst1, dst_full, 16 * 16 * SIZEOF_PIXEL); in check_idct_multiple()
307 call_new(dst1, block_offset, coef1, 16 * SIZEOF_PIXEL, nnzc); in check_idct_multiple()
308 if (memcmp(dst0, dst1, 16 * 16 * SIZEOF_PIXEL) || in check_idct_multiple()
311 bench_new(dst1, block_offset, coef1, 16 * SIZEOF_PIXEL, nnzc); in check_idct_multiple()
[all …]
Dpixblockdsp.c58 call_new(dst1 + dst_offset, src11 + src_offset, 8); \
59 if (memcmp(src10, src11, BUF_SIZE)|| memcmp(dst0, dst1, BUF_SIZE)) \
61 bench_new(dst1 + dst_offset, src11 + src_offset, 8); \
75 call_new(dst1 + dst_offset, src11 + src_offset, src21 + src_offset, 8); \
76 …memcmp(src10, src11, BUF_SIZE) || memcmp(src20, src21, BUF_SIZE) || memcmp(dst0, dst1, BUF_SIZE)) \
78 bench_new(dst1 + dst_offset, src11 + src_offset, src21 + src_offset, 8); \
91 uint16_t *dst1 = (uint16_t *)dst1_; in checkasm_check_pixblockdsp() local
Daudiodsp.c89 LOCAL_ALIGNED(32, int32_t, dst1, [MAX_SIZE]); in checkasm_check_audiodsp()
110 call_new(dst1, src, min, max, len); in checkasm_check_audiodsp()
111 if (memcmp(dst0, dst1, len * sizeof(*dst0))) in checkasm_check_audiodsp()
113 bench_new(dst1, src, min, max, MAX_SIZE); in checkasm_check_audiodsp()
119 LOCAL_ALIGNED(32, float, dst1, [MAX_SIZE]); in checkasm_check_audiodsp()
138 call_new(dst1, src, len, min, max); in checkasm_check_audiodsp()
140 if (!float_near_ulp_array(dst0, dst1, 3, len)) in checkasm_check_audiodsp()
143 bench_new(dst1, src, MAX_SIZE, min, max); in checkasm_check_audiodsp()
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/intrinsics/sse/
DMatMul_Sse.c38 __m128 dst1 = _mm_setzero_ps(); in MatrixMultiplyWinograd() local
54 dst1 = _mm_add_ps(dst1, tmp1); in MatrixMultiplyWinograd()
61 _mm_storeu_ps(matrix_c, dst1); in MatrixMultiplyWinograd()
71 __m128 dst1 = _mm_setzero_ps(); in MatrixMultiplyWinograd() local
81 dst1 = _mm_add_ps(dst1, tmp1); in MatrixMultiplyWinograd()
86 _mm_storeu_ps(matrix_c, dst1); in MatrixMultiplyWinograd()
117 …__m128 dst1 = _mm_setzero_ps(), dst2 = _mm_setzero_ps(), dst3 = _mm_setzero_ps(), dst4 = _mm_setze… in MatmulFloatSse64Opt() local
125 dst1 = _mm_add_ps(dst1, tmp1), dst2 = _mm_add_ps(dst2, tmp2); in MatmulFloatSse64Opt()
136 DoBiasBlock8(bias_d, &dst1, &dst2, &dst3, &dst4, &dst5, &dst6, &dst7, &dst8); in MatmulFloatSse64Opt()
140 ActBlock8(&dst1, &dst2, &dst3, &dst4, &dst5, &dst6, &dst7, &dst8, act_type); in MatmulFloatSse64Opt()
[all …]
DConvDwFp32Row_sse.c31 __m128 dst1 = _mm_loadu_ps(output_ptr); in ConvDwFp32Row() local
43 dst1 = MS_MLAQ_F32(dst1, w1, in1); in ConvDwFp32Row()
47 _mm_storeu_ps(output_ptr, dst1); in ConvDwFp32Row()
56 __m128 dst1 = _mm_loadu_ps(output_ptr); in ConvDwFp32Row() local
62 dst1 = MS_MLAQ_F32(dst1, w1, in1); in ConvDwFp32Row()
64 _mm_storeu_ps(output_ptr, dst1); in ConvDwFp32Row()
71 __m128 dst1 = _mm_loadu_ps(output_ptr); in ConvDwFp32Row() local
74 dst1 = MS_MLAQ_F32(dst1, w1, in1); in ConvDwFp32Row()
75 _mm_storeu_ps(output_ptr, dst1); in ConvDwFp32Row()
Dsse_common.h88 static inline void WriteCol1(float **dst, __m128 *dst1, __m128 *dst2, __m128 *dst3, __m128 *dst4, _… in WriteCol1() argument
90 _mm_store_ss(*dst, *dst1); in WriteCol1()
107 static inline void WriteCol2(float **dst, __m128 *dst1, __m128 *dst2, __m128 *dst3, __m128 *dst4, _… in WriteCol2() argument
109 _mm_store_ss(*dst, *dst1); in WriteCol2()
110 *dst1 = _mm_shuffle_ps(*dst1, *dst1, _MM_SHUFFLE(0, 3, 2, 1)); in WriteCol2()
111 _mm_store_ss(*dst, *dst1); in WriteCol2()
132 static inline void WriteCol2Opt(float **dst, __m128 *dst1, __m128 *dst2, __m128 *dst3, __m128 *dst4… in WriteCol2Opt() argument
134 _mm_store_ss(*dst, *dst1); in WriteCol2Opt()
135 *dst1 = _mm_shuffle_ps(*dst1, *dst1, _MM_SHUFFLE(0, 3, 2, 1)); in WriteCol2Opt()
136 _mm_store_ss(*dst + 1, *dst1); in WriteCol2Opt()
[all …]
DTiledC4MatMulFp32.c20 static inline void TiledC4MatmulFp32_Transfer(__m128 *dst1, __m128 *dst2, __m128 *dst3, __m128 *dst… in TiledC4MatmulFp32_Transfer() argument
23 *dst1 = _mm_add_ps(*dst1, _mm_mul_ps(weight, _mm_set_ps1(v1))); in TiledC4MatmulFp32_Transfer()
54 __m128 dst1 = _mm_mul_ps(weight_data[0], _mm_set_ps1(MS_F32X4_GETI(src1, 0))); in TiledC4MatmulFp32() local
59 TiledC4MatmulFp32_Transfer(&dst1, &dst2, &dst3, &dst4, weight_data[j], MS_F32X4_GETI(src1, j), in TiledC4MatmulFp32()
80 dst1 = _mm_add_ps(dst1, _mm_mul_ps(weight_data[0], _mm_set_ps1(MS_F32X4_GETI(src1, 0)))); in TiledC4MatmulFp32()
86 … TiledC4MatmulFp32_Transfer(&dst1, &dst2, &dst3, &dst4, weight_data[1], MS_F32X4_GETI(src1, 1), in TiledC4MatmulFp32()
93 … TiledC4MatmulFp32_Transfer(&dst1, &dst2, &dst3, &dst4, weight_data[2], MS_F32X4_GETI(src1, 2), in TiledC4MatmulFp32()
96 dst1 = _mm_add_ps(dst1, _mm_mul_ps(weight_data[3], _mm_set_ps1(MS_F32X4_GETI(src1, 3)))); in TiledC4MatmulFp32()
124 dst1 = _mm_add_ps(dst1, _mm_mul_ps(weight_data[0], _mm_set_ps1(MS_F32X4_GETI(src1, 0)))); in TiledC4MatmulFp32()
130 TiledC4MatmulFp32_Transfer(&dst1, &dst2, &dst3, &dst4, weight_data[1], MS_F32X4_GETI(src1, 1), in TiledC4MatmulFp32()
[all …]
/third_party/flutter/skia/third_party/externals/sdl/src/video/
DSDL_blit_A.c185 __m64 src1, src2, dst1, dst2, lmask, hmask, dsta; in BlitRGBtoRGBSurfaceAlpha128MMX() local
202 dst1 = *(__m64 *) dstp; /* 2 x dst -> dst1(ARGBARGB) */ in BlitRGBtoRGBSurfaceAlpha128MMX()
203 dst2 = dst1; /* 2 x dst -> dst2(ARGBARGB) */ in BlitRGBtoRGBSurfaceAlpha128MMX()
213 dst1 = _mm_and_si64(dst1, src1); /* src & dst -> dst1 */ in BlitRGBtoRGBSurfaceAlpha128MMX()
214 dst1 = _mm_and_si64(dst1, lmask); /* dst1 & !mask -> dst1 */ in BlitRGBtoRGBSurfaceAlpha128MMX()
215 dst1 = _mm_add_pi32(dst1, src2); /* src2 + dst1 -> dst1 */ in BlitRGBtoRGBSurfaceAlpha128MMX()
216 dst1 = _mm_or_si64(dst1, dsta); /* dsta(full alpha) | dst1 -> dst1 */ in BlitRGBtoRGBSurfaceAlpha128MMX()
218 *(__m64 *) dstp = dst1; /* dst1 -> 2 x dst pixels */ in BlitRGBtoRGBSurfaceAlpha128MMX()
250 __m64 src1, src2, dst1, dst2, mm_alpha, mm_zero, dsta; in BlitRGBtoRGBSurfaceAlphaMMX() local
271 dst1 = _mm_cvtsi32_si64(*dstp); /* dst(ARGB) -> dst1 (0000ARGB) */ in BlitRGBtoRGBSurfaceAlphaMMX()
[all …]
/third_party/boost/libs/multi_index/test/
Dtest_node_handling.cpp297 Src& src,Dst& dst0,Dst& dst1,Dst& /* dst2 */,Dst& /* dst3 */, in test_transfer() argument
302 test_transfer(src,src.begin(),dst1,dst1.find(*src.begin())); in test_transfer()
303 test_transfer(src,src.begin(),dst1,dst1.find(*src.begin())); in test_transfer()
309 Src& src,Dst& dst0,Dst& dst1,Dst& /* dst2 */,Dst& /* dst3 */, in test_transfer() argument
314 test_transfer(src,*src.begin(),dst1,dst1.begin()); in test_transfer()
315 test_transfer(src,*src.begin(),dst1,dst1.begin()); in test_transfer()
316 test_transfer(src, -1,dst1,dst1.begin()); in test_transfer()
322 Src& src,Dst& dst0,Dst& dst1,Dst& dst2,Dst& dst3, in test_transfer() argument
327 test_transfer(src,*src.begin(),dst1); in test_transfer()
328 test_transfer(src,*src.begin(),dst1); in test_transfer()
[all …]
/third_party/flutter/skia/third_party/externals/libwebp/src/dsp/
Dlossless_msa.c25 v16u8 src0, src1, src2, src3, dst0, dst1, dst2; \
27 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \
29 ST_UB2(dst0, dst1, pdst, 16); \
35 v16u8 src0, src1, src2, dst0, dst1, dst2; \
37 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \
39 ST_UB2(dst0, dst1, pdst, 16); \
46 v16u8 src0, src1, src2 = { 0 }, dst0, dst1; \
48 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \
50 pix_d = __msa_copy_s_d((v2i64)dst1, 0); \
81 #define TRANSFORM_COLOR_INVERSE_8(src0, src1, dst0, dst1, \ argument
[all …]
/third_party/skia/third_party/externals/libwebp/src/dsp/
Dlossless_msa.c25 v16u8 src0, src1, src2, src3, dst0, dst1, dst2; \
27 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \
29 ST_UB2(dst0, dst1, pdst, 16); \
35 v16u8 src0, src1, src2, dst0, dst1, dst2; \
37 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \
39 ST_UB2(dst0, dst1, pdst, 16); \
46 v16u8 src0, src1, src2 = { 0 }, dst0, dst1; \
48 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \
50 pix_d = __msa_copy_s_d((v2i64)dst1, 0); \
81 #define TRANSFORM_COLOR_INVERSE_8(src0, src1, dst0, dst1, \ argument
[all …]
/third_party/mesa3d/src/intel/compiler/
Dtest_fs_saturate_propagation.cpp117 fs_reg dst1 = v->vgrf(glsl_type::float_type); in TEST_F() local
121 set_saturate(true, bld.MOV(dst1, dst0)); in TEST_F()
152 fs_reg dst1 = v->vgrf(glsl_type::float_type); in TEST_F() local
157 set_saturate(true, bld.MOV(dst1, dst0)); in TEST_F()
190 fs_reg dst1 = v->vgrf(glsl_type::float_type); in TEST_F() local
195 set_saturate(true, bld.MOV(dst1, dst0)); in TEST_F()
225 fs_reg dst1 = v->vgrf(glsl_type::float_type); in TEST_F() local
229 set_saturate(true, bld.MOV(dst1, dst0)); in TEST_F()
259 fs_reg dst1 = v->vgrf(glsl_type::float_type); in TEST_F() local
264 set_saturate(true, bld.MOV(dst1, dst0)); in TEST_F()
[all …]
/third_party/ffmpeg/libavcodec/mips/
Dhevc_mc_bi_msa.c74 v8i16 dst0, dst1, dst2, dst3; in hevc_bi_copy_4w_msa() local
96 ILVRL_B2_SH(zero, src0, dst0, dst1); in hevc_bi_copy_4w_msa()
97 SLLI_2V(dst0, dst1, 6); in hevc_bi_copy_4w_msa()
98 HEVC_BI_RND_CLIP2_MAX_SATU(in0, in1, dst0, dst1, 7, dst0, dst1); in hevc_bi_copy_4w_msa()
99 dst0 = (v8i16) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_bi_copy_4w_msa()
117 ILVRL_B2_SH(zero, src0, dst0, dst1); in hevc_bi_copy_4w_msa()
119 SLLI_4V(dst0, dst1, dst2, dst3, 6); in hevc_bi_copy_4w_msa()
120 HEVC_BI_RND_CLIP4_MAX_SATU(in0, in1, in2, in3, dst0, dst1, dst2, in hevc_bi_copy_4w_msa()
121 dst3, 7, dst0, dst1, dst2, dst3); in hevc_bi_copy_4w_msa()
122 PCKEV_B2_SH(dst1, dst0, dst3, dst2, dst0, dst1); in hevc_bi_copy_4w_msa()
[all …]
Dhevc_mc_uniw_msa.c70 v8i16 dst0, dst1, dst2, dst3, offset_vec; in hevc_uniwgt_copy_4w_msa() local
97 ILVRL_B2_SH(zero, src0, dst0, dst1); in hevc_uniwgt_copy_4w_msa()
98 SLLI_2V(dst0, dst1, 6); in hevc_uniwgt_copy_4w_msa()
99 HEVC_UNIW_RND_CLIP2_MAX_SATU_H(dst0, dst1, weight_vec, offset_vec, in hevc_uniwgt_copy_4w_msa()
100 rnd_vec, dst0, dst1); in hevc_uniwgt_copy_4w_msa()
101 out0 = (v16u8) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); in hevc_uniwgt_copy_4w_msa()
111 ILVRL_B2_SH(zero, src0, dst0, dst1); in hevc_uniwgt_copy_4w_msa()
113 SLLI_4V(dst0, dst1, dst2, dst3, 6); in hevc_uniwgt_copy_4w_msa()
114 HEVC_UNIW_RND_CLIP4_MAX_SATU_H(dst0, dst1, dst2, dst3, weight_vec, in hevc_uniwgt_copy_4w_msa()
115 offset_vec, rnd_vec, dst0, dst1, in hevc_uniwgt_copy_4w_msa()
[all …]
/third_party/ffmpeg/libavfilter/x86/
Dvf_spp.c38 #define REQUANT_CORE(dst0, dst1, dst2, dst3, src0, src1, src2, src3) \ in hardthresh_mmx() argument
77 "movq %%mm7, " #dst1 " \n" \ in hardthresh_mmx()
109 #define REQUANT_CORE(dst0, dst1, dst2, dst3, src0, src1, src2, src3) \ in softthresh_mmx() argument
157 "movq %%mm7, " #dst1 " \n" \ in softthresh_mmx()
186 uint8_t *dst1 = dst; in store_slice_mmx() local
211 : "+r" (src1), "+r"(dst1) in store_slice_mmx()

1234567