/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | idct_blk_neon.c | 17 unsigned char *dst0; in idct_dequant_0_2x_neon() local 31 dst0 = dst; in idct_dequant_0_2x_neon() 32 d2s32 = vld1_lane_s32((const int32_t *)dst0, d2s32, 0); in idct_dequant_0_2x_neon() 33 dst0 += stride; in idct_dequant_0_2x_neon() 34 d2s32 = vld1_lane_s32((const int32_t *)dst0, d2s32, 1); in idct_dequant_0_2x_neon() 35 dst0 += stride; in idct_dequant_0_2x_neon() 36 d4s32 = vld1_lane_s32((const int32_t *)dst0, d4s32, 0); in idct_dequant_0_2x_neon() 37 dst0 += stride; in idct_dequant_0_2x_neon() 38 d4s32 = vld1_lane_s32((const int32_t *)dst0, d4s32, 1); in idct_dequant_0_2x_neon() 51 dst0 = dst; in idct_dequant_0_2x_neon() [all …]
|
D | dequant_idct_neon.c | 24 unsigned char *dst0; in vp8_dequant_idct_add_neon() local 47 dst0 = dst; in vp8_dequant_idct_add_neon() 48 d14 = vld1_lane_s32((const int32_t *)dst0, d14, 0); in vp8_dequant_idct_add_neon() 49 dst0 += stride; in vp8_dequant_idct_add_neon() 50 d14 = vld1_lane_s32((const int32_t *)dst0, d14, 1); in vp8_dequant_idct_add_neon() 51 dst0 += stride; in vp8_dequant_idct_add_neon() 52 d15 = vld1_lane_s32((const int32_t *)dst0, d15, 0); in vp8_dequant_idct_add_neon() 53 dst0 += stride; in vp8_dequant_idct_add_neon() 54 d15 = vld1_lane_s32((const int32_t *)dst0, d15, 1); in vp8_dequant_idct_add_neon() 132 dst0 = dst; in vp8_dequant_idct_add_neon() [all …]
|
/external/mesa3d/src/intel/compiler/ |
D | test_fs_saturate_propagation.cpp | 104 fs_reg dst0 = v->vgrf(glsl_type::float_type); in TEST_F() local 108 bld.ADD(dst0, src0, src1); in TEST_F() 109 set_saturate(true, bld.MOV(dst1, dst0)); in TEST_F() 139 fs_reg dst0 = v->vgrf(glsl_type::float_type); in TEST_F() local 144 bld.ADD(dst0, src0, src1); in TEST_F() 145 set_saturate(true, bld.MOV(dst1, dst0)); in TEST_F() 146 bld.ADD(dst2, dst0, src0); in TEST_F() 177 fs_reg dst0 = v->vgrf(glsl_type::float_type); in TEST_F() local 181 bld.ADD(dst0, src0, src1) in TEST_F() 183 set_saturate(true, bld.MOV(dst1, dst0)); in TEST_F() [all …]
|
/external/webp/src/dsp/ |
D | lossless_msa.c | 25 v16u8 src0, src1, src2, src3, dst0, dst1, dst2; \ 27 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \ 29 ST_UB2(dst0, dst1, pdst, 16); \ 35 v16u8 src0, src1, src2, dst0, dst1, dst2; \ 37 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \ 39 ST_UB2(dst0, dst1, pdst, 16); \ 46 v16u8 src0, src1, src2 = { 0 }, dst0, dst1; \ 48 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \ 49 ST_UB(dst0, pdst); \ 56 const v16u8 dst0 = VSHF_UB(src0, src0, m); \ [all …]
|
D | lossless_enc_msa.c | 21 #define TRANSFORM_COLOR_8(src0, src1, dst0, dst1, c0, c1, mask0, mask1) do { \ argument 34 VSHF_B2_UB(src0, t0, src1, t1, mask1, mask1, dst0, dst1); \ 53 v16u8 src0, dst0; in TransformColor_MSA() local 65 TRANSFORM_COLOR_8(src0, src1, dst0, dst1, g2br, r2b, mask0, mask1); in TransformColor_MSA() 66 ST_UB2(dst0, dst1, data, 4); in TransformColor_MSA() 73 TRANSFORM_COLOR_4(src0, dst0, g2br, r2b, mask0, mask1); in TransformColor_MSA() 74 ST_UB(dst0, data); in TransformColor_MSA() 80 TRANSFORM_COLOR_4(src0, dst0, g2br, r2b, mask0, mask1); in TransformColor_MSA() 82 const uint64_t pix_d = __msa_copy_s_d((v2i64)dst0, 0); in TransformColor_MSA() 83 const uint32_t pix_w = __msa_copy_s_w((v4i32)dst0, 2); in TransformColor_MSA() [all …]
|
D | filters_msa.c | 25 v16u8 src0, pred0, dst0; in PredictLineInverse0() local 31 SUB2(src0, pred0, src1, pred1, dst0, dst1); in PredictLineInverse0() 32 ST_UB2(dst0, dst1, dst, 16); in PredictLineInverse0() 43 dst0 = src0 - pred0; in PredictLineInverse0() 44 ST_UB(dst0, dst); in PredictLineInverse0() 105 v16u8 pred0, dst0; in PredictLineGradient() local 118 dst0 = src0 - pred0; in PredictLineGradient() 119 ST_UB(dst0, poutput); in PredictLineGradient()
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
D | vpx_convolve_avg_msa.c | 19 v16u8 dst0, dst1, dst2, dst3; in avg_width4_msa() local 26 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in avg_width4_msa() 28 AVER_UB4_UB(src0, dst0, src1, dst1, src2, dst2, src3, dst3, dst0, dst1, in avg_width4_msa() 31 out0 = __msa_copy_u_w((v4i32)dst0, 0); in avg_width4_msa() 43 LD_UB2(dst, dst_stride, dst0, dst1); in avg_width4_msa() 45 AVER_UB2_UB(src0, dst0, src1, dst1, dst0, dst1); in avg_width4_msa() 47 out0 = __msa_copy_u_w((v4i32)dst0, 0); in avg_width4_msa() 62 v16u8 dst0, dst1, dst2, dst3; in avg_width8_msa() local 67 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in avg_width8_msa() 69 AVER_UB4_UB(src0, dst0, src1, dst1, src2, dst2, src3, dst3, dst0, dst1, in avg_width8_msa() [all …]
|
D | vpx_convolve8_avg_horiz_msa.c | 21 v16u8 dst0 = { 0 }, res; in common_hz_8t_and_aver_dst_4x4_msa() local 41 INSERT_W4_UB(tp0, tp1, tp2, tp3, dst0); in common_hz_8t_and_aver_dst_4x4_msa() 45 res = (v16u8)__msa_aver_u_b(res, dst0); in common_hz_8t_and_aver_dst_4x4_msa() 56 v16u8 dst0 = { 0 }, dst1 = { 0 }; in common_hz_8t_and_aver_dst_4x8_msa() local 74 INSERT_W4_UB(tp0, tp1, tp2, tp3, dst0); in common_hz_8t_and_aver_dst_4x8_msa() 89 AVER_UB2_UB(res0, dst0, res2, dst1, res0, res2); in common_hz_8t_and_aver_dst_4x8_msa() 111 v16u8 mask0, mask1, mask2, mask3, dst0 = { 0 }, dst1 = { 0 }; in common_hz_8t_and_aver_dst_8w_msa() local 133 INSERT_D2_UB(tp0, tp1, dst0); in common_hz_8t_and_aver_dst_8w_msa() 137 CONVERT_UB_AVG_ST8x4_UB(out0, out1, out2, out3, dst0, dst1, dst, in common_hz_8t_and_aver_dst_8w_msa() 149 v16u8 mask0, mask1, mask2, mask3, dst0, dst1; in common_hz_8t_and_aver_dst_16w_msa() local [all …]
|
D | vpx_convolve8_avg_vert_msa.c | 22 v16u8 dst0 = { 0 }, out; in common_vt_8t_and_aver_dst_4w_msa() local 48 INSERT_W4_UB(tp0, tp1, tp2, tp3, dst0); in common_vt_8t_and_aver_dst_4w_msa() 60 out = __msa_aver_u_b(out, dst0); in common_vt_8t_and_aver_dst_4w_msa() 79 v16u8 dst0 = { 0 }, dst1 = { 0 }; in common_vt_8t_and_aver_dst_8w_msa() local 102 INSERT_D2_UB(tp0, tp1, dst0); in common_vt_8t_and_aver_dst_8w_msa() 117 CONVERT_UB_AVG_ST8x4_UB(out0, out1, out2, out3, dst0, dst1, dst, in common_vt_8t_and_aver_dst_8w_msa() 142 v16u8 dst0, dst1, dst2, dst3, tmp0, tmp1, tmp2, tmp3; in common_vt_8t_and_aver_dst_16w_mult_msa() local 169 LD_UB4(dst_tmp, dst_stride, dst0, dst1, dst2, dst3); in common_vt_8t_and_aver_dst_16w_mult_msa() 198 AVER_UB4_UB(tmp0, dst0, tmp1, dst1, tmp2, dst2, tmp3, dst3, dst0, dst1, in common_vt_8t_and_aver_dst_16w_mult_msa() 200 ST_UB4(dst0, dst1, dst2, dst3, dst_tmp, dst_stride); in common_vt_8t_and_aver_dst_16w_mult_msa() [all …]
|
/external/ltp/testcases/kernel/syscalls/string/ |
D | string01.c | 70 char dst0[LONGSTR + 1]; /* place holders for various tests */ variable 142 dst0, "", "", 0}, { 143 dst0, tiat, tiat, 0}, { 144 dst0, "", tiat, 0}, { 145 dst0, yat, tiatyat, 0}, { 159 dst0, "", 0}, { 160 dst0, tiat, 0}, { 161 dst0, longstr, 0}, { 199 dst0, "", LONGSTR, "", 0}, { 200 dst0, tiat, LONGSTR, tiat, 0}, { [all …]
|
/external/mesa3d/src/gallium/auxiliary/gallivm/ |
D | lp_bld_tgsi_aos.c | 436 LLVMValueRef dst0 = NULL; in lp_emit_instruction_aos() local 453 dst0 = bld->bld_base.base.undef; in lp_emit_instruction_aos() 459 dst0 = lp_build_floor(&bld->bld_base.base, src0); in lp_emit_instruction_aos() 463 dst0 = lp_build_emit_fetch(&bld->bld_base, inst, 0, LP_CHAN_ALL); in lp_emit_instruction_aos() 472 dst0 = lp_build_rcp(&bld->bld_base.base, src0); in lp_emit_instruction_aos() 479 dst0 = lp_build_rsqrt(&bld->bld_base.base, tmp0); in lp_emit_instruction_aos() 491 dst0 = lp_build_mul(&bld->bld_base.base, src0, src1); in lp_emit_instruction_aos() 497 dst0 = lp_build_add(&bld->bld_base.base, src0, src1); in lp_emit_instruction_aos() 514 dst0 = lp_build_min(&bld->bld_base.base, src0, src1); in lp_emit_instruction_aos() 520 dst0 = lp_build_max(&bld->bld_base.base, src0, src1); in lp_emit_instruction_aos() [all …]
|
/external/libyuv/files/source/ |
D | scale_msa.cc | 29 v16u8 src0, src1, dst0; in ScaleARGBRowDown2_MSA() local 35 dst0 = (v16u8)__msa_pckod_w((v4i32)src1, (v4i32)src0); in ScaleARGBRowDown2_MSA() 36 ST_UB(dst0, dst_argb); in ScaleARGBRowDown2_MSA() 47 v16u8 src0, src1, vec0, vec1, dst0; in ScaleARGBRowDown2Linear_MSA() local 55 dst0 = (v16u8)__msa_aver_u_b((v16u8)vec0, (v16u8)vec1); in ScaleARGBRowDown2Linear_MSA() 56 ST_UB(dst0, dst_argb); in ScaleARGBRowDown2Linear_MSA() 69 v16u8 src0, src1, src2, src3, vec0, vec1, vec2, vec3, dst0; in ScaleARGBRowDown2Box_MSA() local 90 dst0 = (v16u8)__msa_pckev_b((v16i8)reg1, (v16i8)reg0); in ScaleARGBRowDown2Box_MSA() 91 ST_UB(dst0, dst_argb); in ScaleARGBRowDown2Box_MSA() 134 v16u8 dst0; in ScaleARGBRowDownEvenBox_MSA() local [all …]
|
D | row_msa.cc | 291 v16u8 dst0, dst1, dst2, dst3; in MirrorRow_MSA() local 298 VSHF_B2_UB(src1, src1, src0, src0, shuffler, shuffler, dst1, dst0); in MirrorRow_MSA() 299 ST_UB4(dst0, dst1, dst2, dst3, dst, 16); in MirrorRow_MSA() 308 v16u8 dst0, dst1, dst2, dst3; in ARGBMirrorRow_MSA() local 315 VSHF_B2_UB(src1, src1, src0, src0, shuffler, shuffler, dst1, dst0); in ARGBMirrorRow_MSA() 316 ST_UB4(dst0, dst1, dst2, dst3, dst, 16); in ARGBMirrorRow_MSA() 477 v16u8 src0, src1, src2, src3, src4, dst0, dst1, dst2; in I422ToRGB24Row_MSA() local 510 dst0 = (v16u8)__msa_vshf_b(shuffler0, (v16i8)reg3, (v16i8)reg0); in I422ToRGB24Row_MSA() 513 ST_UB2(dst0, dst1, rgb_buf, 16); in I422ToRGB24Row_MSA() 530 v16u8 src0, src1, src2, dst0; in I422ToRGB565Row_MSA() local [all …]
|
D | rotate_msa.cc | 84 v16u8 src0, src1, src2, src3, dst0, dst1, dst2, dst3, vec0, vec1, vec2, vec3; in TransposeWx16_MSA() local 134 ILVRL_D(res0, res8, res1, res9, dst0, dst1, dst2, dst3); in TransposeWx16_MSA() 135 ST_UB4(dst0, dst1, dst2, dst3, dst, dst_stride); in TransposeWx16_MSA() 139 ILVRL_D(res2, res8, res3, res9, dst0, dst1, dst2, dst3); in TransposeWx16_MSA() 140 ST_UB4(dst0, dst1, dst2, dst3, dst, dst_stride); in TransposeWx16_MSA() 144 ILVRL_D(res4, res8, res5, res9, dst0, dst1, dst2, dst3); in TransposeWx16_MSA() 145 ST_UB4(dst0, dst1, dst2, dst3, dst, dst_stride); in TransposeWx16_MSA() 149 ILVRL_D(res6, res8, res7, res9, dst0, dst1, dst2, dst3); in TransposeWx16_MSA() 150 ST_UB4(dst0, dst1, dst2, dst3, dst, dst_stride); in TransposeWx16_MSA() 165 v16u8 src0, src1, src2, src3, dst0, dst1, dst2, dst3, vec0, vec1, vec2, vec3; in TransposeUVWx16_MSA() local [all …]
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | row_msa.cc | 291 v16u8 dst0, dst1, dst2, dst3; in MirrorRow_MSA() local 298 VSHF_B2_UB(src1, src1, src0, src0, shuffler, shuffler, dst1, dst0); in MirrorRow_MSA() 299 ST_UB4(dst0, dst1, dst2, dst3, dst, 16); in MirrorRow_MSA() 308 v16u8 dst0, dst1, dst2, dst3; in ARGBMirrorRow_MSA() local 315 VSHF_B2_UB(src1, src1, src0, src0, shuffler, shuffler, dst1, dst0); in ARGBMirrorRow_MSA() 316 ST_UB4(dst0, dst1, dst2, dst3, dst, 16); in ARGBMirrorRow_MSA() 477 v16u8 src0, src1, src2, src3, src4, dst0, dst1, dst2; in I422ToRGB24Row_MSA() local 510 dst0 = (v16u8)__msa_vshf_b(shuffler0, (v16i8)reg3, (v16i8)reg0); in I422ToRGB24Row_MSA() 513 ST_UB2(dst0, dst1, dst_argb, 16); in I422ToRGB24Row_MSA() 530 v16u8 src0, src1, src2, dst0; in I422ToRGB565Row_MSA() local [all …]
|
D | scale_msa.cc | 37 v16u8 src0, src1, dst0; in ScaleARGBRowDown2_MSA() local 43 dst0 = (v16u8)__msa_pckod_w((v4i32)src1, (v4i32)src0); in ScaleARGBRowDown2_MSA() 44 ST_UB(dst0, dst_argb); in ScaleARGBRowDown2_MSA() 55 v16u8 src0, src1, vec0, vec1, dst0; in ScaleARGBRowDown2Linear_MSA() local 63 dst0 = (v16u8)__msa_aver_u_b((v16u8)vec0, (v16u8)vec1); in ScaleARGBRowDown2Linear_MSA() 64 ST_UB(dst0, dst_argb); in ScaleARGBRowDown2Linear_MSA() 77 v16u8 src0, src1, src2, src3, vec0, vec1, vec2, vec3, dst0; in ScaleARGBRowDown2Box_MSA() local 98 dst0 = (v16u8)__msa_pckev_b((v16i8)reg1, (v16i8)reg0); in ScaleARGBRowDown2Box_MSA() 99 ST_UB(dst0, dst_argb); in ScaleARGBRowDown2Box_MSA() 142 v16u8 dst0; in ScaleARGBRowDownEvenBox_MSA() local [all …]
|
D | rotate_msa.cc | 84 v16u8 src0, src1, src2, src3, dst0, dst1, dst2, dst3, vec0, vec1, vec2, vec3; in TransposeWx16_MSA() local 134 ILVRL_D(res0, res8, res1, res9, dst0, dst1, dst2, dst3); in TransposeWx16_MSA() 135 ST_UB4(dst0, dst1, dst2, dst3, dst, dst_stride); in TransposeWx16_MSA() 139 ILVRL_D(res2, res8, res3, res9, dst0, dst1, dst2, dst3); in TransposeWx16_MSA() 140 ST_UB4(dst0, dst1, dst2, dst3, dst, dst_stride); in TransposeWx16_MSA() 144 ILVRL_D(res4, res8, res5, res9, dst0, dst1, dst2, dst3); in TransposeWx16_MSA() 145 ST_UB4(dst0, dst1, dst2, dst3, dst, dst_stride); in TransposeWx16_MSA() 149 ILVRL_D(res6, res8, res7, res9, dst0, dst1, dst2, dst3); in TransposeWx16_MSA() 150 ST_UB4(dst0, dst1, dst2, dst3, dst, dst_stride); in TransposeWx16_MSA() 165 v16u8 src0, src1, src2, src3, dst0, dst1, dst2, dst3, vec0, vec1, vec2, vec3; in TransposeUVWx16_MSA() local [all …]
|
/external/mesa3d/src/gallium/drivers/nouveau/codegen/ |
D | nv50_ir_from_tgsi.cpp | 1705 void handleTEX(Value *dst0[4], int R, int S, int L, int C, int Dx, int Dy); 1706 void handleTXF(Value *dst0[4], int R, int L_M); 1707 void handleTXQ(Value *dst0[4], enum TexQuery, int R); 1708 void handleFBFETCH(Value *dst0[4]); 1709 void handleLIT(Value *dst0[4]); 1715 void handleLOAD(Value *dst0[4]); 1717 void handleATOM(Value *dst0[4], DataType, uint16_t subOp); 1719 void handleINTERP(Value *dst0[4]); 2306 Converter::handleTXQ(Value *dst0[4], enum TexQuery query, int R) in handleTXQ() 2313 if (!dst0[c]) in handleTXQ() [all …]
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | idct_neon.asm | 16 ; $dst0..3 are d registers with the pairs assumed to be contiguous in 19 LOAD_TRAN_LOW_TO_S16 $dst0, $dst1, $dst2, $dst3, $src 23 vmovn.i32 $dst0, q0 28 vld1.s16 {$dst0-$dst1,$dst2-$dst3}, [$src]! 32 ; $dst0..3 are d registers. q0-q3 are used as temporaries in high-bitdepth. 34 LOAD_TRAN_LOW_TO_S16X2 $dst0, $dst1, $dst2, $dst3, $src 38 vmovn.i32 $dst0, q0 43 vld2.s16 {$dst0,$dst1,$dst2,$dst3}, [$src]!
|
/external/libvpx/config/arm-neon/vpx_dsp/arm/ |
D | idct_neon.asm.S | 20 @ $dst0..3 are d registers with the pairs assumed to be contiguous in 22 .macro LOAD_TRAN_LOW_TO_S16 dst0, dst1, dst2, dst3, src 26 vmovn.i32 \dst0, q0 31 vld1.s16 {\dst0-\dst1,\dst2-\dst3}, [\src]! 35 @ $dst0..3 are d registers. q0-q3 are used as temporaries in high-bitdepth. 36 .macro LOAD_TRAN_LOW_TO_S16X2 dst0, dst1, dst2, dst3, src 40 vmovn.i32 \dst0, q0 45 vld2.s16 {\dst0,\dst1,\dst2,\dst3}, [\src]!
|
/external/libpng/mips/ |
D | filter_msa_intrinsics.c | 466 v16u8 dst0, dst1; in png_read_filter_row_sub4_msa() local 488 ILVEV_W2_UB(src1, src2, src3, src4, dst0, dst1); in png_read_filter_row_sub4_msa() 489 dst0 = (v16u8) __msa_pckev_d((v2i64) dst1, (v2i64) dst0); in png_read_filter_row_sub4_msa() 491 ST_UB(dst0, nxt); in png_read_filter_row_sub4_msa() 505 v16u8 src0, src1, src2, src3, src4, dst0, dst1; in png_read_filter_row_sub3_msa() local 529 VSHF_B2_UB(src1, src2, src3, src4, mask0, mask0, dst0, dst1); in png_read_filter_row_sub3_msa() 530 dst0 = (v16u8) __msa_vshf_b(mask1, (v16i8) dst1, (v16i8) dst0); in png_read_filter_row_sub3_msa() 531 out0 = __msa_copy_s_d((v2i64) dst0, 0); in png_read_filter_row_sub3_msa() 532 out1 = __msa_copy_s_w((v4i32) dst0, 2); in png_read_filter_row_sub3_msa() 550 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9, dst0, dst1; in png_read_filter_row_avg4_msa() local [all …]
|
/external/libaom/libaom/av1/common/x86/ |
D | jnt_convolve_avx2.c | 39 uint8_t *dst0, int dst_stride0, int w, int h, in av1_dist_wtd_convolve_x_avx2() argument 113 _mm_storel_epi64((__m128i *)(&dst0[i * dst_stride0 + j]), res_0); in av1_dist_wtd_convolve_x_avx2() 115 (__m128i *)((&dst0[i * dst_stride0 + j + dst_stride0])), res_1); in av1_dist_wtd_convolve_x_avx2() 117 *(uint32_t *)(&dst0[i * dst_stride0 + j]) = in av1_dist_wtd_convolve_x_avx2() 119 *(uint32_t *)(&dst0[i * dst_stride0 + j + dst_stride0]) = in av1_dist_wtd_convolve_x_avx2() 168 _mm_storel_epi64((__m128i *)(&dst0[i * dst_stride0 + j]), res_0); in av1_dist_wtd_convolve_x_avx2() 170 (__m128i *)((&dst0[i * dst_stride0 + j + dst_stride0])), res_1); in av1_dist_wtd_convolve_x_avx2() 172 *(uint32_t *)(&dst0[i * dst_stride0 + j]) = in av1_dist_wtd_convolve_x_avx2() 174 *(uint32_t *)(&dst0[i * dst_stride0 + j + dst_stride0]) = in av1_dist_wtd_convolve_x_avx2() 191 uint8_t *dst0, int dst_stride0, int w, int h, in av1_dist_wtd_convolve_y_avx2() argument [all …]
|
/external/libvpx/libvpx/vp9/common/mips/msa/ |
D | vp9_mfqe_msa.c | 23 v16i8 dst0 = { 0 }; in filter_by_weight8x8_msa() local 35 INSERT_D2_SB(dst0_d, dst1_d, dst0); in filter_by_weight8x8_msa() 44 UNPCK_UB_SH(dst0, dst_r, dst_l); in filter_by_weight8x8_msa() 50 dst0 = (v16i8)__msa_pckev_b((v16i8)res_h_l, (v16i8)res_h_r); in filter_by_weight8x8_msa() 51 ST8x2_UB(dst0, dst_ptr, dst_stride); in filter_by_weight8x8_msa() 72 v16i8 src0, src1, src2, src3, dst0, dst1, dst2, dst3; in filter_by_weight16x16_msa() local 81 LD_SB4(dst_ptr, dst_stride, dst0, dst1, dst2, dst3); in filter_by_weight16x16_msa() 84 UNPCK_UB_SH(dst0, dst_r, dst_l); in filter_by_weight16x16_msa()
|
/external/libvpx/libvpx/vp8/common/mips/msa/ |
D | mfqe_msa.c | 23 v16i8 dst0 = { 0 }; in filter_by_weight8x8_msa() local 35 INSERT_D2_SB(dst0_d, dst1_d, dst0); in filter_by_weight8x8_msa() 44 UNPCK_UB_SH(dst0, dst_r, dst_l); in filter_by_weight8x8_msa() 50 dst0 = (v16i8)__msa_pckev_b((v16i8)res_h_l, (v16i8)res_h_r); in filter_by_weight8x8_msa() 51 ST8x2_UB(dst0, dst_ptr, dst_stride); in filter_by_weight8x8_msa() 73 v16i8 dst0, dst1, dst2, dst3; in filter_by_weight16x16_msa() local 84 LD_SB4(dst_ptr, dst_stride, dst0, dst1, dst2, dst3); in filter_by_weight16x16_msa() 87 UNPCK_UB_SH(dst0, dst_r, dst_l); in filter_by_weight16x16_msa()
|
/external/skqp/tests/ |
D | EncodeTest.cpp | 63 SkDynamicMemoryWStream dst0, dst1, dst2, dst3; in test_encode() local 64 success = encode(format, &dst0, src); in test_encode() 83 sk_sp<SkData> data0 = dst0.detachAsData(); in test_encode() 148 SkDynamicMemoryWStream dst0, dst1, dst2; in DEF_TEST() local 150 success = SkJpegEncoder::Encode(&dst0, src, options); in DEF_TEST() 161 sk_sp<SkData> data0 = dst0.detachAsData(); in DEF_TEST() 260 SkDynamicMemoryWStream dst0, dst1, dst2; in DEF_TEST() local 262 success = SkPngEncoder::Encode(&dst0, src, options); in DEF_TEST() 275 sk_sp<SkData> data0 = dst0.detachAsData(); in DEF_TEST() 303 SkDynamicMemoryWStream dst0, dst1, dst2, dst3; in DEF_TEST() local [all …]
|