• Home
  • Raw
  • Download

Lines Matching refs:vec0

55   v16u8 src0, src1, vec0, vec1, dst0;  in ScaleARGBRowDown2Linear_MSA()  local
61 vec0 = (v16u8)__msa_pckev_w((v4i32)src1, (v4i32)src0); in ScaleARGBRowDown2Linear_MSA()
63 dst0 = (v16u8)__msa_aver_u_b((v16u8)vec0, (v16u8)vec1); in ScaleARGBRowDown2Linear_MSA()
77 v16u8 src0, src1, src2, src3, vec0, vec1, vec2, vec3, dst0; in ScaleARGBRowDown2Box_MSA() local
86 vec0 = (v16u8)__msa_vshf_b(shuffler, (v16i8)src0, (v16i8)src0); in ScaleARGBRowDown2Box_MSA()
90 reg0 = __msa_hadd_u_h(vec0, vec0); in ScaleARGBRowDown2Box_MSA()
140 v16u8 vec0, vec1, vec2, vec3; in ScaleARGBRowDownEvenBox_MSA() local
161 vec0 = (v16u8)__msa_ilvr_b((v16i8)src2, (v16i8)src0); in ScaleARGBRowDownEvenBox_MSA()
165 reg0 = __msa_hadd_u_h(vec0, vec0); in ScaleARGBRowDownEvenBox_MSA()
211 v16u8 src0, src1, src2, src3, vec0, vec1, vec2, vec3, dst0, dst1; in ScaleRowDown2Linear_MSA() local
219 vec0 = (v16u8)__msa_pckev_b((v16i8)src1, (v16i8)src0); in ScaleRowDown2Linear_MSA()
223 dst0 = __msa_aver_u_b(vec1, vec0); in ScaleRowDown2Linear_MSA()
239 v8u16 vec0, vec1, vec2, vec3; in ScaleRowDown2Box_MSA() local
250 vec0 = __msa_hadd_u_h(src0, src0); in ScaleRowDown2Box_MSA()
254 vec0 += __msa_hadd_u_h(src4, src4); in ScaleRowDown2Box_MSA()
258 vec0 = (v8u16)__msa_srari_h((v8i16)vec0, 2); in ScaleRowDown2Box_MSA()
262 dst0 = (v16u8)__msa_pckev_b((v16i8)vec1, (v16i8)vec0); in ScaleRowDown2Box_MSA()
276 v16u8 src0, src1, src2, src3, vec0, vec1, dst0; in ScaleRowDown4_MSA() local
284 vec0 = (v16u8)__msa_pckev_b((v16i8)src1, (v16i8)src0); in ScaleRowDown4_MSA()
286 dst0 = (v16u8)__msa_pckod_b((v16i8)vec1, (v16i8)vec0); in ScaleRowDown4_MSA()
303 v8u16 vec0, vec1, vec2, vec3; in ScaleRowDown4Box_MSA() local
315 vec0 = __msa_hadd_u_h(src0, src0); in ScaleRowDown4Box_MSA()
319 vec0 += __msa_hadd_u_h(src4, src4); in ScaleRowDown4Box_MSA()
331 vec0 += __msa_hadd_u_h(src0, src0); in ScaleRowDown4Box_MSA()
335 vec0 += __msa_hadd_u_h(src4, src4); in ScaleRowDown4Box_MSA()
339 reg0 = __msa_hadd_u_w(vec0, vec0); in ScaleRowDown4Box_MSA()
347 vec0 = (v8u16)__msa_pckev_h((v8i16)reg1, (v8i16)reg0); in ScaleRowDown4Box_MSA()
349 dst0 = (v16u8)__msa_pckev_b((v16i8)vec1, (v16i8)vec0); in ScaleRowDown4Box_MSA()
366 v16u8 src0, src1, vec0; in ScaleRowDown38_MSA() local
376 vec0 = (v16u8)__msa_vshf_b(mask, (v16i8)src1, (v16i8)src0); in ScaleRowDown38_MSA()
377 dst0 = __msa_copy_u_d((v2i64)vec0, 0); in ScaleRowDown38_MSA()
378 dst1 = __msa_copy_u_w((v4i32)vec0, 2); in ScaleRowDown38_MSA()
396 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in ScaleRowDown38_2_Box_MSA() local
412 vec0 = (v8u16)__msa_ilvr_b((v16i8)src2, (v16i8)src0); in ScaleRowDown38_2_Box_MSA()
416 vec0 = __msa_hadd_u_h((v16u8)vec0, (v16u8)vec0); in ScaleRowDown38_2_Box_MSA()
420 vec4 = (v8u16)__msa_vshf_h(mask, zero, (v8i16)vec0); in ScaleRowDown38_2_Box_MSA()
424 vec0 = (v8u16)__msa_pckod_w((v4i32)vec1, (v4i32)vec0); in ScaleRowDown38_2_Box_MSA()
426 vec0 = (v8u16)__msa_pckod_w((v4i32)vec1, (v4i32)vec0); in ScaleRowDown38_2_Box_MSA()
431 tmp4 = __msa_hadd_u_w(vec0, vec0); in ScaleRowDown38_2_Box_MSA()
432 vec0 = (v8u16)__msa_pckev_h((v8i16)tmp1, (v8i16)tmp0); in ScaleRowDown38_2_Box_MSA()
434 tmp0 = __msa_hadd_u_w(vec0, vec0); in ScaleRowDown38_2_Box_MSA()
442 vec0 = (v8u16)__msa_pckev_h((v8i16)tmp1, (v8i16)tmp0); in ScaleRowDown38_2_Box_MSA()
444 out = (v16u8)__msa_vshf_b(dst_mask, (v16i8)vec1, (v16i8)vec0); in ScaleRowDown38_2_Box_MSA()
466 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in ScaleRowDown38_3_Box_MSA() local
484 vec0 = (v8u16)__msa_ilvr_b((v16i8)src2, (v16i8)src0); in ScaleRowDown38_3_Box_MSA()
492 vec0 = __msa_hadd_u_h((v16u8)vec0, (v16u8)vec0); in ScaleRowDown38_3_Box_MSA()
496 vec0 += __msa_hadd_u_h((v16u8)vec4, (v16u8)vec4); in ScaleRowDown38_3_Box_MSA()
500 vec4 = (v8u16)__msa_vshf_h(mask, (v8i16)zero, (v8i16)vec0); in ScaleRowDown38_3_Box_MSA()
504 vec0 = (v8u16)__msa_pckod_w((v4i32)vec1, (v4i32)vec0); in ScaleRowDown38_3_Box_MSA()
506 vec0 = (v8u16)__msa_pckod_w((v4i32)vec1, (v4i32)vec0); in ScaleRowDown38_3_Box_MSA()
511 tmp4 = __msa_hadd_u_w(vec0, vec0); in ScaleRowDown38_3_Box_MSA()
512 vec0 = (v8u16)__msa_pckev_h((v8i16)tmp1, (v8i16)tmp0); in ScaleRowDown38_3_Box_MSA()
514 tmp0 = __msa_hadd_u_w(vec0, vec0); in ScaleRowDown38_3_Box_MSA()
522 vec0 = (v8u16)__msa_pckev_h((v8i16)tmp1, (v8i16)tmp0); in ScaleRowDown38_3_Box_MSA()
524 out = (v16u8)__msa_vshf_b(dst_mask, (v16i8)vec1, (v16i8)vec0); in ScaleRowDown38_3_Box_MSA()
565 v4i32 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9; in ScaleFilterCols_MSA() local
572 vec0 = vec_dx * vec_const; in ScaleFilterCols_MSA()
574 vec_x += vec0; in ScaleFilterCols_MSA()
644 v4i32 vec0, vec1, vec2; in ScaleARGBCols_MSA() local
647 vec0 = dx_vec * const_vec; in ScaleARGBCols_MSA()
649 x_vec += vec0; in ScaleARGBCols_MSA()
668 v4u32 vec0, vec1, vec2, vec3; in ScaleARGBFilterCols_MSA() local
678 vec0 = vec_dx * vec_const; in ScaleARGBFilterCols_MSA()
680 vec_x += vec0; in ScaleARGBFilterCols_MSA()
732 v16u8 vec0, vec1, vec2; in ScaleRowDown34_MSA() local
745 vec0 = (v16u8)__msa_vshf_b(mask0, (v16i8)src1, (v16i8)src0); in ScaleRowDown34_MSA()
748 __msa_st_b((v16i8)vec0, dst, 0); in ScaleRowDown34_MSA()
764 v16u8 vec0, vec1, vec2, vec3, vec4, vec5; in ScaleRowDown34_0_Box_MSA() local
790 vec0 = (v16u8)__msa_vshf_b(mask0, (v16i8)src0, (v16i8)src0); in ScaleRowDown34_0_Box_MSA()
802 reg0 = (v8i16)__msa_dotp_u_h(vec0, const0); in ScaleRowDown34_0_Box_MSA()
858 v16u8 vec0, vec1, vec2, vec3, vec4, vec5; in ScaleRowDown34_1_Box_MSA() local
884 vec0 = (v16u8)__msa_vshf_b(mask0, (v16i8)src0, (v16i8)src0); in ScaleRowDown34_1_Box_MSA()
896 reg0 = (v8i16)__msa_dotp_u_h(vec0, const0); in ScaleRowDown34_1_Box_MSA()