• Home
  • Raw
  • Download

Lines Matching refs:vec0

33 #define HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1,          \  argument
39 MUL4(mul_val_h0, vec0, mul_val_h2, vec0, mul_val_h0, vec1, \
68 v8i16 vec0, vec1, vec2; in hevc_intra_pred_vert_4x4_msa() local
79 vec0 = __msa_fill_h(src_left[-1]); in hevc_intra_pred_vert_4x4_msa()
83 vec2 -= vec0; in hevc_intra_pred_vert_4x4_msa()
103 v8i16 vec0, vec1, vec2; in hevc_intra_pred_vert_8x8_msa() local
118 vec0 = __msa_fill_h(src_left[-1]); in hevc_intra_pred_vert_8x8_msa()
122 vec2 -= vec0; in hevc_intra_pred_vert_8x8_msa()
158 v8i16 vec0, vec1, vec2, vec3; in hevc_intra_pred_vert_16x16_msa() local
170 vec0 = __msa_fill_h(src_left[-1]); in hevc_intra_pred_vert_16x16_msa()
174 SUB2(vec2, vec0, vec3, vec0, vec2, vec3); in hevc_intra_pred_vert_16x16_msa()
351 v8u16 sum, vec0, vec1; in hevc_intra_pred_dc_4x4_msa() local
366 ILVR_B2_UH(zero, store, zero, src, vec0, vec1); in hevc_intra_pred_dc_4x4_msa()
368 vec1 += vec0; in hevc_intra_pred_dc_4x4_msa()
369 vec0 += vec0; in hevc_intra_pred_dc_4x4_msa()
370 vec1 += vec0; in hevc_intra_pred_dc_4x4_msa()
412 v8u16 sum, vec0, vec1; in hevc_intra_pred_dc_8x8_msa() local
434 ILVR_B2_UH(zero, store, zero, src, vec0, vec1); in hevc_intra_pred_dc_8x8_msa()
436 vec1 += vec0; in hevc_intra_pred_dc_8x8_msa()
437 vec0 += vec0; in hevc_intra_pred_dc_8x8_msa()
438 vec1 += vec0; in hevc_intra_pred_dc_8x8_msa()
449 vec0 = (v8u16) __msa_fill_h(addition); in hevc_intra_pred_dc_8x8_msa()
450 vec0 *= 3; in hevc_intra_pred_dc_8x8_msa()
451 vec1 += vec0; in hevc_intra_pred_dc_8x8_msa()
470 v8u16 vec0, vec1, vec2; in hevc_intra_pred_dc_16x16_msa() local
492 vec0 = (v8u16) __msa_ilvr_b(zero, (v16i8) store); in hevc_intra_pred_dc_16x16_msa()
494 ADD2(vec1, vec0, vec2, vec0, vec1, vec2); in hevc_intra_pred_dc_16x16_msa()
495 vec0 += vec0; in hevc_intra_pred_dc_16x16_msa()
496 ADD2(vec1, vec0, vec2, vec0, vec1, vec2); in hevc_intra_pred_dc_16x16_msa()
504 vec0 = (v8u16) __msa_fill_h(addition); in hevc_intra_pred_dc_16x16_msa()
505 vec0 *= 3; in hevc_intra_pred_dc_16x16_msa()
506 ADD2(vec1, vec0, vec2, vec0, vec1, vec2); in hevc_intra_pred_dc_16x16_msa()
555 v8i16 vec0, vec1, vec2, vec3, res0, res1, res2, res3; in hevc_intra_pred_plane_4x4_msa() local
568 SPLATI_H4_SH(src1_r, 0, 1, 2, 3, vec0, vec1, vec2, vec3); in hevc_intra_pred_plane_4x4_msa()
573 MUL4(mul_val0, vec0, mul_val0, vec1, mul_val0, vec2, mul_val0, vec3, in hevc_intra_pred_plane_4x4_msa()
602 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in hevc_intra_pred_plane_8x8_msa() local
616 SPLATI_H4_SH(src_vec1_r, 0, 1, 2, 3, vec0, vec1, vec2, vec3); in hevc_intra_pred_plane_8x8_msa()
622 MUL4(mul_val0, vec0, mul_val0, vec1, mul_val0, vec2, mul_val0, vec3, in hevc_intra_pred_plane_8x8_msa()
669 v8i16 vec0, vec1; in hevc_intra_pred_plane_16x16_msa() local
687 SPLATI_H2_SH(src1_r, 0, 1, vec0, vec1); in hevc_intra_pred_plane_16x16_msa()
688 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in hevc_intra_pred_plane_16x16_msa()
694 SPLATI_H2_SH(src1_r, 2, 3, vec0, vec1); in hevc_intra_pred_plane_16x16_msa()
695 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in hevc_intra_pred_plane_16x16_msa()
701 SPLATI_H2_SH(src1_r, 4, 5, vec0, vec1); in hevc_intra_pred_plane_16x16_msa()
702 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in hevc_intra_pred_plane_16x16_msa()
708 SPLATI_H2_SH(src1_r, 6, 7, vec0, vec1); in hevc_intra_pred_plane_16x16_msa()
709 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in hevc_intra_pred_plane_16x16_msa()
715 SPLATI_H2_SH(src1_l, 0, 1, vec0, vec1); in hevc_intra_pred_plane_16x16_msa()
716 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in hevc_intra_pred_plane_16x16_msa()
722 SPLATI_H2_SH(src1_l, 2, 3, vec0, vec1); in hevc_intra_pred_plane_16x16_msa()
723 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in hevc_intra_pred_plane_16x16_msa()
729 SPLATI_H2_SH(src1_l, 4, 5, vec0, vec1); in hevc_intra_pred_plane_16x16_msa()
730 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in hevc_intra_pred_plane_16x16_msa()
736 SPLATI_H2_SH(src1_l, 6, 7, vec0, vec1); in hevc_intra_pred_plane_16x16_msa()
737 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in hevc_intra_pred_plane_16x16_msa()
750 v8i16 vec0, vec1, res0, res1; in process_intra_upper_16x16_msa() local
770 SPLATI_H2_SH(src1_r, 0, 1, vec0, vec1); in process_intra_upper_16x16_msa()
771 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_upper_16x16_msa()
777 SPLATI_H2_SH(src1_r, 2, 3, vec0, vec1); in process_intra_upper_16x16_msa()
778 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_upper_16x16_msa()
784 SPLATI_H2_SH(src1_r, 4, 5, vec0, vec1); in process_intra_upper_16x16_msa()
785 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_upper_16x16_msa()
791 SPLATI_H2_SH(src1_r, 6, 7, vec0, vec1); in process_intra_upper_16x16_msa()
792 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_upper_16x16_msa()
798 SPLATI_H2_SH(src1_l, 0, 1, vec0, vec1); in process_intra_upper_16x16_msa()
799 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_upper_16x16_msa()
805 SPLATI_H2_SH(src1_l, 2, 3, vec0, vec1); in process_intra_upper_16x16_msa()
806 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_upper_16x16_msa()
812 SPLATI_H2_SH(src1_l, 4, 5, vec0, vec1); in process_intra_upper_16x16_msa()
813 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_upper_16x16_msa()
819 SPLATI_H2_SH(src1_l, 6, 7, vec0, vec1); in process_intra_upper_16x16_msa()
820 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_upper_16x16_msa()
833 v8i16 vec0, vec1, res0, res1, tmp0, tmp1; in process_intra_lower_16x16_msa() local
852 SPLATI_H2_SH(src1_r, 0, 1, vec0, vec1); in process_intra_lower_16x16_msa()
853 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_lower_16x16_msa()
859 SPLATI_H2_SH(src1_r, 2, 3, vec0, vec1); in process_intra_lower_16x16_msa()
860 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_lower_16x16_msa()
866 SPLATI_H2_SH(src1_r, 4, 5, vec0, vec1); in process_intra_lower_16x16_msa()
867 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_lower_16x16_msa()
873 SPLATI_H2_SH(src1_r, 6, 7, vec0, vec1); in process_intra_lower_16x16_msa()
874 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_lower_16x16_msa()
880 SPLATI_H2_SH(src1_l, 0, 1, vec0, vec1); in process_intra_lower_16x16_msa()
881 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_lower_16x16_msa()
887 SPLATI_H2_SH(src1_l, 2, 3, vec0, vec1); in process_intra_lower_16x16_msa()
888 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_lower_16x16_msa()
894 SPLATI_H2_SH(src1_l, 4, 5, vec0, vec1); in process_intra_lower_16x16_msa()
895 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_lower_16x16_msa()
901 SPLATI_H2_SH(src1_l, 6, 7, vec0, vec1); in process_intra_lower_16x16_msa()
902 HEVC_PRED_PLANAR_16x2(src0_r, src0_l, tmp0, tmp1, vec0, vec1, in process_intra_lower_16x16_msa()
1908 v16u8 vec0; in ff_intra_pred_8_16x16_msa() local
2044 vec0 = (v16u8) __msa_ldi_b(128); in ff_intra_pred_8_16x16_msa()
2046 ST_UB4(vec0, vec0, vec0, vec0, left, 16); in ff_intra_pred_8_16x16_msa()
2048 ST_UB4(vec0, vec0, vec0, vec0, top, 16); in ff_intra_pred_8_16x16_msa()
2057 vec0 = LD_UB(src - stride); in ff_intra_pred_8_16x16_msa()
2058 ST_UB(vec0, top); in ff_intra_pred_8_16x16_msa()
2061 vec0 = LD_UB(src - stride + 16); in ff_intra_pred_8_16x16_msa()
2062 ST_UB(vec0, (top + 16)); in ff_intra_pred_8_16x16_msa()
2227 vec0 = (v16u8) __msa_fill_b(left[-1]); in ff_intra_pred_8_16x16_msa()
2229 ST_UB(vec0, left); in ff_intra_pred_8_16x16_msa()
2233 vec0 = (v16u8) __msa_fill_b(left[15]); in ff_intra_pred_8_16x16_msa()
2235 ST_UB(vec0, (left + 16)); in ff_intra_pred_8_16x16_msa()
2313 vec0 = (v16u8) __msa_fill_b(left[15]); in ff_intra_pred_8_16x16_msa()
2315 ST_UB(vec0, (left + 16)); in ff_intra_pred_8_16x16_msa()
2318 vec0 = (v16u8) __msa_fill_b(left[-1]); in ff_intra_pred_8_16x16_msa()
2320 ST_UB2(vec0, vec0, left, 16); in ff_intra_pred_8_16x16_msa()
2326 vec0 = (v16u8) __msa_fill_b(left[-1]); in ff_intra_pred_8_16x16_msa()
2328 ST_UB2(vec0, vec0, left, 16); in ff_intra_pred_8_16x16_msa()
2333 vec0 = (v16u8) __msa_fill_b(top[16]); in ff_intra_pred_8_16x16_msa()
2335 ST_UB(vec0, top); in ff_intra_pred_8_16x16_msa()
2339 ST_UB2(vec0, vec0, left, 16); in ff_intra_pred_8_16x16_msa()
2346 vec0 = (v16u8) __msa_ldi_b(128); in ff_intra_pred_8_16x16_msa()
2348 ST_UB2(vec0, vec0, top, 16); in ff_intra_pred_8_16x16_msa()
2349 ST_UB2(vec0, vec0, left, 16); in ff_intra_pred_8_16x16_msa()
2354 vec0 = (v16u8) __msa_fill_b(left[16]); in ff_intra_pred_8_16x16_msa()
2355 ST_UB(vec0, left); in ff_intra_pred_8_16x16_msa()
2361 vec0 = (v16u8) __msa_fill_b(left[-1]); in ff_intra_pred_8_16x16_msa()
2362 ST_UB(vec0, top); in ff_intra_pred_8_16x16_msa()
2365 vec0 = (v16u8) __msa_fill_b(top[15]); in ff_intra_pred_8_16x16_msa()
2366 ST_UB(vec0, (top + 16)); in ff_intra_pred_8_16x16_msa()
2421 v16u8 vec0, vec1; in ff_intra_pred_8_32x32_msa() local
2560 vec0 = (v16u8) __msa_ldi_b(128); in ff_intra_pred_8_32x32_msa()
2562 ST_UB4(vec0, vec0, vec0, vec0, left, 16); in ff_intra_pred_8_32x32_msa()
2563 ST_UB4(vec0, vec0, vec0, vec0, top, 16); in ff_intra_pred_8_32x32_msa()
2572 LD_UB2(src - stride, 16, vec0, vec1); in ff_intra_pred_8_32x32_msa()
2573 ST_UB2(vec0, vec1, top, 16); in ff_intra_pred_8_32x32_msa()
2577 LD_UB2(src - stride + 32, 16, vec0, vec1); in ff_intra_pred_8_32x32_msa()
2578 ST_UB2(vec0, vec1, (top + 32), 16); in ff_intra_pred_8_32x32_msa()
2742 vec0 = (v16u8) __msa_fill_b(left[-1]); in ff_intra_pred_8_32x32_msa()
2744 ST_UB2(vec0, vec0, left, 16); in ff_intra_pred_8_32x32_msa()
2747 vec0 = (v16u8) __msa_fill_b(left[31]); in ff_intra_pred_8_32x32_msa()
2749 ST_UB2(vec0, vec0, (left + 32), 16); in ff_intra_pred_8_32x32_msa()
2827 vec0 = (v16u8) __msa_fill_b(left[31]); in ff_intra_pred_8_32x32_msa()
2829 ST_UB2(vec0, vec0, (left + 32), 16); in ff_intra_pred_8_32x32_msa()
2831 vec0 = (v16u8) __msa_fill_b(left[-1]); in ff_intra_pred_8_32x32_msa()
2833 ST_UB4(vec0, vec0, vec0, vec0, left, 16); in ff_intra_pred_8_32x32_msa()
2839 vec0 = (v16u8) __msa_fill_b(left[-1]); in ff_intra_pred_8_32x32_msa()
2841 ST_UB4(vec0, vec0, vec0, vec0, left, 16); in ff_intra_pred_8_32x32_msa()
2846 vec0 = (v16u8) __msa_fill_b(top[32]); in ff_intra_pred_8_32x32_msa()
2848 ST_UB2(vec0, vec0, top, 16); in ff_intra_pred_8_32x32_msa()
2852 ST_UB4(vec0, vec0, vec0, vec0, left, 16); in ff_intra_pred_8_32x32_msa()
2860 vec0 = (v16u8) __msa_ldi_b(128); in ff_intra_pred_8_32x32_msa()
2862 ST_UB4(vec0, vec0, vec0, vec0, top, 16); in ff_intra_pred_8_32x32_msa()
2863 ST_UB4(vec0, vec0, vec0, vec0, left, 16); in ff_intra_pred_8_32x32_msa()
2868 vec0 = (v16u8) __msa_fill_b(left[32]); in ff_intra_pred_8_32x32_msa()
2870 ST_UB2(vec0, vec0, left, 16); in ff_intra_pred_8_32x32_msa()
2876 vec0 = (v16u8) __msa_fill_b(left[-1]); in ff_intra_pred_8_32x32_msa()
2878 ST_UB2(vec0, vec0, top, 16); in ff_intra_pred_8_32x32_msa()
2881 vec0 = (v16u8) __msa_fill_b(top[31]); in ff_intra_pred_8_32x32_msa()
2883 ST_UB2(vec0, vec0, (top + 32), 16); in ff_intra_pred_8_32x32_msa()
2949 vec0 = (v16u8) __msa_pckev_b((v16i8) res1, (v16i8) res0); in ff_intra_pred_8_32x32_msa()
2952 ST_UB2(vec0, vec1, filtered_top, 16); in ff_intra_pred_8_32x32_msa()
2977 vec0 = (v16u8) __msa_pckev_b((v16i8) res1, (v16i8) res0); in ff_intra_pred_8_32x32_msa()
2980 ST_UB2(vec0, vec1, (filtered_top + 32), 16); in ff_intra_pred_8_32x32_msa()
3008 vec0 = (v16u8) __msa_pckev_b((v16i8) res1, (v16i8) res0); in ff_intra_pred_8_32x32_msa()
3011 ST_UB2(vec0, vec1, left, 16); in ff_intra_pred_8_32x32_msa()
3036 vec0 = (v16u8) __msa_pckev_b((v16i8) res1, (v16i8) res0); in ff_intra_pred_8_32x32_msa()
3039 ST_UB2(vec0, vec1, (left + 32), 16); in ff_intra_pred_8_32x32_msa()