Lines Matching refs:in0
245 #define SW4(in0, in1, in2, in3, pdst, stride) \ argument
247 SW(in0, (pdst)) \
260 #define SD4(in0, in1, in2, in3, pdst, stride) \ argument
262 SD(in0, (pdst)) \
362 #define ST_V2(RTYPE, in0, in1, pdst, stride) \ argument
364 ST_V(RTYPE, in0, (pdst)); \
373 #define ST_V4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument
375 ST_V2(RTYPE, in0, in1, (pdst), stride); \
383 #define ST_V6(RTYPE, in0, in1, in2, in3, in4, in5, pdst, stride) \ argument
385 ST_V4(RTYPE, in0, in1, in2, in3, (pdst), stride); \
390 #define ST_V8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
392 ST_V4(RTYPE, in0, in1, in2, in3, (pdst), stride); \
474 #define ST_W8(in0, in1, idx0, idx1, idx2, idx3, \ argument
477 ST_W4(in0, idx0, idx1, idx2, idx3, pdst, stride) \
503 #define ST_D4(in0, in1, idx0, idx1, idx2, idx3, pdst, stride) \ argument
506 out0_m = __msa_copy_u_d((v2i64) in0, idx0); \
507 out1_m = __msa_copy_u_d((v2i64) in0, idx1); \
515 #define ST_D8(in0, in1, in2, in3, idx0, idx1, idx2, idx3, \ argument
518 ST_D4(in0, in1, idx0, idx1, idx2, idx3, pdst, stride) \
531 #define ST12x8_UB(in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
539 out0_m = __msa_copy_u_d((v2i64) in0, 0); \
548 out8_m = __msa_copy_u_w((v4i32) in0, 2); \
594 #define AVER_UB2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
596 out0 = (RTYPE) __msa_aver_u_b((v16u8) in0, (v16u8) in1); \
601 #define AVER_UB4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
604 AVER_UB2(RTYPE, in0, in1, in2, in3, out0, out1) \
660 #define VSHF_B2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1) \ argument
662 out0 = (RTYPE) __msa_vshf_b((v16i8) mask0, (v16i8) in1, (v16i8) in0); \
670 #define VSHF_B3(RTYPE, in0, in1, in2, in3, in4, in5, mask0, mask1, mask2, \ argument
673 VSHF_B2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1); \
678 #define VSHF_B4(RTYPE, in0, in1, mask0, mask1, mask2, mask3, \ argument
681 VSHF_B2(RTYPE, in0, in1, in0, in1, mask0, mask1, out0, out1); \
682 VSHF_B2(RTYPE, in0, in1, in0, in1, mask2, mask3, out2, out3); \
696 #define VSHF_H2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1) \ argument
698 out0 = (RTYPE) __msa_vshf_h((v8i16) mask0, (v8i16) in1, (v8i16) in0); \
703 #define VSHF_H3(RTYPE, in0, in1, in2, in3, in4, in5, mask0, mask1, mask2, \ argument
706 VSHF_H2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1); \
720 #define VSHF_W2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1) \ argument
722 out0 = (RTYPE) __msa_vshf_w((v4i32) mask0, (v4i32) in1, (v4i32) in0); \
905 #define MIN_UH2(RTYPE, in0, in1, min_vec) \ argument
907 in0 = (RTYPE) __msa_min_u_h((v8u16) in0, min_vec); \
912 #define MIN_UH4(RTYPE, in0, in1, in2, in3, min_vec) \ argument
914 MIN_UH2(RTYPE, in0, in1, min_vec); \
945 #define CLIP_SH2_0_255(in0, in1) \ argument
947 CLIP_SH_0_255(in0); \
951 #define CLIP_SH4_0_255(in0, in1, in2, in3) \ argument
953 CLIP_SH2_0_255(in0, in1); \
957 #define CLIP_SH8_0_255(in0, in1, in2, in3, \ argument
960 CLIP_SH4_0_255(in0, in1, in2, in3); \
976 #define CLIP_SW2_0_255(in0, in1) \ argument
978 CLIP_SW_0_255(in0); \
982 #define CLIP_SW4_0_255(in0, in1, in2, in3) \ argument
984 CLIP_SW2_0_255(in0, in1); \
988 #define CLIP_SW8_0_255(in0, in1, in2, in3, \ argument
991 CLIP_SW4_0_255(in0, in1, in2, in3); \
1043 #define HADD_SB2(RTYPE, in0, in1, out0, out1) \ argument
1045 out0 = (RTYPE) __msa_hadd_s_h((v16i8) in0, (v16i8) in0); \
1050 #define HADD_SB4(RTYPE, in0, in1, in2, in3, out0, out1, out2, out3) \ argument
1052 HADD_SB2(RTYPE, in0, in1, out0, out1); \
1066 #define HADD_UB2(RTYPE, in0, in1, out0, out1) \ argument
1068 out0 = (RTYPE) __msa_hadd_u_h((v16u8) in0, (v16u8) in0); \
1073 #define HADD_UB3(RTYPE, in0, in1, in2, out0, out1, out2) \ argument
1075 HADD_UB2(RTYPE, in0, in1, out0, out1); \
1080 #define HADD_UB4(RTYPE, in0, in1, in2, in3, out0, out1, out2, out3) \ argument
1082 HADD_UB2(RTYPE, in0, in1, out0, out1); \
1097 #define HSUB_UB2(RTYPE, in0, in1, out0, out1) \ argument
1099 out0 = (RTYPE) __msa_hsub_u_h((v16u8) in0, (v16u8) in0); \
1105 #define HSUB_UB4(RTYPE, in0, in1, in2, in3, out0, out1, out2, out3) \ argument
1107 HSUB_UB2(RTYPE, in0, in1, out0, out1); \
1123 #define SAD_UB2_UH(in0, in1, ref0, ref1) \ argument
1126 sad_m += __builtin_msa2_sad_adj2_u_w2x_b((v16u8) in0, (v16u8) ref0); \
1131 #define SAD_UB2_UH(in0, in1, ref0, ref1) \ argument
1136 diff0_m = __msa_asub_u_b((v16u8) in0, (v16u8) ref0); \
1152 #define INSERT_W2(RTYPE, in0, in1, out) \ argument
1154 out = (RTYPE) __msa_insert_w((v4i32) out, 0, in0); \
1160 #define INSERT_W4(RTYPE, in0, in1, in2, in3, out) \ argument
1162 out = (RTYPE) __msa_insert_w((v4i32) out, 0, in0); \
1178 #define INSERT_D2(RTYPE, in0, in1, out) \ argument
1180 out = (RTYPE) __msa_insert_d((v2i64) out, 0, in0); \
1197 #define ILVEV_B2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1199 out0 = (RTYPE) __msa_ilvev_b((v16i8) in1, (v16i8) in0); \
1216 #define ILVEV_H2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1218 out0 = (RTYPE) __msa_ilvev_h((v8i16) in1, (v8i16) in0); \
1234 #define ILVEV_W2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1236 out0 = (RTYPE) __msa_ilvev_w((v4i32) in1, (v4i32) in0); \
1253 #define ILVEV_D2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1255 out0 = (RTYPE) __msa_ilvev_d((v2i64) in1, (v2i64) in0); \
1271 #define ILVL_B2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1273 out0 = (RTYPE) __msa_ilvl_b((v16i8) in0, (v16i8) in1); \
1281 #define ILVL_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1284 ILVL_B2(RTYPE, in0, in1, in2, in3, out0, out1); \
1301 #define ILVL_H2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1303 out0 = (RTYPE) __msa_ilvl_h((v8i16) in0, (v8i16) in1); \
1309 #define ILVL_H4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1312 ILVL_H2(RTYPE, in0, in1, in2, in3, out0, out1); \
1327 #define ILVL_W2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1329 out0 = (RTYPE) __msa_ilvl_w((v4i32) in0, (v4i32) in1); \
1346 #define ILVR_B2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1348 out0 = (RTYPE) __msa_ilvr_b((v16i8) in0, (v16i8) in1); \
1357 #define ILVR_B3(RTYPE, in0, in1, in2, in3, in4, in5, out0, out1, out2) \ argument
1359 ILVR_B2(RTYPE, in0, in1, in2, in3, out0, out1); \
1367 #define ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1370 ILVR_B2(RTYPE, in0, in1, in2, in3, out0, out1); \
1379 #define ILVR_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1383 ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \
1401 #define ILVR_H2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1403 out0 = (RTYPE) __msa_ilvr_h((v8i16) in0, (v8i16) in1); \
1409 #define ILVR_H3(RTYPE, in0, in1, in2, in3, in4, in5, out0, out1, out2) \ argument
1411 ILVR_H2(RTYPE, in0, in1, in2, in3, out0, out1); \
1416 #define ILVR_H4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1419 ILVR_H2(RTYPE, in0, in1, in2, in3, out0, out1); \
1425 #define ILVR_W2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1427 out0 = (RTYPE) __msa_ilvr_w((v4i32) in0, (v4i32) in1); \
1434 #define ILVR_W4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1437 ILVR_W2(RTYPE, in0, in1, in2, in3, out0, out1); \
1452 #define ILVR_D2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1454 out0 = (RTYPE) __msa_ilvr_d((v2i64) in0, (v2i64) in1); \
1461 #define ILVR_D3(RTYPE, in0, in1, in2, in3, in4, in5, out0, out1, out2) \ argument
1463 ILVR_D2(RTYPE, in0, in1, in2, in3, out0, out1); \
1468 #define ILVR_D4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1471 ILVR_D2(RTYPE, in0, in1, in2, in3, out0, out1); \
1486 #define ILVL_D2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1488 out0 = (RTYPE) __msa_ilvl_d((v2i64) in0, (v2i64) in1); \
1504 #define ILVRL_B2(RTYPE, in0, in1, out0, out1) \ argument
1506 out0 = (RTYPE) __msa_ilvr_b((v16i8) in0, (v16i8) in1); \
1507 out1 = (RTYPE) __msa_ilvl_b((v16i8) in0, (v16i8) in1); \
1515 #define ILVRL_H2(RTYPE, in0, in1, out0, out1) \ argument
1517 out0 = (RTYPE) __msa_ilvr_h((v8i16) in0, (v8i16) in1); \
1518 out1 = (RTYPE) __msa_ilvl_h((v8i16) in0, (v8i16) in1); \
1525 #define ILVRL_W2(RTYPE, in0, in1, out0, out1) \ argument
1527 out0 = (RTYPE) __msa_ilvr_w((v4i32) in0, (v4i32) in1); \
1528 out1 = (RTYPE) __msa_ilvl_w((v4i32) in0, (v4i32) in1); \
1542 #define MAXI_SH2(RTYPE, in0, in1, max_val) \ argument
1544 in0 = (RTYPE) __msa_maxi_s_h((v8i16) in0, max_val); \
1550 #define MAXI_SH4(RTYPE, in0, in1, in2, in3, max_val) \ argument
1552 MAXI_SH2(RTYPE, in0, in1, max_val); \
1558 #define MAXI_SH8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, max_val) \ argument
1560 MAXI_SH4(RTYPE, in0, in1, in2, in3, max_val); \
1576 #define SAT_UH2(RTYPE, in0, in1, sat_val) \ argument
1578 in0 = (RTYPE) __msa_sat_u_h((v8u16) in0, sat_val); \
1584 #define SAT_UH4(RTYPE, in0, in1, in2, in3, sat_val) \ argument
1586 SAT_UH2(RTYPE, in0, in1, sat_val); \
1592 #define SAT_UH8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, sat_val) \ argument
1594 SAT_UH4(RTYPE, in0, in1, in2, in3, sat_val); \
1610 #define SAT_SH2(RTYPE, in0, in1, sat_val) \ argument
1612 in0 = (RTYPE) __msa_sat_s_h((v8i16) in0, sat_val); \
1617 #define SAT_SH3(RTYPE, in0, in1, in2, sat_val) \ argument
1619 SAT_SH2(RTYPE, in0, in1, sat_val); \
1624 #define SAT_SH4(RTYPE, in0, in1, in2, in3, sat_val) \ argument
1626 SAT_SH2(RTYPE, in0, in1, sat_val); \
1641 #define SAT_SW2(RTYPE, in0, in1, sat_val) \ argument
1643 in0 = (RTYPE) __msa_sat_s_w((v4i32) in0, sat_val); \
1648 #define SAT_SW4(RTYPE, in0, in1, in2, in3, sat_val) \ argument
1650 SAT_SW2(RTYPE, in0, in1, sat_val); \
1728 #define PCKEV_B2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1730 out0 = (RTYPE) __msa_pckev_b((v16i8) in0, (v16i8) in1); \
1738 #define PCKEV_B3(RTYPE, in0, in1, in2, in3, in4, in5, out0, out1, out2) \ argument
1740 PCKEV_B2(RTYPE, in0, in1, in2, in3, out0, out1); \
1746 #define PCKEV_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1749 PCKEV_B2(RTYPE, in0, in1, in2, in3, out0, out1); \
1768 #define PCKEV_H2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1770 out0 = (RTYPE) __msa_pckev_h((v8i16) in0, (v8i16) in1); \
1776 #define PCKEV_H4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1779 PCKEV_H2(RTYPE, in0, in1, in2, in3, out0, out1); \
1796 #define PCKEV_D2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1798 out0 = (RTYPE) __msa_pckev_d((v2i64) in0, (v2i64) in1); \
1805 #define PCKEV_D4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1808 PCKEV_D2(RTYPE, in0, in1, in2, in3, out0, out1); \
1822 #define PCKOD_D2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1824 out0 = (RTYPE) __msa_pckod_d((v2i64) in0, (v2i64) in1); \
1843 #define XORI_B2_128(RTYPE, in0, in1) \ argument
1845 in0 = (RTYPE) __msa_xori_b((v16u8) in0, 128); \
1852 #define XORI_B3_128(RTYPE, in0, in1, in2) \ argument
1854 XORI_B2_128(RTYPE, in0, in1); \
1859 #define XORI_B4_128(RTYPE, in0, in1, in2, in3) \ argument
1861 XORI_B2_128(RTYPE, in0, in1); \
1868 #define XORI_B5_128(RTYPE, in0, in1, in2, in3, in4) \ argument
1870 XORI_B3_128(RTYPE, in0, in1, in2); \
1875 #define XORI_B6_128(RTYPE, in0, in1, in2, in3, in4, in5) \ argument
1877 XORI_B4_128(RTYPE, in0, in1, in2, in3); \
1882 #define XORI_B7_128(RTYPE, in0, in1, in2, in3, in4, in5, in6) \ argument
1884 XORI_B4_128(RTYPE, in0, in1, in2, in3); \
1889 #define XORI_B8_128(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7) \ argument
1891 XORI_B4_128(RTYPE, in0, in1, in2, in3); \
1906 #define ADDS_SH2(RTYPE, in0, in1, in2, in3, out0, out1) \ argument
1908 out0 = (RTYPE) __msa_adds_s_h((v8i16) in0, (v8i16) in1); \
1913 #define ADDS_SH4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1916 ADDS_SH2(RTYPE, in0, in1, in2, in3, out0, out1); \
1930 #define SLLI_2V(in0, in1, shift) \ argument
1932 in0 = in0 << shift; \
1935 #define SLLI_4V(in0, in1, in2, in3, shift) \ argument
1937 in0 = in0 << shift; \
1953 #define SRA_4V(in0, in1, in2, in3, shift) \ argument
1955 in0 = in0 >> shift; \
1971 #define SRL_H4(RTYPE, in0, in1, in2, in3, shift) \ argument
1973 in0 = (RTYPE) __msa_srl_h((v8i16) in0, (v8i16) shift); \
1980 #define SRLR_H4(RTYPE, in0, in1, in2, in3, shift) \ argument
1982 in0 = (RTYPE) __msa_srlr_h((v8i16) in0, (v8i16) shift); \
1990 #define SRLR_H8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, shift) \ argument
1992 SRLR_H4(RTYPE, in0, in1, in2, in3, shift); \
2009 #define SRAR_H2(RTYPE, in0, in1, shift) \ argument
2011 in0 = (RTYPE) __msa_srar_h((v8i16) in0, (v8i16) shift); \
2017 #define SRAR_H3(RTYPE, in0, in1, in2, shift) \ argument
2019 SRAR_H2(RTYPE, in0, in1, shift) \
2024 #define SRAR_H4(RTYPE, in0, in1, in2, in3, shift) \ argument
2026 SRAR_H2(RTYPE, in0, in1, shift) \
2043 #define SRAR_W2(RTYPE, in0, in1, shift) \ argument
2045 in0 = (RTYPE) __msa_srar_w((v4i32) in0, (v4i32) shift); \
2050 #define SRAR_W4(RTYPE, in0, in1, in2, in3, shift) \ argument
2052 SRAR_W2(RTYPE, in0, in1, shift) \
2067 #define SRARI_H2(RTYPE, in0, in1, shift) \ argument
2069 in0 = (RTYPE) __msa_srari_h((v8i16) in0, shift); \
2075 #define SRARI_H4(RTYPE, in0, in1, in2, in3, shift) \ argument
2077 SRARI_H2(RTYPE, in0, in1, shift); \
2093 #define SRARI_W2(RTYPE, in0, in1, shift) \ argument
2095 in0 = (RTYPE) __msa_srari_w((v4i32) in0, shift); \
2100 #define SRARI_W4(RTYPE, in0, in1, in2, in3, shift) \ argument
2102 SRARI_W2(RTYPE, in0, in1, shift); \
2115 #define MUL2(in0, in1, in2, in3, out0, out1) \ argument
2117 out0 = in0 * in1; \
2120 #define MUL4(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, out3) \ argument
2122 MUL2(in0, in1, in2, in3, out0, out1); \
2132 #define ADD2(in0, in1, in2, in3, out0, out1) \ argument
2134 out0 = in0 + in1; \
2137 #define ADD4(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, out3) \ argument
2139 ADD2(in0, in1, in2, in3, out0, out1); \
2149 #define SUB2(in0, in1, in2, in3, out0, out1) \ argument
2151 out0 = in0 - in1; \
2154 #define SUB4(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, out3) \ argument
2156 out0 = in0 - in1; \
2274 #define SWAP(in0, in1) \ argument
2276 in0 = in0 ^ in1; \
2277 in1 = in0 ^ in1; \
2278 in0 = in0 ^ in1; \
2286 #define BUTTERFLY_4(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
2288 out0 = in0 + in3; \
2292 out3 = in0 - in3; \
2300 #define BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2303 out0 = in0 + in7; \
2311 out7 = in0 - in7; \
2319 #define BUTTERFLY_16(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2324 out0 = in0 + in15; \
2340 out15 = in0 - in15; \
2349 #define TRANSPOSE4x4_UB_UB(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
2354 ILVR_D2_SB(in1, in0, in3, in2, s0_m, s1_m); \
2369 #define TRANSPOSE8x4_UB(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2374 ILVEV_W2_SB(in0, in4, in1, in5, tmp0_m, tmp1_m); \
2396 #define TRANSPOSE8x8_UB(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2403 ILVR_B4_SB(in2, in0, in3, in1, in6, in4, in7, in5, \
2422 #define TRANSPOSE16x4_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2428 ILVEV_W2_SD(in0, in4, in8, in12, tmp0_m, tmp1_m); \
2457 #define TRANSPOSE16x8_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2464 ILVEV_D2_UB(in0, in8, in1, in9, out7, out6); \
2503 #define TRANSPOSE4x4_SH_SH(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
2507 ILVR_H2_SH(in1, in0, in3, in2, s0_m, s1_m); \
2519 #define TRANSPOSE8x8_H(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2530 ILVR_H2_SH(in2, in0, in3, in1, s0_m, s1_m); \
2532 ILVL_H2_SH(in2, in0, in3, in1, s0_m, s1_m); \
2550 #define TRANSPOSE4x4_SW_SW(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
2554 ILVRL_W2_SW(in1, in0, s0_m, s1_m); \
2577 #define AVE_ST8x4_UB(in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
2582 tmp0_m = __msa_ave_u_b((v16u8) in0, (v16u8) in1); \
2608 #define AVE_ST16x4_UB(in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
2612 tmp0_m = __msa_ave_u_b((v16u8) in0, (v16u8) in1); \
2634 #define AVER_ST8x4_UB(in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
2639 AVER_UB4_UB(in0, in1, in2, in3, in4, in5, in6, in7, \
2663 #define AVER_ST16x4_UB(in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
2667 AVER_UB4_UB(in0, in1, in2, in3, in4, in5, in6, in7, \
2687 #define AVER_DST_ST8x4_UB(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2694 AVER_UB4_UB(in0, in1, in2, in3, in4, in5, in6, in7, \
2715 #define AVER_DST_ST16x4_UB(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2722 AVER_UB4_UB(in0, in1, in2, in3, in4, in5, in6, in7, \
2733 #define ADDBLK_ST4x4_UB(in0, in1, in2, in3, pdst, stride) \ argument
2742 ILVR_D2_SH(in1, in0, in3, in2, inp0_m, inp1_m) \
2769 #define DPADD_SH3_SH(in0, in1, in2, coeff0, coeff1, coeff2) \ argument
2773 out0_m = __msa_dotp_s_h((v16i8) in0, (v16i8) coeff0); \
2788 #define PCKEV_XORI128_UB(in0, in1) \ argument
2791 out_m = (v16u8) __msa_pckev_b((v16i8) in1, (v16i8) in0); \
2800 #define CONVERT_UB_AVG_ST8x4_UB(in0, in1, in2, in3, \ argument
2806 tmp0_m = PCKEV_XORI128_UB(in0, in1); \
2817 #define PCKEV_ST4x4_UB(in0, in1, in2, in3, pdst, stride) \ argument
2822 PCKEV_B2_SB(in1, in0, in3, in2, tmp0_m, tmp1_m); \
2836 #define PCKEV_ST_SB(in0, in1, pdst) \ argument
2839 tmp_m = __msa_pckev_b((v16i8) in1, (v16i8) in0); \
2846 #define HORIZ_2TAP_FILT_UH(in0, in1, mask, coeff, shift) \ argument
2851 tmp0_m = __msa_vshf_b((v16i8) mask, (v16i8) in1, (v16i8) in0); \