• Home
  • Raw
  • Download

Lines Matching refs:in7

390 #define ST_V8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride)  \  argument
393 ST_V4(RTYPE, in4, in5, in6, in7, (pdst) + 4 * stride, stride); \
531 #define ST12x8_UB(in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
546 out7_m = __msa_copy_u_d((v2i64) in7, 0); \
555 out15_m = __msa_copy_u_w((v4i32) in7, 2); \
601 #define AVER_UB4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
605 AVER_UB2(RTYPE, in4, in5, in6, in7, out2, out3) \
958 in4, in5, in6, in7) \ argument
961 CLIP_SH4_0_255(in4, in5, in6, in7); \
989 in4, in5, in6, in7) \ argument
992 CLIP_SW4_0_255(in4, in5, in6, in7); \
1281 #define ILVL_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1285 ILVL_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1309 #define ILVL_H4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1313 ILVL_H2(RTYPE, in4, in5, in6, in7, out2, out3); \
1367 #define ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1371 ILVR_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1379 #define ILVR_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1383 ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \
1416 #define ILVR_H4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1420 ILVR_H2(RTYPE, in4, in5, in6, in7, out2, out3); \
1434 #define ILVR_W4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1438 ILVR_W2(RTYPE, in4, in5, in6, in7, out2, out3); \
1468 #define ILVR_D4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1472 ILVR_D2(RTYPE, in4, in5, in6, in7, out2, out3); \
1558 #define MAXI_SH8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, max_val) \ argument
1561 MAXI_SH4(RTYPE, in4, in5, in6, in7, max_val); \
1592 #define SAT_UH8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, sat_val) \ argument
1595 SAT_UH4(RTYPE, in4, in5, in6, in7, sat_val); \
1746 #define PCKEV_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1750 PCKEV_B2(RTYPE, in4, in5, in6, in7, out2, out3); \
1776 #define PCKEV_H4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1780 PCKEV_H2(RTYPE, in4, in5, in6, in7, out2, out3); \
1805 #define PCKEV_D4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1809 PCKEV_D2(RTYPE, in4, in5, in6, in7, out2, out3); \
1889 #define XORI_B8_128(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7) \ argument
1892 XORI_B4_128(RTYPE, in4, in5, in6, in7); \
1913 #define ADDS_SH4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
1917 ADDS_SH2(RTYPE, in4, in5, in6, in7, out2, out3); \
1990 #define SRLR_H8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, shift) \ argument
1993 SRLR_H4(RTYPE, in4, in5, in6, in7, shift); \
2120 #define MUL4(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, out3) \ argument
2123 MUL2(in4, in5, in6, in7, out2, out3); \
2137 #define ADD4(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, out3) \ argument
2140 ADD2(in4, in5, in6, in7, out2, out3); \
2154 #define SUB4(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, out3) \ argument
2159 out3 = in6 - in7; \
2300 #define BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2303 out0 = in0 + in7; \
2311 out7 = in0 - in7; \
2319 #define BUTTERFLY_16(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2331 out7 = in7 + in8; \
2333 out8 = in7 - in8; \
2369 #define TRANSPOSE8x4_UB(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2376 ILVEV_W2_SB(in2, in6, in3, in7, tmp0_m, tmp1_m); \
2396 #define TRANSPOSE8x8_UB(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2403 ILVR_B4_SB(in2, in0, in3, in1, in6, in4, in7, in5, \
2422 #define TRANSPOSE16x4_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2437 ILVEV_W2_SD(in3, in7, in11, in15, tmp0_m, tmp1_m); \
2457 #define TRANSPOSE16x8_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2467 ILVEV_D2_UB(in6, in14, in7, in15, out1, out0); \
2519 #define TRANSPOSE8x8_H(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2526 ILVR_H2_SH(in6, in4, in7, in5, s0_m, s1_m); \
2528 ILVL_H2_SH(in6, in4, in7, in5, s0_m, s1_m); \
2577 #define AVE_ST8x4_UB(in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
2585 tmp3_m = __msa_ave_u_b((v16u8) in6, (v16u8) in7); \
2608 #define AVE_ST16x4_UB(in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
2615 tmp3_m = __msa_ave_u_b((v16u8) in6, (v16u8) in7); \
2634 #define AVER_ST8x4_UB(in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
2639 AVER_UB4_UB(in0, in1, in2, in3, in4, in5, in6, in7, \
2663 #define AVER_ST16x4_UB(in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
2667 AVER_UB4_UB(in0, in1, in2, in3, in4, in5, in6, in7, \
2687 #define AVER_DST_ST8x4_UB(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2694 AVER_UB4_UB(in0, in1, in2, in3, in4, in5, in6, in7, \
2715 #define AVER_DST_ST16x4_UB(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
2722 AVER_UB4_UB(in0, in1, in2, in3, in4, in5, in6, in7, \