Lines Matching refs:vec_s16
108 vec_s16 vi, vileft, ditherleft, ditherright; in yuv2plane1_8_vsx()
219 const vec_s16 vzero = vec_splat_s16(0); in yuv2planeX_nbps_vsx()
221 vec_s16 vfilter[MAX_FILTER_SIZE], vin; in yuv2planeX_nbps_vsx()
227 vfilter[i] = (vec_s16) {filter[i], filter[i], filter[i], filter[i], in yuv2planeX_nbps_vsx()
247 v = (vec_u16) vec_max((vec_s16) v, vzero); in yuv2planeX_nbps_vsx()
536 vec_s16 vv; in yuv2rgb_full_X_vsx_template()
542 vec_s16 vlumFilter[MAX_FILTER_SIZE], vchrFilter[MAX_FILTER_SIZE]; in yuv2rgb_full_X_vsx_template()
693 vec_s16 vy, vu, vv, A = vec_splat_s16(0); in yuv2rgb_full_2_vsx_template()
699 const vec_s16 vyalpha1 = vec_splats(yalpha1); in yuv2rgb_full_2_vsx_template()
700 const vec_s16 vuvalpha1 = vec_splats(uvalpha1); in yuv2rgb_full_2_vsx_template()
701 const vec_s16 vyalpha = vec_splats((int16_t) yalpha); in yuv2rgb_full_2_vsx_template()
702 const vec_s16 vuvalpha = vec_splats((int16_t) uvalpha); in yuv2rgb_full_2_vsx_template()
767 ad = vec_packsu(A, (vec_s16) zero16); in yuv2rgb_full_2_vsx_template()
815 vec_s16 vy, vu, vv, A = vec_splat_s16(0); in yuv2rgb_2_vsx_template()
821 const vec_s16 vyalpha1 = vec_splats(yalpha1); in yuv2rgb_2_vsx_template()
822 const vec_s16 vuvalpha1 = vec_splats(uvalpha1); in yuv2rgb_2_vsx_template()
823 const vec_s16 vyalpha = vec_splats((int16_t) yalpha); in yuv2rgb_2_vsx_template()
824 const vec_s16 vuvalpha = vec_splats((int16_t) uvalpha); in yuv2rgb_2_vsx_template()
897 ad = vec_packsu(A, (vec_s16) zero16); in yuv2rgb_2_vsx_template()
986 vec_s16 vy, vu, vv, A = vec_splat_s16(0), tmp16; in yuv2rgb_full_1_vsx_template()
1005 const vec_s16 mul4 = vec_splat_s16(4); in yuv2rgb_full_1_vsx_template()
1006 const vec_s16 mul8 = vec_splat_s16(8); in yuv2rgb_full_1_vsx_template()
1007 const vec_s16 add64 = vec_splat_s16(64); in yuv2rgb_full_1_vsx_template()
1009 const vec_s16 max255 = vec_splat_s16(255); in yuv2rgb_full_1_vsx_template()
1042 vu = (vec_s16) vec_sub((vec_u16) vu, sub7); in yuv2rgb_full_1_vsx_template()
1043 vv = (vec_s16) vec_sub((vec_u16) vv, sub7); in yuv2rgb_full_1_vsx_template()
1056 vu = (vec_s16) vec_sub((vec_u16) vu, sub8); in yuv2rgb_full_1_vsx_template()
1059 vv = (vec_s16) vec_sub((vec_u16) vv, sub8); in yuv2rgb_full_1_vsx_template()
1072 ad = vec_packsu(A, (vec_s16) zero16); in yuv2rgb_full_1_vsx_template()
1115 vec_s16 vy, vu, vv, A = vec_splat_s16(0), tmp16; in yuv2rgb_1_vsx_template()
1135 const vec_s16 mul4 = vec_splat_s16(4); in yuv2rgb_1_vsx_template()
1136 const vec_s16 mul8 = vec_splat_s16(8); in yuv2rgb_1_vsx_template()
1137 const vec_s16 add64 = vec_splat_s16(64); in yuv2rgb_1_vsx_template()
1139 const vec_s16 max255 = vec_splat_s16(255); in yuv2rgb_1_vsx_template()
1180 vu = (vec_s16) vec_sub((vec_u16) vu, sub7); in yuv2rgb_1_vsx_template()
1181 vv = (vec_s16) vec_sub((vec_u16) vv, sub7); in yuv2rgb_1_vsx_template()
1194 vu = (vec_s16) vec_sub((vec_u16) vu, sub8); in yuv2rgb_1_vsx_template()
1197 vv = (vec_s16) vec_sub((vec_u16) vv, sub8); in yuv2rgb_1_vsx_template()
1210 ad = vec_packsu(A, (vec_s16) zero16); in yuv2rgb_1_vsx_template()
1366 write422(const vec_s16 vy1, const vec_s16 vy2, in write422()
1367 const vec_s16 vu, const vec_s16 vv, in write422()
1436 vec_s16 vy1, vy2, vu, vv; in yuv2422_X_vsx_template()
1438 vec_s16 vlumFilter[MAX_FILTER_SIZE], vchrFilter[MAX_FILTER_SIZE]; in yuv2422_X_vsx_template()
1547 vec_s16 vy1, vy2, vu, vv; in yuv2422_2_vsx_template()
1549 const vec_s16 vyalpha1 = vec_splats(yalpha1); in yuv2422_2_vsx_template()
1550 const vec_s16 vuvalpha1 = vec_splats(uvalpha1); in yuv2422_2_vsx_template()
1576 vec_s16 vy1, vy2, vu, vv, tmp; in yuv2422_1_vsx_template()
1577 const vec_s16 add64 = vec_splats((int16_t) 64); in yuv2422_1_vsx_template()
1578 const vec_s16 add128 = vec_splats((int16_t) 128); in yuv2422_1_vsx_template()
1676 vec_s16 vtmp, vtmp2, vtmp3, vtmp4; in hyscale_fast_vsx()
1752 vec_st((vec_s16) vd_l, 0, &dst[i]); in hyscale_fast_vsx()
1753 vec_st((vec_s16) vd_r, 0, &dst[i + 8]); in hyscale_fast_vsx()
1781 vec_st((vec_s16) vd_l, 0, &out[i]); \
1782 vec_st((vec_s16) vd_r, 0, &out[i + 8])
1867 vec_s16 vfilter, vin; in hScale8To19_vsx()
1906 vin = (vec_s16) vec_mergeh(vin8, vzero); in hScale8To19_vsx()
1908 vin = vec_perm(vin, (vec_s16) vzero, vunused); in hScale8To19_vsx()
1929 vec_s16 vfilter, vin; in hScale16To19_vsx()
1974 vin = (vec_s16) vec_vsx_ld(0, &src[srcPos + j]); in hScale16To19_vsx()
1976 vin = vec_perm(vin, (vec_s16) vzero, vunused); in hScale16To19_vsx()
1982 vtmp = (vec_s32) vec_mergeh(vin, (vec_s16) vzero); in hScale16To19_vsx()
1983 vtmp2 = (vec_s32) vec_mergel(vin, (vec_s16) vzero); in hScale16To19_vsx()
2005 vec_s16 vfilter, vin; in hScale16To15_vsx()
2050 vin = (vec_s16) vec_vsx_ld(0, &src[srcPos + j]); in hScale16To15_vsx()
2052 vin = vec_perm(vin, (vec_s16) vzero, vunused); in hScale16To15_vsx()
2058 vtmp = (vec_s32) vec_mergeh(vin, (vec_s16) vzero); in hScale16To15_vsx()
2059 vtmp2 = (vec_s32) vec_mergel(vin, (vec_s16) vzero); in hScale16To15_vsx()