• Home
  • Raw
  • Download

Lines Matching refs:vuc

17 vector unsigned char vuc = { 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16 };  variable
123 res_vuc = vec_add(vuc, vuc); in test1()
127 res_vuc = vec_add(vbc, vuc); in test1()
131 res_vuc = vec_add(vuc, vbc); in test1()
199 res_vuc = vec_vaddubm(vuc, vuc); in test1()
203 res_vuc = vec_vaddubm(vbc, vuc); in test1()
207 res_vuc = vec_vaddubm(vuc, vbc); in test1()
285 res_vuc = vec_adds(vuc, vuc); in test1()
289 res_vuc = vec_adds(vbc, vuc); in test1()
293 res_vuc = vec_adds(vuc, vbc); in test1()
357 res_vuc = vec_vaddubs(vuc, vuc); in test1()
361 res_vuc = vec_vaddubs(vbc, vuc); in test1()
365 res_vuc = vec_vaddubs(vuc, vbc); in test1()
430 res_vuc = vec_and(vuc, vuc); in test1()
434 res_vuc = vec_and(vbc, vuc); in test1()
438 res_vuc = vec_and(vuc, vbc); in test1()
514 res_vuc = vec_vand(vuc, vuc); in test1()
518 res_vuc = vec_vand(vbc, vuc); in test1()
522 res_vuc = vec_vand(vuc, vbc); in test1()
605 res_vuc = vec_andc(vuc, vuc); in test1()
611 res_vuc = vec_andc(vbc, vuc); in test1()
617 res_vuc = vec_andc(vuc, vbc); in test1()
741 res_vuc = vec_vandc(vuc, vuc); in test1()
747 res_vuc = vec_vandc(vbc, vuc); in test1()
753 res_vuc = vec_vandc(vuc, vbc); in test1()
870 res_vuc = vec_avg(vuc, vuc); in test2()
894 res_vuc = vec_vavgub(vuc, vuc); in test2()
937 res_vbc = vec_cmpeq(vuc, vuc); in test2()
966 res_vbc = vec_cmpge(vuc, vuc); in test2()
1003 res_vbc = vec_cmpgt(vuc, vuc); in test5()
1031 res_vbc = vec_vcmpgtub(vuc, vuc); in test5()
1060 res_vbc = vec_cmple(vuc, vuc); in test5()
1092 res_vbc = vec_cmplt(vuc, vuc); in test6()
1156 res_vuc = vec_div(vuc, vuc); in test6()
1233 res_vuc = vec_ld(0, &vuc); in test6()
1305 res_vuc = vec_lvx(0, &vuc); in test6()
1435 res_vuc = vec_ldl(0, &vuc); in test6()
1507 res_vuc = vec_lvxl(0, &vuc); in test6()
1621 res_vuc = vec_max(vuc, vuc); in test6()
1625 res_vuc = vec_max(vbc, vuc); in test6()
1629 res_vuc = vec_max(vuc, vbc); in test6()
1697 res_vuc = vec_vmaxub(vuc, vuc); in test6()
1701 res_vuc = vec_vmaxub(vbc, vuc); in test6()
1705 res_vuc = vec_vmaxub(vuc, vbc); in test6()
1766 res_vuc = vec_mergeh(vuc, vuc); in test6()
1810 res_vuc = vec_vmrghb(vuc, vuc); in test6()
1855 res_vuc = vec_mergel(vuc, vuc); in test6()
1899 res_vuc = vec_vmrglb(vuc, vuc); in test6()
1957 res_vuc = vec_min(vuc, vuc); in test6()
1961 res_vuc = vec_min(vbc, vuc); in test6()
1965 res_vuc = vec_min(vuc, vbc); in test6()
2033 res_vuc = vec_vminub(vuc, vuc); in test6()
2037 res_vuc = vec_vminub(vbc, vuc); in test6()
2041 res_vuc = vec_vminub(vuc, vbc); in test6()
2132 res_vi = vec_msum(vsc, vuc, vi); in test6()
2136 res_vui = vec_msum(vuc, vuc, vui); in test6()
2148 res_vi = vec_vmsummbm(vsc, vuc, vi); in test6()
2152 res_vui = vec_vmsumubm(vuc, vuc, vui); in test6()
2186 vec_mtvscr(vuc); in test6()
2227 res_vuc = vec_mul(vuc, vuc); in test6()
2252 res_vus = vec_mule(vuc, vuc); in test6()
2268 res_vus = vec_vmuleub(vuc, vuc); in test6()
2285 res_vus = vec_mulo(vuc, vuc); in test6()
2301 res_vus = vec_vmuloub(vuc, vuc); in test6()
2329 res_vuc = vec_nor(vuc, vuc); in test6()
2389 res_vuc = vec_vnor(vuc, vuc); in test6()
2456 res_vuc = vec_or(vuc, vuc); in test6()
2460 res_vuc = vec_or(vbc, vuc); in test6()
2464 res_vuc = vec_or(vuc, vbc); in test6()
2552 res_vuc = vec_vor(vuc, vuc); in test6()
2556 res_vuc = vec_vor(vbc, vuc); in test6()
2560 res_vuc = vec_vor(vuc, vbc); in test6()
2761 res_vsc = vec_perm(vsc, vsc, vuc); in test6()
2765 res_vuc = vec_perm(vuc, vuc, vuc); in test6()
2769 res_vbc = vec_perm(vbc, vbc, vuc); in test6()
2773 res_vs = vec_perm(vs, vs, vuc); in test6()
2777 res_vus = vec_perm(vus, vus, vuc); in test6()
2781 res_vbs = vec_perm(vbs, vbs, vuc); in test6()
2785 res_vp = vec_perm(vp, vp, vuc); in test6()
2789 res_vi = vec_perm(vi, vi, vuc); in test6()
2793 res_vui = vec_perm(vui, vui, vuc); in test6()
2797 res_vbi = vec_perm(vbi, vbi, vuc); in test6()
2801 res_vf = vec_perm(vf, vf, vuc); in test6()
2805 res_vsc = vec_vperm(vsc, vsc, vuc); in test6()
2809 res_vuc = vec_vperm(vuc, vuc, vuc); in test6()
2813 res_vbc = vec_vperm(vbc, vbc, vuc); in test6()
2817 res_vs = vec_vperm(vs, vs, vuc); in test6()
2821 res_vus = vec_vperm(vus, vus, vuc); in test6()
2825 res_vbs = vec_vperm(vbs, vbs, vuc); in test6()
2829 res_vp = vec_vperm(vp, vp, vuc); in test6()
2833 res_vi = vec_vperm(vi, vi, vuc); in test6()
2837 res_vui = vec_vperm(vui, vui, vuc); in test6()
2841 res_vbi = vec_vperm(vbi, vbi, vuc); in test6()
2845 res_vf = vec_vperm(vf, vf, vuc); in test6()
2859 res_vsc = vec_rl(vsc, vuc); in test6()
2863 res_vuc = vec_rl(vuc, vuc); in test6()
2883 res_vsc = vec_vrlb(vsc, vuc); in test6()
2887 res_vuc = vec_vrlb(vuc, vuc); in test6()
2926 res_vsc = vec_sel(vsc, vsc, vuc); in test6()
2946 res_vuc = vec_sel(vuc, vuc, vuc); in test6()
2956 res_vuc = vec_sel(vuc, vuc, vbc); in test6()
2966 res_vbc = vec_sel(vbc, vbc, vuc); in test6()
3126 res_vsc = vec_vsel(vsc, vsc, vuc); in test6()
3146 res_vuc = vec_vsel(vuc, vuc, vuc); in test6()
3156 res_vuc = vec_vsel(vuc, vuc, vbc); in test6()
3166 res_vbc = vec_vsel(vbc, vbc, vuc); in test6()
3327 res_vsc = vec_sl(vsc, vuc); in test6()
3331 res_vuc = vec_sl(vuc, vuc); in test6()
3351 res_vsc = vec_vslb(vsc, vuc); in test6()
3355 res_vuc = vec_vslb(vuc, vuc); in test6()
3388 res_vuc = vec_sld(vuc, vuc, 0); in test6()
3514 res_vuc = vec_vsldoi(vuc, vuc, 0); in test6()
3599 res_vsc = vec_sll(vsc, vuc); in test6()
3611 res_vuc = vec_sll(vuc, vuc); in test6()
3615 res_vuc = vec_sll(vuc, vus); in test6()
3619 res_vuc = vec_sll(vuc, vui); in test6()
3623 res_vbc = vec_sll(vbc, vuc); in test6()
3635 res_vs = vec_sll(vs, vuc); in test6()
3647 res_vus = vec_sll(vus, vuc); in test6()
3659 res_vbs = vec_sll(vbs, vuc); in test6()
3671 res_vp = vec_sll(vp, vuc); in test6()
3683 res_vi = vec_sll(vi, vuc); in test6()
3695 res_vui = vec_sll(vui, vuc); in test6()
3707 res_vbi = vec_sll(vbi, vuc); in test6()
3719 res_vsc = vec_vsl(vsc, vuc); in test6()
3731 res_vuc = vec_vsl(vuc, vuc); in test6()
3735 res_vuc = vec_vsl(vuc, vus); in test6()
3739 res_vuc = vec_vsl(vuc, vui); in test6()
3743 res_vbc = vec_vsl(vbc, vuc); in test6()
3755 res_vs = vec_vsl(vs, vuc); in test6()
3767 res_vus = vec_vsl(vus, vuc); in test6()
3779 res_vbs = vec_vsl(vbs, vuc); in test6()
3791 res_vp = vec_vsl(vp, vuc); in test6()
3803 res_vi = vec_vsl(vi, vuc); in test6()
3815 res_vui = vec_vsl(vui, vuc); in test6()
3827 res_vbi = vec_vsl(vbi, vuc); in test6()
3844 res_vsc = vec_slo(vsc, vuc); in test6()
3848 res_vuc = vec_slo(vuc, vsc); in test6()
3852 res_vuc = vec_slo(vuc, vuc); in test6()
3860 res_vs = vec_slo(vs, vuc); in test6()
3868 res_vus = vec_slo(vus, vuc); in test6()
3876 res_vp = vec_slo(vp, vuc); in test6()
3884 res_vi = vec_slo(vi, vuc); in test6()
3892 res_vui = vec_slo(vui, vuc); in test6()
3900 res_vf = vec_slo(vf, vuc); in test6()
3908 res_vsc = vec_vslo(vsc, vuc); in test6()
3912 res_vuc = vec_vslo(vuc, vsc); in test6()
3916 res_vuc = vec_vslo(vuc, vuc); in test6()
3924 res_vs = vec_vslo(vs, vuc); in test6()
3932 res_vus = vec_vslo(vus, vuc); in test6()
3940 res_vp = vec_vslo(vp, vuc); in test6()
3948 res_vi = vec_vslo(vi, vuc); in test6()
3956 res_vui = vec_vslo(vui, vuc); in test6()
3964 res_vf = vec_vslo(vf, vuc); in test6()
3973 res_vuc = vec_splat(vuc, 0); in test6()
4017 res_vuc = vec_vspltb(vuc, 0); in test6()
4079 res_vsc = vec_sr(vsc, vuc); in test6()
4083 res_vuc = vec_sr(vuc, vuc); in test6()
4103 res_vsc = vec_vsrb(vsc, vuc); in test6()
4107 res_vuc = vec_vsrb(vuc, vuc); in test6()
4128 res_vsc = vec_sra(vsc, vuc); in test6()
4132 res_vuc = vec_sra(vuc, vuc); in test6()
4152 res_vsc = vec_vsrab(vsc, vuc); in test6()
4156 res_vuc = vec_vsrab(vuc, vuc); in test6()
4177 res_vsc = vec_srl(vsc, vuc); in test6()
4189 res_vuc = vec_srl(vuc, vuc); in test6()
4193 res_vuc = vec_srl(vuc, vus); in test6()
4197 res_vuc = vec_srl(vuc, vui); in test6()
4201 res_vbc = vec_srl(vbc, vuc); in test6()
4213 res_vs = vec_srl(vs, vuc); in test6()
4225 res_vus = vec_srl(vus, vuc); in test6()
4237 res_vbs = vec_srl(vbs, vuc); in test6()
4249 res_vp = vec_srl(vp, vuc); in test6()
4261 res_vi = vec_srl(vi, vuc); in test6()
4273 res_vui = vec_srl(vui, vuc); in test6()
4285 res_vbi = vec_srl(vbi, vuc); in test6()
4297 res_vsc = vec_vsr(vsc, vuc); in test6()
4309 res_vuc = vec_vsr(vuc, vuc); in test6()
4313 res_vuc = vec_vsr(vuc, vus); in test6()
4317 res_vuc = vec_vsr(vuc, vui); in test6()
4321 res_vbc = vec_vsr(vbc, vuc); in test6()
4333 res_vs = vec_vsr(vs, vuc); in test6()
4345 res_vus = vec_vsr(vus, vuc); in test6()
4357 res_vbs = vec_vsr(vbs, vuc); in test6()
4369 res_vp = vec_vsr(vp, vuc); in test6()
4381 res_vi = vec_vsr(vi, vuc); in test6()
4393 res_vui = vec_vsr(vui, vuc); in test6()
4405 res_vbi = vec_vsr(vbi, vuc); in test6()
4422 res_vsc = vec_sro(vsc, vuc); in test6()
4426 res_vuc = vec_sro(vuc, vsc); in test6()
4430 res_vuc = vec_sro(vuc, vuc); in test6()
4438 res_vs = vec_sro(vs, vuc); in test6()
4446 res_vus = vec_sro(vus, vuc); in test6()
4454 res_vp = vec_sro(vp, vuc); in test6()
4462 res_vi = vec_sro(vi, vuc); in test6()
4470 res_vui = vec_sro(vui, vuc); in test6()
4478 res_vf = vec_sro(vf, vuc); in test6()
4486 res_vsc = vec_vsro(vsc, vuc); in test6()
4490 res_vuc = vec_vsro(vuc, vsc); in test6()
4494 res_vuc = vec_vsro(vuc, vuc); in test6()
4502 res_vs = vec_vsro(vs, vuc); in test6()
4510 res_vus = vec_vsro(vus, vuc); in test6()
4518 res_vp = vec_vsro(vp, vuc); in test6()
4526 res_vi = vec_vsro(vi, vuc); in test6()
4534 res_vui = vec_vsro(vui, vuc); in test6()
4542 res_vf = vec_vsro(vf, vuc); in test6()
4555 vec_st(vuc, 0, &vuc); in test6()
4559 vec_st(vuc, 0, &param_uc); in test6()
4659 vec_stvx(vuc, 0, &vuc); in test6()
4663 vec_stvx(vuc, 0, &param_uc); in test6()
4760 vec_ste(vuc, 0, &param_uc); in test6()
4820 vec_stvebx(vuc, 0, &param_uc); in test6()
4885 vec_stl(vuc, 0, &vuc); in test6()
4889 vec_stl(vuc, 0, &param_uc); in test6()
4989 vec_stvxl(vuc, 0, &vuc); in test6()
4993 vec_stvxl(vuc, 0, &param_uc); in test6()
5098 res_vuc = vec_sub(vuc, vuc); in test6()
5102 res_vuc = vec_sub(vbc, vuc); in test6()
5106 res_vuc = vec_sub(vuc, vbc); in test6()
5174 res_vuc = vec_vsububm(vuc, vuc); in test6()
5178 res_vuc = vec_vsububm(vbc, vuc); in test6()
5182 res_vuc = vec_vsububm(vuc, vbc); in test6()
5260 res_vuc = vec_subs(vuc, vuc); in test6()
5264 res_vuc = vec_subs(vbc, vuc); in test6()
5268 res_vuc = vec_subs(vuc, vbc); in test6()
5332 res_vuc = vec_vsububs(vuc, vuc); in test6()
5336 res_vuc = vec_vsububs(vbc, vuc); in test6()
5340 res_vuc = vec_vsububs(vuc, vbc); in test6()
5397 res_vui = vec_sum4s(vuc, vui); in test6()
5409 res_vui = vec_vsum4ubs(vuc, vui); in test6()
5545 res_vuc = vec_xor(vuc, vuc); in test6()
5549 res_vuc = vec_xor(vbc, vuc); in test6()
5553 res_vuc = vec_xor(vuc, vbc); in test6()
5641 res_vuc = vec_vxor(vuc, vuc); in test6()
5645 res_vuc = vec_vxor(vbc, vuc); in test6()
5649 res_vuc = vec_vxor(vuc, vbc); in test6()
5732 res_uc = vec_extract(vuc, param_i); in test6()
5773 res_vuc = vec_insert(param_uc, vuc, param_i); in test6()
5840 res_vuc = vec_lvlx(0, &vuc); in test6()
6011 res_vuc = vec_lvlxl(0, &vuc); in test6()
6182 res_vuc = vec_lvrx(0, &vuc); in test6()
6353 res_vuc = vec_lvrxl(0, &vuc); in test6()
6526 vec_stvlx(vuc, 0, &param_uc); in test6()
6542 vec_stvlx(vuc, 0, &vuc); in test6()
6799 vec_stvlxl(vuc, 0, &param_uc); in test6()
6815 vec_stvlxl(vuc, 0, &vuc); in test6()
7072 vec_stvrx(vuc, 0, &param_uc); in test6()
7088 vec_stvrx(vuc, 0, &vuc); in test6()
7345 vec_stvrxl(vuc, 0, &param_uc); in test6()
7361 vec_stvrxl(vuc, 0, &vuc); in test6()
7668 res_i = vec_all_eq(vuc, vuc); in test6()
7672 res_i = vec_all_eq(vuc, vbc); in test6()
7680 res_i = vec_all_eq(vbc, vuc); in test6()
7761 res_i = vec_all_ge(vuc, vuc); in test6()
7765 res_i = vec_all_ge(vuc, vbc); in test6()
7773 res_i = vec_all_ge(vbc, vuc); in test6()
7850 res_i = vec_all_gt(vuc, vuc); in test6()
7854 res_i = vec_all_gt(vuc, vbc); in test6()
7862 res_i = vec_all_gt(vbc, vuc); in test6()
7944 res_i = vec_all_le(vuc, vuc); in test6()
7948 res_i = vec_all_le(vuc, vbc); in test6()
7956 res_i = vec_all_le(vbc, vuc); in test6()
8033 res_i = vec_all_lt(vuc, vuc); in test6()
8037 res_i = vec_all_lt(vuc, vbc); in test6()
8045 res_i = vec_all_lt(vbc, vuc); in test6()
8127 res_i = vec_all_ne(vuc, vuc); in test6()
8131 res_i = vec_all_ne(vuc, vbc); in test6()
8139 res_i = vec_all_ne(vbc, vuc); in test6()
8245 res_i = vec_any_eq(vuc, vuc); in test6()
8249 res_i = vec_any_eq(vuc, vbc); in test6()
8257 res_i = vec_any_eq(vbc, vuc); in test6()
8338 res_i = vec_any_ge(vuc, vuc); in test6()
8342 res_i = vec_any_ge(vuc, vbc); in test6()
8350 res_i = vec_any_ge(vbc, vuc); in test6()
8427 res_i = vec_any_gt(vuc, vuc); in test6()
8431 res_i = vec_any_gt(vuc, vbc); in test6()
8439 res_i = vec_any_gt(vbc, vuc); in test6()
8516 res_i = vec_any_le(vuc, vuc); in test6()
8520 res_i = vec_any_le(vuc, vbc); in test6()
8528 res_i = vec_any_le(vbc, vuc); in test6()
8605 res_i = vec_any_lt(vuc, vuc); in test6()
8609 res_i = vec_any_lt(vuc, vbc); in test6()
8617 res_i = vec_any_lt(vbc, vuc); in test6()
8699 res_i = vec_any_ne(vuc, vuc); in test6()
8703 res_i = vec_any_ne(vuc, vbc); in test6()
8711 res_i = vec_any_ne(vbc, vuc); in test6()