• Home
  • Raw
  • Download

Lines Matching refs:vuc

12 volatile vector unsigned char vuc;  variable
78 uc = vec_extract(vuc, idx); in test_core()
105 vuc = vec_insert(uc, vuc, idx); in test_core()
107 vuc = vec_insert(uc, vbc, idx); in test_core()
132 vuc = vec_promote(uc, idx); in test_core()
151 vuc = vec_insert_and_zero(cptruc); in test_core()
168 vsc = vec_perm(vsc, vsc, vuc); in test_core()
170 vuc = vec_perm(vuc, vuc, vuc); in test_core()
172 vbc = vec_perm(vbc, vbc, vuc); in test_core()
174 vss = vec_perm(vss, vss, vuc); in test_core()
176 vus = vec_perm(vus, vus, vuc); in test_core()
178 vbs = vec_perm(vbs, vbs, vuc); in test_core()
180 vsi = vec_perm(vsi, vsi, vuc); in test_core()
182 vui = vec_perm(vui, vui, vuc); in test_core()
184 vbi = vec_perm(vbi, vbi, vuc); in test_core()
186 vsl = vec_perm(vsl, vsl, vuc); in test_core()
188 vul = vec_perm(vul, vul, vuc); in test_core()
190 vbl = vec_perm(vbl, vbl, vuc); in test_core()
192 vd = vec_perm(vd, vd, vuc); in test_core()
228 vsc = vec_sel(vsc, vsc, vuc); in test_core()
230 vuc = vec_sel(vuc, vuc, vuc); in test_core()
231 vuc = vec_sel(vuc, vuc, vbc); in test_core()
232 vbc = vec_sel(vbc, vbc, vuc); in test_core()
298 vuc = vec_xld2(idx, cptruc); in test_core()
308 vuc = vec_xlw4(idx, cptruc); in test_core()
315 vec_xstd2(vuc, idx, ptruc); in test_core()
325 vec_xstw4(vuc, idx, ptruc); in test_core()
333 vuc = vec_load_bndry(cptruc, 64); in test_core()
364 vuc = vec_load_len(cptruc, idx); in test_core()
383 vec_store_len(vuc, ptruc, idx); in test_core()
403 vuc = vec_genmask(0); in test_core()
405 vuc = vec_genmask(0x8000); in test_core()
407 vuc = vec_genmask(0xffff); in test_core()
410 vuc = vec_genmasks_8(0, 7); in test_core()
412 vuc = vec_genmasks_8(1, 4); in test_core()
414 vuc = vec_genmasks_8(6, 2); in test_core()
439 vuc = vec_splat(vuc, 0); in test_core()
441 vuc = vec_splat(vuc, 15); in test_core()
492 vuc = vec_splat_u8(1); in test_core()
494 vuc = vec_splat_u8(254); in test_core()
523 vuc = vec_splats(uc); in test_core()
546 vuc = vec_mergeh(vuc, vuc); in test_core()
573 vuc = vec_mergel(vuc, vuc); in test_core()
600 vuc = vec_pack(vus, vus); in test_core()
619 vuc = vec_packs(vus, vus); in test_core()
632 vuc = vec_packs_cc(vus, vus, &cc); in test_core()
643 vuc = vec_packsu(vss, vss); in test_core()
645 vuc = vec_packsu(vus, vus); in test_core()
656 vuc = vec_packsu_cc(vus, vus, &cc); in test_core()
665 vus = vec_unpackh(vuc); in test_core()
684 vus = vec_unpackl(vuc); in test_core()
705 vbc = vec_cmpeq(vuc, vuc); in test_compare()
732 vbc = vec_cmpge(vuc, vuc); in test_compare()
751 vbc = vec_cmpgt(vuc, vuc); in test_compare()
770 vbc = vec_cmple(vuc, vuc); in test_compare()
789 vbc = vec_cmplt(vuc, vuc); in test_compare()
812 idx = vec_all_eq(vuc, vuc); in test_compare()
814 idx = vec_all_eq(vuc, vbc); in test_compare()
816 idx = vec_all_eq(vbc, vuc); in test_compare()
871 idx = vec_all_ne(vuc, vuc); in test_compare()
873 idx = vec_all_ne(vuc, vbc); in test_compare()
875 idx = vec_all_ne(vbc, vuc); in test_compare()
930 idx = vec_all_ge(vuc, vuc); in test_compare()
932 idx = vec_all_ge(vuc, vbc); in test_compare()
934 idx = vec_all_ge(vbc, vuc); in test_compare()
989 idx = vec_all_gt(vuc, vuc); in test_compare()
991 idx = vec_all_gt(vuc, vbc); in test_compare()
993 idx = vec_all_gt(vbc, vuc); in test_compare()
1048 idx = vec_all_le(vuc, vuc); in test_compare()
1050 idx = vec_all_le(vuc, vbc); in test_compare()
1052 idx = vec_all_le(vbc, vuc); in test_compare()
1107 idx = vec_all_lt(vuc, vuc); in test_compare()
1109 idx = vec_all_lt(vuc, vbc); in test_compare()
1111 idx = vec_all_lt(vbc, vuc); in test_compare()
1180 idx = vec_any_eq(vuc, vuc); in test_compare()
1182 idx = vec_any_eq(vuc, vbc); in test_compare()
1184 idx = vec_any_eq(vbc, vuc); in test_compare()
1239 idx = vec_any_ne(vuc, vuc); in test_compare()
1241 idx = vec_any_ne(vuc, vbc); in test_compare()
1243 idx = vec_any_ne(vbc, vuc); in test_compare()
1298 idx = vec_any_ge(vuc, vuc); in test_compare()
1300 idx = vec_any_ge(vuc, vbc); in test_compare()
1302 idx = vec_any_ge(vbc, vuc); in test_compare()
1357 idx = vec_any_gt(vuc, vuc); in test_compare()
1359 idx = vec_any_gt(vuc, vbc); in test_compare()
1361 idx = vec_any_gt(vbc, vuc); in test_compare()
1416 idx = vec_any_le(vuc, vuc); in test_compare()
1418 idx = vec_any_le(vuc, vbc); in test_compare()
1420 idx = vec_any_le(vbc, vuc); in test_compare()
1475 idx = vec_any_lt(vuc, vuc); in test_compare()
1477 idx = vec_any_lt(vuc, vbc); in test_compare()
1479 idx = vec_any_lt(vbc, vuc); in test_compare()
1547 vuc = vec_andc(vuc, vuc); in test_integer()
1548 vuc = vec_andc(vuc, vbc); in test_integer()
1549 vuc = vec_andc(vbc, vuc); in test_integer()
1579 vuc = vec_nor(vuc, vuc); in test_integer()
1580 vuc = vec_nor(vuc, vbc); in test_integer()
1581 vuc = vec_nor(vbc, vuc); in test_integer()
1608 vuc = vec_cntlz(vsc); in test_integer()
1610 vuc = vec_cntlz(vuc); in test_integer()
1625 vuc = vec_cnttz(vsc); in test_integer()
1627 vuc = vec_cnttz(vuc); in test_integer()
1642 vuc = vec_popcnt(vsc); in test_integer()
1644 vuc = vec_popcnt(vuc); in test_integer()
1659 vsc = vec_rl(vsc, vuc); in test_integer()
1661 vuc = vec_rl(vuc, vuc); in test_integer()
1678 vuc = vec_rli(vuc, ul); in test_integer()
1693 vsc = vec_rl_mask(vsc, vuc, 0); in test_integer()
1695 vsc = vec_rl_mask(vsc, vuc, 255); in test_integer()
1697 vuc = vec_rl_mask(vuc, vuc, 0); in test_integer()
1699 vuc = vec_rl_mask(vuc, vuc, 255); in test_integer()
1726 vsc = vec_sll(vsc, vuc); in test_integer()
1732 vuc = vec_sll(vuc, vuc); in test_integer()
1734 vuc = vec_sll(vuc, vus); in test_integer()
1736 vuc = vec_sll(vuc, vui); in test_integer()
1738 vbc = vec_sll(vbc, vuc); in test_integer()
1744 vss = vec_sll(vss, vuc); in test_integer()
1750 vus = vec_sll(vus, vuc); in test_integer()
1756 vbs = vec_sll(vbs, vuc); in test_integer()
1762 vsi = vec_sll(vsi, vuc); in test_integer()
1768 vui = vec_sll(vui, vuc); in test_integer()
1774 vbi = vec_sll(vbi, vuc); in test_integer()
1780 vsl = vec_sll(vsl, vuc); in test_integer()
1786 vul = vec_sll(vul, vuc); in test_integer()
1792 vbl = vec_sll(vbl, vuc); in test_integer()
1801 vsc = vec_slb(vsc, vuc); in test_integer()
1803 vuc = vec_slb(vuc, vsc); in test_integer()
1805 vuc = vec_slb(vuc, vuc); in test_integer()
1840 vuc = vec_sld(vuc, vuc, 0); in test_integer()
1842 vuc = vec_sld(vuc, vuc, 15); in test_integer()
1877 vuc = vec_sldw(vuc, vuc, 0); in test_integer()
1879 vuc = vec_sldw(vuc, vuc, 3); in test_integer()
1910 vsc = vec_sral(vsc, vuc); in test_integer()
1916 vuc = vec_sral(vuc, vuc); in test_integer()
1918 vuc = vec_sral(vuc, vus); in test_integer()
1920 vuc = vec_sral(vuc, vui); in test_integer()
1922 vbc = vec_sral(vbc, vuc); in test_integer()
1928 vss = vec_sral(vss, vuc); in test_integer()
1934 vus = vec_sral(vus, vuc); in test_integer()
1940 vbs = vec_sral(vbs, vuc); in test_integer()
1946 vsi = vec_sral(vsi, vuc); in test_integer()
1952 vui = vec_sral(vui, vuc); in test_integer()
1958 vbi = vec_sral(vbi, vuc); in test_integer()
1964 vsl = vec_sral(vsl, vuc); in test_integer()
1970 vul = vec_sral(vul, vuc); in test_integer()
1976 vbl = vec_sral(vbl, vuc); in test_integer()
1985 vsc = vec_srab(vsc, vuc); in test_integer()
1987 vuc = vec_srab(vuc, vsc); in test_integer()
1989 vuc = vec_srab(vuc, vuc); in test_integer()
2020 vsc = vec_srl(vsc, vuc); in test_integer()
2026 vuc = vec_srl(vuc, vuc); in test_integer()
2028 vuc = vec_srl(vuc, vus); in test_integer()
2030 vuc = vec_srl(vuc, vui); in test_integer()
2032 vbc = vec_srl(vbc, vuc); in test_integer()
2038 vss = vec_srl(vss, vuc); in test_integer()
2044 vus = vec_srl(vus, vuc); in test_integer()
2050 vbs = vec_srl(vbs, vuc); in test_integer()
2056 vsi = vec_srl(vsi, vuc); in test_integer()
2062 vui = vec_srl(vui, vuc); in test_integer()
2068 vbi = vec_srl(vbi, vuc); in test_integer()
2074 vsl = vec_srl(vsl, vuc); in test_integer()
2080 vul = vec_srl(vul, vuc); in test_integer()
2086 vbl = vec_srl(vbl, vuc); in test_integer()
2095 vsc = vec_srb(vsc, vuc); in test_integer()
2097 vuc = vec_srb(vuc, vsc); in test_integer()
2099 vuc = vec_srb(vuc, vuc); in test_integer()
2138 vuc = vec_max(vuc, vuc); in test_integer()
2139 vuc = vec_max(vuc, vbc); in test_integer()
2140 vuc = vec_max(vbc, vuc); in test_integer()
2164 vuc = vec_min(vuc, vuc); in test_integer()
2165 vuc = vec_min(vuc, vbc); in test_integer()
2166 vuc = vec_min(vbc, vuc); in test_integer()
2187 vuc = vec_addc(vuc, vuc); in test_integer()
2196 vuc = vec_add_u128(vuc, vuc); in test_integer()
2198 vuc = vec_addc_u128(vuc, vuc); in test_integer()
2200 vuc = vec_adde_u128(vuc, vuc, vuc); in test_integer()
2202 vuc = vec_addec_u128(vuc, vuc, vuc); in test_integer()
2207 vuc = vec_avg(vuc, vuc); in test_integer()
2225 vus = vec_gfmsum(vuc, vuc); in test_integer()
2231 vuc = vec_gfmsum_128(vul, vul); in test_integer()
2234 vus = vec_gfmsum_accum(vuc, vuc, vus); in test_integer()
2240 vuc = vec_gfmsum_accum_128(vul, vul, vuc); in test_integer()
2244 vsc = vec_mladd(vuc, vsc, vsc); in test_integer()
2245 vsc = vec_mladd(vsc, vuc, vuc); in test_integer()
2246 vuc = vec_mladd(vuc, vuc, vuc); in test_integer()
2258 vuc = vec_mhadd(vuc, vuc, vuc); in test_integer()
2271 vus = vec_meadd(vuc, vuc, vus); in test_integer()
2284 vus = vec_moadd(vuc, vuc, vus); in test_integer()
2297 vuc = vec_mulh(vuc, vuc); in test_integer()
2310 vus = vec_mule(vuc, vuc); in test_integer()
2323 vus = vec_mulo(vuc, vuc); in test_integer()
2334 vuc = vec_subc(vuc, vuc); in test_integer()
2343 vuc = vec_sub_u128(vuc, vuc); in test_integer()
2345 vuc = vec_subc_u128(vuc, vuc); in test_integer()
2347 vuc = vec_sube_u128(vuc, vuc, vuc); in test_integer()
2349 vuc = vec_subec_u128(vuc, vuc, vuc); in test_integer()
2352 vui = vec_sum4(vuc, vuc); in test_integer()
2360 vuc = vec_sum_u128(vui, vui); in test_integer()
2362 vuc = vec_sum_u128(vul, vul); in test_integer()
2365 idx = vec_test_mask(vsc, vuc); in test_integer()
2367 idx = vec_test_mask(vuc, vuc); in test_integer()
2388 vuc = vec_cp_until_zero(vuc); in test_string()
2407 vuc = vec_cp_until_zero_cc(vuc, &cc); in test_string()
2426 vuc = vec_cmpeq_idx(vuc, vuc); in test_string()
2428 vuc = vec_cmpeq_idx(vbc, vbc); in test_string()
2445 vuc = vec_cmpeq_idx_cc(vuc, vuc, &cc); in test_string()
2447 vuc = vec_cmpeq_idx_cc(vbc, vbc, &cc); in test_string()
2464 vuc = vec_cmpeq_or_0_idx(vuc, vuc); in test_string()
2466 vuc = vec_cmpeq_or_0_idx(vbc, vbc); in test_string()
2483 vuc = vec_cmpeq_or_0_idx_cc(vuc, vuc, &cc); in test_string()
2485 vuc = vec_cmpeq_or_0_idx_cc(vbc, vbc, &cc); in test_string()
2502 vuc = vec_cmpne_idx(vuc, vuc); in test_string()
2504 vuc = vec_cmpne_idx(vbc, vbc); in test_string()
2521 vuc = vec_cmpne_idx_cc(vuc, vuc, &cc); in test_string()
2523 vuc = vec_cmpne_idx_cc(vbc, vbc, &cc); in test_string()
2540 vuc = vec_cmpne_or_0_idx(vuc, vuc); in test_string()
2542 vuc = vec_cmpne_or_0_idx(vbc, vbc); in test_string()
2559 vuc = vec_cmpne_or_0_idx_cc(vuc, vuc, &cc); in test_string()
2561 vuc = vec_cmpne_or_0_idx_cc(vbc, vbc, &cc); in test_string()
2576 vbc = vec_cmprg(vuc, vuc, vuc); in test_string()
2583 vbc = vec_cmprg_cc(vuc, vuc, vuc, &cc); in test_string()
2590 vuc = vec_cmprg_idx(vuc, vuc, vuc); in test_string()
2597 vuc = vec_cmprg_idx_cc(vuc, vuc, vuc, &cc); in test_string()
2604 vuc = vec_cmprg_or_0_idx(vuc, vuc, vuc); in test_string()
2611 vuc = vec_cmprg_or_0_idx_cc(vuc, vuc, vuc, &cc); in test_string()
2618 vbc = vec_cmpnrg(vuc, vuc, vuc); in test_string()
2625 vbc = vec_cmpnrg_cc(vuc, vuc, vuc, &cc); in test_string()
2632 vuc = vec_cmpnrg_idx(vuc, vuc, vuc); in test_string()
2639 vuc = vec_cmpnrg_idx_cc(vuc, vuc, vuc, &cc); in test_string()
2646 vuc = vec_cmpnrg_or_0_idx(vuc, vuc, vuc); in test_string()
2653 vuc = vec_cmpnrg_or_0_idx_cc(vuc, vuc, vuc, &cc); in test_string()
2662 vbc = vec_find_any_eq(vuc, vuc); in test_string()
2681 vbc = vec_find_any_eq_cc(vuc, vuc, &cc); in test_string()
2700 vuc = vec_find_any_eq_idx(vuc, vuc); in test_string()
2702 vuc = vec_find_any_eq_idx(vbc, vbc); in test_string()
2719 vuc = vec_find_any_eq_idx_cc(vuc, vuc, &cc); in test_string()
2721 vuc = vec_find_any_eq_idx_cc(vbc, vbc, &cc); in test_string()
2738 vuc = vec_find_any_eq_or_0_idx(vuc, vuc); in test_string()
2740 vuc = vec_find_any_eq_or_0_idx(vbc, vbc); in test_string()
2757 vuc = vec_find_any_eq_or_0_idx_cc(vuc, vuc, &cc); in test_string()
2759 vuc = vec_find_any_eq_or_0_idx_cc(vbc, vbc, &cc); in test_string()
2776 vbc = vec_find_any_ne(vuc, vuc); in test_string()
2795 vbc = vec_find_any_ne_cc(vuc, vuc, &cc); in test_string()
2814 vuc = vec_find_any_ne_idx(vuc, vuc); in test_string()
2816 vuc = vec_find_any_ne_idx(vbc, vbc); in test_string()
2833 vuc = vec_find_any_ne_idx_cc(vuc, vuc, &cc); in test_string()
2835 vuc = vec_find_any_ne_idx_cc(vbc, vbc, &cc); in test_string()
2852 vuc = vec_find_any_ne_or_0_idx(vuc, vuc); in test_string()
2854 vuc = vec_find_any_ne_or_0_idx(vbc, vbc); in test_string()
2871 vuc = vec_find_any_ne_or_0_idx_cc(vuc, vuc, &cc); in test_string()
2873 vuc = vec_find_any_ne_or_0_idx_cc(vbc, vbc, &cc); in test_string()