Lines Matching refs:vz0
56 vec_f vz0, vzo1, vzo2, vzo3; in pass_vsx_interleave() local
81 vz0 = vec_ld(0, &(out[0])); // z0.r z0.i z1.r z1.i in pass_vsx_interleave()
127 x11 = vec_add(vz0, x9); in pass_vsx_interleave()
128 x12 = vec_sub(vz0, x9); in pass_vsx_interleave()
170 vz0 = vec_ld(0, &(out[0])); // z0.r z0.i z1.r z1.i in pass_vsx_interleave()
219 x18 = vec_add(vz0, x16); in pass_vsx_interleave()
220 x19 = vec_sub(vz0, x16); in pass_vsx_interleave()
282 vec_f vz0, vz1, vz2, vz3; in fft8_vsx_interleave() local
296 vz0 = vec_ld(0, &(out[0])); in fft8_vsx_interleave()
301 x0 = vec_perm(vz0, vz1, vcprm(0,1,s2,s1)); in fft8_vsx_interleave()
302 x1 = vec_perm(vz0, vz1, vcprm(2,3,s0,s3)); in fft8_vsx_interleave()
355 vec_f vz0, vz1, vz2, vz3; in fft16_vsx_interleave() local
380 vz0 = vec_ld(0, &(out[0])); in fft16_vsx_interleave()
389 x0 = vec_perm(vz0, vz1, vcprm(0,1,s2,s1)); in fft16_vsx_interleave()
390 x1 = vec_perm(vz0, vz1, vcprm(2,3,s0,s3)); in fft16_vsx_interleave()
529 vec_f vz0, vz1, vz2, vz3; in fft8_vsx() local
537 vz0 = vec_ld(0, &(out[0])); in fft8_vsx()
544 vz4 = vec_perm(vz0, vz1, vcprm(0,1,s2,s1)); in fft8_vsx()
545 vz5 = vec_perm(vz0, vz1, vcprm(2,3,s0,s3)); in fft8_vsx()
551 vz0 = vec_add(vz4, vz5); in fft8_vsx()
557 vz4 = vec_perm(vz0, vz1, vcprm(0,s0,1,s1)); in fft8_vsx()
558 vz5 = vec_perm(vz0, vz1, vcprm(2,s3,3,s2)); in fft8_vsx()
562 vz0 = vec_add(vz4, vz5); in fft8_vsx()
567 vz4 = vec_perm(vz0, vz1, vcprm(0,1,s0,s1)); in fft8_vsx()
568 vz5 = vec_perm(vz0, vz1, vcprm(2,3,s2,s3)); in fft8_vsx()
576 vz0 = vec_add(vz4, vz6); in fft8_vsx()
579 vec_st(vz0, 0, &(out[0])); in fft8_vsx()
596 vec_f vz0, vz1, vz2, vz3; in fft16_vsx() local
601 vz0 = vec_ld(byte_8complex, &(out[0])); in fft16_vsx()
606 vz4 = vec_perm(vz0, vz1, vcprm(0,1,s2,s1)); in fft16_vsx()
607 vz5 = vec_perm(vz0, vz1, vcprm(2,3,s0,s3)); in fft16_vsx()
611 vz0 = vec_add(vz4, vz5); in fft16_vsx()
616 vz4 = vec_perm(vz0, vz1, vcprm(0,s0,1,s1)); in fft16_vsx()
617 vz5 = vec_perm(vz0, vz1, vcprm(2,s3,3,s2)); in fft16_vsx()
621 vz0 = vec_add(vz4, vz5); in fft16_vsx()
626 vz4 = vec_perm(vz0, vz1, vcprm(0,1,s0,s1)); in fft16_vsx()
627 vz5 = vec_perm(vz0, vz1, vcprm(2,3,s2,s3)); in fft16_vsx()
632 vz0 = vec_ld(0, &(out[0])); in fft16_vsx()
638 vz8 = vec_perm(vz0, vz1, vcprm(0,1,s2,s1)); in fft16_vsx()
639 vz9 = vec_perm(vz0, vz1, vcprm(2,3,s0,s3)); in fft16_vsx()
644 vz0 = vec_add(vz8, vz9); in fft16_vsx()
649 vz8 = vec_perm(vz0, vz1, vcprm(0,s0,1,s1)); in fft16_vsx()
650 vz9 = vec_perm(vz0, vz1, vcprm(2,s3,3,s2)); in fft16_vsx()
654 vz0 = vec_add(vz8, vz9); in fft16_vsx()
659 vz8 = vec_perm(vz0, vz1, vcprm(0,1,s0,s1)); in fft16_vsx()
660 vz9 = vec_perm(vz0, vz1, vcprm(2,3,s2,s3)); in fft16_vsx()
666 vz0 = vec_add(vz8, vz10); in fft16_vsx()
685 vz4 = vec_sub(vz0, vz10); in fft16_vsx()
686 vz0 = vec_add(vz0, vz10); in fft16_vsx()
697 vec_st(vz0, 0, &(out[0])); in fft16_vsx()