• Home
  • Raw
  • Download

Lines Matching refs:xmm1

65 ; SSE-NEXT:    pshufd {{.*#+}} xmm0 = xmm1[0,1,0,1]
70 ; AVX1-NEXT: vpshufd {{.*#+}} xmm0 = xmm1[0,1,0,1]
75 ; AVX2-NEXT: vpbroadcastq %xmm1, %xmm0
80 ; AVX512VL-NEXT: vpbroadcastq %xmm1, %xmm0
88 ; SSE-NEXT: pshufd {{.*#+}} xmm0 = xmm1[2,3,0,1]
93 ; AVX-NEXT: vpshufd {{.*#+}} xmm0 = xmm1[2,3,0,1]
101 ; SSE-NEXT: pshufd {{.*#+}} xmm0 = xmm1[2,3,2,3]
106 ; AVX-NEXT: vpshufd {{.*#+}} xmm0 = xmm1[2,3,2,3]
170 ; SSE2-NEXT: movlhps {{.*#+}} xmm1 = xmm1[0,0]
171 ; SSE2-NEXT: movaps %xmm1, %xmm0
176 ; SSE3-NEXT: movddup {{.*#+}} xmm0 = xmm1[0,0]
181 ; SSSE3-NEXT: movddup {{.*#+}} xmm0 = xmm1[0,0]
186 ; SSE41-NEXT: movddup {{.*#+}} xmm0 = xmm1[0,0]
191 ; AVX-NEXT: vmovddup {{.*#+}} xmm0 = xmm1[0,0]
199 ; SSE-NEXT: shufpd {{.*#+}} xmm1 = xmm1[1,0]
200 ; SSE-NEXT: movapd %xmm1, %xmm0
205 ; AVX-NEXT: vpermilpd {{.*#+}} xmm0 = xmm1[1,0]
214 ; SSE-NEXT: movhlps {{.*#+}} xmm1 = xmm1[1,1]
215 ; SSE-NEXT: movaps %xmm1, %xmm0
220 ; AVX-NEXT: vmovhlps {{.*#+}} xmm0 = xmm1[1,1]
228 ; SSE2-NEXT: movsd {{.*#+}} xmm1 = xmm0[0],xmm1[1]
229 ; SSE2-NEXT: movapd %xmm1, %xmm0
234 ; SSE3-NEXT: movsd {{.*#+}} xmm1 = xmm0[0],xmm1[1]
235 ; SSE3-NEXT: movapd %xmm1, %xmm0
240 ; SSSE3-NEXT: movsd {{.*#+}} xmm1 = xmm0[0],xmm1[1]
241 ; SSSE3-NEXT: movapd %xmm1, %xmm0
246 ; SSE41-NEXT: blendpd {{.*#+}} xmm0 = xmm0[0],xmm1[1]
251 ; AVX-NEXT: vblendpd {{.*#+}} xmm0 = xmm0[0],xmm1[1]
259 ; SSE2-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
264 ; SSE3-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
269 ; SSSE3-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
274 ; SSE41-NEXT: blendpd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
279 ; AVX-NEXT: vblendpd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
289 ; SSE-NEXT: punpcklqdq {{.*#+}} xmm0 = xmm0[0],xmm1[0]
294 ; AVX-NEXT: vpunpcklqdq {{.*#+}} xmm0 = xmm0[0],xmm1[0]
302 ; SSE-NEXT: punpcklqdq {{.*#+}} xmm1 = xmm1[0],xmm2[0]
303 ; SSE-NEXT: movdqa %xmm1, %xmm0
308 ; AVX-NEXT: vpunpcklqdq {{.*#+}} xmm0 = xmm1[0],xmm2[0]
316 ; SSE2-NEXT: movsd {{.*#+}} xmm1 = xmm0[0],xmm1[1]
317 ; SSE2-NEXT: movapd %xmm1, %xmm0
322 ; SSE3-NEXT: movsd {{.*#+}} xmm1 = xmm0[0],xmm1[1]
323 ; SSE3-NEXT: movapd %xmm1, %xmm0
328 ; SSSE3-NEXT: movsd {{.*#+}} xmm1 = xmm0[0],xmm1[1]
329 ; SSSE3-NEXT: movapd %xmm1, %xmm0
334 ; SSE41-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1,2,3],xmm1[4,5,6,7]
339 ; AVX1-NEXT: vpblendw {{.*#+}} xmm0 = xmm0[0,1,2,3],xmm1[4,5,6,7]
344 ; AVX2-NEXT: vpblendd {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,3]
349 ; AVX512VL-NEXT: vpblendd {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,3]
357 ; SSE2-NEXT: movsd {{.*#+}} xmm2 = xmm1[0],xmm2[1]
363 ; SSE3-NEXT: movsd {{.*#+}} xmm2 = xmm1[0],xmm2[1]
369 ; SSSE3-NEXT: movsd {{.*#+}} xmm2 = xmm1[0],xmm2[1]
375 ; SSE41-NEXT: pblendw {{.*#+}} xmm1 = xmm1[0,1,2,3],xmm2[4,5,6,7]
376 ; SSE41-NEXT: movdqa %xmm1, %xmm0
381 ; AVX1-NEXT: vpblendw {{.*#+}} xmm0 = xmm1[0,1,2,3],xmm2[4,5,6,7]
386 ; AVX2-NEXT: vpblendd {{.*#+}} xmm0 = xmm1[0,1],xmm2[2,3]
391 ; AVX512VL-NEXT: vpblendd {{.*#+}} xmm0 = xmm1[0,1],xmm2[2,3]
399 ; SSE2-NEXT: shufpd {{.*#+}} xmm0 = xmm0[1],xmm1[0]
404 ; SSE3-NEXT: shufpd {{.*#+}} xmm0 = xmm0[1],xmm1[0]
409 ; SSSE3-NEXT: palignr {{.*#+}} xmm1 = xmm0[8,9,10,11,12,13,14,15],xmm1[0,1,2,3,4,5,6,7]
410 ; SSSE3-NEXT: movdqa %xmm1, %xmm0
415 ; SSE41-NEXT: palignr {{.*#+}} xmm1 = xmm0[8,9,10,11,12,13,14,15],xmm1[0,1,2,3,4,5,6,7]
416 ; SSE41-NEXT: movdqa %xmm1, %xmm0
421 ; AVX-NEXT: vpalignr {{.*#+}} xmm0 = xmm0[8,9,10,11,12,13,14,15],xmm1[0,1,2,3,4,5,6,7]
429 ; SSE2-NEXT: shufpd {{.*#+}} xmm1 = xmm1[1],xmm2[0]
430 ; SSE2-NEXT: movapd %xmm1, %xmm0
435 ; SSE3-NEXT: shufpd {{.*#+}} xmm1 = xmm1[1],xmm2[0]
436 ; SSE3-NEXT: movapd %xmm1, %xmm0
441 ; SSSE3-NEXT: palignr {{.*#+}} xmm2 = xmm1[8,9,10,11,12,13,14,15],xmm2[0,1,2,3,4,5,6,7]
447 ; SSE41-NEXT: palignr {{.*#+}} xmm2 = xmm1[8,9,10,11,12,13,14,15],xmm2[0,1,2,3,4,5,6,7]
453 ; AVX-NEXT: vpalignr {{.*#+}} xmm0 = xmm1[8,9,10,11,12,13,14,15],xmm2[0,1,2,3,4,5,6,7]
461 ; SSE-NEXT: punpckhqdq {{.*#+}} xmm0 = xmm0[1],xmm1[1]
466 ; AVX-NEXT: vpunpckhqdq {{.*#+}} xmm0 = xmm0[1],xmm1[1]
474 ; SSE-NEXT: punpckhqdq {{.*#+}} xmm1 = xmm1[1],xmm2[1]
475 ; SSE-NEXT: movdqa %xmm1, %xmm0
480 ; AVX-NEXT: vpunpckhqdq {{.*#+}} xmm0 = xmm1[1],xmm2[1]
488 ; SSE-NEXT: punpcklqdq {{.*#+}} xmm1 = xmm1[0],xmm0[0]
489 ; SSE-NEXT: movdqa %xmm1, %xmm0
494 ; AVX-NEXT: vpunpcklqdq {{.*#+}} xmm0 = xmm1[0],xmm0[0]
502 ; SSE-NEXT: punpcklqdq {{.*#+}} xmm2 = xmm2[0],xmm1[0]
508 ; AVX-NEXT: vpunpcklqdq {{.*#+}} xmm0 = xmm2[0],xmm1[0]
516 ; SSE2-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
521 ; SSE3-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
526 ; SSSE3-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
531 ; SSE41-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1,2,3],xmm0[4,5,6,7]
536 ; AVX1-NEXT: vpblendw {{.*#+}} xmm0 = xmm1[0,1,2,3],xmm0[4,5,6,7]
541 ; AVX2-NEXT: vpblendd {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3]
546 ; AVX512VL-NEXT: vpblendd {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3]
554 ; SSE2-NEXT: movsd {{.*#+}} xmm1 = xmm2[0],xmm1[1]
555 ; SSE2-NEXT: movapd %xmm1, %xmm0
560 ; SSE3-NEXT: movsd {{.*#+}} xmm1 = xmm2[0],xmm1[1]
561 ; SSE3-NEXT: movapd %xmm1, %xmm0
566 ; SSSE3-NEXT: movsd {{.*#+}} xmm1 = xmm2[0],xmm1[1]
567 ; SSSE3-NEXT: movapd %xmm1, %xmm0
572 ; SSE41-NEXT: pblendw {{.*#+}} xmm1 = xmm2[0,1,2,3],xmm1[4,5,6,7]
573 ; SSE41-NEXT: movdqa %xmm1, %xmm0
578 ; AVX1-NEXT: vpblendw {{.*#+}} xmm0 = xmm2[0,1,2,3],xmm1[4,5,6,7]
583 ; AVX2-NEXT: vpblendd {{.*#+}} xmm0 = xmm2[0,1],xmm1[2,3]
588 ; AVX512VL-NEXT: vpblendd {{.*#+}} xmm0 = xmm2[0,1],xmm1[2,3]
596 ; SSE2-NEXT: shufpd {{.*#+}} xmm1 = xmm1[1],xmm0[0]
597 ; SSE2-NEXT: movapd %xmm1, %xmm0
602 ; SSE3-NEXT: shufpd {{.*#+}} xmm1 = xmm1[1],xmm0[0]
603 ; SSE3-NEXT: movapd %xmm1, %xmm0
608 ; SSSE3-NEXT: palignr {{.*#+}} xmm0 = xmm1[8,9,10,11,12,13,14,15],xmm0[0,1,2,3,4,5,6,7]
613 ; SSE41-NEXT: palignr {{.*#+}} xmm0 = xmm1[8,9,10,11,12,13,14,15],xmm0[0,1,2,3,4,5,6,7]
618 ; AVX-NEXT: vpalignr {{.*#+}} xmm0 = xmm1[8,9,10,11,12,13,14,15],xmm0[0,1,2,3,4,5,6,7]
626 ; SSE2-NEXT: shufpd {{.*#+}} xmm2 = xmm2[1],xmm1[0]
632 ; SSE3-NEXT: shufpd {{.*#+}} xmm2 = xmm2[1],xmm1[0]
638 ; SSSE3-NEXT: palignr {{.*#+}} xmm1 = xmm2[8,9,10,11,12,13,14,15],xmm1[0,1,2,3,4,5,6,7]
639 ; SSSE3-NEXT: movdqa %xmm1, %xmm0
644 ; SSE41-NEXT: palignr {{.*#+}} xmm1 = xmm2[8,9,10,11,12,13,14,15],xmm1[0,1,2,3,4,5,6,7]
645 ; SSE41-NEXT: movdqa %xmm1, %xmm0
650 ; AVX-NEXT: vpalignr {{.*#+}} xmm0 = xmm2[8,9,10,11,12,13,14,15],xmm1[0,1,2,3,4,5,6,7]
658 ; SSE-NEXT: punpckhqdq {{.*#+}} xmm1 = xmm1[1],xmm0[1]
659 ; SSE-NEXT: movdqa %xmm1, %xmm0
664 ; AVX-NEXT: vpunpckhqdq {{.*#+}} xmm0 = xmm1[1],xmm0[1]
672 ; SSE-NEXT: punpckhqdq {{.*#+}} xmm2 = xmm2[1],xmm1[1]
678 ; AVX-NEXT: vpunpckhqdq {{.*#+}} xmm0 = xmm2[1],xmm1[1]
729 ; SSE2-NEXT: xorpd %xmm1, %xmm1
730 ; SSE2-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
735 ; SSE3-NEXT: xorpd %xmm1, %xmm1
736 ; SSE3-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
741 ; SSSE3-NEXT: xorpd %xmm1, %xmm1
742 ; SSSE3-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
747 ; SSE41-NEXT: pxor %xmm1, %xmm1
748 ; SSE41-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1,2,3],xmm0[4,5,6,7]
753 ; AVX1-NEXT: vpxor %xmm1, %xmm1, %xmm1
754 ; AVX1-NEXT: vpblendw {{.*#+}} xmm0 = xmm1[0,1,2,3],xmm0[4,5,6,7]
759 ; AVX2-NEXT: vpxor %xmm1, %xmm1, %xmm1
760 ; AVX2-NEXT: vpblendd {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3]
765 ; AVX512VL-NEXT: vpxor %xmm1, %xmm1, %xmm1
766 ; AVX512VL-NEXT: vpblendd {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3]
789 ; SSE-NEXT: xorpd %xmm1, %xmm1
790 ; SSE-NEXT: unpckhpd {{.*#+}} xmm0 = xmm0[1],xmm1[1]
795 ; AVX1-NEXT: vxorpd %xmm1, %xmm1, %xmm1
796 ; AVX1-NEXT: vunpckhpd {{.*#+}} xmm0 = xmm0[1],xmm1[1]
801 ; AVX2-NEXT: vxorpd %xmm1, %xmm1, %xmm1
802 ; AVX2-NEXT: vunpckhpd {{.*#+}} xmm0 = xmm0[1],xmm1[1]
807 ; AVX512VL-NEXT: vxorps %xmm1, %xmm1, %xmm1
808 ; AVX512VL-NEXT: vunpckhpd {{.*#+}} xmm0 = xmm0[1],xmm1[1]
817 ; SSE-NEXT: xorpd %xmm1, %xmm1
818 ; SSE-NEXT: unpcklpd {{.*#+}} xmm1 = xmm1[0],xmm0[0]
819 ; SSE-NEXT: movapd %xmm1, %xmm0
824 ; AVX1-NEXT: vxorpd %xmm1, %xmm1, %xmm1
825 ; AVX1-NEXT: vunpcklpd {{.*#+}} xmm0 = xmm1[0],xmm0[0]
830 ; AVX2-NEXT: vxorpd %xmm1, %xmm1, %xmm1
831 ; AVX2-NEXT: vunpcklpd {{.*#+}} xmm0 = xmm1[0],xmm0[0]
836 ; AVX512VL-NEXT: vxorps %xmm1, %xmm1, %xmm1
837 ; AVX512VL-NEXT: vunpcklpd {{.*#+}} xmm0 = xmm1[0],xmm0[0]
846 ; SSE2-NEXT: xorpd %xmm1, %xmm1
847 ; SSE2-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
852 ; SSE3-NEXT: xorpd %xmm1, %xmm1
853 ; SSE3-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
858 ; SSSE3-NEXT: xorpd %xmm1, %xmm1
859 ; SSSE3-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
864 ; SSE41-NEXT: xorpd %xmm1, %xmm1
865 ; SSE41-NEXT: blendpd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
870 ; AVX-NEXT: vxorpd %xmm1, %xmm1, %xmm1
871 ; AVX-NEXT: vblendpd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
880 ; SSE-NEXT: xorpd %xmm1, %xmm1
881 ; SSE-NEXT: shufpd {{.*#+}} xmm0 = xmm0[1],xmm1[0]
886 ; AVX1-NEXT: vxorpd %xmm1, %xmm1, %xmm1
887 ; AVX1-NEXT: vshufpd {{.*#+}} xmm0 = xmm0[1],xmm1[0]
892 ; AVX2-NEXT: vxorpd %xmm1, %xmm1, %xmm1
893 ; AVX2-NEXT: vshufpd {{.*#+}} xmm0 = xmm0[1],xmm1[0]
898 ; AVX512VL-NEXT: vpxor %xmm1, %xmm1, %xmm1
899 ; AVX512VL-NEXT: vshufpd {{.*#+}} xmm0 = xmm0[1],xmm1[0]
911 ; SSE2-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
912 ; SSE2-NEXT: movss {{.*#+}} xmm0 = xmm1[0],xmm0[1,2,3]
918 ; SSE3-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
919 ; SSE3-NEXT: movss {{.*#+}} xmm0 = xmm1[0],xmm0[1,2,3]
925 ; SSSE3-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
926 ; SSSE3-NEXT: movss {{.*#+}} xmm0 = xmm1[0],xmm0[1,2,3]
932 ; SSE41-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
933 ; SSE41-NEXT: blendps {{.*#+}} xmm0 = xmm1[0],xmm0[1,2,3]
934 ; SSE41-NEXT: xorps %xmm1, %xmm1
935 ; SSE41-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3,4,5,6,7]
940 ; AVX1-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero
941 ; AVX1-NEXT: vblendps {{.*#+}} xmm0 = xmm1[0],xmm0[1,2,3]
942 ; AVX1-NEXT: vxorps %xmm1, %xmm1, %xmm1
943 ; AVX1-NEXT: vpblendw {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3,4,5,6,7]
948 ; AVX2-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero
949 ; AVX2-NEXT: vblendps {{.*#+}} xmm0 = xmm1[0],xmm0[1,2,3]
950 ; AVX2-NEXT: vxorps %xmm1, %xmm1, %xmm1
951 ; AVX2-NEXT: vpblendd {{.*#+}} xmm0 = xmm1[0],xmm0[1,2,3]
956 ; AVX512VL-NEXT: vmovss {{.*}}(%rip), %xmm1
957 ; AVX512VL-NEXT: vblendps {{.*#+}} xmm0 = xmm1[0],xmm0[1,2,3]
958 ; AVX512VL-NEXT: vxorps %xmm1, %xmm1, %xmm1
959 ; AVX512VL-NEXT: vpblendd {{.*#+}} xmm0 = xmm1[0],xmm0[1,2,3]
1053 ; SSE2-NEXT: movd %rdi, %xmm1
1054 ; SSE2-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
1059 ; SSE3-NEXT: movd %rdi, %xmm1
1060 ; SSE3-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
1065 ; SSSE3-NEXT: movd %rdi, %xmm1
1066 ; SSSE3-NEXT: movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
1071 ; SSE41-NEXT: movd %rdi, %xmm1
1072 ; SSE41-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1,2,3],xmm0[4,5,6,7]
1077 ; AVX1-NEXT: vmovq %rdi, %xmm1
1078 ; AVX1-NEXT: vpblendw {{.*#+}} xmm0 = xmm1[0,1,2,3],xmm0[4,5,6,7]
1083 ; AVX2-NEXT: vmovq %rdi, %xmm1
1084 ; AVX2-NEXT: vpblendd {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3]
1089 ; AVX512VL-NEXT: vmovq %rdi, %xmm1
1090 ; AVX512VL-NEXT: vpblendd {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3]
1115 ; SSE41-NEXT: movq {{.*#+}} xmm1 = mem[0],zero
1116 ; SSE41-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1,2,3],xmm0[4,5,6,7]
1121 ; AVX1-NEXT: vmovq {{.*#+}} xmm1 = mem[0],zero
1122 ; AVX1-NEXT: vpblendw {{.*#+}} xmm0 = xmm1[0,1,2,3],xmm0[4,5,6,7]
1127 ; AVX2-NEXT: vmovq {{.*#+}} xmm1 = mem[0],zero
1128 ; AVX2-NEXT: vpblendd {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3]
1133 ; AVX512VL-NEXT: vmovq (%rdi), %xmm1
1134 ; AVX512VL-NEXT: vpblendd {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3]
1145 ; SSE-NEXT: movd %rdi, %xmm1
1146 ; SSE-NEXT: punpcklqdq {{.*#+}} xmm0 = xmm0[0],xmm1[0]
1151 ; AVX-NEXT: vmovq %rdi, %xmm1
1152 ; AVX-NEXT: vpunpcklqdq {{.*#+}} xmm0 = xmm0[0],xmm1[0]
1162 ; SSE-NEXT: movq {{.*#+}} xmm1 = mem[0],zero
1163 ; SSE-NEXT: punpcklqdq {{.*#+}} xmm0 = xmm0[0],xmm1[0]
1168 ; AVX1-NEXT: vmovq {{.*#+}} xmm1 = mem[0],zero
1169 ; AVX1-NEXT: vpunpcklqdq {{.*#+}} xmm0 = xmm0[0],xmm1[0]
1174 ; AVX2-NEXT: vmovq {{.*#+}} xmm1 = mem[0],zero
1175 ; AVX2-NEXT: vpunpcklqdq {{.*#+}} xmm0 = xmm0[0],xmm1[0]
1180 ; AVX512VL-NEXT: vmovq (%rdi), %xmm1
1181 ; AVX512VL-NEXT: vpunpcklqdq {{.*#+}} xmm0 = xmm0[0],xmm1[0]
1192 ; SSE-NEXT: movsd {{.*#+}} xmm1 = xmm0[0],xmm1[1]
1193 ; SSE-NEXT: movapd %xmm1, %xmm0
1198 ; AVX1-NEXT: vmovsd {{.*#+}} xmm0 = xmm0[0],xmm1[1]
1203 ; AVX2-NEXT: vmovsd {{.*#+}} xmm0 = xmm0[0],xmm1[1]
1208 ; AVX512VL-NEXT: vmovsd %xmm0, %xmm1, %xmm0
1234 ; SSE-NEXT: unpcklpd {{.*#+}} xmm1 = xmm1[0],xmm0[0]
1235 ; SSE-NEXT: movapd %xmm1, %xmm0
1240 ; AVX-NEXT: vunpcklpd {{.*#+}} xmm0 = xmm1[0],xmm0[0]