• Home
  • Raw
  • Download

Lines Matching refs:xmm2

231 ; SSE2-NEXT:    movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
232 ; SSE2-NEXT: pxor %xmm2, %xmm0
233 ; SSE2-NEXT: pxor %xmm2, %xmm1
234 ; SSE2-NEXT: movdqa %xmm1, %xmm2
235 ; SSE2-NEXT: pcmpgtd %xmm0, %xmm2
236 ; SSE2-NEXT: pshufd {{.*#+}} xmm3 = xmm2[0,0,2,2]
240 ; SSE2-NEXT: pshufd {{.*#+}} xmm1 = xmm2[1,1,3,3]
248 ; SSE41-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
249 ; SSE41-NEXT: pxor %xmm2, %xmm0
250 ; SSE41-NEXT: pxor %xmm2, %xmm1
251 ; SSE41-NEXT: movdqa %xmm1, %xmm2
252 ; SSE41-NEXT: pcmpgtd %xmm0, %xmm2
253 ; SSE41-NEXT: pshufd {{.*#+}} xmm3 = xmm2[0,0,2,2]
257 ; SSE41-NEXT: pshufd {{.*#+}} xmm1 = xmm2[1,1,3,3]
265 ; SSE42-NEXT: movdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808]
266 ; SSE42-NEXT: pxor %xmm2, %xmm0
267 ; SSE42-NEXT: pxor %xmm1, %xmm2
268 ; SSE42-NEXT: pcmpgtq %xmm0, %xmm2
270 ; SSE42-NEXT: pxor %xmm2, %xmm0
275 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808]
276 ; AVX-NEXT: vpxor %xmm2, %xmm0, %xmm0
277 ; AVX-NEXT: vpxor %xmm2, %xmm1, %xmm1
295 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
296 ; SSE2-NEXT: pxor %xmm2, %xmm0
297 ; SSE2-NEXT: pxor %xmm1, %xmm2
298 ; SSE2-NEXT: pcmpgtd %xmm0, %xmm2
300 ; SSE2-NEXT: pxor %xmm2, %xmm0
394 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
395 ; SSE2-NEXT: pxor %xmm2, %xmm1
396 ; SSE2-NEXT: pxor %xmm2, %xmm0
397 ; SSE2-NEXT: movdqa %xmm0, %xmm2
398 ; SSE2-NEXT: pcmpgtd %xmm1, %xmm2
399 ; SSE2-NEXT: pshufd {{.*#+}} xmm3 = xmm2[0,0,2,2]
403 ; SSE2-NEXT: pshufd {{.*#+}} xmm0 = xmm2[1,1,3,3]
409 ; SSE41-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
410 ; SSE41-NEXT: pxor %xmm2, %xmm1
411 ; SSE41-NEXT: pxor %xmm2, %xmm0
412 ; SSE41-NEXT: movdqa %xmm0, %xmm2
413 ; SSE41-NEXT: pcmpgtd %xmm1, %xmm2
414 ; SSE41-NEXT: pshufd {{.*#+}} xmm3 = xmm2[0,0,2,2]
418 ; SSE41-NEXT: pshufd {{.*#+}} xmm0 = xmm2[1,1,3,3]
424 ; SSE42-NEXT: movdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808]
425 ; SSE42-NEXT: pxor %xmm2, %xmm1
426 ; SSE42-NEXT: pxor %xmm2, %xmm0
432 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808]
433 ; AVX-NEXT: vpxor %xmm2, %xmm1, %xmm1
434 ; AVX-NEXT: vpxor %xmm2, %xmm0, %xmm0
450 ; SSE-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
451 ; SSE-NEXT: pxor %xmm2, %xmm1
452 ; SSE-NEXT: pxor %xmm2, %xmm0
458 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
459 ; AVX1-NEXT: vpxor %xmm2, %xmm1, %xmm1
460 ; AVX1-NEXT: vpxor %xmm2, %xmm0, %xmm0
466 ; AVX2-NEXT: vpbroadcastd {{.*}}(%rip), %xmm2
467 ; AVX2-NEXT: vpxor %xmm2, %xmm1, %xmm1
468 ; AVX2-NEXT: vpxor %xmm2, %xmm0, %xmm0
479 ; AVX512-NEXT: vpbroadcastd {{.*}}(%rip), %xmm2
480 ; AVX512-NEXT: vpxor %xmm2, %xmm1, %xmm1
481 ; AVX512-NEXT: vpxor %xmm2, %xmm0, %xmm0
492 ; SSE-NEXT: movdqa {{.*#+}} xmm2 = [32768,32768,32768,32768,32768,32768,32768,32768]
493 ; SSE-NEXT: pxor %xmm2, %xmm1
494 ; SSE-NEXT: pxor %xmm2, %xmm0
500 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [32768,32768,32768,32768,32768,32768,32768,32768]
501 ; AVX-NEXT: vpxor %xmm2, %xmm1, %xmm1
502 ; AVX-NEXT: vpxor %xmm2, %xmm0, %xmm0
518 ; SSE-NEXT: movdqa {{.*#+}} xmm2 = [128,128,128,128,128,128,128,128,128,128,128,128,128,128,128,…
519 ; SSE-NEXT: pxor %xmm2, %xmm1
520 ; SSE-NEXT: pxor %xmm2, %xmm0
526 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [128,128,128,128,128,128,128,128,128,128,128,128,128,128,128…
527 ; AVX-NEXT: vpxor %xmm2, %xmm1, %xmm1
528 ; AVX-NEXT: vpxor %xmm2, %xmm0, %xmm0
548 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
549 ; SSE2-NEXT: pxor %xmm2, %xmm1
550 ; SSE2-NEXT: pxor %xmm2, %xmm0
551 ; SSE2-NEXT: movdqa %xmm0, %xmm2
552 ; SSE2-NEXT: pcmpgtd %xmm1, %xmm2
553 ; SSE2-NEXT: pshufd {{.*#+}} xmm3 = xmm2[0,0,2,2]
557 ; SSE2-NEXT: pshufd {{.*#+}} xmm1 = xmm2[1,1,3,3]
565 ; SSE41-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
566 ; SSE41-NEXT: pxor %xmm2, %xmm1
567 ; SSE41-NEXT: pxor %xmm2, %xmm0
568 ; SSE41-NEXT: movdqa %xmm0, %xmm2
569 ; SSE41-NEXT: pcmpgtd %xmm1, %xmm2
570 ; SSE41-NEXT: pshufd {{.*#+}} xmm3 = xmm2[0,0,2,2]
574 ; SSE41-NEXT: pshufd {{.*#+}} xmm1 = xmm2[1,1,3,3]
582 ; SSE42-NEXT: movdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808]
583 ; SSE42-NEXT: pxor %xmm2, %xmm1
584 ; SSE42-NEXT: pxor %xmm2, %xmm0
592 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808]
593 ; AVX-NEXT: vpxor %xmm2, %xmm1, %xmm1
594 ; AVX-NEXT: vpxor %xmm2, %xmm0, %xmm0
612 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
613 ; SSE2-NEXT: pxor %xmm2, %xmm1
614 ; SSE2-NEXT: pxor %xmm2, %xmm0
711 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
712 ; SSE2-NEXT: pxor %xmm2, %xmm0
713 ; SSE2-NEXT: pxor %xmm2, %xmm1
714 ; SSE2-NEXT: movdqa %xmm1, %xmm2
715 ; SSE2-NEXT: pcmpgtd %xmm0, %xmm2
716 ; SSE2-NEXT: pshufd {{.*#+}} xmm3 = xmm2[0,0,2,2]
720 ; SSE2-NEXT: pshufd {{.*#+}} xmm0 = xmm2[1,1,3,3]
726 ; SSE41-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
727 ; SSE41-NEXT: pxor %xmm2, %xmm0
728 ; SSE41-NEXT: pxor %xmm2, %xmm1
729 ; SSE41-NEXT: movdqa %xmm1, %xmm2
730 ; SSE41-NEXT: pcmpgtd %xmm0, %xmm2
731 ; SSE41-NEXT: pshufd {{.*#+}} xmm3 = xmm2[0,0,2,2]
735 ; SSE41-NEXT: pshufd {{.*#+}} xmm0 = xmm2[1,1,3,3]
741 ; SSE42-NEXT: movdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808]
742 ; SSE42-NEXT: pxor %xmm2, %xmm0
743 ; SSE42-NEXT: pxor %xmm1, %xmm2
744 ; SSE42-NEXT: pcmpgtq %xmm0, %xmm2
745 ; SSE42-NEXT: movdqa %xmm2, %xmm0
750 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808]
751 ; AVX-NEXT: vpxor %xmm2, %xmm0, %xmm0
752 ; AVX-NEXT: vpxor %xmm2, %xmm1, %xmm1
768 ; SSE-NEXT: movdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
769 ; SSE-NEXT: pxor %xmm2, %xmm0
770 ; SSE-NEXT: pxor %xmm1, %xmm2
771 ; SSE-NEXT: pcmpgtd %xmm0, %xmm2
772 ; SSE-NEXT: movdqa %xmm2, %xmm0
777 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [2147483648,2147483648,2147483648,2147483648]
778 ; AVX1-NEXT: vpxor %xmm2, %xmm0, %xmm0
779 ; AVX1-NEXT: vpxor %xmm2, %xmm1, %xmm1
785 ; AVX2-NEXT: vpbroadcastd {{.*}}(%rip), %xmm2
786 ; AVX2-NEXT: vpxor %xmm2, %xmm0, %xmm0
787 ; AVX2-NEXT: vpxor %xmm2, %xmm1, %xmm1
798 ; AVX512-NEXT: vpbroadcastd {{.*}}(%rip), %xmm2
799 ; AVX512-NEXT: vpxor %xmm2, %xmm0, %xmm0
800 ; AVX512-NEXT: vpxor %xmm2, %xmm1, %xmm1
811 ; SSE-NEXT: movdqa {{.*#+}} xmm2 = [32768,32768,32768,32768,32768,32768,32768,32768]
812 ; SSE-NEXT: pxor %xmm2, %xmm0
813 ; SSE-NEXT: pxor %xmm1, %xmm2
814 ; SSE-NEXT: pcmpgtw %xmm0, %xmm2
815 ; SSE-NEXT: movdqa %xmm2, %xmm0
820 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [32768,32768,32768,32768,32768,32768,32768,32768]
821 ; AVX-NEXT: vpxor %xmm2, %xmm0, %xmm0
822 ; AVX-NEXT: vpxor %xmm2, %xmm1, %xmm1
838 ; SSE-NEXT: movdqa {{.*#+}} xmm2 = [128,128,128,128,128,128,128,128,128,128,128,128,128,128,128,…
839 ; SSE-NEXT: pxor %xmm2, %xmm0
840 ; SSE-NEXT: pxor %xmm1, %xmm2
841 ; SSE-NEXT: pcmpgtb %xmm0, %xmm2
842 ; SSE-NEXT: movdqa %xmm2, %xmm0
847 ; AVX-NEXT: vmovdqa {{.*#+}} xmm2 = [128,128,128,128,128,128,128,128,128,128,128,128,128,128,128…
848 ; AVX-NEXT: vpxor %xmm2, %xmm0, %xmm0
849 ; AVX-NEXT: vpxor %xmm2, %xmm1, %xmm1