; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc < %s -mtriple=i686-apple-darwin -mattr=+avx2 | FileCheck %s --check-prefix=X32 --check-prefix=X32-AVX2 ; RUN: llc < %s -mtriple=x86_64-apple-darwin -mattr=+avx2 | FileCheck %s --check-prefix=X64 --check-prefix=X64-AVX2 ; RUN: llc < %s -mtriple=i686-apple-darwin -mattr=+avx512vl,+avx512dq | FileCheck %s --check-prefix=X32 --check-prefix=X32-AVX512VL ; RUN: llc < %s -mtriple=x86_64-apple-darwin -mattr=+avx512vl,+avx512dq | FileCheck %s --check-prefix=X64 --check-prefix=X64-AVX512VL define <16 x i8> @BB16(i8* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: BB16: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vpbroadcastb (%eax), %xmm0 ; X32-NEXT: retl ; ; X64-LABEL: BB16: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vpbroadcastb (%rdi), %xmm0 ; X64-NEXT: retq entry: %q = load i8, i8* %ptr, align 4 %q0 = insertelement <16 x i8> undef, i8 %q, i32 0 %q1 = insertelement <16 x i8> %q0, i8 %q, i32 1 %q2 = insertelement <16 x i8> %q1, i8 %q, i32 2 %q3 = insertelement <16 x i8> %q2, i8 %q, i32 3 %q4 = insertelement <16 x i8> %q3, i8 %q, i32 4 %q5 = insertelement <16 x i8> %q4, i8 %q, i32 5 %q6 = insertelement <16 x i8> %q5, i8 %q, i32 6 %q7 = insertelement <16 x i8> %q6, i8 %q, i32 7 %q8 = insertelement <16 x i8> %q7, i8 %q, i32 8 %q9 = insertelement <16 x i8> %q8, i8 %q, i32 9 %qa = insertelement <16 x i8> %q9, i8 %q, i32 10 %qb = insertelement <16 x i8> %qa, i8 %q, i32 11 %qc = insertelement <16 x i8> %qb, i8 %q, i32 12 %qd = insertelement <16 x i8> %qc, i8 %q, i32 13 %qe = insertelement <16 x i8> %qd, i8 %q, i32 14 %qf = insertelement <16 x i8> %qe, i8 %q, i32 15 ret <16 x i8> %qf } define <32 x i8> @BB32(i8* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: BB32: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vpbroadcastb (%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: BB32: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vpbroadcastb (%rdi), %ymm0 ; X64-NEXT: retq entry: %q = load i8, i8* %ptr, align 4 %q0 = insertelement <32 x i8> undef, i8 %q, i32 0 %q1 = insertelement <32 x i8> %q0, i8 %q, i32 1 %q2 = insertelement <32 x i8> %q1, i8 %q, i32 2 %q3 = insertelement <32 x i8> %q2, i8 %q, i32 3 %q4 = insertelement <32 x i8> %q3, i8 %q, i32 4 %q5 = insertelement <32 x i8> %q4, i8 %q, i32 5 %q6 = insertelement <32 x i8> %q5, i8 %q, i32 6 %q7 = insertelement <32 x i8> %q6, i8 %q, i32 7 %q8 = insertelement <32 x i8> %q7, i8 %q, i32 8 %q9 = insertelement <32 x i8> %q8, i8 %q, i32 9 %qa = insertelement <32 x i8> %q9, i8 %q, i32 10 %qb = insertelement <32 x i8> %qa, i8 %q, i32 11 %qc = insertelement <32 x i8> %qb, i8 %q, i32 12 %qd = insertelement <32 x i8> %qc, i8 %q, i32 13 %qe = insertelement <32 x i8> %qd, i8 %q, i32 14 %qf = insertelement <32 x i8> %qe, i8 %q, i32 15 %q20 = insertelement <32 x i8> %qf, i8 %q, i32 16 %q21 = insertelement <32 x i8> %q20, i8 %q, i32 17 %q22 = insertelement <32 x i8> %q21, i8 %q, i32 18 %q23 = insertelement <32 x i8> %q22, i8 %q, i32 19 %q24 = insertelement <32 x i8> %q23, i8 %q, i32 20 %q25 = insertelement <32 x i8> %q24, i8 %q, i32 21 %q26 = insertelement <32 x i8> %q25, i8 %q, i32 22 %q27 = insertelement <32 x i8> %q26, i8 %q, i32 23 %q28 = insertelement <32 x i8> %q27, i8 %q, i32 24 %q29 = insertelement <32 x i8> %q28, i8 %q, i32 25 %q2a = insertelement <32 x i8> %q29, i8 %q, i32 26 %q2b = insertelement <32 x i8> %q2a, i8 %q, i32 27 %q2c = insertelement <32 x i8> %q2b, i8 %q, i32 28 %q2d = insertelement <32 x i8> %q2c, i8 %q, i32 29 %q2e = insertelement <32 x i8> %q2d, i8 %q, i32 30 %q2f = insertelement <32 x i8> %q2e, i8 %q, i32 31 ret <32 x i8> %q2f } define <8 x i16> @W16(i16* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: W16: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vpbroadcastw (%eax), %xmm0 ; X32-NEXT: retl ; ; X64-LABEL: W16: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vpbroadcastw (%rdi), %xmm0 ; X64-NEXT: retq entry: %q = load i16, i16* %ptr, align 4 %q0 = insertelement <8 x i16> undef, i16 %q, i32 0 %q1 = insertelement <8 x i16> %q0, i16 %q, i32 1 %q2 = insertelement <8 x i16> %q1, i16 %q, i32 2 %q3 = insertelement <8 x i16> %q2, i16 %q, i32 3 %q4 = insertelement <8 x i16> %q3, i16 %q, i32 4 %q5 = insertelement <8 x i16> %q4, i16 %q, i32 5 %q6 = insertelement <8 x i16> %q5, i16 %q, i32 6 %q7 = insertelement <8 x i16> %q6, i16 %q, i32 7 ret <8 x i16> %q7 } define <16 x i16> @WW16(i16* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: WW16: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vpbroadcastw (%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: WW16: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vpbroadcastw (%rdi), %ymm0 ; X64-NEXT: retq entry: %q = load i16, i16* %ptr, align 4 %q0 = insertelement <16 x i16> undef, i16 %q, i32 0 %q1 = insertelement <16 x i16> %q0, i16 %q, i32 1 %q2 = insertelement <16 x i16> %q1, i16 %q, i32 2 %q3 = insertelement <16 x i16> %q2, i16 %q, i32 3 %q4 = insertelement <16 x i16> %q3, i16 %q, i32 4 %q5 = insertelement <16 x i16> %q4, i16 %q, i32 5 %q6 = insertelement <16 x i16> %q5, i16 %q, i32 6 %q7 = insertelement <16 x i16> %q6, i16 %q, i32 7 %q8 = insertelement <16 x i16> %q7, i16 %q, i32 8 %q9 = insertelement <16 x i16> %q8, i16 %q, i32 9 %qa = insertelement <16 x i16> %q9, i16 %q, i32 10 %qb = insertelement <16 x i16> %qa, i16 %q, i32 11 %qc = insertelement <16 x i16> %qb, i16 %q, i32 12 %qd = insertelement <16 x i16> %qc, i16 %q, i32 13 %qe = insertelement <16 x i16> %qd, i16 %q, i32 14 %qf = insertelement <16 x i16> %qe, i16 %q, i32 15 ret <16 x i16> %qf } define <4 x i32> @D32(i32* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: D32: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vbroadcastss (%eax), %xmm0 ; X32-NEXT: retl ; ; X64-LABEL: D32: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vbroadcastss (%rdi), %xmm0 ; X64-NEXT: retq entry: %q = load i32, i32* %ptr, align 4 %q0 = insertelement <4 x i32> undef, i32 %q, i32 0 %q1 = insertelement <4 x i32> %q0, i32 %q, i32 1 %q2 = insertelement <4 x i32> %q1, i32 %q, i32 2 %q3 = insertelement <4 x i32> %q2, i32 %q, i32 3 ret <4 x i32> %q3 } define <8 x i32> @DD32(i32* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: DD32: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vbroadcastss (%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: DD32: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vbroadcastss (%rdi), %ymm0 ; X64-NEXT: retq entry: %q = load i32, i32* %ptr, align 4 %q0 = insertelement <8 x i32> undef, i32 %q, i32 0 %q1 = insertelement <8 x i32> %q0, i32 %q, i32 1 %q2 = insertelement <8 x i32> %q1, i32 %q, i32 2 %q3 = insertelement <8 x i32> %q2, i32 %q, i32 3 %q4 = insertelement <8 x i32> %q3, i32 %q, i32 4 %q5 = insertelement <8 x i32> %q4, i32 %q, i32 5 %q6 = insertelement <8 x i32> %q5, i32 %q, i32 6 %q7 = insertelement <8 x i32> %q6, i32 %q, i32 7 ret <8 x i32> %q7 } define <2 x i64> @Q64(i64* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: Q64: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vpbroadcastq (%eax), %xmm0 ; X32-NEXT: retl ; ; X64-LABEL: Q64: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vpbroadcastq (%rdi), %xmm0 ; X64-NEXT: retq entry: %q = load i64, i64* %ptr, align 4 %q0 = insertelement <2 x i64> undef, i64 %q, i32 0 %q1 = insertelement <2 x i64> %q0, i64 %q, i32 1 ret <2 x i64> %q1 } define <4 x i64> @QQ64(i64* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: QQ64: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vmovsd {{.*#+}} xmm0 = mem[0],zero ; X32-NEXT: vbroadcastsd %xmm0, %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: QQ64: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vbroadcastsd (%rdi), %ymm0 ; X64-NEXT: retq entry: %q = load i64, i64* %ptr, align 4 %q0 = insertelement <4 x i64> undef, i64 %q, i32 0 %q1 = insertelement <4 x i64> %q0, i64 %q, i32 1 %q2 = insertelement <4 x i64> %q1, i64 %q, i32 2 %q3 = insertelement <4 x i64> %q2, i64 %q, i32 3 ret <4 x i64> %q3 } define <8 x i16> @broadcast_mem_v4i16_v8i16(<4 x i16>* %ptr) { ; X32-LABEL: broadcast_mem_v4i16_v8i16: ; X32: ## %bb.0: ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vmovddup {{.*#+}} xmm0 = mem[0,0] ; X32-NEXT: retl ; ; X64-LABEL: broadcast_mem_v4i16_v8i16: ; X64: ## %bb.0: ; X64-NEXT: vpbroadcastq (%rdi), %xmm0 ; X64-NEXT: retq %load = load <4 x i16>, <4 x i16>* %ptr %shuf = shufflevector <4 x i16> %load, <4 x i16> undef, <8 x i32> ret <8 x i16> %shuf } define <16 x i16> @broadcast_mem_v4i16_v16i16(<4 x i16>* %ptr) { ; X32-LABEL: broadcast_mem_v4i16_v16i16: ; X32: ## %bb.0: ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vmovsd {{.*#+}} xmm0 = mem[0],zero ; X32-NEXT: vbroadcastsd %xmm0, %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: broadcast_mem_v4i16_v16i16: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastsd (%rdi), %ymm0 ; X64-NEXT: retq %load = load <4 x i16>, <4 x i16>* %ptr %shuf = shufflevector <4 x i16> %load, <4 x i16> undef, <16 x i32> ret <16 x i16> %shuf } ; FIXME: Pointer adjusted broadcasts define <16 x i8> @load_splat_16i8_16i8_1111111111111111(<16 x i8>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_16i8_16i8_1111111111111111: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vpbroadcastb 1(%eax), %xmm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_16i8_16i8_1111111111111111: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vpbroadcastb 1(%rdi), %xmm0 ; X64-NEXT: retq entry: %ld = load <16 x i8>, <16 x i8>* %ptr %ret = shufflevector <16 x i8> %ld, <16 x i8> undef, <16 x i32> ret <16 x i8> %ret } define <32 x i8> @load_splat_32i8_16i8_11111111111111111111111111111111(<16 x i8>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_32i8_16i8_11111111111111111111111111111111: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vpbroadcastb 1(%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_32i8_16i8_11111111111111111111111111111111: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vpbroadcastb 1(%rdi), %ymm0 ; X64-NEXT: retq entry: %ld = load <16 x i8>, <16 x i8>* %ptr %ret = shufflevector <16 x i8> %ld, <16 x i8> undef, <32 x i32> ret <32 x i8> %ret } define <32 x i8> @load_splat_32i8_32i8_11111111111111111111111111111111(<32 x i8>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_32i8_32i8_11111111111111111111111111111111: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vpbroadcastb 1(%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_32i8_32i8_11111111111111111111111111111111: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vpbroadcastb 1(%rdi), %ymm0 ; X64-NEXT: retq entry: %ld = load <32 x i8>, <32 x i8>* %ptr %ret = shufflevector <32 x i8> %ld, <32 x i8> undef, <32 x i32> ret <32 x i8> %ret } define <8 x i16> @load_splat_8i16_8i16_11111111(<8 x i16>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_8i16_8i16_11111111: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vpbroadcastw 2(%eax), %xmm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_8i16_8i16_11111111: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vpbroadcastw 2(%rdi), %xmm0 ; X64-NEXT: retq entry: %ld = load <8 x i16>, <8 x i16>* %ptr %ret = shufflevector <8 x i16> %ld, <8 x i16> undef, <8 x i32> ret <8 x i16> %ret } define <16 x i16> @load_splat_16i16_8i16_1111111111111111(<8 x i16>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_16i16_8i16_1111111111111111: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vpbroadcastw 2(%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_16i16_8i16_1111111111111111: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vpbroadcastw 2(%rdi), %ymm0 ; X64-NEXT: retq entry: %ld = load <8 x i16>, <8 x i16>* %ptr %ret = shufflevector <8 x i16> %ld, <8 x i16> undef, <16 x i32> ret <16 x i16> %ret } define <16 x i16> @load_splat_16i16_16i16_1111111111111111(<16 x i16>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_16i16_16i16_1111111111111111: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vpbroadcastw 2(%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_16i16_16i16_1111111111111111: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vpbroadcastw 2(%rdi), %ymm0 ; X64-NEXT: retq entry: %ld = load <16 x i16>, <16 x i16>* %ptr %ret = shufflevector <16 x i16> %ld, <16 x i16> undef, <16 x i32> ret <16 x i16> %ret } define <4 x i32> @load_splat_4i32_4i32_1111(<4 x i32>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_4i32_4i32_1111: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vbroadcastss 4(%eax), %xmm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_4i32_4i32_1111: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vbroadcastss 4(%rdi), %xmm0 ; X64-NEXT: retq entry: %ld = load <4 x i32>, <4 x i32>* %ptr %ret = shufflevector <4 x i32> %ld, <4 x i32> undef, <4 x i32> ret <4 x i32> %ret } define <8 x i32> @load_splat_8i32_4i32_33333333(<4 x i32>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_8i32_4i32_33333333: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vbroadcastss 12(%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_8i32_4i32_33333333: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vbroadcastss 12(%rdi), %ymm0 ; X64-NEXT: retq entry: %ld = load <4 x i32>, <4 x i32>* %ptr %ret = shufflevector <4 x i32> %ld, <4 x i32> undef, <8 x i32> ret <8 x i32> %ret } define <8 x i32> @load_splat_8i32_8i32_55555555(<8 x i32>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_8i32_8i32_55555555: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vbroadcastss 20(%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_8i32_8i32_55555555: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vbroadcastss 20(%rdi), %ymm0 ; X64-NEXT: retq entry: %ld = load <8 x i32>, <8 x i32>* %ptr %ret = shufflevector <8 x i32> %ld, <8 x i32> undef, <8 x i32> ret <8 x i32> %ret } define <4 x float> @load_splat_4f32_4f32_1111(<4 x float>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_4f32_4f32_1111: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vbroadcastss 4(%eax), %xmm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_4f32_4f32_1111: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vbroadcastss 4(%rdi), %xmm0 ; X64-NEXT: retq entry: %ld = load <4 x float>, <4 x float>* %ptr %ret = shufflevector <4 x float> %ld, <4 x float> undef, <4 x i32> ret <4 x float> %ret } define <8 x float> @load_splat_8f32_4f32_33333333(<4 x float>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_8f32_4f32_33333333: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vbroadcastss 12(%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_8f32_4f32_33333333: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vbroadcastss 12(%rdi), %ymm0 ; X64-NEXT: retq entry: %ld = load <4 x float>, <4 x float>* %ptr %ret = shufflevector <4 x float> %ld, <4 x float> undef, <8 x i32> ret <8 x float> %ret } define <8 x float> @load_splat_8f32_8f32_55555555(<8 x float>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_8f32_8f32_55555555: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vbroadcastss 20(%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_8f32_8f32_55555555: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vbroadcastss 20(%rdi), %ymm0 ; X64-NEXT: retq entry: %ld = load <8 x float>, <8 x float>* %ptr %ret = shufflevector <8 x float> %ld, <8 x float> undef, <8 x i32> ret <8 x float> %ret } define <2 x i64> @load_splat_2i64_2i64_1111(<2 x i64>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_2i64_2i64_1111: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vmovddup {{.*#+}} xmm0 = mem[0,0] ; X32-NEXT: retl ; ; X64-LABEL: load_splat_2i64_2i64_1111: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vpbroadcastq 8(%rdi), %xmm0 ; X64-NEXT: retq entry: %ld = load <2 x i64>, <2 x i64>* %ptr %ret = shufflevector <2 x i64> %ld, <2 x i64> undef, <2 x i32> ret <2 x i64> %ret } define <4 x i64> @load_splat_4i64_2i64_1111(<2 x i64>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_4i64_2i64_1111: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vbroadcastsd 8(%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_4i64_2i64_1111: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vbroadcastsd 8(%rdi), %ymm0 ; X64-NEXT: retq entry: %ld = load <2 x i64>, <2 x i64>* %ptr %ret = shufflevector <2 x i64> %ld, <2 x i64> undef, <4 x i32> ret <4 x i64> %ret } define <4 x i64> @load_splat_4i64_4i64_2222(<4 x i64>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_4i64_4i64_2222: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vbroadcastsd 16(%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_4i64_4i64_2222: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vbroadcastsd 16(%rdi), %ymm0 ; X64-NEXT: retq entry: %ld = load <4 x i64>, <4 x i64>* %ptr %ret = shufflevector <4 x i64> %ld, <4 x i64> undef, <4 x i32> ret <4 x i64> %ret } define <2 x double> @load_splat_2f64_2f64_1111(<2 x double>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_2f64_2f64_1111: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vmovddup {{.*#+}} xmm0 = mem[0,0] ; X32-NEXT: retl ; ; X64-LABEL: load_splat_2f64_2f64_1111: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vmovddup {{.*#+}} xmm0 = mem[0,0] ; X64-NEXT: retq entry: %ld = load <2 x double>, <2 x double>* %ptr %ret = shufflevector <2 x double> %ld, <2 x double> undef, <2 x i32> ret <2 x double> %ret } define <4 x double> @load_splat_4f64_2f64_1111(<2 x double>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_4f64_2f64_1111: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vbroadcastsd 8(%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_4f64_2f64_1111: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vbroadcastsd 8(%rdi), %ymm0 ; X64-NEXT: retq entry: %ld = load <2 x double>, <2 x double>* %ptr %ret = shufflevector <2 x double> %ld, <2 x double> undef, <4 x i32> ret <4 x double> %ret } define <4 x double> @load_splat_4f64_4f64_2222(<4 x double>* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: load_splat_4f64_4f64_2222: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vbroadcastsd 16(%eax), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: load_splat_4f64_4f64_2222: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vbroadcastsd 16(%rdi), %ymm0 ; X64-NEXT: retq entry: %ld = load <4 x double>, <4 x double>* %ptr %ret = shufflevector <4 x double> %ld, <4 x double> undef, <4 x i32> ret <4 x double> %ret } ; make sure that we still don't support broadcast double into 128-bit vector ; this used to crash define <2 x double> @I(double* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: I: ; X32: ## %bb.0: ## %entry ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vmovddup {{.*#+}} xmm0 = mem[0,0] ; X32-NEXT: retl ; ; X64-LABEL: I: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vmovddup {{.*#+}} xmm0 = mem[0,0] ; X64-NEXT: retq entry: %q = load double, double* %ptr, align 4 %vecinit.i = insertelement <2 x double> undef, double %q, i32 0 %vecinit2.i = insertelement <2 x double> %vecinit.i, double %q, i32 1 ret <2 x double> %vecinit2.i } define <8 x i32> @V111(<8 x i32> %in) nounwind uwtable readnone ssp { ; X32-AVX2-LABEL: V111: ; X32-AVX2: ## %bb.0: ## %entry ; X32-AVX2-NEXT: vpbroadcastd {{.*#+}} ymm1 = [2,2,2,2,2,2,2,2] ; X32-AVX2-NEXT: vpaddd %ymm1, %ymm0, %ymm0 ; X32-AVX2-NEXT: retl ; ; X64-AVX2-LABEL: V111: ; X64-AVX2: ## %bb.0: ## %entry ; X64-AVX2-NEXT: vpbroadcastd {{.*#+}} ymm1 = [2,2,2,2,2,2,2,2] ; X64-AVX2-NEXT: vpaddd %ymm1, %ymm0, %ymm0 ; X64-AVX2-NEXT: retq ; ; X32-AVX512VL-LABEL: V111: ; X32-AVX512VL: ## %bb.0: ## %entry ; X32-AVX512VL-NEXT: vpaddd LCPI29_0{1to8}, %ymm0, %ymm0 ; X32-AVX512VL-NEXT: retl ; ; X64-AVX512VL-LABEL: V111: ; X64-AVX512VL: ## %bb.0: ## %entry ; X64-AVX512VL-NEXT: vpaddd {{.*}}(%rip){1to8}, %ymm0, %ymm0 ; X64-AVX512VL-NEXT: retq entry: %g = add <8 x i32> %in, ret <8 x i32> %g } define <8 x float> @V113(<8 x float> %in) nounwind uwtable readnone ssp { ; X32-AVX2-LABEL: V113: ; X32-AVX2: ## %bb.0: ## %entry ; X32-AVX2-NEXT: vbroadcastss {{.*#+}} ymm1 = [-0.0078125,-0.0078125,-0.0078125,-0.0078125,-0.0078125,-0.0078125,-0.0078125,-0.0078125] ; X32-AVX2-NEXT: vaddps %ymm1, %ymm0, %ymm0 ; X32-AVX2-NEXT: retl ; ; X64-AVX2-LABEL: V113: ; X64-AVX2: ## %bb.0: ## %entry ; X64-AVX2-NEXT: vbroadcastss {{.*#+}} ymm1 = [-0.0078125,-0.0078125,-0.0078125,-0.0078125,-0.0078125,-0.0078125,-0.0078125,-0.0078125] ; X64-AVX2-NEXT: vaddps %ymm1, %ymm0, %ymm0 ; X64-AVX2-NEXT: retq ; ; X32-AVX512VL-LABEL: V113: ; X32-AVX512VL: ## %bb.0: ## %entry ; X32-AVX512VL-NEXT: vaddps LCPI30_0{1to8}, %ymm0, %ymm0 ; X32-AVX512VL-NEXT: retl ; ; X64-AVX512VL-LABEL: V113: ; X64-AVX512VL: ## %bb.0: ## %entry ; X64-AVX512VL-NEXT: vaddps {{.*}}(%rip){1to8}, %ymm0, %ymm0 ; X64-AVX512VL-NEXT: retq entry: %g = fadd <8 x float> %in, ret <8 x float> %g } define <4 x float> @_e2(float* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: _e2: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastss {{.*#+}} xmm0 = [-0.0078125,-0.0078125,-0.0078125,-0.0078125] ; X32-NEXT: retl ; ; X64-LABEL: _e2: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastss {{.*#+}} xmm0 = [-0.0078125,-0.0078125,-0.0078125,-0.0078125] ; X64-NEXT: retq %vecinit.i = insertelement <4 x float> undef, float 0xbf80000000000000, i32 0 %vecinit2.i = insertelement <4 x float> %vecinit.i, float 0xbf80000000000000, i32 1 %vecinit4.i = insertelement <4 x float> %vecinit2.i, float 0xbf80000000000000, i32 2 %vecinit6.i = insertelement <4 x float> %vecinit4.i, float 0xbf80000000000000, i32 3 ret <4 x float> %vecinit6.i } define <8 x i8> @_e4(i8* %ptr) nounwind uwtable readnone ssp { ; X32-LABEL: _e4: ; X32: ## %bb.0: ; X32-NEXT: vmovaps {{.*#+}} xmm0 = [52,52,52,52,52,52,52,52] ; X32-NEXT: retl ; ; X64-LABEL: _e4: ; X64: ## %bb.0: ; X64-NEXT: vmovaps {{.*#+}} xmm0 = [52,52,52,52,52,52,52,52] ; X64-NEXT: retq %vecinit0.i = insertelement <8 x i8> undef, i8 52, i32 0 %vecinit1.i = insertelement <8 x i8> %vecinit0.i, i8 52, i32 1 %vecinit2.i = insertelement <8 x i8> %vecinit1.i, i8 52, i32 2 %vecinit3.i = insertelement <8 x i8> %vecinit2.i, i8 52, i32 3 %vecinit4.i = insertelement <8 x i8> %vecinit3.i, i8 52, i32 4 %vecinit5.i = insertelement <8 x i8> %vecinit4.i, i8 52, i32 5 %vecinit6.i = insertelement <8 x i8> %vecinit5.i, i8 52, i32 6 %vecinit7.i = insertelement <8 x i8> %vecinit6.i, i8 52, i32 7 ret <8 x i8> %vecinit7.i } define void @crash() nounwind alwaysinline { ; X32-LABEL: crash: ; X32: ## %bb.0: ## %WGLoopsEntry ; X32-NEXT: xorl %eax, %eax ; X32-NEXT: testb %al, %al ; X32-NEXT: je LBB33_1 ; X32-NEXT: ## %bb.2: ## %ret ; X32-NEXT: retl ; X32-NEXT: .p2align 4, 0x90 ; X32-NEXT: LBB33_1: ## %footer329VF ; X32-NEXT: ## =>This Inner Loop Header: Depth=1 ; X32-NEXT: jmp LBB33_1 ; ; X64-LABEL: crash: ; X64: ## %bb.0: ## %WGLoopsEntry ; X64-NEXT: xorl %eax, %eax ; X64-NEXT: testb %al, %al ; X64-NEXT: je LBB33_1 ; X64-NEXT: ## %bb.2: ## %ret ; X64-NEXT: retq ; X64-NEXT: .p2align 4, 0x90 ; X64-NEXT: LBB33_1: ## %footer329VF ; X64-NEXT: ## =>This Inner Loop Header: Depth=1 ; X64-NEXT: jmp LBB33_1 WGLoopsEntry: br i1 undef, label %ret, label %footer329VF footer329VF: %A.0.inVF = fmul float undef, 6.553600e+04 %B.0.in407VF = fmul <8 x float> undef, %A.0VF = fptosi float %A.0.inVF to i32 %B.0408VF = fptosi <8 x float> %B.0.in407VF to <8 x i32> %0 = and <8 x i32> %B.0408VF, %1 = and i32 %A.0VF, 65535 %temp1098VF = insertelement <8 x i32> undef, i32 %1, i32 0 %vector1099VF = shufflevector <8 x i32> %temp1098VF, <8 x i32> undef, <8 x i32> zeroinitializer br i1 undef, label %preload1201VF, label %footer349VF preload1201VF: br label %footer349VF footer349VF: %2 = mul nsw <8 x i32> undef, %0 %3 = mul nsw <8 x i32> undef, %vector1099VF br label %footer329VF ret: ret void } define <8 x i32> @_inreg0(i32 %scalar) nounwind uwtable readnone ssp { ; X32-LABEL: _inreg0: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastss {{[0-9]+}}(%esp), %ymm0 ; X32-NEXT: retl ; ; X64-AVX2-LABEL: _inreg0: ; X64-AVX2: ## %bb.0: ; X64-AVX2-NEXT: vmovd %edi, %xmm0 ; X64-AVX2-NEXT: vpbroadcastd %xmm0, %ymm0 ; X64-AVX2-NEXT: retq ; ; X64-AVX512VL-LABEL: _inreg0: ; X64-AVX512VL: ## %bb.0: ; X64-AVX512VL-NEXT: vpbroadcastd %edi, %ymm0 ; X64-AVX512VL-NEXT: retq %in = insertelement <8 x i32> undef, i32 %scalar, i32 0 %wide = shufflevector <8 x i32> %in, <8 x i32> undef, <8 x i32> zeroinitializer ret <8 x i32> %wide } define <8 x float> @_inreg1(float %scalar) nounwind uwtable readnone ssp { ; X32-LABEL: _inreg1: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastss {{[0-9]+}}(%esp), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: _inreg1: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastss %xmm0, %ymm0 ; X64-NEXT: retq %in = insertelement <8 x float> undef, float %scalar, i32 0 %wide = shufflevector <8 x float> %in, <8 x float> undef, <8 x i32> zeroinitializer ret <8 x float> %wide } define <4 x float> @_inreg2(float %scalar) nounwind uwtable readnone ssp { ; X32-LABEL: _inreg2: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastss {{[0-9]+}}(%esp), %xmm0 ; X32-NEXT: retl ; ; X64-LABEL: _inreg2: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastss %xmm0, %xmm0 ; X64-NEXT: retq %in = insertelement <4 x float> undef, float %scalar, i32 0 %wide = shufflevector <4 x float> %in, <4 x float> undef, <4 x i32> zeroinitializer ret <4 x float> %wide } define <4 x double> @_inreg3(double %scalar) nounwind uwtable readnone ssp { ; X32-LABEL: _inreg3: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastsd {{[0-9]+}}(%esp), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: _inreg3: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastsd %xmm0, %ymm0 ; X64-NEXT: retq %in = insertelement <4 x double> undef, double %scalar, i32 0 %wide = shufflevector <4 x double> %in, <4 x double> undef, <4 x i32> zeroinitializer ret <4 x double> %wide } define <8 x float> @_inreg8xfloat(<8 x float> %a) { ; X32-LABEL: _inreg8xfloat: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastss %xmm0, %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: _inreg8xfloat: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastss %xmm0, %ymm0 ; X64-NEXT: retq %b = shufflevector <8 x float> %a, <8 x float> undef, <8 x i32> zeroinitializer ret <8 x float> %b } define <4 x float> @_inreg4xfloat(<4 x float> %a) { ; X32-LABEL: _inreg4xfloat: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastss %xmm0, %xmm0 ; X32-NEXT: retl ; ; X64-LABEL: _inreg4xfloat: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastss %xmm0, %xmm0 ; X64-NEXT: retq %b = shufflevector <4 x float> %a, <4 x float> undef, <4 x i32> zeroinitializer ret <4 x float> %b } define <16 x i16> @_inreg16xi16(<16 x i16> %a) { ; X32-LABEL: _inreg16xi16: ; X32: ## %bb.0: ; X32-NEXT: vpbroadcastw %xmm0, %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: _inreg16xi16: ; X64: ## %bb.0: ; X64-NEXT: vpbroadcastw %xmm0, %ymm0 ; X64-NEXT: retq %b = shufflevector <16 x i16> %a, <16 x i16> undef, <16 x i32> zeroinitializer ret <16 x i16> %b } define <8 x i16> @_inreg8xi16(<8 x i16> %a) { ; X32-LABEL: _inreg8xi16: ; X32: ## %bb.0: ; X32-NEXT: vpbroadcastw %xmm0, %xmm0 ; X32-NEXT: retl ; ; X64-LABEL: _inreg8xi16: ; X64: ## %bb.0: ; X64-NEXT: vpbroadcastw %xmm0, %xmm0 ; X64-NEXT: retq %b = shufflevector <8 x i16> %a, <8 x i16> undef, <8 x i32> zeroinitializer ret <8 x i16> %b } define <4 x i64> @_inreg4xi64(<4 x i64> %a) { ; X32-LABEL: _inreg4xi64: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastsd %xmm0, %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: _inreg4xi64: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastsd %xmm0, %ymm0 ; X64-NEXT: retq %b = shufflevector <4 x i64> %a, <4 x i64> undef, <4 x i32> zeroinitializer ret <4 x i64> %b } define <2 x i64> @_inreg2xi64(<2 x i64> %a) { ; X32-LABEL: _inreg2xi64: ; X32: ## %bb.0: ; X32-NEXT: vpbroadcastq %xmm0, %xmm0 ; X32-NEXT: retl ; ; X64-LABEL: _inreg2xi64: ; X64: ## %bb.0: ; X64-NEXT: vpbroadcastq %xmm0, %xmm0 ; X64-NEXT: retq %b = shufflevector <2 x i64> %a, <2 x i64> undef, <2 x i32> zeroinitializer ret <2 x i64> %b } define <4 x double> @_inreg4xdouble(<4 x double> %a) { ; X32-LABEL: _inreg4xdouble: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastsd %xmm0, %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: _inreg4xdouble: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastsd %xmm0, %ymm0 ; X64-NEXT: retq %b = shufflevector <4 x double> %a, <4 x double> undef, <4 x i32> zeroinitializer ret <4 x double> %b } define <2 x double> @_inreg2xdouble(<2 x double> %a) { ; X32-LABEL: _inreg2xdouble: ; X32: ## %bb.0: ; X32-NEXT: vmovddup {{.*#+}} xmm0 = xmm0[0,0] ; X32-NEXT: retl ; ; X64-LABEL: _inreg2xdouble: ; X64: ## %bb.0: ; X64-NEXT: vmovddup {{.*#+}} xmm0 = xmm0[0,0] ; X64-NEXT: retq %b = shufflevector <2 x double> %a, <2 x double> undef, <2 x i32> zeroinitializer ret <2 x double> %b } define <8 x i32> @_inreg8xi32(<8 x i32> %a) { ; X32-LABEL: _inreg8xi32: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastss %xmm0, %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: _inreg8xi32: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastss %xmm0, %ymm0 ; X64-NEXT: retq %b = shufflevector <8 x i32> %a, <8 x i32> undef, <8 x i32> zeroinitializer ret <8 x i32> %b } define <4 x i32> @_inreg4xi32(<4 x i32> %a) { ; X32-LABEL: _inreg4xi32: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastss %xmm0, %xmm0 ; X32-NEXT: retl ; ; X64-LABEL: _inreg4xi32: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastss %xmm0, %xmm0 ; X64-NEXT: retq %b = shufflevector <4 x i32> %a, <4 x i32> undef, <4 x i32> zeroinitializer ret <4 x i32> %b } define <32 x i8> @_inreg32xi8(<32 x i8> %a) { ; X32-LABEL: _inreg32xi8: ; X32: ## %bb.0: ; X32-NEXT: vpbroadcastb %xmm0, %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: _inreg32xi8: ; X64: ## %bb.0: ; X64-NEXT: vpbroadcastb %xmm0, %ymm0 ; X64-NEXT: retq %b = shufflevector <32 x i8> %a, <32 x i8> undef, <32 x i32> zeroinitializer ret <32 x i8> %b } define <16 x i8> @_inreg16xi8(<16 x i8> %a) { ; X32-LABEL: _inreg16xi8: ; X32: ## %bb.0: ; X32-NEXT: vpbroadcastb %xmm0, %xmm0 ; X32-NEXT: retl ; ; X64-LABEL: _inreg16xi8: ; X64: ## %bb.0: ; X64-NEXT: vpbroadcastb %xmm0, %xmm0 ; X64-NEXT: retq %b = shufflevector <16 x i8> %a, <16 x i8> undef, <16 x i32> zeroinitializer ret <16 x i8> %b } ; These tests check that a vbroadcast instruction is used when we have a splat ; formed from a concat_vectors (via the shufflevector) of two BUILD_VECTORs ; (via the insertelements). define <8 x float> @splat_concat1(float %f) { ; X32-LABEL: splat_concat1: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastss {{[0-9]+}}(%esp), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: splat_concat1: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastss %xmm0, %ymm0 ; X64-NEXT: retq %1 = insertelement <4 x float> undef, float %f, i32 0 %2 = insertelement <4 x float> %1, float %f, i32 1 %3 = insertelement <4 x float> %2, float %f, i32 2 %4 = insertelement <4 x float> %3, float %f, i32 3 %5 = shufflevector <4 x float> %4, <4 x float> undef, <8 x i32> ret <8 x float> %5 } define <8 x float> @splat_concat2(float %f) { ; X32-LABEL: splat_concat2: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastss {{[0-9]+}}(%esp), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: splat_concat2: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastss %xmm0, %ymm0 ; X64-NEXT: retq %1 = insertelement <4 x float> undef, float %f, i32 0 %2 = insertelement <4 x float> %1, float %f, i32 1 %3 = insertelement <4 x float> %2, float %f, i32 2 %4 = insertelement <4 x float> %3, float %f, i32 3 %5 = insertelement <4 x float> undef, float %f, i32 0 %6 = insertelement <4 x float> %5, float %f, i32 1 %7 = insertelement <4 x float> %6, float %f, i32 2 %8 = insertelement <4 x float> %7, float %f, i32 3 %9 = shufflevector <4 x float> %4, <4 x float> %8, <8 x i32> ret <8 x float> %9 } define <4 x double> @splat_concat3(double %d) { ; X32-LABEL: splat_concat3: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastsd {{[0-9]+}}(%esp), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: splat_concat3: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastsd %xmm0, %ymm0 ; X64-NEXT: retq %1 = insertelement <2 x double> undef, double %d, i32 0 %2 = insertelement <2 x double> %1, double %d, i32 1 %3 = shufflevector <2 x double> %2, <2 x double> undef, <4 x i32> ret <4 x double> %3 } define <4 x double> @splat_concat4(double %d) { ; X32-LABEL: splat_concat4: ; X32: ## %bb.0: ; X32-NEXT: vbroadcastsd {{[0-9]+}}(%esp), %ymm0 ; X32-NEXT: retl ; ; X64-LABEL: splat_concat4: ; X64: ## %bb.0: ; X64-NEXT: vbroadcastsd %xmm0, %ymm0 ; X64-NEXT: retq %1 = insertelement <2 x double> undef, double %d, i32 0 %2 = insertelement <2 x double> %1, double %d, i32 1 %3 = insertelement <2 x double> undef, double %d, i32 0 %4 = insertelement <2 x double> %3, double %d, i32 1 %5 = shufflevector <2 x double> %2, <2 x double> %4, <4 x i32> ret <4 x double> %5 } ; Test cases for . ; Instruction selection for broacast instruction fails if ; the load cannot be folded into the broadcast. ; This happens if the load has initial one use but other uses are ; created later, or if selection DAG cannot prove that folding the ; load will not create a cycle in the DAG. ; Those test cases exerce the latter. define void @isel_crash_16b(i8* %cV_R.addr) { ; X32-LABEL: isel_crash_16b: ; X32: ## %bb.0: ## %eintry ; X32-NEXT: subl $60, %esp ; X32-NEXT: .cfi_def_cfa_offset 64 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X32-NEXT: vmovaps %xmm0, (%esp) ; X32-NEXT: vpbroadcastb (%eax), %xmm1 ; X32-NEXT: vmovaps %xmm0, {{[0-9]+}}(%esp) ; X32-NEXT: vmovdqa %xmm1, {{[0-9]+}}(%esp) ; X32-NEXT: addl $60, %esp ; X32-NEXT: retl ; ; X64-LABEL: isel_crash_16b: ; X64: ## %bb.0: ## %eintry ; X64-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X64-NEXT: vmovaps %xmm0, -{{[0-9]+}}(%rsp) ; X64-NEXT: vpbroadcastb (%rdi), %xmm1 ; X64-NEXT: vmovaps %xmm0, -{{[0-9]+}}(%rsp) ; X64-NEXT: vmovdqa %xmm1, -{{[0-9]+}}(%rsp) ; X64-NEXT: retq eintry: %__a.addr.i = alloca <2 x i64>, align 16 %__b.addr.i = alloca <2 x i64>, align 16 %vCr = alloca <2 x i64>, align 16 store <2 x i64> zeroinitializer, <2 x i64>* %vCr, align 16 %tmp = load <2 x i64>, <2 x i64>* %vCr, align 16 %tmp2 = load i8, i8* %cV_R.addr, align 4 %splat.splatinsert = insertelement <16 x i8> undef, i8 %tmp2, i32 0 %splat.splat = shufflevector <16 x i8> %splat.splatinsert, <16 x i8> undef, <16 x i32> zeroinitializer %tmp3 = bitcast <16 x i8> %splat.splat to <2 x i64> store <2 x i64> %tmp, <2 x i64>* %__a.addr.i, align 16 store <2 x i64> %tmp3, <2 x i64>* %__b.addr.i, align 16 ret void } define void @isel_crash_32b(i8* %cV_R.addr) { ; X32-LABEL: isel_crash_32b: ; X32: ## %bb.0: ## %eintry ; X32-NEXT: pushl %ebp ; X32-NEXT: .cfi_def_cfa_offset 8 ; X32-NEXT: .cfi_offset %ebp, -8 ; X32-NEXT: movl %esp, %ebp ; X32-NEXT: .cfi_def_cfa_register %ebp ; X32-NEXT: andl $-32, %esp ; X32-NEXT: subl $128, %esp ; X32-NEXT: movl 8(%ebp), %eax ; X32-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X32-NEXT: vmovaps %ymm0, (%esp) ; X32-NEXT: vpbroadcastb (%eax), %ymm1 ; X32-NEXT: vmovaps %ymm0, {{[0-9]+}}(%esp) ; X32-NEXT: vmovdqa %ymm1, {{[0-9]+}}(%esp) ; X32-NEXT: movl %ebp, %esp ; X32-NEXT: popl %ebp ; X32-NEXT: vzeroupper ; X32-NEXT: retl ; ; X64-LABEL: isel_crash_32b: ; X64: ## %bb.0: ## %eintry ; X64-NEXT: pushq %rbp ; X64-NEXT: .cfi_def_cfa_offset 16 ; X64-NEXT: .cfi_offset %rbp, -16 ; X64-NEXT: movq %rsp, %rbp ; X64-NEXT: .cfi_def_cfa_register %rbp ; X64-NEXT: andq $-32, %rsp ; X64-NEXT: subq $128, %rsp ; X64-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X64-NEXT: vmovaps %ymm0, (%rsp) ; X64-NEXT: vpbroadcastb (%rdi), %ymm1 ; X64-NEXT: vmovaps %ymm0, {{[0-9]+}}(%rsp) ; X64-NEXT: vmovdqa %ymm1, {{[0-9]+}}(%rsp) ; X64-NEXT: movq %rbp, %rsp ; X64-NEXT: popq %rbp ; X64-NEXT: vzeroupper ; X64-NEXT: retq eintry: %__a.addr.i = alloca <4 x i64>, align 16 %__b.addr.i = alloca <4 x i64>, align 16 %vCr = alloca <4 x i64>, align 16 store <4 x i64> zeroinitializer, <4 x i64>* %vCr, align 16 %tmp = load <4 x i64>, <4 x i64>* %vCr, align 16 %tmp2 = load i8, i8* %cV_R.addr, align 4 %splat.splatinsert = insertelement <32 x i8> undef, i8 %tmp2, i32 0 %splat.splat = shufflevector <32 x i8> %splat.splatinsert, <32 x i8> undef, <32 x i32> zeroinitializer %tmp3 = bitcast <32 x i8> %splat.splat to <4 x i64> store <4 x i64> %tmp, <4 x i64>* %__a.addr.i, align 16 store <4 x i64> %tmp3, <4 x i64>* %__b.addr.i, align 16 ret void } define void @isel_crash_8w(i16* %cV_R.addr) { ; X32-LABEL: isel_crash_8w: ; X32: ## %bb.0: ## %entry ; X32-NEXT: subl $60, %esp ; X32-NEXT: .cfi_def_cfa_offset 64 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X32-NEXT: vmovaps %xmm0, (%esp) ; X32-NEXT: vpbroadcastw (%eax), %xmm1 ; X32-NEXT: vmovaps %xmm0, {{[0-9]+}}(%esp) ; X32-NEXT: vmovdqa %xmm1, {{[0-9]+}}(%esp) ; X32-NEXT: addl $60, %esp ; X32-NEXT: retl ; ; X64-LABEL: isel_crash_8w: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X64-NEXT: vmovaps %xmm0, -{{[0-9]+}}(%rsp) ; X64-NEXT: vpbroadcastw (%rdi), %xmm1 ; X64-NEXT: vmovaps %xmm0, -{{[0-9]+}}(%rsp) ; X64-NEXT: vmovdqa %xmm1, -{{[0-9]+}}(%rsp) ; X64-NEXT: retq entry: %__a.addr.i = alloca <2 x i64>, align 16 %__b.addr.i = alloca <2 x i64>, align 16 %vCr = alloca <2 x i64>, align 16 store <2 x i64> zeroinitializer, <2 x i64>* %vCr, align 16 %tmp = load <2 x i64>, <2 x i64>* %vCr, align 16 %tmp2 = load i16, i16* %cV_R.addr, align 4 %splat.splatinsert = insertelement <8 x i16> undef, i16 %tmp2, i32 0 %splat.splat = shufflevector <8 x i16> %splat.splatinsert, <8 x i16> undef, <8 x i32> zeroinitializer %tmp3 = bitcast <8 x i16> %splat.splat to <2 x i64> store <2 x i64> %tmp, <2 x i64>* %__a.addr.i, align 16 store <2 x i64> %tmp3, <2 x i64>* %__b.addr.i, align 16 ret void } define void @isel_crash_16w(i16* %cV_R.addr) { ; X32-LABEL: isel_crash_16w: ; X32: ## %bb.0: ## %eintry ; X32-NEXT: pushl %ebp ; X32-NEXT: .cfi_def_cfa_offset 8 ; X32-NEXT: .cfi_offset %ebp, -8 ; X32-NEXT: movl %esp, %ebp ; X32-NEXT: .cfi_def_cfa_register %ebp ; X32-NEXT: andl $-32, %esp ; X32-NEXT: subl $128, %esp ; X32-NEXT: movl 8(%ebp), %eax ; X32-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X32-NEXT: vmovaps %ymm0, (%esp) ; X32-NEXT: vpbroadcastw (%eax), %ymm1 ; X32-NEXT: vmovaps %ymm0, {{[0-9]+}}(%esp) ; X32-NEXT: vmovdqa %ymm1, {{[0-9]+}}(%esp) ; X32-NEXT: movl %ebp, %esp ; X32-NEXT: popl %ebp ; X32-NEXT: vzeroupper ; X32-NEXT: retl ; ; X64-LABEL: isel_crash_16w: ; X64: ## %bb.0: ## %eintry ; X64-NEXT: pushq %rbp ; X64-NEXT: .cfi_def_cfa_offset 16 ; X64-NEXT: .cfi_offset %rbp, -16 ; X64-NEXT: movq %rsp, %rbp ; X64-NEXT: .cfi_def_cfa_register %rbp ; X64-NEXT: andq $-32, %rsp ; X64-NEXT: subq $128, %rsp ; X64-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X64-NEXT: vmovaps %ymm0, (%rsp) ; X64-NEXT: vpbroadcastw (%rdi), %ymm1 ; X64-NEXT: vmovaps %ymm0, {{[0-9]+}}(%rsp) ; X64-NEXT: vmovdqa %ymm1, {{[0-9]+}}(%rsp) ; X64-NEXT: movq %rbp, %rsp ; X64-NEXT: popq %rbp ; X64-NEXT: vzeroupper ; X64-NEXT: retq eintry: %__a.addr.i = alloca <4 x i64>, align 16 %__b.addr.i = alloca <4 x i64>, align 16 %vCr = alloca <4 x i64>, align 16 store <4 x i64> zeroinitializer, <4 x i64>* %vCr, align 16 %tmp = load <4 x i64>, <4 x i64>* %vCr, align 16 %tmp2 = load i16, i16* %cV_R.addr, align 4 %splat.splatinsert = insertelement <16 x i16> undef, i16 %tmp2, i32 0 %splat.splat = shufflevector <16 x i16> %splat.splatinsert, <16 x i16> undef, <16 x i32> zeroinitializer %tmp3 = bitcast <16 x i16> %splat.splat to <4 x i64> store <4 x i64> %tmp, <4 x i64>* %__a.addr.i, align 16 store <4 x i64> %tmp3, <4 x i64>* %__b.addr.i, align 16 ret void } define void @isel_crash_4d(i32* %cV_R.addr) { ; X32-LABEL: isel_crash_4d: ; X32: ## %bb.0: ## %entry ; X32-NEXT: subl $60, %esp ; X32-NEXT: .cfi_def_cfa_offset 64 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X32-NEXT: vmovaps %xmm0, (%esp) ; X32-NEXT: vbroadcastss (%eax), %xmm1 ; X32-NEXT: vmovaps %xmm0, {{[0-9]+}}(%esp) ; X32-NEXT: vmovaps %xmm1, {{[0-9]+}}(%esp) ; X32-NEXT: addl $60, %esp ; X32-NEXT: retl ; ; X64-LABEL: isel_crash_4d: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X64-NEXT: vmovaps %xmm0, -{{[0-9]+}}(%rsp) ; X64-NEXT: vbroadcastss (%rdi), %xmm1 ; X64-NEXT: vmovaps %xmm0, -{{[0-9]+}}(%rsp) ; X64-NEXT: vmovaps %xmm1, -{{[0-9]+}}(%rsp) ; X64-NEXT: retq entry: %__a.addr.i = alloca <2 x i64>, align 16 %__b.addr.i = alloca <2 x i64>, align 16 %vCr = alloca <2 x i64>, align 16 store <2 x i64> zeroinitializer, <2 x i64>* %vCr, align 16 %tmp = load <2 x i64>, <2 x i64>* %vCr, align 16 %tmp2 = load i32, i32* %cV_R.addr, align 4 %splat.splatinsert = insertelement <4 x i32> undef, i32 %tmp2, i32 0 %splat.splat = shufflevector <4 x i32> %splat.splatinsert, <4 x i32> undef, <4 x i32> zeroinitializer %tmp3 = bitcast <4 x i32> %splat.splat to <2 x i64> store <2 x i64> %tmp, <2 x i64>* %__a.addr.i, align 16 store <2 x i64> %tmp3, <2 x i64>* %__b.addr.i, align 16 ret void } define void @isel_crash_8d(i32* %cV_R.addr) { ; X32-LABEL: isel_crash_8d: ; X32: ## %bb.0: ## %eintry ; X32-NEXT: pushl %ebp ; X32-NEXT: .cfi_def_cfa_offset 8 ; X32-NEXT: .cfi_offset %ebp, -8 ; X32-NEXT: movl %esp, %ebp ; X32-NEXT: .cfi_def_cfa_register %ebp ; X32-NEXT: andl $-32, %esp ; X32-NEXT: subl $128, %esp ; X32-NEXT: movl 8(%ebp), %eax ; X32-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X32-NEXT: vmovaps %ymm0, (%esp) ; X32-NEXT: vbroadcastss (%eax), %ymm1 ; X32-NEXT: vmovaps %ymm0, {{[0-9]+}}(%esp) ; X32-NEXT: vmovaps %ymm1, {{[0-9]+}}(%esp) ; X32-NEXT: movl %ebp, %esp ; X32-NEXT: popl %ebp ; X32-NEXT: vzeroupper ; X32-NEXT: retl ; ; X64-LABEL: isel_crash_8d: ; X64: ## %bb.0: ## %eintry ; X64-NEXT: pushq %rbp ; X64-NEXT: .cfi_def_cfa_offset 16 ; X64-NEXT: .cfi_offset %rbp, -16 ; X64-NEXT: movq %rsp, %rbp ; X64-NEXT: .cfi_def_cfa_register %rbp ; X64-NEXT: andq $-32, %rsp ; X64-NEXT: subq $128, %rsp ; X64-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X64-NEXT: vmovaps %ymm0, (%rsp) ; X64-NEXT: vbroadcastss (%rdi), %ymm1 ; X64-NEXT: vmovaps %ymm0, {{[0-9]+}}(%rsp) ; X64-NEXT: vmovaps %ymm1, {{[0-9]+}}(%rsp) ; X64-NEXT: movq %rbp, %rsp ; X64-NEXT: popq %rbp ; X64-NEXT: vzeroupper ; X64-NEXT: retq eintry: %__a.addr.i = alloca <4 x i64>, align 16 %__b.addr.i = alloca <4 x i64>, align 16 %vCr = alloca <4 x i64>, align 16 store <4 x i64> zeroinitializer, <4 x i64>* %vCr, align 16 %tmp = load <4 x i64>, <4 x i64>* %vCr, align 16 %tmp2 = load i32, i32* %cV_R.addr, align 4 %splat.splatinsert = insertelement <8 x i32> undef, i32 %tmp2, i32 0 %splat.splat = shufflevector <8 x i32> %splat.splatinsert, <8 x i32> undef, <8 x i32> zeroinitializer %tmp3 = bitcast <8 x i32> %splat.splat to <4 x i64> store <4 x i64> %tmp, <4 x i64>* %__a.addr.i, align 16 store <4 x i64> %tmp3, <4 x i64>* %__b.addr.i, align 16 ret void } define void @isel_crash_2q(i64* %cV_R.addr) { ; X32-LABEL: isel_crash_2q: ; X32: ## %bb.0: ## %entry ; X32-NEXT: subl $60, %esp ; X32-NEXT: .cfi_def_cfa_offset 64 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax ; X32-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X32-NEXT: vmovaps %xmm0, (%esp) ; X32-NEXT: vpbroadcastq (%eax), %xmm1 ; X32-NEXT: vmovaps %xmm0, {{[0-9]+}}(%esp) ; X32-NEXT: vmovdqa %xmm1, {{[0-9]+}}(%esp) ; X32-NEXT: addl $60, %esp ; X32-NEXT: retl ; ; X64-LABEL: isel_crash_2q: ; X64: ## %bb.0: ## %entry ; X64-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X64-NEXT: vmovaps %xmm0, -{{[0-9]+}}(%rsp) ; X64-NEXT: vpbroadcastq (%rdi), %xmm1 ; X64-NEXT: vmovaps %xmm0, -{{[0-9]+}}(%rsp) ; X64-NEXT: vmovdqa %xmm1, -{{[0-9]+}}(%rsp) ; X64-NEXT: retq entry: %__a.addr.i = alloca <2 x i64>, align 16 %__b.addr.i = alloca <2 x i64>, align 16 %vCr = alloca <2 x i64>, align 16 store <2 x i64> zeroinitializer, <2 x i64>* %vCr, align 16 %tmp = load <2 x i64>, <2 x i64>* %vCr, align 16 %tmp2 = load i64, i64* %cV_R.addr, align 4 %splat.splatinsert = insertelement <2 x i64> undef, i64 %tmp2, i32 0 %splat.splat = shufflevector <2 x i64> %splat.splatinsert, <2 x i64> undef, <2 x i32> zeroinitializer store <2 x i64> %tmp, <2 x i64>* %__a.addr.i, align 16 store <2 x i64> %splat.splat, <2 x i64>* %__b.addr.i, align 16 ret void } define void @isel_crash_4q(i64* %cV_R.addr) { ; X32-LABEL: isel_crash_4q: ; X32: ## %bb.0: ## %eintry ; X32-NEXT: pushl %ebp ; X32-NEXT: .cfi_def_cfa_offset 8 ; X32-NEXT: .cfi_offset %ebp, -8 ; X32-NEXT: movl %esp, %ebp ; X32-NEXT: .cfi_def_cfa_register %ebp ; X32-NEXT: andl $-32, %esp ; X32-NEXT: subl $128, %esp ; X32-NEXT: movl 8(%ebp), %eax ; X32-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X32-NEXT: vmovaps %ymm0, (%esp) ; X32-NEXT: vmovsd {{.*#+}} xmm1 = mem[0],zero ; X32-NEXT: vbroadcastsd %xmm1, %ymm1 ; X32-NEXT: vmovaps %ymm0, {{[0-9]+}}(%esp) ; X32-NEXT: vmovaps %ymm1, {{[0-9]+}}(%esp) ; X32-NEXT: movl %ebp, %esp ; X32-NEXT: popl %ebp ; X32-NEXT: vzeroupper ; X32-NEXT: retl ; ; X64-LABEL: isel_crash_4q: ; X64: ## %bb.0: ## %eintry ; X64-NEXT: pushq %rbp ; X64-NEXT: .cfi_def_cfa_offset 16 ; X64-NEXT: .cfi_offset %rbp, -16 ; X64-NEXT: movq %rsp, %rbp ; X64-NEXT: .cfi_def_cfa_register %rbp ; X64-NEXT: andq $-32, %rsp ; X64-NEXT: subq $128, %rsp ; X64-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; X64-NEXT: vmovaps %ymm0, (%rsp) ; X64-NEXT: vbroadcastsd (%rdi), %ymm1 ; X64-NEXT: vmovaps %ymm0, {{[0-9]+}}(%rsp) ; X64-NEXT: vmovaps %ymm1, {{[0-9]+}}(%rsp) ; X64-NEXT: movq %rbp, %rsp ; X64-NEXT: popq %rbp ; X64-NEXT: vzeroupper ; X64-NEXT: retq eintry: %__a.addr.i = alloca <4 x i64>, align 16 %__b.addr.i = alloca <4 x i64>, align 16 %vCr = alloca <4 x i64>, align 16 store <4 x i64> zeroinitializer, <4 x i64>* %vCr, align 16 %tmp = load <4 x i64>, <4 x i64>* %vCr, align 16 %tmp2 = load i64, i64* %cV_R.addr, align 4 %splat.splatinsert = insertelement <4 x i64> undef, i64 %tmp2, i32 0 %splat.splat = shufflevector <4 x i64> %splat.splatinsert, <4 x i64> undef, <4 x i32> zeroinitializer store <4 x i64> %tmp, <4 x i64>* %__a.addr.i, align 16 store <4 x i64> %splat.splat, <4 x i64>* %__b.addr.i, align 16 ret void }