/external/llvm/test/CodeGen/X86/ |
D | 2008-07-19-movups-spills.ll | 7 @0 = external global <4 x float>, align 1 ; <<4 x float>*>:0 [#uses=2] 8 @1 = external global <4 x float>, align 1 ; <<4 x float>*>:1 [#uses=1] 9 @2 = external global <4 x float>, align 1 ; <<4 x float>*>:2 [#uses=1] 10 @3 = external global <4 x float>, align 1 ; <<4 x float>*>:3 [#uses=1] 11 @4 = external global <4 x float>, align 1 ; <<4 x float>*>:4 [#uses=1] 12 @5 = external global <4 x float>, align 1 ; <<4 x float>*>:5 [#uses=1] 13 @6 = external global <4 x float>, align 1 ; <<4 x float>*>:6 [#uses=1] 14 @7 = external global <4 x float>, align 1 ; <<4 x float>*>:7 [#uses=1] 15 @8 = external global <4 x float>, align 1 ; <<4 x float>*>:8 [#uses=1] 16 @9 = external global <4 x float>, align 1 ; <<4 x float>*>:9 [#uses=1] [all …]
|
D | large-gep-chain.ll | 4 %0 = type { i32, float* } 24 %tmp = getelementptr inbounds float, float* null, i64 1 25 %tmp3 = getelementptr inbounds float, float* %tmp, i64 1 26 %tmp4 = getelementptr inbounds float, float* %tmp3, i64 1 27 %tmp5 = getelementptr inbounds float, float* %tmp4, i64 1 28 %tmp6 = getelementptr inbounds float, float* %tmp5, i64 1 29 %tmp7 = getelementptr inbounds float, float* %tmp6, i64 1 30 %tmp8 = getelementptr inbounds float, float* %tmp7, i64 1 31 %tmp9 = getelementptr inbounds float, float* %tmp8, i64 1 32 %tmp10 = getelementptr inbounds float, float* %tmp9, i64 1 [all …]
|
/external/llvm-project/llvm/test/CodeGen/X86/ |
D | 2008-07-19-movups-spills.ll | 7 @0 = external dso_local global <4 x float>, align 1 ; <<4 x float>*>:0 [#uses=2] 8 @1 = external dso_local global <4 x float>, align 1 ; <<4 x float>*>:1 [#uses=1] 9 @2 = external dso_local global <4 x float>, align 1 ; <<4 x float>*>:2 [#uses=1] 10 @3 = external dso_local global <4 x float>, align 1 ; <<4 x float>*>:3 [#uses=1] 11 @4 = external dso_local global <4 x float>, align 1 ; <<4 x float>*>:4 [#uses=1] 12 @5 = external dso_local global <4 x float>, align 1 ; <<4 x float>*>:5 [#uses=1] 13 @6 = external dso_local global <4 x float>, align 1 ; <<4 x float>*>:6 [#uses=1] 14 @7 = external dso_local global <4 x float>, align 1 ; <<4 x float>*>:7 [#uses=1] 15 @8 = external dso_local global <4 x float>, align 1 ; <<4 x float>*>:8 [#uses=1] 16 @9 = external dso_local global <4 x float>, align 1 ; <<4 x float>*>:9 [#uses=1] [all …]
|
D | large-gep-chain.ll | 4 %0 = type { i32, float* } 24 %tmp = getelementptr inbounds float, float* null, i64 1 25 %tmp3 = getelementptr inbounds float, float* %tmp, i64 1 26 %tmp4 = getelementptr inbounds float, float* %tmp3, i64 1 27 %tmp5 = getelementptr inbounds float, float* %tmp4, i64 1 28 %tmp6 = getelementptr inbounds float, float* %tmp5, i64 1 29 %tmp7 = getelementptr inbounds float, float* %tmp6, i64 1 30 %tmp8 = getelementptr inbounds float, float* %tmp7, i64 1 31 %tmp9 = getelementptr inbounds float, float* %tmp8, i64 1 32 %tmp10 = getelementptr inbounds float, float* %tmp9, i64 1 [all …]
|
/external/llvm-project/llvm/test/Transforms/InstSimplify/ConstProp/AMDGPU/ |
D | cubetc.ll | 4 declare float @llvm.amdgcn.cubetc(float, float, float) 6 define void @test(float* %p) { 8 ; CHECK-NEXT: store volatile float -4.000000e+00, float* [[P:%.*]] 9 ; CHECK-NEXT: store volatile float 4.000000e+00, float* [[P]] 10 ; CHECK-NEXT: store volatile float -3.000000e+00, float* [[P]] 11 ; CHECK-NEXT: store volatile float 3.000000e+00, float* [[P]] 12 ; CHECK-NEXT: store volatile float -3.000000e+00, float* [[P]] 13 ; CHECK-NEXT: store volatile float -4.000000e+00, float* [[P]] 14 ; CHECK-NEXT: store volatile float -4.000000e+00, float* [[P]] 15 ; CHECK-NEXT: store volatile float -4.000000e+00, float* [[P]] [all …]
|
D | cubesc.ll | 4 declare float @llvm.amdgcn.cubesc(float, float, float) 6 define void @test(float* %p) { 8 ; CHECK-NEXT: store volatile float 3.000000e+00, float* [[P:%.*]] 9 ; CHECK-NEXT: store volatile float 3.000000e+00, float* [[P]] 10 ; CHECK-NEXT: store volatile float 4.000000e+00, float* [[P]] 11 ; CHECK-NEXT: store volatile float 4.000000e+00, float* [[P]] 12 ; CHECK-NEXT: store volatile float -4.000000e+00, float* [[P]] 13 ; CHECK-NEXT: store volatile float -3.000000e+00, float* [[P]] 14 ; CHECK-NEXT: store volatile float -3.000000e+00, float* [[P]] 15 ; CHECK-NEXT: store volatile float 3.000000e+00, float* [[P]] [all …]
|
D | cubema.ll | 4 declare float @llvm.amdgcn.cubema(float, float, float) 6 define void @test(float* %p) { 8 ; CHECK-NEXT: store volatile float 1.000000e+01, float* [[P:%.*]] 9 ; CHECK-NEXT: store volatile float 1.000000e+01, float* [[P]] 10 ; CHECK-NEXT: store volatile float 1.000000e+01, float* [[P]] 11 ; CHECK-NEXT: store volatile float 1.000000e+01, float* [[P]] 12 ; CHECK-NEXT: store volatile float 1.000000e+01, float* [[P]] 13 ; CHECK-NEXT: store volatile float 1.000000e+01, float* [[P]] 14 ; CHECK-NEXT: store volatile float -1.000000e+01, float* [[P]] 15 ; CHECK-NEXT: store volatile float 1.000000e+01, float* [[P]] [all …]
|
D | cubeid.ll | 4 declare float @llvm.amdgcn.cubeid(float, float, float) 6 define void @test(float* %p) { 8 ; CHECK-NEXT: store volatile float 4.000000e+00, float* [[P:%.*]] 9 ; CHECK-NEXT: store volatile float 2.000000e+00, float* [[P]] 10 ; CHECK-NEXT: store volatile float 4.000000e+00, float* [[P]] 11 ; CHECK-NEXT: store volatile float 2.000000e+00, float* [[P]] 12 ; CHECK-NEXT: store volatile float 0.000000e+00, float* [[P]] 13 ; CHECK-NEXT: store volatile float 0.000000e+00, float* [[P]] 14 ; CHECK-NEXT: store volatile float 5.000000e+00, float* [[P]] 15 ; CHECK-NEXT: store volatile float 2.000000e+00, float* [[P]] [all …]
|
/external/llvm-project/llvm/test/CodeGen/AMDGPU/ |
D | scratch-simple.ll | 80 define amdgpu_ps float @ps_main(i32 %idx) { 81 …float> <float undef, float undef, float undef, float undef, float undef, float undef, float undef,… 82 …float> <float undef, float undef, float undef, float undef, float undef, float undef, float undef,… 83 %r = fadd float %v1, %v2 84 ret float %r 131 define amdgpu_vs float @vs_main(i32 %idx) { 132 …float> <float undef, float undef, float undef, float undef, float undef, float undef, float undef,… 133 …float> <float undef, float undef, float undef, float undef, float undef, float undef, float undef,… 134 %r = fadd float %v1, %v2 135 ret float %r [all …]
|
D | schedule-regpressure-limit3.ll | 7 define amdgpu_kernel void @load_fma_store(float addrspace(3)* nocapture readonly %arg, float addrsp… 9 %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 1 10 %tmp2 = load float, float addrspace(3)* %tmp, align 4 11 %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 2 12 %tmp4 = load float, float addrspace(3)* %tmp3, align 4 13 %tmp5 = getelementptr inbounds float, float addrspace(3)* %arg, i32 3 14 %tmp6 = load float, float addrspace(3)* %tmp5, align 4 15 %tmp7 = tail call float @llvm.fmuladd.f32(float %tmp2, float %tmp4, float %tmp6) 16 %tmp8 = getelementptr inbounds float, float addrspace(3)* %arg, i32 5 17 %tmp9 = load float, float addrspace(3)* %tmp8, align 4 [all …]
|
D | schedule-regpressure-limit.ll | 8 define amdgpu_kernel void @load_fma_store(float addrspace(3)* nocapture readonly %arg, float addrsp… 10 %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 1 11 %tmp2 = load float, float addrspace(3)* %tmp, align 4 12 %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 2 13 %tmp4 = load float, float addrspace(3)* %tmp3, align 4 14 %tmp5 = getelementptr inbounds float, float addrspace(3)* %arg, i32 3 15 %tmp6 = load float, float addrspace(3)* %tmp5, align 4 16 %tmp7 = tail call float @llvm.fmuladd.f32(float %tmp2, float %tmp4, float %tmp6) 17 %tmp8 = getelementptr inbounds float, float addrspace(3)* %arg, i32 5 18 %tmp9 = load float, float addrspace(3)* %tmp8, align 4 [all …]
|
D | schedule-ilp.ll | 5 define amdgpu_kernel void @load_fma_store(float addrspace(3)* nocapture readonly %arg, float addrsp… 7 %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 1 8 %tmp2 = load float, float addrspace(3)* %tmp, align 4 9 %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 2 10 %tmp4 = load float, float addrspace(3)* %tmp3, align 4 11 %tmp5 = getelementptr inbounds float, float addrspace(3)* %arg, i32 3 12 %tmp6 = load float, float addrspace(3)* %tmp5, align 4 13 %tmp7 = tail call float @llvm.fmuladd.f32(float %tmp2, float %tmp4, float %tmp6) 14 %tmp8 = getelementptr inbounds float, float addrspace(3)* %arg, i32 5 15 %tmp9 = load float, float addrspace(3)* %tmp8, align 4 [all …]
|
D | schedule-regpressure-limit2.ll | 16 define amdgpu_kernel void @load_fma_store(float addrspace(3)* nocapture readonly %in_arg, float add… 18 %adr.a.0 = getelementptr inbounds float, float addrspace(3)* %in_arg, i32 20004 19 %adr.b.0 = getelementptr inbounds float, float addrspace(3)* %in_arg, i32 20252 20 %adr.c.0 = getelementptr inbounds float, float addrspace(3)* %in_arg, i32 20508 21 %adr.a.1 = getelementptr inbounds float, float addrspace(3)* %in_arg, i32 20772 22 %adr.b.1 = getelementptr inbounds float, float addrspace(3)* %in_arg, i32 21020 23 %adr.c.1 = getelementptr inbounds float, float addrspace(3)* %in_arg, i32 21276 24 %adr.a.2 = getelementptr inbounds float, float addrspace(3)* %in_arg, i32 21540 25 %adr.b.2 = getelementptr inbounds float, float addrspace(3)* %in_arg, i32 21788 26 %adr.c.2 = getelementptr inbounds float, float addrspace(3)* %in_arg, i32 22044 [all …]
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | pv.ll | 6 …float> inreg %reg0, <4 x float> inreg %reg1, <4 x float> inreg %reg2, <4 x float> inreg %reg3, <4 … 8 %0 = extractelement <4 x float> %reg1, i32 0 9 %1 = extractelement <4 x float> %reg1, i32 1 10 %2 = extractelement <4 x float> %reg1, i32 2 11 %3 = extractelement <4 x float> %reg1, i32 3 12 %4 = extractelement <4 x float> %reg2, i32 0 13 %5 = extractelement <4 x float> %reg2, i32 1 14 %6 = extractelement <4 x float> %reg2, i32 2 15 %7 = extractelement <4 x float> %reg2, i32 3 16 %8 = extractelement <4 x float> %reg3, i32 0 [all …]
|
D | local-stack-slot-bug.ll | 16 define amdgpu_ps float @main(i32 %idx) { 18 …float> <float undef, float undef, float undef, float undef, float undef, float undef, float undef,… 19 …float> <float undef, float undef, float undef, float undef, float undef, float undef, float undef,… 20 %r = fadd float %v1, %v2 21 ret float %r
|
/external/llvm/test/Transforms/InstCombine/ |
D | x86-sse.ll | 5 define float @test_rcp_ss_0(float %a) { 7 ; CHECK-NEXT: [[TMP1:%.*]] = insertelement <4 x float> undef, float %a, i32 0 8 ; CHECK-NEXT: [[TMP2:%.*]] = tail call <4 x float> @llvm.x86.sse.rcp.ss(<4 x float> [[TMP1]]) 9 ; CHECK-NEXT: [[TMP3:%.*]] = extractelement <4 x float> [[TMP2]], i32 0 10 ; CHECK-NEXT: ret float [[TMP3]] 12 %1 = insertelement <4 x float> undef, float %a, i32 0 13 %2 = insertelement <4 x float> %1, float 1.000000e+00, i32 1 14 %3 = insertelement <4 x float> %2, float 2.000000e+00, i32 2 15 %4 = insertelement <4 x float> %3, float 3.000000e+00, i32 3 16 %5 = tail call <4 x float> @llvm.x86.sse.rcp.ss(<4 x float> %4) [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | 2009-11-13-ScavengerAssert2.ll | 4 %bar = type { %quad, float, float, [3 x %quuz*], [3 x %bar*], [2 x %bar*], [3 x i8], i8 } 6 %foo = type { i8, %quux, %quad, float, [64 x %quuz], [128 x %bar], i32, %baz, %baz } 7 %quad = type { [4 x float] } 8 %quux = type { [4 x %quuz*], [4 x float], i32 } 22 %0 = getelementptr inbounds %quuz, %quuz* %a, i32 0, i32 1, i32 0, i32 0 ; <float*> [#uses=0] 23 %1 = fsub float 0.000000e+00, undef ; <float> [#uses=1] 24 %2 = getelementptr inbounds %quuz, %quuz* %b, i32 0, i32 1, i32 0, i32 1 ; <float*> [#uses=2] 25 %3 = load float, float* %2, align 4 ; <float> [#uses=1] 26 %4 = getelementptr inbounds %quuz, %quuz* %a, i32 0, i32 1, i32 0, i32 1 ; <float*> [#uses=1] 27 %5 = fsub float %3, undef ; <float> [#uses=2] [all …]
|
/external/llvm-project/llvm/test/CodeGen/ARM/ |
D | 2009-11-13-ScavengerAssert2.ll | 4 %bar = type { %quad, float, float, [3 x %quuz*], [3 x %bar*], [2 x %bar*], [3 x i8], i8 } 6 %foo = type { i8, %quux, %quad, float, [64 x %quuz], [128 x %bar], i32, %baz, %baz } 7 %quad = type { [4 x float] } 8 %quux = type { [4 x %quuz*], [4 x float], i32 } 22 %0 = getelementptr inbounds %quuz, %quuz* %a, i32 0, i32 1, i32 0, i32 0 ; <float*> [#uses=0] 23 %1 = fsub float 0.000000e+00, undef ; <float> [#uses=1] 24 %2 = getelementptr inbounds %quuz, %quuz* %b, i32 0, i32 1, i32 0, i32 1 ; <float*> [#uses=2] 25 %3 = load float, float* %2, align 4 ; <float> [#uses=1] 26 %4 = getelementptr inbounds %quuz, %quuz* %a, i32 0, i32 1, i32 0, i32 1 ; <float*> [#uses=1] 27 %5 = fsub float %3, undef ; <float> [#uses=2] [all …]
|
D | 2009-11-07-SubRegAsmPrinting.ll | 9 %0 = load float, float* null, align 4 ; <float> [#uses=2] 10 %1 = fmul float %0, %0 ; <float> [#uses=2] 11 %2 = fmul float 0.000000e+00, %1 ; <float> [#uses=2] 12 %3 = fmul float %0, %1 ; <float> [#uses=1] 13 %4 = fadd float 0.000000e+00, %3 ; <float> [#uses=1] 14 %5 = fsub float 1.000000e+00, %4 ; <float> [#uses=1] 17 %6 = fsub float 1.000000e+00, %2 ; <float> [#uses=2] 18 %7 = fsub float %2, %2 ; <float> [#uses=1] 19 %8 = fsub float 0.000000e+00, %7 ; <float> [#uses=3] 20 %9 = fadd float %2, %2 ; <float> [#uses=3] [all …]
|
/external/llvm-project/llvm/test/Transforms/InstCombine/X86/ |
D | x86-sse.ll | 5 define float @test_rcp_ss_0(float %a) { 7 ; CHECK-NEXT: [[TMP1:%.*]] = insertelement <4 x float> undef, float [[A:%.*]], i32 0 8 ; CHECK-NEXT: [[TMP2:%.*]] = tail call <4 x float> @llvm.x86.sse.rcp.ss(<4 x float> [[TMP1]]) 9 ; CHECK-NEXT: [[TMP3:%.*]] = extractelement <4 x float> [[TMP2]], i32 0 10 ; CHECK-NEXT: ret float [[TMP3]] 12 %1 = insertelement <4 x float> undef, float %a, i32 0 13 %2 = insertelement <4 x float> %1, float 1.000000e+00, i32 1 14 %3 = insertelement <4 x float> %2, float 2.000000e+00, i32 2 15 %4 = insertelement <4 x float> %3, float 3.000000e+00, i32 3 16 %5 = tail call <4 x float> @llvm.x86.sse.rcp.ss(<4 x float> %4) [all …]
|
/external/llvm/test/CodeGen/Generic/ |
D | 2003-05-28-ManyArgs.ll | 21 %struct..s_annealing_sched = type { i32, float, float, float, float } 22 %struct..s_chan = type { i32, float, float, float, float } 23 …%struct..s_det_routing_arch = type { i32, float, float, float, i32, i32, i16, i16, i16, float, flo… 24 %struct..s_placer_opts = type { i32, float, i32, i32, i8*, i32, i32 } 25 %struct..s_router_opts = type { float, float, float, float, float, i32, i32, i32, i32 } 26 %struct..s_segment_inf = type { float, i32, i16, i16, float, float, i32, float, float } 27 %struct..s_switch_inf = type { i32, float, float, float, float } 44 … float, float, float, float, float, float, float, float, float, float } ; <{ i32, float, float, f… 50 …2, float, float, float, float, float, float, float, float, float, float }, { i32, float, float, fl… 56 …tr %struct..s_placer_opts, %struct..s_placer_opts* %placer_opts, i64 0, i32 1 ; <float*> [#uses=1] [all …]
|
/external/llvm-project/llvm/test/CodeGen/Generic/ |
D | 2003-05-28-ManyArgs.ll | 21 %struct..s_annealing_sched = type { i32, float, float, float, float } 22 %struct..s_chan = type { i32, float, float, float, float } 23 …%struct..s_det_routing_arch = type { i32, float, float, float, i32, i32, i16, i16, i16, float, flo… 24 %struct..s_placer_opts = type { i32, float, i32, i32, i8*, i32, i32 } 25 %struct..s_router_opts = type { float, float, float, float, float, i32, i32, i32, i32 } 26 %struct..s_segment_inf = type { float, i32, i16, i16, float, float, i32, float, float } 27 %struct..s_switch_inf = type { i32, float, float, float, float } 44 … float, float, float, float, float, float, float, float, float, float } ; <{ i32, float, float, f… 50 …2, float, float, float, float, float, float, float, float, float, float }, { i32, float, float, fl… 56 …tr %struct..s_placer_opts, %struct..s_placer_opts* %placer_opts, i64 0, i32 1 ; <float*> [#uses=1] [all …]
|
/external/llvm-project/llvm/test/Transforms/SLPVectorizer/X86/ |
D | horizontal-list.ll | 6 @arr = common local_unnamed_addr global [20 x float] zeroinitializer, align 16 7 @arr1 = common local_unnamed_addr global [20 x float] zeroinitializer, align 16 8 @res = external local_unnamed_addr global float, align 4 10 define float @baz() { 15 ; CHECK-NEXT: [[CONV:%.*]] = sitofp i32 [[MUL]] to float 16 ; CHECK-NEXT: [[TMP1:%.*]] = load <2 x float>, <2 x float>* bitcast ([20 x float]* @arr to <2 x … 17 ; CHECK-NEXT: [[TMP2:%.*]] = load <2 x float>, <2 x float>* bitcast ([20 x float]* @arr1 to <2 x… 18 ; CHECK-NEXT: [[TMP3:%.*]] = fmul fast <2 x float> [[TMP2]], [[TMP1]] 19 ; CHECK-NEXT: [[TMP4:%.*]] = extractelement <2 x float> [[TMP3]], i32 0 20 ; CHECK-NEXT: [[ADD:%.*]] = fadd fast float [[TMP4]], [[CONV]] [all …]
|
/external/llvm-project/llvm/test/Transforms/InstSimplify/ |
D | fminmax-folds.ll | 4 declare float @llvm.minnum.f32(float, float) 5 declare float @llvm.maxnum.f32(float, float) 6 declare float @llvm.minimum.f32(float, float) 7 declare float @llvm.maximum.f32(float, float) 8 declare <2 x float> @llvm.minnum.v2f32(<2 x float>, <2 x float>) 9 declare <2 x float> @llvm.maxnum.v2f32(<2 x float>, <2 x float>) 10 declare <2 x float> @llvm.minimum.v2f32(<2 x float>, <2 x float>) 11 declare <2 x float> @llvm.maximum.v2f32(<2 x float>, <2 x float>) 22 define float @test_minnum_const_nan(float %x) { 24 ; CHECK-NEXT: ret float [[X:%.*]] [all …]
|
/external/llvm-project/llvm/test/CodeGen/Hexagon/ |
D | large-number-of-preds.ll | 6 @g0 = external global void (float*, i32, i32, float*, float*)** 9 define void @f0(float* nocapture %a0, float* nocapture %a1, float* %a2) #0 { 11 %v0 = alloca [64 x float], align 16 12 %v1 = alloca [8 x float], align 8 13 %v2 = bitcast [64 x float]* %v0 to i8* 15 %v3 = load float, float* %a0, align 4, !tbaa !0 16 %v4 = getelementptr inbounds [64 x float], [64 x float]* %v0, i32 0, i32 35 17 store float %v3, float* %v4, align 4, !tbaa !0 18 %v5 = getelementptr inbounds [64 x float], [64 x float]* %v0, i32 0, i32 0 19 store float %v3, float* %v5, align 16, !tbaa !0 [all …]
|