/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/StructurizeCFG/ |
D | nested-loop-order.ll | 14 %temp4.0.ph = phi i32 [ 0, %main_body ], [ %tmp20, %ENDIF28 ] 20 %temp4.0 = phi i32 [ %temp4.0.ph, %LOOP.outer ], [ %tmp20, %IF29 ] 21 %tmp20 = add i32 %temp4.0, 1 22 %tmp22 = icmp sgt i32 %tmp20, 3 32 %tmp23 = icmp eq i32 %tmp20, 3 40 %tmp31 = icmp sgt i32 %tmp20, 1 49 %tmp32 = icmp sgt i32 %tmp20, 2 62 %tmp36 = icmp sgt i32 %tmp20, 2
|
/external/llvm/test/Transforms/StructurizeCFG/ |
D | nested-loop-order.ll | 14 %temp4.0.ph = phi i32 [ 0, %main_body ], [ %tmp20, %ENDIF28 ] 20 %temp4.0 = phi i32 [ %temp4.0.ph, %LOOP.outer ], [ %tmp20, %IF29 ] 21 %tmp20 = add i32 %temp4.0, 1 22 %tmp22 = icmp sgt i32 %tmp20, 3 32 %tmp23 = icmp eq i32 %tmp20, 3 40 %tmp31 = icmp sgt i32 %tmp20, 1 49 %tmp32 = icmp sgt i32 %tmp20, 2 62 %tmp36 = icmp sgt i32 %tmp20, 2
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Analysis/ScalarEvolution/ |
D | truncate.ll | 43 %tmp20 = add i64 %tmp19, %tmp18 44 %tmp21 = trunc i64 %tmp20 to i32 57 %tmp28 = sub i64 %tmp27, %tmp20 82 ; CHECK: %tmp27 = add i64 %tmp20, -1 83 ; CHECK-NOT: (-1 + %tmp20)<nuw> 84 ; CHECK-NEXT: --> (-1 + %tmp20) U: 91 %tmp20 = phi i64 [ %shift, %bb ], [ 0, %bb36 ] 96 %tmp27 = add i64 %tmp20, -1
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/ |
D | 2008-02-27-DeadSlotElimBug.ll | 39 %tmp20.i39 = fadd double %tmp17.i, %tmp17.i63 ; <double> [#uses=1] 40 %tmp20.i23 = fadd double %tmp20.i39, %tmp17.i76 ; <double> [#uses=1] 48 %tmp20.i7 = getelementptr %struct.CompAtom, %struct.CompAtom* %d, i32 0, i32 2 ; <i32*> [#uses=2] 50 %tmp74.i = load i32, i32* %tmp20.i7, align 1 ; <i32> [#uses=1] 54 %tmp88.i = load i32, i32* %tmp20.i7, align 1 ; <i32> [#uses=1] 58 %tmp6.i = fadd double 0.000000e+00, %tmp20.i23 ; <double> [#uses=0]
|
D | fp-stack-compare-cmov.ll | 9 %tmp20 = fsub float -0.000000e+00, %tmp 10 %iftmp.2.0 = select i1 %tmp16, float %tmp20, float %tmp
|
D | fp-stack-compare.ll | 11 %tmp20 = fsub float -0.000000e+00, %tmp 12 %iftmp.2.0 = select i1 %tmp16, float %tmp20, float %tmp
|
/external/llvm/test/CodeGen/X86/ |
D | 2008-02-27-DeadSlotElimBug.ll | 39 %tmp20.i39 = fadd double %tmp17.i, %tmp17.i63 ; <double> [#uses=1] 40 %tmp20.i23 = fadd double %tmp20.i39, %tmp17.i76 ; <double> [#uses=1] 48 %tmp20.i7 = getelementptr %struct.CompAtom, %struct.CompAtom* %d, i32 0, i32 2 ; <i32*> [#uses=2] 50 %tmp74.i = load i32, i32* %tmp20.i7, align 1 ; <i32> [#uses=1] 54 %tmp88.i = load i32, i32* %tmp20.i7, align 1 ; <i32> [#uses=1] 58 %tmp6.i = fadd double 0.000000e+00, %tmp20.i23 ; <double> [#uses=0]
|
D | fp-stack-compare-cmov.ll | 9 %tmp20 = fsub float -0.000000e+00, %tmp 10 %iftmp.2.0 = select i1 %tmp16, float %tmp20, float %tmp
|
D | fp-stack-compare.ll | 11 %tmp20 = fsub float -0.000000e+00, %tmp 12 %iftmp.2.0 = select i1 %tmp16, float %tmp20, float %tmp
|
/external/swiftshader/third_party/LLVM/test/CodeGen/X86/ |
D | 2008-02-27-DeadSlotElimBug.ll | 39 %tmp20.i39 = fadd double %tmp17.i, %tmp17.i63 ; <double> [#uses=1] 40 %tmp20.i23 = fadd double %tmp20.i39, %tmp17.i76 ; <double> [#uses=1] 48 %tmp20.i7 = getelementptr %struct.CompAtom* %d, i32 0, i32 2 ; <i32*> [#uses=2] 50 %tmp74.i = load i32* %tmp20.i7, align 1 ; <i32> [#uses=1] 54 %tmp88.i = load i32* %tmp20.i7, align 1 ; <i32> [#uses=1] 58 %tmp6.i = fadd double 0.000000e+00, %tmp20.i23 ; <double> [#uses=0]
|
D | fp-stack-compare.ll | 8 %tmp20 = fsub float -0.000000e+00, %tmp 9 %iftmp.2.0 = select i1 %tmp16, float %tmp20, float %tmp
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Analysis/IVUsers/ |
D | deep_recursion_in_scev.ll | 27 %tmp20 = load i32, i32 addrspace(1)* %tmp19, align 8, !range !0 49 %tmp36 = icmp ult i32 %tmp35, %tmp20 80 %tmp55 = icmp ult i32 %tmp53, %tmp20 95 %tmp65 = icmp ult i32 %tmp63, %tmp20 107 %tmp73 = icmp ult i32 %tmp71, %tmp20 119 %tmp81 = icmp ult i32 %tmp79, %tmp20 131 %tmp89 = icmp ult i32 %tmp87, %tmp20 143 %tmp97 = icmp ult i32 %tmp95, %tmp20 155 %tmp105 = icmp ult i32 %tmp103, %tmp20 167 %tmp113 = icmp ult i32 %tmp111, %tmp20
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | sgpr-copy.ll | 17 %tmp20 = load <16 x i8>, <16 x i8> addrspace(2)* %tmp, !tbaa !0 18 %tmp21 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 0) 19 %tmp22 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 16) 20 %tmp23 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 32) 41 %tmp20 = load <16 x i8>, <16 x i8> addrspace(2)* %tmp, !tbaa !0 42 %tmp21 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 16) 43 %tmp22 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 32) 44 %tmp23 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 36) 45 %tmp24 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 40) 46 %tmp25 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 48) [all …]
|
/external/libjpeg-turbo/ |
D | jidctint.c | 1078 JLONG tmp20, tmp21, tmp22, tmp23, tmp24; in jpeg_idct_10x10() local 1117 tmp20 = tmp10 + tmp12; in jpeg_idct_10x10() 1151 wsptr[8 * 0] = (int)RIGHT_SHIFT(tmp20 + tmp10, CONST_BITS - PASS1_BITS); in jpeg_idct_10x10() 1152 wsptr[8 * 9] = (int)RIGHT_SHIFT(tmp20 - tmp10, CONST_BITS - PASS1_BITS); in jpeg_idct_10x10() 1189 tmp20 = tmp10 + tmp12; in jpeg_idct_10x10() 1223 outptr[0] = range_limit[(int)RIGHT_SHIFT(tmp20 + tmp10, in jpeg_idct_10x10() 1226 outptr[9] = range_limit[(int)RIGHT_SHIFT(tmp20 - tmp10, in jpeg_idct_10x10() 1273 JLONG tmp20, tmp21, tmp22, tmp23, tmp24, tmp25; in jpeg_idct_11x11() local 1301 tmp20 = MULTIPLY(z2 - z3, FIX(2.546640132)); /* c2+c4 */ in jpeg_idct_11x11() 1307 tmp21 = tmp20 + tmp23 + tmp25 - in jpeg_idct_11x11() [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/GVNHoist/ |
D | pr35222-hoist-load.ll | 22 %tmp20 = load i32, i32* @heap, align 4 23 ret i32 %tmp20 65 %tmp20 = load i32, i32* getelementptr inbounds ([573 x i32], [573 x i32]* @j, 67 ret i32 %tmp20
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AMDGPU/ |
D | sgpr-copy.ll | 10 %tmp20 = load <4 x i32>, <4 x i32> addrspace(4)* %tmp, !tbaa !0 11 %tmp21 = call float @llvm.SI.load.const.v4i32(<4 x i32> %tmp20, i32 0) 12 %tmp22 = call float @llvm.SI.load.const.v4i32(<4 x i32> %tmp20, i32 16) 13 %tmp23 = call float @llvm.SI.load.const.v4i32(<4 x i32> %tmp20, i32 32) 34 %tmp20 = load <4 x i32>, <4 x i32> addrspace(4)* %tmp, !tbaa !0 35 %tmp21 = call float @llvm.SI.load.const.v4i32(<4 x i32> %tmp20, i32 16) 36 %tmp22 = call float @llvm.SI.load.const.v4i32(<4 x i32> %tmp20, i32 32) 37 %tmp23 = call float @llvm.SI.load.const.v4i32(<4 x i32> %tmp20, i32 36) 38 %tmp24 = call float @llvm.SI.load.const.v4i32(<4 x i32> %tmp20, i32 40) 39 %tmp25 = call float @llvm.SI.load.const.v4i32(<4 x i32> %tmp20, i32 48) [all …]
|
D | waitcnt-looptest.ll | 47 %tmp20 = phi i32 [ 0, %bb ], [ %tmp100, %bb18 ] 55 %tmp28 = or i32 %tmp20, 1 63 %tmp36 = add nuw nsw i32 %tmp20, 2 71 %tmp44 = add nuw nsw i32 %tmp20, 3 79 %tmp52 = add nuw nsw i32 %tmp20, 4 87 %tmp60 = add nuw nsw i32 %tmp20, 5 95 %tmp68 = add nuw nsw i32 %tmp20, 6 103 %tmp76 = add nuw nsw i32 %tmp20, 7 111 %tmp84 = add nuw nsw i32 %tmp20, 8 119 %tmp92 = add nuw nsw i32 %tmp20, 9 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/InstCombine/ |
D | gepphigep.ll | 21 %tmp20 = getelementptr inbounds %struct2, %struct2* %tmp1, i64 %tmp19 22 %tmp21 = getelementptr inbounds %struct2, %struct2* %tmp20, i64 0, i32 0 27 %phi = phi %struct2* [ %tmp10, %bb1 ], [ %tmp20, %bb2 ] 47 %tmp20 = getelementptr inbounds %struct2, %struct2* %tmp1, i64 %tmp19 48 %tmp21 = getelementptr inbounds %struct2, %struct2* %tmp20, i64 0, i32 0 62 define i32 @test3(%struct3* %dm, i1 %tmp4, i64 %tmp9, i64 %tmp19, i64 %tmp20, i64 %tmp21) personali… 74 %tmp2 = getelementptr inbounds %struct3, %struct3* %tmp, i64 %tmp20, i32 1
|
D | 2007-05-10-icmp-or.ll | 3 %tmp20 = icmp ugt i32 %tmp9, 255 ; <i1> [#uses=1] 5 %bothcond = or i1 %tmp20, %tmp11.not ; <i1> [#uses=1]
|
/external/llvm/test/Transforms/InstCombine/ |
D | gepphigep.ll | 21 %tmp20 = getelementptr inbounds %struct2, %struct2* %tmp1, i64 %tmp19 22 %tmp21 = getelementptr inbounds %struct2, %struct2* %tmp20, i64 0, i32 0 27 %phi = phi %struct2* [ %tmp10, %bb1 ], [ %tmp20, %bb2 ] 47 %tmp20 = getelementptr inbounds %struct2, %struct2* %tmp1, i64 %tmp19 48 %tmp21 = getelementptr inbounds %struct2, %struct2* %tmp20, i64 0, i32 0 62 define i32 @test3(%struct3* %dm, i1 %tmp4, i64 %tmp9, i64 %tmp19, i64 %tmp20, i64 %tmp21) personali… 74 %tmp2 = getelementptr inbounds %struct3, %struct3* %tmp, i64 %tmp20, i32 1
|
D | 2007-05-10-icmp-or.ll | 3 %tmp20 = icmp ugt i32 %tmp9, 255 ; <i1> [#uses=1] 5 %bothcond = or i1 %tmp20, %tmp11.not ; <i1> [#uses=1]
|
/external/swiftshader/third_party/LLVM/test/Transforms/InstCombine/ |
D | 2007-05-10-icmp-or.ll | 3 %tmp20 = icmp ugt i32 %tmp9, 255 ; <i1> [#uses=1] 5 %bothcond = or i1 %tmp20, %tmp11.not ; <i1> [#uses=1]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/Reassociate/ |
D | long-chains.ll | 24 %tmp20 = add i8 %tmp19, %in19 25 %tmp21 = add i8 %tmp20, %in20 26 ret i8 %tmp20
|
/external/llvm/test/Transforms/SeparateConstOffsetFromGEP/AMDGPU/ |
D | split-gep-and-gvn-addrspace-addressing-modes.ll | 29 %tmp20 = load float, float addrspace(2)* %tmp18, align 4 30 %tmp21 = fadd float %tmp17, %tmp20 62 %tmp20 = load float, float addrspace(2)* %tmp18, align 4 63 %tmp21 = fadd float %tmp17, %tmp20 90 %tmp20 = load float, float addrspace(3)* %tmp18, align 4 91 %tmp21 = fadd float %tmp17, %tmp20
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/CorrelatedValuePropagation/ |
D | pr35807.ll | 51 %tmp20 = select i1 %tmp19, i64 %tmp20, i64 0 64 %tmp41 = phi i64 [ 4, %bb4 ], [ %tmp20, %bb14 ], [ undef, %bb32 ]
|