Home
last modified time | relevance | path

Searched refs:tmp20 (Results 1 – 25 of 101) sorted by relevance

12345

/external/llvm/test/Transforms/StructurizeCFG/
Dnested-loop-order.ll14 %temp4.0.ph = phi i32 [ 0, %main_body ], [ %tmp20, %ENDIF28 ]
20 %temp4.0 = phi i32 [ %temp4.0.ph, %LOOP.outer ], [ %tmp20, %IF29 ]
21 %tmp20 = add i32 %temp4.0, 1
22 %tmp22 = icmp sgt i32 %tmp20, 3
32 %tmp23 = icmp eq i32 %tmp20, 3
40 %tmp31 = icmp sgt i32 %tmp20, 1
49 %tmp32 = icmp sgt i32 %tmp20, 2
62 %tmp36 = icmp sgt i32 %tmp20, 2
/external/llvm/test/CodeGen/X86/
D2008-02-27-DeadSlotElimBug.ll39 %tmp20.i39 = fadd double %tmp17.i, %tmp17.i63 ; <double> [#uses=1]
40 %tmp20.i23 = fadd double %tmp20.i39, %tmp17.i76 ; <double> [#uses=1]
48 %tmp20.i7 = getelementptr %struct.CompAtom, %struct.CompAtom* %d, i32 0, i32 2 ; <i32*> [#uses=2]
50 %tmp74.i = load i32, i32* %tmp20.i7, align 1 ; <i32> [#uses=1]
54 %tmp88.i = load i32, i32* %tmp20.i7, align 1 ; <i32> [#uses=1]
58 %tmp6.i = fadd double 0.000000e+00, %tmp20.i23 ; <double> [#uses=0]
Dfp-stack-compare-cmov.ll9 %tmp20 = fsub float -0.000000e+00, %tmp
10 %iftmp.2.0 = select i1 %tmp16, float %tmp20, float %tmp
Dfp-stack-compare.ll11 %tmp20 = fsub float -0.000000e+00, %tmp
12 %iftmp.2.0 = select i1 %tmp16, float %tmp20, float %tmp
Dtailcall-fastisel.ll7 %tmp20 = tail call fastcc i8* @"visit_array_aux<`Reference>"(%0 %arg, i32 undef) ; <i8*> [#uses=1]
9 ret i8* %tmp20
Dphi-immediate-factoring.ll29 %i.248.1 = phi i32 [ 0, %entry ], [ %tmp20, %cond_next18 ] ; <i32> [#uses=2]
35 %tmp20 = add i32 %i.248.1, 1 ; <i32> [#uses=2]
36 icmp slt i32 %tmp20, 1000 ; <i1>:3 [#uses=1]
Dpr2326.ll12 %tmp20 = load i32, i32* %l_108, align 4 ; <i32> [#uses=1]
13 %tmp21 = icmp ule i32 %tmp19, %tmp20 ; <i1> [#uses=1]
Dcse-add-with-overflow.ll39 %tmp20 = or i64 %tmp19, -281474976710656
40 ret i64 %tmp20
D2008-02-06-LoadFoldingBug.ll10 %tmp20 = load double, double* %tmp19, align 8 ; <double> [#uses=1]
13 %tmp32 = fsub double -0.000000e+00, %tmp20 ; <double> [#uses=1]
Ddagcombine-cse.ll12 %tmp20.sum = add i32 %tmp18, %tmp7 ; <i32> [#uses=1]
13 %tmp21 = getelementptr i8, i8* %ref_frame_ptr, i32 %tmp20.sum ; <i8*> [#uses=1]
Dnarrow-shl-load.ll23 %tmp20 = load i64, i64* undef, align 8
24 %add21 = add i64 %xor, %tmp20
D2011-06-12-FastAllocSpill.ll38 %tmp20 = bitcast %6* %tmp18 to i8*
41 call void %tmp22(i8* %tmp20)
/external/llvm/test/CodeGen/AMDGPU/
Dsgpr-copy.ll17 %tmp20 = load <16 x i8>, <16 x i8> addrspace(2)* %tmp, !tbaa !0
18 %tmp21 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 0)
19 %tmp22 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 16)
20 %tmp23 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 32)
41 %tmp20 = load <16 x i8>, <16 x i8> addrspace(2)* %tmp, !tbaa !0
42 %tmp21 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 16)
43 %tmp22 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 32)
44 %tmp23 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 36)
45 %tmp24 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 40)
46 %tmp25 = call float @llvm.SI.load.const(<16 x i8> %tmp20, i32 48)
[all …]
/external/libjpeg-turbo/
Djidctint.c1078 JLONG tmp20, tmp21, tmp22, tmp23, tmp24; in jpeg_idct_10x10() local
1117 tmp20 = tmp10 + tmp12; in jpeg_idct_10x10()
1151 wsptr[8 * 0] = (int)RIGHT_SHIFT(tmp20 + tmp10, CONST_BITS - PASS1_BITS); in jpeg_idct_10x10()
1152 wsptr[8 * 9] = (int)RIGHT_SHIFT(tmp20 - tmp10, CONST_BITS - PASS1_BITS); in jpeg_idct_10x10()
1189 tmp20 = tmp10 + tmp12; in jpeg_idct_10x10()
1223 outptr[0] = range_limit[(int)RIGHT_SHIFT(tmp20 + tmp10, in jpeg_idct_10x10()
1226 outptr[9] = range_limit[(int)RIGHT_SHIFT(tmp20 - tmp10, in jpeg_idct_10x10()
1273 JLONG tmp20, tmp21, tmp22, tmp23, tmp24, tmp25; in jpeg_idct_11x11() local
1301 tmp20 = MULTIPLY(z2 - z3, FIX(2.546640132)); /* c2+c4 */ in jpeg_idct_11x11()
1307 tmp21 = tmp20 + tmp23 + tmp25 - in jpeg_idct_11x11()
[all …]
/external/llvm/test/Transforms/InstCombine/
D2007-05-10-icmp-or.ll3 %tmp20 = icmp ugt i32 %tmp9, 255 ; <i1> [#uses=1]
5 %bothcond = or i1 %tmp20, %tmp11.not ; <i1> [#uses=1]
Dgepphigep.ll21 %tmp20 = getelementptr inbounds %struct2, %struct2* %tmp1, i64 %tmp19
22 %tmp21 = getelementptr inbounds %struct2, %struct2* %tmp20, i64 0, i32 0
27 %phi = phi %struct2* [ %tmp10, %bb1 ], [ %tmp20, %bb2 ]
47 %tmp20 = getelementptr inbounds %struct2, %struct2* %tmp1, i64 %tmp19
48 %tmp21 = getelementptr inbounds %struct2, %struct2* %tmp20, i64 0, i32 0
62 define i32 @test3(%struct3* %dm, i1 %tmp4, i64 %tmp9, i64 %tmp19, i64 %tmp20, i64 %tmp21) personali…
74 %tmp2 = getelementptr inbounds %struct3, %struct3* %tmp, i64 %tmp20, i32 1
D2009-02-25-CrashZeroSizeArray.ll21 %tmp20 = bitcast [0 x i32]* %tmp19 to i32* ; <i32*> [#uses=1]
22 %tmp21 = getelementptr i32, i32* %tmp20, i32 0 ; <i32*> [#uses=1]
/external/llvm/test/Transforms/SeparateConstOffsetFromGEP/AMDGPU/
Dsplit-gep-and-gvn-addrspace-addressing-modes.ll29 %tmp20 = load float, float addrspace(2)* %tmp18, align 4
30 %tmp21 = fadd float %tmp17, %tmp20
62 %tmp20 = load float, float addrspace(2)* %tmp18, align 4
63 %tmp21 = fadd float %tmp17, %tmp20
90 %tmp20 = load float, float addrspace(3)* %tmp18, align 4
91 %tmp21 = fadd float %tmp17, %tmp20
/external/llvm/test/Bitcode/
Dmetadata-2.ll42 %tmp20 = add i32 %tmp18, %tmp16 ; <i32> [#uses=1]
43 %tmp22 = and i32 %tmp20, 16711935 ; <i32> [#uses=2]
66 %tmp20 = or i32 %tmp16, %tmp19 ; <i32> [#uses=2]
67 %tmp22 = lshr i32 %tmp20, 8 ; <i32> [#uses=1]
69 %tmp25 = shl i32 %tmp20, 8 ; <i32> [#uses=1]
/external/llvm/test/CodeGen/ARM/
D2010-06-11-vmovdrr-bitcast.ll14 %tmp20.i = bitcast i64 %tmp9 to <8 x i8> ; <<8 x i8>> [#uses=1]
15 …tail call void @llvm.arm.neon.vst2.p0i8.v8i8(i8* %b, <8 x i8> %tmp16.i, <8 x i8> %tmp20.i, i32 1) …
D2011-08-29-SchedCycle.ll40 %tmp20 = urem i64 %tmp19, 3
43 store i64 %tmp20, i64* undef, align 4
/external/llvm/test/Transforms/SROA/
Dvector-lifetime-intrinsic.ll22 %tmp20 = bitcast <4 x float>* %tmp to i8*
23 call void @llvm.lifetime.end(i64 16, i8* %tmp20)
/external/llvm/test/Transforms/IndVarSimplify/
Dlrev-existing-umin.ll17 %tmp20 = or i32 %tmp19, %tmp10
18 %tmp21 = trunc i32 %tmp20 to i8
/external/llvm/test/CodeGen/PowerPC/
Dpr28130.ll28 %tmp20 = load double, double* %tmp9, align 16
29 %tmp21 = fsub double 1.210000e+04, %tmp20
31 %tmp23 = fadd double %tmp20, %tmp22
/external/llvm/test/Analysis/ScalarEvolution/
D2008-06-12-BinomialInt64.ll15 %tmp20 = icmp slt i32 %tmp16, 0 ; <i1> [#uses=1]
16 br i1 %tmp20, label %bb10, label %bb23

12345