Home
last modified time | relevance | path

Searched refs:tmp21 (Results 1 – 25 of 178) sorted by relevance

12345678

/external/swiftshader/third_party/LLVM/test/Transforms/InstCombine/
Dapint-cast.ll8 %tmp21 = lshr i37 %tmp, 8 ; <i37> [#uses=1]
9 ; CHECK: %tmp21 = lshr i17 %a, 8
12 %tmp.upgrd.32 = or i37 %tmp21, %tmp5 ; <i37> [#uses=1]
13 ; CHECK: %tmp.upgrd.32 = or i17 %tmp21, %tmp5
21 %tmp21 = lshr i577 %tmp, 9 ; <i577> [#uses=1]
22 ; CHECK: %tmp21 = lshr i167 %a, 9
25 %tmp.upgrd.32 = or i577 %tmp21, %tmp5 ; <i577> [#uses=1]
26 ; CHECK: %tmp.upgrd.32 = or i167 %tmp21, %tmp5
D2008-01-21-MulTrunc.ll7 %tmp21 = lshr i32 %tmp, 8 ; <i32> [#uses=1]
8 ; CHECK: %tmp21 = lshr i16 %a, 8
11 %tmp.upgrd.32 = or i32 %tmp21, %tmp5 ; <i32> [#uses=1]
12 ; CHECK: %tmp.upgrd.32 = or i16 %tmp21, %tmp5
/external/llvm/test/Transforms/InstCombine/
Dapint-cast.ll8 %tmp21 = lshr i37 %tmp, 8 ; <i37> [#uses=1]
9 ; CHECK: %tmp21 = lshr i17 %a, 8
12 %tmp.upgrd.32 = or i37 %tmp21, %tmp5 ; <i37> [#uses=1]
13 ; CHECK: %tmp.upgrd.32 = or i17 %tmp21, %tmp5
21 %tmp21 = lshr i577 %tmp, 9 ; <i577> [#uses=1]
22 ; CHECK: %tmp21 = lshr i167 %a, 9
25 %tmp.upgrd.32 = or i577 %tmp21, %tmp5 ; <i577> [#uses=1]
26 ; CHECK: %tmp.upgrd.32 = or i167 %tmp21, %tmp5
D2008-01-21-MulTrunc.ll7 %tmp21 = lshr i32 %tmp, 8 ; <i32> [#uses=1]
8 ; CHECK: %tmp21 = lshr i16 %a, 8
11 %tmp.upgrd.32 = or i32 %tmp21, %tmp5 ; <i32> [#uses=1]
12 ; CHECK: %tmp.upgrd.32 = or i16 %tmp21, %tmp5
Dgepphigep.ll22 %tmp21 = getelementptr inbounds %struct2, %struct2* %tmp20, i64 0, i32 0
23 store i32 0, i32* %tmp21, align 4
48 %tmp21 = getelementptr inbounds %struct2, %struct2* %tmp20, i64 0, i32 0
49 store i32 0, i32* %tmp21, align 4
62 define i32 @test3(%struct3* %dm, i1 %tmp4, i64 %tmp9, i64 %tmp19, i64 %tmp20, i64 %tmp21) personali…
88 %tmp34 = getelementptr inbounds %struct4, %struct4* %phi, i64 %tmp21, i32 1
/external/llvm/test/CodeGen/AMDGPU/
Dsi-sgpr-spill.ll28 %tmp21 = load <16 x i8>, <16 x i8> addrspace(2)* %tmp, !tbaa !0
29 %tmp22 = call float @llvm.SI.load.const(<16 x i8> %tmp21, i32 96)
30 %tmp23 = call float @llvm.SI.load.const(<16 x i8> %tmp21, i32 100)
31 %tmp24 = call float @llvm.SI.load.const(<16 x i8> %tmp21, i32 104)
32 %tmp25 = call float @llvm.SI.load.const(<16 x i8> %tmp21, i32 112)
33 %tmp26 = call float @llvm.SI.load.const(<16 x i8> %tmp21, i32 116)
34 %tmp27 = call float @llvm.SI.load.const(<16 x i8> %tmp21, i32 120)
35 %tmp28 = call float @llvm.SI.load.const(<16 x i8> %tmp21, i32 128)
36 %tmp29 = call float @llvm.SI.load.const(<16 x i8> %tmp21, i32 132)
37 %tmp30 = call float @llvm.SI.load.const(<16 x i8> %tmp21, i32 140)
[all …]
/external/swiftshader/third_party/LLVM/test/CodeGen/PowerPC/
Drlwimi3.ll17 %tmp21 = lshr i32 %tmp19, 5 ; <i32> [#uses=1]
18 %tmp21.upgrd.1 = trunc i32 %tmp21 to i16 ; <i16> [#uses=1]
19 %tmp = and i16 %tmp21.upgrd.1, 31775 ; <i16> [#uses=1]
Dmask64.ll13 %tmp21 = load i8* %tmp19, align 1 ; <i8> [#uses=1]
14 switch i8 %tmp21, label %bb49 [
/external/llvm/test/CodeGen/PowerPC/
Drlwimi3.ll18 %tmp21 = lshr i32 %tmp19, 5 ; <i32> [#uses=1]
19 %tmp21.upgrd.1 = trunc i32 %tmp21 to i16 ; <i16> [#uses=1]
20 %tmp = and i16 %tmp21.upgrd.1, 31775 ; <i16> [#uses=1]
Dmask64.ll13 %tmp21 = load i8, i8* %tmp19, align 1 ; <i8> [#uses=1]
14 switch i8 %tmp21, label %bb49 [
/external/llvm/test/Transforms/NaryReassociate/
Dpr24301.ll9 %tmp21 = add i32 119, %tmp4
11 %tmp23 = add i32 %tmp21, -128
12 ; CHECK: %tmp23 = add i32 %tmp21, -128
/external/swiftshader/third_party/LLVM/test/Transforms/SimplifyCFG/
D2009-01-19-UnconditionalTrappingConstantExpr.ll7 define i32 @test(i32 %tmp21, i32 %tmp24) {
8 %tmp25 = icmp sle i32 %tmp21, %tmp24
21 define i32 @test2(i32 %tmp21, i32 %tmp24, i1 %tmp34) {
/external/swiftshader/third_party/LLVM/test/CodeGen/Generic/
Di128-addsub.ll16 %tmp21 = lshr i128 %tmp15, 64 ; <i128> [#uses=1]
17 %tmp2122 = trunc i128 %tmp21 to i64 ; <i64> [#uses=1]
35 %tmp21 = lshr i128 %tmp15, 64 ; <i128> [#uses=1]
36 %tmp2122 = trunc i128 %tmp21 to i64 ; <i64> [#uses=1]
D2008-02-04-ExtractSubvector.ll8 %tmp21 = fadd <8 x double> zeroinitializer, zeroinitializer ; <<8 x double>> [#uses=1]
12 store <8 x double> %tmp21, <8 x double>* null, align 64
/external/llvm/test/CodeGen/Generic/
Di128-addsub.ll16 %tmp21 = lshr i128 %tmp15, 64 ; <i128> [#uses=1]
17 %tmp2122 = trunc i128 %tmp21 to i64 ; <i64> [#uses=1]
35 %tmp21 = lshr i128 %tmp15, 64 ; <i128> [#uses=1]
36 %tmp2122 = trunc i128 %tmp21 to i64 ; <i64> [#uses=1]
D2008-02-04-ExtractSubvector.ll8 %tmp21 = fadd <8 x double> zeroinitializer, zeroinitializer ; <<8 x double>> [#uses=1]
12 store <8 x double> %tmp21, <8 x double>* null, align 64
/external/swiftshader/third_party/LLVM/test/CodeGen/Blackfin/
Daddsub-i128.ll19 %tmp21 = lshr i128 %tmp15, 64 ; <i128> [#uses=1]
20 %tmp2122 = trunc i128 %tmp21 to i64 ; <i64> [#uses=1]
38 %tmp21 = lshr i128 %tmp15, 64 ; <i128> [#uses=1]
39 %tmp2122 = trunc i128 %tmp21 to i64 ; <i64> [#uses=1]
/external/llvm/test/Transforms/SimplifyCFG/
D2009-01-19-UnconditionalTrappingConstantExpr.ll12 define i32 @test(i32 %tmp21, i32 %tmp24) {
13 %tmp25 = icmp sle i32 %tmp21, %tmp24
31 define i32 @test2(i32 %tmp21, i32 %tmp24, i1 %tmp34) {
/external/libjpeg-turbo/
Djidctint.c1078 JLONG tmp20, tmp21, tmp22, tmp23, tmp24; in jpeg_idct_10x10() local
1119 tmp21 = tmp11 + tmp13; in jpeg_idct_10x10()
1153 wsptr[8*1] = (int) RIGHT_SHIFT(tmp21 + tmp11, CONST_BITS-PASS1_BITS); in jpeg_idct_10x10()
1154 wsptr[8*8] = (int) RIGHT_SHIFT(tmp21 - tmp11, CONST_BITS-PASS1_BITS); in jpeg_idct_10x10()
1191 tmp21 = tmp11 + tmp13; in jpeg_idct_10x10()
1229 outptr[1] = range_limit[(int) RIGHT_SHIFT(tmp21 + tmp11, in jpeg_idct_10x10()
1232 outptr[8] = range_limit[(int) RIGHT_SHIFT(tmp21 - tmp11, in jpeg_idct_10x10()
1273 JLONG tmp20, tmp21, tmp22, tmp23, tmp24, tmp25; in jpeg_idct_11x11() local
1307 tmp21 = tmp20 + tmp23 + tmp25 - in jpeg_idct_11x11()
1345 wsptr[8*1] = (int) RIGHT_SHIFT(tmp21 + tmp11, CONST_BITS-PASS1_BITS); in jpeg_idct_11x11()
[all …]
/external/webrtc/webrtc/common_audio/signal_processing/
Dresample_by_2_mips.c59 int32_t tmp11, tmp12, tmp21, tmp22; in WebRtcSpl_DownsampleBy2() local
108 : [tmp22] "=r" (tmp22), [tmp21] "=&r" (tmp21), in WebRtcSpl_DownsampleBy2()
142 : [tmp22] "r" (tmp22), [tmp21] "r" (tmp21), in WebRtcSpl_DownsampleBy2()
/external/swiftshader/third_party/LLVM/test/CodeGen/X86/
Dmmx-arith.ll19 …%tmp21 = tail call x86_mmx @llvm.x86.mmx.paddus.b( x86_mmx %tmp12, x86_mmx %tmp16 ) ; <x86_mmx> […
20 store x86_mmx %tmp21, x86_mmx* %A
22 %tmp21a = bitcast x86_mmx %tmp21 to <8 x i8>
79 %tmp21 = load x86_mmx* %B ; <x86_mmx> [#uses=1]
81 %tmp21a = bitcast x86_mmx %tmp21 to <2 x i32>
114 …%tmp21 = tail call x86_mmx @llvm.x86.mmx.paddus.w( x86_mmx %tmp12, x86_mmx %tmp16 ) ; <x86_mmx> […
115 store x86_mmx %tmp21, x86_mmx* %A
117 %tmp21a = bitcast x86_mmx %tmp21 to <4 x i16>
175 …%tmp21 = tail call x86_mmx @llvm.x86.mmx.paddus.b( x86_mmx %tmp12, x86_mmx %tmp16 ) ; <x86_mmx> […
176 store x86_mmx %tmp21, x86_mmx* %A
[all …]
D2007-04-25-MMX-PADDQ.ll20 %tmp21 = add <1 x i64> %tmp19, %tmp14 ; <<1 x i64>> [#uses=1]
21 %tmp22 = add <1 x i64> %tmp21, %sum.035.0 ; <<1 x i64>> [#uses=2]
52 %tmp21 = call x86_mmx @llvm.x86.mmx.padd.q (x86_mmx %tmp19, x86_mmx %tmp14) ; <x86_mmx> [#uses=1]
53 …%tmp22 = call x86_mmx @llvm.x86.mmx.padd.q (x86_mmx %tmp21, x86_mmx %sum.035.0) ; <x86_mmx> [#use…
/external/llvm/test/Transforms/LCSSA/
Dindirectbr.ll553 %tmp21 = add i32 undef, 677038203
560 ; CHECK: %tmp21.lcssa1 = phi i32 [ %tmp21.lcssa1, %lab4 ], [ %tmp21, %lab2 ]
561 %tmp12 = phi i32 [ %tmp21, %lab2 ], [ %tmp12, %lab4 ]
568 ; CHECK: %tmp21.lcssa1.lcssa = phi i32 [ %tmp21.lcssa1, %lab3 ]
/external/llvm/test/Transforms/SeparateConstOffsetFromGEP/AMDGPU/
Dsplit-gep-and-gvn-addrspace-addressing-modes.ll30 %tmp21 = fadd float %tmp17, %tmp20
31 store float %tmp21, float addrspace(1)* %output, align 4
63 %tmp21 = fadd float %tmp17, %tmp20
64 store float %tmp21, float addrspace(1)* %output, align 4
91 %tmp21 = fadd float %tmp17, %tmp20
92 store float %tmp21, float addrspace(1)* %output, align 4
/external/llvm/test/Transforms/PhaseOrdering/
DPR6627.ll33 %tmp21 = load i8, i8* %arrayidx20, align 1
34 %conv22 = zext i8 %tmp21 to i32
77 %tmp21 = load i8, i8* %arrayidx20, align 1
78 %conv22 = zext i8 %tmp21 to i32

12345678