Home
last modified time | relevance | path

Searched refs:overflow (Results 1 – 25 of 2087) sorted by relevance

12345678910>>...84

/external/llvm-project/llvm/test/Transforms/InstCombine/
Dumul-sign-check.ll4 ; Check that we simplify llvm.umul.with.overflow, if the overflow check is
5 ; weakened by or (icmp ne %res, 0) %overflow. This is generated by code using
12 declare { i64, i1 } @llvm.umul.with.overflow.i64(i64, i64) #0
24 %res = tail call { i64, i1 } @llvm.umul.with.overflow.i64(i64 %a, i64 %b)
25 %overflow = extractvalue { i64, i1 } %res, 1
28 %overflow.1 = or i1 %overflow, %cmp
30 ret i1 %overflow.1
44 %res = tail call { i64, i1 } @llvm.umul.with.overflow.i64(i64 %a, i64 %b)
45 %overflow = extractvalue { i64, i1 } %res, 1
48 %overflow.1 = or i1 %cmp, %overflow
[all …]
Dssub-with-overflow.ll4 declare { <2 x i32>, <2 x i1> } @llvm.ssub.with.overflow.v2i32(<2 x i32>, <2 x i32>)
6 declare { <2 x i8>, <2 x i1> } @llvm.ssub.with.overflow.v2i8(<2 x i8>, <2 x i8>)
8 declare { i32, i1 } @llvm.ssub.with.overflow.i32(i32, i32)
10 declare { i8, i1 } @llvm.ssub.with.overflow.i8(i8, i8)
14 ; CHECK-NEXT: [[TMP1:%.*]] = call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 [[X:%.*]], i32 -2…
18 %b = tail call { i32, i1 } @llvm.ssub.with.overflow.i32(i32 %a, i32 13)
29 %b = tail call { i32, i1 } @llvm.ssub.with.overflow.i32(i32 %a, i32 -7)
35 ; CHECK-NEXT: [[TMP1:%.*]] = call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 [[X:%.*]], i8 -128)
39 %b = tail call { i8, i1 } @llvm.ssub.with.overflow.i8(i8 %a, i8 28)
46 ; CHECK-NEXT: [[TMP1:%.*]] = call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 [[A]], i8 -29)
[all …]
Doverflow-mul.ll22 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y)
23 %overflow = icmp ugt i64 %mul64, 4294967295
25 %retval = zext i1 %overflow to i32
37 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y)
38 %overflow = icmp uge i64 %mul64, 4294967296
40 %retval = zext i1 %overflow to i32
53 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y)
54 %overflow = icmp ugt i64 %mul64, 4294967295
58 %retval = select i1 %overflow, i32 %mul32, i32 111
71 ; CHECK-NOT: umul.with.overflow.i32
[all …]
Dsadd-with-overflow.ll4 declare { <2 x i32>, <2 x i1> } @llvm.sadd.with.overflow.v2i32(<2 x i32>, <2 x i32>)
6 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32)
8 declare { i8, i1 } @llvm.sadd.with.overflow.i8(i8, i8)
12 ; CHECK-NEXT: [[TMP1:%.*]] = call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 [[X:%.*]], i32 20)
16 %b = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 13)
27 %b = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 -7)
33 ; CHECK-NEXT: [[TMP1:%.*]] = call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 [[X:%.*]], i8 127)
37 %b = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 %a, i8 27)
44 ; CHECK-NEXT: [[B:%.*]] = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 [[A]], i8 28)
48 %b = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 %a, i8 28)
[all …]
Duadd-with-overflow.ll4 declare { <2 x i32>, <2 x i1> } @llvm.uadd.with.overflow.v2i32(<2 x i32>, <2 x i32>)
6 declare { <2 x i8>, <2 x i1> } @llvm.uadd.with.overflow.v2i8(<2 x i8>, <2 x i8>)
8 declare { i32, i1 } @llvm.uadd.with.overflow.i32(i32, i32)
10 declare { i8, i1 } @llvm.uadd.with.overflow.i8(i8, i8)
14 ; CHECK-NEXT: [[TMP1:%.*]] = call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 [[X:%.*]], i32 20)
18 %b = tail call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 %a, i32 13)
24 ; CHECK-NEXT: [[TMP1:%.*]] = call { i8, i1 } @llvm.uadd.with.overflow.i8(i8 [[X:%.*]], i8 -1)
28 %b = tail call { i8, i1 } @llvm.uadd.with.overflow.i8(i8 %a, i8 55)
38 %b = tail call { i8, i1 } @llvm.uadd.with.overflow.i8(i8 %a, i8 56)
45 ; CHECK-NEXT: [[B:%.*]] = tail call { <2 x i8>, <2 x i1> } @llvm.uadd.with.overflow.v2i8(<2 x i8…
[all …]
Dwith_overflow.ll4 declare { i8, i1 } @llvm.uadd.with.overflow.i8(i8, i8) nounwind readnone
5 declare { i8, i1 } @llvm.sadd.with.overflow.i8(i8, i8) nounwind readnone
6 declare { i8, i1 } @llvm.usub.with.overflow.i8(i8, i8) nounwind readnone
7 declare { i8, i1 } @llvm.ssub.with.overflow.i8(i8, i8) nounwind readnone
8 declare { i8, i1 } @llvm.umul.with.overflow.i8(i8, i8) nounwind readnone
9 declare { i8, i1 } @llvm.smul.with.overflow.i8(i8, i8) nounwind readnone
10 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) nounwind readnone
11 declare { i32, i1 } @llvm.uadd.with.overflow.i32(i32, i32) nounwind readnone
12 declare { i32, i1 } @llvm.ssub.with.overflow.i32(i32, i32) nounwind readnone
13 declare { i32, i1 } @llvm.usub.with.overflow.i32(i32, i32) nounwind readnone
[all …]
/external/llvm-project/llvm/test/Analysis/CostModel/ARM/
Darith-overflow.ll9 declare {i64, i1} @llvm.sadd.with.overflow.i64(i64, i64)
10 declare {<2 x i64>, <2 x i1>} @llvm.sadd.with.overflow.v2i64(<2 x i64>, <2 x i64>)
11 declare {<4 x i64>, <4 x i1>} @llvm.sadd.with.overflow.v4i64(<4 x i64>, <4 x i64>)
12 declare {<8 x i64>, <8 x i1>} @llvm.sadd.with.overflow.v8i64(<8 x i64>, <8 x i64>)
14 declare {i32, i1} @llvm.sadd.with.overflow.i32(i32, i32)
15 declare {<4 x i32>, <4 x i1>} @llvm.sadd.with.overflow.v4i32(<4 x i32>, <4 x i32>)
16 declare {<8 x i32>, <8 x i1>} @llvm.sadd.with.overflow.v8i32(<8 x i32>, <8 x i32>)
17 declare {<16 x i32>, <16 x i1>} @llvm.sadd.with.overflow.v16i32(<16 x i32>, <16 x i32>)
19 declare {i16, i1} @llvm.sadd.with.overflow.i16(i16, i16)
20 declare {<8 x i16>, <8 x i1>} @llvm.sadd.with.overflow.v8i16(<8 x i16>, <8 x i16>)
[all …]
/external/llvm-project/llvm/test/Analysis/CostModel/X86/
Darith-overflow.ll15 ; sadd.with.overflow
18 declare {i64, i1} @llvm.sadd.with.overflow.i64(i64, i64)
19 declare {<2 x i64>, <2 x i1>} @llvm.sadd.with.overflow.v2i64(<2 x i64>, <2 x i64>)
20 declare {<4 x i64>, <4 x i1>} @llvm.sadd.with.overflow.v4i64(<4 x i64>, <4 x i64>)
21 declare {<8 x i64>, <8 x i1>} @llvm.sadd.with.overflow.v8i64(<8 x i64>, <8 x i64>)
23 declare {i32, i1} @llvm.sadd.with.overflow.i32(i32, i32)
24 declare {<4 x i32>, <4 x i1>} @llvm.sadd.with.overflow.v4i32(<4 x i32>, <4 x i32>)
25 declare {<8 x i32>, <8 x i1>} @llvm.sadd.with.overflow.v8i32(<8 x i32>, <8 x i32>)
26 declare {<16 x i32>, <16 x i1>} @llvm.sadd.with.overflow.v16i32(<16 x i32>, <16 x i32>)
28 declare {i16, i1} @llvm.sadd.with.overflow.i16(i16, i16)
[all …]
/external/llvm/test/Transforms/ConstProp/
Doverflow-ops.ll4 declare {i8, i1} @llvm.uadd.with.overflow.i8(i8, i8)
5 declare {i8, i1} @llvm.usub.with.overflow.i8(i8, i8)
6 declare {i8, i1} @llvm.umul.with.overflow.i8(i8, i8)
8 declare {i8, i1} @llvm.sadd.with.overflow.i8(i8, i8)
9 declare {i8, i1} @llvm.ssub.with.overflow.i8(i8, i8)
10 declare {i8, i1} @llvm.smul.with.overflow.i8(i8, i8)
18 %t = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 42, i8 100)
27 %t = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 142, i8 120)
40 %t = call {i8, i1} @llvm.usub.with.overflow.i8(i8 4, i8 2)
49 %t = call {i8, i1} @llvm.usub.with.overflow.i8(i8 4, i8 6)
[all …]
/external/llvm/test/CodeGen/AArch64/
Darm64-xaluo.ll5 ; Get the actual value of the overflow bit.
12 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %v2)
25 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 4)
38 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 -4)
51 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 16777215)
65 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %lsl)
77 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 %v2)
89 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 4)
101 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 -4)
113 %t = call {i32, i1} @llvm.uadd.with.overflow.i32(i32 %v1, i32 %v2)
[all …]
/external/llvm/test/CodeGen/X86/
Dxaluo.ll5 ; Get the actual value of the overflow bit.
13 %t = call {i8, i1} @llvm.sadd.with.overflow.i8(i8 %v1, i8 %v2)
25 %t = call {i16, i1} @llvm.sadd.with.overflow.i16(i16 %v1, i16 %v2)
37 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %v2)
49 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 %v2)
62 %t = call {i8, i1} @llvm.sadd.with.overflow.i8(i8 %v1, i8 1)
74 %t = call {i16, i1} @llvm.sadd.with.overflow.i16(i16 %v1, i16 1)
86 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 1)
98 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 1)
116 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 2, i64 %v1)
[all …]
Dsmul-with-overflow.ll8 %t = call {i32, i1} @llvm.smul.with.overflow.i32(i32 %v1, i32 %v2)
11 br i1 %obit, label %overflow, label %normal
17 overflow:
27 %t = call {i32, i1} @llvm.smul.with.overflow.i32(i32 %v1, i32 %v2)
30 br i1 %obit, label %overflow, label %normal
32 overflow:
45 declare {i32, i1} @llvm.smul.with.overflow.i32(i32, i32)
50 %tmp1 = call { i32, i1 } @llvm.smul.with.overflow.i32(i32 %tmp0, i32 2)
62 %tmp1 = call { i32, i1 } @llvm.smul.with.overflow.i32(i32 %tmp0, i32 4)
71 declare { i63, i1 } @llvm.smul.with.overflow.i63(i63, i63) nounwind readnone
[all …]
/external/llvm/test/CodeGen/Generic/
Doverflow.ll2 ; Verify codegen's don't crash on overflow intrinsics.
8 %sadd = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 %a, i8 %b)
15 declare { i8, i1 } @llvm.sadd.with.overflow.i8(i8, i8) nounwind readnone
19 %sadd = tail call { i16, i1 } @llvm.sadd.with.overflow.i16(i16 %a, i16 %b)
26 declare { i16, i1 } @llvm.sadd.with.overflow.i16(i16, i16) nounwind readnone
30 %sadd = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 %b)
37 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) nounwind readnone
44 %uadd = tail call { i8, i1 } @llvm.uadd.with.overflow.i8(i8 %a, i8 %b)
51 declare { i8, i1 } @llvm.uadd.with.overflow.i8(i8, i8) nounwind readnone
55 %uadd = tail call { i16, i1 } @llvm.uadd.with.overflow.i16(i16 %a, i16 %b)
[all …]
/external/llvm-project/llvm/test/CodeGen/Generic/
Doverflow.ll2 ; Verify codegen's don't crash on overflow intrinsics.
8 %sadd = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 %a, i8 %b)
15 declare { i8, i1 } @llvm.sadd.with.overflow.i8(i8, i8) nounwind readnone
19 %sadd = tail call { i16, i1 } @llvm.sadd.with.overflow.i16(i16 %a, i16 %b)
26 declare { i16, i1 } @llvm.sadd.with.overflow.i16(i16, i16) nounwind readnone
30 %sadd = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 %b)
37 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) nounwind readnone
44 %uadd = tail call { i8, i1 } @llvm.uadd.with.overflow.i8(i8 %a, i8 %b)
51 declare { i8, i1 } @llvm.uadd.with.overflow.i8(i8, i8) nounwind readnone
55 %uadd = tail call { i16, i1 } @llvm.uadd.with.overflow.i16(i16 %a, i16 %b)
[all …]
/external/llvm-project/llvm/test/Transforms/InstSimplify/ConstProp/
Doverflow-ops.ll4 declare {i8, i1} @llvm.uadd.with.overflow.i8(i8, i8)
5 declare {i8, i1} @llvm.usub.with.overflow.i8(i8, i8)
6 declare {i8, i1} @llvm.umul.with.overflow.i8(i8, i8)
8 declare {i8, i1} @llvm.sadd.with.overflow.i8(i8, i8)
9 declare {i8, i1} @llvm.ssub.with.overflow.i8(i8, i8)
10 declare {i8, i1} @llvm.smul.with.overflow.i8(i8, i8)
20 %t = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 42, i8 100)
28 %t = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 142, i8 120)
36 %t = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 142, i8 undef)
48 %t = call {i8, i1} @llvm.usub.with.overflow.i8(i8 4, i8 2)
[all …]
/external/python/cpython3/Modules/_decimal/libmpdec/
Dmpalloc.c62 mpd_size_t overflow; in mpd_callocfunc_em() local
72 &overflow); in mpd_callocfunc_em()
73 if (overflow) { in mpd_callocfunc_em()
92 mpd_size_t req, overflow; in mpd_alloc() local
94 req = mul_size_t_overflow(nmemb, size, &overflow); in mpd_alloc()
95 if (overflow) { in mpd_alloc()
106 mpd_size_t overflow; in mpd_calloc() local
108 (void)mul_size_t_overflow(nmemb, size, &overflow); in mpd_calloc()
109 if (overflow) { in mpd_calloc()
121 mpd_size_t req, overflow; in mpd_realloc() local
[all …]
/external/llvm-project/llvm/test/tools/llvm-profdata/
Doverflow-instr.test1 Tests for overflow when merging instrumented profiles.
3 1- Merge profile having maximum counts with itself and verify overflow detected and saturation occu…
4 RUN: llvm-profdata merge -instr %p/Inputs/overflow-instr.proftext %p/Inputs/overflow-instr.proftext…
6 MERGE_OVERFLOW: {{.*}}: overflow: Counter overflow
10 2- Merge profile having maximum counts by itself and verify no overflow
11 RUN: llvm-profdata merge -instr %p/Inputs/overflow-instr.proftext -o %t.out 2>&1 | FileCheck %s -ch…
13 MERGE_NO_OVERFLOW-NOT: {{.*}}: overflow: Counter overflow
/external/llvm/test/tools/llvm-profdata/
Doverflow-instr.test1 Tests for overflow when merging instrumented profiles.
3 1- Merge profile having maximum counts with itself and verify overflow detected and saturation occu…
4 RUN: llvm-profdata merge -instr %p/Inputs/overflow-instr.proftext %p/Inputs/overflow-instr.proftext…
6 MERGE_OVERFLOW: {{.*}}: overflow: Counter overflow
11 2- Merge profile having maximum counts by itself and verify no overflow
12 RUN: llvm-profdata merge -instr %p/Inputs/overflow-instr.proftext -o %t.out 2>&1 | FileCheck %s -ch…
14 MERGE_NO_OVERFLOW-NOT: {{.*}}: overflow: Counter overflow
/external/llvm-project/llvm/test/CodeGen/AArch64/
Darm64-xaluo.ll6 ; Get the actual value of the overflow bit.
13 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %v2)
26 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 4)
39 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 -4)
52 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 16777215)
66 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %lsl)
78 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 %v2)
90 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 4)
102 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 -4)
118 %t = call {i32, i1} @llvm.uadd.with.overflow.i32(i32 %v1, i32 %v2)
[all …]
/external/llvm/test/Transforms/InstCombine/
Doverflow-mul.ll11 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y)
12 %overflow = icmp ugt i64 %mul64, 4294967295
14 %retval = zext i1 %overflow to i32
26 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y)
27 %overflow = icmp uge i64 %mul64, 4294967296
29 %retval = zext i1 %overflow to i32
42 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y)
43 %overflow = icmp ugt i64 %mul64, 4294967295
47 %retval = select i1 %overflow, i32 %mul32, i32 111
60 ; CHECK-NOT: umul.with.overflow.i32
[all …]
Dintrinsics.ll3 %overflow.result = type {i8, i1}
7 declare %overflow.result @llvm.uadd.with.overflow.i8(i8, i8) nounwind readnone
8 declare %overflow.result @llvm.umul.with.overflow.i8(i8, i8) nounwind readnone
9 declare %ov.result.32 @llvm.sadd.with.overflow.i32(i32, i32) nounwind readnone
10 declare %ov.result.32 @llvm.uadd.with.overflow.i32(i32, i32) nounwind readnone
11 declare %ov.result.32 @llvm.ssub.with.overflow.i32(i32, i32) nounwind readnone
12 declare %ov.result.32 @llvm.usub.with.overflow.i32(i32, i32) nounwind readnone
13 declare %ov.result.32 @llvm.smul.with.overflow.i32(i32, i32) nounwind readnone
14 declare %ov.result.32 @llvm.umul.with.overflow.i32(i32, i32) nounwind readnone
29 %x = call %overflow.result @llvm.uadd.with.overflow.i8(i8 %A, i8 %B)
[all …]
/external/llvm-project/llvm/test/CodeGen/ARM/
Dunschedule-first-call.ll20 %13 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 undef, i64 undef)
22 %15 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %14, i64 1)
27 %20 = call { i64, i1 } @llvm.smul.with.overflow.i64(i64 %18, i64 %19)
29 %22 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %21, i64 0)
31 %24 = call { i64, i1 } @llvm.smul.with.overflow.i64(i64 %23, i64 undef)
33 %26 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %25, i64 0)
41 %34 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 0, i64 undef)
43 %36 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %35, i64 1)
48 %41 = call { i64, i1 } @llvm.smul.with.overflow.i64(i64 %39, i64 %40)
50 %43 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %42, i64 0)
[all …]
/external/llvm-project/llvm/test/CodeGen/X86/
Doverflow-intrinsic-setcc-fold.ll11 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %v2)
24 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 %v2)
37 %t = call {i32, i1} @llvm.uadd.with.overflow.i32(i32 %v1, i32 %v2)
50 %t = call {i64, i1} @llvm.uadd.with.overflow.i64(i64 %v1, i64 %v2)
63 %t = call {i32, i1} @llvm.ssub.with.overflow.i32(i32 %v1, i32 %v2)
76 %t = call {i64, i1} @llvm.ssub.with.overflow.i64(i64 %v1, i64 %v2)
89 %t = call {i32, i1} @llvm.usub.with.overflow.i32(i32 %v1, i32 %v2)
102 %t = call {i64, i1} @llvm.usub.with.overflow.i64(i64 %v1, i64 %v2)
115 %t = call {i32, i1} @llvm.smul.with.overflow.i32(i32 %v1, i32 %v2)
128 %t = call {i64, i1} @llvm.smul.with.overflow.i64(i64 %v1, i64 %v2)
[all …]
Dsmul-with-overflow.ll8 %t = call {i32, i1} @llvm.smul.with.overflow.i32(i32 %v1, i32 %v2)
11 br i1 %obit, label %overflow, label %normal
17 overflow:
27 %t = call {i32, i1} @llvm.smul.with.overflow.i32(i32 %v1, i32 %v2)
30 br i1 %obit, label %overflow, label %normal
32 overflow:
45 declare {i32, i1} @llvm.smul.with.overflow.i32(i32, i32)
50 %tmp1 = call { i32, i1 } @llvm.smul.with.overflow.i32(i32 %tmp0, i32 2)
62 %tmp1 = call { i32, i1 } @llvm.smul.with.overflow.i32(i32 %tmp0, i32 4)
71 declare { i63, i1 } @llvm.smul.with.overflow.i63(i63, i63) nounwind readnone
[all …]
/external/llvm/test/Transforms/GVN/
D2011-07-07-MatchIntrinsicExtract.ll8 %uadd = tail call %0 @llvm.uadd.with.overflow.i64(i64 %a, i64 %b)
20 %usub = tail call %0 @llvm.usub.with.overflow.i64(i64 %a, i64 %b)
32 %umul = tail call %0 @llvm.umul.with.overflow.i64(i64 %a, i64 %b)
44 %sadd = tail call %0 @llvm.sadd.with.overflow.i64(i64 %a, i64 %b)
56 %ssub = tail call %0 @llvm.ssub.with.overflow.i64(i64 %a, i64 %b)
68 %smul = tail call %0 @llvm.smul.with.overflow.i64(i64 %a, i64 %b)
79 declare %0 @llvm.uadd.with.overflow.i64(i64, i64) nounwind readnone
80 declare %0 @llvm.usub.with.overflow.i64(i64, i64) nounwind readnone
81 declare %0 @llvm.umul.with.overflow.i64(i64, i64) nounwind readnone
82 declare %0 @llvm.sadd.with.overflow.i64(i64, i64) nounwind readnone
[all …]

12345678910>>...84