| /external/llvm-project/llvm/test/Transforms/InstCombine/ |
| D | umul-sign-check.ll | 4 ; Check that we simplify llvm.umul.with.overflow, if the overflow check is 5 ; weakened by or (icmp ne %res, 0) %overflow. This is generated by code using 12 declare { i64, i1 } @llvm.umul.with.overflow.i64(i64, i64) #0 24 %res = tail call { i64, i1 } @llvm.umul.with.overflow.i64(i64 %a, i64 %b) 25 %overflow = extractvalue { i64, i1 } %res, 1 28 %overflow.1 = or i1 %overflow, %cmp 30 ret i1 %overflow.1 44 %res = tail call { i64, i1 } @llvm.umul.with.overflow.i64(i64 %a, i64 %b) 45 %overflow = extractvalue { i64, i1 } %res, 1 48 %overflow.1 = or i1 %cmp, %overflow [all …]
|
| D | ssub-with-overflow.ll | 4 declare { <2 x i32>, <2 x i1> } @llvm.ssub.with.overflow.v2i32(<2 x i32>, <2 x i32>) 6 declare { <2 x i8>, <2 x i1> } @llvm.ssub.with.overflow.v2i8(<2 x i8>, <2 x i8>) 8 declare { i32, i1 } @llvm.ssub.with.overflow.i32(i32, i32) 10 declare { i8, i1 } @llvm.ssub.with.overflow.i8(i8, i8) 14 ; CHECK-NEXT: [[TMP1:%.*]] = call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 [[X:%.*]], i32 -2… 18 %b = tail call { i32, i1 } @llvm.ssub.with.overflow.i32(i32 %a, i32 13) 29 %b = tail call { i32, i1 } @llvm.ssub.with.overflow.i32(i32 %a, i32 -7) 35 ; CHECK-NEXT: [[TMP1:%.*]] = call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 [[X:%.*]], i8 -128) 39 %b = tail call { i8, i1 } @llvm.ssub.with.overflow.i8(i8 %a, i8 28) 46 ; CHECK-NEXT: [[TMP1:%.*]] = call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 [[A]], i8 -29) [all …]
|
| D | overflow-mul.ll | 22 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y) 23 %overflow = icmp ugt i64 %mul64, 4294967295 25 %retval = zext i1 %overflow to i32 37 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y) 38 %overflow = icmp uge i64 %mul64, 4294967296 40 %retval = zext i1 %overflow to i32 53 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y) 54 %overflow = icmp ugt i64 %mul64, 4294967295 58 %retval = select i1 %overflow, i32 %mul32, i32 111 71 ; CHECK-NOT: umul.with.overflow.i32 [all …]
|
| D | sadd-with-overflow.ll | 4 declare { <2 x i32>, <2 x i1> } @llvm.sadd.with.overflow.v2i32(<2 x i32>, <2 x i32>) 6 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) 8 declare { i8, i1 } @llvm.sadd.with.overflow.i8(i8, i8) 12 ; CHECK-NEXT: [[TMP1:%.*]] = call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 [[X:%.*]], i32 20) 16 %b = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 13) 27 %b = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 -7) 33 ; CHECK-NEXT: [[TMP1:%.*]] = call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 [[X:%.*]], i8 127) 37 %b = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 %a, i8 27) 44 ; CHECK-NEXT: [[B:%.*]] = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 [[A]], i8 28) 48 %b = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 %a, i8 28) [all …]
|
| D | uadd-with-overflow.ll | 4 declare { <2 x i32>, <2 x i1> } @llvm.uadd.with.overflow.v2i32(<2 x i32>, <2 x i32>) 6 declare { <2 x i8>, <2 x i1> } @llvm.uadd.with.overflow.v2i8(<2 x i8>, <2 x i8>) 8 declare { i32, i1 } @llvm.uadd.with.overflow.i32(i32, i32) 10 declare { i8, i1 } @llvm.uadd.with.overflow.i8(i8, i8) 14 ; CHECK-NEXT: [[TMP1:%.*]] = call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 [[X:%.*]], i32 20) 18 %b = tail call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 %a, i32 13) 24 ; CHECK-NEXT: [[TMP1:%.*]] = call { i8, i1 } @llvm.uadd.with.overflow.i8(i8 [[X:%.*]], i8 -1) 28 %b = tail call { i8, i1 } @llvm.uadd.with.overflow.i8(i8 %a, i8 55) 38 %b = tail call { i8, i1 } @llvm.uadd.with.overflow.i8(i8 %a, i8 56) 45 ; CHECK-NEXT: [[B:%.*]] = tail call { <2 x i8>, <2 x i1> } @llvm.uadd.with.overflow.v2i8(<2 x i8… [all …]
|
| D | with_overflow.ll | 4 declare { i8, i1 } @llvm.uadd.with.overflow.i8(i8, i8) nounwind readnone 5 declare { i8, i1 } @llvm.sadd.with.overflow.i8(i8, i8) nounwind readnone 6 declare { i8, i1 } @llvm.usub.with.overflow.i8(i8, i8) nounwind readnone 7 declare { i8, i1 } @llvm.ssub.with.overflow.i8(i8, i8) nounwind readnone 8 declare { i8, i1 } @llvm.umul.with.overflow.i8(i8, i8) nounwind readnone 9 declare { i8, i1 } @llvm.smul.with.overflow.i8(i8, i8) nounwind readnone 10 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) nounwind readnone 11 declare { i32, i1 } @llvm.uadd.with.overflow.i32(i32, i32) nounwind readnone 12 declare { i32, i1 } @llvm.ssub.with.overflow.i32(i32, i32) nounwind readnone 13 declare { i32, i1 } @llvm.usub.with.overflow.i32(i32, i32) nounwind readnone [all …]
|
| /external/llvm-project/llvm/test/Analysis/CostModel/ARM/ |
| D | arith-overflow.ll | 9 declare {i64, i1} @llvm.sadd.with.overflow.i64(i64, i64) 10 declare {<2 x i64>, <2 x i1>} @llvm.sadd.with.overflow.v2i64(<2 x i64>, <2 x i64>) 11 declare {<4 x i64>, <4 x i1>} @llvm.sadd.with.overflow.v4i64(<4 x i64>, <4 x i64>) 12 declare {<8 x i64>, <8 x i1>} @llvm.sadd.with.overflow.v8i64(<8 x i64>, <8 x i64>) 14 declare {i32, i1} @llvm.sadd.with.overflow.i32(i32, i32) 15 declare {<4 x i32>, <4 x i1>} @llvm.sadd.with.overflow.v4i32(<4 x i32>, <4 x i32>) 16 declare {<8 x i32>, <8 x i1>} @llvm.sadd.with.overflow.v8i32(<8 x i32>, <8 x i32>) 17 declare {<16 x i32>, <16 x i1>} @llvm.sadd.with.overflow.v16i32(<16 x i32>, <16 x i32>) 19 declare {i16, i1} @llvm.sadd.with.overflow.i16(i16, i16) 20 declare {<8 x i16>, <8 x i1>} @llvm.sadd.with.overflow.v8i16(<8 x i16>, <8 x i16>) [all …]
|
| /external/llvm-project/llvm/test/Analysis/CostModel/X86/ |
| D | arith-overflow.ll | 15 ; sadd.with.overflow 18 declare {i64, i1} @llvm.sadd.with.overflow.i64(i64, i64) 19 declare {<2 x i64>, <2 x i1>} @llvm.sadd.with.overflow.v2i64(<2 x i64>, <2 x i64>) 20 declare {<4 x i64>, <4 x i1>} @llvm.sadd.with.overflow.v4i64(<4 x i64>, <4 x i64>) 21 declare {<8 x i64>, <8 x i1>} @llvm.sadd.with.overflow.v8i64(<8 x i64>, <8 x i64>) 23 declare {i32, i1} @llvm.sadd.with.overflow.i32(i32, i32) 24 declare {<4 x i32>, <4 x i1>} @llvm.sadd.with.overflow.v4i32(<4 x i32>, <4 x i32>) 25 declare {<8 x i32>, <8 x i1>} @llvm.sadd.with.overflow.v8i32(<8 x i32>, <8 x i32>) 26 declare {<16 x i32>, <16 x i1>} @llvm.sadd.with.overflow.v16i32(<16 x i32>, <16 x i32>) 28 declare {i16, i1} @llvm.sadd.with.overflow.i16(i16, i16) [all …]
|
| /external/llvm/test/Transforms/ConstProp/ |
| D | overflow-ops.ll | 4 declare {i8, i1} @llvm.uadd.with.overflow.i8(i8, i8) 5 declare {i8, i1} @llvm.usub.with.overflow.i8(i8, i8) 6 declare {i8, i1} @llvm.umul.with.overflow.i8(i8, i8) 8 declare {i8, i1} @llvm.sadd.with.overflow.i8(i8, i8) 9 declare {i8, i1} @llvm.ssub.with.overflow.i8(i8, i8) 10 declare {i8, i1} @llvm.smul.with.overflow.i8(i8, i8) 18 %t = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 42, i8 100) 27 %t = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 142, i8 120) 40 %t = call {i8, i1} @llvm.usub.with.overflow.i8(i8 4, i8 2) 49 %t = call {i8, i1} @llvm.usub.with.overflow.i8(i8 4, i8 6) [all …]
|
| /external/llvm/test/CodeGen/AArch64/ |
| D | arm64-xaluo.ll | 5 ; Get the actual value of the overflow bit. 12 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %v2) 25 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 4) 38 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 -4) 51 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 16777215) 65 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %lsl) 77 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 %v2) 89 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 4) 101 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 -4) 113 %t = call {i32, i1} @llvm.uadd.with.overflow.i32(i32 %v1, i32 %v2) [all …]
|
| /external/llvm/test/CodeGen/X86/ |
| D | xaluo.ll | 5 ; Get the actual value of the overflow bit. 13 %t = call {i8, i1} @llvm.sadd.with.overflow.i8(i8 %v1, i8 %v2) 25 %t = call {i16, i1} @llvm.sadd.with.overflow.i16(i16 %v1, i16 %v2) 37 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %v2) 49 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 %v2) 62 %t = call {i8, i1} @llvm.sadd.with.overflow.i8(i8 %v1, i8 1) 74 %t = call {i16, i1} @llvm.sadd.with.overflow.i16(i16 %v1, i16 1) 86 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 1) 98 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 1) 116 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 2, i64 %v1) [all …]
|
| D | smul-with-overflow.ll | 8 %t = call {i32, i1} @llvm.smul.with.overflow.i32(i32 %v1, i32 %v2) 11 br i1 %obit, label %overflow, label %normal 17 overflow: 27 %t = call {i32, i1} @llvm.smul.with.overflow.i32(i32 %v1, i32 %v2) 30 br i1 %obit, label %overflow, label %normal 32 overflow: 45 declare {i32, i1} @llvm.smul.with.overflow.i32(i32, i32) 50 %tmp1 = call { i32, i1 } @llvm.smul.with.overflow.i32(i32 %tmp0, i32 2) 62 %tmp1 = call { i32, i1 } @llvm.smul.with.overflow.i32(i32 %tmp0, i32 4) 71 declare { i63, i1 } @llvm.smul.with.overflow.i63(i63, i63) nounwind readnone [all …]
|
| /external/llvm/test/CodeGen/Generic/ |
| D | overflow.ll | 2 ; Verify codegen's don't crash on overflow intrinsics. 8 %sadd = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 %a, i8 %b) 15 declare { i8, i1 } @llvm.sadd.with.overflow.i8(i8, i8) nounwind readnone 19 %sadd = tail call { i16, i1 } @llvm.sadd.with.overflow.i16(i16 %a, i16 %b) 26 declare { i16, i1 } @llvm.sadd.with.overflow.i16(i16, i16) nounwind readnone 30 %sadd = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 37 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) nounwind readnone 44 %uadd = tail call { i8, i1 } @llvm.uadd.with.overflow.i8(i8 %a, i8 %b) 51 declare { i8, i1 } @llvm.uadd.with.overflow.i8(i8, i8) nounwind readnone 55 %uadd = tail call { i16, i1 } @llvm.uadd.with.overflow.i16(i16 %a, i16 %b) [all …]
|
| /external/llvm-project/llvm/test/CodeGen/Generic/ |
| D | overflow.ll | 2 ; Verify codegen's don't crash on overflow intrinsics. 8 %sadd = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 %a, i8 %b) 15 declare { i8, i1 } @llvm.sadd.with.overflow.i8(i8, i8) nounwind readnone 19 %sadd = tail call { i16, i1 } @llvm.sadd.with.overflow.i16(i16 %a, i16 %b) 26 declare { i16, i1 } @llvm.sadd.with.overflow.i16(i16, i16) nounwind readnone 30 %sadd = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 37 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) nounwind readnone 44 %uadd = tail call { i8, i1 } @llvm.uadd.with.overflow.i8(i8 %a, i8 %b) 51 declare { i8, i1 } @llvm.uadd.with.overflow.i8(i8, i8) nounwind readnone 55 %uadd = tail call { i16, i1 } @llvm.uadd.with.overflow.i16(i16 %a, i16 %b) [all …]
|
| /external/llvm-project/llvm/test/Transforms/InstSimplify/ConstProp/ |
| D | overflow-ops.ll | 4 declare {i8, i1} @llvm.uadd.with.overflow.i8(i8, i8) 5 declare {i8, i1} @llvm.usub.with.overflow.i8(i8, i8) 6 declare {i8, i1} @llvm.umul.with.overflow.i8(i8, i8) 8 declare {i8, i1} @llvm.sadd.with.overflow.i8(i8, i8) 9 declare {i8, i1} @llvm.ssub.with.overflow.i8(i8, i8) 10 declare {i8, i1} @llvm.smul.with.overflow.i8(i8, i8) 20 %t = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 42, i8 100) 28 %t = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 142, i8 120) 36 %t = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 142, i8 undef) 48 %t = call {i8, i1} @llvm.usub.with.overflow.i8(i8 4, i8 2) [all …]
|
| /external/python/cpython3/Modules/_decimal/libmpdec/ |
| D | mpalloc.c | 62 mpd_size_t overflow; in mpd_callocfunc_em() local 72 &overflow); in mpd_callocfunc_em() 73 if (overflow) { in mpd_callocfunc_em() 92 mpd_size_t req, overflow; in mpd_alloc() local 94 req = mul_size_t_overflow(nmemb, size, &overflow); in mpd_alloc() 95 if (overflow) { in mpd_alloc() 106 mpd_size_t overflow; in mpd_calloc() local 108 (void)mul_size_t_overflow(nmemb, size, &overflow); in mpd_calloc() 109 if (overflow) { in mpd_calloc() 121 mpd_size_t req, overflow; in mpd_realloc() local [all …]
|
| /external/llvm-project/llvm/test/tools/llvm-profdata/ |
| D | overflow-instr.test | 1 Tests for overflow when merging instrumented profiles. 3 1- Merge profile having maximum counts with itself and verify overflow detected and saturation occu… 4 RUN: llvm-profdata merge -instr %p/Inputs/overflow-instr.proftext %p/Inputs/overflow-instr.proftext… 6 MERGE_OVERFLOW: {{.*}}: overflow: Counter overflow 10 2- Merge profile having maximum counts by itself and verify no overflow 11 RUN: llvm-profdata merge -instr %p/Inputs/overflow-instr.proftext -o %t.out 2>&1 | FileCheck %s -ch… 13 MERGE_NO_OVERFLOW-NOT: {{.*}}: overflow: Counter overflow
|
| /external/llvm/test/tools/llvm-profdata/ |
| D | overflow-instr.test | 1 Tests for overflow when merging instrumented profiles. 3 1- Merge profile having maximum counts with itself and verify overflow detected and saturation occu… 4 RUN: llvm-profdata merge -instr %p/Inputs/overflow-instr.proftext %p/Inputs/overflow-instr.proftext… 6 MERGE_OVERFLOW: {{.*}}: overflow: Counter overflow 11 2- Merge profile having maximum counts by itself and verify no overflow 12 RUN: llvm-profdata merge -instr %p/Inputs/overflow-instr.proftext -o %t.out 2>&1 | FileCheck %s -ch… 14 MERGE_NO_OVERFLOW-NOT: {{.*}}: overflow: Counter overflow
|
| /external/llvm-project/llvm/test/CodeGen/AArch64/ |
| D | arm64-xaluo.ll | 6 ; Get the actual value of the overflow bit. 13 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %v2) 26 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 4) 39 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 -4) 52 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 16777215) 66 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %lsl) 78 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 %v2) 90 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 4) 102 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 -4) 118 %t = call {i32, i1} @llvm.uadd.with.overflow.i32(i32 %v1, i32 %v2) [all …]
|
| /external/llvm/test/Transforms/InstCombine/ |
| D | overflow-mul.ll | 11 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y) 12 %overflow = icmp ugt i64 %mul64, 4294967295 14 %retval = zext i1 %overflow to i32 26 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y) 27 %overflow = icmp uge i64 %mul64, 4294967296 29 %retval = zext i1 %overflow to i32 42 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y) 43 %overflow = icmp ugt i64 %mul64, 4294967295 47 %retval = select i1 %overflow, i32 %mul32, i32 111 60 ; CHECK-NOT: umul.with.overflow.i32 [all …]
|
| D | intrinsics.ll | 3 %overflow.result = type {i8, i1} 7 declare %overflow.result @llvm.uadd.with.overflow.i8(i8, i8) nounwind readnone 8 declare %overflow.result @llvm.umul.with.overflow.i8(i8, i8) nounwind readnone 9 declare %ov.result.32 @llvm.sadd.with.overflow.i32(i32, i32) nounwind readnone 10 declare %ov.result.32 @llvm.uadd.with.overflow.i32(i32, i32) nounwind readnone 11 declare %ov.result.32 @llvm.ssub.with.overflow.i32(i32, i32) nounwind readnone 12 declare %ov.result.32 @llvm.usub.with.overflow.i32(i32, i32) nounwind readnone 13 declare %ov.result.32 @llvm.smul.with.overflow.i32(i32, i32) nounwind readnone 14 declare %ov.result.32 @llvm.umul.with.overflow.i32(i32, i32) nounwind readnone 29 %x = call %overflow.result @llvm.uadd.with.overflow.i8(i8 %A, i8 %B) [all …]
|
| /external/llvm-project/llvm/test/CodeGen/ARM/ |
| D | unschedule-first-call.ll | 20 %13 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 undef, i64 undef) 22 %15 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %14, i64 1) 27 %20 = call { i64, i1 } @llvm.smul.with.overflow.i64(i64 %18, i64 %19) 29 %22 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %21, i64 0) 31 %24 = call { i64, i1 } @llvm.smul.with.overflow.i64(i64 %23, i64 undef) 33 %26 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %25, i64 0) 41 %34 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 0, i64 undef) 43 %36 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %35, i64 1) 48 %41 = call { i64, i1 } @llvm.smul.with.overflow.i64(i64 %39, i64 %40) 50 %43 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %42, i64 0) [all …]
|
| /external/llvm-project/llvm/test/CodeGen/X86/ |
| D | overflow-intrinsic-setcc-fold.ll | 11 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %v2) 24 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 %v2) 37 %t = call {i32, i1} @llvm.uadd.with.overflow.i32(i32 %v1, i32 %v2) 50 %t = call {i64, i1} @llvm.uadd.with.overflow.i64(i64 %v1, i64 %v2) 63 %t = call {i32, i1} @llvm.ssub.with.overflow.i32(i32 %v1, i32 %v2) 76 %t = call {i64, i1} @llvm.ssub.with.overflow.i64(i64 %v1, i64 %v2) 89 %t = call {i32, i1} @llvm.usub.with.overflow.i32(i32 %v1, i32 %v2) 102 %t = call {i64, i1} @llvm.usub.with.overflow.i64(i64 %v1, i64 %v2) 115 %t = call {i32, i1} @llvm.smul.with.overflow.i32(i32 %v1, i32 %v2) 128 %t = call {i64, i1} @llvm.smul.with.overflow.i64(i64 %v1, i64 %v2) [all …]
|
| D | smul-with-overflow.ll | 8 %t = call {i32, i1} @llvm.smul.with.overflow.i32(i32 %v1, i32 %v2) 11 br i1 %obit, label %overflow, label %normal 17 overflow: 27 %t = call {i32, i1} @llvm.smul.with.overflow.i32(i32 %v1, i32 %v2) 30 br i1 %obit, label %overflow, label %normal 32 overflow: 45 declare {i32, i1} @llvm.smul.with.overflow.i32(i32, i32) 50 %tmp1 = call { i32, i1 } @llvm.smul.with.overflow.i32(i32 %tmp0, i32 2) 62 %tmp1 = call { i32, i1 } @llvm.smul.with.overflow.i32(i32 %tmp0, i32 4) 71 declare { i63, i1 } @llvm.smul.with.overflow.i63(i63, i63) nounwind readnone [all …]
|
| /external/llvm/test/Transforms/GVN/ |
| D | 2011-07-07-MatchIntrinsicExtract.ll | 8 %uadd = tail call %0 @llvm.uadd.with.overflow.i64(i64 %a, i64 %b) 20 %usub = tail call %0 @llvm.usub.with.overflow.i64(i64 %a, i64 %b) 32 %umul = tail call %0 @llvm.umul.with.overflow.i64(i64 %a, i64 %b) 44 %sadd = tail call %0 @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) 56 %ssub = tail call %0 @llvm.ssub.with.overflow.i64(i64 %a, i64 %b) 68 %smul = tail call %0 @llvm.smul.with.overflow.i64(i64 %a, i64 %b) 79 declare %0 @llvm.uadd.with.overflow.i64(i64, i64) nounwind readnone 80 declare %0 @llvm.usub.with.overflow.i64(i64, i64) nounwind readnone 81 declare %0 @llvm.umul.with.overflow.i64(i64, i64) nounwind readnone 82 declare %0 @llvm.sadd.with.overflow.i64(i64, i64) nounwind readnone [all …]
|