/external/swiftshader/third_party/LLVM/test/Transforms/InstCombine/ |
D | intrinsics.ll | 3 %overflow.result = type {i8, i1} 5 declare %overflow.result @llvm.uadd.with.overflow.i8(i8, i8) 6 declare %overflow.result @llvm.umul.with.overflow.i8(i8, i8) 14 %x = call %overflow.result @llvm.uadd.with.overflow.i8(i8 %A, i8 %B) 15 %y = extractvalue %overflow.result %x, 0 25 %x = call %overflow.result @llvm.uadd.with.overflow.i8(i8 %and.A, i8 %and.B) 26 %y = extractvalue %overflow.result %x, 0 27 %z = extractvalue %overflow.result %x, 1 41 %x = call %overflow.result @llvm.uadd.with.overflow.i8(i8 %or.A, i8 %or.B) 42 %y = extractvalue %overflow.result %x, 0 [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | xaluo.ll | 5 ; Get the actual value of the overflow bit. 13 %t = call {i8, i1} @llvm.sadd.with.overflow.i8(i8 %v1, i8 %v2) 25 %t = call {i16, i1} @llvm.sadd.with.overflow.i16(i16 %v1, i16 %v2) 37 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %v2) 49 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 %v2) 62 %t = call {i8, i1} @llvm.sadd.with.overflow.i8(i8 %v1, i8 1) 74 %t = call {i16, i1} @llvm.sadd.with.overflow.i16(i16 %v1, i16 1) 86 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 1) 98 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 1) 116 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 2, i64 %v1) [all …]
|
D | smul-with-overflow.ll | 8 %t = call {i32, i1} @llvm.smul.with.overflow.i32(i32 %v1, i32 %v2) 11 br i1 %obit, label %overflow, label %normal 17 overflow: 27 %t = call {i32, i1} @llvm.smul.with.overflow.i32(i32 %v1, i32 %v2) 30 br i1 %obit, label %overflow, label %normal 32 overflow: 45 declare {i32, i1} @llvm.smul.with.overflow.i32(i32, i32) 50 %tmp1 = call { i32, i1 } @llvm.smul.with.overflow.i32(i32 %tmp0, i32 2) 62 %tmp1 = call { i32, i1 } @llvm.smul.with.overflow.i32(i32 %tmp0, i32 4) 71 declare { i63, i1 } @llvm.smul.with.overflow.i63(i63, i63) nounwind readnone [all …]
|
/external/llvm/test/CodeGen/AArch64/ |
D | arm64-xaluo.ll | 5 ; Get the actual value of the overflow bit. 12 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %v2) 25 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 4) 38 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 -4) 51 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 16777215) 65 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %lsl) 77 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 %v2) 89 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 4) 101 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %v1, i64 -4) 113 %t = call {i32, i1} @llvm.uadd.with.overflow.i32(i32 %v1, i32 %v2) [all …]
|
/external/python/cpython3/Modules/_decimal/libmpdec/ |
D | memory.c | 58 mpd_size_t overflow; in mpd_callocfunc_em() local 68 &overflow); in mpd_callocfunc_em() 69 if (overflow) { in mpd_callocfunc_em() 88 mpd_size_t req, overflow; in mpd_alloc() local 90 req = mul_size_t_overflow(nmemb, size, &overflow); in mpd_alloc() 91 if (overflow) { in mpd_alloc() 102 mpd_size_t overflow; in mpd_calloc() local 104 (void)mul_size_t_overflow(nmemb, size, &overflow); in mpd_calloc() 105 if (overflow) { in mpd_calloc() 117 mpd_size_t req, overflow; in mpd_realloc() local [all …]
|
/external/llvm/test/Transforms/ConstProp/ |
D | overflow-ops.ll | 4 declare {i8, i1} @llvm.uadd.with.overflow.i8(i8, i8) 5 declare {i8, i1} @llvm.usub.with.overflow.i8(i8, i8) 6 declare {i8, i1} @llvm.umul.with.overflow.i8(i8, i8) 8 declare {i8, i1} @llvm.sadd.with.overflow.i8(i8, i8) 9 declare {i8, i1} @llvm.ssub.with.overflow.i8(i8, i8) 10 declare {i8, i1} @llvm.smul.with.overflow.i8(i8, i8) 18 %t = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 42, i8 100) 27 %t = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 142, i8 120) 40 %t = call {i8, i1} @llvm.usub.with.overflow.i8(i8 4, i8 2) 49 %t = call {i8, i1} @llvm.usub.with.overflow.i8(i8 4, i8 6) [all …]
|
/external/swiftshader/third_party/LLVM/test/Transforms/ConstProp/ |
D | overflow-ops.ll | 4 declare {i8, i1} @llvm.uadd.with.overflow.i8(i8, i8) 5 declare {i8, i1} @llvm.usub.with.overflow.i8(i8, i8) 6 declare {i8, i1} @llvm.umul.with.overflow.i8(i8, i8) 8 declare {i8, i1} @llvm.sadd.with.overflow.i8(i8, i8) 9 declare {i8, i1} @llvm.ssub.with.overflow.i8(i8, i8) 10 declare {i8, i1} @llvm.smul.with.overflow.i8(i8, i8) 18 %t = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 42, i8 100) 27 %t = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 142, i8 120) 40 %t = call {i8, i1} @llvm.usub.with.overflow.i8(i8 4, i8 2) 49 %t = call {i8, i1} @llvm.usub.with.overflow.i8(i8 4, i8 6) [all …]
|
/external/llvm/test/CodeGen/Generic/ |
D | overflow.ll | 2 ; Verify codegen's don't crash on overflow intrinsics. 8 %sadd = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 %a, i8 %b) 15 declare { i8, i1 } @llvm.sadd.with.overflow.i8(i8, i8) nounwind readnone 19 %sadd = tail call { i16, i1 } @llvm.sadd.with.overflow.i16(i16 %a, i16 %b) 26 declare { i16, i1 } @llvm.sadd.with.overflow.i16(i16, i16) nounwind readnone 30 %sadd = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 37 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) nounwind readnone 44 %uadd = tail call { i8, i1 } @llvm.uadd.with.overflow.i8(i8 %a, i8 %b) 51 declare { i8, i1 } @llvm.uadd.with.overflow.i8(i8, i8) nounwind readnone 55 %uadd = tail call { i16, i1 } @llvm.uadd.with.overflow.i16(i16 %a, i16 %b) [all …]
|
D | add-with-overflow.ll | 9 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %v2) 12 br i1 %obit, label %overflow, label %normal 18 overflow: 25 %t = call {i32, i1} @llvm.uadd.with.overflow.i32(i32 %v1, i32 %v2) 28 br i1 %obit, label %overflow, label %normal 34 overflow: 40 declare {i32, i1} @llvm.sadd.with.overflow.i32(i32, i32) 41 declare {i32, i1} @llvm.uadd.with.overflow.i32(i32, i32)
|
/external/swiftshader/third_party/LLVM/test/CodeGen/Generic/ |
D | overflow.ll | 2 ; Verify codegen's don't crash on overflow intrinsics. 8 %sadd = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 %a, i8 %b) 15 declare { i8, i1 } @llvm.sadd.with.overflow.i8(i8, i8) nounwind readnone 19 %sadd = tail call { i16, i1 } @llvm.sadd.with.overflow.i16(i16 %a, i16 %b) 26 declare { i16, i1 } @llvm.sadd.with.overflow.i16(i16, i16) nounwind readnone 30 %sadd = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 37 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) nounwind readnone 44 %uadd = tail call { i8, i1 } @llvm.uadd.with.overflow.i8(i8 %a, i8 %b) 51 declare { i8, i1 } @llvm.uadd.with.overflow.i8(i8, i8) nounwind readnone 55 %uadd = tail call { i16, i1 } @llvm.uadd.with.overflow.i16(i16 %a, i16 %b) [all …]
|
D | add-with-overflow.ll | 9 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %v1, i32 %v2) 12 br i1 %obit, label %overflow, label %normal 18 overflow: 25 %t = call {i32, i1} @llvm.uadd.with.overflow.i32(i32 %v1, i32 %v2) 28 br i1 %obit, label %overflow, label %normal 34 overflow: 40 declare {i32, i1} @llvm.sadd.with.overflow.i32(i32, i32) 41 declare {i32, i1} @llvm.uadd.with.overflow.i32(i32, i32)
|
/external/llvm/test/Transforms/InstCombine/ |
D | overflow-mul.ll | 11 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y) 12 %overflow = icmp ugt i64 %mul64, 4294967295 14 %retval = zext i1 %overflow to i32 26 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y) 27 %overflow = icmp uge i64 %mul64, 4294967296 29 %retval = zext i1 %overflow to i32 42 ; CHECK: [[MUL:%.*]] = call { i32, i1 } @llvm.umul.with.overflow.i32(i32 %x, i32 %y) 43 %overflow = icmp ugt i64 %mul64, 4294967295 47 %retval = select i1 %overflow, i32 %mul32, i32 111 60 ; CHECK-NOT: umul.with.overflow.i32 [all …]
|
D | intrinsics.ll | 3 %overflow.result = type {i8, i1} 7 declare %overflow.result @llvm.uadd.with.overflow.i8(i8, i8) nounwind readnone 8 declare %overflow.result @llvm.umul.with.overflow.i8(i8, i8) nounwind readnone 9 declare %ov.result.32 @llvm.sadd.with.overflow.i32(i32, i32) nounwind readnone 10 declare %ov.result.32 @llvm.uadd.with.overflow.i32(i32, i32) nounwind readnone 11 declare %ov.result.32 @llvm.ssub.with.overflow.i32(i32, i32) nounwind readnone 12 declare %ov.result.32 @llvm.usub.with.overflow.i32(i32, i32) nounwind readnone 13 declare %ov.result.32 @llvm.smul.with.overflow.i32(i32, i32) nounwind readnone 14 declare %ov.result.32 @llvm.umul.with.overflow.i32(i32, i32) nounwind readnone 29 %x = call %overflow.result @llvm.uadd.with.overflow.i8(i8 %A, i8 %B) [all …]
|
/external/llvm/test/tools/llvm-profdata/ |
D | overflow-instr.test | 1 Tests for overflow when merging instrumented profiles. 3 1- Merge profile having maximum counts with itself and verify overflow detected and saturation occu… 4 RUN: llvm-profdata merge -instr %p/Inputs/overflow-instr.proftext %p/Inputs/overflow-instr.proftext… 6 MERGE_OVERFLOW: {{.*}}: overflow: Counter overflow 11 2- Merge profile having maximum counts by itself and verify no overflow 12 RUN: llvm-profdata merge -instr %p/Inputs/overflow-instr.proftext -o %t.out 2>&1 | FileCheck %s -ch… 14 MERGE_NO_OVERFLOW-NOT: {{.*}}: overflow: Counter overflow
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | fwd_txfm_sse2.h | 176 int overflow = k_check_epi32_overflow_4(preg0, preg1, preg2, preg3, zero); in k_check_epi32_overflow_8() local 177 if (!overflow) { in k_check_epi32_overflow_8() 178 overflow = k_check_epi32_overflow_4(preg4, preg5, preg6, preg7, zero); in k_check_epi32_overflow_8() 180 return overflow; in k_check_epi32_overflow_8() 190 int overflow = k_check_epi32_overflow_4(preg0, preg1, preg2, preg3, zero); in k_check_epi32_overflow_16() local 191 if (!overflow) { in k_check_epi32_overflow_16() 192 overflow = k_check_epi32_overflow_4(preg4, preg5, preg6, preg7, zero); in k_check_epi32_overflow_16() 193 if (!overflow) { in k_check_epi32_overflow_16() 194 overflow = k_check_epi32_overflow_4(preg8, preg9, preg10, preg11, zero); in k_check_epi32_overflow_16() 195 if (!overflow) { in k_check_epi32_overflow_16() [all …]
|
D | fwd_txfm_impl_sse2.h | 81 int test, overflow; in FDCT4x4_2D() local 170 overflow = check_epi16_overflow_x2(&x0, &x1); in FDCT4x4_2D() 171 if (overflow) { in FDCT4x4_2D() 192 overflow = check_epi16_overflow_x2(&t0, &t1); in FDCT4x4_2D() 193 if (overflow) { in FDCT4x4_2D() 231 overflow = check_epi16_overflow_x2(&x0, &x1); in FDCT4x4_2D() 232 if (overflow) { in FDCT4x4_2D() 274 int overflow; in FDCT8x8_2D() local 313 overflow = in FDCT8x8_2D() 315 if (overflow) { in FDCT8x8_2D() [all …]
|
/external/arm-neon-tests/ |
D | ref-rvct-all.txt | 2243 VRSHL/VRSHLQ (checking round_const overflow: shift by -1) output: 2269 VRSHL/VRSHLQ (checking round_const overflow: shift by -3) output: 4639 VABAL test intermediate overflow output: 6207 VRSHR_N (overflow test: max shift amount, positive input) output: 6233 VRSHR_N (overflow test: shift by 1, with negative input) output: 6259 VRSHR_N (overflow test: shift by 3, positive input) output: 6285 VRSHR_N (overflow test: shift by 1, with negative input) output: 6311 VRSHR_N (overflow test: shift by 3, with negative input) output: 6363 VRSRA_N (checking overflow: shift by 1, positive input) output: 6389 VRSRA_N (checking overflow: shift by 3, positive input) output: [all …]
|
/external/swiftshader/third_party/LLVM/test/Transforms/GVN/ |
D | 2011-07-07-MatchIntrinsicExtract.ll | 8 %uadd = tail call %0 @llvm.uadd.with.overflow.i64(i64 %a, i64 %b) 20 %usub = tail call %0 @llvm.usub.with.overflow.i64(i64 %a, i64 %b) 32 %umul = tail call %0 @llvm.umul.with.overflow.i64(i64 %a, i64 %b) 44 %sadd = tail call %0 @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) 56 %ssub = tail call %0 @llvm.ssub.with.overflow.i64(i64 %a, i64 %b) 68 %smul = tail call %0 @llvm.smul.with.overflow.i64(i64 %a, i64 %b) 79 declare %0 @llvm.uadd.with.overflow.i64(i64, i64) nounwind readnone 80 declare %0 @llvm.usub.with.overflow.i64(i64, i64) nounwind readnone 81 declare %0 @llvm.umul.with.overflow.i64(i64, i64) nounwind readnone 82 declare %0 @llvm.sadd.with.overflow.i64(i64, i64) nounwind readnone [all …]
|
/external/llvm/test/Transforms/GVN/ |
D | 2011-07-07-MatchIntrinsicExtract.ll | 8 %uadd = tail call %0 @llvm.uadd.with.overflow.i64(i64 %a, i64 %b) 20 %usub = tail call %0 @llvm.usub.with.overflow.i64(i64 %a, i64 %b) 32 %umul = tail call %0 @llvm.umul.with.overflow.i64(i64 %a, i64 %b) 44 %sadd = tail call %0 @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) 56 %ssub = tail call %0 @llvm.ssub.with.overflow.i64(i64 %a, i64 %b) 68 %smul = tail call %0 @llvm.smul.with.overflow.i64(i64 %a, i64 %b) 79 declare %0 @llvm.uadd.with.overflow.i64(i64, i64) nounwind readnone 80 declare %0 @llvm.usub.with.overflow.i64(i64, i64) nounwind readnone 81 declare %0 @llvm.umul.with.overflow.i64(i64, i64) nounwind readnone 82 declare %0 @llvm.sadd.with.overflow.i64(i64, i64) nounwind readnone [all …]
|
/external/libgsm/add-test/ |
D | add_test.dta | 36 ' positive overflow 76 ' positive overflow 115 ' positive overflow 155 ' positive overflow 184 ' overflow 212 ' overflow 248 ' overflow 289 ' overflow 311 ' overflow 359 ' overflow [all …]
|
/external/clang/test/SemaCXX/ |
D | integer-overflow.cpp | 14 static const uint64_t overflow = 1 * 4608 * 1024 * 1024; // expected-warning {{overflow in expressi… variable 18 uint64_t overflow = 4608 * 1024 * 1024, in check_integer_overflows() local 31 overflow += overflow2 = overflow3 = (uint64_t)(4608 * 1024 * 1024); in check_integer_overflows() 33 overflow += overflow2 = overflow3 = 4608 * 1024 * 1024; in check_integer_overflows() 36 overflow += overflow2 = overflow3 = static_cast<uint64_t>(4608 * 1024 * 1024); in check_integer_overflows() 42 overflow = 4608 * 1024 * 1024 ? 4608 * 1024 * 1024 : 0; in check_integer_overflows() 45 overflow = 0 ? 0 : 4608 * 1024 * 1024; in check_integer_overflows()
|
/external/compiler-rt/lib/builtins/ |
D | mulodi4.c | 22 __mulodi4(di_int a, di_int b, int* overflow) in __mulodi4() argument 27 *overflow = 0; in __mulodi4() 32 *overflow = 1; in __mulodi4() 38 *overflow = 1; in __mulodi4() 50 *overflow = 1; in __mulodi4() 55 *overflow = 1; in __mulodi4()
|
D | mulosi4.c | 22 __mulosi4(si_int a, si_int b, int* overflow) in __mulosi4() argument 27 *overflow = 0; in __mulosi4() 32 *overflow = 1; in __mulosi4() 38 *overflow = 1; in __mulosi4() 50 *overflow = 1; in __mulosi4() 55 *overflow = 1; in __mulosi4()
|
D | muloti4.c | 24 __muloti4(ti_int a, ti_int b, int* overflow) in __muloti4() argument 29 *overflow = 0; in __muloti4() 34 *overflow = 1; in __muloti4() 40 *overflow = 1; in __muloti4() 52 *overflow = 1; in __muloti4() 57 *overflow = 1; in __muloti4()
|
/external/python/cpython2/Misc/Vim/ |
D | vim_syntax.py | 132 overflow = None 137 if overflow: 138 buffer_.append(overflow) 139 total_len += len(overflow) + 1 140 overflow = None 149 if overflow: 150 yield overflow 153 overflow = buffer_.pop() 154 total_len -= len(overflow) - 1
|