/external/llvm/test/CodeGen/AMDGPU/ |
D | usubo.ll | 5 declare { i32, i1 } @llvm.usub.with.overflow.i32(i32, i32) nounwind readnone 6 declare { i64, i1 } @llvm.usub.with.overflow.i64(i64, i64) nounwind readnone 13 %usub = call { i64, i1 } @llvm.usub.with.overflow.i64(i64 %a, i64 %b) nounwind 14 %val = extractvalue { i64, i1 } %usub, 0 15 %carry = extractvalue { i64, i1 } %usub, 1 28 %usub = call { i32, i1 } @llvm.usub.with.overflow.i32(i32 %a, i32 %b) nounwind 29 %val = extractvalue { i32, i1 } %usub, 0 30 %carry = extractvalue { i32, i1 } %usub, 1 44 %usub = call { i32, i1 } @llvm.usub.with.overflow.i32(i32 %a, i32 %b) nounwind 45 %val = extractvalue { i32, i1 } %usub, 0 [all …]
|
/external/swiftshader/third_party/LLVM/test/CodeGen/Generic/ |
D | overflow.ll | 113 ;; usub 117 %usub = tail call { i8, i1 } @llvm.usub.with.overflow.i8(i8 %a, i8 %b) 118 %cmp = extractvalue { i8, i1 } %usub, 1 119 %usub.result = extractvalue { i8, i1 } %usub, 0 120 %X = select i1 %cmp, i8 %usub.result, i8 42 124 declare { i8, i1 } @llvm.usub.with.overflow.i8(i8, i8) nounwind readnone 128 %usub = tail call { i16, i1 } @llvm.usub.with.overflow.i16(i16 %a, i16 %b) 129 %cmp = extractvalue { i16, i1 } %usub, 1 130 %usub.result = extractvalue { i16, i1 } %usub, 0 131 %X = select i1 %cmp, i16 %usub.result, i16 42 [all …]
|
/external/llvm/test/CodeGen/Generic/ |
D | overflow.ll | 113 ;; usub 117 %usub = tail call { i8, i1 } @llvm.usub.with.overflow.i8(i8 %a, i8 %b) 118 %cmp = extractvalue { i8, i1 } %usub, 1 119 %usub.result = extractvalue { i8, i1 } %usub, 0 120 %X = select i1 %cmp, i8 %usub.result, i8 42 124 declare { i8, i1 } @llvm.usub.with.overflow.i8(i8, i8) nounwind readnone 128 %usub = tail call { i16, i1 } @llvm.usub.with.overflow.i16(i16 %a, i16 %b) 129 %cmp = extractvalue { i16, i1 } %usub, 1 130 %usub.result = extractvalue { i16, i1 } %usub, 0 131 %X = select i1 %cmp, i16 %usub.result, i16 42 [all …]
|
/external/swiftshader/third_party/LLVM/test/Transforms/GVN/ |
D | 2011-07-07-MatchIntrinsicExtract.ll | 20 %usub = tail call %0 @llvm.usub.with.overflow.i64(i64 %a, i64 %b) 21 %usub.0 = extractvalue %0 %usub, 0 80 declare %0 @llvm.usub.with.overflow.i64(i64, i64) nounwind readnone
|
/external/llvm/test/Transforms/GVN/ |
D | 2011-07-07-MatchIntrinsicExtract.ll | 20 %usub = tail call %0 @llvm.usub.with.overflow.i64(i64 %a, i64 %b) 21 %usub.0 = extractvalue %0 %usub, 0 80 declare %0 @llvm.usub.with.overflow.i64(i64, i64) nounwind readnone
|
/external/autotest/client/site_tests/graphics_GpuReset/src/ |
D | gpureset.c | 115 const char *usub, *dnode; in drm_open_matching() local 130 usub = udev_device_get_subsystem(parent); in drm_open_matching() 132 if (!usub || (strcmp(usub, "pci") != 0)) in drm_open_matching()
|
/external/eigen/Eigen/src/SparseLU/ |
D | SparseLU_copy_to_ucol.h | 80 mem = memXpand<IndexVector>(glu.usub, glu.nzumax, nextu, USUB, glu.num_expansions); in copy_to_ucol() 88 … glu.usub(nextu) = perm_r(irow); // Unlike the L part, the U part is stored in its final order in copy_to_ucol()
|
D | SparseLU_Memory.h | 183 || (expand<IndexVector> (glu.usub, glu.nzumax, 0, 1, num_expansions)<0) ) in memInit() 191 } while (!glu.lusup.size() || !glu.ucol.size() || !glu.lsub.size() || !glu.usub.size()); in memInit()
|
D | SparseLU_Structs.h | 88 IndexVector usub; // row indices of U columns in ucol member
|
D | SparseLU.h | 697 …Scalar, ColMajor, StorageIndex> ( m, n, m_nnzU, m_glu.xusub.data(), m_glu.usub.data(), m_glu.ucol.… in factorize()
|
/external/swiftshader/third_party/LLVM/test/Transforms/ConstProp/ |
D | overflow-ops.ll | 5 declare {i8, i1} @llvm.usub.with.overflow.i8(i8, i8) 35 ;; usub 40 %t = call {i8, i1} @llvm.usub.with.overflow.i8(i8 4, i8 2) 49 %t = call {i8, i1} @llvm.usub.with.overflow.i8(i8 4, i8 6)
|
/external/llvm/test/Transforms/ConstProp/ |
D | overflow-ops.ll | 5 declare {i8, i1} @llvm.usub.with.overflow.i8(i8, i8) 35 ;; usub 40 %t = call {i8, i1} @llvm.usub.with.overflow.i8(i8 4, i8 2) 49 %t = call {i8, i1} @llvm.usub.with.overflow.i8(i8 4, i8 6)
|
/external/llvm/test/CodeGen/ARM/ |
D | intrinsics-overflow.ll | 31 %sadd = tail call { i32, i1 } @llvm.usub.with.overflow.i32(i32 %a, i32 %b) 56 declare { i32, i1 } @llvm.usub.with.overflow.i32(i32, i32) #3
|
/external/swiftshader/third_party/LLVM/test/CodeGen/X86/ |
D | sub-with-overflow.ll | 28 %t = call {i32, i1} @llvm.usub.with.overflow.i32(i32 %v1, i32 %v2) 48 declare {i32, i1} @llvm.usub.with.overflow.i32(i32, i32)
|
/external/llvm/test/CodeGen/X86/ |
D | sub-with-overflow.ll | 28 %t = call {i32, i1} @llvm.usub.with.overflow.i32(i32 %v1, i32 %v2) 48 declare {i32, i1} @llvm.usub.with.overflow.i32(i32, i32)
|
D | xaluo.ll | 275 %t = call {i32, i1} @llvm.usub.with.overflow.i32(i32 %v1, i32 %v2) 287 %t = call {i64, i1} @llvm.usub.with.overflow.i64(i64 %v1, i64 %v2) 468 %t = call {i32, i1} @llvm.usub.with.overflow.i32(i32 %v1, i32 %v2) 479 %t = call {i64, i1} @llvm.usub.with.overflow.i64(i64 %v1, i64 %v2) 640 %t = call {i32, i1} @llvm.usub.with.overflow.i32(i32 %v1, i32 %v2) 657 %t = call {i64, i1} @llvm.usub.with.overflow.i64(i64 %v1, i64 %v2) 767 declare {i32, i1} @llvm.usub.with.overflow.i32(i32, i32) nounwind readnone 768 declare {i64, i1} @llvm.usub.with.overflow.i64(i64, i64) nounwind readnone
|
/external/llvm/test/CodeGen/AArch64/ |
D | arm64-xaluo.ll | 173 %t = call {i32, i1} @llvm.usub.with.overflow.i32(i32 %v1, i32 %v2) 185 %t = call {i64, i1} @llvm.usub.with.overflow.i64(i64 %v1, i64 %v2) 345 %t = call {i32, i1} @llvm.usub.with.overflow.i32(i32 %v1, i32 %v2) 356 %t = call {i64, i1} @llvm.usub.with.overflow.i64(i64 %v1, i64 %v2) 523 %t = call {i32, i1} @llvm.usub.with.overflow.i32(i32 %v1, i32 %v2) 540 %t = call {i64, i1} @llvm.usub.with.overflow.i64(i64 %v1, i64 %v2) 665 declare {i32, i1} @llvm.usub.with.overflow.i32(i32, i32) nounwind readnone 666 declare {i64, i1} @llvm.usub.with.overflow.i64(i64, i64) nounwind readnone
|
/external/llvm/test/Transforms/IndVarSimplify/ |
D | overflow-intrinsics.ll | 109 %0 = tail call { i32, i1 } @llvm.usub.with.overflow.i32(i32 %i.04, i32 1) 133 declare { i32, i1 } @llvm.usub.with.overflow.i32(i32, i32) nounwind readnone
|
/external/llvm/test/Transforms/InstSimplify/ |
D | call.ll | 5 declare {i8, i1} @llvm.usub.with.overflow.i8(i8 %a, i8 %b) 27 %x = call {i8, i1} @llvm.usub.with.overflow.i8(i8 %V, i8 %V)
|
/external/llvm/test/Analysis/ScalarEvolution/ |
D | overflow-intrinsics.ll | 269 %tmp0 = tail call { i32, i1 } @llvm.usub.with.overflow.i32(i32 %i.04, i32 1) 305 declare { i32, i1 } @llvm.usub.with.overflow.i32(i32, i32) nounwind readnone
|
/external/llvm/test/Transforms/InstCombine/ |
D | intrinsics.ll | 12 declare %ov.result.32 @llvm.usub.with.overflow.i32(i32, i32) nounwind readnone 144 %x = call %ov.result.32 @llvm.usub.with.overflow.i32(i32 %A, i32 %B)
|
/external/python/cpython2/Lib/test/ |
D | test_unicode.py | 65 class usub(unicode): class 68 object = usub(object)
|
/external/llvm/docs/ |
D | LangRef.rst | 11065 '``llvm.usub.with.overflow.*``' Intrinsics 11071 This is an overloaded intrinsic. You can use ``llvm.usub.with.overflow`` 11076 declare {i16, i1} @llvm.usub.with.overflow.i16(i16 %a, i16 %b) 11077 declare {i32, i1} @llvm.usub.with.overflow.i32(i32 %a, i32 %b) 11078 declare {i64, i1} @llvm.usub.with.overflow.i64(i64 %a, i64 %b) 11083 The '``llvm.usub.with.overflow``' family of intrinsic functions perform 11099 The '``llvm.usub.with.overflow``' family of intrinsic functions perform 11110 %res = call {i32, i1} @llvm.usub.with.overflow.i32(i32 %a, i32 %b)
|
/external/swiftshader/third_party/LLVM/include/llvm/ |
D | Intrinsics.gen | 96 usub_with_overflow, // llvm.usub.with.overflow 623 "llvm.usub.with.overflow", 5577 case Intrinsic::usub_with_overflow: // llvm.usub.with.overflow 6653 case Intrinsic::usub_with_overflow: // llvm.usub.with.overflow
|
/external/swiftshader/third_party/llvm-subzero/build/Android/include/llvm/IR/ |
D | Intrinsics.gen | 156 usub_with_overflow, // llvm.usub.with.overflow 6214 "llvm.usub.with.overflow", 14154 1, // llvm.usub.with.overflow
|