/external/swiftshader/third_party/llvm-7.0/llvm/test/Analysis/BasicAA/ |
D | sequential-gep.ll | 5 define void @t1([8 x i32]* %p, i32 %addend, i32* %q) { 7 %add = add nsw nuw i32 %addend, %knownnonzero 8 %gep1 = getelementptr [8 x i32], [8 x i32]* %p, i32 2, i32 %addend 15 define void @t2([8 x i32]* %p, i32 %addend, i32* %q) { 17 %add = add nsw nuw i32 %addend, %knownnonzero 18 %gep1 = getelementptr [8 x i32], [8 x i32]* %p, i32 1, i32 %addend 25 define void @t3([8 x i32]* %p, i32 %addend, i32* %q) { 27 %add = add nsw nuw i32 %addend, %knownnonzero 35 define void @t4([8 x i32]* %p, i32 %addend, i32* %q) { 37 %add = add nsw nuw i32 %addend, %knownnonzero [all …]
|
/external/llvm/test/Analysis/BasicAA/ |
D | sequential-gep.ll | 5 define void @t1([8 x i32]* %p, i32 %addend, i32* %q) { 7 %add = add nsw nuw i32 %addend, %knownnonzero 8 %gep1 = getelementptr [8 x i32], [8 x i32]* %p, i32 2, i32 %addend 15 define void @t2([8 x i32]* %p, i32 %addend, i32* %q) { 17 %add = add nsw nuw i32 %addend, %knownnonzero 18 %gep1 = getelementptr [8 x i32], [8 x i32]* %p, i32 1, i32 %addend 25 define void @t3([8 x i32]* %p, i32 %addend, i32* %q) { 27 %add = add nsw nuw i32 %addend, %knownnonzero 35 define void @t4([8 x i32]* %p, i32 %addend, i32* %q) { 37 %add = add nsw nuw i32 %addend, %knownnonzero [all …]
|
/external/google-breakpad/src/common/ |
D | test_assembler.cc | 67 Label Label::operator+(uint64_t addend) const { in operator +() 69 l.value_->Set(this->value_, addend); in operator +() 106 uint64_t addend; in IsKnownConstant() local 107 value_->Get(&base, &addend); in IsKnownConstant() 109 if (value_p) *value_p = addend; in IsKnownConstant() 129 Label::Binding::Binding(uint64_t addend) in Binding() argument 130 : base_(NULL), addend_(addend), reference_count_(1) { } in Binding() 138 void Label::Binding::Set(Binding *binding, uint64_t addend) { in Set() argument 141 assert(addend_ == addend); in Set() 145 binding->Set(NULL, addend_ - addend); in Set() [all …]
|
/external/swiftshader/third_party/subzero/tests_lit/llvm2ice_tests/ |
D | vector-align.ll | 16 define internal <4 x i32> @test_add(i32 %addr_i, <4 x i32> %addend) { 20 %result = add <4 x i32> %addend, %loaded 33 define internal <4 x i32> @test_and(i32 %addr_i, <4 x i32> %addend) { 37 %result = and <4 x i32> %addend, %loaded 50 define internal <4 x i32> @test_or(i32 %addr_i, <4 x i32> %addend) { 54 %result = or <4 x i32> %addend, %loaded 67 define internal <4 x i32> @test_xor(i32 %addr_i, <4 x i32> %addend) { 71 %result = xor <4 x i32> %addend, %loaded 84 define internal <4 x i32> @test_sub(i32 %addr_i, <4 x i32> %addend) { 88 %result = sub <4 x i32> %addend, %loaded [all …]
|
/external/brotli/c/fuzz/ |
D | decode_fuzzer.c | 13 size_t addend = 0; in LLVMFuzzerTestOneInput() local 15 addend = data[size - 1] & 7; in LLVMFuzzerTestOneInput() 26 const size_t total_out_limit = (addend == 0) ? (1 << 26) : (1 << 24); in LLVMFuzzerTestOneInput() 31 if (addend == 0) in LLVMFuzzerTestOneInput() 32 addend = size; in LLVMFuzzerTestOneInput() 35 size_t next_i = i + addend; in LLVMFuzzerTestOneInput()
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/LICM/ |
D | dropped-tbaa.ll | 6 ; void add(double *restrict data, int *restrict addend) { 7 ; *data += *addend; 10 ; void foo(double *data, int *addend) { 12 ; *data += *addend; 13 ; add(data, addend); 16 ; We want to make sure the load of addend gets hoisted, independent of the second load 19 define void @foo(double* %data, i32* %addend) #0 { 44 %2 = load i32, i32* %addend, align 4, !tbaa !1 54 %5 = load i32, i32* %addend, align 4, !tbaa !1, !alias.scope !7, !noalias !10 87 !8 = distinct !{!8, !9, !"add: %addend"}
|
/external/swiftshader/third_party/LLVM/test/CodeGen/ARM/ |
D | fast-isel-static.ll | 4 define void @myadd(float* %sum, float* %addend) nounwind { 7 %addend.addr = alloca float*, align 4 9 store float* %addend, float** %addend.addr, align 4 12 %tmp2 = load float** %addend.addr, align 4
|
/external/jemalloc_new/src/ |
D | nstime.c | 52 nstime_add(nstime_t *time, const nstime_t *addend) { in nstime_add() argument 53 assert(UINT64_MAX - time->ns >= addend->ns); in nstime_add() 55 time->ns += addend->ns; in nstime_add() 59 nstime_iadd(nstime_t *time, uint64_t addend) { in nstime_iadd() argument 60 assert(UINT64_MAX - time->ns >= addend); in nstime_iadd() 62 time->ns += addend; in nstime_iadd()
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/ARM/ |
D | fast-isel-static.ll | 6 define void @myadd(float* %sum, float* %addend) nounwind { 9 %addend.addr = alloca float*, align 4 11 store float* %addend, float** %addend.addr, align 4 14 %tmp2 = load float*, float** %addend.addr, align 4
|
/external/llvm/test/CodeGen/ARM/ |
D | fast-isel-static.ll | 6 define void @myadd(float* %sum, float* %addend) nounwind { 9 %addend.addr = alloca float*, align 4 11 store float* %addend, float** %addend.addr, align 4 14 %tmp2 = load float*, float** %addend.addr, align 4
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/LoopStrengthReduce/ |
D | sext-ind-var.ll | 33 %addend = load float, float* %ptr, align 4 34 %nextsum = fadd float %sum, %addend 65 %addend = load float, float* %ptr, align 4 66 %nextsum = fadd float %sum, %addend 97 %addend = load float, float* %ptr, align 4 98 %nextsum = fadd float %sum, %addend 131 %addend = load float, float* %ptr, align 4 132 %nextsum = fadd float %sum, %addend
|
/external/grpc-grpc/test/core/gpr/ |
D | time_test.cc | 185 gpr_timespec addend[3]; in test_sticky_infinities() local 188 addend[0] = gpr_inf_future(GPR_TIMESPAN); in test_sticky_infinities() 189 addend[1] = gpr_inf_past(GPR_TIMESPAN); in test_sticky_infinities() 190 addend[2] = gpr_time_0(GPR_TIMESPAN); in test_sticky_infinities() 194 for (j = 0; j != sizeof(addend) / sizeof(addend[0]); j++) { in test_sticky_infinities() 195 gpr_timespec x = gpr_time_add(infinity[i], addend[j]); in test_sticky_infinities() 197 x = gpr_time_sub(infinity[i], addend[j]); in test_sticky_infinities()
|
/external/llvm/test/Transforms/LoopStrengthReduce/ |
D | sext-ind-var.ll | 33 %addend = load float, float* %ptr, align 4 34 %nextsum = fadd float %sum, %addend 66 %addend = load float, float* %ptr, align 4 67 %nextsum = fadd float %sum, %addend 98 %addend = load float, float* %ptr, align 4 99 %nextsum = fadd float %sum, %addend 132 %addend = load float, float* %ptr, align 4 133 %nextsum = fadd float %sum, %addend
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
D | tiled_dot_emitter.cc | 233 llvm::Value* rhs, llvm::Value* addend, in ColumnMajorMatrixVectorProductEmitter() argument 239 addend_(addend), in ColumnMajorMatrixVectorProductEmitter() 245 CHECK(!has_addend() || addend != nullptr); in ColumnMajorMatrixVectorProductEmitter() 459 llvm::Value* rhs, llvm::Value* addend, in RowMajorMatrixVectorProductEmitter() argument 464 addend_(addend), in RowMajorMatrixVectorProductEmitter() 470 CHECK(!has_addend() || addend != nullptr); in RowMajorMatrixVectorProductEmitter() 959 llvm::Value* lhs, llvm::Value* rhs, llvm::Value* addend, in GetGemvBuffersWithCanonicalType() argument 972 llvm::Type* addend_type = addend ? addend->getType() : nullptr; in GetGemvBuffersWithCanonicalType() 980 addend ? b->CreateBitCast(addend, GetPointerToElementType(addend_type)) in GetGemvBuffersWithCanonicalType() 992 llvm::Value* rhs, llvm::Value* addend, in EmitRowMajorGemv() argument [all …]
|
D | tiled_dot_emitter.h | 32 llvm::Value* addend, llvm::Value* result, 39 llvm::Value* rhs, llvm::Value* addend,
|
/external/swiftshader/third_party/llvm-7.0/llvm/lib/ExecutionEngine/RuntimeDyld/ |
D | RuntimeDyldImpl.h | 155 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend) in RelocationEntry() argument 156 : SectionID(id), Offset(offset), RelType(type), Addend(addend), in RelocationEntry() 159 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend, in RelocationEntry() argument 161 : SectionID(id), Offset(offset), RelType(type), Addend(addend), in RelocationEntry() 165 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend, in RelocationEntry() argument 167 : SectionID(id), Offset(offset), RelType(type), Addend(addend), in RelocationEntry() 170 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend, in RelocationEntry() argument 174 Addend(SectionAOffset - SectionBOffset + addend), IsPCRel(IsPCRel), in RelocationEntry() 180 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend, in RelocationEntry() argument 185 Addend(SectionAOffset - SectionBOffset + addend), IsPCRel(IsPCRel), in RelocationEntry()
|
/external/llvm/lib/ExecutionEngine/RuntimeDyld/ |
D | RuntimeDyldImpl.h | 152 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend) in RelocationEntry() argument 153 : SectionID(id), Offset(offset), RelType(type), Addend(addend), in RelocationEntry() 156 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend, in RelocationEntry() argument 158 : SectionID(id), Offset(offset), RelType(type), Addend(addend), in RelocationEntry() 161 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend, in RelocationEntry() argument 163 : SectionID(id), Offset(offset), RelType(type), Addend(addend), in RelocationEntry() 166 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend, in RelocationEntry() argument 170 Addend(SectionAOffset - SectionBOffset + addend), IsPCRel(IsPCRel), in RelocationEntry()
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/Thumb2/ |
D | thumb2-uxt_rot.ll | 44 %addend = zext i8 %X.trunc to i32 45 %sum = add i32 %A, %addend 55 %addend = zext i16 %X.trunc to i32 56 %sum = add i32 %A, %addend 66 %addend = zext i8 %X.trunc to i32 67 %sum = add i32 %A, %addend
|
/external/llvm/test/CodeGen/Thumb2/ |
D | thumb2-uxt_rot.ll | 42 %addend = zext i8 %X.trunc to i32 44 %sum = add i32 %A, %addend 53 %addend = zext i16 %X.trunc to i32 55 %sum = add i32 %A, %addend
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | cudnn_fused_conv_rewriter.cc | 118 for (auto addend : addends) { in FindConvWithRelu() local 119 if (addend) { in FindConvWithRelu() 120 if (bias == nullptr && Match(addend, bias_pattern)) { in FindConvWithRelu() 122 } else if (conv_instr == nullptr && Match(addend, conv_pattern)) { in FindConvWithRelu() 124 } else if (side_input == nullptr && Match(addend, side_input_pattern)) { in FindConvWithRelu()
|
/external/clang/test/CodeGen/ |
D | cleanup-stack.c | 6 int addend; member 10 *p->var += p->addend; in f0()
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/tests/ |
D | cpu_infeed_test.cc | 139 auto addend = Reduce(infeed, ConstantR0<float>(&builder, 0.0f), in TEST_F() local 141 Add(prev, addend); in TEST_F() 228 auto addend = in TEST_F() local 231 auto result = Add(GetTupleElement(prev, 0), addend); in TEST_F()
|
/external/caliper/caliper/src/main/java/com/google/caliper/util/ |
D | ShortDuration.java | 121 public abstract ShortDuration plus(ShortDuration addend); in plus() argument 168 @Override public ShortDuration plus(ShortDuration addend) { 169 return addend; 219 @Override public ShortDuration plus(ShortDuration addend) { in plus() argument 220 return new PositiveShortDuration(picos + addend.picos); in plus()
|
/external/jemalloc/src/ |
D | nstime.c | 55 nstime_add(nstime_t *time, const nstime_t *addend) in nstime_add() argument 58 assert(UINT64_MAX - time->ns >= addend->ns); in nstime_add() 60 time->ns += addend->ns; in nstime_add()
|
/external/jemalloc_new/include/jemalloc/internal/ |
D | nstime.h | 20 void nstime_add(nstime_t *time, const nstime_t *addend); 21 void nstime_iadd(nstime_t *time, uint64_t addend);
|