Home
last modified time | relevance | path

Searched refs:addend (Results 1 – 25 of 119) sorted by relevance

12345

/external/swiftshader/third_party/llvm-7.0/llvm/test/Analysis/BasicAA/
Dsequential-gep.ll5 define void @t1([8 x i32]* %p, i32 %addend, i32* %q) {
7 %add = add nsw nuw i32 %addend, %knownnonzero
8 %gep1 = getelementptr [8 x i32], [8 x i32]* %p, i32 2, i32 %addend
15 define void @t2([8 x i32]* %p, i32 %addend, i32* %q) {
17 %add = add nsw nuw i32 %addend, %knownnonzero
18 %gep1 = getelementptr [8 x i32], [8 x i32]* %p, i32 1, i32 %addend
25 define void @t3([8 x i32]* %p, i32 %addend, i32* %q) {
27 %add = add nsw nuw i32 %addend, %knownnonzero
35 define void @t4([8 x i32]* %p, i32 %addend, i32* %q) {
37 %add = add nsw nuw i32 %addend, %knownnonzero
[all …]
/external/llvm/test/Analysis/BasicAA/
Dsequential-gep.ll5 define void @t1([8 x i32]* %p, i32 %addend, i32* %q) {
7 %add = add nsw nuw i32 %addend, %knownnonzero
8 %gep1 = getelementptr [8 x i32], [8 x i32]* %p, i32 2, i32 %addend
15 define void @t2([8 x i32]* %p, i32 %addend, i32* %q) {
17 %add = add nsw nuw i32 %addend, %knownnonzero
18 %gep1 = getelementptr [8 x i32], [8 x i32]* %p, i32 1, i32 %addend
25 define void @t3([8 x i32]* %p, i32 %addend, i32* %q) {
27 %add = add nsw nuw i32 %addend, %knownnonzero
35 define void @t4([8 x i32]* %p, i32 %addend, i32* %q) {
37 %add = add nsw nuw i32 %addend, %knownnonzero
[all …]
/external/google-breakpad/src/common/
Dtest_assembler.cc67 Label Label::operator+(uint64_t addend) const { in operator +()
69 l.value_->Set(this->value_, addend); in operator +()
106 uint64_t addend; in IsKnownConstant() local
107 value_->Get(&base, &addend); in IsKnownConstant()
109 if (value_p) *value_p = addend; in IsKnownConstant()
129 Label::Binding::Binding(uint64_t addend) in Binding() argument
130 : base_(NULL), addend_(addend), reference_count_(1) { } in Binding()
138 void Label::Binding::Set(Binding *binding, uint64_t addend) { in Set() argument
141 assert(addend_ == addend); in Set()
145 binding->Set(NULL, addend_ - addend); in Set()
[all …]
/external/swiftshader/third_party/subzero/tests_lit/llvm2ice_tests/
Dvector-align.ll16 define internal <4 x i32> @test_add(i32 %addr_i, <4 x i32> %addend) {
20 %result = add <4 x i32> %addend, %loaded
33 define internal <4 x i32> @test_and(i32 %addr_i, <4 x i32> %addend) {
37 %result = and <4 x i32> %addend, %loaded
50 define internal <4 x i32> @test_or(i32 %addr_i, <4 x i32> %addend) {
54 %result = or <4 x i32> %addend, %loaded
67 define internal <4 x i32> @test_xor(i32 %addr_i, <4 x i32> %addend) {
71 %result = xor <4 x i32> %addend, %loaded
84 define internal <4 x i32> @test_sub(i32 %addr_i, <4 x i32> %addend) {
88 %result = sub <4 x i32> %addend, %loaded
[all …]
/external/brotli/c/fuzz/
Ddecode_fuzzer.c13 size_t addend = 0; in LLVMFuzzerTestOneInput() local
15 addend = data[size - 1] & 7; in LLVMFuzzerTestOneInput()
26 const size_t total_out_limit = (addend == 0) ? (1 << 26) : (1 << 24); in LLVMFuzzerTestOneInput()
31 if (addend == 0) in LLVMFuzzerTestOneInput()
32 addend = size; in LLVMFuzzerTestOneInput()
35 size_t next_i = i + addend; in LLVMFuzzerTestOneInput()
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/LICM/
Ddropped-tbaa.ll6 ; void add(double *restrict data, int *restrict addend) {
7 ; *data += *addend;
10 ; void foo(double *data, int *addend) {
12 ; *data += *addend;
13 ; add(data, addend);
16 ; We want to make sure the load of addend gets hoisted, independent of the second load
19 define void @foo(double* %data, i32* %addend) #0 {
44 %2 = load i32, i32* %addend, align 4, !tbaa !1
54 %5 = load i32, i32* %addend, align 4, !tbaa !1, !alias.scope !7, !noalias !10
87 !8 = distinct !{!8, !9, !"add: %addend"}
/external/swiftshader/third_party/LLVM/test/CodeGen/ARM/
Dfast-isel-static.ll4 define void @myadd(float* %sum, float* %addend) nounwind {
7 %addend.addr = alloca float*, align 4
9 store float* %addend, float** %addend.addr, align 4
12 %tmp2 = load float** %addend.addr, align 4
/external/jemalloc_new/src/
Dnstime.c52 nstime_add(nstime_t *time, const nstime_t *addend) { in nstime_add() argument
53 assert(UINT64_MAX - time->ns >= addend->ns); in nstime_add()
55 time->ns += addend->ns; in nstime_add()
59 nstime_iadd(nstime_t *time, uint64_t addend) { in nstime_iadd() argument
60 assert(UINT64_MAX - time->ns >= addend); in nstime_iadd()
62 time->ns += addend; in nstime_iadd()
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/ARM/
Dfast-isel-static.ll6 define void @myadd(float* %sum, float* %addend) nounwind {
9 %addend.addr = alloca float*, align 4
11 store float* %addend, float** %addend.addr, align 4
14 %tmp2 = load float*, float** %addend.addr, align 4
/external/llvm/test/CodeGen/ARM/
Dfast-isel-static.ll6 define void @myadd(float* %sum, float* %addend) nounwind {
9 %addend.addr = alloca float*, align 4
11 store float* %addend, float** %addend.addr, align 4
14 %tmp2 = load float*, float** %addend.addr, align 4
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/LoopStrengthReduce/
Dsext-ind-var.ll33 %addend = load float, float* %ptr, align 4
34 %nextsum = fadd float %sum, %addend
65 %addend = load float, float* %ptr, align 4
66 %nextsum = fadd float %sum, %addend
97 %addend = load float, float* %ptr, align 4
98 %nextsum = fadd float %sum, %addend
131 %addend = load float, float* %ptr, align 4
132 %nextsum = fadd float %sum, %addend
/external/grpc-grpc/test/core/gpr/
Dtime_test.cc185 gpr_timespec addend[3]; in test_sticky_infinities() local
188 addend[0] = gpr_inf_future(GPR_TIMESPAN); in test_sticky_infinities()
189 addend[1] = gpr_inf_past(GPR_TIMESPAN); in test_sticky_infinities()
190 addend[2] = gpr_time_0(GPR_TIMESPAN); in test_sticky_infinities()
194 for (j = 0; j != sizeof(addend) / sizeof(addend[0]); j++) { in test_sticky_infinities()
195 gpr_timespec x = gpr_time_add(infinity[i], addend[j]); in test_sticky_infinities()
197 x = gpr_time_sub(infinity[i], addend[j]); in test_sticky_infinities()
/external/llvm/test/Transforms/LoopStrengthReduce/
Dsext-ind-var.ll33 %addend = load float, float* %ptr, align 4
34 %nextsum = fadd float %sum, %addend
66 %addend = load float, float* %ptr, align 4
67 %nextsum = fadd float %sum, %addend
98 %addend = load float, float* %ptr, align 4
99 %nextsum = fadd float %sum, %addend
132 %addend = load float, float* %ptr, align 4
133 %nextsum = fadd float %sum, %addend
/external/tensorflow/tensorflow/compiler/xla/service/cpu/
Dtiled_dot_emitter.cc233 llvm::Value* rhs, llvm::Value* addend, in ColumnMajorMatrixVectorProductEmitter() argument
239 addend_(addend), in ColumnMajorMatrixVectorProductEmitter()
245 CHECK(!has_addend() || addend != nullptr); in ColumnMajorMatrixVectorProductEmitter()
459 llvm::Value* rhs, llvm::Value* addend, in RowMajorMatrixVectorProductEmitter() argument
464 addend_(addend), in RowMajorMatrixVectorProductEmitter()
470 CHECK(!has_addend() || addend != nullptr); in RowMajorMatrixVectorProductEmitter()
959 llvm::Value* lhs, llvm::Value* rhs, llvm::Value* addend, in GetGemvBuffersWithCanonicalType() argument
972 llvm::Type* addend_type = addend ? addend->getType() : nullptr; in GetGemvBuffersWithCanonicalType()
980 addend ? b->CreateBitCast(addend, GetPointerToElementType(addend_type)) in GetGemvBuffersWithCanonicalType()
992 llvm::Value* rhs, llvm::Value* addend, in EmitRowMajorGemv() argument
[all …]
Dtiled_dot_emitter.h32 llvm::Value* addend, llvm::Value* result,
39 llvm::Value* rhs, llvm::Value* addend,
/external/swiftshader/third_party/llvm-7.0/llvm/lib/ExecutionEngine/RuntimeDyld/
DRuntimeDyldImpl.h155 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend) in RelocationEntry() argument
156 : SectionID(id), Offset(offset), RelType(type), Addend(addend), in RelocationEntry()
159 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend, in RelocationEntry() argument
161 : SectionID(id), Offset(offset), RelType(type), Addend(addend), in RelocationEntry()
165 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend, in RelocationEntry() argument
167 : SectionID(id), Offset(offset), RelType(type), Addend(addend), in RelocationEntry()
170 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend, in RelocationEntry() argument
174 Addend(SectionAOffset - SectionBOffset + addend), IsPCRel(IsPCRel), in RelocationEntry()
180 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend, in RelocationEntry() argument
185 Addend(SectionAOffset - SectionBOffset + addend), IsPCRel(IsPCRel), in RelocationEntry()
/external/llvm/lib/ExecutionEngine/RuntimeDyld/
DRuntimeDyldImpl.h152 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend) in RelocationEntry() argument
153 : SectionID(id), Offset(offset), RelType(type), Addend(addend), in RelocationEntry()
156 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend, in RelocationEntry() argument
158 : SectionID(id), Offset(offset), RelType(type), Addend(addend), in RelocationEntry()
161 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend, in RelocationEntry() argument
163 : SectionID(id), Offset(offset), RelType(type), Addend(addend), in RelocationEntry()
166 RelocationEntry(unsigned id, uint64_t offset, uint32_t type, int64_t addend, in RelocationEntry() argument
170 Addend(SectionAOffset - SectionBOffset + addend), IsPCRel(IsPCRel), in RelocationEntry()
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/Thumb2/
Dthumb2-uxt_rot.ll44 %addend = zext i8 %X.trunc to i32
45 %sum = add i32 %A, %addend
55 %addend = zext i16 %X.trunc to i32
56 %sum = add i32 %A, %addend
66 %addend = zext i8 %X.trunc to i32
67 %sum = add i32 %A, %addend
/external/llvm/test/CodeGen/Thumb2/
Dthumb2-uxt_rot.ll42 %addend = zext i8 %X.trunc to i32
44 %sum = add i32 %A, %addend
53 %addend = zext i16 %X.trunc to i32
55 %sum = add i32 %A, %addend
/external/tensorflow/tensorflow/compiler/xla/service/gpu/
Dcudnn_fused_conv_rewriter.cc118 for (auto addend : addends) { in FindConvWithRelu() local
119 if (addend) { in FindConvWithRelu()
120 if (bias == nullptr && Match(addend, bias_pattern)) { in FindConvWithRelu()
122 } else if (conv_instr == nullptr && Match(addend, conv_pattern)) { in FindConvWithRelu()
124 } else if (side_input == nullptr && Match(addend, side_input_pattern)) { in FindConvWithRelu()
/external/clang/test/CodeGen/
Dcleanup-stack.c6 int addend; member
10 *p->var += p->addend; in f0()
/external/tensorflow/tensorflow/compiler/xla/service/cpu/tests/
Dcpu_infeed_test.cc139 auto addend = Reduce(infeed, ConstantR0<float>(&builder, 0.0f), in TEST_F() local
141 Add(prev, addend); in TEST_F()
228 auto addend = in TEST_F() local
231 auto result = Add(GetTupleElement(prev, 0), addend); in TEST_F()
/external/caliper/caliper/src/main/java/com/google/caliper/util/
DShortDuration.java121 public abstract ShortDuration plus(ShortDuration addend); in plus() argument
168 @Override public ShortDuration plus(ShortDuration addend) {
169 return addend;
219 @Override public ShortDuration plus(ShortDuration addend) { in plus() argument
220 return new PositiveShortDuration(picos + addend.picos); in plus()
/external/jemalloc/src/
Dnstime.c55 nstime_add(nstime_t *time, const nstime_t *addend) in nstime_add() argument
58 assert(UINT64_MAX - time->ns >= addend->ns); in nstime_add()
60 time->ns += addend->ns; in nstime_add()
/external/jemalloc_new/include/jemalloc/internal/
Dnstime.h20 void nstime_add(nstime_t *time, const nstime_t *addend);
21 void nstime_iadd(nstime_t *time, uint64_t addend);

12345