/external/ltp/testcases/open_posix_testsuite/stress/threads/pthread_exit/ |
D | stress.c | 146 int rets[NSCENAR * SCALABILITY_FACTOR * FACTOR]; in main() local 176 rets[i * NSCENAR + sc] = in main() 182 if (rets[i * NSCENAR + sc] != 0) { in main() 183 UNRESOLVED(rets in main() 190 if (rets[i * NSCENAR + sc] == 0) { in main() 199 if (rets[i * NSCENAR + sc] == 0) { in main() 205 strerror(rets in main() 212 if (rets[i * NSCENAR + sc] == 0) { in main() 232 && (rets[i * NSCENAR + sc] == 0)) { in main()
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
D | function_ops.cc | 85 std::vector<Tensor>* rets = new std::vector<Tensor>; in ComputeAsync() local 87 opts, handle_, args, rets, [ctx, done, rets](const Status& status) { in ComputeAsync() 90 } else if (rets->size() != ctx->num_outputs()) { in ComputeAsync() 93 " tensor(s), but get ", rets->size(), " tensor(s) instead.")); in ComputeAsync() 95 for (size_t i = 0; i < rets->size(); ++i) { in ComputeAsync() 96 ctx->set_output(i, (*rets)[i]); in ComputeAsync() 99 delete rets; in ComputeAsync()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | function_ops.cc | 264 std::vector<Tensor>* rets = new std::vector<Tensor>; in ComputeAsync() local 265 lib->Run(opts, handle, args, rets, [ctx, done, rets](const Status& status) { in ComputeAsync() 268 } else if (rets->size() != ctx->num_outputs()) { in ComputeAsync() 271 " tensor(s), but get ", rets->size(), " tensor(s) instead.")); in ComputeAsync() 273 for (size_t i = 0; i < rets->size(); ++i) { in ComputeAsync() 274 ctx->set_output(i, (*rets)[i]); in ComputeAsync() 277 delete rets; in ComputeAsync() 376 auto* rets = new std::vector<Tensor>; in ComputeAsync() local 381 lib->Run(opts, handle, args, rets, in ComputeAsync() 382 [rets, activity, done, ctx](const Status& status) { in ComputeAsync() [all …]
|
D | quantization_utils_test.cc | 751 int64_t rets[2]; in TestDivide64x2Pow() local 752 vst1q_s64(rets, ret); in TestDivide64x2Pow() 753 EXPECT_EQ(rets[0], ref); in TestDivide64x2Pow() 754 EXPECT_EQ(rets[1], ref); in TestDivide64x2Pow() 763 int64_t rets[2]; in TestDivide64x2PowRound() local 764 vst1q_s64(rets, shifted); in TestDivide64x2PowRound() 765 EXPECT_EQ(rets[0], ref) << "in = " << val << ", " << POW in TestDivide64x2PowRound() 766 << ", act = " << rets[0] << ", ref = " << ref; in TestDivide64x2PowRound() 767 EXPECT_EQ(rets[1], ref); in TestDivide64x2PowRound() 768 VLOG(1) << "div round: " << val << ", " << rets[0]; in TestDivide64x2PowRound() [all …]
|
D | partitioned_function_ops.cc | 231 std::vector<Tensor>* rets = new std::vector<Tensor>; in RunFunction() local 233 lib->Run(run_opts, handle, inputs, rets, in RunFunction() 234 [rets, rendez, done, ctx, func_name](const Status& status) { in RunFunction() 241 for (int i = 0; i < rets->size(); ++i) { in RunFunction() 242 ctx->set_output(i, (*rets)[i]); in RunFunction() 245 delete rets; in RunFunction()
|
D | functional_ops.cc | 99 gtl::ArraySlice<Tensor> rets) { in SetOutputs() argument 100 if (rets.size() != ctx->num_outputs()) { in SetOutputs() 102 " tensors, but only get ", rets.size()); in SetOutputs() 104 for (int i = 0; i < rets.size(); ++i) { in SetOutputs() 105 if (rets[i].dtype() != kernel->output_type(i)) { in SetOutputs() 108 " but get ", DataTypeString(rets[i].dtype())); in SetOutputs() 110 ctx->set_output(i, rets[i]); in SetOutputs()
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | function_threadpool_test.cc | 74 const std::vector<Tensor>& args, std::vector<Tensor*> rets, in Run() argument 98 CHECK_EQ(rets.size(), out.size()); in Run() 99 for (size_t i = 0; i < rets.size(); ++i) { in Run() 100 *rets[i] = out[i]; in Run() 126 std::vector<Tensor*> rets, bool add_runner = true) { in InstantiateAndRun() argument 129 std::move(rets), add_runner); in InstantiateAndRun() 136 const std::vector<Tensor>& args, std::vector<Tensor*> rets, in InstantiateAndRun() argument 144 status = Run(flr, handle, opts, args, rets, add_runner); in InstantiateAndRun() 151 Status status2 = Run(flr, handle, opts, args, std::move(rets)); in InstantiateAndRun()
|
D | function.cc | 183 std::vector<Tensor>* rets, DoneCallback done) override; 238 std::vector<Tensor>* rets, in Run() argument 240 base_flr_->Run(opts, handle, args, rets, std::move(done)); in Run() 328 std::vector<Tensor>* rets, DoneCallback done) override; 414 gtl::ArraySlice<Tensor> args, std::vector<Tensor>* rets, 495 std::vector<Tensor>* rets = new std::vector<Tensor>; in ComputeAsync() local 496 lib->Run(opts, handle_, args, rets, in ComputeAsync() 497 [ctx, done, rets](const Status& status) { in ComputeAsync() 501 const int ret_size = static_cast<int>(rets->size()); in ComputeAsync() 504 ctx->set_output(i, (*rets)[i]); in ComputeAsync() [all …]
|
D | process_function_library_runtime.cc | 738 std::vector<Tensor>* rets, in RunMultiDevice() argument 785 rets->resize(data->num_outputs_); in RunMultiDevice() 788 [comp_rets, rets, comp_data, refcounted_done](const Status& status) { in RunMultiDevice() 794 (*rets)[comp_data.ret_indices_[i]] = (*comp_rets)[i]; in RunMultiDevice() 912 std::vector<Tensor>* rets, in Run() argument 920 return RunMultiDevice(opts, handle, args, rets, done); in Run() 978 rets](const Status& status, in Run() 991 rendezvous, rets, std::move(done)); in Run() 997 parent_->Run(opts, local_handle, args, rets, std::move(done)); in Run()
|
D | function_test.cc | 104 void Run(const std::vector<Tensor>& args, std::vector<Tensor*> rets) { in Run() argument 113 CHECK_EQ(computed.size(), rets.size()); in Run() 114 for (int i = 0; i < rets.size(); ++i) { in Run() 115 *(rets[i]) = computed[i]; in Run() 171 const std::vector<Tensor>& args, std::vector<Tensor*> rets, in Run() argument 195 CHECK_EQ(rets.size(), out.size()); in Run() 196 for (size_t i = 0; i < rets.size(); ++i) { in Run() 197 *rets[i] = out[i]; in Run() 223 std::vector<Tensor*> rets, bool add_runner = true) { in InstantiateAndRun() argument 226 std::move(rets), add_runner); in InstantiateAndRun() [all …]
|
D | process_function_library_runtime_test.cc | 58 gtl::ArraySlice<Tensor> args, std::vector<Tensor>* rets, in Run() argument 157 const std::vector<Tensor>& args, std::vector<Tensor*> rets) { in Run() argument 183 CHECK_EQ(rets.size(), out.size()); in Run() 184 for (size_t i = 0; i < rets.size(); ++i) { in Run() 185 *rets[i] = out[i]; in Run()
|
D | process_function_library_runtime.h | 137 std::vector<Tensor>* rets, 248 gtl::ArraySlice<Tensor> args, std::vector<Tensor>* rets,
|
/external/tensorflow/tensorflow/core/kernels/data/ |
D | captured_function.h | 58 std::vector<Tensor>* rets) const; 65 std::vector<Tensor>* rets) const; 73 std::vector<Tensor>* rets); 80 std::vector<Tensor>* rets,
|
D | captured_function.cc | 302 std::vector<Tensor>* rets) const { in Run() 333 return frame.ConsumeRetvals(rets); in Run() 338 std::vector<Tensor>* rets) const { in RunWithBorrowedArgs() 370 return frame.ConsumeRetvals(rets); in RunWithBorrowedArgs() 374 const std::vector<Tensor>& args, std::vector<Tensor>* rets) { in RunInstantiated() argument 406 return frame.ConsumeRetvals(rets); in RunInstantiated() 410 IteratorContext* ctx, std::vector<Tensor>&& args, std::vector<Tensor>* rets, in RunAsync() argument 445 [this, rets, step_container, c_mgr, frame]( in RunAsync() 456 s = frame->ConsumeRetvals(rets); in RunAsync()
|
/external/tensorflow/tensorflow/core/distributed_runtime/ |
D | cluster_function_library_runtime_test.cc | 72 const std::vector<Tensor>& args, std::vector<Tensor*> rets) { in InstantiateAndRun() argument 90 CHECK_EQ(rets.size(), out.size()); in InstantiateAndRun() 91 for (size_t i = 0; i < rets.size(); ++i) { in InstantiateAndRun() 92 *rets[i] = out[i]; in InstantiateAndRun()
|
D | cluster_function_library_runtime.cc | 169 std::vector<Tensor>* rets, FunctionLibraryRuntime::DoneCallback done) { in Run() argument 207 [call_options, req, resp, rets, recv_keys, done](const Status& status) { in Run() 231 rets->push_back(t); in Run()
|
/external/python/cpython2/Lib/ |
D | poplib.py | 206 rets = retval.split() 207 if self._debugging: print '*stat*', repr(rets) 208 numMessages = int(rets[1]) 209 sizeMessages = int(rets[2])
|
/external/python/cpython3/Lib/ |
D | poplib.py | 222 rets = retval.split() 223 if self._debugging: print('*stat*', repr(rets)) 224 numMessages = int(rets[1]) 225 sizeMessages = int(rets[2])
|
/external/libxml2/ |
D | gentest.py | 525 rets = ctxt.xpathEval("/api/symbols/function/return") variable 526 for ret in rets: 703 rets = node.xpathEval("return") 705 rets = [] 707 for ret in rets:
|
/external/tensorflow/tensorflow/core/framework/ |
D | function_test.cc | 915 std::vector<Tensor> rets; in TEST() local 916 TF_EXPECT_OK(frame.GetRetvals(&rets)); in TEST() 917 EXPECT_EQ(rets.size(), 0); in TEST() 944 std::vector<Tensor> rets; in TEST() local 945 HasError(frame.GetRetvals(&rets), "does not have value"); in TEST() 949 TF_EXPECT_OK(frame.GetRetvals(&rets)); in TEST() 950 EXPECT_EQ(rets.size(), 1); in TEST() 951 test::ExpectTensorEqual<float>(rets[0], v); in TEST()
|
D | function.h | 282 Status GetRetvals(std::vector<Tensor>* rets) const; 286 Status ConsumeRetvals(std::vector<Tensor>* rets, bool allow_dead_tensors); 644 gtl::ArraySlice<Tensor> args, std::vector<Tensor>* rets, 737 gtl::ArraySlice<Tensor> args, std::vector<Tensor>* rets,
|
/external/llvm/test/CodeGen/Mips/Fast-ISel/ |
D | retabi.ll | 39 define signext i16 @rets() { 41 ; CHECK-LABEL: rets:
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/Mips/Fast-ISel/ |
D | retabi.ll | 39 define signext i16 @rets() { 41 ; CHECK-LABEL: rets:
|
/external/v8/src/compiler/ |
D | wasm-compiler.h | 215 Node* CallDirect(uint32_t index, Node** args, Node*** rets, 217 Node* CallIndirect(uint32_t index, Node** args, Node*** rets, 370 Node* BuildWasmCall(wasm::FunctionSig* sig, Node** args, Node*** rets, 373 Node* BuildImportWasmCall(wasm::FunctionSig* sig, Node** args, Node*** rets, 375 Node* BuildImportWasmCall(wasm::FunctionSig* sig, Node** args, Node*** rets,
|
/external/tensorflow/tensorflow/contrib/recurrent/python/ops/ |
D | recurrent.py | 233 rets = [] 236 rets.append(array_ops.zeros_like(x)) 238 rets.append(dx) 240 return nest.pack_sequence_as(dxs, rets)
|