/external/tensorflow/tensorflow/core/kernels/ |
D | queue_op.h | 70 void ComputeAsync(OpKernelContext* ctx, DoneCallback callback) final; 73 virtual void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 99 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 124 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 145 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 170 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 213 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 231 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 250 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 262 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue,
|
D | queue_op.cc | 53 void QueueOpKernel::ComputeAsync(OpKernelContext* ctx, DoneCallback callback) { in ComputeAsync() function in tensorflow::QueueOpKernel 62 ComputeAsync(ctx, queue, [callback, queue]() { in ComputeAsync() 88 void EnqueueOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::EnqueueOp 129 void EnqueueManyOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::EnqueueManyOp 166 void DequeueOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::DequeueOp 209 void DequeueManyOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::DequeueManyOp 282 void DequeueUpToOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::DequeueUpToOp 333 void QueueCloseOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::QueueCloseOp 347 void QueueSizeOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::QueueSizeOp 358 void QueueIsClosedOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::QueueIsClosedOp
|
D | nccl_ops.cc | 93 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon49c74ebe0111::NcclAllReduceOpKernel 130 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon49c74ebe0111::NcclReduceSendKernel 162 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon49c74ebe0111::NcclReduceRecvKernel 202 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon49c74ebe0111::NcclBroadcastSendKernel 232 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon49c74ebe0111::NcclBroadcastRecvKernel 266 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon49c74ebe0111::NcclStubKernel
|
D | conditional_accumulator_base_op.h | 129 void ComputeAsync(OpKernelContext* ctx, DoneCallback callback) final { in ComputeAsync() function 133 ComputeAsync(ctx, accumulator, [callback, accumulator]() { in ComputeAsync() 140 virtual void ComputeAsync(OpKernelContext* ctx, 215 void ComputeAsync(OpKernelContext* ctx, in ComputeAsync() function
|
D | barrier_ops.cc | 507 void ComputeAsync(OpKernelContext* ctx, DoneCallback callback) final { in ComputeAsync() function in tensorflow::barrier::BarrierOpKernel 511 ComputeAsync(ctx, barrier, [callback, barrier]() { in ComputeAsync() 518 virtual void ComputeAsync(OpKernelContext* ctx, Barrier* barrier, 532 void ComputeAsync(OpKernelContext* ctx, Barrier* barrier, in ComputeAsync() function in tensorflow::barrier::InsertManyOp 581 void ComputeAsync(OpKernelContext* ctx, Barrier* barrier, in ComputeAsync() function in tensorflow::barrier::TakeManyOp 640 void ComputeAsync(OpKernelContext* ctx, Barrier* barrier, in ComputeAsync() function in tensorflow::barrier::BarrierCloseOp 659 void ComputeAsync(OpKernelContext* ctx, Barrier* barrier, in ComputeAsync() function in tensorflow::barrier::BarrierIncompleteSizeOp 678 void ComputeAsync(OpKernelContext* ctx, Barrier* barrier, in ComputeAsync() function in tensorflow::barrier::BarrierReadySizeOp
|
D | stack.h | 44 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override; 63 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override;
|
D | collective_ops.cc | 55 ComputeAsync(c, done); in CanProceedWithCompute() 87 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon947389680111::CollectiveGatherOpKernel 185 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon947389680111::CollectiveReduceOpKernel 250 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon947389680111::CollectiveBcastSendOpKernel 319 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon947389680111::CollectiveBcastRecvOpKernel
|
D | sendrecv_ops.h | 40 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override;
|
D | functional_ops.cc | 138 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anonf4cf7ed10111::IfOp 225 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anonf4cf7ed10111::CaseOp 332 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anonf4cf7ed10111::WhileOp 510 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anonf4cf7ed10111::ForOp
|
D | partitioned_function_ops.h | 41 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override;
|
D | function_ops.h | 65 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override;
|
D | stack.cc | 220 void StackPushOp::ComputeAsync(OpKernelContext* ctx, DoneCallback done) { in ComputeAsync() function in tensorflow::StackPushOp 289 void StackPopOp::ComputeAsync(OpKernelContext* ctx, DoneCallback done) { in ComputeAsync() function in tensorflow::StackPopOp
|
D | rpc_op.cc | 61 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::RpcOp
|
D | cholesky_op.cc | 100 void ComputeAsync(OpKernelContext* context, DoneCallback done) final { in ComputeAsync() function in tensorflow::CholeskyOpGpu
|
D | determinant_op.cc | 133 void ComputeAsync(OpKernelContext* context, DoneCallback done) final { in ComputeAsync() function in tensorflow::DeterminantOpGpu 273 void ComputeAsync(OpKernelContext* context, DoneCallback done) final { in ComputeAsync() function in tensorflow::LogDeterminantOpGpu
|
/external/tensorflow/tensorflow/core/grappler/optimizers/ |
D | gpu_swapping_kernels.cc | 29 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon145ac5220111::CopyFromGpuToHostKernel 62 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon145ac5220111::CopyFromHostToGpuKernel
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | device.h | 93 virtual void ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context, in ComputeAsync() function 95 op_kernel->ComputeAsync(context, std::move(done)); in ComputeAsync()
|
D | renamed_device.h | 103 void ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context, in ComputeAsync() function 105 underlying_->ComputeAsync(op_kernel, context, std::move(done)); in ComputeAsync()
|
D | testlib_ops.cc | 84 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::test::DelayOp
|
/external/tensorflow/tensorflow/core/kernels/data/ |
D | iterator_ops.h | 114 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override; 130 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override;
|
D | iterator_ops.cc | 570 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::data::__anon9c98ff710411::ToSingleElementOp 650 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::data::__anon9c98ff710411::ReduceDatasetOp 794 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::data::__anon9c98ff710411::OneShotIteratorOp 947 void IteratorGetNextOp::ComputeAsync(OpKernelContext* ctx, DoneCallback done) { in ComputeAsync() function in tensorflow::data::IteratorGetNextOp 997 void IteratorGetNextAsOptionalOp::ComputeAsync(OpKernelContext* ctx, in ComputeAsync() function in tensorflow::data::IteratorGetNextAsOptionalOp
|
/external/tensorflow/tensorflow/compiler/xrt/kernels/ |
D | xrt_execute_op.cc | 122 void ComputeAsync(OpKernelContext* context, DoneCallback done) override; 131 void XRTExecuteOp::ComputeAsync(OpKernelContext* context, DoneCallback done) { in ComputeAsync() function in tensorflow::__anon7384b19a0111::XRTExecuteOp
|
/external/tensorflow/tensorflow/core/kernels/data/experimental/ |
D | to_tf_record_op.cc | 46 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::data::__anon18e6484b0111::ToTFRecordOp
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
D | function_ops.cc | 66 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon0de038b60111::SymbolicGradientOp
|
/external/tensorflow/tensorflow/compiler/jit/ |
D | xla_device.cc | 379 void XlaDevice::ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context, in ComputeAsync() function in tensorflow::XlaDevice 385 op_kernel->ComputeAsync(context, done); in ComputeAsync()
|