/external/tensorflow/tensorflow/core/kernels/ |
D | queue_op.h | 70 void ComputeAsync(OpKernelContext* ctx, DoneCallback callback) final; 73 virtual void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 99 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 124 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 145 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 170 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 213 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 231 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 250 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, 262 void ComputeAsync(OpKernelContext* ctx, QueueInterface* queue,
|
D | queue_op.cc | 53 void QueueOpKernel::ComputeAsync(OpKernelContext* ctx, DoneCallback callback) { in ComputeAsync() function in tensorflow::QueueOpKernel 62 ComputeAsync(ctx, queue, [callback, queue]() { in ComputeAsync() 88 void EnqueueOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::EnqueueOp 129 void EnqueueManyOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::EnqueueManyOp 166 void DequeueOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::DequeueOp 209 void DequeueManyOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::DequeueManyOp 282 void DequeueUpToOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::DequeueUpToOp 333 void QueueCloseOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::QueueCloseOp 347 void QueueSizeOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::QueueSizeOp 358 void QueueIsClosedOp::ComputeAsync(OpKernelContext* ctx, QueueInterface* queue, in ComputeAsync() function in tensorflow::QueueIsClosedOp
|
D | nccl_ops.cc | 97 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon27d3229c0111::NcclAllReduceOpKernel 133 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon27d3229c0111::NcclReduceSendKernel 164 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon27d3229c0111::NcclReduceRecvKernel 203 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon27d3229c0111::NcclBroadcastSendKernel 232 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon27d3229c0111::NcclBroadcastRecvKernel 265 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon27d3229c0111::NcclStubKernel
|
D | conditional_accumulator_base_op.h | 197 void ComputeAsync(OpKernelContext* ctx, DoneCallback callback) final { in ComputeAsync() function 201 ComputeAsync(ctx, accumulator, [callback, accumulator]() { in ComputeAsync() 208 virtual void ComputeAsync(OpKernelContext* ctx, 238 void ComputeAsync(OpKernelContext* ctx, in ComputeAsync() function
|
D | stack.h | 44 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override; 63 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override;
|
D | barrier_ops.cc | 507 void ComputeAsync(OpKernelContext* ctx, DoneCallback callback) final { in ComputeAsync() function in tensorflow::barrier::BarrierOpKernel 511 ComputeAsync(ctx, barrier, [callback, barrier]() { in ComputeAsync() 518 virtual void ComputeAsync(OpKernelContext* ctx, Barrier* barrier, 532 void ComputeAsync(OpKernelContext* ctx, Barrier* barrier, in ComputeAsync() function in tensorflow::barrier::InsertManyOp 581 void ComputeAsync(OpKernelContext* ctx, Barrier* barrier, in ComputeAsync() function in tensorflow::barrier::TakeManyOp 640 void ComputeAsync(OpKernelContext* ctx, Barrier* barrier, in ComputeAsync() function in tensorflow::barrier::BarrierCloseOp 659 void ComputeAsync(OpKernelContext* ctx, Barrier* barrier, in ComputeAsync() function in tensorflow::barrier::BarrierIncompleteSizeOp 678 void ComputeAsync(OpKernelContext* ctx, Barrier* barrier, in ComputeAsync() function in tensorflow::barrier::BarrierReadySizeOp
|
D | sendrecv_ops.h | 42 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override;
|
D | partitioned_function_ops.h | 41 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override;
|
D | function_ops.h | 65 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override;
|
D | stack.cc | 221 void StackPushOp::ComputeAsync(OpKernelContext* ctx, DoneCallback done) { in ComputeAsync() function in tensorflow::StackPushOp 290 void StackPopOp::ComputeAsync(OpKernelContext* ctx, DoneCallback done) { in ComputeAsync() function in tensorflow::StackPopOp
|
D | collective_ops.cc | 60 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon8fd24bc60111::CollectiveOpV1Kernel 116 ComputeAsync(c, done); in CanProceedWithCompute() 500 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon8fd24bc60111::CollectiveReduceV2OpKernel 629 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon8fd24bc60111::CollectiveGatherV2OpKernel 763 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon8fd24bc60111::CollectiveBcastSendV2OpKernel 889 void ComputeAsync(OpKernelContext* c, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anon8fd24bc60111::CollectiveBcastRecvV2OpKernel
|
D | batch_kernels.cc | 246 void ComputeAsync(OpKernelContext* c, DoneCallback done) final { in ComputeAsync() function in tensorflow::BatchFunctionKernel 480 void ComputeAsync(OpKernelContext* c, DoneCallback done) final { in ComputeAsync() function in tensorflow::BatchKernel 735 void ComputeAsync(OpKernelContext* c, DoneCallback done) final { in ComputeAsync() function in tensorflow::UnbatchKernel 921 void ComputeAsync(OpKernelContext* c, DoneCallback done) final { in ComputeAsync() function in tensorflow::UnbatchGradKernel
|
/external/tensorflow/tensorflow/core/grappler/optimizers/ |
D | gpu_swapping_kernels.cc | 29 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anoncce793000111::CopyFromGpuToHostKernel 62 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::__anoncce793000111::CopyFromHostToGpuKernel
|
/external/tensorflow/tensorflow/core/framework/ |
D | device.h | 92 virtual void ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context, in ComputeAsync() function 94 op_kernel->ComputeAsync(context, std::move(done)); in ComputeAsync()
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | renamed_device.h | 126 void ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context, in ComputeAsync() function 128 underlying_device_->ComputeAsync(op_kernel, context, std::move(done)); in ComputeAsync()
|
D | testlib_ops.cc | 92 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::test::DelayOp
|
/external/tensorflow/tensorflow/core/tpu/kernels/ |
D | tpu_execute_op.h | 37 void ComputeAsync(OpKernelContext* context, DoneCallback done) override;
|
D | transfer_ops.h | 34 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override;
|
D | transfer_ops.cc | 43 void TpuTransferAsyncOpKernelBase::ComputeAsync(OpKernelContext* ctx, in ComputeAsync() function in tensorflow::TpuTransferAsyncOpKernelBase
|
/external/tensorflow/tensorflow/core/kernels/data/ |
D | map_defun_op.h | 51 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override;
|
/external/tensorflow/tensorflow/compiler/xrt/kernels/ |
D | tpu_execute_op.cc | 253 void ComputeAsync(OpKernelContext* context, DoneCallback done) override; 262 void XRTExecuteOp::ComputeAsync(OpKernelContext* context, DoneCallback done) { in ComputeAsync() function in tensorflow::__anon3fac8bf30111::XRTExecuteOp 375 void ComputeAsync(OpKernelContext* context, DoneCallback done) override; 384 void XRTExecuteChainedOp::ComputeAsync(OpKernelContext* context, in ComputeAsync() function in tensorflow::__anon3fac8bf30111::XRTExecuteChainedOp
|
D | xrt_execute_op.cc | 418 void ComputeAsync(OpKernelContext* context, DoneCallback done) override; 427 void XRTExecuteOp::ComputeAsync(OpKernelContext* context, DoneCallback done) { in ComputeAsync() function in tensorflow::__anon0036f1f80111::XRTExecuteOp 505 void ComputeAsync(OpKernelContext* context, DoneCallback done) override; 514 void XRTExecuteChainedOp::ComputeAsync(OpKernelContext* context, in ComputeAsync() function in tensorflow::__anon0036f1f80111::XRTExecuteChainedOp
|
/external/tensorflow/tensorflow/core/kernels/data/experimental/ |
D | to_tf_record_op.cc | 49 void ComputeAsync(OpKernelContext* ctx, DoneCallback done) override { in ComputeAsync() function in tensorflow::data::experimental::__anone30cd1e90111::ToTFRecordOp
|
/external/tensorflow/tensorflow/core/kernels/linalg/ |
D | self_adjoint_eig_v2_op_gpu.cc | 48 void ComputeAsync(OpKernelContext* context, DoneCallback done) final { in ComputeAsync() function in tensorflow::SelfAdjointEigV2OpGpu
|
D | determinant_op.cc | 133 void ComputeAsync(OpKernelContext* context, DoneCallback done) final { in ComputeAsync() function in tensorflow::DeterminantOpGpu 273 void ComputeAsync(OpKernelContext* context, DoneCallback done) final { in ComputeAsync() function in tensorflow::LogDeterminantOpGpu
|