/external/tensorflow/tensorflow/core/common_runtime/gpu/ |
D | gpu_util.h | 39 const DeviceContext* device_context, 48 const DeviceContext* device_context, 81 const DeviceContext* device_context, 89 const DeviceContext* device_context, 94 DeviceContext* send_dev_context, DeviceContext* recv_dev_context, 104 const DeviceContext* device_context,
|
D | gpu_util.cc | 63 Status PrepareCopy(Device* device, const DeviceContext* ctx, const Tensor& src, in PrepareCopy() 116 const DeviceContext* device_context, in SetProtoFromGPU() 189 DeviceContext* send_dev_context, DeviceContext* recv_dev_context, in DeviceToDeviceCopy() 256 const DeviceContext* device_context, in CopyGPUTensorToCPU() 301 const DeviceContext* device_context, in CopyCPUTensorToGPU() 394 const DeviceContext* device_context, in Checksum() 423 const DeviceContext* device_context, in CopyGPUTensorToSameGPU()
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | copy_tensor.h | 32 DeviceContext* send_dev_context, DeviceContext* recv_dev_context, 43 static void ViaDMA(StringPiece edge_name, DeviceContext* send_dev_context, 44 DeviceContext* recv_dev_context, Device* src, Device* dst,
|
D | collective_rma_local.h | 41 DeviceContext* to_device_ctx, 49 DeviceContext* from_device_ctx, 76 static void MemCpyAsync(DeviceContext* src_dev_ctx, 77 DeviceContext* dst_dev_ctx, Device* src_dev,
|
D | collective_rma_local.cc | 28 const string& key, Device* to_device, DeviceContext* to_device_ctx, in RecvFromPeer() 73 Device* from_device, DeviceContext* from_device_ctx, in PostToPeer() 84 DeviceContext* src_dev_ctx, DeviceContext* dst_dev_ctx, Device* src_dev, in MemCpyAsync()
|
D | buf_rendezvous.h | 28 class DeviceContext; variable 61 DeviceContext* prod_ctx; 77 void ProvideBuf(const string& key, Device* dev, DeviceContext* dev_ctx,
|
D | rendezvous_util.h | 34 Rendezvous* rendezvous, DeviceContext* device_context, 43 Rendezvous* rendezvous, DeviceContext* device_context,
|
D | copy_tensor.cc | 53 DeviceContext* recv_dev_context, StatusCallback done) { in CopyHostToDevice() 118 DeviceContext* send_dev_context, StatusCallback done) { in CopyDeviceToHost() 182 DeviceContext* send_dev_context, in CopyDeviceToDevice() 183 DeviceContext* recv_dev_context, Device* src, in CopyDeviceToDevice() 249 void CopyTensor::ViaDMA(StringPiece edge_name, DeviceContext* send_dev_context, in ViaDMA() 250 DeviceContext* recv_dev_context, Device* src, in ViaDMA()
|
D | test_collective_executor_mgr.h | 33 DeviceContext* to_device_ctx, in RecvFromPeer() 43 DeviceContext* from_device_ctx, in PostToPeer()
|
D | renamed_device.h | 86 DeviceContext* dc, in ReinitializeGpuDevice() 109 DeviceContext* context, const TensorReferenceVector& tensors) override { in ConsumeListOfAccessedTensors()
|
D | process_function_library_runtime.h | 60 DeviceContext* device_context, 73 DeviceContext* device_context, 207 DeviceContext** device_context) const;
|
D | base_collective_executor.h | 123 DeviceContext* to_device_ctx, in RecvFromPeer() 134 DeviceContext* from_device_ctx, in PostToPeer()
|
D | rendezvous_util.cc | 23 Rendezvous* rendezvous, DeviceContext* device_context, in SendTensorsToRendezvous() 57 Rendezvous* rendezvous, DeviceContext* device_context, in RecvOutputsFromRendezvousAsync()
|
D | buf_rendezvous_test.cc | 34 fake_dev_ctx_ = reinterpret_cast<DeviceContext*>(1024LLU); in BufRendezvousTest() 40 DeviceContext* fake_dev_ctx_ = nullptr;
|
D | ring_gatherer_test.cc | 68 DeviceContext* to_device_ctx, in RecvFromPeer() 82 DeviceContext* from_device_ctx, in PostToPeer() 452 gtl::InlinedVector<DeviceContext*, 4> input_dc; in DoGather() 453 DeviceContext* dev_ctx = nullptr; in DoGather() 459 dev_ctx = new DeviceContext; in DoGather()
|
D | collective_util.h | 42 gtl::InlinedVector<DeviceContext*, 4> sub_input_dc_;
|
/external/tensorflow/tensorflow/core/framework/ |
D | device_base.h | 68 class DeviceContext : public core::RefCounted { 70 ~DeviceContext() override {} in ~DeviceContext() 112 typedef std::vector<DeviceContext*> DeviceContextMap; 153 DeviceContext* default_context = nullptr; 227 DeviceContext* /*dc*/, in ReinitializeGpuDevice() argument
|
D | op_kernel.h | 658 const gtl::InlinedVector<DeviceContext*, 4>* input_device_contexts = 660 DeviceContext* op_device_context = nullptr; 1022 DeviceContext* input_device_context(int index); 1032 DeviceContext* op_device_context() { in op_device_context() 1033 DeviceContext* ret = params_->op_device_context; in op_device_context() 1612 static_assert(std::is_base_of<DeviceContext, T>::value, in op_device_context() 1622 static_assert(std::is_base_of<DeviceContext, T>::value, in input_device_context() 1627 inline DeviceContext* OpKernelContext::input_device_context(int index) { in input_device_context()
|
/external/tensorflow/tensorflow/contrib/gdr/ |
D | gdr_memory_manager.h | 25 class DeviceContext; variable 44 Device* device, DeviceContext* device_context, bool on_host, 51 Device* device, DeviceContext* device_context, bool on_host,
|
D | gdr_collective_executor_mgr.cc | 38 DeviceContext* to_device_ctx, in RecvBufCall() 80 DeviceContext* to_device_ctx, in RecvFromPeer()
|
/external/tensorflow/tensorflow/core/distributed_runtime/ |
D | collective_rma_distributed.cc | 37 DeviceContext* to_device_ctx, in RecvBufCall() 77 const string& key, Device* to_device, DeviceContext* to_device_ctx, in RecvFromPeer()
|
D | collective_rma_distributed_test.cc | 276 DeviceContext* to_device_ctx = nullptr; in TEST_F() 304 DeviceContext* to_device_ctx = nullptr; in TEST_F() 336 DeviceContext* to_device_ctx = nullptr; in TEST_F()
|
D | collective_rma_distributed.h | 37 DeviceContext* to_device_ctx,
|
/external/tensorflow/tensorflow/core/common_runtime/sycl/ |
D | sycl_device_context.h | 28 class SYCLDeviceContext : public DeviceContext {
|
/external/tensorflow/tensorflow/core/kernels/ |
D | collective_nccl_reducer_test.cc | 259 gtl::InlinedVector<DeviceContext*, 4> input_dc; in DoReduce() 260 DeviceContext* dev_ctx = nullptr; in DoReduce() 266 dev_ctx = new DeviceContext; in DoReduce()
|