/external/tensorflow/tensorflow/core/kernels/ |
D | mkl_input_conversion_op.cc | 123 Tensor* tensor_out; in Compute() local 135 AllocateOutputSetMklShape(context, kInputIndex_0, &tensor_out, in Compute() 148 tensor_out, &net), in Compute() 221 Tensor* tensor_out; in Compute() local 233 AllocateOutputSetMklShape(context, tf_tensor_index, &tensor_out, in Compute() 246 memory::primitive_desc(output_mkl_md, cpu_engine), tensor_out, &net); in Compute() 254 CHECK(tensor_out->CopyFrom(*tf_tensor, tensor_out->shape())); in Compute()
|
D | pooling_ops_common.cc | 161 Tensor* tensor_out = nullptr; in Compute() local 163 context->allocate_output(0, tensor_out_shape, &tensor_out)); in Compute() 200 transformed_output = *tensor_out; in Compute() 205 auto& transformed_output = *tensor_out; in Compute() 279 tensor_out->tensor<RT, 4>()); in Compute() 289 const Tensor* tensor_out, const Tensor& out_backprop, in Compute() argument 292 (tensor_in && tensor_out)) in Compute() 324 if (data_format == FORMAT_NHWC || !tensor_out) { in Compute() 331 transformed_output = *tensor_out; in Compute() 362 if (tensor_out) { in Compute() [all …]
|
D | cudnn_pooling_gpu.cc | 124 const Tensor* tensor_in, const Tensor* tensor_out, Tensor* input_backprop) { in Compute() argument 126 (tensor_in && tensor_out)) in Compute() 146 if (data_format == FORMAT_NHWC || tensor_out == nullptr) { in Compute() 153 transformed_output = *tensor_out; in Compute() 179 if (tensor_out != nullptr) { in Compute() 181 tensor_out->tensor<T, 5>(), in Compute()
|
D | pooling_ops_3d.cc | 209 const Tensor& tensor_out, const Tensor& out_backprop, in launch() 276 tensor_out.tensor<T, 5>().slice(src_indices, src_sizes); in launch() 341 const Tensor& tensor_out = context->input(1); in Compute() local 345 OP_REQUIRES(context, tensor_out.dims() == 5, in Compute() 369 context, tensor_in, tensor_out, out_backprop, window, stride, out, in Compute() 544 const Tensor& tensor_in, const Tensor& tensor_out, in launch() 561 ConstEigenMatrixMap out_mat(tensor_out.flat<T>().data(), params.depth, in launch() 685 const Tensor& tensor_out = context->input(1); in Compute() local 691 OP_REQUIRES(context, tensor_out.dims() == 5, in Compute() 703 {2}, 0, tensor_out.shape(), &output)); in Compute() [all …]
|
D | fractional_max_pool_op.cc | 231 const Tensor& tensor_out = context->input(1); in Compute() local 244 output_size[i] = tensor_out.dim_size(i); in Compute() 252 {1}, DataTypeToEnum<T>::v(), tensor_out.shape(), in Compute() 256 tensor_out.shape(), in Compute() 327 tensor_out.flat<T>().data(), output_size[3], in Compute()
|
D | maxpooling_op.cc | 248 const Tensor& tensor_out = context->input(1); in Compute() local 254 OP_REQUIRES(context, tensor_out.dims() == 4, in Compute() 264 {1}, DataTypeToEnum<T>::v(), tensor_out.shape(), in Compute() 268 tensor_out.shape(), in Compute() 380 const Tensor& tensor_out = context->input(1); in Compute() local 386 OP_REQUIRES(context, tensor_out.dims() == 4, in Compute() 422 &tensor_in, &tensor_out, out_backprop, in Compute() 486 const Tensor& tensor_out = context->input(1); in Compute() local 492 OP_REQUIRES(context, tensor_out.dims() == 4, in Compute() 531 {2}, 0, tensor_out.shape(), &output)); in Compute() [all …]
|
D | cudnn_pooling_gpu.h | 62 const Tensor* tensor_in, const Tensor* tensor_out,
|
D | pooling_ops_common_gpu.h | 62 const Tensor* tensor_out, const Tensor& out_backprop,
|
D | pooling_ops_3d_sycl.h | 348 const Tensor& tensor_out, const Tensor& out_backprop, 366 device.get_sycl_buffer(tensor_out.template flat<T>().data()); 477 const Tensor& tensor_in, const Tensor& tensor_out, 486 device.get_sycl_buffer(tensor_out.template flat<T>().data());
|
/external/tensorflow/tensorflow/core/util/ |
D | mkl_util.h | 1040 inline void AllocTmpBuffer(OpKernelContext* context, Tensor* tensor_out, in AllocTmpBuffer() argument 1046 tf_shape, tensor_out)); in AllocTmpBuffer() 1047 *buf_out = static_cast<void*>(tensor_out->flat<T>().data()); in AllocTmpBuffer() 1050 inline void AllocTmpBuffer(OpKernelContext* context, Tensor* tensor_out, in AllocTmpBuffer() argument 1059 tf_shape, tensor_out)); in AllocTmpBuffer() 1060 *buf_out = static_cast<void*>(tensor_out->flat<float>().data()); in AllocTmpBuffer() 1065 inline void AllocTmpBuffer(OpKernelContext* context, Tensor* tensor_out, in AllocTmpBuffer() argument 1068 tf_shape, tensor_out)); in AllocTmpBuffer()
|