/external/executorch/backends/arm/runtime/ |
D | ArmBackendEthosU.cpp | 246 auto tensor_out = args[handles.inputs->count + i]->toTensor(); in execute() local 250 tensor_out, in execute() 254 if (tensor_out.scalar_type() == ScalarType::Char and in execute() 259 tensor_out.mutable_data_ptr<char>(), in execute() 260 tensor_out.size(1), in execute() 261 tensor_out.size(2), in execute() 262 tensor_out.size(3)); in execute() 264 for (int j = 0; j < tensor_out.numel(); j++) { in execute() 265 if (tensor_out.scalar_type() == ScalarType::Char) { in execute() 267 tensor_out.mutable_data_ptr<char>()[j] = output_address[j]; in execute() [all …]
|
/external/tensorflow/tensorflow/core/kernels/mkl/ |
D | mkl_input_conversion_op.cc | 110 Tensor* tensor_out; in Compute() local 122 AllocateOutputSetMklShape(context, kInputIndex_0, &tensor_out, in Compute() 137 input.CheckReorderToOpMem(input1_md, tensor_out, net, net_args, in Compute() 211 Tensor* tensor_out; in Compute() local 223 AllocateOutputSetMklShape(context, tf_tensor_index, &tensor_out, in Compute() 235 bool reordered = tf_input.CheckReorderToOpMem(output_mkl_md, tensor_out, in Compute() 244 tensor_out->CopyFrom(*tf_tensor, tensor_out->shape()), in Compute()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | pooling_ops_3d.cc | 210 const Tensor& tensor_out, const Tensor& out_backprop, in launch() 273 tensor_out.tensor<T, 5>().slice(src_indices, src_sizes); in launch() 338 const Tensor& tensor_out = context->input(1); in Compute() local 342 OP_REQUIRES(context, tensor_out.dims() == 5, in Compute() 371 context, tensor_out.shape() == out_shape, in Compute() 373 ", but got ", tensor_out.shape())); in Compute() 379 context, tensor_in, tensor_out, out_backprop, window, stride, out, in Compute() 563 const Tensor& tensor_in, const Tensor& tensor_out, in launch() 580 ConstEigenMatrixMap out_mat(tensor_out.flat<T>().data(), params.depth, in launch() 704 const Tensor& tensor_out = context->input(1); in Compute() local [all …]
|
D | pooling_ops_common.cc | 224 Tensor* tensor_out = nullptr; in Compute() local 226 context->allocate_output(0, tensor_out_shape, &tensor_out)); in Compute() 267 transformed_output = *tensor_out; in Compute() 272 auto& transformed_output = *tensor_out; in Compute() 416 tensor_out->tensor<RT, 4>()); in Compute() 444 TensorFormat data_format, const Tensor* tensor_in, const Tensor* tensor_out, in Compute() argument 448 (tensor_in && tensor_out)) in Compute() 464 if (tensor_out) { in Compute() 465 OP_REQUIRES(context, tensor_out->shape() == params.forward_output_shape(), in Compute() 468 ", but got ", tensor_out->shape())); in Compute() [all …]
|
D | fractional_max_pool_op.cc | 240 const Tensor& tensor_out = context->input(1); in Compute() local 254 OP_REQUIRES(context, tensor_out.dims() == tensor_in_and_out_dims, in Compute() 257 tensor_out.DebugString())); in Compute() 258 OP_REQUIRES(context, tensor_out.NumElements() > 0, in Compute() 260 tensor_out.DebugString())); in Compute() 267 output_size[i] = tensor_out.dim_size(i); in Compute() 275 {1}, DataTypeToEnum<T>::v(), tensor_out.shape(), in Compute() 279 tensor_out.shape(), in Compute() 350 tensor_out.flat<T>().data(), output_size[3], in Compute()
|
D | cudnn_pooling_gpu.cc | 134 const Tensor* tensor_in, const Tensor* tensor_out, Tensor* input_backprop) { in Compute() argument 136 (tensor_in && tensor_out)) in Compute() 156 if (data_format == FORMAT_NHWC || tensor_out == nullptr) { in Compute() 163 transformed_output = *tensor_out; in Compute() 189 if (tensor_out != nullptr) { in Compute() 191 tensor_out->tensor<T, 5>(), in Compute()
|
D | maxpooling_op.cc | 268 const Tensor& tensor_out = context->input(1); in Compute() local 274 OP_REQUIRES(context, tensor_out.dims() == 4, in Compute() 284 {1}, DataTypeToEnum<T>::v(), tensor_out.shape(), in Compute() 288 tensor_out.shape(), in Compute() 328 OP_REQUIRES(context, tensor_out.shape() == params.forward_output_shape(), in Compute() 331 ", but got ", tensor_out.shape())); in Compute() 394 const Tensor& tensor_out = context->input(1); in Compute() local 400 OP_REQUIRES(context, tensor_out.dims() == 4, in Compute() 443 data_format_, &tensor_in, &tensor_out, in Compute() 501 const Tensor& tensor_out = context->input(1); in Compute() local [all …]
|
D | cudnn_pooling_gpu.h | 62 const Tensor* tensor_in, const Tensor* tensor_out,
|
D | pooling_ops_common_gpu.h | 64 const Tensor* tensor_out, const Tensor& out_backprop,
|
/external/pytorch/test/jit/ |
D | test_complexity.py | 47 tensor_out = False 50 tensor_out = True 52 num_non_tensor += int(not tensor_out)
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | batch_matmul.cc | 393 const scalar* input, TfLiteTensor* tensor_out, in TransposeRowsColumnsImpl() argument 413 TfLiteTensor* tensor_out) { in TransposeRowsColumns() argument 416 tensor_out, in TransposeRowsColumns() 417 GetTensorData<float>(tensor_out)); in TransposeRowsColumns() 421 tensor_in, GetTensorData<int8_t>(tensor_in), tensor_out, in TransposeRowsColumns() 422 GetTensorData<int8_t>(tensor_out)); in TransposeRowsColumns() 426 tensor_in, GetTensorData<int16_t>(tensor_in), tensor_out, in TransposeRowsColumns() 427 GetTensorData<int16_t>(tensor_out)); in TransposeRowsColumns()
|
/external/tensorflow/tensorflow/core/util/ |
D | mkl_util.h | 789 inline void AllocTmpBuffer(OpKernelContext* context, Tensor* tensor_out, in AllocTmpBuffer() argument 795 tf_shape, tensor_out)); in AllocTmpBuffer() 796 *buf_out = static_cast<void*>(tensor_out->flat<T>().data()); in AllocTmpBuffer() 800 inline void AllocTmpBuffer(OpKernelContext* context, Tensor* tensor_out, in AllocTmpBuffer() argument 803 tf_shape, tensor_out)); in AllocTmpBuffer()
|
/external/pytorch/torch/testing/_internal/distributed/ |
D | distributed_test.py | 2528 self, tensor_out, tensor_in, group_id, rank, cuda=True, rank_to_GPU=None argument 2532 tensor_out = tensor_out.cuda(rank_to_GPU[rank][0]) 2533 tensor_shapes = [tensor_out.shape] 2538 tensor_out, 2546 return tensor_out 2556 tensor_out = torch.zeros(size, dtype=torch.int64) 2560 tensor_out = self._reduce_scatter_tensor_helper( 2561 tensor_out, tensor_in, group_id, rank, True, rank_to_GPU 2565 self.assertEqual(tensor_out, expected_tensor) 2570 tensor_out = self._reduce_scatter_tensor_helper( [all …]
|
/external/pytorch/test/onnx/ |
D | test_pytorch_onnx_onnxruntime.py | 508 tensor_out: Tensor 515 tensor_out=a,
|