Home
last modified time | relevance | path

Searched refs:tensor_out (Results 1 – 14 of 14) sorted by relevance

/external/executorch/backends/arm/runtime/
DArmBackendEthosU.cpp246 auto tensor_out = args[handles.inputs->count + i]->toTensor(); in execute() local
250 tensor_out, in execute()
254 if (tensor_out.scalar_type() == ScalarType::Char and in execute()
259 tensor_out.mutable_data_ptr<char>(), in execute()
260 tensor_out.size(1), in execute()
261 tensor_out.size(2), in execute()
262 tensor_out.size(3)); in execute()
264 for (int j = 0; j < tensor_out.numel(); j++) { in execute()
265 if (tensor_out.scalar_type() == ScalarType::Char) { in execute()
267 tensor_out.mutable_data_ptr<char>()[j] = output_address[j]; in execute()
[all …]
/external/tensorflow/tensorflow/core/kernels/mkl/
Dmkl_input_conversion_op.cc110 Tensor* tensor_out; in Compute() local
122 AllocateOutputSetMklShape(context, kInputIndex_0, &tensor_out, in Compute()
137 input.CheckReorderToOpMem(input1_md, tensor_out, net, net_args, in Compute()
211 Tensor* tensor_out; in Compute() local
223 AllocateOutputSetMklShape(context, tf_tensor_index, &tensor_out, in Compute()
235 bool reordered = tf_input.CheckReorderToOpMem(output_mkl_md, tensor_out, in Compute()
244 tensor_out->CopyFrom(*tf_tensor, tensor_out->shape()), in Compute()
/external/tensorflow/tensorflow/core/kernels/
Dpooling_ops_3d.cc210 const Tensor& tensor_out, const Tensor& out_backprop, in launch()
273 tensor_out.tensor<T, 5>().slice(src_indices, src_sizes); in launch()
338 const Tensor& tensor_out = context->input(1); in Compute() local
342 OP_REQUIRES(context, tensor_out.dims() == 5, in Compute()
371 context, tensor_out.shape() == out_shape, in Compute()
373 ", but got ", tensor_out.shape())); in Compute()
379 context, tensor_in, tensor_out, out_backprop, window, stride, out, in Compute()
563 const Tensor& tensor_in, const Tensor& tensor_out, in launch()
580 ConstEigenMatrixMap out_mat(tensor_out.flat<T>().data(), params.depth, in launch()
704 const Tensor& tensor_out = context->input(1); in Compute() local
[all …]
Dpooling_ops_common.cc224 Tensor* tensor_out = nullptr; in Compute() local
226 context->allocate_output(0, tensor_out_shape, &tensor_out)); in Compute()
267 transformed_output = *tensor_out; in Compute()
272 auto& transformed_output = *tensor_out; in Compute()
416 tensor_out->tensor<RT, 4>()); in Compute()
444 TensorFormat data_format, const Tensor* tensor_in, const Tensor* tensor_out, in Compute() argument
448 (tensor_in && tensor_out)) in Compute()
464 if (tensor_out) { in Compute()
465 OP_REQUIRES(context, tensor_out->shape() == params.forward_output_shape(), in Compute()
468 ", but got ", tensor_out->shape())); in Compute()
[all …]
Dfractional_max_pool_op.cc240 const Tensor& tensor_out = context->input(1); in Compute() local
254 OP_REQUIRES(context, tensor_out.dims() == tensor_in_and_out_dims, in Compute()
257 tensor_out.DebugString())); in Compute()
258 OP_REQUIRES(context, tensor_out.NumElements() > 0, in Compute()
260 tensor_out.DebugString())); in Compute()
267 output_size[i] = tensor_out.dim_size(i); in Compute()
275 {1}, DataTypeToEnum<T>::v(), tensor_out.shape(), in Compute()
279 tensor_out.shape(), in Compute()
350 tensor_out.flat<T>().data(), output_size[3], in Compute()
Dcudnn_pooling_gpu.cc134 const Tensor* tensor_in, const Tensor* tensor_out, Tensor* input_backprop) { in Compute() argument
136 (tensor_in && tensor_out)) in Compute()
156 if (data_format == FORMAT_NHWC || tensor_out == nullptr) { in Compute()
163 transformed_output = *tensor_out; in Compute()
189 if (tensor_out != nullptr) { in Compute()
191 tensor_out->tensor<T, 5>(), in Compute()
Dmaxpooling_op.cc268 const Tensor& tensor_out = context->input(1); in Compute() local
274 OP_REQUIRES(context, tensor_out.dims() == 4, in Compute()
284 {1}, DataTypeToEnum<T>::v(), tensor_out.shape(), in Compute()
288 tensor_out.shape(), in Compute()
328 OP_REQUIRES(context, tensor_out.shape() == params.forward_output_shape(), in Compute()
331 ", but got ", tensor_out.shape())); in Compute()
394 const Tensor& tensor_out = context->input(1); in Compute() local
400 OP_REQUIRES(context, tensor_out.dims() == 4, in Compute()
443 data_format_, &tensor_in, &tensor_out, in Compute()
501 const Tensor& tensor_out = context->input(1); in Compute() local
[all …]
Dcudnn_pooling_gpu.h62 const Tensor* tensor_in, const Tensor* tensor_out,
Dpooling_ops_common_gpu.h64 const Tensor* tensor_out, const Tensor& out_backprop,
/external/pytorch/test/jit/
Dtest_complexity.py47 tensor_out = False
50 tensor_out = True
52 num_non_tensor += int(not tensor_out)
/external/tensorflow/tensorflow/lite/kernels/
Dbatch_matmul.cc393 const scalar* input, TfLiteTensor* tensor_out, in TransposeRowsColumnsImpl() argument
413 TfLiteTensor* tensor_out) { in TransposeRowsColumns() argument
416 tensor_out, in TransposeRowsColumns()
417 GetTensorData<float>(tensor_out)); in TransposeRowsColumns()
421 tensor_in, GetTensorData<int8_t>(tensor_in), tensor_out, in TransposeRowsColumns()
422 GetTensorData<int8_t>(tensor_out)); in TransposeRowsColumns()
426 tensor_in, GetTensorData<int16_t>(tensor_in), tensor_out, in TransposeRowsColumns()
427 GetTensorData<int16_t>(tensor_out)); in TransposeRowsColumns()
/external/tensorflow/tensorflow/core/util/
Dmkl_util.h789 inline void AllocTmpBuffer(OpKernelContext* context, Tensor* tensor_out, in AllocTmpBuffer() argument
795 tf_shape, tensor_out)); in AllocTmpBuffer()
796 *buf_out = static_cast<void*>(tensor_out->flat<T>().data()); in AllocTmpBuffer()
800 inline void AllocTmpBuffer(OpKernelContext* context, Tensor* tensor_out, in AllocTmpBuffer() argument
803 tf_shape, tensor_out)); in AllocTmpBuffer()
/external/pytorch/torch/testing/_internal/distributed/
Ddistributed_test.py2528 self, tensor_out, tensor_in, group_id, rank, cuda=True, rank_to_GPU=None argument
2532 tensor_out = tensor_out.cuda(rank_to_GPU[rank][0])
2533 tensor_shapes = [tensor_out.shape]
2538 tensor_out,
2546 return tensor_out
2556 tensor_out = torch.zeros(size, dtype=torch.int64)
2560 tensor_out = self._reduce_scatter_tensor_helper(
2561 tensor_out, tensor_in, group_id, rank, True, rank_to_GPU
2565 self.assertEqual(tensor_out, expected_tensor)
2570 tensor_out = self._reduce_scatter_tensor_helper(
[all …]
/external/pytorch/test/onnx/
Dtest_pytorch_onnx_onnxruntime.py508 tensor_out: Tensor
515 tensor_out=a,