/external/ComputeLibrary/src/core/CPP/kernels/ |
D | CPPNonMaximumSuppressionKernel.cpp | 38 …t ITensorInfo *bboxes, const ITensorInfo *scores, const ITensorInfo *output_indices, unsigned int … in validate_arguments() argument 41 ARM_COMPUTE_RETURN_ERROR_ON_NULLPTR(bboxes, scores, output_indices); in validate_arguments() 43 ARM_COMPUTE_RETURN_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(output_indices, 1, DataType::S32); in validate_arguments() 46 …ARM_COMPUTE_RETURN_ERROR_ON_MSG(output_indices->num_dimensions() > 1, "The indices must be 1-D int… in validate_arguments() 48 …ARM_COMPUTE_RETURN_ERROR_ON_MSG(output_indices->dimension(0) == 0, "Indices tensor must be bigger … in validate_arguments() 62 …ernel::configure(const ITensor *input_bboxes, const ITensor *input_scores, ITensor *output_indices, in configure() argument 65 ARM_COMPUTE_ERROR_ON_NULLPTR(input_bboxes, input_scores, output_indices); in configure() 66 …OW_ON(validate_arguments(input_bboxes->info(), input_scores->info(), output_indices->info(), max_o… in configure() 68 …auto_init_if_empty(*output_indices->info(), TensorShape(max_output_size), 1, DataType::U8, Quantiz… in configure() 72 _output_indices = output_indices; in configure() [all …]
|
/external/tensorflow/tensorflow/core/kernels/ |
D | sparse_slice_grad_op.cc | 80 const Tensor *backprop_val_grad, *input_indices, *output_indices, *input_start; in Compute() local 84 OP_REQUIRES_OK(ctx, ctx->input("output_indices", &output_indices)); in Compute() 88 TensorShapeUtils::IsMatrix(output_indices->shape()), in Compute() 93 output_indices->shape().DebugString())); in Compute() 101 input_indices->dim_size(1) == output_indices->dim_size(1), in Compute() 104 output_indices->dim_size(1))); in Compute() 106 ctx, output_indices->dim_size(0) <= input_indices->dim_size(0), in Compute() 109 output_indices->dim_size(0), " and ", in Compute() 112 ctx, backprop_val_grad->NumElements() == output_indices->dim_size(0), in Compute() 116 output_indices->dim_size(0))); in Compute() [all …]
|
D | deserialize_sparse_variant_op.cc | 59 const Tensor* output_indices; in Compute() local 69 output_shape->NumElements(), &output_indices, in Compute() 71 context->set_output(0, *output_indices); in Compute() 152 Tensor* output_indices; in Compute() local 159 &output_indices)); in Compute() 169 int64_t* output_indices_data = output_indices->matrix<int64_t>().data(); in Compute() 304 int index, int expected_rank, const Tensor** output_indices, in GetAndValidateSparseTensorIndicesAndValues() argument 307 *output_indices = serialized_indices.get<Tensor>(); in GetAndValidateSparseTensorIndicesAndValues() 308 if (*output_indices == nullptr) { in GetAndValidateSparseTensorIndicesAndValues() 312 if ((*output_indices)->dtype() != DT_INT64) { in GetAndValidateSparseTensorIndicesAndValues() [all …]
|
D | deserialize_sparse_string_op.cc | 82 Tensor output_indices; in Compute() local 89 &output_indices, &output_values, &output_shape)); in Compute() 90 context->set_output(0, output_indices); in Compute() 105 Tensor output_indices; in Compute() local 111 serialized_sparse_t(i, 2), dtype_, i, &output_indices, in Compute() 113 int64_t num_entries = output_indices.dim_size(0); in Compute() 114 int rank = output_indices.dim_size(1); in Compute() 119 const auto& output_indices_t = output_indices.matrix<int64_t>(); in Compute() 232 Tensor* output_indices, Tensor* output_values, Tensor* output_shape) { in GetAndValidateSparseTensor() argument 234 TF_RETURN_IF_ERROR(this->Deserialize(serialized_indices, output_indices)); in GetAndValidateSparseTensor() [all …]
|
D | sparse_slice_op_gpu.cu.cc | 94 int64_t* __restrict__ output_indices, T* __restrict__ output_values) { in SparseSliceGatherKernel() argument 101 output_indices[output_nz * dims + dim] = in SparseSliceGatherKernel() 152 Tensor* output_indices = nullptr; in operator ()() local 154 context, context->allocate_output(0, {0, dims}, &output_indices), in operator ()() 216 Tensor* output_indices = nullptr; in operator ()() local 219 context->allocate_output(0, {output_nnz, dims}, &output_indices), in operator ()() 221 int64_t* output_indices_ptr = output_indices->matrix<int64_t>().data(); in operator ()()
|
D | sparse_fill_empty_rows_op_gpu.cu.cc | 132 const Tindex* num_new_rows_before, Tindex* output_indices, T* output_values, in ScatterInputElementsKernel() argument 139 output_indices[output_i * rank + dim] = indices[input_i * rank + dim]; in ScatterInputElementsKernel() 154 const bool* empty_row_indicator, Tindex* output_indices, T* output_values) { in ScatterNewElementsKernel() argument 160 output_indices[output_i * rank + dim] = (dim == 0) ? row : 0; in ScatterNewElementsKernel() 190 Tindex* output_indices; in operator ()() local 194 context, N, rank, /*num_empty_rows=*/0, &output_indices, in operator ()() 374 Tindex* output_indices; in operator ()() local 380 context, N, rank, num_empty_rows, &output_indices, &output_values, in operator ()() 403 num_empty_rows_through, output_indices, in operator ()() 413 input_row_ends, empty_row_indicator, output_indices, in operator ()() [all …]
|
D | sparse_split_op_gpu.cu.cc | 168 Index* __restrict__* __restrict__ output_indices = in SparseSplitScatterKernel() local 183 output_indices[slice_index][slice_nz * rank + dim] = in SparseSplitScatterKernel() 313 GpuDeviceArrayOnHost<Index*> output_indices(context, num_split); in operator ()() local 319 &output_indices, &output_values), in operator ()() 330 input_values_ptr, output_indices.data(), output_values.data()), in operator ()() 347 GpuDeviceArrayOnHost<Index*>* output_indices, in AllocateOutputs() 349 TF_RETURN_IF_ERROR(output_indices->Init()); in AllocateOutputs() 358 output_indices->Set(slice_index, output_inds->matrix<Index>().data()); in AllocateOutputs() 372 TF_RETURN_IF_ERROR(output_indices->Finalize()); in AllocateOutputs()
|
D | reshape_util_gpu.cu.cc | 35 Tindex* __restrict__ output_indices) { in ReshapeSparseTensorKernel() argument 38 Tindex* output_index = &output_indices[sparse_index * output_rank]; in ReshapeSparseTensorKernel() 62 typename TTypes<int64_t>::Matrix output_indices) const { in operator ()() 103 /*output_indices=*/output_indices.data()); in operator ()()
|
D | sparse_tensors_map_ops.cc | 307 Tensor output_indices = Tensor(DT_INT64, {num_entries, rank - 1}); in Compute() local 310 auto output_indices_t = output_indices.matrix<int64_t>(); in Compute() 322 SparseTensor::Create(output_indices, output_values, in Compute() 399 const Tensor& output_indices = st.indices(); in Compute() local 403 OP_REQUIRES(context, TensorShapeUtils::IsMatrix(output_indices.shape()), in Compute() 407 output_indices.shape().DebugString())); in Compute() 420 int64_t num_entries = output_indices.dim_size(0); in Compute() 427 int rank = output_indices.dim_size(1); in Compute() 439 DT_INT64, TensorShape({num_entries, 1 + output_indices.dim_size(1)})); in Compute() 441 const auto& output_indices_t = output_indices.matrix<int64_t>(); in Compute()
|
D | sparse_slice_grad_op_gpu.cu.cc | 97 MultiIndexSearchIterator output_indices, in SparseSliceGradKernel() argument 104 const int64 output_nz = gpu_helper::lower_bound(output_indices, output_nnz, in SparseSliceGradKernel() 107 output_indices[output_nz] == input_indices[input_nz]) { in SparseSliceGradKernel()
|
/external/libtextclassifier/native/utils/tflite/ |
D | dist_diversification.cc | 83 TfLiteTensor& output_indices = in AllocateOutputIndexes() local 86 return context->ResizeTensor(context, &output_indices, in AllocateOutputIndexes() 97 TfLiteTensor& output_indices = in Prepare() local 100 tflite::SetTensorToDynamic(&output_indices); in Prepare() 110 TfLiteTensor& output_indices = in Eval() local 113 if (tflite::IsDynamicTensor(&output_indices)) { in Eval() 133 std::copy(indices.begin(), indices.end(), output_indices.data.i32); in Eval() 134 std::fill_n(output_indices.data.i32 + indices.size(), in Eval()
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/tasks/ |
D | pooling.cc | 130 bool output_indices, GPUOperation* op) { in GetMaxPoolingKernelCode() argument 133 if (output_indices) { in GetMaxPoolingKernelCode() 174 if (output_indices) { in GetMaxPoolingKernelCode() 193 if (output_indices) { in GetMaxPoolingKernelCode() 211 if (output_indices) { in GetMaxPoolingKernelCode() 243 if (output_indices) { in GetMaxPoolingKernelCode() 265 op.code_ = GetMaxPoolingKernelCode(definition, attr.output_indices, &op); in CreatePooling() 287 op.code_ = GetMaxPoolingKernelCode(definition, attr.output_indices, &op); in CreatePooling()
|
/external/tensorflow/tensorflow/compiler/xla/tests/ |
D | reverse_test.cc | 93 std::vector<int64_t> output_indices(spec.input_dims.size()); in TEST_P() local 96 output_indices[i] = indices[i]; in TEST_P() 100 output_indices[dim] = (spec.input_dims[dim] - 1) - indices[dim]; in TEST_P() 102 expected.Set<float>(output_indices, value); in TEST_P()
|
/external/tensorflow/tensorflow/core/kernels/image/ |
D | non_max_suppression_op.cu.cc | 328 Tensor* output_indices = nullptr; in DoNMS() local 330 context->allocate_output(0, TensorShape({0}), &output_indices)); in DoNMS() 402 Tensor* output_indices = nullptr; in DoNMS() local 408 &output_indices)); in DoNMS() 427 Tensor* output_indices = nullptr; in DoNMS() local 431 0, TensorShape({max_output_size}), &output_indices)); in DoNMS() 436 output_indices->flat<int>().data())); in DoNMS() 440 &output_indices)); in DoNMS() 451 (*output_indices).flat<int>().data())); in DoNMS() 546 Tensor* output_indices = nullptr; in Compute() local [all …]
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/ |
D | quantization_util.cc | 108 TfLiteContext* context, const std::vector<uint32_t>& output_indices, in QuantizeOutputs() argument 110 for (auto index : output_indices) { in QuantizeOutputs() 118 TfLiteContext* context, const std::vector<int64_t>& output_indices, in QuantizeOutputs() argument 120 for (auto index : output_indices) { in QuantizeOutputs()
|
D | quantization_util.h | 49 TfLiteContext* context, const std::vector<uint32_t>& output_indices, 53 TfLiteContext* context, const std::vector<int64_t>& output_indices,
|
/external/tensorflow/tensorflow/core/ops/compat/ops_history_v1/ |
D | SparseReduceMaxSparse.pbtxt | 20 name: "output_indices" 75 name: "output_indices" 132 name: "output_indices" 190 name: "output_indices"
|
D | SparseSparseMaximum.pbtxt | 28 name: "output_indices" 80 name: "output_indices" 134 name: "output_indices" 189 name: "output_indices"
|
D | SparseReduceSumSparse.pbtxt | 20 name: "output_indices" 80 name: "output_indices" 142 name: "output_indices" 205 name: "output_indices"
|
D | SparseSparseMinimum.pbtxt | 28 name: "output_indices" 85 name: "output_indices" 144 name: "output_indices" 204 name: "output_indices"
|
/external/tensorflow/tensorflow/core/ops/compat/ops_history_v2/ |
D | SparseSparseMaximum.pbtxt | 28 name: "output_indices" 80 name: "output_indices" 134 name: "output_indices" 189 name: "output_indices"
|
D | SparseReduceMaxSparse.pbtxt | 20 name: "output_indices" 75 name: "output_indices" 132 name: "output_indices" 190 name: "output_indices"
|
D | SparseReduceSumSparse.pbtxt | 20 name: "output_indices" 80 name: "output_indices" 142 name: "output_indices" 205 name: "output_indices"
|
D | SparseSparseMinimum.pbtxt | 28 name: "output_indices" 85 name: "output_indices" 144 name: "output_indices" 204 name: "output_indices"
|
/external/ComputeLibrary/arm_compute/core/CPP/kernels/ |
D | CPPNonMaximumSuppressionKernel.h | 68 …e(const ITensor *input_bboxes, const ITensor *input_scores, ITensor *output_indices, unsigned int … 80 …o *input_bboxes, const ITensorInfo *input_scores, const ITensorInfo *output_indices, unsigned int …
|