/external/ComputeLibrary/src/core/ |
D | Validate.cpp | 26 arm_compute::Status arm_compute::error_on_mismatching_windows(const char *function, const char *fil… in error_on_mismatching_windows() 27 … const arm_compute::Window &full, const arm_compute::Window &win) in error_on_mismatching_windows() 32 for(size_t i = 0; i < arm_compute::Coordinates::num_max_dimensions; ++i) in error_on_mismatching_windows() 38 return arm_compute::Status{}; in error_on_mismatching_windows() 41 arm_compute::Status arm_compute::error_on_invalid_subwindow(const char *function, const char *file,… in error_on_invalid_subwindow() 42 … const arm_compute::Window &full, const arm_compute::Window &sub) in error_on_invalid_subwindow() 47 for(size_t i = 0; i < arm_compute::Coordinates::num_max_dimensions; ++i) in error_on_invalid_subwindow() 54 return arm_compute::Status{}; in error_on_invalid_subwindow() 57 arm_compute::Status arm_compute::error_on_window_not_collapsable_at_dimension(const char *function,… in error_on_window_not_collapsable_at_dimension() 58 … const arm_compute::Window &full, const arm_compute::Window &window, const int dim) in error_on_window_not_collapsable_at_dimension() [all …]
|
/external/armnn/src/backends/aclCommon/ |
D | ArmComputeTensorUtils.cpp | 19 arm_compute::DataType GetArmComputeDataType(armnn::DataType dataType, bool multiScales) in GetArmComputeDataType() 24 return arm_compute::DataType::BFLOAT16; in GetArmComputeDataType() 26 return arm_compute::DataType::U8; in GetArmComputeDataType() 28 return arm_compute::DataType::F16; in GetArmComputeDataType() 30 return arm_compute::DataType::F32; in GetArmComputeDataType() 32 return arm_compute::DataType::QASYMM8_SIGNED; in GetArmComputeDataType() 34 return arm_compute::DataType::QASYMM8; in GetArmComputeDataType() 36 return arm_compute::DataType::QSYMM16; in GetArmComputeDataType() 38 return arm_compute::DataType::S64; in GetArmComputeDataType() 41 … return multiScales ? arm_compute::DataType::QSYMM8_PER_CHANNEL : arm_compute::DataType::QSYMM8; in GetArmComputeDataType() [all …]
|
D | ArmComputeUtils.hpp | 27 inline arm_compute::NormalizationLayerInfo 57 …return arm_compute::NormalizationLayerInfo(arm_compute::NormType::CROSS_MAP, normSize, alpha, beta… in CreateAclNormalizationLayerInfoForL2Normalization() 60 inline arm_compute::ActivationLayerInfo::ActivationFunction 63 using AclActivationFunction = arm_compute::ActivationLayerInfo::ActivationFunction; in ConvertActivationFunctionToAclActivationFunction() 84 inline arm_compute::ActivationLayerInfo 87 …return arm_compute::ActivationLayerInfo(ConvertActivationFunctionToAclActivationFunction(actDesc.m… in ConvertActivationDescriptorToAclActivationLayerInfo() 91 inline arm_compute::ActivationLayerInfo 99 return arm_compute::ActivationLayerInfo(); in ConvertActivationDescriptorToAclActivationLayerInfo() 102 inline arm_compute::ActivationLayerInfo 112 return arm_compute::ActivationLayerInfo(); in ConvertAdditionalInfoToAclActivationLayerInfo() [all …]
|
/external/ComputeLibrary/src/core/utils/ |
D | AssemblyUtils.cpp | 26 namespace arm_compute namespace 70 arm_gemm::WeightFormat map_to_arm_gemm_weight_format(const arm_compute::WeightFormat &weight_format) in map_to_arm_gemm_weight_format() 76 case arm_compute::WeightFormat::UNSPECIFIED: in map_to_arm_gemm_weight_format() 79 case arm_compute::WeightFormat::ANY: in map_to_arm_gemm_weight_format() 82 case arm_compute::WeightFormat::OHWI: in map_to_arm_gemm_weight_format() 85 case arm_compute::WeightFormat::OHWIo2: in map_to_arm_gemm_weight_format() 88 case arm_compute::WeightFormat::OHWIo4: in map_to_arm_gemm_weight_format() 91 case arm_compute::WeightFormat::OHWIo8: in map_to_arm_gemm_weight_format() 94 case arm_compute::WeightFormat::OHWIo16: in map_to_arm_gemm_weight_format() 97 case arm_compute::WeightFormat::OHWIo32: in map_to_arm_gemm_weight_format() [all …]
|
/external/armnn/src/backends/neon/workloads/ |
D | NeonUnidirectionalSequenceLstmFloatWorkload.hpp | 35 mutable std::unique_ptr<arm_compute::NEPermute> m_Permute1; 36 mutable std::unique_ptr<arm_compute::IFunction> m_Splitter; 37 mutable std::vector<std::unique_ptr<arm_compute::NELSTMLayer>> m_Layers; 38 mutable std::unique_ptr<arm_compute::NEConcatenateLayer> m_Concat; 40 mutable std::unique_ptr<arm_compute::NEPermute> m_Permute2; 45 std::unique_ptr<arm_compute::Tensor> m_InputToInputWeightsTensor; 46 std::unique_ptr<arm_compute::Tensor> m_InputToForgetWeightsTensor; 47 std::unique_ptr<arm_compute::Tensor> m_InputToCellWeightsTensor; 48 std::unique_ptr<arm_compute::Tensor> m_InputToOutputWeightsTensor; 49 std::unique_ptr<arm_compute::Tensor> m_RecurrentToInputWeightsTensor; [all …]
|
D | NeonUnidirectionalSequenceLstmWorkload.hpp | 34 mutable std::unique_ptr<arm_compute::NEPermute> m_Permute1; 35 mutable std::unique_ptr<arm_compute::IFunction> m_Splitter; 36 mutable std::vector<std::unique_ptr<arm_compute::NEQLSTMLayer>> m_Layers; 37 mutable std::unique_ptr<arm_compute::NEConcatenateLayer> m_Concat; 38 mutable std::unique_ptr<arm_compute::NEPermute> m_Permute2; 43 std::unique_ptr<arm_compute::Tensor> m_InputToInputWeightsTensor; 44 std::unique_ptr<arm_compute::Tensor> m_InputToForgetWeightsTensor; 45 std::unique_ptr<arm_compute::Tensor> m_InputToCellWeightsTensor; 46 std::unique_ptr<arm_compute::Tensor> m_InputToOutputWeightsTensor; 47 std::unique_ptr<arm_compute::Tensor> m_RecurrentToInputWeightsTensor; [all …]
|
D | NeonUnidirectionalSequenceLstmFloatWorkload.cpp | 41 …const arm_compute::ITensor& input = static_cast<IAclTensorHandle*>(m_Data.m_Inputs[0])->GetTensor(… in NeonUnidirectionalSequenceLstmFloatWorkload() 42 arm_compute::ITensor& output = static_cast<IAclTensorHandle*>(m_Data.m_Outputs[2])->GetTensor(); in NeonUnidirectionalSequenceLstmFloatWorkload() 47 …arm_compute::DataType armComputeDataType = static_cast<IAclTensorHandle*>(m_Data.m_Inputs[0])->Get… in NeonUnidirectionalSequenceLstmFloatWorkload() 68 std::unique_ptr<arm_compute::NEPermute> layer(new arm_compute::NEPermute()); in NeonUnidirectionalSequenceLstmFloatWorkload() 76 layer->configure(&input, &m_PermuteFirstOut, arm_compute::PermutationVector(0U,2U,1U)); in NeonUnidirectionalSequenceLstmFloatWorkload() 85 arm_compute::Tensor splitter_out; in NeonUnidirectionalSequenceLstmFloatWorkload() 86 arm_compute::Tensor concat_in; in NeonUnidirectionalSequenceLstmFloatWorkload() 131 std::unique_ptr<arm_compute::NESplit> split_layer(new arm_compute::NESplit()); in NeonUnidirectionalSequenceLstmFloatWorkload() 149 arm_compute::LSTMParams<arm_compute::ITensor> lstm_param; in NeonUnidirectionalSequenceLstmFloatWorkload() 151 m_InputToForgetWeightsTensor = std::make_unique<arm_compute::Tensor>(); in NeonUnidirectionalSequenceLstmFloatWorkload() [all …]
|
D | NeonUnidirectionalSequenceLstmWorkload.cpp | 43 …const arm_compute::ITensor& input = static_cast<IAclTensorHandle*>(m_Data.m_Inputs[0])->Ge… in NeonUnidirectionalSequenceLstmWorkload() 44 …arm_compute::ITensor& outputStateIn = static_cast<IAclTensorHandle*>(m_Data.m_Inputs[1])->Ge… in NeonUnidirectionalSequenceLstmWorkload() 45 …const arm_compute::ITensor& cellStateIn = static_cast<IAclTensorHandle*>(m_Data.m_Inputs[2])->Ge… in NeonUnidirectionalSequenceLstmWorkload() 47 …arm_compute::ITensor& outputStateOut = static_cast<IAclTensorHandle*>(m_Data.m_Outputs[0])->GetTen… in NeonUnidirectionalSequenceLstmWorkload() 48 …arm_compute::ITensor& cellStateOut = static_cast<IAclTensorHandle*>(m_Data.m_Outputs[1])->GetTen… in NeonUnidirectionalSequenceLstmWorkload() 49 …arm_compute::ITensor& output = static_cast<IAclTensorHandle*>(m_Data.m_Outputs[2])->GetTen… in NeonUnidirectionalSequenceLstmWorkload() 70 std::unique_ptr<arm_compute::NEPermute> layer(new arm_compute::NEPermute()); in NeonUnidirectionalSequenceLstmWorkload() 78 layer->configure(&input, &m_PermuteFirstOut, arm_compute::PermutationVector(0U,2U,1U)); in NeonUnidirectionalSequenceLstmWorkload() 87 arm_compute::Tensor splitter_out; in NeonUnidirectionalSequenceLstmWorkload() 88 arm_compute::Tensor concat_in; in NeonUnidirectionalSequenceLstmWorkload() [all …]
|
D | NeonConstantWorkload.cpp | 20 arm_compute::Status NeonConstantWorkloadValidate(const TensorInfo& output) in NeonConstantWorkloadValidate() 22 …const arm_compute::TensorInfo neonOutputInfo = armcomputetensorutils::BuildArmComputeTensorInfo(ou… in NeonConstantWorkloadValidate() 24 std::array<arm_compute::DataType,9> supportedTypes = { in NeonConstantWorkloadValidate() 25 arm_compute::DataType::BFLOAT16, in NeonConstantWorkloadValidate() 26 arm_compute::DataType::F16, in NeonConstantWorkloadValidate() 27 arm_compute::DataType::F32, in NeonConstantWorkloadValidate() 28 arm_compute::DataType::QASYMM8, in NeonConstantWorkloadValidate() 29 arm_compute::DataType::QASYMM8_SIGNED, in NeonConstantWorkloadValidate() 30 arm_compute::DataType::QSYMM16, in NeonConstantWorkloadValidate() 31 arm_compute::DataType::QSYMM8, in NeonConstantWorkloadValidate() [all …]
|
D | NeonQLstmWorkload.hpp | 26 mutable arm_compute::NEQLSTMLayer m_QLstmLayer; 28 std::unique_ptr<arm_compute::Tensor> m_InputToInputWeightsTensor; 29 std::unique_ptr<arm_compute::Tensor> m_InputToForgetWeightsTensor; 30 std::unique_ptr<arm_compute::Tensor> m_InputToCellWeightsTensor; 31 std::unique_ptr<arm_compute::Tensor> m_InputToOutputWeightsTensor; 33 std::unique_ptr<arm_compute::Tensor> m_RecurrentToInputWeightsTensor; 34 std::unique_ptr<arm_compute::Tensor> m_RecurrentToForgetWeightsTensor; 35 std::unique_ptr<arm_compute::Tensor> m_RecurrentToCellWeightsTensor; 36 std::unique_ptr<arm_compute::Tensor> m_RecurrentToOutputWeightsTensor; 38 std::unique_ptr<arm_compute::Tensor> m_CellToInputWeightsTensor; [all …]
|
D | NeonLstmFloatWorkload.hpp | 30 mutable arm_compute::NELSTMLayer m_LstmLayer; 32 std::unique_ptr<arm_compute::Tensor> m_InputToInputWeightsTensor; 33 std::unique_ptr<arm_compute::Tensor> m_InputToForgetWeightsTensor; 34 std::unique_ptr<arm_compute::Tensor> m_InputToCellWeightsTensor; 35 std::unique_ptr<arm_compute::Tensor> m_InputToOutputWeightsTensor; 36 std::unique_ptr<arm_compute::Tensor> m_RecurrentToInputWeightsTensor; 37 std::unique_ptr<arm_compute::Tensor> m_RecurrentToForgetWeightsTensor; 38 std::unique_ptr<arm_compute::Tensor> m_RecurrentToCellWeightsTensor; 39 std::unique_ptr<arm_compute::Tensor> m_RecurrentToOutputWeightsTensor; 40 std::unique_ptr<arm_compute::Tensor> m_CellToInputWeightsTensor; [all …]
|
/external/armnn/src/backends/cl/workloads/ |
D | ClUnidirectionalSequenceLstmFloatWorkload.hpp | 27 … const arm_compute::CLCompileContext& clCompileContext); 37 mutable std::unique_ptr<arm_compute::CLPermute> m_Permute1; 38 mutable std::unique_ptr<arm_compute::IFunction> m_Splitter; 39 mutable std::vector<std::unique_ptr<arm_compute::CLLSTMLayer>> m_Layers; 40 mutable std::unique_ptr<arm_compute::CLConcatenateLayer> m_Concat; 42 mutable std::unique_ptr<arm_compute::CLPermute> m_Permute2; 47 std::unique_ptr<arm_compute::CLTensor> m_InputToInputWeightsTensor; 48 std::unique_ptr<arm_compute::CLTensor> m_InputToForgetWeightsTensor; 49 std::unique_ptr<arm_compute::CLTensor> m_InputToCellWeightsTensor; 50 std::unique_ptr<arm_compute::CLTensor> m_InputToOutputWeightsTensor; [all …]
|
D | ClUnidirectionalSequenceLstmFloatWorkload.cpp | 34 const arm_compute::CLCompileContext& clCompileContext) in ClUnidirectionalSequenceLstmFloatWorkload() 43 …const arm_compute::ICLTensor& input = static_cast<IClTensorHandle*>(m_Data.m_Inputs[0])->GetTensor… in ClUnidirectionalSequenceLstmFloatWorkload() 44 … arm_compute::ICLTensor& output = static_cast<IClTensorHandle*>(m_Data.m_Outputs[2])->GetTensor(); in ClUnidirectionalSequenceLstmFloatWorkload() 49 …arm_compute::DataType armComputeDataType = static_cast<IClTensorHandle*>(m_Data.m_Inputs[0])->GetD… in ClUnidirectionalSequenceLstmFloatWorkload() 70 std::unique_ptr<arm_compute::CLPermute> layer(new arm_compute::CLPermute()); in ClUnidirectionalSequenceLstmFloatWorkload() 78 …layer->configure(clCompileContext, &input, &m_PermuteFirstOut, arm_compute::PermutationVector(0U,2… in ClUnidirectionalSequenceLstmFloatWorkload() 87 arm_compute::CLTensor splitter_out; in ClUnidirectionalSequenceLstmFloatWorkload() 88 arm_compute::CLTensor concat_in; in ClUnidirectionalSequenceLstmFloatWorkload() 133 std::unique_ptr<arm_compute::CLSplit> split_layer(new arm_compute::CLSplit()); in ClUnidirectionalSequenceLstmFloatWorkload() 151 arm_compute::LSTMParams<arm_compute::ICLTensor> lstm_param; in ClUnidirectionalSequenceLstmFloatWorkload() [all …]
|
D | ClQLstmWorkload.hpp | 24 const arm_compute::CLCompileContext& clCompileContext); 28 mutable arm_compute::CLQLSTMLayer m_QLstmLayer; 30 std::unique_ptr<arm_compute::CLTensor> m_InputToInputWeightsTensor; 31 std::unique_ptr<arm_compute::CLTensor> m_InputToForgetWeightsTensor; 32 std::unique_ptr<arm_compute::CLTensor> m_InputToCellWeightsTensor; 33 std::unique_ptr<arm_compute::CLTensor> m_InputToOutputWeightsTensor; 35 std::unique_ptr<arm_compute::CLTensor> m_RecurrentToInputWeightsTensor; 36 std::unique_ptr<arm_compute::CLTensor> m_RecurrentToForgetWeightsTensor; 37 std::unique_ptr<arm_compute::CLTensor> m_RecurrentToCellWeightsTensor; 38 std::unique_ptr<arm_compute::CLTensor> m_RecurrentToOutputWeightsTensor; [all …]
|
D | ClLstmFloatWorkload.hpp | 23 const arm_compute::CLCompileContext& clCompileContext); 31 mutable arm_compute::CLLSTMLayer m_LstmLayer; 34 std::unique_ptr<arm_compute::CLTensor> m_InputToInputWeightsTensor; 35 std::unique_ptr<arm_compute::CLTensor> m_InputToForgetWeightsTensor; 36 std::unique_ptr<arm_compute::CLTensor> m_InputToCellWeightsTensor; 37 std::unique_ptr<arm_compute::CLTensor> m_InputToOutputWeightsTensor; 38 std::unique_ptr<arm_compute::CLTensor> m_RecurrentToInputWeightsTensor; 39 std::unique_ptr<arm_compute::CLTensor> m_RecurrentToForgetWeightsTensor; 40 std::unique_ptr<arm_compute::CLTensor> m_RecurrentToCellWeightsTensor; 41 std::unique_ptr<arm_compute::CLTensor> m_RecurrentToOutputWeightsTensor; [all …]
|
D | ClConstantWorkload.cpp | 18 arm_compute::Status ClConstantWorkloadValidate(const TensorInfo& output) in ClConstantWorkloadValidate() 20 …const arm_compute::TensorInfo neonOutputInfo = armcomputetensorutils::BuildArmComputeTensorInfo(ou… in ClConstantWorkloadValidate() 22 std::array<arm_compute::DataType,8> supportedTypes = { in ClConstantWorkloadValidate() 23 arm_compute::DataType::F16, in ClConstantWorkloadValidate() 24 arm_compute::DataType::F32, in ClConstantWorkloadValidate() 25 arm_compute::DataType::QASYMM8, in ClConstantWorkloadValidate() 26 arm_compute::DataType::QASYMM8_SIGNED, in ClConstantWorkloadValidate() 27 arm_compute::DataType::QSYMM16, in ClConstantWorkloadValidate() 28 arm_compute::DataType::QSYMM8, in ClConstantWorkloadValidate() 29 arm_compute::DataType::QSYMM8_PER_CHANNEL, in ClConstantWorkloadValidate() [all …]
|
D | ClBatchMatMulWorkload.cpp | 29 arm_compute::Status ClBatchMatMulValidate(const TensorInfo& inputX, in ClBatchMatMulValidate() 43 arm_compute::Status statusGEMM = arm_compute::Status(arm_compute::ErrorCode::OK); in ClBatchMatMulValidate() 44 arm_compute::Status statusPermuteX = arm_compute::Status(arm_compute::ErrorCode::OK); in ClBatchMatMulValidate() 45 arm_compute::Status statusPermuteY = arm_compute::Status(arm_compute::ErrorCode::OK); in ClBatchMatMulValidate() 53 arm_compute::TensorInfo aclPermutedXInfo = arm_compute::TensorInfo(); in ClBatchMatMulValidate() 54 arm_compute::TensorInfo aclPermutedYInfo = arm_compute::TensorInfo(); in ClBatchMatMulValidate() 65 statusPermuteX = arm_compute::CLPermute::validate(&aclInputXInfo, in ClBatchMatMulValidate() 79 statusPermuteY = arm_compute::CLPermute::validate(&aclInputYInfo, in ClBatchMatMulValidate() 84 const arm_compute::GEMMInfo& gemm_info = arm_compute::GEMMInfo(false, // is inputX reshaped in ClBatchMatMulValidate() 89 …statusGEMM = arm_compute::CLGEMM::validate(descriptor.m_TransposeX ? &aclPermutedXInfo : &aclInput… in ClBatchMatMulValidate() [all …]
|
D | ClResizeWorkload.cpp | 22 arm_compute::Status ClResizeWorkloadValidate(const TensorInfo& input, in ClResizeWorkloadValidate() 26 arm_compute::TensorInfo aclInputInfo = BuildArmComputeTensorInfo(input); in ClResizeWorkloadValidate() 27 arm_compute::TensorInfo aclOutputInfo = BuildArmComputeTensorInfo(output); in ClResizeWorkloadValidate() 29 arm_compute::DataLayout aclDataLayout = ConvertDataLayout(descriptor.m_DataLayout); in ClResizeWorkloadValidate() 33 arm_compute::InterpolationPolicy aclInterpolationPolicy = in ClResizeWorkloadValidate() 36 …arm_compute::SamplingPolicy samplingPolicy = descriptor.m_HalfPixelCenters ? arm_compute::Sampling… in ClResizeWorkloadValidate() 37 … arm_compute::SamplingPolicy::TOP_LEFT; in ClResizeWorkloadValidate() 39 return arm_compute::CLScale::validate(&aclInputInfo, in ClResizeWorkloadValidate() 41 arm_compute::ScaleKernelInfo(aclInterpolationPolicy, in ClResizeWorkloadValidate() 42 … arm_compute::BorderMode::REPLICATE, in ClResizeWorkloadValidate() [all …]
|
D | ClQLstmWorkload.cpp | 19 const arm_compute::CLCompileContext& clCompileContext) in ClQLstmWorkload() 28 arm_compute::LSTMParams<arm_compute::ICLTensor> qLstmParams; in ClQLstmWorkload() 31 m_InputToForgetWeightsTensor = std::make_unique<arm_compute::CLTensor>(); in ClQLstmWorkload() 34 m_InputToCellWeightsTensor = std::make_unique<arm_compute::CLTensor>(); in ClQLstmWorkload() 37 m_InputToOutputWeightsTensor = std::make_unique<arm_compute::CLTensor>(); in ClQLstmWorkload() 40 m_RecurrentToForgetWeightsTensor = std::make_unique<arm_compute::CLTensor>(); in ClQLstmWorkload() 43 m_RecurrentToCellWeightsTensor = std::make_unique<arm_compute::CLTensor>(); in ClQLstmWorkload() 46 m_RecurrentToOutputWeightsTensor = std::make_unique<arm_compute::CLTensor>(); in ClQLstmWorkload() 49 m_ForgetGateBiasTensor = std::make_unique<arm_compute::CLTensor>(); in ClQLstmWorkload() 52 m_CellBiasTensor = std::make_unique<arm_compute::CLTensor>(); in ClQLstmWorkload() [all …]
|
/external/armnn/src/backends/cl/ |
D | ClTensorHandle.hpp | 47 arm_compute::CLTensor& GetTensor() override { return m_Tensor; } in GetTensor() 48 arm_compute::CLTensor const& GetTensor() const override { return m_Tensor; } in GetTensor() 79 const_cast<arm_compute::CLTensor*>(&m_Tensor)->map(blocking); in Map() 83 virtual void Unmap() const override { const_cast<arm_compute::CLTensor*>(&m_Tensor)->unmap(); } in Unmap() 87 virtual arm_compute::DataType GetDataType() const override in GetDataType() 92 …virtual void SetMemoryGroup(const std::shared_ptr<arm_compute::IMemoryGroup>& memoryGroup) override in SetMemoryGroup() 94 m_MemoryGroup = PolymorphicPointerDowncast<arm_compute::MemoryGroup>(memoryGroup); in SetMemoryGroup() 147 case arm_compute::DataType::F32: in CopyOutTo() 151 case arm_compute::DataType::U8: in CopyOutTo() 152 case arm_compute::DataType::QASYMM8: in CopyOutTo() [all …]
|
D | ClImportTensorHandle.hpp | 47 arm_compute::CLTensor& GetTensor() override { return m_Tensor; } in GetTensor() 48 arm_compute::CLTensor const& GetTensor() const override { return m_Tensor; } in GetTensor() 62 virtual arm_compute::DataType GetDataType() const override in GetDataType() 67 …virtual void SetMemoryGroup(const std::shared_ptr<arm_compute::IMemoryGroup>& memoryGroup) override in SetMemoryGroup() 141 arm_compute::Status status = in Import() 162 arm_compute::Status status = in Import() 212 … arm_compute::CLKernelLibrary::get().get_device().getInfo<CL_DEVICE_GLOBAL_MEM_CACHELINE_SIZE>(); in ClImport() 223 buffer = clImportMemoryARM(arm_compute::CLKernelLibrary::get().context().get(), in ClImport() 228 buffer = clImportMemoryARM(arm_compute::CLKernelLibrary::get().context().get(), in ClImport() 238 arm_compute::Status status = m_Tensor.allocator()->import_memory(wrappedBuffer); in ClImport() [all …]
|
/external/armnn/src/backends/neon/ |
D | NeonTensorHandle.hpp | 51 arm_compute::ITensor& GetTensor() override { return m_Tensor; } in GetTensor() 52 arm_compute::ITensor const& GetTensor() const override { return m_Tensor; } in GetTensor() 75 virtual arm_compute::DataType GetDataType() const override in GetDataType() 80 …virtual void SetMemoryGroup(const std::shared_ptr<arm_compute::IMemoryGroup>& memoryGroup) override in SetMemoryGroup() 82 m_MemoryGroup = PolymorphicPointerDowncast<arm_compute::MemoryGroup>(memoryGroup); in SetMemoryGroup() 140 arm_compute::Status status = m_Tensor.allocator()->import_memory(memory); in Import() 161 arm_compute::Status status = m_Tensor.allocator()->import_memory(memory); in Import() 190 case arm_compute::DataType::F32: in CopyOutTo() 194 case arm_compute::DataType::U8: in CopyOutTo() 195 case arm_compute::DataType::QASYMM8: in CopyOutTo() [all …]
|
/external/ComputeLibrary/tests/framework/ |
D | Asserts.h | 35 namespace arm_compute 47 inline std::string make_printable(const arm_compute::WeightFormat wf) in make_printable() 49 return arm_compute::to_string(wf); in make_printable() 67 arm_compute::test::framework::Framework::get().print_test_info(msg); in ARM_COMPUTE_PRINT_INFO() 68 arm_compute::test::framework::Framework::get().log_info(msg.str()); in ARM_COMPUTE_PRINT_INFO() 69 arm_compute::test::framework::Framework::get().clear_test_info(); in ARM_COMPUTE_PRINT_INFO() 76 arm_compute::test::framework::Framework::get().add_test_info(info.str()); \ 89 …<< std::boolalpha << arm_compute::test::framework::make_printable(x) … 91 …<< std::boolalpha << arm_compute::test::framework::make_printable(y) … 93 …arm_compute::test::framework::Framework::get().print_test_info(msg); … [all …]
|
/external/ComputeLibrary/arm_compute/core/ |
D | Validate.h | 34 namespace arm_compute 49 for(unsigned int i = upper_dim; i < arm_compute::Dimensions<T>::num_max_dimensions; ++i) in have_different_dimensions() 89 arm_compute::Status operator()(const Dimensions<T> &dim) in operator() 93 return arm_compute::Status{}; in operator() 104 inline arm_compute::Status for_each_error(F &&) in for_each_error() 106 return arm_compute::Status{}; in for_each_error() 110 inline arm_compute::Status for_each_error(F &&func, T &&arg, Ts &&... args) in for_each_error() 114 return arm_compute::Status{}; in for_each_error() 147 inline arm_compute::Status error_on_nullptr(const char *function, const char *file, const int line,… 155 return arm_compute::Status{}; [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
D | runtime_conv2d_acl.h | 28 arm_compute::NEDepthwiseConvolutionLayer depthwise_conv; 29 arm_compute::NEArithmeticAddition add; 30 arm_compute::NEActivationLayer act; 31 arm_compute::Tensor input_tensor; 32 arm_compute::Tensor kernel_tensor; 33 arm_compute::Tensor bia_tensor; 34 arm_compute::Tensor output_tensor; 35 arm_compute::Tensor output_acc_tensor; 39 arm_compute::NEGEMMConvolutionLayer gemm_conv; 40 arm_compute::NEArithmeticAddition add; [all …]
|