• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 //
2 // Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
3 // SPDX-License-Identifier: MIT
4 //
5 #pragma once
6 
7 #include <armnn/Tensor.hpp>
8 
9 #include <armnn/backends/IBackendInternal.hpp>
10 #include <armnn/backends/IMemoryManager.hpp>
11 #include <armnn/backends/Workload.hpp>
12 #include <armnn/backends/WorkloadInfo.hpp>
13 
14 namespace armnn
15 {
16 class ITensorHandle;
17 } // namespace armnn
18 
19 namespace
20 {
21 
22 template <typename QueueDescriptor>
AddInputToWorkload(QueueDescriptor & descriptor,armnn::WorkloadInfo & info,const armnn::TensorInfo & tensorInfo,armnn::ITensorHandle * tensorHandle)23 void AddInputToWorkload(QueueDescriptor& descriptor,
24     armnn::WorkloadInfo& info,
25     const armnn::TensorInfo& tensorInfo,
26     armnn::ITensorHandle* tensorHandle)
27 {
28     descriptor.m_Inputs.push_back(tensorHandle);
29     info.m_InputTensorInfos.push_back(tensorInfo);
30 }
31 
32 template <typename QueueDescriptor>
AddOutputToWorkload(QueueDescriptor & descriptor,armnn::WorkloadInfo & info,const armnn::TensorInfo & tensorInfo,armnn::ITensorHandle * tensorHandle)33 void AddOutputToWorkload(QueueDescriptor& descriptor,
34     armnn::WorkloadInfo& info,
35     const armnn::TensorInfo& tensorInfo,
36     armnn::ITensorHandle* tensorHandle)
37 {
38     descriptor.m_Outputs.push_back(tensorHandle);
39     info.m_OutputTensorInfos.push_back(tensorInfo);
40 }
41 
42 template <typename QueueDescriptor>
SetWorkloadInput(QueueDescriptor & descriptor,armnn::WorkloadInfo & info,unsigned int index,const armnn::TensorInfo & tensorInfo,armnn::ITensorHandle * tensorHandle)43 void SetWorkloadInput(QueueDescriptor& descriptor,
44     armnn::WorkloadInfo& info,
45     unsigned int index,
46     const armnn::TensorInfo& tensorInfo,
47     armnn::ITensorHandle* tensorHandle)
48 {
49     descriptor.m_Inputs[index] = tensorHandle;
50     info.m_InputTensorInfos[index] = tensorInfo;
51 }
52 
53 template <typename QueueDescriptor>
SetWorkloadOutput(QueueDescriptor & descriptor,armnn::WorkloadInfo & info,unsigned int index,const armnn::TensorInfo & tensorInfo,armnn::ITensorHandle * tensorHandle)54 void SetWorkloadOutput(QueueDescriptor& descriptor,
55     armnn::WorkloadInfo& info,
56     unsigned int index,
57     const armnn::TensorInfo& tensorInfo,
58     armnn::ITensorHandle* tensorHandle)
59 {
60     descriptor.m_Outputs[index] = tensorHandle;
61     info.m_OutputTensorInfos[index] = tensorInfo;
62 }
63 
ExecuteWorkload(armnn::IWorkload & workload,const armnn::IBackendInternal::IMemoryManagerSharedPtr & memoryManager,bool memoryManagementRequested=true)64 inline void ExecuteWorkload(armnn::IWorkload& workload,
65                             const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager,
66                             bool memoryManagementRequested = true)
67 {
68     const bool manageMemory = memoryManager && memoryManagementRequested;
69 
70     // Acquire working memory (if needed)
71     if (manageMemory)
72     {
73         memoryManager->Acquire();
74     }
75 
76     // Perform PostAllocationConfiguration
77     workload.PostAllocationConfigure();
78 
79     // Execute the workload
80     workload.Execute();
81 
82     // Release working memory (if needed)
83     if (manageMemory)
84     {
85         memoryManager->Release();
86     }
87 }
88 
GetBiasTypeFromWeightsType(armnn::Optional<armnn::DataType> weightsType)89 inline armnn::Optional<armnn::DataType> GetBiasTypeFromWeightsType(armnn::Optional<armnn::DataType> weightsType)
90 {
91     if (!weightsType)
92     {
93         return weightsType;
94     }
95 
96     switch(weightsType.value())
97     {
98         case armnn::DataType::BFloat16:
99         case armnn::DataType::Float16:
100         case armnn::DataType::Float32:
101             return weightsType;
102         case armnn::DataType::QAsymmS8:
103         case armnn::DataType::QAsymmU8:
104         case armnn::DataType::QSymmS8:
105         case armnn::DataType::QSymmS16:
106             return armnn::DataType::Signed32;
107         default:
108             ARMNN_ASSERT_MSG(false, "GetBiasTypeFromWeightsType(): Unsupported data type.");
109     }
110     return armnn::EmptyOptional();
111 }
112 
113 } // anonymous namespace
114