/packages/modules/NeuralNetworks/runtime/test/ |
D | PreparedModelCallback.cpp | 22 bool deadObject, ErrorStatus errorStatus, const sp<V1_0::IPreparedModel>& preparedModel) { in notifyInternal() 43 V1_0::ErrorStatus errorStatus, const sp<V1_0::IPreparedModel>& preparedModel) { in notify() 48 V1_0::ErrorStatus errorStatus, const sp<V1_2::IPreparedModel>& preparedModel) { in notify_1_2() 53 V1_3::ErrorStatus errorStatus, const sp<V1_3::IPreparedModel>& preparedModel) { in notify_1_3()
|
D | TestExecution.cpp | 80 TestPreparedModelLatest(sp<V1_0::IPreparedModel> preparedModel, V1_3::ErrorStatus errorStatus) in TestPreparedModelLatest() 276 TestPreparedModel12(sp<V1_0::IPreparedModel> preparedModel, V1_3::ErrorStatus errorStatus) in TestPreparedModel12() 312 TestPreparedModel10(sp<V1_0::IPreparedModel> preparedModel, V1_3::ErrorStatus errorStatus) in TestPreparedModel10()
|
D | TestPartitioning.cpp | 3402 index, [&actual](const auto* preparedModel, IOType type, uint32_t) { in checkStepRolesOfInput() 3413 index, [&actual](const auto* preparedModel, IOType type, uint32_t) { in checkStepRolesOfOutput() 3425 index, [&actual](const auto* preparedModel, IOType type, uint32_t) { in checkStepRolesOfSourceOperand()
|
/packages/modules/NeuralNetworks/driver/sample/ |
D | SampleDriverUtils.cpp | 28 const sp<SamplePreparedModel>& preparedModel) { in notify() 36 const sp<SamplePreparedModel>& preparedModel) { in notify() 45 const sp<SamplePreparedModel>& preparedModel) { in notify()
|
D | CanonicalBurst.cpp | 33 Burst::Burst(std::shared_ptr<const PreparedModel> preparedModel) in Burst()
|
D | SampleDriver.cpp | 237 const sp<V1_3::IPreparedModel>& preparedModel) { in castToSamplePreparedModel() 257 auto getModel = [](const sp<V1_3::IPreparedModel>& preparedModel) -> const V1_3::Model* { in allocate() 383 const SamplePreparedModel* preparedModel) { in createRunTimePoolInfos() 462 const SamplePreparedModel* preparedModel, in asyncExecute() 518 const SamplePreparedModel* preparedModel, in executeBase() 582 const SamplePreparedModel* preparedModel, in executeSynchronouslyBase()
|
D | CanonicalDevice.cpp | 211 auto getModel = [](const SharedPreparedModel& preparedModel) -> const Model* { in allocate()
|
D | CanonicalPreparedModel.cpp | 40 const PreparedModel& preparedModel) { in createRunTimePoolInfos()
|
D | SampleDriverFloatXNNPACK.cpp | 1888 sp<SamplePreparedModelXNNPACK> preparedModel = new SamplePreparedModelXNNPACK( in prepareModelXNNPACK() local
|
/packages/modules/NeuralNetworks/driver/sample_aidl/ |
D | SampleDriverUtils.cpp | 40 const std::shared_ptr<aidl_hal::IPreparedModel>& preparedModel) { in notify() 120 std::shared_ptr<SamplePreparedModel> preparedModel = in prepareModelBase() local
|
D | SampleDriver.cpp | 131 const std::shared_ptr<aidl_hal::IPreparedModel>& preparedModel) { in castToSamplePreparedModel() 148 -> const aidl_hal::Model* { in allocate() 309 const SamplePreparedModel* preparedModel) { in createRunTimePoolInfos() 602 SampleBurst::SampleBurst(std::shared_ptr<SamplePreparedModel> preparedModel) in SampleBurst()
|
/packages/modules/NeuralNetworks/common/ |
D | AidlValidateHal.cpp | 53 const auto& preparedModel = preparedModels[role.modelIndex]; in validateMemoryDesc() local 67 const auto& preparedModel = preparedModels[role.modelIndex]; in validateMemoryDesc() local
|
D | ExecutionBurstServer.cpp | 56 DefaultBurstExecutorWithCache(V1_2::IPreparedModel* preparedModel) in DefaultBurstExecutorWithCache() 526 const MQDescriptorSync<FmqResultDatum>& resultChannel, V1_2::IPreparedModel* preparedModel, in create()
|
D | ValidateHal.cpp | 884 const auto& preparedModel = preparedModels[role.modelIndex]; in validateMemoryDesc() local 898 const auto& preparedModel = preparedModels[role.modelIndex]; in validateMemoryDesc() local
|
D | ExecutionBurstController.cpp | 478 const sp<V1_2::IPreparedModel>& preparedModel, in create()
|
D | Validation.cpp | 1280 const auto& preparedModel = preparedModels[role.modelIndex]; in validateMemoryDescImpl() local 1294 const auto& preparedModel = preparedModels[role.modelIndex]; in validateMemoryDescImpl() local
|
/packages/modules/NeuralNetworks/common/include/ |
D | DefaultExecution.h | 34 DefaultExecution(SharedPreparedModel preparedModel, Request request, MeasureTiming measure, in DefaultExecution()
|
/packages/modules/NeuralNetworks/shim_and_sl/ |
D | ShimDevice.cpp | 88 const ShimPreparedModel* castToShimPreparedModel(IPreparedModel* preparedModel) { in castToShimPreparedModel() 342 constexpr auto getCompilation = [](IPreparedModel* preparedModel) -> const ShimPreparedModel* { in allocate() 360 auto preparedModel = preparedModels[role.modelIndex]; in allocate() local 398 auto preparedModel = preparedModels[role.modelIndex]; in allocate() local 590 const std::shared_ptr<ShimPreparedModel> preparedModel = in prepareModel() local
|
D | ShimPreparedModel.cpp | 367 ShimBurst::ShimBurst(std::shared_ptr<ShimPreparedModel> preparedModel) in ShimBurst()
|
/packages/modules/NeuralNetworks/runtime/test/android_fuzzing/ |
D | DriverFuzzTest.cpp | 263 const sp<V1_3::IPreparedModel>& preparedModel) override { in notify_1_3() 296 void execute(const sp<V1_3::IPreparedModel>& preparedModel, const V1_3::Request& request) { in execute() 316 const auto preparedModel = prepareModel(device, model); in nnapiFuzzTest() local
|
/packages/modules/NeuralNetworks/runtime/ |
D | Memory.cpp | 322 auto callback = [&roles](const auto* preparedModel, IOType type, uint32_t index) { in addRole() 415 for (const auto* preparedModel : desc.preparedModels) { in logMemoryDescriptorToInfo() local 431 for (const auto* preparedModel : desc.preparedModels) { in getDevices() local
|
D | Manager.cpp | 141 DriverPreparedModel(const Device* device, const SharedPreparedModel& preparedModel) in DriverPreparedModel() 475 SharedPreparedModel preparedModel = std::move(result).value(); in prepareModel() local 487 [](const auto* preparedModel) { in allocate() 943 CpuExecution(const CpuPreparedModel& preparedModel, Request request, in CpuExecution() 1022 std::shared_ptr<RuntimePreparedModel> preparedModel = in create() local
|
D | ExecutionPlan.cpp | 72 TokenHasher* token, std::shared_ptr<RuntimePreparedModel>* preparedModel) { in compile() 1171 if (const auto preparedModel = in makeBursts() local 1189 if (const auto preparedModel = simpleBody->mPreparedModel) { in makeBursts() local 2053 index, [&alignment, &padding](const auto* preparedModel, IOType, uint32_t) { in getMemoryPreferenceOfSourceOperand()
|
D | ExecutionBuilder.cpp | 1315 std::shared_ptr<RuntimePreparedModel> preparedModel, bool reusable, in StepExecutor()
|