1 /*
2  * Copyright (C) 2020 The Android Open Source Project
3  *
4  * Licensed under the Apache License, Version 2.0 (the "License");
5  * you may not use this file except in compliance with the License.
6  * You may obtain a copy of the License at
7  *
8  *      http://www.apache.org/licenses/LICENSE-2.0
9  *
10  * Unless required by applicable law or agreed to in writing, software
11  * distributed under the License is distributed on an "AS IS" BASIS,
12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  * See the License for the specific language governing permissions and
14  * limitations under the License.
15  */
16 
17 #include "ResilientPreparedModel.h"
18 
19 #include "InvalidBurst.h"
20 #include "InvalidExecution.h"
21 #include "ResilientBurst.h"
22 #include "ResilientExecution.h"
23 
24 #include <android-base/logging.h>
25 #include <android-base/thread_annotations.h>
26 #include <nnapi/IPreparedModel.h>
27 #include <nnapi/Result.h>
28 #include <nnapi/TypeUtils.h>
29 #include <nnapi/Types.h>
30 
31 #include <functional>
32 #include <memory>
33 #include <mutex>
34 #include <sstream>
35 #include <utility>
36 #include <vector>
37 
38 namespace android::hardware::neuralnetworks::utils {
39 namespace {
40 
41 template <typename FnType>
protect(const ResilientPreparedModel & resilientPreparedModel,const FnType & fn)42 auto protect(const ResilientPreparedModel& resilientPreparedModel, const FnType& fn)
43         -> decltype(fn(*resilientPreparedModel.getPreparedModel())) {
44     auto preparedModel = resilientPreparedModel.getPreparedModel();
45     auto result = fn(*preparedModel);
46 
47     // Immediately return if prepared model is not dead.
48     if (result.has_value() || result.error().code != nn::ErrorStatus::DEAD_OBJECT) {
49         return result;
50     }
51 
52     // Attempt recovery and return if it fails.
53     auto maybePreparedModel = resilientPreparedModel.recover(preparedModel.get());
54     if (!maybePreparedModel.has_value()) {
55         const auto& [message, code] = maybePreparedModel.error();
56         std::ostringstream oss;
57         oss << ", and failed to recover dead prepared model with error " << code << ": " << message;
58         result.error().message += oss.str();
59         return result;
60     }
61     preparedModel = std::move(maybePreparedModel).value();
62 
63     return fn(*preparedModel);
64 }
65 
66 }  // namespace
67 
create(Factory makePreparedModel)68 nn::GeneralResult<std::shared_ptr<const ResilientPreparedModel>> ResilientPreparedModel::create(
69         Factory makePreparedModel) {
70     if (makePreparedModel == nullptr) {
71         return NN_ERROR(nn::ErrorStatus::INVALID_ARGUMENT)
72                << "utils::ResilientPreparedModel::create must have non-empty makePreparedModel";
73     }
74     auto preparedModel = NN_TRY(makePreparedModel());
75     CHECK(preparedModel != nullptr);
76     return std::make_shared<ResilientPreparedModel>(
77             PrivateConstructorTag{}, std::move(makePreparedModel), std::move(preparedModel));
78 }
79 
ResilientPreparedModel(PrivateConstructorTag,Factory makePreparedModel,nn::SharedPreparedModel preparedModel)80 ResilientPreparedModel::ResilientPreparedModel(PrivateConstructorTag /*tag*/,
81                                                Factory makePreparedModel,
82                                                nn::SharedPreparedModel preparedModel)
83     : kMakePreparedModel(std::move(makePreparedModel)), mPreparedModel(std::move(preparedModel)) {
84     CHECK(kMakePreparedModel != nullptr);
85     CHECK(mPreparedModel != nullptr);
86 }
87 
getPreparedModel() const88 nn::SharedPreparedModel ResilientPreparedModel::getPreparedModel() const {
89     std::lock_guard guard(mMutex);
90     return mPreparedModel;
91 }
92 
recover(const nn::IPreparedModel * failingPreparedModel) const93 nn::GeneralResult<nn::SharedPreparedModel> ResilientPreparedModel::recover(
94         const nn::IPreparedModel* failingPreparedModel) const {
95     std::lock_guard guard(mMutex);
96 
97     // Another caller updated the failing prepared model.
98     if (mPreparedModel.get() != failingPreparedModel) {
99         return mPreparedModel;
100     }
101 
102     mPreparedModel = NN_TRY(kMakePreparedModel());
103     return mPreparedModel;
104 }
105 
106 nn::ExecutionResult<std::pair<std::vector<nn::OutputShape>, nn::Timing>>
execute(const nn::Request & request,nn::MeasureTiming measure,const nn::OptionalTimePoint & deadline,const nn::OptionalDuration & loopTimeoutDuration,const std::vector<nn::TokenValuePair> & hints,const std::vector<nn::ExtensionNameAndPrefix> & extensionNameToPrefix) const107 ResilientPreparedModel::execute(
108         const nn::Request& request, nn::MeasureTiming measure,
109         const nn::OptionalTimePoint& deadline, const nn::OptionalDuration& loopTimeoutDuration,
110         const std::vector<nn::TokenValuePair>& hints,
111         const std::vector<nn::ExtensionNameAndPrefix>& extensionNameToPrefix) const {
112     const auto fn = [&request, measure, &deadline, &loopTimeoutDuration, &hints,
113                      &extensionNameToPrefix](const nn::IPreparedModel& preparedModel) {
114         return preparedModel.execute(request, measure, deadline, loopTimeoutDuration, hints,
115                                      extensionNameToPrefix);
116     };
117     return protect(*this, fn);
118 }
119 
120 nn::GeneralResult<std::pair<nn::SyncFence, nn::ExecuteFencedInfoCallback>>
executeFenced(const nn::Request & request,const std::vector<nn::SyncFence> & waitFor,nn::MeasureTiming measure,const nn::OptionalTimePoint & deadline,const nn::OptionalDuration & loopTimeoutDuration,const nn::OptionalDuration & timeoutDurationAfterFence,const std::vector<nn::TokenValuePair> & hints,const std::vector<nn::ExtensionNameAndPrefix> & extensionNameToPrefix) const121 ResilientPreparedModel::executeFenced(
122         const nn::Request& request, const std::vector<nn::SyncFence>& waitFor,
123         nn::MeasureTiming measure, const nn::OptionalTimePoint& deadline,
124         const nn::OptionalDuration& loopTimeoutDuration,
125         const nn::OptionalDuration& timeoutDurationAfterFence,
126         const std::vector<nn::TokenValuePair>& hints,
127         const std::vector<nn::ExtensionNameAndPrefix>& extensionNameToPrefix) const {
128     const auto fn = [&request, &waitFor, measure, &deadline, &loopTimeoutDuration,
129                      &timeoutDurationAfterFence, &hints,
130                      &extensionNameToPrefix](const nn::IPreparedModel& preparedModel) {
131         return preparedModel.executeFenced(request, waitFor, measure, deadline, loopTimeoutDuration,
132                                            timeoutDurationAfterFence, hints, extensionNameToPrefix);
133     };
134     return protect(*this, fn);
135 }
136 
createReusableExecution(const nn::Request & request,nn::MeasureTiming measure,const nn::OptionalDuration & loopTimeoutDuration,const std::vector<nn::TokenValuePair> & hints,const std::vector<nn::ExtensionNameAndPrefix> & extensionNameToPrefix) const137 nn::GeneralResult<nn::SharedExecution> ResilientPreparedModel::createReusableExecution(
138         const nn::Request& request, nn::MeasureTiming measure,
139         const nn::OptionalDuration& loopTimeoutDuration,
140         const std::vector<nn::TokenValuePair>& hints,
141         const std::vector<nn::ExtensionNameAndPrefix>& extensionNameToPrefix) const {
142 #if 0
143     auto self = shared_from_this();
144     ResilientExecution::Factory makeExecution = [preparedModel = std::move(self), request, measure,
145                                                  loopTimeoutDuration, hints,
146                                                  extensionNameToPrefix] {
147         return preparedModel->createReusableExecutionInternal(request, measure, loopTimeoutDuration,
148                                                               hints, extensionNameToPrefix);
149     };
150     return ResilientExecution::create(std::move(makeExecution));
151 #else
152     return createReusableExecutionInternal(request, measure, loopTimeoutDuration, hints,
153                                            extensionNameToPrefix);
154 #endif
155 }
156 
configureExecutionBurst() const157 nn::GeneralResult<nn::SharedBurst> ResilientPreparedModel::configureExecutionBurst() const {
158 #if 0
159     auto self = shared_from_this();
160     ResilientBurst::Factory makeBurst =
161             [preparedModel = std::move(self)]() -> nn::GeneralResult<nn::SharedBurst> {
162         return preparedModel->configureExecutionBurst();
163     };
164     return ResilientBurst::create(std::move(makeBurst));
165 #else
166     return configureExecutionBurstInternal();
167 #endif
168 }
169 
createReusableExecutionInternal(const nn::Request & request,nn::MeasureTiming measure,const nn::OptionalDuration & loopTimeoutDuration,const std::vector<nn::TokenValuePair> & hints,const std::vector<nn::ExtensionNameAndPrefix> & extensionNameToPrefix) const170 nn::GeneralResult<nn::SharedExecution> ResilientPreparedModel::createReusableExecutionInternal(
171         const nn::Request& request, nn::MeasureTiming measure,
172         const nn::OptionalDuration& loopTimeoutDuration,
173         const std::vector<nn::TokenValuePair>& hints,
174         const std::vector<nn::ExtensionNameAndPrefix>& extensionNameToPrefix) const {
175     if (!isValidInternal()) {
176         return std::make_shared<const InvalidExecution>();
177     }
178     const auto fn = [&request, measure, &loopTimeoutDuration, &hints,
179                      &extensionNameToPrefix](const nn::IPreparedModel& preparedModel) {
180         return preparedModel.createReusableExecution(request, measure, loopTimeoutDuration, hints,
181                                                      extensionNameToPrefix);
182     };
183     return protect(*this, fn);
184 }
185 
getUnderlyingResource() const186 std::any ResilientPreparedModel::getUnderlyingResource() const {
187     return getPreparedModel()->getUnderlyingResource();
188 }
189 
isValidInternal() const190 bool ResilientPreparedModel::isValidInternal() const {
191     return true;
192 }
193 
configureExecutionBurstInternal() const194 nn::GeneralResult<nn::SharedBurst> ResilientPreparedModel::configureExecutionBurstInternal() const {
195     if (!isValidInternal()) {
196         return std::make_shared<const InvalidBurst>();
197     }
198     const auto fn = [](const nn::IPreparedModel& preparedModel) {
199         return preparedModel.configureExecutionBurst();
200     };
201     return protect(*this, fn);
202 }
203 
204 }  // namespace android::hardware::neuralnetworks::utils
205