/packages/modules/NeuralNetworks/driver/sample/ |
D | CanonicalPreparedModel.cpp | 41 std::vector<RunTimePoolInfo> requestPoolInfos; in createRunTimePoolInfos() local 43 requestPoolInfos.reserve(request.pools.size()); in createRunTimePoolInfos() 53 requestPoolInfos.push_back(std::move(*buffer)); in createRunTimePoolInfos() 65 requestPoolInfos.push_back(bufferWrapper->createRunTimePoolInfo()); in createRunTimePoolInfos() 69 return std::make_pair(std::move(requestPoolInfos), std::move(bufferWrappers)); in createRunTimePoolInfos() 149 const auto [requestPoolInfos, bufferWrappers] = in execute() 163 int n = executor.run(kModel, request, kPoolInfos, requestPoolInfos); in execute() 237 const auto [requestPoolInfos, bufferWrappers] = in executeFenced() 250 int n = executor.run(kModel, request, kPoolInfos, requestPoolInfos); in executeFenced()
|
/packages/modules/NeuralNetworks/driver/sample_hidl/ |
D | SampleDriver.cpp | 373 std::vector<RunTimePoolInfo> requestPoolInfos; in createRunTimePoolInfos() local 375 requestPoolInfos.reserve(request.pools.size()); in createRunTimePoolInfos() 387 requestPoolInfos.push_back(std::move(*buffer)); in createRunTimePoolInfos() 400 requestPoolInfos.push_back(bufferWrapper->createRunTimePoolInfo()); in createRunTimePoolInfos() 405 return {V1_3::ErrorStatus::NONE, std::move(requestPoolInfos), std::move(bufferWrappers)}; in createRunTimePoolInfos() 458 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in asyncExecute() 478 requestPoolInfos); in asyncExecute() 592 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in executeSynchronouslyBase() 610 requestPoolInfos); in executeSynchronouslyBase() 704 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in executeFenced() [all …]
|
D | SampleDriverFloatXNNPACK.cpp | 74 const std::vector<RunTimePoolInfo>& requestPoolInfos, in updateForArguments() argument 95 CHECK_LT(poolIndex, requestPoolInfos.size()); in updateForArguments() 96 auto& r = requestPoolInfos[poolIndex]; in updateForArguments() 1649 std::vector<RunTimePoolInfo> requestPoolInfos; in asyncExecuteXNNPACK() local 1650 if (!setRunTimePoolInfosFromMemoryPools(&requestPoolInfos, uncheckedConvert(request.pools))) { in asyncExecuteXNNPACK() 1653 updateForArguments(model.main.inputIndexes, request.inputs, requestPoolInfos, operands); in asyncExecuteXNNPACK() 1654 updateForArguments(model.main.outputIndexes, request.outputs, requestPoolInfos, operands); in asyncExecuteXNNPACK() 1659 for (auto& runtimeInfo : requestPoolInfos) { in asyncExecuteXNNPACK() 1744 std::vector<RunTimePoolInfo> requestPoolInfos; in executeSynchronouslyXNNPACKBase() local 1745 if (!setRunTimePoolInfosFromMemoryPools(&requestPoolInfos, uncheckedConvert(request.pools))) { in executeSynchronouslyXNNPACKBase() [all …]
|
/packages/modules/NeuralNetworks/runtime/ |
D | Manager.cpp | 980 std::vector<RunTimePoolInfo> requestPoolInfos, in CpuExecution() argument 984 kRequestPoolInfos(std::move(requestPoolInfos)), in CpuExecution() 1079 const std::vector<RunTimePoolInfo>& requestPoolInfos, const OptionalTimePoint& deadline, in computeOnCpu() argument 1089 int err = executor.run(model, request, modelPoolInfos, requestPoolInfos); in computeOnCpu() 1129 std::vector<RunTimePoolInfo> requestPoolInfos; in createCpuRequest() local 1130 requestPoolInfos.reserve(memories.size()); in createCpuRequest() 1133 requestPoolInfos.emplace_back(*poolInfo); in createCpuRequest() 1140 [&requestPoolInfos](const std::vector<ModelArgumentInfo>& argumentInfos) { in createCpuRequest() 1145 {.poolIndex = static_cast<uint32_t>(requestPoolInfos.size()), in createCpuRequest() 1149 requestPoolInfos.emplace_back(RunTimePoolInfo::createFromExistingBuffer( in createCpuRequest() [all …]
|
/packages/modules/NeuralNetworks/driver/sample_aidl/ |
D | SampleDriverAidl.cpp | 318 std::vector<RunTimePoolInfo> requestPoolInfos; in createRunTimePoolInfos() local 320 requestPoolInfos.reserve(request.pools.size()); in createRunTimePoolInfos() 330 requestPoolInfos.push_back(std::move(*buffer)); in createRunTimePoolInfos() 344 requestPoolInfos.push_back(RunTimePoolInfo::createFromExistingBuffer( in createRunTimePoolInfos() 352 return {aidl_hal::ErrorStatus::NONE, std::move(requestPoolInfos), std::move(bufferWrappers)}; in createRunTimePoolInfos() 435 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in executeSynchronously() 451 int n = executor.run(model, request, mPoolInfos, requestPoolInfos); in executeSynchronously() 550 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in executeFenced() 566 int n = executor.run(model, request, mPoolInfos, requestPoolInfos); in executeFenced()
|
/packages/modules/NeuralNetworks/common/include/ |
D | CpuExecutor.h | 151 const std::vector<RunTimePoolInfo>& requestPoolInfos); 168 const std::vector<RunTimePoolInfo>& requestPoolInfos,
|
/packages/modules/NeuralNetworks/common/ |
D | CpuExecutor.cpp | 560 const std::vector<RunTimePoolInfo>& requestPoolInfos) { in run() argument 573 updateForArguments(model.main.inputIndexes, request.inputs, requestPoolInfos, operands.data()); in run() 574 updateForArguments(model.main.outputIndexes, request.outputs, requestPoolInfos, in run() 581 for (auto& runtimeInfo : requestPoolInfos) { in run() 668 const std::vector<RunTimePoolInfo>& requestPoolInfos, in updateForArguments() argument 692 CHECK_LT(poolIndex, requestPoolInfos.size()); in updateForArguments() 693 auto& r = requestPoolInfos[poolIndex]; in updateForArguments()
|