Home
last modified time | relevance | path

Searched refs:requestPoolInfos (Results 1 – 7 of 7) sorted by relevance

/packages/modules/NeuralNetworks/driver/sample/
DCanonicalPreparedModel.cpp41 std::vector<RunTimePoolInfo> requestPoolInfos; in createRunTimePoolInfos() local
43 requestPoolInfos.reserve(request.pools.size()); in createRunTimePoolInfos()
53 requestPoolInfos.push_back(std::move(*buffer)); in createRunTimePoolInfos()
65 requestPoolInfos.push_back(bufferWrapper->createRunTimePoolInfo()); in createRunTimePoolInfos()
69 return std::make_pair(std::move(requestPoolInfos), std::move(bufferWrappers)); in createRunTimePoolInfos()
149 const auto [requestPoolInfos, bufferWrappers] = in execute()
163 int n = executor.run(kModel, request, kPoolInfos, requestPoolInfos); in execute()
237 const auto [requestPoolInfos, bufferWrappers] = in executeFenced()
250 int n = executor.run(kModel, request, kPoolInfos, requestPoolInfos); in executeFenced()
/packages/modules/NeuralNetworks/driver/sample_hidl/
DSampleDriver.cpp373 std::vector<RunTimePoolInfo> requestPoolInfos; in createRunTimePoolInfos() local
375 requestPoolInfos.reserve(request.pools.size()); in createRunTimePoolInfos()
387 requestPoolInfos.push_back(std::move(*buffer)); in createRunTimePoolInfos()
400 requestPoolInfos.push_back(bufferWrapper->createRunTimePoolInfo()); in createRunTimePoolInfos()
405 return {V1_3::ErrorStatus::NONE, std::move(requestPoolInfos), std::move(bufferWrappers)}; in createRunTimePoolInfos()
458 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in asyncExecute()
478 requestPoolInfos); in asyncExecute()
592 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in executeSynchronouslyBase()
610 requestPoolInfos); in executeSynchronouslyBase()
704 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in executeFenced()
[all …]
DSampleDriverFloatXNNPACK.cpp74 const std::vector<RunTimePoolInfo>& requestPoolInfos, in updateForArguments() argument
95 CHECK_LT(poolIndex, requestPoolInfos.size()); in updateForArguments()
96 auto& r = requestPoolInfos[poolIndex]; in updateForArguments()
1649 std::vector<RunTimePoolInfo> requestPoolInfos; in asyncExecuteXNNPACK() local
1650 if (!setRunTimePoolInfosFromMemoryPools(&requestPoolInfos, uncheckedConvert(request.pools))) { in asyncExecuteXNNPACK()
1653 updateForArguments(model.main.inputIndexes, request.inputs, requestPoolInfos, operands); in asyncExecuteXNNPACK()
1654 updateForArguments(model.main.outputIndexes, request.outputs, requestPoolInfos, operands); in asyncExecuteXNNPACK()
1659 for (auto& runtimeInfo : requestPoolInfos) { in asyncExecuteXNNPACK()
1744 std::vector<RunTimePoolInfo> requestPoolInfos; in executeSynchronouslyXNNPACKBase() local
1745 if (!setRunTimePoolInfosFromMemoryPools(&requestPoolInfos, uncheckedConvert(request.pools))) { in executeSynchronouslyXNNPACKBase()
[all …]
/packages/modules/NeuralNetworks/runtime/
DManager.cpp980 std::vector<RunTimePoolInfo> requestPoolInfos, in CpuExecution() argument
984 kRequestPoolInfos(std::move(requestPoolInfos)), in CpuExecution()
1079 const std::vector<RunTimePoolInfo>& requestPoolInfos, const OptionalTimePoint& deadline, in computeOnCpu() argument
1089 int err = executor.run(model, request, modelPoolInfos, requestPoolInfos); in computeOnCpu()
1129 std::vector<RunTimePoolInfo> requestPoolInfos; in createCpuRequest() local
1130 requestPoolInfos.reserve(memories.size()); in createCpuRequest()
1133 requestPoolInfos.emplace_back(*poolInfo); in createCpuRequest()
1140 [&requestPoolInfos](const std::vector<ModelArgumentInfo>& argumentInfos) { in createCpuRequest()
1145 {.poolIndex = static_cast<uint32_t>(requestPoolInfos.size()), in createCpuRequest()
1149 requestPoolInfos.emplace_back(RunTimePoolInfo::createFromExistingBuffer( in createCpuRequest()
[all …]
/packages/modules/NeuralNetworks/driver/sample_aidl/
DSampleDriverAidl.cpp318 std::vector<RunTimePoolInfo> requestPoolInfos; in createRunTimePoolInfos() local
320 requestPoolInfos.reserve(request.pools.size()); in createRunTimePoolInfos()
330 requestPoolInfos.push_back(std::move(*buffer)); in createRunTimePoolInfos()
344 requestPoolInfos.push_back(RunTimePoolInfo::createFromExistingBuffer( in createRunTimePoolInfos()
352 return {aidl_hal::ErrorStatus::NONE, std::move(requestPoolInfos), std::move(bufferWrappers)}; in createRunTimePoolInfos()
435 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in executeSynchronously()
451 int n = executor.run(model, request, mPoolInfos, requestPoolInfos); in executeSynchronously()
550 const auto [poolStatus, requestPoolInfos, bufferWrappers] = in executeFenced()
566 int n = executor.run(model, request, mPoolInfos, requestPoolInfos); in executeFenced()
/packages/modules/NeuralNetworks/common/include/
DCpuExecutor.h151 const std::vector<RunTimePoolInfo>& requestPoolInfos);
168 const std::vector<RunTimePoolInfo>& requestPoolInfos,
/packages/modules/NeuralNetworks/common/
DCpuExecutor.cpp560 const std::vector<RunTimePoolInfo>& requestPoolInfos) { in run() argument
573 updateForArguments(model.main.inputIndexes, request.inputs, requestPoolInfos, operands.data()); in run()
574 updateForArguments(model.main.outputIndexes, request.outputs, requestPoolInfos, in run()
581 for (auto& runtimeInfo : requestPoolInfos) { in run()
668 const std::vector<RunTimePoolInfo>& requestPoolInfos, in updateForArguments() argument
692 CHECK_LT(poolIndex, requestPoolInfos.size()); in updateForArguments()
693 auto& r = requestPoolInfos[poolIndex]; in updateForArguments()