From c4493551009ba5585763b3152c2c50657c00dd79 Mon Sep 17 00:00:00 2001 From: w30052974 Date: Thu, 26 Sep 2024 11:20:54 +0800 Subject: [PATCH] =?UTF-8?q?=E5=8F=96=E6=B6=88DLSA=E6=A8=A1=E5=9E=8B?= =?UTF-8?q?=E5=8A=A0=E8=BD=BD=E5=8D=B8=E8=BD=BD=E4=BC=A0=E9=80=92callingpi?= =?UTF-8?q?d?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: w30052974 --- .../native/neural_network_core/compilation.h | 1 - .../neural_network_core.cpp | 20 ++++--------------- .../native/neural_network_core/nnrt_client.h | 4 ++-- 3 files changed, 6 insertions(+), 19 deletions(-) diff --git a/frameworks/native/neural_network_core/compilation.h b/frameworks/native/neural_network_core/compilation.h index f3a9a36..c327f7a 100644 --- a/frameworks/native/neural_network_core/compilation.h +++ b/frameworks/native/neural_network_core/compilation.h @@ -42,7 +42,6 @@ struct Compilation { std::unordered_map> configs; size_t nnrtModelID {0}; uint32_t hiaiModelId {0}; - int32_t callingPid {-1}; bool isNeedModelLatency {false}; ~Compilation() diff --git a/frameworks/native/neural_network_core/neural_network_core.cpp b/frameworks/native/neural_network_core/neural_network_core.cpp index 05eef3e..6e689f0 100644 --- a/frameworks/native/neural_network_core/neural_network_core.cpp +++ b/frameworks/native/neural_network_core/neural_network_core.cpp @@ -573,7 +573,7 @@ OH_NN_ReturnCode AuthenticateModel(const Compilation* compilation) LOGE("Authentication failed, nnrtService Authentication func is nullptr."); return OH_NN_INVALID_PARAMETER; } - ret = nnrtService.Authentication(compilation->callingPid); + ret = nnrtService.Authentication(); if (ret != static_cast(OH_NN_SUCCESS)) { LOGE("Authentication failed, input model cannot run by npu."); return static_cast(ret); @@ -595,13 +595,6 @@ OH_NN_ReturnCode Authentication(Compilation** compilation) return OH_NN_INVALID_PARAMETER; } - auto iter = compilationImpl->configs.find("callingPid"); - if (iter == compilationImpl->configs.end()) { - LOGE("missing 'callingPid' parameter in compilation configs."); - } else { - compilationImpl->callingPid = std::atoi((iter->second).data()); - } - const NNRtServiceApi& nnrtService = NNRtServiceApi::GetInstance(); if (!nnrtService.IsServiceAvaliable()) { LOGW("Authentication failed, fail to get nnrt service, skip Authentication."); @@ -1296,7 +1289,7 @@ OH_NN_ReturnCode SetModelId(const Compilation* compilation) } int ret = nnrtService.SetModelID( - compilation->callingPid, compilation->hiaiModelId, compilation->nnrtModelID); + compilation->hiaiModelId, compilation->nnrtModelID); if (ret != static_cast(OH_NN_SUCCESS)) { LOGE("SetModelId failed, fail to set modelId."); return static_cast(ret); @@ -1335,8 +1328,8 @@ OH_NN_ReturnCode ExecutorPrepare(Executor** executor, Compilation** compilation) return ret; } - LOGD("ExecutorPrepare parameter, callingPid: %{public}d, hiaiModelId: %{public}u, nnrtModelId: %{public}zu.", - compilationImpl->callingPid, compilationImpl->hiaiModelId, compilationImpl->nnrtModelID); + LOGD("ExecutorPrepare parameter, hiaiModelId: %{public}u, nnrtModelId: %{public}zu.", + compilationImpl->hiaiModelId, compilationImpl->nnrtModelID); ret = Scheduling(&compilationImpl); if (ret != OH_NN_SUCCESS) { @@ -1345,11 +1338,6 @@ OH_NN_ReturnCode ExecutorPrepare(Executor** executor, Compilation** compilation) } std::unordered_map> configMap; - std::string callingPidStr = std::to_string(compilationImpl->callingPid); - std::vector vecCallingPid(callingPidStr.begin(), callingPidStr.end()); - vecCallingPid.emplace_back('\0'); - configMap["callingPid"] = vecCallingPid; - std::string hiaiModelIdStr = std::to_string(compilationImpl->hiaiModelId); std::vector vechiaiModelId(hiaiModelIdStr.begin(), hiaiModelIdStr.end()); vechiaiModelId.emplace_back('\0'); diff --git a/frameworks/native/neural_network_core/nnrt_client.h b/frameworks/native/neural_network_core/nnrt_client.h index dd19084..628ab63 100644 --- a/frameworks/native/neural_network_core/nnrt_client.h +++ b/frameworks/native/neural_network_core/nnrt_client.h @@ -33,10 +33,10 @@ public: size_t (*GetNNRtModelIDFromCache)(const char* path, const char* modelName) = nullptr; size_t (*GetNNRtModelIDFromBuffer)(const void* buffer, size_t size) = nullptr; size_t (*GetNNRtModelIDFromModel)(void* model) = nullptr; - int (*SetModelID)(int callingPid, uint32_t hiaimodelID, size_t nnrtModelID) = nullptr; + int (*SetModelID)(uint32_t hiaimodelID, size_t nnrtModelID) = nullptr; int (*IsSupportAuthentication)(bool* supportStat) = nullptr; int (*IsSupportScheduling)(bool* supportStat) = nullptr; - int (*Authentication)(int callingPid) = nullptr; + int (*Authentication)() = nullptr; int (*Scheduling)(uint32_t hiaiModelId, bool* needModelLatency, const char* cachePath) = nullptr; int (*UpdateModelLatency)(uint32_t hiaiModelId, int modelLatency) = nullptr; int (*Unload)(uint32_t hiaiModelId) = nullptr; -- Gitee