代码拉取完成,页面将自动刷新
/**
* Copyright 2021 Huawei Technologies Co., Ltd
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef MINDSPORE_LITE_SRC_RUNTIME_DELEGATE_NPU_NPU_GRAPH_H_
#define MINDSPORE_LITE_SRC_RUNTIME_DELEGATE_NPU_NPU_GRAPH_H_
#include <vector>
#include <queue>
#include <map>
#include <utility>
#include "include/api/kernel.h"
#include "src/litert/delegate/npu/op/npu_op.h"
#include "src/litert/delegate/npu/npu_executor.h"
namespace mindspore::lite {
class NPUGraph : public kernel::Kernel {
public:
NPUGraph(std::vector<NPUOp *> npu_ops, NPUManager *npu_manager, const std::vector<mindspore::MSTensor> &inputs,
const std::vector<mindspore::MSTensor> &outputs)
: kernel::Kernel(inputs, outputs, nullptr, nullptr), npu_ops_(std::move(npu_ops)), npu_manager_(npu_manager) {}
~NPUGraph() override;
int Init();
int Prepare() override;
int Execute() override;
int ReSize() override {
MS_LOG(ERROR) << "NPU does not support the resize function temporarily.";
return RET_ERROR;
}
void set_input(mindspore::MSTensor in_tensor, int index) override;
void set_output(mindspore::MSTensor out_tensor, int index) override;
int FindPreNextOps();
std::vector<NPUOp *> *GetOps() { return &npu_ops_; }
std::vector<mindspore::MSTensor *> *GetInsertTensors() { return &insert_tensors_; }
protected:
std::vector<NPUOp *> FindPreOps(NPUOp *cur_op);
std::vector<NPUOp *> FindNextOps(NPUOp *cur_op);
int FindValidSubgraphInOps(std::queue<NPUOp *> *valid_in_ops, std::queue<NPUOp *> *candidate_in_ops,
std::map<const NPUOp *, bool> *is_visited);
std::vector<NPUOp *> FindReadySubgraphOps(std::queue<NPUOp *> op_queue, std::queue<NPUOp *> *next_candidate_ops,
std::map<const NPUOp *, bool> *is_visited);
int CreateSubgraphFromReadyOps(std::queue<NPUOp *> *valid_in_ops, std::vector<NPUOp *> ready_ops,
std::map<const NPUOp *, bool> *is_searched);
kernel::Kernel *CreateNPUSubgraphKernel(std::vector<NPUOp *> ops);
kernel::Kernel *CreateNPUTransposeKernel(NPUOp *op);
std::vector<NPUOp *> npu_ops_{};
std::vector<kernel::Kernel *> all_kernels_{};
std::vector<mindspore::MSTensor *> insert_tensors_;
NPUManager *npu_manager_ = nullptr;
};
} // namespace mindspore::lite
#endif // MINDSPORE_LITE_SRC_RUNTIME_DELEGATE_NPU_NPU_GRAPH_H_
此处可能存在不合适展示的内容,页面不予展示。您可通过相关编辑功能自查并修改。
如您确认内容无涉及 不当用语 / 纯广告导流 / 暴力 / 低俗色情 / 侵权 / 盗版 / 虚假 / 无价值内容或违法国家有关法律法规的内容,可点击提交进行申诉,我们将尽快为您处理。