提交 02bfb8f8 编写于 作者: M Megvii Engine Team

feat(lite): add and fix some feature for load and run fitting mode

GitOrigin-RevId: bbddc9bb791297c636e3f3a6294922f8e18cb36b
上级 c62ddba2
......@@ -11,7 +11,7 @@
#pragma once
#include <string>
#include <vector>
namespace lite {
struct FeatureBits32 {
......@@ -45,6 +45,11 @@ public:
std::string model_path, std::string packed_model_path,
std::string info_data_path = "", std::string info_algo_policy_path = "",
std::string info_binary_cache_path = "");
ModelPacker(
std::vector<uint8_t> model_data, std::string packed_model_path,
std::vector<uint8_t> info_data = {},
std::vector<uint8_t> info_algo_policy_data = {},
std::vector<uint8_t> info_binary_cache_data = {});
void set_header(
std::string model_decryption_method = "NONE",
......@@ -53,13 +58,12 @@ public:
void pack_model();
private:
std::string m_packed_model_path;
std::string m_info_data_path;
std::vector<uint8_t> m_info_data;
//! fastrun cache / algo policy
std::string m_info_algo_policy_path;
std::vector<uint8_t> m_algo_policy_data;
//! binary cache
std::string m_info_binary_cache_path;
std::vector<uint8_t> m_binary_cache_data;
std::string m_packed_model_path;
Header m_header;
friend class FbsHelper;
......
......@@ -25,6 +25,7 @@ class FbsHelper {
public:
FbsHelper() = default;
FbsHelper(ModelPacker* packer, std::string model_path);
FbsHelper(ModelPacker* packer, std::vector<uint8_t>& model_data);
flatbuffers::Offset<model_parse::ModelHeader> build_header();
flatbuffers::Offset<model_parse::ModelInfo> build_info();
flatbuffers::Offset<model_parse::ModelData> build_data();
......@@ -58,6 +59,19 @@ std::vector<uint8_t> read_file(std::string path) {
fclose(fin);
return buf;
}
FbsHelper::FbsHelper(ModelPacker* packer, std::vector<uint8_t>& model_data)
: m_packer(packer), m_model_buffer(model_data) {
const char* model_ptr =
static_cast<const char*>(static_cast<void*>(m_model_buffer.data()));
std::string tag(model_ptr, 12);
if (tag == "packed_model") {
uint8_t* buffer = m_model_buffer.data() + 12;
auto model = GetPackModel(buffer)->models()->Get(0);
m_model_header = model->header();
m_model_info = model->info();
m_model_data = model->data();
}
}
FbsHelper::FbsHelper(ModelPacker* packer, std::string model_path) : m_packer(packer) {
m_model_buffer = read_file(model_path);
......@@ -118,21 +132,20 @@ flatbuffers::Offset<ModelData> FbsHelper::build_data() {
flatbuffers::Offset<ModelInfo> FbsHelper::build_info() {
flatbuffers::Offset<flatbuffers::Vector<uint8_t>> fb_data;
if (m_model_info && m_model_info->data() && m_packer->m_info_data_path.empty()) {
if (m_model_info && m_model_info->data() && m_packer->m_info_data.empty()) {
auto data = m_model_info->data()->Data();
auto size = m_model_info->data()->size();
fb_data = m_builder.CreateVector(data, size);
} else if (!m_packer->m_info_data_path.empty()) {
auto info_data = read_file(m_packer->m_info_data_path);
fb_data = m_builder.CreateVector(info_data);
} else if (!m_packer->m_info_data.empty()) {
fb_data = m_builder.CreateVector(m_packer->m_info_data);
}
flatbuffers::Offset<flatbuffers::Vector<uint8_t>> fb_algo_policy;
flatbuffers::Offset<flatbuffers::Vector<uint8_t>> fb_binary_cache;
if (m_packer->m_header.fb32.is_fast_run_cache) {
std::vector<uint8_t> info_algo_policy;
if (!m_packer->m_info_algo_policy_path.empty()) {
info_algo_policy = read_file(m_packer->m_info_algo_policy_path);
if (!m_packer->m_algo_policy_data.empty()) {
info_algo_policy = m_packer->m_algo_policy_data;
if (m_model_info && m_model_info->algo_policy()) {
auto cache = m_model_info->algo_policy()->Data();
auto size = m_model_info->algo_policy()->size();
......@@ -178,11 +191,27 @@ ModelPacker::ModelPacker(
std::string model_path, std::string packed_model_path,
std::string info_data_path, std::string info_algo_policy_path,
std::string info_binary_cache_path)
: m_packed_model_path(packed_model_path),
m_info_data_path(info_data_path),
m_info_algo_policy_path(info_algo_policy_path),
m_info_binary_cache_path(info_binary_cache_path) {
: m_packed_model_path(packed_model_path) {
m_fbs_helper = new FbsHelper(this, model_path);
std::vector<uint8_t> empty_vec;
m_info_data = info_data_path.empty() ? empty_vec : read_file(info_data_path);
m_algo_policy_data = info_algo_policy_path.empty()
? empty_vec
: read_file(info_algo_policy_path);
m_binary_cache_data = info_binary_cache_path.empty()
? empty_vec
: read_file(info_binary_cache_path);
}
ModelPacker::ModelPacker(
std::vector<uint8_t> model_data, std::string packed_model_path,
std::vector<uint8_t> info_data, std::vector<uint8_t> info_algo_policy_data,
std::vector<uint8_t> info_binary_cache_data) {
m_fbs_helper = new FbsHelper(this, model_data);
m_packed_model_path = packed_model_path;
m_info_data = info_data;
m_algo_policy_data = info_algo_policy_data;
m_binary_cache_data = info_binary_cache_data;
}
void ModelPacker::set_header(
......@@ -192,10 +221,10 @@ void ModelPacker::set_header(
m_header.info_decryption_method = info_decryption_method;
memset(&m_header.fb32, 0, sizeof(m_header.fb32));
m_header.fb32.is_fast_run_cache = is_fast_run_cache;
if (!m_info_data_path.empty()) {
auto buf = read_file(m_info_data_path);
if (!m_info_data.empty()) {
std::string json_string(
static_cast<const char*>(static_cast<void*>(buf.data())), buf.size());
static_cast<const char*>(static_cast<void*>(m_info_data.data())),
m_info_data.size());
auto info = nlohmann::json::parse(json_string);
m_header.name = info["name"];
}
......@@ -221,6 +250,9 @@ void ModelPacker::pack_model() {
m_fbs_helper->builder().Finish(pack_model_builder.Finish());
FILE* fptr = fopen(m_packed_model_path.c_str(), "wb");
LITE_ASSERT(
fptr, "failed to open %s: %s", m_packed_model_path.c_str(),
strerror(errno));
std::string packed_model_tag = "packed_model";
auto nr_tag = fwrite(packed_model_tag.c_str(), 1, packed_model_tag.size(), fptr);
LITE_ASSERT(nr_tag == packed_model_tag.size());
......
......@@ -15,7 +15,8 @@ namespace json {
class Value : public std::enable_shared_from_this<Value>, public DynTypeObj {
public:
virtual void writeto(std::string& fout, int indent = 0) const = 0;
MGE_WIN_DECLSPEC_FUC virtual void writeto(
std::string& fout, int indent = 0) const = 0;
MGE_WIN_DECLSPEC_FUC void writeto_fpath(
const std::string& fout_path, int indent = 0) const {
......@@ -38,11 +39,11 @@ class Number final : public Value {
public:
Number(double v) : m_val(v) {}
static std::shared_ptr<Number> make(double v) {
MGE_WIN_DECLSPEC_FUC static std::shared_ptr<Number> make(double v) {
return std::make_shared<Number>(v);
}
void writeto(std::string& fout, int indent = 0) const override;
MGE_WIN_DECLSPEC_FUC void writeto(std::string& fout, int indent = 0) const override;
auto&& get_impl() { return m_val; }
......@@ -57,7 +58,7 @@ class NumberInt final : public Value {
public:
NumberInt(int64_t v) : m_val(v) {}
static std::shared_ptr<NumberInt> make(int64_t v) {
MGE_WIN_DECLSPEC_FUC static std::shared_ptr<NumberInt> make(int64_t v) {
return std::make_shared<NumberInt>(v);
}
......@@ -76,7 +77,7 @@ class Bool final : public Value {
public:
Bool(bool v) : m_val(v) {}
static std::shared_ptr<Bool> make(bool v);
MGE_WIN_DECLSPEC_FUC static std::shared_ptr<Bool> make(bool v);
MGE_WIN_DECLSPEC_FUC void writeto(std::string& fout, int indent = 0) const override;
......@@ -95,11 +96,13 @@ public:
String(char const* v) : m_val(v) {}
static std::shared_ptr<String> make(const std::string& v) {
MGE_WIN_DECLSPEC_FUC static std::shared_ptr<String> make(const std::string& v) {
return std::make_shared<String>(v);
}
bool operator==(const String& rhs) const { return m_val == rhs.m_val; }
MGE_WIN_DECLSPEC_FUC bool operator==(const String& rhs) const {
return m_val == rhs.m_val;
}
MGE_WIN_DECLSPEC_FUC void writeto(std::string& fout, int indent = 0) const override;
......@@ -114,9 +117,11 @@ class Object final : public Value {
std::unordered_map<String, std::shared_ptr<Value>, StdHashAdaptor<String>> m_val;
public:
static std::shared_ptr<Object> make() { return std::make_shared<Object>(); }
MGE_WIN_DECLSPEC_FUC static std::shared_ptr<Object> make() {
return std::make_shared<Object>();
}
static std::shared_ptr<Object> make(
MGE_WIN_DECLSPEC_FUC static std::shared_ptr<Object> make(
const std::vector<std::pair<String, std::shared_ptr<Value>>>& val) {
for (auto&& i : val)
mgb_assert(i.second);
......@@ -125,11 +130,17 @@ public:
return rst;
}
std::shared_ptr<Value>& operator[](const String& s) { return m_val[s]; }
MGE_WIN_DECLSPEC_FUC std::shared_ptr<Value>& operator[](const String& s) {
return m_val[s];
}
std::shared_ptr<Value>& operator[](const std::string& s) { return m_val[s]; }
MGE_WIN_DECLSPEC_FUC std::shared_ptr<Value>& operator[](const std::string& s) {
return m_val[s];
}
std::shared_ptr<Value>& operator[](const char* s) { return m_val[std::string(s)]; }
MGE_WIN_DECLSPEC_FUC std::shared_ptr<Value>& operator[](const char* s) {
return m_val[std::string(s)];
}
MGE_WIN_DECLSPEC_FUC void writeto(std::string& fout, int indent = 0) const override;
......@@ -144,14 +155,18 @@ class Array final : public Value {
std::vector<std::shared_ptr<Value>> m_val;
public:
static std::shared_ptr<Array> make() { return std::make_shared<Array>(); }
MGE_WIN_DECLSPEC_FUC static std::shared_ptr<Array> make() {
return std::make_shared<Array>();
}
void add(std::shared_ptr<Value> val) {
MGE_WIN_DECLSPEC_FUC void add(std::shared_ptr<Value> val) {
mgb_assert(val);
m_val.emplace_back(std::move(val));
}
std::shared_ptr<Value>& operator[](size_t idx) { return m_val.at(idx); }
MGE_WIN_DECLSPEC_FUC std::shared_ptr<Value>& operator[](size_t idx) {
return m_val.at(idx);
}
MGE_WIN_DECLSPEC_FUC void writeto(std::string& fout, int indent = 0) const override;
......@@ -164,7 +179,7 @@ class Null final : public Value {
MGB_DYN_TYPE_OBJ_FINAL_DECL_WITH_EXPORT;
public:
static std::shared_ptr<Value> make() {
MGE_WIN_DECLSPEC_FUC static std::shared_ptr<Value> make() {
static std::shared_ptr<Null> v(new Null);
return v;
}
......
......@@ -524,9 +524,14 @@ void ParamFusePass::apply(OptState& state) const {
{
auto orig_level = cg->options().log_level;
auto orig_record_level = cg->options().comp_node_seq_record_level;
cg->options().comp_node_seq_record_level = 0;
cg->options().log_level = 0;
MGB_TRY { cg->compile({{var, cb}})->execute(); }
MGB_FINALLY(cg->options().log_level = orig_level);
MGB_FINALLY({
cg->options().comp_node_seq_record_level = orig_record_level;
cg->options().log_level = orig_level;
});
}
SymbolVar new_var;
......
......@@ -398,7 +398,7 @@ void ImmutableTensor::Value::setup(CompNode cn, const HostTensorND& val) {
return true;
};
if (one_elem(val.shape())) {
if (!val.empty() && one_elem(val.shape())) {
float v;
static_cast_dtype(&v, val.dtype(), val.raw_ptr());
m_summary = ssprintf("const<%.3g>", v);
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册