// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. // This file contains model format related operations, such as load a model, // parse an operator definitions and so on. #pragma once #include #include #include #ifndef LITE_ON_TINY_PUBLISH #include "lite/core/framework.pb.h" #endif #include "lite/core/scope.h" #include "lite/core/variable.h" #include "lite/model_parser/compatible_pb.h" #include "lite/model_parser/naive_buffer/proto/framework.nb.h" namespace paddle { namespace lite { #ifndef LITE_ON_TINY_PUBLISH // Read a __model__ file. std::unique_ptr LoadProgram( const std::string& path); // Read a single file containing all the parameters. void LoadParams(const std::string& path); // Load a single parameter to an output tensor. void LoadParam(const std::string& path, Variable* out); void LoadCombinedParamsPb(const std::string& path, lite::Scope* scope, const cpp::ProgramDesc& prog); // Read a model and files of parameters in pb format. void LoadModelPb(const std::string& model_dir, const std::string& model_file, const std::string& param_file, Scope* scope, cpp::ProgramDesc* prog, bool combined = false); // Save a model and files of parameters in pb format. void SaveModelPb(const std::string& model_dir, const Scope& scope, const cpp::ProgramDesc& prog, bool combined = false); void SaveCombinedParamsPb(const std::string& path, const lite::Scope& exec_scope, const cpp::ProgramDesc& prog); // Serialize tensors to ostream. void SerializeTensor(std::ostream& os, const lite::Scope& scope, const std::string& var); // LoDTensor to ostream void TensorToStream(std::ostream& os, const lite::Tensor& tensor); void ReadBinaryFile(const std::string& filename, std::string* contents); // For naive buffer void SaveParamNaive(const std::string& path, const lite::Scope& exec_scope, const std::string& var_name); void SaveCombinedParamsNaive(const std::string& path, const lite::Scope& exec_scope, const cpp::ProgramDesc& cpp_prog); void SaveModelNaive(const std::string& model_dir, const Scope& exec_scope, const cpp::ProgramDesc& cpp_prog, bool combined = true); #endif void LoadParamNaive(const std::string& path, lite::Scope* scope, const std::string& name); void LoadCombinedParamsNaive(const std::string& path, lite::Scope* scope, const cpp::ProgramDesc& cpp_prog); void LoadModelNaive(const std::string& model_dir, lite::Scope* scope, cpp::ProgramDesc* prog, bool combined = true); } // namespace lite } // namespace paddle