// Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include "paddle/fluid/inference/utils/io_utils.h" #include #include #include "paddle/fluid/inference/analysis/helper.h" namespace paddle { namespace inference { // ========================================================= // Item | Type | Bytes // --------------------------------------------------------- // Version | uint32_t | 4 // --------------------------------------------------------- // Bytes of `Name` | uint64_t | 8 // Name | char | Bytes of `Name` // --------------------------------------------------------- // LoD Level | uint64_t | 8 // Bytes of `LoD[0]`| uint64_t | 8 // LoD[0] | uint64_t | Bytes of `LoD[0]` // ... | ... | ... // --------------------------------------------------------- // Dims of `Shape` | uint64_t | 8 // Shape | uint64_t | Dims * 4 // --------------------------------------------------------- // Dtype | int32_t | 4 // Bytes of `Data` | uint64_t | 8 // Data | Dtype | Bytes of `Data` // ========================================================= void SerializePDTensorToStream(std::ostream *os, const PaddleTensor &tensor) { // 1. Version os->write(reinterpret_cast(&kCurPDTensorVersion), sizeof(kCurPDTensorVersion)); // 2. Name uint64_t name_bytes = tensor.name.size(); os->write(reinterpret_cast(&name_bytes), sizeof(name_bytes)); os->write(tensor.name.c_str(), name_bytes); // 3. LoD auto lod = tensor.lod; uint64_t lod_size = lod.size(); os->write(reinterpret_cast(&lod_size), sizeof(lod_size)); for (auto &each : lod) { auto size = each.size() * sizeof(size_t); os->write(reinterpret_cast(&size), sizeof(size)); os->write(reinterpret_cast(each.data()), static_cast(size)); } // 4. Shape size_t dims = tensor.shape.size(); os->write(reinterpret_cast(&dims), sizeof(dims)); os->write(reinterpret_cast(tensor.shape.data()), sizeof(int) * dims); // 5. Data os->write(reinterpret_cast(&tensor.dtype), sizeof(tensor.dtype)); uint64_t length = tensor.data.length(); os->write(reinterpret_cast(&length), sizeof(size_t)); os->write(reinterpret_cast(tensor.data.data()), length); } void DeserializePDTensorToStream(std::istream &is, PaddleTensor *tensor) { // 1. Version uint32_t version; is.read(reinterpret_cast(&version), sizeof(version)); // 2. Name uint64_t name_bytes; is.read(reinterpret_cast(&name_bytes), sizeof(name_bytes)); std::vector bytes(name_bytes); is.read(bytes.data(), name_bytes); tensor->name = std::string(bytes.data(), name_bytes); // 3. LoD uint64_t lod_level; is.read(reinterpret_cast(&lod_level), sizeof(lod_level)); auto *lod = &(tensor->lod); lod->resize(lod_level); for (uint64_t i = 0; i < lod_level; ++i) { uint64_t size; is.read(reinterpret_cast(&size), sizeof(size)); std::vector tmp(size / sizeof(size_t)); is.read(reinterpret_cast(tmp.data()), static_cast(size)); (*lod)[i] = tmp; } // 4. Shape size_t dims; is.read(reinterpret_cast(&dims), sizeof(dims)); tensor->shape.resize(dims); is.read(reinterpret_cast(tensor->shape.data()), sizeof(int) * dims); // 5. Data uint64_t length; is.read(reinterpret_cast(&tensor->dtype), sizeof(tensor->dtype)); is.read(reinterpret_cast(&length), sizeof(length)); tensor->data.Resize(length); is.read(reinterpret_cast(tensor->data.data()), length); } // ========================================================= // Item | Type | Bytes // --------------------------------------------------------- // Version | uint32_t | 4 // --------------------------------------------------------- // Size of Tensors | uint64_t | 8 // Tensors | ---- | --- // --------------------------------------------------------- void SerializePDTensorsToStream(std::ostream *os, const std::vector &tensors) { // 1. Version os->write(reinterpret_cast(&kCurPDTensorVersion), sizeof(kCurPDTensorVersion)); // 2. Tensors uint64_t num = tensors.size(); os->write(reinterpret_cast(&num), sizeof(num)); for (const auto &tensor : tensors) { SerializePDTensorToStream(os, tensor); } } void DeserializePDTensorsToStream(std::istream &is, std::vector *tensors) { // 1. Version uint32_t version; is.read(reinterpret_cast(&version), sizeof(version)); // 2. Tensors uint64_t num; is.read(reinterpret_cast(&num), sizeof(num)); tensors->resize(num); for (auto &tensor : *tensors) { DeserializePDTensorToStream(is, &tensor); } } void SerializePDTensorsToFile(const std::string &path, const std::vector &tensors) { std::ofstream fout(path, std::ios::binary); SerializePDTensorsToStream(&fout, tensors); fout.close(); } void DeserializePDTensorsToFile(const std::string &path, std::vector *tensors) { bool is_present = analysis::FileExists(path); PADDLE_ENFORCE_EQ(is_present, true, platform::errors::InvalidArgument( "Cannot open %s to read", path)); std::ifstream fin(path, std::ios::binary); DeserializePDTensorsToStream(fin, tensors); fin.close(); } } // namespace inference } // namespace paddle