general_reader_op.cpp 7.9 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14
// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

M
MRXLT 已提交
15
#include "core/general-server/op/general_reader_op.h"
16 17 18 19
#include <algorithm>
#include <iostream>
#include <memory>
#include <sstream>
20
#include "core/general-server/op/general_infer_helper.h"
21 22
#include "core/predictor/framework/infer.h"
#include "core/predictor/framework/memory.h"
G
guru4elephant 已提交
23
#include "core/util/include/timer.h"
24 25 26 27 28

namespace baidu {
namespace paddle_serving {
namespace serving {

G
guru4elephant 已提交
29
using baidu::paddle_serving::Timer;
30 31 32 33
using baidu::paddle_serving::predictor::MempoolWrapper;
using baidu::paddle_serving::predictor::general_model::Tensor;
using baidu::paddle_serving::predictor::general_model::Request;
using baidu::paddle_serving::predictor::PaddleGeneralModelConfig;
S
ShiningZhang 已提交
34
// support: FLOAT32, INT64, INT32, UINT8, INT8, FLOAT16
S
ShiningZhang 已提交
35 36 37 38 39 40 41 42 43 44 45 46 47 48 49
enum ProtoDataType {
  P_INT64 = 0,
  P_FLOAT32,
  P_INT32,
  P_FP64,
  P_INT16,
  P_FP16,
  P_BF16,
  P_UINT8,
  P_INT8,
  P_BOOL,
  P_COMPLEX64,
  P_COMPLEX128,
  P_STRING = 20,
};
50 51

int GeneralReaderOp::inference() {
H
HexToString 已提交
52
  // read request from client
W
wangjiawei04 已提交
53
  const Request *req = dynamic_cast<const Request *>(get_request_message());
54 55 56 57 58
  if (!req) {
    LOG(ERROR) << "Failed get request message";
    return -1;
  }

W
wangjiawei04 已提交
59 60
  uint64_t log_id = req->log_id();
  int input_var_num = 0;
H
HexToString 已提交
61 62 63

  GeneralBlob *res = mutable_data<GeneralBlob>();
  if (!res) {
64 65
    LOG(ERROR) << "(logid=" << log_id << ") Failed get GeneralBlob";
    return -1;
H
HexToString 已提交
66 67
  }

68 69 70 71 72 73 74
  TensorVector *out = &(res->tensor_vector);
  if (!out) {
    LOG(ERROR) << "(logid=" << log_id << ") Failed get tensor_vector of res";
    return -1;
  }

  res->SetLogId(log_id);
H
HexToString 已提交
75 76
  Timer timeline;
  int64_t start = timeline.TimeStampUS();
77
  // var_num means the number of feed_var.
H
HexToString 已提交
78
  int var_num = req->tensor_size();
H
HexToString 已提交
79

H
HexToString 已提交
80
  VLOG(2) << "(logid=" << log_id << ") var num: " << var_num
H
HexToString 已提交
81
          << ") start to call load general model_conf op";
S
ShiningZhang 已提交
82 83 84 85 86
  if (var_num < 1) {
    LOG(ERROR) << "(logid=" << log_id << ") Failed get feed_var, var_num="
               << var_num;
    return -1;
  }
H
HexToString 已提交
87

W
wangjiawei04 已提交
88 89
  baidu::paddle_serving::predictor::Resource &resource =
      baidu::paddle_serving::predictor::Resource::instance();
H
HexToString 已提交
90 91 92 93 94

  VLOG(2) << "(logid=" << log_id << ") get resource pointer done.";

  // package tensor
  // prepare basic information for input
H
HexToString 已提交
95 96 97
  // specify the memory needed for output tensor_vector
  // fill the data into output general_blob
  int data_len = 0;
H
HexToString 已提交
98 99 100
  int64_t elem_type = 0;
  int64_t elem_size = 0;
  int64_t databuf_size = 0;
S
ShiningZhang 已提交
101
  const void* src_ptr = nullptr;
W
wangjiawei04 已提交
102
  for (int i = 0; i < var_num; ++i) {
H
HexToString 已提交
103
    paddle::PaddleTensor paddleTensor;
H
HexToString 已提交
104
    const Tensor &tensor = req->tensor(i);
H
HexToString 已提交
105
    data_len = 0;
H
HexToString 已提交
106 107 108 109
    elem_type = 0;
    elem_size = 0;
    databuf_size = 0;
    elem_type = tensor.elem_type();
S
ShiningZhang 已提交
110
    src_ptr = nullptr ;
H
HexToString 已提交
111 112 113
    if (elem_type == P_INT64) {  // int64
      elem_size = sizeof(int64_t);
      paddleTensor.dtype = paddle::PaddleDType::INT64;
H
HexToString 已提交
114
      data_len = tensor.int64_data_size();
S
ShiningZhang 已提交
115
      src_ptr = tensor.int64_data().data();
H
HexToString 已提交
116 117 118
    } else if (elem_type == P_FLOAT32) {
      elem_size = sizeof(float);
      paddleTensor.dtype = paddle::PaddleDType::FLOAT32;
H
HexToString 已提交
119
      data_len = tensor.float_data_size();
S
ShiningZhang 已提交
120
      src_ptr = tensor.float_data().data();
H
HexToString 已提交
121 122 123
    } else if (elem_type == P_INT32) {
      elem_size = sizeof(int32_t);
      paddleTensor.dtype = paddle::PaddleDType::INT32;
H
HexToString 已提交
124
      data_len = tensor.int_data_size();
S
ShiningZhang 已提交
125 126 127 128 129 130 131 132 133 134 135 136
      src_ptr = tensor.int_data().data();
    } else if (elem_type == P_UINT8) {
      elem_size = sizeof(uint8_t);
      paddleTensor.dtype = paddle::PaddleDType::UINT8;
      data_len = tensor.tensor_content().size();
      src_ptr = tensor.tensor_content().data();
    } else if (elem_type == P_INT8) {
      elem_size = sizeof(int8_t);
      paddleTensor.dtype = paddle::PaddleDType::INT8;
      data_len = tensor.tensor_content().size();
      src_ptr = tensor.tensor_content().data();
    } else if (elem_type == P_FP16) {
S
ShiningZhang 已提交
137 138 139 140 141
      // copy bytes from tensor content to TensorVector
      elem_size = 1;
      paddleTensor.dtype = paddle::PaddleDType::FLOAT16;
      data_len = tensor.tensor_content().size();
      src_ptr = tensor.tensor_content().data();
H
HexToString 已提交
142
    } else if (elem_type == P_STRING) {
143
      // use paddle::PaddleDType::UINT8 as for String.
H
HexToString 已提交
144 145
      elem_size = sizeof(char);
      paddleTensor.dtype = paddle::PaddleDType::UINT8;
146 147
      // this is for vector<String>, cause the databuf_size !=
      // vector<String>.size()*sizeof(char);
H
HexToString 已提交
148 149
      // data_len should be +1 cause '\0'
      // now only support single string
H
HexToString 已提交
150
      for (int idx = 0; idx < tensor.data_size(); idx++) {
H
HexToString 已提交
151
        data_len += tensor.data()[idx].length() + 1;
S
ShiningZhang 已提交
152
        src_ptr = tensor.data()[idx].data();
H
HexToString 已提交
153
      }
H
HexToString 已提交
154
    }
S
ShiningZhang 已提交
155 156 157 158 159 160 161 162 163
    VLOG(2) << "var[" << i << "] has elem type: " << elem_type << ";"
            << "elem_size=" << elem_size << ";"
            << "dtype=" << paddleTensor.dtype << ";"
            << "data_len=" << data_len;
    if (src_ptr == nullptr) {
      LOG(ERROR) << "Not support var[" << i << "] with elem_type[" 
                 << elem_type << "]";
      continue;
    }
H
HexToString 已提交
164
    // implement lod tensor here
H
HexToString 已提交
165
    // only support 1-D lod
H
HexToString 已提交
166
    // TODO(HexToString): support 2-D lod
H
HexToString 已提交
167
    if (tensor.lod_size() > 0) {
H
HexToString 已提交
168
      VLOG(2) << "(logid=" << log_id << ") var[" << i << "] is lod_tensor";
H
HexToString 已提交
169
      paddleTensor.lod.resize(1);
H
HexToString 已提交
170
      for (int k = 0; k < tensor.lod_size(); ++k) {
H
HexToString 已提交
171
        paddleTensor.lod[0].push_back(tensor.lod(k));
W
wangjiawei04 已提交
172 173
      }
    }
H
HexToString 已提交
174

H
HexToString 已提交
175 176
    for (int k = 0; k < tensor.shape_size(); ++k) {
      int dim = tensor.shape(k);
177
      VLOG(2) << "(logid=" << log_id << ") shape for var[" << i << "]: " << dim;
H
HexToString 已提交
178
      paddleTensor.shape.push_back(dim);
H
HexToString 已提交
179
    }
H
HexToString 已提交
180
    paddleTensor.name = tensor.name();
H
HexToString 已提交
181
    out->push_back(paddleTensor);
H
HexToString 已提交
182

H
HexToString 已提交
183 184
    VLOG(2) << "(logid=" << log_id << ") tensor size for var[" << i
            << "]: " << data_len;
H
HexToString 已提交
185
    databuf_size = data_len * elem_size;
H
HexToString 已提交
186 187 188 189
    void *databuf_char = MempoolWrapper::instance().malloc(databuf_size);
    paddle::PaddleBuf paddleBuf(databuf_char, databuf_size);
    out->at(i).data = paddleBuf;
    // out->at(i).data.Resize(databuf_size);
H
HexToString 已提交
190 191 192 193
    if (out->at(i).lod.size() > 0) {
      VLOG(2) << "(logid=" << log_id << ") var[" << i
              << "] has lod_tensor and len=" << out->at(i).lod[0].back();
    }
S
ShiningZhang 已提交
194 195 196 197 198 199 200 201 202 203 204
    void* dst_ptr = out->at(i).data.data();
    if (!dst_ptr) {
      LOG(ERROR) << "dst_ptr is nullptr";
      return -1;
    }

    // For common data, we just copy from src to dst
    // For string data, we need to iterate through all str
    if (elem_type != P_STRING) {
      memcpy(dst_ptr, src_ptr, databuf_size);
    } else {
H
HexToString 已提交
205
      char *dst_ptr = static_cast<char *>(out->at(i).data.data());
H
HexToString 已提交
206
      VLOG(2) << "(logid=" << log_id << ") first element data in var[" << i
H
HexToString 已提交
207
              << "] is " << tensor.data(0);
H
HexToString 已提交
208 209
      if (!dst_ptr) {
        LOG(ERROR) << "dst_ptr is nullptr";
210
        return -1;
H
HexToString 已提交
211
      }
H
HexToString 已提交
212
      int elem_num = tensor.data_size();
H
HexToString 已提交
213
      int offset = 0;
W
wangjiawei04 已提交
214
      for (int k = 0; k < elem_num; ++k) {
H
HexToString 已提交
215 216 217 218
        memcpy(dst_ptr + offset,
               tensor.data(k).c_str(),
               strlen(tensor.data(k).c_str()) + 1);
        offset += strlen(tensor.data(k).c_str()) + 1;
W
wangjiawei04 已提交
219 220 221
      }
    }
  }
H
HexToString 已提交
222 223 224 225 226 227 228 229 230 231

  VLOG(2) << "(logid=" << log_id << ") output size: " << out->size();
  timeline.Pause();
  int64_t end = timeline.TimeStampUS();
  res->p_size = 0;
  res->_batch_size = 1;
  AddBlobInfo(res, start);
  AddBlobInfo(res, end);

  VLOG(2) << "(logid=" << log_id << ") read data from client success";
232 233 234 235 236
  return 0;
}
DEFINE_OP(GeneralReaderOp);
}  // namespace serving
}  // namespace paddle_serving
H
HexToString 已提交
237
}  // namespace baidu