// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include #include #include #include #include "predictor/builtin_format.pb.h" #include "sdk-cpp/include/common.h" #include "sdk-cpp/include/predictor_sdk.h" #include "sdk-cpp/int64tensor_service.pb.h" using baidu::paddle_serving::sdk_cpp::Predictor; using baidu::paddle_serving::sdk_cpp::PredictorApi; using baidu::paddle_serving::predictor::int64tensor_service::Request; using baidu::paddle_serving::predictor::int64tensor_service::Response; using baidu::paddle_serving::predictor::format::Float32TensorPredictor; using baidu::paddle_serving::predictor::format::Int64TensorInstance; int create_req(Request& req) { // NOLINT Int64TensorInstance* ins = req.mutable_instances()->Add(); ins->add_data(1); ins->add_data(2); ins->add_data(3); ins->add_data(4); ins->add_shape(2); ins->add_shape(2); ins = req.mutable_instances()->Add(); ins->add_data(5); ins->add_data(6); ins->add_data(7); ins->add_data(8); ins->add_data(9); ins->add_shape(5); ins->add_shape(1); return 0; } void print_res(const Request& req, const Response& res, std::string route_tag, uint64_t elapse_ms) { for (uint32_t i = 0; i < res.predictions_size(); ++i) { const Float32TensorPredictor& prediction = res.predictions(i); std::ostringstream oss1; for (uint32_t j = 0; j < prediction.data_size(); ++j) { oss1 << prediction.data(j) << " "; } std::ostringstream oss2; for (uint32_t j = 0; j < prediction.shape_size(); ++j) { oss2 << prediction.shape(j) << " "; } LOG(INFO) << "Receive result " << oss1.str() << ", shape " << oss2.str(); } LOG(INFO) << "Succ call predictor[int64tensor_format], the tag is: " << route_tag << ", elapse_ms: " << elapse_ms; } int main(int argc, char** argv) { PredictorApi api; // initialize logger instance struct stat st_buf; int ret = 0; if ((ret = stat("./log", &st_buf)) != 0) { mkdir("./log", 0777); ret = stat("./log", &st_buf); if (ret != 0) { LOG(WARNING) << "Log path ./log not exist, and create fail"; return -1; } } FLAGS_log_dir = "./log"; google::InitGoogleLogging(strdup(argv[0])); if (api.create("./conf", "predictors.prototxt") != 0) { LOG(ERROR) << "Failed create predictors api!"; return -1; } Request req; Response res; api.thrd_initialize(); while (true) { timeval start; gettimeofday(&start, NULL); api.thrd_clear(); Predictor* predictor = api.fetch_predictor("int64tensor_service"); if (!predictor) { LOG(ERROR) << "Failed fetch predictor: int64tensor_service"; return -1; } req.Clear(); res.Clear(); if (create_req(req) != 0) { return -1; } butil::IOBufBuilder debug_os; if (predictor->debug(&req, &res, &debug_os) != 0) { LOG(ERROR) << "failed call predictor with req:" << req.ShortDebugString(); return -1; } butil::IOBuf debug_buf; debug_os.move_to(debug_buf); LOG(INFO) << "Debug string: " << debug_buf; timeval end; gettimeofday(&end, NULL); uint64_t elapse_ms = (end.tv_sec * 1000 + end.tv_usec / 1000) - (start.tv_sec * 1000 + start.tv_usec / 1000); print_res(req, res, predictor->tag(), elapse_ms); res.Clear(); usleep(50); } // while (true) api.thrd_finalize(); api.destroy(); google::ShutdownGoogleLogging(); return 0; } /* vim: set expandtab ts=4 sw=4 sts=4 tw=100: */