diff --git a/CMakeLists.txt b/CMakeLists.txt
index cf67e6127fabfe998bedefda174fa38e2e358e39..f4e8c64c4ff73d0a417c35159901c2e67d0ca4ef 100644
--- a/CMakeLists.txt
+++ b/CMakeLists.txt
@@ -65,7 +65,7 @@ endif()
if (NOT CLIENT_ONLY)
include(external/jsoncpp)
-include(external/rocksdb)
+#include(external/rocksdb)
endif()
#include(external/gtest)
diff --git a/README.md b/README.md
index 8b0ab325b0b85b8fbcad3b22c07f42041340394d..bae5bb3834035a0b786e963d4e6fe4c72acdd72b 100644
--- a/README.md
+++ b/README.md
@@ -53,8 +53,9 @@ Paddle Serving provides HTTP and RPC based service for users to access
### HTTP service
+Paddle Serving provides a built-in python module called `paddle_serving_server.serve` that can start a rpc service or a http service with one-line command. If we specify the argument `--name uci`, it means that we will have a HTTP service with a url of `$IP:$PORT/uci/prediction`
``` shell
-python -m paddle_serving_server.web_serve --model uci_housing_model --thread 10 --port 9292 --name uci
+python -m paddle_serving_server.serve --model uci_housing_model --thread 10 --port 9292 --name uci
```
@@ -65,6 +66,7 @@ python -m paddle_serving_server.web_serve --model uci_housing_model --thread 10
| `name` | str | `""` | Service name, can be used to generate HTTP request url |
| `model` | str | `""` | Path of paddle model directory to be served |
+Here, we use `curl` to send a HTTP POST request to the service we just started. Users can use any python library to send HTTP POST as well, e.g, [requests](https://requests.readthedocs.io/en/master/).
``` shell
@@ -73,6 +75,7 @@ curl -H "Content-Type:application/json" -X POST -d '{"x": [0.0137, -0.1136, 0.25
### RPC service
+A user can also start a rpc service with `paddle_serving_server.serve`. RPC service is usually faster than HTTP service, although a user needs to do some coding based on Paddle Serving's python client API. Note that we do not specify `--name` here.
``` shell
python -m paddle_serving_server.serve --model uci_housing_model --thread 10 --port 9292
```
@@ -91,17 +94,37 @@ print(fetch_map)
```
-Applications you can do with Paddle Serving
+ Pre-built services with Paddle Serving
-
+Chinese Word Segmentation
-| Model Name | Resnet50 |
-|:--------------------: |:----------------------------------: |
-| Package URL | To be released |
-| Description | Get the representation of an image |
-| Training Data Source | Imagenet |
+- **Description**: Chinese word segmentation HTTP service that can be deployed with one line command.
+
+- **Download**:
+``` shell
+wget --no-check-certificate https://paddle-serving.bj.bcebos.com/lac/lac_model_jieba_web.tar.gz
+```
+- **Host web service**:
+``` shell
+tar -xzf lac_model_jieba_web.tar.gz
+python lac_web_service.py jieba_server_model/ lac_workdir 9292
+```
+- **Request sample**:
+``` shell
+curl -H "Content-Type:application/json" -X POST -d '{"words": "我爱北京天安门", "fetch":["word_seg"]}' http://127.0.0.1:9292/lac/prediction
+```
+- **Request result**:
+``` shell
+{"word_seg":"我|爱|北京|天安门"}
+```
+
+
+Chinese Sentence To Vector
+
+Image To Vector
+
+Image Classification
-
Document
@@ -119,6 +142,7 @@ print(fetch_map)
### About Efficiency
- [How profile serving efficiency?(Chinese)](https://github.com/PaddlePaddle/Serving/tree/develop/python/examples/util)
+- [Benchmarks](doc/BENCHMARK.md)
### FAQ
- [FAQ(Chinese)](doc/FAQ.md)
diff --git a/core/CMakeLists.txt b/core/CMakeLists.txt
index 3c775a9acb2ac17a99557bfd28c4e8a3cf20a8c5..7226a69a87163cf625f49f54fe2a7df996a98efd 100644
--- a/core/CMakeLists.txt
+++ b/core/CMakeLists.txt
@@ -14,7 +14,7 @@
if(NOT CLIENT_ONLY)
add_subdirectory(cube)
-add_subdirectory(kvdb)
+#add_subdirectory(kvdb)
endif()
add_subdirectory(configure)
add_subdirectory(pdcodegen)
diff --git a/core/general-server/CMakeLists.txt b/core/general-server/CMakeLists.txt
index 7cf0a04d837c992dfc2557f442ea4c33e494ed28..9056e229a51f56463dc2eec5629f219d00dc6a38 100644
--- a/core/general-server/CMakeLists.txt
+++ b/core/general-server/CMakeLists.txt
@@ -1,4 +1,3 @@
-include_directories(SYSTEM ${CMAKE_CURRENT_LIST_DIR}/../kvdb/include)
include_directories(SYSTEM ${CMAKE_CURRENT_LIST_DIR}/../../)
include(op/CMakeLists.txt)
include(proto/CMakeLists.txt)
@@ -25,8 +24,6 @@ target_link_libraries(serving pdserving)
target_link_libraries(serving cube-api)
target_link_libraries(serving utils)
-target_link_libraries(serving kvdb rocksdb)
-
if(WITH_GPU)
target_link_libraries(serving ${CUDA_LIBRARIES})
endif()
diff --git a/core/predictor/CMakeLists.txt b/core/predictor/CMakeLists.txt
index 3a0724fe25c83804a2d594bc47b0ac4d263bb2bd..1b9dc7b29845a2b8c7f958c1d8e836cb57e91d41 100644
--- a/core/predictor/CMakeLists.txt
+++ b/core/predictor/CMakeLists.txt
@@ -6,17 +6,16 @@ include(framework/CMakeLists.txt)
include(tools/CMakeLists.txt)
include(src/CMakeLists.txt)
-include_directories(SYSTEM ${CMAKE_CURRENT_LIST_DIR}/../kvdb/include)
add_library(pdserving ${pdserving_srcs})
set_source_files_properties(
${pdserving_srcs}
PROPERTIES
COMPILE_FLAGS "-Wno-strict-aliasing -Wno-unused-variable -Wno-non-virtual-dtor -Wno-error=non-virtual-dtor -Wno-error=delete-non-virtual-dtor")
-add_dependencies(pdserving protobuf kvdb boost brpc leveldb pdcodegen configure)
+add_dependencies(pdserving protobuf boost brpc leveldb pdcodegen configure)
target_link_libraries(pdserving
- brpc protobuf boost leveldb configure kvdb -lpthread -lcrypto -lm -lrt -lssl -ldl -lz)
+ brpc protobuf boost leveldb configure -lpthread -lcrypto -lm -lrt -lssl -ldl -lz)
# install
install(TARGETS pdserving
diff --git a/core/predictor/framework/resource.cpp b/core/predictor/framework/resource.cpp
index 61a132d13f28350b70e03d8113f537d393ee6282..647cd58d041802de27defb6033ba05b5234e0563 100644
--- a/core/predictor/framework/resource.cpp
+++ b/core/predictor/framework/resource.cpp
@@ -39,8 +39,6 @@ DynamicResource::~DynamicResource() {}
int DynamicResource::initialize() { return 0; }
-std::shared_ptr Resource::getDB() { return db; }
-
std::shared_ptr Resource::get_general_model_config() {
return _config;
}
@@ -155,9 +153,6 @@ int Resource::initialize(const std::string& path, const std::string& file) {
this->cube_config_fullpath = cube_config_fullpath;
}
- if (db.get() == nullptr) {
- db = RocksDBWrapper::RocksDBWrapperFactory("kvdb");
- }
THREAD_SETSPECIFIC(_tls_bspec_key, NULL);
return 0;
diff --git a/core/predictor/framework/resource.h b/core/predictor/framework/resource.h
index 70461f48446cd2e41e0531d1863298cf48d2cf0a..1a648f0fa363efa4eb915a64553949206ec96153 100644
--- a/core/predictor/framework/resource.h
+++ b/core/predictor/framework/resource.h
@@ -18,7 +18,6 @@
#include
#include
#include "core/cube/cube-api/include/cube_api.h"
-#include "core/kvdb/include/kvdb/paddle_rocksdb.h"
#include "core/predictor/common/inner_common.h"
#include "core/predictor/framework/infer.h"
#include "core/predictor/framework/memory.h"
@@ -101,8 +100,6 @@ class Resource {
void print_general_model_config(
const std::shared_ptr& config);
- std::shared_ptr getDB();
-
DynamicResource* get_dynamic_resource() {
return reinterpret_cast(
THREAD_GETSPECIFIC(_tls_bspec_key));
@@ -110,7 +107,6 @@ class Resource {
private:
int thread_finalize() { return 0; }
- std::shared_ptr db;
std::shared_ptr _config;
std::string cube_config_fullpath;
diff --git a/paddle_inference/inferencer-fluid-cpu/CMakeLists.txt b/paddle_inference/inferencer-fluid-cpu/CMakeLists.txt
index b3183eca1a0c2bb9524db7d7eb19f924079e0f0d..fe7cd91c534d900f3afb466e48599f60e64e4f9d 100644
--- a/paddle_inference/inferencer-fluid-cpu/CMakeLists.txt
+++ b/paddle_inference/inferencer-fluid-cpu/CMakeLists.txt
@@ -2,7 +2,7 @@ FILE(GLOB fluid_cpu_engine_srcs ${CMAKE_CURRENT_LIST_DIR}/src/*.cpp)
add_library(fluid_cpu_engine ${fluid_cpu_engine_srcs})
target_include_directories(fluid_cpu_engine PUBLIC
${CMAKE_BINARY_DIR}/Paddle/fluid_install_dir/)
-add_dependencies(fluid_cpu_engine pdserving extern_paddle configure kvdb)
+add_dependencies(fluid_cpu_engine pdserving extern_paddle configure)
target_link_libraries(fluid_cpu_engine pdserving paddle_fluid -lpthread -lcrypto -lm -lrt -lssl -ldl -lz)
install(TARGETS fluid_cpu_engine
diff --git a/paddle_inference/inferencer-fluid-gpu/CMakeLists.txt b/paddle_inference/inferencer-fluid-gpu/CMakeLists.txt
index 6c980d6d94ef99585cf956072d8181d0d98cda42..725da85b45ca1070badf5343f340e49dce6b936f 100644
--- a/paddle_inference/inferencer-fluid-gpu/CMakeLists.txt
+++ b/paddle_inference/inferencer-fluid-gpu/CMakeLists.txt
@@ -2,7 +2,7 @@ FILE(GLOB fluid_gpu_engine_srcs ${CMAKE_CURRENT_LIST_DIR}/src/*.cpp)
add_library(fluid_gpu_engine ${fluid_gpu_engine_srcs})
target_include_directories(fluid_gpu_engine PUBLIC
${CMAKE_BINARY_DIR}/Paddle/fluid_install_dir/)
-add_dependencies(fluid_gpu_engine pdserving extern_paddle configure kvdb)
+add_dependencies(fluid_gpu_engine pdserving extern_paddle configure)
target_link_libraries(fluid_gpu_engine pdserving paddle_fluid iomp5 mklml_intel -lpthread -lcrypto -lm -lrt -lssl -ldl -lz)
install(TARGETS fluid_gpu_engine
diff --git a/python/paddle_serving_server/serve.py b/python/paddle_serving_server/serve.py
index c86c3f46b3b3ef83fb5fe630031cf28a95c52649..279e3a895e975473fc5569c4716368c3dda1d9f1 100644
--- a/python/paddle_serving_server/serve.py
+++ b/python/paddle_serving_server/serve.py
@@ -18,6 +18,7 @@ Usage:
python -m paddle_serving_server.serve --model ./serving_server_model --port 9292
"""
import argparse
+from .web_service import WebService
def parse_args(): # pylint: disable=doc-string-missing
@@ -28,6 +29,8 @@ def parse_args(): # pylint: disable=doc-string-missing
"--model", type=str, default="", help="Model for serving")
parser.add_argument(
"--port", type=int, default=9292, help="Port the server")
+ parser.add_argument(
+ "--name", type=str, default="None", help="Web service name")
parser.add_argument(
"--workdir",
type=str,
@@ -71,4 +74,13 @@ def start_standard_model(): # pylint: disable=doc-string-missing
if __name__ == "__main__":
- start_standard_model()
+
+ args = parse_args()
+ if args.name == "None":
+ start_standard_model()
+ else:
+ service = WebService(name=args.name)
+ service.load_model_config(args.model)
+ service.prepare_server(
+ workdir=args.workdir, port=args.port, device=args.device)
+ service.run_server()
diff --git a/python/paddle_serving_server/web_serve.py b/python/paddle_serving_server/web_serve.py
deleted file mode 100644
index 46437ad5e53288c6ab03b32ea8882e1b3cfa66a3..0000000000000000000000000000000000000000
--- a/python/paddle_serving_server/web_serve.py
+++ /dev/null
@@ -1,51 +0,0 @@
-# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
-#
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-"""
-Usage:
- Host a trained paddle model with one line command
- Example:
- python -m paddle_serving_server.web_serve --model ./serving_server_model --port 9292
-"""
-import argparse
-from multiprocessing import Pool, Process
-from .web_service import WebService
-
-
-def parse_args(): # pylint: disable=doc-string-missing
- parser = argparse.ArgumentParser("web_serve")
- parser.add_argument(
- "--thread", type=int, default=10, help="Concurrency of server")
- parser.add_argument(
- "--model", type=str, default="", help="Model for serving")
- parser.add_argument(
- "--port", type=int, default=9292, help="Port the server")
- parser.add_argument(
- "--workdir",
- type=str,
- default="workdir",
- help="Working dir of current service")
- parser.add_argument(
- "--device", type=str, default="cpu", help="Type of device")
- parser.add_argument(
- "--name", type=str, default="default", help="Default service name")
- return parser.parse_args()
-
-
-if __name__ == "__main__":
- args = parse_args()
- service = WebService(name=args.name)
- service.load_model_config(args.model)
- service.prepare_server(
- workdir=args.workdir, port=args.port, device=args.device)
- service.run_server()
diff --git a/python/paddle_serving_server_gpu/__init__.py b/python/paddle_serving_server_gpu/__init__.py
index 8ee1e137fb8fe282d26bda95e4b4bffa6f670f11..02b55801c35fb5d1ed7e35c249ac07e4d3eb45ab 100644
--- a/python/paddle_serving_server_gpu/__init__.py
+++ b/python/paddle_serving_server_gpu/__init__.py
@@ -42,7 +42,7 @@ def serve_args():
"--device", type=str, default="gpu", help="Type of device")
parser.add_argument("--gpu_ids", type=str, default="", help="gpu ids")
parser.add_argument(
- "--name", type=str, default="default", help="Default service name")
+ "--name", type=str, default="None", help="Default service name")
return parser.parse_args()
diff --git a/python/paddle_serving_server_gpu/serve.py b/python/paddle_serving_server_gpu/serve.py
index cc9b18f6920c46c5d0119e8adfaf8f76ecf2ad26..5d9d96d517d64b21313fda0b44a83b34142b014b 100644
--- a/python/paddle_serving_server_gpu/serve.py
+++ b/python/paddle_serving_server_gpu/serve.py
@@ -88,4 +88,18 @@ def start_multi_card(args): # pylint: disable=doc-string-missing
if __name__ == "__main__":
args = serve_args()
- start_multi_card(args)
+ if args.name == "None":
+ start_multi_card(args)
+ else:
+ web_service = WebService(name=args.name)
+ web_service.load_model_config(args.model)
+ gpu_ids = []
+ if args.gpu_ids == "":
+ if "CUDA_VISIBLE_DEVICES" in os.environ:
+ gpu_ids = os.environ["CUDA_VISIBLE_DEVICES"]
+ if len(gpu_ids) > 0:
+ gpus = [int(x) for x in gpu_ids.split(",")]
+ web_service.set_gpus(gpus)
+ web_service.prepare_server(
+ workdir=args.workdir, port=args.port, device=args.device)
+ web_service.run_server()
diff --git a/python/paddle_serving_server_gpu/web_serve.py b/python/paddle_serving_server_gpu/web_serve.py
deleted file mode 100644
index 734e6d7b93b4f3ad22f330b1545b63c6ac6f2838..0000000000000000000000000000000000000000
--- a/python/paddle_serving_server_gpu/web_serve.py
+++ /dev/null
@@ -1,39 +0,0 @@
-# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
-#
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-"""
-Usage:
- Host a trained paddle model with one line command
- Example:
- python -m paddle_serving_server.web_serve --model ./serving_server_model --port 9292
-"""
-import os
-from multiprocessing import Pool, Process
-from .web_service import WebService
-import paddle_serving_server_gpu as serving
-from paddle_serving_server_gpu import serve_args
-
-if __name__ == "__main__":
- args = serve_args()
- web_service = WebService(name=args.name)
- web_service.load_model_config(args.model)
- gpu_ids = []
- if args.gpu_ids == "":
- if "CUDA_VISIBLE_DEVICES" in os.environ:
- gpu_ids = os.environ["CUDA_VISIBLE_DEVICES"]
- if len(gpu_ids) > 0:
- gpus = [int(x) for x in gpu_ids.split(",")]
- web_service.set_gpus(gpus)
- web_service.prepare_server(
- workdir=args.workdir, port=args.port, device=args.device)
- web_service.run_server()
diff --git a/tools/serving_build.sh b/tools/serving_build.sh
index 163f4f74429066581aa17cc78b3ab00947ba4d77..b810e3139803bd363c771c6f655cef6595177dc8 100644
--- a/tools/serving_build.sh
+++ b/tools/serving_build.sh
@@ -83,7 +83,7 @@ function python_test_fit_a_line() {
check_cmd "python test_client.py uci_housing_client/serving_client_conf.prototxt > /dev/null"
ps -ef | grep "paddle_serving_server" | grep -v grep | awk '{print $2}' | xargs kill
# test web
- check_cmd "python -m paddle_serving_server.web_serve --model uci_housing_model/ --name uci --port 9399 --name uci > /dev/null &"
+ check_cmd "python -m paddle_serving_server.serve --model uci_housing_model/ --name uci --port 9399 --name uci > /dev/null &"
sleep 5
check_cmd "curl -H \"Content-Type:application/json\" -X POST -d '{\"x\": [0.0137, -0.1136, 0.2553, -0.0692, 0.0582, -0.0727, -0.1583, -0.0584, 0.6283, 0.4919, 0.1856, 0.0795, -0.0332], \"fetch\":[\"price\"]}' http://127.0.0.1:9399/uci/prediction"
ps -ef | grep "paddle_serving_server" | grep -v grep | awk '{print $2}' | xargs kill