Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
Serving
提交
2710f973
S
Serving
项目概览
PaddlePaddle
/
Serving
大约 1 年 前同步成功
通知
186
Star
833
Fork
253
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
105
列表
看板
标记
里程碑
合并请求
10
Wiki
2
Wiki
分析
仓库
DevOps
项目成员
Pages
S
Serving
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
105
Issue
105
列表
看板
标记
里程碑
合并请求
10
合并请求
10
Pages
分析
分析
仓库分析
DevOps
Wiki
2
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
2710f973
编写于
7月 27, 2021
作者:
J
Jiawei Wang
提交者:
GitHub
7月 27, 2021
浏览文件
操作
浏览文件
下载
差异文件
Merge pull request #1304 from ZhangYulongg/ci-test
update CI scripts and fix example
上级
a2cb6e86
5033b0e0
变更
2
隐藏空白更改
内联
并排
Showing
2 changed file
with
62 addition
and
11 deletion
+62
-11
python/examples/bert/bert_web_service_gpu.py
python/examples/bert/bert_web_service_gpu.py
+1
-0
tools/scripts/ipipe_py3.sh
tools/scripts/ipipe_py3.sh
+61
-11
未找到文件。
python/examples/bert/bert_web_service_gpu.py
浏览文件 @
2710f973
...
...
@@ -42,6 +42,7 @@ class BertService(WebService):
bert_service
=
BertService
(
name
=
"bert"
)
bert_service
.
load
()
bert_service
.
load_model_config
(
sys
.
argv
[
1
])
bert_service
.
set_gpus
(
"0"
)
bert_service
.
prepare_server
(
workdir
=
"workdir"
,
port
=
int
(
sys
.
argv
[
2
]),
device
=
"gpu"
)
bert_service
.
run_rpc_service
()
...
...
tools/scripts/ipipe_py3.sh
浏览文件 @
2710f973
...
...
@@ -120,31 +120,66 @@ function check() {
fi
}
function
check_gpu_memory
()
{
gpu_memory
=
`
nvidia-smi
--id
=
$1
--format
=
csv,noheader
--query-gpu
=
memory.used |
awk
'{print $1}'
`
echo
-e
"
${
GREEN_COLOR
}
-------id-
$1
gpu_memory_used:
${
gpu_memory
}${
RES
}
"
if
[
${
gpu_memory
}
-le
100
]
;
then
echo
"-------GPU-
$1
is not used"
status
=
"GPU-
$1
is not used"
else
echo
"-------GPU_memory used is expected"
fi
}
function
check_result
()
{
if
[
$?
==
0
]
;
then
echo
-e
"
${
GREEN_COLOR
}
$1
execute normally
${
RES
}
"
if
[
$1
==
"server"
]
;
then
sleep
$2
tail
${
dir
}
server_log.txt |
tee
-a
${
log_dir
}
server_total.txt
cat
${
dir
}
server_log.txt |
tee
-a
${
log_dir
}
server_total.txt
fi
if
[
$1
==
"client"
]
;
then
tail
${
dir
}
client_log.txt |
tee
-a
${
log_dir
}
client_total.txt
cat
${
dir
}
client_log.txt |
tee
-a
${
log_dir
}
client_total.txt
grep
-E
"
${
error_words
}
"
${
dir
}
client_log.txt
>
/dev/null
if
[
$?
==
0
]
;
then
if
[
"
${
status
}
"
!=
""
]
;
then
status
=
"
${
status
}
|Failed"
else
status
=
"Failed"
fi
echo
-e
"
${
RED_COLOR
}
$1
error command
${
RES
}
\n
"
|
tee
-a
${
log_dir
}
server_total.txt
${
log_dir
}
client_total.txt
echo
-e
"--------------pipeline.log:----------------
\n
"
echo
"--------------server log:--------------"
cat
${
dir
}
server_log.txt
echo
"--------------client log:--------------"
cat
${
dir
}
client_log.txt
echo
"--------------pipeline.log:----------------"
cat
PipelineServingLogs/pipeline.log
echo
-e
"-------------------------------------------
\n
"
echo
"-------------------------------------------
\n
"
error_log
$2
else
if
[
"
${
status
}
"
!=
""
]
;
then
error_log
$2
fi
echo
-e
"
${
GREEN_COLOR
}
$2
${
RES
}
\n
"
|
tee
-a
${
log_dir
}
server_total.txt
${
log_dir
}
client_total.txt
fi
fi
else
echo
-e
"
${
RED_COLOR
}
$1
error command
${
RES
}
\n
"
|
tee
-a
${
log_dir
}
server_total.txt
${
log_dir
}
client_total.txt
tail
${
dir
}
client_log.txt |
tee
-a
${
log_dir
}
client_total.txt
echo
"--------------server log:--------------"
cat
${
dir
}
server_log.txt
echo
"--------------client log:--------------"
cat
${
dir
}
client_log.txt
echo
"--------------pipeline.log:----------------"
cat
PipelineServingLogs/pipeline.log
echo
"-------------------------------------------
\n
"
if
[
"
${
status
}
"
!=
""
]
;
then
status
=
"
${
status
}
|Failed"
else
status
=
"Failed"
fi
error_log
$2
fi
status
=
""
}
function
error_log
()
{
...
...
@@ -163,7 +198,7 @@ function error_log() {
echo
"deployment:
${
deployment
// /_
}
"
|
tee
-a
${
log_dir
}
error_models.txt
echo
"py_version:
${
py_version
}
"
|
tee
-a
${
log_dir
}
error_models.txt
echo
"cuda_version:
${
cuda_version
}
"
|
tee
-a
${
log_dir
}
error_models.txt
echo
"status:
Failed
"
|
tee
-a
${
log_dir
}
error_models.txt
echo
"status:
${
status
}
"
|
tee
-a
${
log_dir
}
error_models.txt
echo
-e
"-----------------------------
\n\n
"
|
tee
-a
${
log_dir
}
error_models.txt
prefix
=
${
arg
//\//_
}
for
file
in
${
dir
}*
...
...
@@ -192,7 +227,7 @@ function link_data() {
function
before_hook
()
{
setproxy
cd
${
build_path
}
/python
${
py_version
}
-m
pip
install
--upgrade
pip
${
py_version
}
-m
pip
install
--upgrade
pip
==
21.1.3
${
py_version
}
-m
pip
install
requests
${
py_version
}
-m
pip
install
-r
requirements.txt
${
py_version
}
-m
pip
install
numpy
==
1.16.4
...
...
@@ -379,6 +414,7 @@ function bert_rpc_gpu() {
ls
-hlst
${
py_version
}
-m
paddle_serving_server.serve
--model
bert_seq128_model/
--port
8860
--gpu_ids
0
>
${
dir
}
server_log.txt 2>&1 &
check_result server 15
check_gpu_memory 0
nvidia-smi
head
data-c.txt |
${
py_version
}
bert_client.py
--model
bert_seq128_client/serving_client_conf.prototxt
>
${
dir
}
client_log.txt 2>&1
check_result client
"bert_GPU_RPC server test completed"
...
...
@@ -429,6 +465,7 @@ function ResNet50_rpc() {
sed
-i
's/9696/8863/g'
resnet50_rpc_client.py
${
py_version
}
-m
paddle_serving_server.serve
--model
ResNet50_vd_model
--port
8863
--gpu_ids
0
>
${
dir
}
server_log.txt 2>&1 &
check_result server 8
check_gpu_memory 0
nvidia-smi
${
py_version
}
resnet50_rpc_client.py ResNet50_vd_client_config/serving_client_conf.prototxt
>
${
dir
}
client_log.txt 2>&1
check_result client
"ResNet50_GPU_RPC server test completed"
...
...
@@ -446,6 +483,7 @@ function ResNet101_rpc() {
sed
-i
"22cclient.connect(['127.0.0.1:8864'])"
image_rpc_client.py
${
py_version
}
-m
paddle_serving_server.serve
--model
ResNet101_vd_model
--port
8864
--gpu_ids
0
>
${
dir
}
server_log.txt 2>&1 &
check_result server 8
check_gpu_memory 0
nvidia-smi
${
py_version
}
image_rpc_client.py ResNet101_vd_client_config/serving_client_conf.prototxt
>
${
dir
}
client_log.txt 2>&1
check_result client
"ResNet101_GPU_RPC server test completed"
...
...
@@ -536,10 +574,11 @@ function faster_rcnn_model_rpc() {
data_dir
=
${
data
}
detection/faster_rcnn_r50_fpn_1x_coco/
link_data
${
data_dir
}
sed
-i
's/9494/8870/g'
test_client.py
${
py_version
}
-m
paddle_serving_server.serve
--model
serving_server
--port
8870
--gpu_ids
0
--thread
2
>
${
dir
}
server_log.txt 2>&1 &
${
py_version
}
-m
paddle_serving_server.serve
--model
serving_server
--port
8870
--gpu_ids
1
--thread
2
>
${
dir
}
server_log.txt 2>&1 &
echo
"faster rcnn running ..."
nvidia-smi
check_result server 10
check_gpu_memory 1
${
py_version
}
test_client.py 000000570688.jpg
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
check_result client
"faster_rcnn_GPU_RPC server test completed"
...
...
@@ -556,6 +595,7 @@ function cascade_rcnn_rpc() {
sed
-i
"s/9292/8879/g"
test_client.py
${
py_version
}
-m
paddle_serving_server.serve
--model
serving_server
--port
8879
--gpu_ids
0
--thread
2
>
${
dir
}
server_log.txt 2>&1 &
check_result server 8
check_gpu_memory 0
nvidia-smi
${
py_version
}
test_client.py
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
...
...
@@ -573,6 +613,7 @@ function deeplabv3_rpc() {
sed
-i
"s/9494/8880/g"
deeplabv3_client.py
${
py_version
}
-m
paddle_serving_server.serve
--model
deeplabv3_server
--gpu_ids
0
--port
8880
--thread
2
>
${
dir
}
server_log.txt 2>&1 &
check_result server 10
check_gpu_memory 0
nvidia-smi
${
py_version
}
deeplabv3_client.py
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
...
...
@@ -590,6 +631,7 @@ function mobilenet_rpc() {
sed
-i
"s/9393/8881/g"
mobilenet_tutorial.py
${
py_version
}
-m
paddle_serving_server.serve
--model
mobilenet_v2_imagenet_model
--gpu_ids
0
--port
8881
>
${
dir
}
server_log.txt 2>&1 &
check_result server 8
check_gpu_memory 0
nvidia-smi
${
py_version
}
mobilenet_tutorial.py
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
...
...
@@ -605,8 +647,9 @@ function unet_rpc() {
data_dir
=
${
data
}
unet_for_image_seg/
link_data
${
data_dir
}
sed
-i
"s/9494/8882/g"
seg_client.py
${
py_version
}
-m
paddle_serving_server.serve
--model
unet_model
--gpu_ids
0
--port
8882
>
${
dir
}
server_log.txt 2>&1 &
${
py_version
}
-m
paddle_serving_server.serve
--model
unet_model
--gpu_ids
1
--port
8882
>
${
dir
}
server_log.txt 2>&1 &
check_result server 8
check_gpu_memory 1
nvidia-smi
${
py_version
}
seg_client.py
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
...
...
@@ -624,6 +667,7 @@ function resnetv2_rpc() {
sed
-i
's/9393/8883/g'
resnet50_v2_tutorial.py
${
py_version
}
-m
paddle_serving_server.serve
--model
resnet_v2_50_imagenet_model
--gpu_ids
0
--port
8883
>
${
dir
}
server_log.txt 2>&1 &
check_result server 10
check_gpu_memory 0
nvidia-smi
${
py_version
}
resnet50_v2_tutorial.py
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
...
...
@@ -671,8 +715,9 @@ function criteo_ctr_rpc_gpu() {
data_dir
=
${
data
}
criteo_ctr/
link_data
${
data_dir
}
sed
-i
"s/8885/8886/g"
test_client.py
${
py_version
}
-m
paddle_serving_server.serve
--model
ctr_serving_model/
--port
8886
--gpu_ids
0
>
${
dir
}
server_log.txt 2>&1 &
${
py_version
}
-m
paddle_serving_server.serve
--model
ctr_serving_model/
--port
8886
--gpu_ids
1
>
${
dir
}
server_log.txt 2>&1 &
check_result server 8
check_gpu_memory 1
nvidia-smi
${
py_version
}
test_client.py ctr_client_conf/serving_client_conf.prototxt raw_data/part-0
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
...
...
@@ -691,6 +736,7 @@ function yolov4_rpc_gpu() {
${
py_version
}
-m
paddle_serving_server.serve
--model
yolov4_model
--port
8887
--gpu_ids
0
>
${
dir
}
server_log.txt 2>&1 &
nvidia-smi
check_result server 8
check_gpu_memory 0
${
py_version
}
test_client.py 000000570688.jpg
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
check_result client
"yolov4_GPU_RPC server test completed"
...
...
@@ -708,6 +754,7 @@ function senta_rpc_cpu() {
${
py_version
}
-m
paddle_serving_server.serve
--model
yolov4_model
--port
8887
--gpu_ids
0
>
${
dir
}
server_log.txt 2>&1 &
nvidia-smi
check_result server 8
check_gpu_memory 0
${
py_version
}
test_client.py 000000570688.jpg
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
check_result client
"senta_GPU_RPC server test completed"
...
...
@@ -783,13 +830,14 @@ function ResNet50_http() {
cd
${
build_path
}
/python/examples/imagenet
${
py_version
}
resnet50_web_service.py ResNet50_vd_model gpu 8876
>
${
dir
}
server_log.txt 2>&1 &
check_result server 10
check_gpu_memory 0
curl
-H
"Content-Type:application/json"
-X
POST
-d
'{"feed":[{"image": "https://paddle-serving.bj.bcebos.com/imagenet-example/daisy.jpg"}], "fetch": ["score"]}'
http://127.0.0.1:8876/image/prediction
>
${
dir
}
client_log.txt 2>&1
check_result client
"ResNet50_GPU_HTTP server test completed"
kill_server_process
}
function
bert_http
()
{
dir
=
${
log_dir
}
http_model/
ResNet50
_http/
dir
=
${
log_dir
}
http_model/
bert
_http/
check_dir
${
dir
}
unsetproxy
cd
${
build_path
}
/python/examples/bert
...
...
@@ -836,6 +884,7 @@ function grpc_yolov4() {
echo
-e
"
${
GREEN_COLOR
}
grpc_impl_example_yolov4_GPU_gRPC server started
${
RES
}
"
${
py_version
}
-m
paddle_serving_server.serve
--model
yolov4_model
--port
9393
--gpu_ids
0
--use_multilang
>
${
dir
}
server_log.txt 2>&1 &
check_result server 15
check_gpu_memory 0
echo
-e
"
${
GREEN_COLOR
}
grpc_impl_example_yolov4_GPU_gRPC client started
${
RES
}
"
${
py_version
}
test_client.py 000000570688.jpg
>
${
dir
}
client_log.txt 2>&1
check_result client
"grpc_yolov4_GPU_GRPC server test completed"
...
...
@@ -857,6 +906,7 @@ function ocr_c++_service() {
echo
-e
"
${
GREEN_COLOR
}
OCR_C++_Service_GPU_RPC server started
${
RES
}
"
$py_version
-m
paddle_serving_server.serve
--model
ocr_det_model ocr_rec_model
--port
9293
--gpu_id
0
>
${
dir
}
server_log.txt 2>&1 &
check_result server 8
check_gpu_memory 0
echo
-e
"
${
GREEN_COLOR
}
OCR_C++_Service_GPU_RPC client started
${
RES
}
"
echo
"------------------first:"
$py_version
ocr_cpp_client.py ocr_det_client ocr_rec_client
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录