Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
Serving
提交
58415334
S
Serving
项目概览
PaddlePaddle
/
Serving
1 年多 前同步成功
通知
186
Star
833
Fork
253
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
105
列表
看板
标记
里程碑
合并请求
10
Wiki
2
Wiki
分析
仓库
DevOps
项目成员
Pages
S
Serving
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
105
Issue
105
列表
看板
标记
里程碑
合并请求
10
合并请求
10
Pages
分析
分析
仓库分析
DevOps
Wiki
2
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
58415334
编写于
7月 12, 2021
作者:
Z
ZhangYulongg
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
update CI scripts and fix example
上级
30788724
变更
2
隐藏空白更改
内联
并排
Showing
2 changed file
with
51 addition
and
4 deletion
+51
-4
python/examples/bert/bert_web_service_gpu.py
python/examples/bert/bert_web_service_gpu.py
+1
-0
tools/scripts/ipipe_py3.sh
tools/scripts/ipipe_py3.sh
+50
-4
未找到文件。
python/examples/bert/bert_web_service_gpu.py
浏览文件 @
58415334
...
...
@@ -42,6 +42,7 @@ class BertService(WebService):
bert_service
=
BertService
(
name
=
"bert"
)
bert_service
.
load
()
bert_service
.
load_model_config
(
sys
.
argv
[
1
])
bert_service
.
set_gpus
(
"0"
)
bert_service
.
prepare_server
(
workdir
=
"workdir"
,
port
=
int
(
sys
.
argv
[
2
]),
device
=
"gpu"
)
bert_service
.
run_rpc_service
()
...
...
tools/scripts/ipipe_py3.sh
浏览文件 @
58415334
...
...
@@ -120,31 +120,62 @@ function check() {
fi
}
function
check_gpu_memory
()
{
gpu_memory
=
`
nvidia-smi
--id
=
$1
--format
=
csv,noheader
--query-gpu
=
memory.used |
awk
'{print $1}'
`
echo
-e
"
${
GREEN_COLOR
}
-------id-
$1
gpu_memory_used:
${
gpu_memory
}${
RES
}
"
if
[
${
gpu_memory
}
-le
100
]
;
then
echo
"-------GPU-
$1
is not used"
status
=
"GPU-
$1
is not used"
else
echo
"-------GPU_memory used is expected"
fi
}
function
check_result
()
{
if
[
$?
==
0
]
;
then
echo
-e
"
${
GREEN_COLOR
}
$1
execute normally
${
RES
}
"
if
[
$1
==
"server"
]
;
then
sleep
$2
tail
${
dir
}
server_log.txt |
tee
-a
${
log_dir
}
server_total.txt
cat
${
dir
}
server_log.txt |
tee
-a
${
log_dir
}
server_total.txt
fi
if
[
$1
==
"client"
]
;
then
tail
${
dir
}
client_log.txt |
tee
-a
${
log_dir
}
client_total.txt
grep
-E
"
${
error_words
}
"
${
dir
}
client_log.txt
>
/dev/null
if
[
$?
==
0
]
;
then
if
[
"
${
status
}
"
!=
""
]
;
then
status
=
"
${
status
}
|Failed"
else
status
=
"Failed"
fi
echo
-e
"
${
RED_COLOR
}
$1
error command
${
RES
}
\n
"
|
tee
-a
${
log_dir
}
server_total.txt
${
log_dir
}
client_total.txt
echo
"----------server log:"
cat
${
dir
}
server_log.txt
echo
"----------client log:"
cat
${
dir
}
client_log.txt |
tee
-a
${
log_dir
}
client_total.txt
echo
-e
"--------------pipeline.log:----------------
\n
"
cat
PipelineServingLogs/pipeline.log
echo
-e
"-------------------------------------------
\n
"
error_log
$2
else
if
[
"
${
status
}
"
!=
""
]
;
then
error_log
$2
fi
echo
-e
"
${
GREEN_COLOR
}
$2
${
RES
}
\n
"
|
tee
-a
${
log_dir
}
server_total.txt
${
log_dir
}
client_total.txt
fi
fi
else
echo
-e
"
${
RED_COLOR
}
$1
error command
${
RES
}
\n
"
|
tee
-a
${
log_dir
}
server_total.txt
${
log_dir
}
client_total.txt
tail
${
dir
}
client_log.txt |
tee
-a
${
log_dir
}
client_total.txt
echo
"----------server log:"
cat
${
dir
}
server_log.txt
echo
"----------client log:"
cat
${
dir
}
client_log.txt |
tee
-a
${
log_dir
}
client_total.txt
if
[
"
${
status
}
"
!=
""
]
;
then
status
=
"
${
status
}
|Failed"
else
status
=
"Failed"
fi
error_log
$2
fi
status
=
""
}
function
error_log
()
{
...
...
@@ -163,7 +194,7 @@ function error_log() {
echo
"deployment:
${
deployment
// /_
}
"
|
tee
-a
${
log_dir
}
error_models.txt
echo
"py_version:
${
py_version
}
"
|
tee
-a
${
log_dir
}
error_models.txt
echo
"cuda_version:
${
cuda_version
}
"
|
tee
-a
${
log_dir
}
error_models.txt
echo
"status:
Failed
"
|
tee
-a
${
log_dir
}
error_models.txt
echo
"status:
${
status
}
"
|
tee
-a
${
log_dir
}
error_models.txt
echo
-e
"-----------------------------
\n\n
"
|
tee
-a
${
log_dir
}
error_models.txt
prefix
=
${
arg
//\//_
}
for
file
in
${
dir
}*
...
...
@@ -379,6 +410,7 @@ function bert_rpc_gpu() {
ls
-hlst
${
py_version
}
-m
paddle_serving_server.serve
--model
bert_seq128_model/
--port
8860
--gpu_ids
0
>
${
dir
}
server_log.txt 2>&1 &
check_result server 15
check_gpu_memory 0
nvidia-smi
head
data-c.txt |
${
py_version
}
bert_client.py
--model
bert_seq128_client/serving_client_conf.prototxt
>
${
dir
}
client_log.txt 2>&1
check_result client
"bert_GPU_RPC server test completed"
...
...
@@ -429,6 +461,7 @@ function ResNet50_rpc() {
sed
-i
's/9696/8863/g'
resnet50_rpc_client.py
${
py_version
}
-m
paddle_serving_server.serve
--model
ResNet50_vd_model
--port
8863
--gpu_ids
0
>
${
dir
}
server_log.txt 2>&1 &
check_result server 8
check_gpu_memory 0
nvidia-smi
${
py_version
}
resnet50_rpc_client.py ResNet50_vd_client_config/serving_client_conf.prototxt
>
${
dir
}
client_log.txt 2>&1
check_result client
"ResNet50_GPU_RPC server test completed"
...
...
@@ -446,6 +479,7 @@ function ResNet101_rpc() {
sed
-i
"22cclient.connect(['127.0.0.1:8864'])"
image_rpc_client.py
${
py_version
}
-m
paddle_serving_server.serve
--model
ResNet101_vd_model
--port
8864
--gpu_ids
0
>
${
dir
}
server_log.txt 2>&1 &
check_result server 8
check_gpu_memory 0
nvidia-smi
${
py_version
}
image_rpc_client.py ResNet101_vd_client_config/serving_client_conf.prototxt
>
${
dir
}
client_log.txt 2>&1
check_result client
"ResNet101_GPU_RPC server test completed"
...
...
@@ -540,6 +574,7 @@ function faster_rcnn_model_rpc() {
echo
"faster rcnn running ..."
nvidia-smi
check_result server 10
check_gpu_memory 0
${
py_version
}
test_client.py 000000570688.jpg
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
check_result client
"faster_rcnn_GPU_RPC server test completed"
...
...
@@ -556,6 +591,7 @@ function cascade_rcnn_rpc() {
sed
-i
"s/9292/8879/g"
test_client.py
${
py_version
}
-m
paddle_serving_server.serve
--model
serving_server
--port
8879
--gpu_ids
0
--thread
2
>
${
dir
}
server_log.txt 2>&1 &
check_result server 8
check_gpu_memory 0
nvidia-smi
${
py_version
}
test_client.py
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
...
...
@@ -573,6 +609,7 @@ function deeplabv3_rpc() {
sed
-i
"s/9494/8880/g"
deeplabv3_client.py
${
py_version
}
-m
paddle_serving_server.serve
--model
deeplabv3_server
--gpu_ids
0
--port
8880
--thread
2
>
${
dir
}
server_log.txt 2>&1 &
check_result server 10
check_gpu_memory 0
nvidia-smi
${
py_version
}
deeplabv3_client.py
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
...
...
@@ -590,6 +627,7 @@ function mobilenet_rpc() {
sed
-i
"s/9393/8881/g"
mobilenet_tutorial.py
${
py_version
}
-m
paddle_serving_server.serve
--model
mobilenet_v2_imagenet_model
--gpu_ids
0
--port
8881
>
${
dir
}
server_log.txt 2>&1 &
check_result server 8
check_gpu_memory 0
nvidia-smi
${
py_version
}
mobilenet_tutorial.py
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
...
...
@@ -607,6 +645,7 @@ function unet_rpc() {
sed
-i
"s/9494/8882/g"
seg_client.py
${
py_version
}
-m
paddle_serving_server.serve
--model
unet_model
--gpu_ids
0
--port
8882
>
${
dir
}
server_log.txt 2>&1 &
check_result server 8
check_gpu_memory 0
nvidia-smi
${
py_version
}
seg_client.py
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
...
...
@@ -624,6 +663,7 @@ function resnetv2_rpc() {
sed
-i
's/9393/8883/g'
resnet50_v2_tutorial.py
${
py_version
}
-m
paddle_serving_server.serve
--model
resnet_v2_50_imagenet_model
--gpu_ids
0
--port
8883
>
${
dir
}
server_log.txt 2>&1 &
check_result server 10
check_gpu_memory 0
nvidia-smi
${
py_version
}
resnet50_v2_tutorial.py
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
...
...
@@ -673,6 +713,7 @@ function criteo_ctr_rpc_gpu() {
sed
-i
"s/8885/8886/g"
test_client.py
${
py_version
}
-m
paddle_serving_server.serve
--model
ctr_serving_model/
--port
8886
--gpu_ids
0
>
${
dir
}
server_log.txt 2>&1 &
check_result server 8
check_gpu_memory 0
nvidia-smi
${
py_version
}
test_client.py ctr_client_conf/serving_client_conf.prototxt raw_data/part-0
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
...
...
@@ -691,6 +732,7 @@ function yolov4_rpc_gpu() {
${
py_version
}
-m
paddle_serving_server.serve
--model
yolov4_model
--port
8887
--gpu_ids
0
>
${
dir
}
server_log.txt 2>&1 &
nvidia-smi
check_result server 8
check_gpu_memory 0
${
py_version
}
test_client.py 000000570688.jpg
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
check_result client
"yolov4_GPU_RPC server test completed"
...
...
@@ -708,6 +750,7 @@ function senta_rpc_cpu() {
${
py_version
}
-m
paddle_serving_server.serve
--model
yolov4_model
--port
8887
--gpu_ids
0
>
${
dir
}
server_log.txt 2>&1 &
nvidia-smi
check_result server 8
check_gpu_memory 0
${
py_version
}
test_client.py 000000570688.jpg
>
${
dir
}
client_log.txt 2>&1
nvidia-smi
check_result client
"senta_GPU_RPC server test completed"
...
...
@@ -783,6 +826,7 @@ function ResNet50_http() {
cd
${
build_path
}
/python/examples/imagenet
${
py_version
}
resnet50_web_service.py ResNet50_vd_model gpu 8876
>
${
dir
}
server_log.txt 2>&1 &
check_result server 10
check_gpu_memory 0
curl
-H
"Content-Type:application/json"
-X
POST
-d
'{"feed":[{"image": "https://paddle-serving.bj.bcebos.com/imagenet-example/daisy.jpg"}], "fetch": ["score"]}'
http://127.0.0.1:8876/image/prediction
>
${
dir
}
client_log.txt 2>&1
check_result client
"ResNet50_GPU_HTTP server test completed"
kill_server_process
...
...
@@ -836,6 +880,7 @@ function grpc_yolov4() {
echo
-e
"
${
GREEN_COLOR
}
grpc_impl_example_yolov4_GPU_gRPC server started
${
RES
}
"
${
py_version
}
-m
paddle_serving_server.serve
--model
yolov4_model
--port
9393
--gpu_ids
0
--use_multilang
>
${
dir
}
server_log.txt 2>&1 &
check_result server 15
check_gpu_memory 0
echo
-e
"
${
GREEN_COLOR
}
grpc_impl_example_yolov4_GPU_gRPC client started
${
RES
}
"
${
py_version
}
test_client.py 000000570688.jpg
>
${
dir
}
client_log.txt 2>&1
check_result client
"grpc_yolov4_GPU_GRPC server test completed"
...
...
@@ -857,6 +902,7 @@ function ocr_c++_service() {
echo
-e
"
${
GREEN_COLOR
}
OCR_C++_Service_GPU_RPC server started
${
RES
}
"
$py_version
-m
paddle_serving_server.serve
--model
ocr_det_model ocr_rec_model
--port
9293
--gpu_id
0
>
${
dir
}
server_log.txt 2>&1 &
check_result server 8
check_gpu_memory 0
echo
-e
"
${
GREEN_COLOR
}
OCR_C++_Service_GPU_RPC client started
${
RES
}
"
echo
"------------------first:"
$py_version
ocr_cpp_client.py ocr_det_client ocr_rec_client
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录