未验证 提交 ba3ebe20 编写于 作者: S shangliang Xu 提交者: GitHub

[TIPC] fix cpp_infer, not use trt/mkldnn (#6102)

上级 bdf86c58
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -18,7 +18,7 @@ infer_mode:norm ...@@ -18,7 +18,7 @@ infer_mode:norm
infer_quant:False infer_quant:False
inference:./deploy/cpp/build/main inference:./deploy/cpp/build/main
--device:gpu|cpu --device:gpu|cpu
--use_mkldnn:True|False --use_mkldnn:False
--cpu_threads:4 --cpu_threads:4
--batch_size:1|2 --batch_size:1|2
--use_tensorrt:null --use_tensorrt:null
......
...@@ -131,7 +131,9 @@ fi ...@@ -131,7 +131,9 @@ fi
# build program # build program
# TODO: set PADDLE_DIR and TENSORRT_ROOT # TODO: set PADDLE_DIR and TENSORRT_ROOT
if [ -z $PADDLE_DIR ]; then if [ -z $PADDLE_DIR ]; then
PADDLE_DIR=/paddle/Paddle/build/paddle_inference_install_dir/ wget -nc https://paddle-inference-lib.bj.bcebos.com/2.2.2/cxx_c/Linux/GPU/x86-64_gcc8.2_avx_mkl_cuda11.1_cudnn8.1.1_trt7.2.3.4/paddle_inference.tgz --no-check-certificate
tar zxf paddle_inference.tgz
PADDLE_DIR=$(pwd)/paddle_inference
fi fi
if [ -z $TENSORRT_ROOT ]; then if [ -z $TENSORRT_ROOT ]; then
TENSORRT_ROOT=/usr/local/TensorRT6-cuda10.1-cudnn7 TENSORRT_ROOT=/usr/local/TensorRT6-cuda10.1-cudnn7
...@@ -146,8 +148,8 @@ mkdir -p build ...@@ -146,8 +148,8 @@ mkdir -p build
cd ./build cd ./build
cmake .. \ cmake .. \
-DWITH_GPU=ON \ -DWITH_GPU=ON \
-DWITH_MKL=ON \ -DWITH_MKL=OFF \
-DWITH_TENSORRT=ON \ -DWITH_TENSORRT=OFF \
-DPADDLE_LIB_NAME=libpaddle_inference \ -DPADDLE_LIB_NAME=libpaddle_inference \
-DPADDLE_DIR=${PADDLE_DIR} \ -DPADDLE_DIR=${PADDLE_DIR} \
-DCUDA_LIB=${CUDA_LIB} \ -DCUDA_LIB=${CUDA_LIB} \
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册