Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
44b6a5f3
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
44b6a5f3
编写于
7月 17, 2018
作者:
L
Luo Tao
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
fix inference_lib.cmake and make demo_ci pass
上级
af1e54ac
变更
6
显示空白变更内容
内联
并排
Showing
6 changed file
with
22 addition
and
25 deletion
+22
-25
cmake/inference_lib.cmake
cmake/inference_lib.cmake
+14
-17
paddle/fluid/inference/api/demo_ci/CMakeLists.txt
paddle/fluid/inference/api/demo_ci/CMakeLists.txt
+2
-2
paddle/fluid/inference/api/demo_ci/run.sh
paddle/fluid/inference/api/demo_ci/run.sh
+1
-1
paddle/fluid/inference/api/demo_ci/simple_on_word2vec.cc
paddle/fluid/inference/api/demo_ci/simple_on_word2vec.cc
+2
-2
paddle/fluid/inference/api/demo_ci/utils.h
paddle/fluid/inference/api/demo_ci/utils.h
+2
-2
paddle/scripts/paddle_build.sh
paddle/scripts/paddle_build.sh
+1
-1
未找到文件。
cmake/inference_lib.cmake
浏览文件 @
44b6a5f3
...
...
@@ -138,25 +138,22 @@ copy(memory_lib
set
(
inference_deps paddle_fluid_shared paddle_fluid
)
if
(
WITH_CONTRIB
)
message
(
STATUS
"installing contrib"
)
set
(
contrib_dst_dir
"
${
FLUID_INSTALL_DIR
}
/contrib/inference"
)
if
(
WITH_ANAKIN AND WITH_GPU
)
copy
(
contrib_anakin_inference_lib DEPS paddle_inference_api inference_anakin_api
set
(
module
"inference/api"
)
if
(
WITH_ANAKIN AND WITH_GPU
)
copy
(
anakin_inference_lib DEPS paddle_inference_api inference_anakin_api
SRCS
${
PADDLE_BINARY_DIR
}
/paddle/contrib/inference
/libinference_anakin_api*
# compiled anakin api
${
PADDLE_BINARY_DIR
}
/paddle/fluid/inference/api
/libinference_anakin_api*
# compiled anakin api
${
PADDLE_BINARY_DIR
}
/third_party/install/anakin/*.tar.gz
# anakin release
DSTS
${
contrib_dst_dir
}
/anakin
${
contrib_dst_dir
}
/anakin
)
list
(
APPEND inference_deps contrib_anakin_inference_lib
)
endif
()
copy
(
contrib_inference_lib DEPS paddle_inference_api paddle_inference_api_shared
SRCS
${
PADDLE_SOURCE_DIR
}
/paddle/contrib/inference/paddle_inference_api.h
${
PADDLE_BINARY_DIR
}
/paddle/contrib/inference/libpaddle_inference_api*
DSTS
${
contrib_dst_dir
}
${
contrib_dst_dir
}
)
list
(
APPEND inference_deps contrib_inference_lib
)
DSTS
${
dst_dir
}
/inference/anakin
${
dst_dir
}
/inference/anakin
)
list
(
APPEND inference_deps anakin_inference_lib
)
endif
()
copy
(
inference_api_lib DEPS paddle_inference_api paddle_inference_api_shared
SRCS
${
src_dir
}
/
${
module
}
/api.h
${
PADDLE_BINARY_DIR
}
/paddle/fluid/inference/api/libpaddle_inference_api*
DSTS
${
dst_dir
}
/inference
${
dst_dir
}
/inference
)
list
(
APPEND inference_deps inference_api_lib
)
set
(
module
"inference"
)
copy
(
inference_lib DEPS
${
inference_deps
}
SRCS
${
src_dir
}
/
${
module
}
/*.h
${
PADDLE_BINARY_DIR
}
/paddle/fluid/inference/libpaddle_fluid.*
...
...
paddle/fluid/inference/api/demo_ci/CMakeLists.txt
浏览文件 @
44b6a5f3
...
...
@@ -55,11 +55,11 @@ endif()
# Note: libpaddle_inference_api.so/a must put before libpaddle_fluid.so/a
if
(
WITH_STATIC_LIB
)
set
(
DEPS
${
PADDLE_LIB
}
/
contrib
/inference/libpaddle_inference_api.a
${
PADDLE_LIB
}
/
paddle/fluid
/inference/libpaddle_inference_api.a
${
PADDLE_LIB
}
/paddle/fluid/inference/libpaddle_fluid.a
)
else
()
set
(
DEPS
${
PADDLE_LIB
}
/
contrib
/inference/libpaddle_inference_api.so
${
PADDLE_LIB
}
/
paddle/fluid
/inference/libpaddle_inference_api.so
${
PADDLE_LIB
}
/paddle/fluid/inference/libpaddle_fluid.so
)
endif
()
set
(
EXTERNAL_LIB
"-lrt -ldl -lpthread"
)
...
...
paddle/fluid/inference/api/demo_ci/run.sh
浏览文件 @
44b6a5f3
...
...
@@ -64,7 +64,7 @@ for WITH_STATIC_LIB in ON OFF; do
-DWITH_GPU
=
$TEST_GPU_CPU
\
-DWITH_STATIC_LIB
=
$WITH_STATIC_LIB
make
-j
for
use_gpu
in
false
;
do
for
use_gpu
in
$use_gpu_list
;
do
for
vis_demo_name
in
$vis_demo_list
;
do
./vis_demo
\
--modeldir
=
../data/
$vis_demo_name
/model
\
...
...
paddle/fluid/inference/api/demo_ci/simple_on_word2vec.cc
浏览文件 @
44b6a5f3
...
...
@@ -19,8 +19,8 @@ limitations under the License. */
#include <gflags/gflags.h>
#include <glog/logging.h>
#include <memory>
#include <thread>
#include "
contrib/inference/paddle_inference_
api.h"
#include <thread>
//NOLINT
#include "
paddle/fluid/inference/
api.h"
#include "paddle/fluid/platform/enforce.h"
DEFINE_string
(
dirname
,
""
,
"Directory of the inference model."
);
...
...
paddle/fluid/inference/api/demo_ci/utils.h
浏览文件 @
44b6a5f3
...
...
@@ -13,10 +13,10 @@
// limitations under the License.
#pragma once
#include <algorithm>
#include <string>
#include <vector>
#include "contrib/inference/paddle_inference_api.h"
#include "paddle/fluid/inference/api.h"
namespace
paddle
{
namespace
demo
{
...
...
paddle/scripts/paddle_build.sh
浏览文件 @
44b6a5f3
...
...
@@ -531,7 +531,7 @@ function test_fluid_inference_lib() {
Testing fluid inference library ...
========================================
EOF
cd
${
PADDLE_ROOT
}
/paddle/
contrib/inference
/demo_ci
cd
${
PADDLE_ROOT
}
/paddle/
fluid/inference/api
/demo_ci
./run.sh
${
PADDLE_ROOT
}
${
WITH_MKL
:-
ON
}
${
WITH_GPU
:-
OFF
}
fi
}
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录