提交 65d2179a 编写于 作者: S superjomn

disable lite by default

test=develop
上级 0a20a618
...@@ -80,7 +80,7 @@ option(WITH_FAST_MATH "Make use of fast math library, might affect the precisi ...@@ -80,7 +80,7 @@ option(WITH_FAST_MATH "Make use of fast math library, might affect the precisi
option(WITH_DGC "Use DGC(Deep Gradient Compression) or not" ON) option(WITH_DGC "Use DGC(Deep Gradient Compression) or not" ON)
# for lite, both server and mobile framework. # for lite, both server and mobile framework.
option(WITH_LITE "Enable lite framework" ON) option(WITH_LITE "Enable lite framework" OFF)
option(LITE_WITH_CUDA "Enable CUDA in lite mode" OFF) option(LITE_WITH_CUDA "Enable CUDA in lite mode" OFF)
option(LITE_WITH_X86 "Enable X86 in lite mode" ON) option(LITE_WITH_X86 "Enable X86 in lite mode" ON)
option(LITE_WITH_LIGHT_WEIGHT_FRAMEWORK "Enable light-weight framework" OFF) option(LITE_WITH_LIGHT_WEIGHT_FRAMEWORK "Enable light-weight framework" OFF)
......
...@@ -26,7 +26,7 @@ TEST(LightAPI, load) { ...@@ -26,7 +26,7 @@ TEST(LightAPI, load) {
predictor.Build(FLAGS_optimized_model); predictor.Build(FLAGS_optimized_model);
auto* input_tensor = predictor.GetInput(0); auto* input_tensor = predictor.GetInput(0);
input_tensor->Resize(DDimLite(std::vector<int64_t>({100, 100}))); input_tensor->Resize(DDim(std::vector<int64_t>({100, 100})));
auto* data = input_tensor->mutable_data<float>(); auto* data = input_tensor->mutable_data<float>();
for (int i = 0; i < 100 * 100; i++) { for (int i = 0; i < 100 * 100; i++) {
data[i] = i; data[i] = i;
...@@ -39,13 +39,13 @@ TEST(LightAPI, load) { ...@@ -39,13 +39,13 @@ TEST(LightAPI, load) {
} // namespace paddle } // namespace paddle
USE_LITE_OP(mul); USE_LITE_OP(mul);
// USE_LITE_OP(fc); USE_LITE_OP(fc);
// USE_LITE_OP(scale); USE_LITE_OP(scale);
USE_LITE_OP(feed); USE_LITE_OP(feed);
USE_LITE_OP(fetch); USE_LITE_OP(fetch);
USE_LITE_OP(io_copy); USE_LITE_OP(io_copy);
// USE_LITE_KERNEL(fc, kHost, kFloat, kNCHW, def); USE_LITE_KERNEL(fc, kHost, kFloat, kNCHW, def);
// USE_LITE_KERNEL(mul, kHost, kFloat, kNCHW, def); USE_LITE_KERNEL(mul, kHost, kFloat, kNCHW, def);
// USE_LITE_KERNEL(scale, kHost, kFloat, kNCHW, def); USE_LITE_KERNEL(scale, kHost, kFloat, kNCHW, def);
USE_LITE_KERNEL(feed, kHost, kAny, kAny, def); USE_LITE_KERNEL(feed, kHost, kAny, kAny, def);
USE_LITE_KERNEL(fetch, kHost, kAny, kAny, def); USE_LITE_KERNEL(fetch, kHost, kAny, kAny, def);
cc_library(lite_gtest_main SRCS lite_gtest_main.cc) cc_library(lite_gtest_main SRCS lite_gtest_main.cc DEPS gtest)
cc_library(memory_lite SRCS memory.cc DEPS target_wrapper_lite target_wrapper_host) cc_library(memory_lite SRCS memory.cc DEPS target_wrapper_lite target_wrapper_host)
cc_library(target_wrapper_lite SRCS target_wrapper.cc) cc_library(target_wrapper_lite SRCS target_wrapper.cc)
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册