Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Crayon鑫
Paddle
提交
0a20a618
P
Paddle
项目概览
Crayon鑫
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
0a20a618
编写于
5月 13, 2019
作者:
S
superjomn
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
fix ARM compile error
test=develop
上级
9174652b
变更
28
隐藏空白更改
内联
并排
Showing
28 changed file
with
156 addition
and
111 deletion
+156
-111
CMakeLists.txt
CMakeLists.txt
+3
-2
paddle/fluid/framework/CMakeLists.txt
paddle/fluid/framework/CMakeLists.txt
+1
-1
paddle/fluid/lite/CMakeLists.txt
paddle/fluid/lite/CMakeLists.txt
+24
-1
paddle/fluid/lite/api/CMakeLists.txt
paddle/fluid/lite/api/CMakeLists.txt
+17
-3
paddle/fluid/lite/api/cxx_api_test.cc
paddle/fluid/lite/api/cxx_api_test.cc
+4
-3
paddle/fluid/lite/api/light_api.h
paddle/fluid/lite/api/light_api.h
+4
-5
paddle/fluid/lite/api/light_api_test.cc
paddle/fluid/lite/api/light_api_test.cc
+10
-11
paddle/fluid/lite/core/CMakeLists.txt
paddle/fluid/lite/core/CMakeLists.txt
+3
-3
paddle/fluid/lite/core/memory.h
paddle/fluid/lite/core/memory.h
+2
-0
paddle/fluid/lite/core/mir/CMakeLists.txt
paddle/fluid/lite/core/mir/CMakeLists.txt
+5
-2
paddle/fluid/lite/core/mir/ssa_graph_test.cc
paddle/fluid/lite/core/mir/ssa_graph_test.cc
+4
-2
paddle/fluid/lite/core/mir/variable_place_inference_pass.h
paddle/fluid/lite/core/mir/variable_place_inference_pass.h
+0
-6
paddle/fluid/lite/core/mir/variable_place_inference_pass_test.cc
...fluid/lite/core/mir/variable_place_inference_pass_test.cc
+1
-1
paddle/fluid/lite/core/op_registry.h
paddle/fluid/lite/core/op_registry.h
+2
-1
paddle/fluid/lite/core/optimizer_test.cc
paddle/fluid/lite/core/optimizer_test.cc
+7
-1
paddle/fluid/lite/core/program.h
paddle/fluid/lite/core/program.h
+1
-1
paddle/fluid/lite/core/program_fake_utils.h
paddle/fluid/lite/core/program_fake_utils.h
+11
-10
paddle/fluid/lite/kernels/host/CMakeLists.txt
paddle/fluid/lite/kernels/host/CMakeLists.txt
+11
-12
paddle/fluid/lite/kernels/host/fc_compute_test.cc
paddle/fluid/lite/kernels/host/fc_compute_test.cc
+11
-11
paddle/fluid/lite/model_parser/CMakeLists.txt
paddle/fluid/lite/model_parser/CMakeLists.txt
+10
-2
paddle/fluid/lite/model_parser/compatible_pb.h
paddle/fluid/lite/model_parser/compatible_pb.h
+1
-1
paddle/fluid/lite/model_parser/model_parser.h
paddle/fluid/lite/model_parser/model_parser.h
+1
-1
paddle/fluid/lite/model_parser/model_parser_test.cc
paddle/fluid/lite/model_parser/model_parser_test.cc
+8
-5
paddle/fluid/lite/model_parser/pb/op_desc.h
paddle/fluid/lite/model_parser/pb/op_desc.h
+1
-1
paddle/fluid/lite/model_parser/pb/var_desc.h
paddle/fluid/lite/model_parser/pb/var_desc.h
+1
-1
paddle/fluid/lite/operators/fc_op.h
paddle/fluid/lite/operators/fc_op.h
+0
-3
paddle/fluid/lite/operators/fc_op_test.cc
paddle/fluid/lite/operators/fc_op_test.cc
+12
-18
paddle/fluid/lite/utils/CMakeLists.txt
paddle/fluid/lite/utils/CMakeLists.txt
+1
-3
未找到文件。
CMakeLists.txt
浏览文件 @
0a20a618
...
...
@@ -80,9 +80,10 @@ option(WITH_FAST_MATH "Make use of fast math library, might affect the precisi
option
(
WITH_DGC
"Use DGC(Deep Gradient Compression) or not"
ON
)
# for lite, both server and mobile framework.
option
(
LITE_WITH_CUDA
"Enable CUDA in lite mode"
ON
)
option
(
WITH_LITE
"Enable lite framework"
ON
)
option
(
LITE_WITH_CUDA
"Enable CUDA in lite mode"
OFF
)
option
(
LITE_WITH_X86
"Enable X86 in lite mode"
ON
)
option
(
LITE_WITH_LIGHT_WEIGHT_FRAMEWORK
"Enable light-weight framework"
O
N
)
option
(
LITE_WITH_LIGHT_WEIGHT_FRAMEWORK
"Enable light-weight framework"
O
FF
)
set
(
THIRD_PARTY_PATH
"
${
CMAKE_BINARY_DIR
}
/third_party"
CACHE STRING
...
...
paddle/fluid/framework/CMakeLists.txt
浏览文件 @
0a20a618
...
...
@@ -124,7 +124,7 @@ cc_library(shape_inference SRCS shape_inference.cc DEPS ddim attribute device_co
cc_library
(
transfer_scope_cache SRCS transfer_scope_cache.cc DEPS scope framework_proto device_context
)
cc_library
(
op_kernel_type SRCS op_kernel_type.cc DEPS device_context place
)
cc_library
(
operator SRCS operator.cc DEPS op_info device_context tensor scope glog
shape_inference data_transform lod_tensor profiler transfer_scope_cache op_kernel_type
)
shape_inference data_transform lod_tensor profiler transfer_scope_cache op_kernel_type
data_feed_proto
)
cc_test
(
operator_test SRCS operator_test.cc DEPS operator op_registry device_context
)
...
...
paddle/fluid/lite/CMakeLists.txt
浏览文件 @
0a20a618
...
...
@@ -2,11 +2,34 @@ if (NOT WITH_LITE)
return
()
endif
()
message
(
WARNING
"
Enable Lite
"
)
message
(
WARNING
"
Lite enabled!
"
)
message
(
STATUS
"LIGHT_FRAMEWORK:
${
LITE_WITH_LIGHT_WEIGHT_FRAMEWORK
}
"
)
message
(
STATUS
"LITE_WITH_CUDA:
${
LITE_WITH_CUDA
}
"
)
message
(
STATUS
"LITE_WITH_X86:
${
LITE_WITH_X86
}
"
)
set
(
LITE_MODEL_DIR
"
${
THIRD_PARTY_PATH
}
/install"
)
function
(
lite_download_and_uncompress INSTALL_DIR URL FILENAME
)
message
(
STATUS
"Download inference test stuff from
${
URL
}
/
${
FILENAME
}
"
)
string
(
REGEX REPLACE
"[-%.]"
"_"
FILENAME_EX
${
FILENAME
}
)
set
(
EXTERNAL_PROJECT_NAME
"extern_lite_download_
${
FILENAME_EX
}
"
)
set
(
UNPACK_DIR
"
${
INSTALL_DIR
}
/src/
${
EXTERNAL_PROJECT_NAME
}
"
)
ExternalProject_Add
(
${
EXTERNAL_PROJECT_NAME
}
${
EXTERNAL_PROJECT_LOG_ARGS
}
PREFIX
${
INSTALL_DIR
}
DOWNLOAD_COMMAND wget --no-check-certificate -q -O
${
INSTALL_DIR
}
/
${
FILENAME
}
${
URL
}
/
${
FILENAME
}
&&
${
CMAKE_COMMAND
}
-E tar xzf
${
INSTALL_DIR
}
/
${
FILENAME
}
DOWNLOAD_DIR
${
INSTALL_DIR
}
DOWNLOAD_NO_PROGRESS 1
CONFIGURE_COMMAND
""
BUILD_COMMAND
""
UPDATE_COMMAND
""
INSTALL_COMMAND
""
)
endfunction
()
add_subdirectory
(
core
)
add_subdirectory
(
x86
)
add_subdirectory
(
host
)
...
...
paddle/fluid/lite/api/CMakeLists.txt
浏览文件 @
0a20a618
...
...
@@ -18,8 +18,22 @@ cc_library(light_api_lite SRCS light_api.cc DEPS ${light_api_deps} ${ops_lite} $
message
(
STATUS
"get ops
${
ops_lite
}
"
)
message
(
STATUS
"get kernels
${
host_kernels
}
"
)
lite_cc_test
(
test_cxx_api_lite SRCS cxx_api_test.cc DEPS cxx_api_lite model_parser_lite target_wrapper_host
${
ops_lite
}
${
host_kernels
}
)
lite_cc_test
(
test_light_api SRCS light_api_test.cc DEPS light_api_lite
)
include
(
ExternalProject
)
set
(
LITE_URL
"http://paddle-inference-dist.bj.bcebos.com"
CACHE STRING
"inference download url"
)
set
(
LITE_DEMO_INSTALL_DIR
"
${
THIRD_PARTY_PATH
}
/inference_demo"
CACHE STRING
"A path setting inference demo download directories."
)
lite_cc_test
(
test_cxx_api_lite SRCS cxx_api_test.cc
DEPS cxx_api_lite model_parser_lite target_wrapper_host
${
ops_lite
}
${
host_kernels
}
ARGS --model_dir=
${
LITE_MODEL_DIR
}
/lite_naive_model
--optimized_model=
${
LITE_MODEL_DIR
}
/lite_naive_model_opt SERIAL
)
if
(
WITH_TESTING
)
lite_download_and_uncompress
(
${
LITE_MODEL_DIR
}
${
LITE_URL
}
"lite_naive_model.tar.gz"
)
add_dependencies
(
test_cxx_api_lite extern_lite_download_lite_naive_model_tar_gz
)
endif
(
WITH_TESTING
)
lite_cc_test
(
test_light_api SRCS light_api_test.cc DEPS light_api_lite ARGS --optimized_model=
${
LITE_MODEL_DIR
}
/lite_naive_model_opt SERIAL
)
cc_binary
(
cxx_api_lite_bin SRCS cxx_api_bin.cc DEPS cxx_api_lite model_parser_lite target_wrapper_host
${
ops_lite
}
${
host_kernels
}
)
paddle/fluid/lite/api/cxx_api_test.cc
浏览文件 @
0a20a618
...
...
@@ -19,6 +19,7 @@
#include "paddle/fluid/lite/core/op_registry.h"
DEFINE_string
(
model_dir
,
""
,
""
);
DEFINE_string
(
optimized_model
,
""
,
""
);
namespace
paddle
{
namespace
lite
{
...
...
@@ -48,7 +49,7 @@ TEST(CXXApi, test) {
data
[
i
]
=
i
;
}
LOG
(
INFO
)
<<
"input "
<<
*
input_tensor
;
//
LOG(INFO) << "input " << *input_tensor;
predictor
.
Run
();
...
...
@@ -57,7 +58,7 @@ TEST(CXXApi, test) {
LOG
(
INFO
)
<<
"out "
<<
out
->
data
<
float
>
()[
0
];
LOG
(
INFO
)
<<
"out "
<<
out
->
data
<
float
>
()[
1
];
LOG
(
INFO
)
<<
"dims "
<<
out
->
dims
();
LOG
(
INFO
)
<<
"out "
<<
*
out
;
//
LOG(INFO) << "out " << *out;
}
#ifndef LITE_WITH_LIGHT_WEIGHT_FRAMEWORK
...
...
@@ -67,7 +68,7 @@ TEST(CXXApi, save_model) {
predictor
.
Build
(
FLAGS_model_dir
,
Place
{
TARGET
(
kCUDA
),
PRECISION
(
kFloat
)},
valid_places
);
predictor
.
SaveModel
(
"./optimized_model"
);
predictor
.
SaveModel
(
FLAGS_optimized_model
);
}
#endif
...
...
paddle/fluid/lite/api/light_api.h
浏览文件 @
0a20a618
...
...
@@ -41,21 +41,20 @@ class LightPredictor {
void
Run
()
{
program_
->
Run
();
}
// Get offset-th col of feed.
Tensor
Base
*
GetInput
(
size_t
offset
)
{
Tensor
*
GetInput
(
size_t
offset
)
{
auto
*
_feed_list
=
program_
->
exec_scope
()
->
FindVar
(
"feed"
);
CHECK
(
_feed_list
)
<<
"no feed variable in exec_scope"
;
auto
*
feed_list
=
_feed_list
->
GetMutable
<
std
::
vector
<
Tensor
Base
>>
();
auto
*
feed_list
=
_feed_list
->
GetMutable
<
std
::
vector
<
Tensor
>>
();
if
(
offset
>=
feed_list
->
size
())
{
feed_list
->
resize
(
offset
+
1
);
}
return
&
feed_list
->
at
(
offset
);
}
const
Tensor
Base
*
GetOutput
(
size_t
offset
)
{
const
Tensor
*
GetOutput
(
size_t
offset
)
{
auto
*
_fetch_list
=
program_
->
exec_scope
()
->
FindVar
(
"fetch"
);
CHECK
(
_fetch_list
)
<<
"no fatch variable in exec_scope"
;
auto
&
fetch_list
=
*
_fetch_list
->
GetMutable
<
std
::
vector
<
lite
::
TensorBase
>>
();
auto
&
fetch_list
=
*
_fetch_list
->
GetMutable
<
std
::
vector
<
lite
::
Tensor
>>
();
CHECK_LT
(
offset
,
fetch_list
.
size
())
<<
"offset "
<<
offset
<<
" overflow"
;
return
&
fetch_list
.
at
(
offset
);
}
...
...
paddle/fluid/lite/api/light_api_test.cc
浏览文件 @
0a20a618
...
...
@@ -13,21 +13,20 @@
// limitations under the License.
#include "paddle/fluid/lite/api/light_api.h"
#include <gflags/gflags.h>
#include <gtest/gtest.h>
DEFINE_string
(
optimized_model
,
""
,
""
);
namespace
paddle
{
namespace
lite
{
const
std
::
string
model_dir
=
"/home/chunwei/project/Paddle/cmake-build-relwithdebinfo/paddle/fluid/lite/"
"api/optimized_model"
;
TEST
(
LightAPI
,
load
)
{
LightPredictor
predictor
;
predictor
.
Build
(
model_dir
);
predictor
.
Build
(
FLAGS_optimized_model
);
auto
*
input_tensor
=
predictor
.
GetInput
(
0
);
input_tensor
->
Resize
(
{
100
,
100
}
);
input_tensor
->
Resize
(
DDimLite
(
std
::
vector
<
int64_t
>
({
100
,
100
}))
);
auto
*
data
=
input_tensor
->
mutable_data
<
float
>
();
for
(
int
i
=
0
;
i
<
100
*
100
;
i
++
)
{
data
[
i
]
=
i
;
...
...
@@ -40,13 +39,13 @@ TEST(LightAPI, load) {
}
// namespace paddle
USE_LITE_OP
(
mul
);
USE_LITE_OP
(
fc
);
USE_LITE_OP
(
scale
);
//
USE_LITE_OP(fc);
//
USE_LITE_OP(scale);
USE_LITE_OP
(
feed
);
USE_LITE_OP
(
fetch
);
USE_LITE_OP
(
io_copy
);
USE_LITE_KERNEL
(
fc
,
kHost
,
kFloat
,
kNCHW
,
def
);
USE_LITE_KERNEL
(
mul
,
kHost
,
kFloat
,
kNCHW
,
def
);
USE_LITE_KERNEL
(
scale
,
kHost
,
kFloat
,
kNCHW
,
def
);
//
USE_LITE_KERNEL(fc, kHost, kFloat, kNCHW, def);
//
USE_LITE_KERNEL(mul, kHost, kFloat, kNCHW, def);
//
USE_LITE_KERNEL(scale, kHost, kFloat, kNCHW, def);
USE_LITE_KERNEL
(
feed
,
kHost
,
kAny
,
kAny
,
def
);
USE_LITE_KERNEL
(
fetch
,
kHost
,
kAny
,
kAny
,
def
);
paddle/fluid/lite/core/CMakeLists.txt
浏览文件 @
0a20a618
cc_library
(
lite_gtest_main SRCS lite_gtest_main.cc
)
cc_library
(
memory_lite SRCS memory.cc
)
cc_library
(
memory_lite SRCS memory.cc
DEPS target_wrapper_lite target_wrapper_host
)
cc_library
(
target_wrapper_lite SRCS target_wrapper.cc
)
cc_library
(
lite_tensor SRCS lite_tensor.cc DEPS memory_lite target_wrapper_lite
)
...
...
@@ -24,7 +24,7 @@ cc_library(op_lite SRCS op_lite.cc DEPS scope_lite op_registry_lite compatible_p
cc_library
(
types_lite SRCS types.cc
)
cc_library
(
type_system SRCS type_system.cc DEPS
${
tensor_lite
}
target_wrapper_lite
)
cc_library
(
kernel_executor_lite SRCS kernel_executor.cc DEPS mir_ssa_graph kernel_lite
)
#
cc_library(kernel_executor_lite SRCS kernel_executor.cc DEPS mir_ssa_graph kernel_lite)
cc_library
(
program_lite SRCS program.cc DEPS op_lite kernel_lite
)
cc_library
(
optimizer_lite SRCS optimizer.cc DEPS mir_pass_manager model_parser_lite program_lite
)
...
...
@@ -46,5 +46,5 @@ lite_cc_test(test_kernel_lite SRCS kernel_test.cc DEPS kernel_lite target_wrappe
lite_cc_test
(
test_op_lite SRCS op_lite_test.cc DEPS op_lite
)
lite_cc_test
(
test_tensor_lite SRCS lite_tensor_test.cc DEPS lite_tensor
)
lite_cc_test
(
test_type_system SRCS type_system_test.cc DEPS type_system utils_lite
)
lite_cc_test
(
test_optimizer_lite SRCS optimizer_test.cc DEPS mir_pass_manager program_fake_utils mir_passes
)
#lite_cc_test(test_optimizer_lite SRCS optimizer_test.cc DEPS mir_pass_manager program_fake_utils mir_passes optimizer_lite fc_op_lite
)
lite_cc_test
(
test_types_lite SRCS types_test.cc DEPS types_lite
)
paddle/fluid/lite/core/memory.h
浏览文件 @
0a20a618
...
...
@@ -22,9 +22,11 @@ static void* TargetMalloc(TargetType target, size_t size) {
void
*
data
{
nullptr
};
switch
(
target
)
{
case
TargetType
::
kHost
:
#ifdef LITE_WITH_X86
case
TargetType
::
kX86
:
data
=
TargetWrapper
<
TARGET
(
kHost
)
>::
Malloc
(
size
);
break
;
#endif
#ifdef LITE_WITH_CUDA
case
TargetType
::
kCUDA
:
data
=
...
...
paddle/fluid/lite/core/mir/CMakeLists.txt
浏览文件 @
0a20a618
...
...
@@ -22,14 +22,17 @@ endif()
cc_test
(
test_mir_pass_manager SRCS pass_manager_test.cc DEPS mir_pass_manager mir_passes
)
cc_test
(
test_ssa_graph SRCS ssa_graph_test.cc DEPS
mir_ssa_graph scope_lite op_lite
${
ops_lite
}
fc_op_lite
${
host_kernels
}
mir_passes
mir_pass_manager
program_fake_utils
)
set
(
test_variable_place_infrence_pass_DEPS
${
ops_lite
}
mul_op_lite
feed_op_lite
fetch_op_lite
io_copy_op_lite
${
host_kernels
}
mir_passes
mir_pass_manager
...
...
paddle/fluid/lite/core/mir/ssa_graph_test.cc
浏览文件 @
0a20a618
...
...
@@ -35,10 +35,12 @@ void BuildFc(framework::ProgramDesc* desc, const std::string& x,
}
TEST
(
SSAGraph
,
test
)
{
auto
program
=
ProgramFaker
();
auto
program
_faker
=
ProgramFaker
();
SSAGraph
graph
;
std
::
vector
<
Place
>
places
{{
TARGET
(
kHost
),
PRECISION
(
kFloat
)}};
auto
scope
=
std
::
make_shared
<
lite
::
Scope
>
();
lite
::
Program
program
(
*
program_faker
.
program
()
->
Proto
(),
scope
,
places
);
graph
.
Build
(
program
,
places
);
Visualize
(
&
graph
);
...
...
@@ -49,4 +51,4 @@ TEST(SSAGraph, test) {
}
// namespace paddle
USE_LITE_OP
(
fc
);
USE_LITE_KERNEL
(
fc
,
kHost
,
kFloat
);
USE_LITE_KERNEL
(
fc
,
kHost
,
kFloat
,
kNCHW
,
def
);
paddle/fluid/lite/core/mir/variable_place_inference_pass.h
浏览文件 @
0a20a618
...
...
@@ -38,12 +38,6 @@ class VariablePlaceInferencePass : public DebugPass {
LOG
(
INFO
)
<<
"found kernel in inputs "
<<
v
->
AsStmt
().
op_type
;
continue
;
}
// auto& arg = v->AsArgument();
// LOG(INFO) << "get graph input " << arg.name << " " << *arg.type;
// arg.type.target = argument_default_target_;
// the other place description can't be determined yet, until their first
// usage by some kernel.
}
}
...
...
paddle/fluid/lite/core/mir/variable_place_inference_pass_test.cc
浏览文件 @
0a20a618
...
...
@@ -44,7 +44,7 @@ TEST(variable_place_inference_pass, test) {
},
});
Program
program
(
*
desc
,
scope
,
places
);
Program
program
(
*
desc
->
Proto
()
,
scope
,
places
);
core
::
KernelPickFactor
factor
;
factor
.
ConsiderTarget
();
...
...
paddle/fluid/lite/core/op_registry.h
浏览文件 @
0a20a618
...
...
@@ -93,7 +93,8 @@ class KernelRegistry final {
std
::
move
(
creator
));
}
template
<
TargetType
Target
,
PrecisionType
Precision
,
DataLayoutType
Layout
>
template
<
TargetType
Target
,
PrecisionType
Precision
=
PRECISION
(
kFloat
),
DataLayoutType
Layout
=
DATALAYOUT
(
kNCHW
)>
std
::
list
<
std
::
unique_ptr
<
KernelBase
>>
Create
(
const
std
::
string
&
op_type
)
{
using
kernel_registor_t
=
KernelRegistryForTarget
<
Target
,
Precision
,
Layout
>
;
...
...
paddle/fluid/lite/core/optimizer_test.cc
浏览文件 @
0a20a618
...
...
@@ -25,12 +25,18 @@ namespace lite {
TEST
(
Optimizer
,
test
)
{
Optimizer
optimizer
;
auto
program
=
ProgramFaker
();
auto
program_faker
=
ProgramFaker
();
program_faker
.
AddFeed
(
"X"
,
0
);
program_faker
.
AddFetch
(
"X"
,
0
);
std
::
vector
<
Place
>
places
({
Place
{
TARGET
(
kHost
),
PRECISION
(
kFloat
)}});
core
::
KernelPickFactor
factor
;
factor
.
ConsiderTarget
();
auto
scope
=
std
::
make_shared
<
lite
::
Scope
>
();
auto
program_proto
=
*
program_faker
.
program
()
->
Proto
();
Program
program
(
program_proto
,
scope
,
places
);
optimizer
.
Run
(
std
::
move
(
program
),
places
,
factor
);
auto
runtime_program
=
optimizer
.
GenRuntimeProgram
();
LOG
(
INFO
)
<<
"num statements "
<<
runtime_program
->
num_instructions
();
...
...
paddle/fluid/lite/core/program.h
浏览文件 @
0a20a618
...
...
@@ -70,7 +70,7 @@ struct Program {
VLOG
(
4
)
<<
"create Op ["
<<
op_type
<<
"]"
;
auto
op
=
LiteOpRegistry
::
Global
().
Create
(
op_type
);
CHECK
(
op
)
<<
"no Op found for "
<<
op_type
;
ops
.
emplace_back
(
op
);
ops
.
emplace_back
(
std
::
move
(
op
)
);
ops
.
back
()
->
Attach
(
op_desc
,
exec_scope
);
}
}
...
...
paddle/fluid/lite/core/program_fake_utils.h
浏览文件 @
0a20a618
...
...
@@ -14,6 +14,7 @@
#pragma once
#include <string>
#include "paddle/fluid/framework/program_desc.h"
#include "paddle/fluid/lite/core/mir/ssa_graph.h"
#include "paddle/fluid/lite/core/op_registry.h"
...
...
@@ -28,9 +29,9 @@ Program FakeProgram() {
std
::
string
w1
=
"w"
+
std
::
to_string
(
id
);
std
::
string
b1
=
"b"
+
std
::
to_string
(
id
);
std
::
string
out1
=
"out"
+
std
::
to_string
(
id
);
auto
w1v
=
program
.
scope
->
Var
(
w1
)
->
GetMutable
<
TensorBase
>
();
auto
b1v
=
program
.
scope
->
Var
(
b1
)
->
GetMutable
<
TensorBase
>
();
auto
out1v
=
program
.
scope
->
Var
(
out1
)
->
GetMutable
<
TensorBase
>
();
auto
w1v
=
program
.
scope
->
Var
(
w1
)
->
GetMutable
<
lite
::
Tensor
>
();
auto
b1v
=
program
.
scope
->
Var
(
b1
)
->
GetMutable
<
lite
::
Tensor
>
();
auto
out1v
=
program
.
scope
->
Var
(
out1
)
->
GetMutable
<
lite
::
Tensor
>
();
lite
::
OpDesc
desc
;
desc
.
SetInput
(
"Input"
,
{
x
});
...
...
@@ -38,7 +39,7 @@ Program FakeProgram() {
desc
.
SetInput
(
"Bias"
,
{
b1
});
desc
.
SetOutput
(
"Out"
,
{
out1
});
desc
.
SetType
(
"fc"
);
desc
.
SetAttr
<
int
>
(
"in_num_col_dims"
,
1
);
desc
.
SetAttr
(
"in_num_col_dims"
,
1
);
// add to input
program
.
tmp_vars
.
push_back
(
w1
);
...
...
@@ -48,9 +49,9 @@ Program FakeProgram() {
fc_op
->
Attach
(
desc
,
program
.
scope
.
get
());
program
.
ops
.
emplace_back
(
std
::
move
(
fc_op
));
w1v
->
Resize
(
{
100
,
100
}
);
b1v
->
Resize
(
{
100
,
1
}
);
out1v
->
Resize
(
{
100
,
100
}
);
w1v
->
Resize
(
DDimHvy
(
std
::
vector
<
int64_t
>
({
100
,
100
}))
);
b1v
->
Resize
(
DDimHvy
(
std
::
vector
<
int64_t
>
({
100
,
1
}))
);
out1v
->
Resize
(
DDimHvy
(
std
::
vector
<
int64_t
>
({
100
,
100
}))
);
return
out1
;
};
...
...
@@ -60,8 +61,8 @@ Program FakeProgram() {
std
::
string
x
=
"x"
;
program
.
tmp_vars
.
push_back
(
x
);
auto
*
xv
=
program
.
scope
->
Var
(
x
)
->
GetMutable
<
TensorBase
>
();
xv
->
Resize
(
{
100
,
100
}
);
auto
*
xv
=
program
.
scope
->
Var
(
x
)
->
GetMutable
<
lite
::
Tensor
>
();
xv
->
Resize
(
DDimHvy
(
std
::
vector
<
int64_t
>
({
100
,
100
}))
);
for
(
int
i
=
0
;
i
<
3
;
i
++
)
{
x
=
add_fc
(
i
,
x
);
...
...
@@ -81,7 +82,7 @@ class ProgramFaker {
void
CreateVars
(
lite
::
Scope
*
scope
)
{
for
(
auto
&
var
:
tmp_vars_
)
{
auto
*
x
=
scope
->
Var
(
var
);
x
->
GetMutable
<
lite
::
Tensor
Base
>
();
x
->
GetMutable
<
lite
::
Tensor
>
();
}
for
(
auto
&
x
:
tmp_vars_
)
{
...
...
paddle/fluid/lite/kernels/host/CMakeLists.txt
浏览文件 @
0a20a618
message
(
STATUS
"compile with lite host kernels"
)
cc_library
(
fc_compute_host SRCS fc_compute.cc DEPS
${
lite_kernel_deps
}
)
cc_library
(
fc_compute_host SRCS fc_compute.cc DEPS
${
lite_kernel_deps
}
eigen3
)
cc_library
(
relu_compute_host SRCS relu_compute.cc DEPS
${
lite_kernel_deps
}
)
cc_library
(
mul_compute_host SRCS mul_compute.cc DEPS
${
lite_kernel_deps
}
)
cc_library
(
scale_compute_host SRCS scale_compute.cc DEPS
${
lite_kernel_deps
}
)
cc_library
(
mul_compute_host SRCS mul_compute.cc DEPS
${
lite_kernel_deps
}
eigen3
)
cc_library
(
scale_compute_host SRCS scale_compute.cc DEPS
${
lite_kernel_deps
}
eigen3
)
cc_library
(
feed_compute_host SRCS feed_compute.cc DEPS
${
lite_kernel_deps
}
)
cc_library
(
fetch_compute_host SRCS fetch_compute.cc DEPS
${
lite_kernel_deps
}
)
set
(
host_kernels
feed_compute_host
fetch_compute_host
fc_compute_host
relu_compute_host
mul_compute_host
scale_compute_host
)
feed_compute_host
fetch_compute_host
fc_compute_host
relu_compute_host
mul_compute_host
scale_compute_host
)
set
(
host_kernels
"
${
host_kernels
}
"
CACHE INTERNAL
"host kernels"
)
lite_cc_test
(
test_fc_compute SRCS fc_compute_test.cc DEPS fc_compute_host fc_op_lite
)
paddle/fluid/lite/kernels/host/fc_compute_test.cc
浏览文件 @
0a20a618
...
...
@@ -23,7 +23,7 @@ namespace kernels {
namespace
host
{
TEST
(
fc_compute_naive
,
test
)
{
TensorBase
x
,
w
,
b
,
out
,
out1
;
lite
::
Tensor
x
,
w
,
b
,
out
,
out1
;
const
int
batch_size
=
2
;
x
.
Resize
({
batch_size
,
3
});
w
.
Resize
({
4
,
3
});
...
...
@@ -79,15 +79,15 @@ TEST(fc_host, compute) {
FcCompute
fc
;
operators
::
FcParam
param
;
TensorBase
x
;
TensorBase
w
;
TensorBase
bias
;
TensorBase
output
;
lite
::
Tensor
x
;
lite
::
Tensor
w
;
lite
::
Tensor
bias
;
lite
::
Tensor
output
;
x
.
Resize
(
{
1
,
10
,
20
}
);
w
.
Resize
(
{
20
,
20
}
);
bias
.
Resize
(
{
1
,
10
}
);
output
.
Resize
(
{
10
,
20
}
);
x
.
Resize
(
DDim
(
std
::
vector
<
int64_t
>
({
1
,
10
,
20
}))
);
w
.
Resize
(
DDim
(
std
::
vector
<
int64_t
>
({
20
,
20
}))
);
bias
.
Resize
(
DDim
(
std
::
vector
<
int64_t
>
({
1
,
10
}))
);
output
.
Resize
(
DDim
(
std
::
vector
<
int64_t
>
({
10
,
20
}))
);
auto
*
x_data
=
x
.
mutable_data
<
float
>
();
auto
*
w_data
=
w
.
mutable_data
<
float
>
();
...
...
@@ -119,7 +119,7 @@ TEST(fc_host, compute) {
TEST
(
fc
,
retrive_op
)
{
auto
fc
=
KernelRegistry
::
Global
().
Create
<
TARGET
(
kHost
),
PRECISION
(
kFloat
)
>
(
"fc"
);
ASSERT_TRUE
(
fc
.
get
()
);
ASSERT_TRUE
(
fc
);
}
}
// namespace host
...
...
@@ -127,4 +127,4 @@ TEST(fc, retrive_op) {
}
// namespace lite
}
// namespace paddle
USE_LITE_KERNEL
(
fc
,
kHost
,
kFloat
);
USE_LITE_KERNEL
(
fc
,
kHost
,
kFloat
,
kNCHW
,
def
);
paddle/fluid/lite/model_parser/CMakeLists.txt
浏览文件 @
0a20a618
cc_library
(
runtime_lite SRCS runtime.cc
)
lite_cc_test
(
test_model_parser_lite SRCS model_parser_test.cc DEPS model_parser_lite
)
#cc_library(runtime_lite SRCS runtime.cc)
lite_cc_test
(
test_model_parser_lite SRCS model_parser_test.cc
DEPS model_parser_lite framework_proto_lite
ARGS --model_dir=
${
LITE_MODEL_DIR
}
/lite_naive_model
)
if
(
WITH_TESTING
)
add_dependencies
(
test_model_parser_lite extern_lite_download_lite_naive_model_tar_gz
)
endif
(
WITH_TESTING
)
if
(
LITE_WITH_LIGHT_WEIGHT_FRAMEWORK
)
cc_library
(
compatible_pb_lite SRCS compatible_pb.cc DEPS op_desc_lite var_desc_lite
)
else
()
...
...
paddle/fluid/lite/model_parser/compatible_pb.h
浏览文件 @
0a20a618
...
...
@@ -20,8 +20,8 @@
* lite::pb::XXDesc.
*/
#include "paddle/fluid/framework/framework.pb.h"
#ifdef LITE_WITH_LIGHT_WEIGHT_FRAMEWORK
#include "paddle/fluid/lite/core/framework.pb.h"
#include "paddle/fluid/lite/model_parser/pb/op_desc.h"
#include "paddle/fluid/lite/model_parser/pb/var_desc.h"
#else
...
...
paddle/fluid/lite/model_parser/model_parser.h
浏览文件 @
0a20a618
...
...
@@ -18,7 +18,7 @@
#include <memory>
#include <string>
#include <vector>
#include "paddle/fluid/
framework
/framework.pb.h"
#include "paddle/fluid/
lite/core
/framework.pb.h"
#include "paddle/fluid/lite/core/scope.h"
#include "paddle/fluid/lite/core/variable.h"
...
...
paddle/fluid/lite/model_parser/model_parser_test.cc
浏览文件 @
0a20a618
...
...
@@ -13,22 +13,25 @@
// limitations under the License.
#include "paddle/fluid/lite/model_parser/model_parser.h"
#include <gflags/gflags.h>
#include <gtest/gtest.h>
#include "paddle/fluid/lite/core/scope.h"
DEFINE_string
(
model_dir
,
""
,
""
);
namespace
paddle
{
namespace
lite
{
TEST
(
ModelParser
,
LoadProgram
)
{
auto
program
=
LoadProgram
(
"/home/chunwei/project2/models/fc/fluid_checkpoint
/__model__"
);
CHECK
(
!
FLAGS_model_dir
.
empty
());
auto
program
=
LoadProgram
(
FLAGS_model_dir
+
"
/__model__"
);
}
TEST
(
ModelParser
,
LoadParam
)
{
Scope
scope
;
auto
*
v
=
scope
.
Var
(
"xxx"
);
LoadParam
(
"/home/chunwei/project2/models/fc/fluid_checkpoint/b1
"
,
v
);
const
auto
&
t
=
v
->
Get
<
Tensor
Base
>
();
LoadParam
(
FLAGS_model_dir
+
"/fc_0.b_0
"
,
v
);
const
auto
&
t
=
v
->
Get
<
Tensor
>
();
LOG
(
INFO
)
<<
"loaded
\n
"
;
LOG
(
INFO
)
<<
t
;
}
...
...
@@ -36,7 +39,7 @@ TEST(ModelParser, LoadParam) {
TEST
(
ModelParser
,
LoadModel
)
{
Scope
scope
;
framework
::
proto
::
ProgramDesc
prog
;
LoadModel
(
"/home/chunwei/project2/models/fc/fluid_checkpoint"
,
&
scope
,
&
prog
);
LoadModel
(
FLAGS_model_dir
,
&
scope
,
&
prog
);
}
}
// namespace lite
...
...
paddle/fluid/lite/model_parser/pb/op_desc.h
浏览文件 @
0a20a618
...
...
@@ -24,7 +24,7 @@
#include <string>
#include <unordered_map>
#include <vector>
#include "paddle/fluid/
framework
/framework.pb.h"
#include "paddle/fluid/
lite/core
/framework.pb.h"
#include "paddle/fluid/lite/utils/all.h"
namespace
paddle
{
...
...
paddle/fluid/lite/model_parser/pb/var_desc.h
浏览文件 @
0a20a618
...
...
@@ -17,7 +17,7 @@
#include <algorithm>
#include <string>
#include <vector>
#include "paddle/fluid/
framework
/framework.pb.h"
#include "paddle/fluid/
lite/core
/framework.pb.h"
#include "paddle/fluid/lite/utils/cp_logging.h"
namespace
paddle
{
...
...
paddle/fluid/lite/operators/fc_op.h
浏览文件 @
0a20a618
...
...
@@ -59,9 +59,6 @@ class FcOpLite : public OpLite {
param_
.
output
=
scope
->
FindVar
(
out
)
->
GetMutable
<
lite
::
Tensor
>
();
param_
.
in_num_col_dims
=
GetAttr
<
int
>
(
op_desc
.
GetAttr
(
"in_num_col_dims"
));
CHECK
(
kernel_
);
kernel_
->
SetParam
(
param_
);
return
true
;
}
...
...
paddle/fluid/lite/operators/fc_op_test.cc
浏览文件 @
0a20a618
...
...
@@ -21,17 +21,16 @@ namespace lite {
namespace
operators
{
TEST
(
fc_op_lite
,
test
)
{
LOG
(
INFO
)
<<
"
\n
"
<<
KernelRegistry
::
Global
().
DebugString
();
// prepare variables
Scope
scope
;
auto
*
x
=
scope
.
Var
(
"x"
)
->
GetMutable
<
Tensor
Base
>
();
auto
*
w
=
scope
.
Var
(
"w"
)
->
GetMutable
<
Tensor
Base
>
();
auto
*
bias
=
scope
.
Var
(
"bias"
)
->
GetMutable
<
Tensor
Base
>
();
auto
*
output
=
scope
.
Var
(
"output"
)
->
GetMutable
<
Tensor
Base
>
();
x
->
Resize
(
{
1
,
10
,
20
}
);
w
->
Resize
(
{
20
,
20
}
);
bias
->
Resize
(
{
1
,
10
}
);
output
->
Resize
(
{
10
,
20
}
);
auto
*
x
=
scope
.
Var
(
"x"
)
->
GetMutable
<
Tensor
>
();
auto
*
w
=
scope
.
Var
(
"w"
)
->
GetMutable
<
Tensor
>
();
auto
*
bias
=
scope
.
Var
(
"bias"
)
->
GetMutable
<
Tensor
>
();
auto
*
output
=
scope
.
Var
(
"output"
)
->
GetMutable
<
Tensor
>
();
x
->
Resize
(
DDim
(
std
::
vector
<
int64_t
>
({
1
,
10
,
20
}))
);
w
->
Resize
(
DDim
(
std
::
vector
<
int64_t
>
{
20
,
20
})
);
bias
->
Resize
(
DDim
(
std
::
vector
<
int64_t
>
{
1
,
10
})
);
output
->
Resize
(
DDim
(
std
::
vector
<
int64_t
>
{
10
,
20
})
);
// set data
for
(
int
i
=
0
;
i
<
10
*
20
;
i
++
)
{
...
...
@@ -59,18 +58,13 @@ TEST(fc_op_lite, test) {
FcOpLite
fc
(
"fc"
);
fc
.
SetValidPlaces
({
Place
{
TARGET
(
kHost
),
PRECISION
(
kFloat
)}});
fc
.
PickKernel
({
Place
{
TARGET
(
kHost
),
PRECISION
(
kFloat
)}});
fc
.
AttachImpl
(
desc
,
&
scope
);
fc
.
Run
();
for
(
int
i
=
0
;
i
<
10
*
20
;
i
++
)
{
LOG
(
INFO
)
<<
output
->
data
<
float
>
()[
i
];
}
fc
.
Attach
(
desc
,
&
scope
);
auto
kernels
=
fc
.
CreateKernels
({
Place
{
TARGET
(
kHost
),
PRECISION
(
kFloat
)}});
ASSERT_FALSE
(
kernels
.
empty
());
}
}
// namespace operators
}
// namespace lite
}
// namespace paddle
USE_LITE_KERNEL
(
fc
,
kHost
,
kFloat
);
USE_LITE_KERNEL
(
fc
,
kHost
,
kFloat
,
kNCHW
,
def
);
paddle/fluid/lite/utils/CMakeLists.txt
浏览文件 @
0a20a618
if
(
LITE_WITH_LIGHT_WEIGHT_FRAMEWORK
)
set
(
utils_DEPS
)
lite_cc_test
(
test_logging_lite SRCS logging_test.cc
)
else
()
set
(
utils_DEPS glog
)
endif
()
lite_cc_test
(
test_logging_lite SRCS logging_test.cc
)
lite_cc_test
(
test_varient SRCS varient_test.cc DEPS utils_lite
)
cc_library
(
any_lite SRCS any.cc
)
cc_library
(
utils_lite SRCS cp_logging.cc DEPS
${
utils_DEPS
}
any_lite
)
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录