Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
8864f459
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
8864f459
编写于
6月 15, 2019
作者:
N
nhzlx
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
1. delete unused file
2. refine after mrege from github
上级
0c3191f4
变更
9
隐藏空白更改
内联
并排
Showing
9 changed file
with
28 addition
and
145 deletion
+28
-145
paddle/fluid/lite/api/CMakeLists.txt
paddle/fluid/lite/api/CMakeLists.txt
+0
-2
paddle/fluid/lite/core/mir/fusion/CMakeLists.txt
paddle/fluid/lite/core/mir/fusion/CMakeLists.txt
+4
-0
paddle/fluid/lite/core/mir/fusion/conv_bn_fuser.cc
paddle/fluid/lite/core/mir/fusion/conv_bn_fuser.cc
+3
-3
paddle/fluid/lite/core/mir/pattern_matcher.cc
paddle/fluid/lite/core/mir/pattern_matcher.cc
+4
-4
paddle/fluid/lite/core/mir/ssa_graph.cc
paddle/fluid/lite/core/mir/ssa_graph.cc
+0
-7
paddle/fluid/lite/core/optimizer.h
paddle/fluid/lite/core/optimizer.h
+16
-10
paddle/fluid/lite/operators/batch_norm.cc
paddle/fluid/lite/operators/batch_norm.cc
+0
-31
paddle/fluid/lite/operators/batch_norm.h
paddle/fluid/lite/operators/batch_norm.h
+0
-87
paddle/fluid/lite/operators/batch_norm_op.cc
paddle/fluid/lite/operators/batch_norm_op.cc
+1
-1
未找到文件。
paddle/fluid/lite/api/CMakeLists.txt
浏览文件 @
8864f459
...
...
@@ -43,8 +43,6 @@ endif()
# lite_cc_test(test_light_api SRCS light_api_test.cc DEPS light_api_lite ARGS --optimized_model=${LITE_MODEL_DIR}/lite_naive_model_opt SERIAL)
# endif()
message
(
STATUS
"!!!"
${
ops_lite
}
)
message
(
STATUS
"!!!"
${
arm_kernels
}
)
lite_cc_binary
(
cxx_api_lite_bin SRCS cxx_api_bin.cc
DEPS
cxx_api_lite
...
...
paddle/fluid/lite/core/mir/fusion/CMakeLists.txt
浏览文件 @
8864f459
...
...
@@ -14,5 +14,9 @@ set(mir_fusers
fuse_conv_bn
CACHE INTERNAL
"fusers"
)
if
(
LITE_WITH_LIGHT_WEIGHT_FRAMEWORK
)
return
()
endif
()
lite_cc_test
(
test_lite_conv_bn_fuse SRCS conv_bn_fuse_pass_test.cc
DEPS elementwise_ops_lite batch_norm_op_lite conv_op_lite proto_desc compatible_pb_lite program_lite mir_pass mir_pass_manager pattern_matcher_high_api
)
paddle/fluid/lite/core/mir/fusion/conv_bn_fuser.cc
浏览文件 @
8864f459
...
...
@@ -84,9 +84,9 @@ void ConvBNFuser::InsertNewNode(SSAGraph* graph, const key2nodes_t& matched) {
->
GetMutable
<
lite
::
Tensor
>
();
size_t
bias_size
=
bn_scale_t
->
data_size
();
auto
bn_scale_d
=
bn_scale_t
->
mutable_data
<
float
>
();
PADDLE_ENFORCE
(
bias_size
==
conv_weight_dims
[
0
],
"The BN bias's size should be equal to the size of the first "
"dim size of the conv weights"
)
;
CHECK
(
bias_size
==
conv_weight_dims
[
0
])
<<
"The BN bias's size should be equal to the size of the first "
<<
"dim size of the conv weights"
;
auto
bn_mean_t
=
scope
->
FindVar
(
matched
.
at
(
"bn_mean"
)
->
arg
()
->
name
)
->
GetMutable
<
lite
::
Tensor
>
();
...
...
paddle/fluid/lite/core/mir/pattern_matcher.cc
浏览文件 @
8864f459
...
...
@@ -409,8 +409,8 @@ PMNode *PMNode::assert_is_op_output(const std::string &op_type) {
bool
IsNthOutput
(
const
Node
*
var
,
const
Node
*
op
,
const
std
::
string
&
argument
,
size_t
nth
)
{
PADDLE_ENFORCE
(
var
->
IsArg
());
PADDLE_ENFORCE
(
op
->
IsStmt
());
CHECK
(
var
->
IsArg
());
CHECK
(
op
->
IsStmt
());
auto
op_info
=
op
->
stmt
()
->
op_info
();
if
(
op_info
->
Output
(
argument
).
size
()
<=
nth
)
return
false
;
return
var
->
arg
()
->
name
==
op_info
->
Output
(
argument
)[
nth
];
...
...
@@ -418,8 +418,8 @@ bool IsNthOutput(const Node *var, const Node *op, const std::string &argument,
bool
IsNthInput
(
const
Node
*
var
,
const
Node
*
op
,
const
std
::
string
&
argument
,
size_t
nth
)
{
PADDLE_ENFORCE
(
var
->
IsArg
());
PADDLE_ENFORCE
(
op
->
IsStmt
());
CHECK
(
var
->
IsArg
());
CHECK
(
op
->
IsStmt
());
auto
op_info
=
op
->
stmt
()
->
op_info
();
if
(
op_info
->
Input
(
argument
).
size
()
<=
nth
)
return
false
;
return
var
->
arg
()
->
name
==
op_info
->
Input
(
argument
)[
nth
];
...
...
paddle/fluid/lite/core/mir/ssa_graph.cc
浏览文件 @
8864f459
...
...
@@ -26,8 +26,6 @@ namespace mir {
bool
SSAGraph
::
CheckBidirectionalConnection
()
{
LOG
(
INFO
)
<<
"node count "
<<
node_storage_
.
size
();
for
(
auto
&
node
:
node_storage_
)
{
if
(
node
.
IsStmt
())
LOG
(
INFO
)
<<
node
.
AsStmt
().
op_info
()
->
Type
();
if
(
node
.
IsArg
())
LOG
(
INFO
)
<<
node
.
AsArg
().
name
<<
" "
<<
node
.
AsArg
().
id
;
for
(
auto
*
in
:
node
.
inlinks
)
{
CHECK
(
in
->
outlinks
.
end
()
!=
std
::
find
(
in
->
outlinks
.
begin
(),
in
->
outlinks
.
end
(),
&
node
));
...
...
@@ -123,11 +121,8 @@ void SSAGraph::Build(const Program &program,
std
::
unordered_map
<
std
::
string
,
mir
::
Node
*>
arg_update_node_map_
;
for
(
auto
&
op
:
program
.
ops
())
{
LOG
(
INFO
)
<<
op
->
op_info
()
->
Type
();
auto
*
op_node
=
GraphCreateInstructNode
(
op
,
valid_places
);
LOG
(
INFO
)
<<
"input:"
;
for
(
const
std
::
string
&
name
:
op
->
op_info
()
->
input_names
())
{
LOG
(
INFO
)
<<
name
;
mir
::
Node
*
arg_node
=
nullptr
;
if
(
arg_update_node_map_
.
count
(
name
))
{
arg_node
=
arg_update_node_map_
.
at
(
name
);
...
...
@@ -141,9 +136,7 @@ void SSAGraph::Build(const Program &program,
CHECK
(
arg_node
->
IsRoleSet
());
DirectedLink
(
arg_node
,
op_node
);
}
LOG
(
INFO
)
<<
"output:"
;
for
(
const
std
::
string
&
name
:
op
->
op_info
()
->
output_names
())
{
LOG
(
INFO
)
<<
name
;
node_storage_
.
emplace_back
();
auto
*
arg_node
=
&
node_storage_
.
back
();
arg_node
->
AsArg
(
name
,
node_storage_
.
size
()
-
1
);
...
...
paddle/fluid/lite/core/optimizer.h
浏览文件 @
8864f459
...
...
@@ -46,27 +46,33 @@ class Optimizer {
SpecifyKernelPickTactic
(
kernel_pick_factor
);
InitTargetTypeTransformPass
();
#ifndef LITE_WITH_LIGHT_WEIGHT_FRAMEWORK
//
#ifndef LITE_WITH_LIGHT_WEIGHT_FRAMEWORK
if
(
passes
.
empty
())
{
RunPasses
(
std
::
vector
<
std
::
string
>
{{
"graph_visualze"
,
//
"lite_conv_bn_fuse_pass"
,
//
"graph_visualze"
,
//
"lite_conv_elementwise_add_act_fuse_pass"
,
//
"graph_visualze"
,
//
"lite_fc_fuse_pass"
,
//
"static_kernel_pick_pass"
,
//
"variable_place_inference_pass"
,
//
"argument_type_display_pass"
,
//
"type_target_transform_pass"
,
//
"argument_type_display_pass"
,
//
"variable_place_inference_pass"
,
//
"argument_type_display_pass"
,
//
"io_copy_kernel_pick_pass"
,
//
"variable_place_inference_pass"
,
//
/*
"variable_place_inference_pass", //
"argument_type_display_pass", //
"type_target_transform_pass", //
// "argument_type_display_pass", //
"variable_place_inference_pass", //
"argument_type_display_pass", //
"io_copy_kernel_pick_pass", //
"variable_place_inference_pass", //
"graph_visualze", //
*/
"runtime_context_assign_pass"
,
//
}});
}
else
{
RunPasses
(
passes
);
}
#endif
//
#endif
exec_scope_
=
program
.
exec_scope
();
}
...
...
paddle/fluid/lite/operators/batch_norm.cc
已删除
100644 → 0
浏览文件 @
0c3191f4
// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "paddle/fluid/lite/operators/batch_norm.h"
#include <vector>
#include "paddle/fluid/lite/core/op_registry.h"
namespace
paddle
{
namespace
lite
{
namespace
operators
{
bool
BatchNormOpLite
::
CheckShape
()
const
{
return
true
;
}
bool
BatchNormOpLite
::
InferShape
()
const
{
return
true
;
}
}
// namespace operators
}
// namespace lite
}
// namespace paddle
REGISTER_LITE_OP
(
batch_norm
,
paddle
::
lite
::
operators
::
BatchNormOpLite
);
paddle/fluid/lite/operators/batch_norm.h
已删除
100644 → 0
浏览文件 @
0c3191f4
// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#pragma once
#include <string>
#include <vector>
#include "paddle/fluid/lite/core/kernel.h"
#include "paddle/fluid/lite/core/op_lite.h"
#include "paddle/fluid/lite/core/scope.h"
#include "paddle/fluid/lite/operators/op_params.h"
#include "paddle/fluid/lite/utils/all.h"
namespace
paddle
{
namespace
lite
{
namespace
operators
{
class
BatchNormOpLite
:
public
OpLite
{
public:
BatchNormOpLite
()
{}
explicit
BatchNormOpLite
(
const
std
::
string
&
type
)
:
OpLite
(
type
)
{}
bool
CheckShape
()
const
override
;
bool
InferShape
()
const
override
;
void
AttachKernel
(
KernelBase
*
kernel
)
override
{
kernel
->
SetParam
(
param_
);
}
// TODO(Superjomn) replace framework::OpDesc with a lite one.
bool
AttachImpl
(
const
cpp
::
OpDesc
&
op_desc
,
lite
::
Scope
*
scope
)
override
{
auto
x
=
op_desc
.
Input
(
"X"
).
front
();
auto
bias
=
op_desc
.
Input
(
"Bias"
).
front
();
auto
mean
=
op_desc
.
Input
(
"Mean"
).
front
();
auto
scale
=
op_desc
.
Input
(
"Scale"
).
front
();
auto
variance
=
op_desc
.
Input
(
"Variance"
).
front
();
auto
out
=
op_desc
.
Output
(
"Y"
).
front
();
auto
mean_out
=
op_desc
.
Output
(
"MeanOut"
).
front
();
auto
var_out
=
op_desc
.
Output
(
"VarianceOut"
).
front
();
auto
saved_mean
=
op_desc
.
Output
(
"SavedMean"
).
front
();
auto
saved_var
=
op_desc
.
Output
(
"SavedVariance"
).
front
();
auto
*
var
=
scope
->
FindVar
(
x
);
param_
.
x
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
bias
);
param_
.
bias
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
mean
);
param_
.
mean
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
scale
);
param_
.
scale
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
variance
);
param_
.
var
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
out
);
param_
.
out
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
mean_out
);
param_
.
mean_out
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
var_out
);
param_
.
var_out
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
saved_mean
);
param_
.
saved_mean
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
saved_var
);
param_
.
saved_var
=
var
->
GetMutable
<
Tensor
>
();
param_
.
eps
=
op_desc
.
GetAttr
<
float
>
(
"epsilon"
);
return
true
;
}
std
::
string
DebugString
()
const
override
{
return
"batch_norm"
;
}
private:
mutable
BatchNormParam
param_
;
};
}
// namespace operators
}
// namespace lite
}
// namespace paddle
paddle/fluid/lite/operators/batch_norm_op.cc
浏览文件 @
8864f459
...
...
@@ -82,7 +82,7 @@ bool BatchNormOp::AttachImpl(const cpp::OpDesc &op_desc, lite::Scope *scope) {
param_
.
variance
=
scope
->
FindVar
(
op_desc
.
Input
(
"Variance"
).
front
())
->
GetMutable
<
Tensor
>
();
param_
.
y
=
scope
->
FindVar
(
op_desc
.
Output
(
"Y"
).
front
())
->
GetMutable
<
Tensor
>
();
param_
.
is_test
=
op_desc
.
GetAttr
<
bool
>
(
"is_test"
);
param_
.
is_test
=
op_desc
.
GetAttr
<
int
>
(
"is_test"
);
param_
.
use_global_stats
=
op_desc
.
GetAttr
<
bool
>
(
"use_global_stats"
);
if
(
!
param_
.
is_test
)
{
param_
.
mean_out
=
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录