Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
PaddleDetection
提交
8864f459
P
PaddleDetection
项目概览
PaddlePaddle
/
PaddleDetection
1 年多 前同步成功
通知
696
Star
11112
Fork
2696
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
184
列表
看板
标记
里程碑
合并请求
40
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
PaddleDetection
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
184
Issue
184
列表
看板
标记
里程碑
合并请求
40
合并请求
40
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
8864f459
编写于
6月 15, 2019
作者:
N
nhzlx
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
1. delete unused file
2. refine after mrege from github
上级
0c3191f4
变更
9
隐藏空白更改
内联
并排
Showing
9 changed file
with
28 addition
and
145 deletion
+28
-145
paddle/fluid/lite/api/CMakeLists.txt
paddle/fluid/lite/api/CMakeLists.txt
+0
-2
paddle/fluid/lite/core/mir/fusion/CMakeLists.txt
paddle/fluid/lite/core/mir/fusion/CMakeLists.txt
+4
-0
paddle/fluid/lite/core/mir/fusion/conv_bn_fuser.cc
paddle/fluid/lite/core/mir/fusion/conv_bn_fuser.cc
+3
-3
paddle/fluid/lite/core/mir/pattern_matcher.cc
paddle/fluid/lite/core/mir/pattern_matcher.cc
+4
-4
paddle/fluid/lite/core/mir/ssa_graph.cc
paddle/fluid/lite/core/mir/ssa_graph.cc
+0
-7
paddle/fluid/lite/core/optimizer.h
paddle/fluid/lite/core/optimizer.h
+16
-10
paddle/fluid/lite/operators/batch_norm.cc
paddle/fluid/lite/operators/batch_norm.cc
+0
-31
paddle/fluid/lite/operators/batch_norm.h
paddle/fluid/lite/operators/batch_norm.h
+0
-87
paddle/fluid/lite/operators/batch_norm_op.cc
paddle/fluid/lite/operators/batch_norm_op.cc
+1
-1
未找到文件。
paddle/fluid/lite/api/CMakeLists.txt
浏览文件 @
8864f459
...
...
@@ -43,8 +43,6 @@ endif()
# lite_cc_test(test_light_api SRCS light_api_test.cc DEPS light_api_lite ARGS --optimized_model=${LITE_MODEL_DIR}/lite_naive_model_opt SERIAL)
# endif()
message
(
STATUS
"!!!"
${
ops_lite
}
)
message
(
STATUS
"!!!"
${
arm_kernels
}
)
lite_cc_binary
(
cxx_api_lite_bin SRCS cxx_api_bin.cc
DEPS
cxx_api_lite
...
...
paddle/fluid/lite/core/mir/fusion/CMakeLists.txt
浏览文件 @
8864f459
...
...
@@ -14,5 +14,9 @@ set(mir_fusers
fuse_conv_bn
CACHE INTERNAL
"fusers"
)
if
(
LITE_WITH_LIGHT_WEIGHT_FRAMEWORK
)
return
()
endif
()
lite_cc_test
(
test_lite_conv_bn_fuse SRCS conv_bn_fuse_pass_test.cc
DEPS elementwise_ops_lite batch_norm_op_lite conv_op_lite proto_desc compatible_pb_lite program_lite mir_pass mir_pass_manager pattern_matcher_high_api
)
paddle/fluid/lite/core/mir/fusion/conv_bn_fuser.cc
浏览文件 @
8864f459
...
...
@@ -84,9 +84,9 @@ void ConvBNFuser::InsertNewNode(SSAGraph* graph, const key2nodes_t& matched) {
->
GetMutable
<
lite
::
Tensor
>
();
size_t
bias_size
=
bn_scale_t
->
data_size
();
auto
bn_scale_d
=
bn_scale_t
->
mutable_data
<
float
>
();
PADDLE_ENFORCE
(
bias_size
==
conv_weight_dims
[
0
],
"The BN bias's size should be equal to the size of the first "
"dim size of the conv weights"
)
;
CHECK
(
bias_size
==
conv_weight_dims
[
0
])
<<
"The BN bias's size should be equal to the size of the first "
<<
"dim size of the conv weights"
;
auto
bn_mean_t
=
scope
->
FindVar
(
matched
.
at
(
"bn_mean"
)
->
arg
()
->
name
)
->
GetMutable
<
lite
::
Tensor
>
();
...
...
paddle/fluid/lite/core/mir/pattern_matcher.cc
浏览文件 @
8864f459
...
...
@@ -409,8 +409,8 @@ PMNode *PMNode::assert_is_op_output(const std::string &op_type) {
bool
IsNthOutput
(
const
Node
*
var
,
const
Node
*
op
,
const
std
::
string
&
argument
,
size_t
nth
)
{
PADDLE_ENFORCE
(
var
->
IsArg
());
PADDLE_ENFORCE
(
op
->
IsStmt
());
CHECK
(
var
->
IsArg
());
CHECK
(
op
->
IsStmt
());
auto
op_info
=
op
->
stmt
()
->
op_info
();
if
(
op_info
->
Output
(
argument
).
size
()
<=
nth
)
return
false
;
return
var
->
arg
()
->
name
==
op_info
->
Output
(
argument
)[
nth
];
...
...
@@ -418,8 +418,8 @@ bool IsNthOutput(const Node *var, const Node *op, const std::string &argument,
bool
IsNthInput
(
const
Node
*
var
,
const
Node
*
op
,
const
std
::
string
&
argument
,
size_t
nth
)
{
PADDLE_ENFORCE
(
var
->
IsArg
());
PADDLE_ENFORCE
(
op
->
IsStmt
());
CHECK
(
var
->
IsArg
());
CHECK
(
op
->
IsStmt
());
auto
op_info
=
op
->
stmt
()
->
op_info
();
if
(
op_info
->
Input
(
argument
).
size
()
<=
nth
)
return
false
;
return
var
->
arg
()
->
name
==
op_info
->
Input
(
argument
)[
nth
];
...
...
paddle/fluid/lite/core/mir/ssa_graph.cc
浏览文件 @
8864f459
...
...
@@ -26,8 +26,6 @@ namespace mir {
bool
SSAGraph
::
CheckBidirectionalConnection
()
{
LOG
(
INFO
)
<<
"node count "
<<
node_storage_
.
size
();
for
(
auto
&
node
:
node_storage_
)
{
if
(
node
.
IsStmt
())
LOG
(
INFO
)
<<
node
.
AsStmt
().
op_info
()
->
Type
();
if
(
node
.
IsArg
())
LOG
(
INFO
)
<<
node
.
AsArg
().
name
<<
" "
<<
node
.
AsArg
().
id
;
for
(
auto
*
in
:
node
.
inlinks
)
{
CHECK
(
in
->
outlinks
.
end
()
!=
std
::
find
(
in
->
outlinks
.
begin
(),
in
->
outlinks
.
end
(),
&
node
));
...
...
@@ -123,11 +121,8 @@ void SSAGraph::Build(const Program &program,
std
::
unordered_map
<
std
::
string
,
mir
::
Node
*>
arg_update_node_map_
;
for
(
auto
&
op
:
program
.
ops
())
{
LOG
(
INFO
)
<<
op
->
op_info
()
->
Type
();
auto
*
op_node
=
GraphCreateInstructNode
(
op
,
valid_places
);
LOG
(
INFO
)
<<
"input:"
;
for
(
const
std
::
string
&
name
:
op
->
op_info
()
->
input_names
())
{
LOG
(
INFO
)
<<
name
;
mir
::
Node
*
arg_node
=
nullptr
;
if
(
arg_update_node_map_
.
count
(
name
))
{
arg_node
=
arg_update_node_map_
.
at
(
name
);
...
...
@@ -141,9 +136,7 @@ void SSAGraph::Build(const Program &program,
CHECK
(
arg_node
->
IsRoleSet
());
DirectedLink
(
arg_node
,
op_node
);
}
LOG
(
INFO
)
<<
"output:"
;
for
(
const
std
::
string
&
name
:
op
->
op_info
()
->
output_names
())
{
LOG
(
INFO
)
<<
name
;
node_storage_
.
emplace_back
();
auto
*
arg_node
=
&
node_storage_
.
back
();
arg_node
->
AsArg
(
name
,
node_storage_
.
size
()
-
1
);
...
...
paddle/fluid/lite/core/optimizer.h
浏览文件 @
8864f459
...
...
@@ -46,27 +46,33 @@ class Optimizer {
SpecifyKernelPickTactic
(
kernel_pick_factor
);
InitTargetTypeTransformPass
();
#ifndef LITE_WITH_LIGHT_WEIGHT_FRAMEWORK
//
#ifndef LITE_WITH_LIGHT_WEIGHT_FRAMEWORK
if
(
passes
.
empty
())
{
RunPasses
(
std
::
vector
<
std
::
string
>
{{
"graph_visualze"
,
//
"lite_conv_bn_fuse_pass"
,
//
"graph_visualze"
,
//
"lite_conv_elementwise_add_act_fuse_pass"
,
//
"graph_visualze"
,
//
"lite_fc_fuse_pass"
,
//
"static_kernel_pick_pass"
,
//
"variable_place_inference_pass"
,
//
"argument_type_display_pass"
,
//
"type_target_transform_pass"
,
//
"argument_type_display_pass"
,
//
"variable_place_inference_pass"
,
//
"argument_type_display_pass"
,
//
"io_copy_kernel_pick_pass"
,
//
"variable_place_inference_pass"
,
//
/*
"variable_place_inference_pass", //
"argument_type_display_pass", //
"type_target_transform_pass", //
// "argument_type_display_pass", //
"variable_place_inference_pass", //
"argument_type_display_pass", //
"io_copy_kernel_pick_pass", //
"variable_place_inference_pass", //
"graph_visualze", //
*/
"runtime_context_assign_pass"
,
//
}});
}
else
{
RunPasses
(
passes
);
}
#endif
//
#endif
exec_scope_
=
program
.
exec_scope
();
}
...
...
paddle/fluid/lite/operators/batch_norm.cc
已删除
100644 → 0
浏览文件 @
0c3191f4
// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "paddle/fluid/lite/operators/batch_norm.h"
#include <vector>
#include "paddle/fluid/lite/core/op_registry.h"
namespace
paddle
{
namespace
lite
{
namespace
operators
{
bool
BatchNormOpLite
::
CheckShape
()
const
{
return
true
;
}
bool
BatchNormOpLite
::
InferShape
()
const
{
return
true
;
}
}
// namespace operators
}
// namespace lite
}
// namespace paddle
REGISTER_LITE_OP
(
batch_norm
,
paddle
::
lite
::
operators
::
BatchNormOpLite
);
paddle/fluid/lite/operators/batch_norm.h
已删除
100644 → 0
浏览文件 @
0c3191f4
// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#pragma once
#include <string>
#include <vector>
#include "paddle/fluid/lite/core/kernel.h"
#include "paddle/fluid/lite/core/op_lite.h"
#include "paddle/fluid/lite/core/scope.h"
#include "paddle/fluid/lite/operators/op_params.h"
#include "paddle/fluid/lite/utils/all.h"
namespace
paddle
{
namespace
lite
{
namespace
operators
{
class
BatchNormOpLite
:
public
OpLite
{
public:
BatchNormOpLite
()
{}
explicit
BatchNormOpLite
(
const
std
::
string
&
type
)
:
OpLite
(
type
)
{}
bool
CheckShape
()
const
override
;
bool
InferShape
()
const
override
;
void
AttachKernel
(
KernelBase
*
kernel
)
override
{
kernel
->
SetParam
(
param_
);
}
// TODO(Superjomn) replace framework::OpDesc with a lite one.
bool
AttachImpl
(
const
cpp
::
OpDesc
&
op_desc
,
lite
::
Scope
*
scope
)
override
{
auto
x
=
op_desc
.
Input
(
"X"
).
front
();
auto
bias
=
op_desc
.
Input
(
"Bias"
).
front
();
auto
mean
=
op_desc
.
Input
(
"Mean"
).
front
();
auto
scale
=
op_desc
.
Input
(
"Scale"
).
front
();
auto
variance
=
op_desc
.
Input
(
"Variance"
).
front
();
auto
out
=
op_desc
.
Output
(
"Y"
).
front
();
auto
mean_out
=
op_desc
.
Output
(
"MeanOut"
).
front
();
auto
var_out
=
op_desc
.
Output
(
"VarianceOut"
).
front
();
auto
saved_mean
=
op_desc
.
Output
(
"SavedMean"
).
front
();
auto
saved_var
=
op_desc
.
Output
(
"SavedVariance"
).
front
();
auto
*
var
=
scope
->
FindVar
(
x
);
param_
.
x
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
bias
);
param_
.
bias
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
mean
);
param_
.
mean
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
scale
);
param_
.
scale
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
variance
);
param_
.
var
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
out
);
param_
.
out
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
mean_out
);
param_
.
mean_out
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
var_out
);
param_
.
var_out
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
saved_mean
);
param_
.
saved_mean
=
var
->
GetMutable
<
Tensor
>
();
var
=
scope
->
FindVar
(
saved_var
);
param_
.
saved_var
=
var
->
GetMutable
<
Tensor
>
();
param_
.
eps
=
op_desc
.
GetAttr
<
float
>
(
"epsilon"
);
return
true
;
}
std
::
string
DebugString
()
const
override
{
return
"batch_norm"
;
}
private:
mutable
BatchNormParam
param_
;
};
}
// namespace operators
}
// namespace lite
}
// namespace paddle
paddle/fluid/lite/operators/batch_norm_op.cc
浏览文件 @
8864f459
...
...
@@ -82,7 +82,7 @@ bool BatchNormOp::AttachImpl(const cpp::OpDesc &op_desc, lite::Scope *scope) {
param_
.
variance
=
scope
->
FindVar
(
op_desc
.
Input
(
"Variance"
).
front
())
->
GetMutable
<
Tensor
>
();
param_
.
y
=
scope
->
FindVar
(
op_desc
.
Output
(
"Y"
).
front
())
->
GetMutable
<
Tensor
>
();
param_
.
is_test
=
op_desc
.
GetAttr
<
bool
>
(
"is_test"
);
param_
.
is_test
=
op_desc
.
GetAttr
<
int
>
(
"is_test"
);
param_
.
use_global_stats
=
op_desc
.
GetAttr
<
bool
>
(
"use_global_stats"
);
if
(
!
param_
.
is_test
)
{
param_
.
mean_out
=
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录