Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
43036686
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
43036686
编写于
4月 20, 2019
作者:
S
Superjomn
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
init io complement pass
上级
64f10504
变更
15
隐藏空白更改
内联
并排
Showing
15 changed file
with
140 addition
and
32 deletion
+140
-32
paddle/fluid/lite/api/cxx_api.h
paddle/fluid/lite/api/cxx_api.h
+8
-17
paddle/fluid/lite/core/kernel.h
paddle/fluid/lite/core/kernel.h
+0
-1
paddle/fluid/lite/core/mir/CMakeLists.txt
paddle/fluid/lite/core/mir/CMakeLists.txt
+1
-1
paddle/fluid/lite/core/mir/generate_program_pass.cc
paddle/fluid/lite/core/mir/generate_program_pass.cc
+2
-1
paddle/fluid/lite/core/mir/generate_program_pass.h
paddle/fluid/lite/core/mir/generate_program_pass.h
+6
-2
paddle/fluid/lite/core/mir/io_complement_pass.cc
paddle/fluid/lite/core/mir/io_complement_pass.cc
+22
-2
paddle/fluid/lite/core/mir/node.h
paddle/fluid/lite/core/mir/node.h
+5
-0
paddle/fluid/lite/core/op_executor.cc
paddle/fluid/lite/core/op_executor.cc
+6
-0
paddle/fluid/lite/core/op_executor.h
paddle/fluid/lite/core/op_executor.h
+14
-0
paddle/fluid/lite/core/op_lite.cc
paddle/fluid/lite/core/op_lite.cc
+20
-0
paddle/fluid/lite/core/op_lite.h
paddle/fluid/lite/core/op_lite.h
+2
-0
paddle/fluid/lite/core/optimizer.h
paddle/fluid/lite/core/optimizer.h
+6
-2
paddle/fluid/lite/core/optimizer_test.cc
paddle/fluid/lite/core/optimizer_test.cc
+2
-6
paddle/fluid/lite/core/program.h
paddle/fluid/lite/core/program.h
+39
-0
paddle/fluid/lite/core/target_wrapper.h
paddle/fluid/lite/core/target_wrapper.h
+7
-0
未找到文件。
paddle/fluid/lite/api/cxx_api.h
浏览文件 @
43036686
...
...
@@ -15,6 +15,8 @@
#pragma once
#include "paddle/fluid/lite/core/op_executor.h"
#include "paddle/fluid/lite/core/op_lite.h"
#include "paddle/fluid/lite/core/optimizer.h"
#include "paddle/fluid/lite/core/program.h"
#include "paddle/fluid/lite/model_parser/model_parser.h"
namespace
paddle
{
...
...
@@ -28,34 +30,23 @@ class Predictor {
void
Build
(
const
std
::
string
&
model_path
,
const
std
::
vector
<
Place
>&
valid_places
)
{
CHECK
(
!
executor_
.
get
())
<<
"duplicate build found"
;
CHECK
(
!
scope_
.
get
())
<<
"duplicate build found"
;
framework
::
proto
::
ProgramDesc
prog
;
LoadModel
(
model_path
,
scope_
.
get
(),
&
prog
);
framework
::
ProgramDesc
prog_desc
(
prog
);
executor_
.
reset
(
new
Executor
(
prog_desc
,
scope_
.
get
(),
valid_places
));
}
// Get a tensor for input from scope directly.
Tensor
*
GetInputTensor
(
const
std
::
string
&
name
)
{
auto
*
var
=
executor_
->
exec_scope
()
->
FindVar
(
name
);
CHECK
(
var
)
<<
"no tensor called "
<<
name
<<
" exists"
;
return
var
->
GetMutable
<
Tensor
>
();
}
Program
program
(
prog_desc
,
scope_
,
valid_places
);
// Get a tensor for output from scope directly.
const
Tensor
*
GetOutputTensor
(
const
std
::
string
&
name
)
{
auto
*
var
=
executor_
->
exec_scope
()
->
FindVar
(
name
);
CHECK
(
var
)
<<
"no tensor called "
<<
name
<<
" exists"
;
return
&
var
->
Get
<
Tensor
>
();
Optimizer
optimizer
;
optimizer
.
Run
(
std
::
move
(
program
),
valid_places
);
program_
=
optimizer
.
GenRuntimeProgram
();
}
void
Run
()
{
executor
_
->
Run
();
}
void
Run
()
{
program
_
->
Run
();
}
private:
std
::
shared_ptr
<
Scope
>
scope_
;
std
::
unique_ptr
<
lite
::
Executor
>
executor
_
;
std
::
unique_ptr
<
RuntimeProgram
>
program
_
;
};
}
// namespace lite
...
...
paddle/fluid/lite/core/kernel.h
浏览文件 @
43036686
...
...
@@ -177,7 +177,6 @@ class ParamTypeRegistry {
const
ParamType
*
Retrieve
(
const
Place
&
place
,
const
std
::
string
&
op_type
,
const
std
::
string
&
arg_name
)
{
KernelIdTy
key
{
op_type
,
place
,
io
,
arg_name
};
LOG
(
INFO
)
<<
"Looking for "
<<
key
;
auto
it
=
types_
.
find
(
key
);
if
(
it
==
types_
.
end
())
return
nullptr
;
return
&
it
->
second
;
...
...
paddle/fluid/lite/core/mir/CMakeLists.txt
浏览文件 @
43036686
...
...
@@ -5,10 +5,10 @@ cc_library(mir_pass_manager SRCS pass_manager.cc DEPS mir_pass mir_ssa_graph)
cc_library
(
mir_pass_registry SRCS pass_registry.cc DEPS mir_pass_manager
)
cc_library
(
mir_passes
SRCS static_kernel_pick_pass.cc
variable_place_inference_pass.cc
io_complement_pass.cc
graph_visualize_pass.cc
generate_program_pass.cc
variable_place_inference_pass.cc
demo_pass.cc
DEPS mir_pass types_lite
)
...
...
paddle/fluid/lite/core/mir/generate_program_pass.cc
浏览文件 @
43036686
...
...
@@ -23,7 +23,8 @@ void GenerateProgramPass::Apply(std::unique_ptr<mir::SSAGraph>& graph) {
for
(
auto
&
item
:
graph
->
InstructTopologicalOrder
())
{
if
(
item
->
IsInstruct
())
{
auto
&
instruct
=
item
->
AsInstruct
();
kernels_
.
emplace_back
(
std
::
move
(
instruct
.
valid_kernels
.
front
()));
insts_
.
emplace_back
(
instruct
.
op
,
std
::
move
(
instruct
.
valid_kernels
.
front
()));
}
}
}
...
...
paddle/fluid/lite/core/mir/generate_program_pass.h
浏览文件 @
43036686
...
...
@@ -30,10 +30,14 @@ class GenerateProgramPass : public ProgramPass {
public:
void
Apply
(
std
::
unique_ptr
<
mir
::
SSAGraph
>&
graph
)
override
;
std
::
list
<
std
::
unique_ptr
<
KernelBase
>>&
kernels
()
{
return
kernels_
;
}
std
::
unique_ptr
<
RuntimeProgram
>
GenProgram
()
{
std
::
unique_ptr
<
RuntimeProgram
>
program
(
new
RuntimeProgram
(
std
::
move
(
insts_
)));
return
program
;
}
private:
std
::
list
<
std
::
unique_ptr
<
KernelBase
>>
kernel
s_
;
std
::
vector
<
Instruction
>
inst
s_
;
};
}
// namespace mir
...
...
paddle/fluid/lite/core/mir/io_complement_pass.cc
浏览文件 @
43036686
...
...
@@ -19,8 +19,28 @@ namespace paddle {
namespace
lite
{
namespace
mir
{
void
IoComplementPass
::
Apply
(
std
::
unique_ptr
<
mir
::
SSAGraph
>
&
graph
)
{
// Start from inputs of the graph, those should should have place set.
void
IoComplementPass
::
Apply
(
std
::
unique_ptr
<
mir
::
SSAGraph
>&
graph
)
{
// Start from inputs of the graph, those should have place set.
for
(
auto
&
node
:
graph
->
mutable_nodes
())
{
if
(
!
node
.
IsInstruct
())
continue
;
auto
&
inst
=
node
.
AsInstruct
();
// inputs
for
(
auto
*
in
:
node
.
inlinks
)
{
CHECK
(
in
->
IsArgument
());
auto
name
=
in
->
AsArgument
().
name
;
std
::
string
tmp
;
CHECK
(
inst
.
op_info
->
GetInputArgname
(
name
,
&
tmp
));
auto
type
=
ParamTypeRegistry
::
Global
().
Retrieve
<
ParamTypeRegistry
::
IO
::
kInput
>
(
inst
.
place
,
inst
.
op_type
,
tmp
);
CHECK
(
type
)
<<
"no param type found for "
<<
inst
.
op_type
<<
":"
<<
name
<<
" "
<<
inst
.
place
;
if
(
type
->
tensor_place
!=
inst
.
place
)
{
LOG
(
INFO
)
<<
"found IO unmatched tensor"
;
}
}
}
}
}
// namespace mir
...
...
paddle/fluid/lite/core/mir/node.h
浏览文件 @
43036686
...
...
@@ -48,6 +48,11 @@ class Node {
std
::
shared_ptr
<
OpInfo
>
op_info
;
// TODO(Superjomn) make this a shared_ptr for resource safety.
std
::
shared_ptr
<
OpLite
>
op
;
// we hold op to run InferShape
KernelBase
&
picked_kernel
()
{
CHECK
(
!
valid_kernels
.
empty
());
return
*
valid_kernels
.
front
();
}
};
struct
Argument
{
...
...
paddle/fluid/lite/core/op_executor.cc
浏览文件 @
43036686
...
...
@@ -11,3 +11,9 @@
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "paddle/fluid/lite/core/op_executor.h"
namespace
paddle
{
namespace
lite
{}
// namespace lite
}
// namespace paddle
paddle/fluid/lite/core/op_executor.h
浏览文件 @
43036686
...
...
@@ -17,6 +17,7 @@
#include "paddle/fluid/lite/core/op_lite.h"
#include "paddle/fluid/lite/core/op_registry.h"
#include "paddle/fluid/lite/core/program.h"
#include "paddle/fluid/lite/core/program.h"
#include "paddle/fluid/lite/core/scope.h"
namespace
paddle
{
...
...
@@ -50,5 +51,18 @@ class Executor {
std
::
unique_ptr
<
Program
>
program_
;
};
class
RuntimeExecutor
{
public:
RuntimeExecutor
(
RuntimeProgram
*
program
)
:
program_
(
program
)
{}
void
Run
()
{
CHECK
(
program_
);
program_
->
Run
();
}
private:
RuntimeProgram
*
program_
{};
};
}
// namespace lite
}
// namespace paddle
paddle/fluid/lite/core/op_lite.cc
浏览文件 @
43036686
...
...
@@ -61,5 +61,25 @@ bool OpLite::Attach(const framework::OpDesc &opdesc, lite::Scope *scope) {
return
AttachImpl
(
opdesc
,
scope
);
}
bool
OpInfo
::
GetInputArgname
(
const
std
::
string
&
value_name
,
std
::
string
*
out
)
{
for
(
auto
&
item
:
input_argument_
)
{
auto
it
=
std
::
find
(
item
.
second
.
begin
(),
item
.
second
.
end
(),
value_name
);
if
(
it
!=
item
.
second
.
end
())
{
*
out
=
item
.
first
;
return
true
;
}
}
return
false
;
}
bool
OpInfo
::
GetOutputArgname
(
const
std
::
string
&
value_name
,
std
::
string
*
out
)
{
for
(
auto
&
item
:
output_argument_
)
{
auto
it
=
std
::
find
(
item
.
second
.
begin
(),
item
.
second
.
end
(),
value_name
);
if
(
it
!=
item
.
second
.
end
())
{
*
out
=
item
.
first
;
return
true
;
}
}
return
false
;
}
}
// namespace lite
}
// namespace paddle
paddle/fluid/lite/core/op_lite.h
浏览文件 @
43036686
...
...
@@ -151,6 +151,8 @@ class OpInfo {
const
std
::
map
<
std
::
string
,
std
::
list
<
std
::
string
>>
&
output_argument
()
{
return
output_argument_
;
}
bool
GetInputArgname
(
const
std
::
string
&
value_name
,
std
::
string
*
out
);
bool
GetOutputArgname
(
const
std
::
string
&
value_name
,
std
::
string
*
out
);
const
std
::
list
<
std
::
string
>
&
input_argnames
()
const
{
return
input_argnames_
;
...
...
paddle/fluid/lite/core/optimizer.h
浏览文件 @
43036686
...
...
@@ -15,8 +15,10 @@
#pragma once
#include <string>
#include <vector>
#include "paddle/fluid/lite/core/mir/generate_program_pass.h"
#include "paddle/fluid/lite/core/mir/pass_manager.h"
#include "paddle/fluid/lite/core/mir/ssa_graph.h"
#include "paddle/fluid/lite/core/program.h"
namespace
paddle
{
namespace
lite
{
...
...
@@ -36,9 +38,11 @@ class Optimizer {
}
// Generate a new program based on the mir graph.
std
::
unique_ptr
<
Program
>
Gen
Program
()
{
std
::
unique_ptr
<
RuntimeProgram
>
GenRuntime
Program
()
{
std
::
unique_ptr
<
Program
>
res
;
return
res
;
auto
pass
=
mir
::
PassManager
::
Global
().
LookUp
<
mir
::
GenerateProgramPass
>
(
"generate_program_pass"
);
return
pass
->
GenProgram
();
}
// Generate C++ code which combines the inference program, model and weights.
...
...
paddle/fluid/lite/core/optimizer_test.cc
浏览文件 @
43036686
...
...
@@ -37,12 +37,8 @@ TEST(Optimizer, test) {
.
ConsiderPrecision
();
optimizer
.
Run
(
std
::
move
(
program
),
places
);
auto
*
program_pass
=
mir
::
PassManager
::
Global
().
LookUp
<
mir
::
GenerateProgramPass
>
(
"generate_program_pass"
);
auto
&
kernels
=
program_pass
->
kernels
();
LOG
(
INFO
)
<<
"get kernels: "
<<
kernels
.
size
();
auto
runtime_program
=
optimizer
.
GenRuntimeProgram
();
LOG
(
INFO
)
<<
"num instructions "
<<
runtime_program
->
num_instructions
();
}
}
// namespace lite
...
...
paddle/fluid/lite/core/program.h
浏览文件 @
43036686
...
...
@@ -18,6 +18,7 @@
#include <vector>
#include "paddle/fluid/framework/program_desc.h"
#include "paddle/fluid/lite/core/kernel.h"
#include "paddle/fluid/lite/core/kernel.h"
#include "paddle/fluid/lite/core/op_lite.h"
#include "paddle/fluid/lite/core/op_registry.h"
...
...
@@ -86,5 +87,43 @@ struct Program {
}
};
struct
Instruction
{
Instruction
(
const
std
::
shared_ptr
<
OpLite
>&
op
,
std
::
unique_ptr
<
KernelBase
>&&
kernel
)
:
op_
(
op
),
kernel_
(
std
::
move
(
kernel
))
{}
void
Run
()
{
CHECK
(
op_
);
CHECK
(
kernel_
);
op_
->
InferShape
();
kernel_
->
Run
();
}
private:
std
::
shared_ptr
<
OpLite
>
op_
;
std
::
unique_ptr
<
KernelBase
>
kernel_
;
};
/*
* A program contains kernels for runtime.
*/
class
RuntimeProgram
{
public:
explicit
RuntimeProgram
(
std
::
vector
<
Instruction
>&&
instruction
)
:
instructions_
(
std
::
move
(
instruction
))
{}
void
Run
()
{
for
(
auto
&
inst
:
instructions_
)
{
inst
.
Run
();
}
}
size_t
num_instructions
()
const
{
return
instructions_
.
size
();
}
private:
RuntimeProgram
(
const
RuntimeProgram
&
)
=
delete
;
std
::
vector
<
Instruction
>
instructions_
;
};
}
// namespace lite
}
// namespace paddle
paddle/fluid/lite/core/target_wrapper.h
浏览文件 @
43036686
...
...
@@ -84,6 +84,8 @@ struct Place {
layout
==
other
.
layout
&&
device
==
other
.
device
;
}
bool
operator
!=
(
const
Place
&
other
)
const
{
return
!
(
*
this
==
other
);
}
friend
bool
operator
<
(
const
Place
&
a
,
const
Place
&
b
)
{
if
(
a
.
target
!=
b
.
target
)
return
a
.
target
<
b
.
target
;
if
(
a
.
precision
!=
b
.
precision
)
return
a
.
precision
<
b
.
precision
;
...
...
@@ -92,6 +94,11 @@ struct Place {
return
true
;
}
friend
std
::
ostream
&
operator
<<
(
std
::
ostream
&
os
,
const
Place
&
other
)
{
os
<<
other
.
DebugString
();
return
os
;
}
std
::
string
DebugString
()
const
{
std
::
stringstream
os
;
os
<<
TargetToStr
(
target
)
<<
"/"
<<
PrecisionToStr
(
precision
)
<<
"/"
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录