Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
Paddle-Lite
提交
b66052dd
P
Paddle-Lite
项目概览
PaddlePaddle
/
Paddle-Lite
通知
331
Star
4
Fork
1
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
271
列表
看板
标记
里程碑
合并请求
78
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle-Lite
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
271
Issue
271
列表
看板
标记
里程碑
合并请求
78
合并请求
78
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
b66052dd
编写于
3月 22, 2020
作者:
D
DannyIsFunny
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
test=develop
上级
3896590b
变更
7
显示空白变更内容
内联
并排
Showing
7 changed file
with
121 addition
and
24 deletion
+121
-24
lite/api/cxx_api.cc
lite/api/cxx_api.cc
+1
-2
lite/api/cxx_api.h
lite/api/cxx_api.h
+26
-1
lite/api/cxx_api_impl.cc
lite/api/cxx_api_impl.cc
+25
-19
lite/api/cxx_api_test.cc
lite/api/cxx_api_test.cc
+38
-0
lite/core/optimizer.h
lite/core/optimizer.h
+14
-0
lite/core/scope.cc
lite/core/scope.cc
+14
-2
lite/core/scope.h
lite/core/scope.h
+3
-0
未找到文件。
lite/api/cxx_api.cc
浏览文件 @
b66052dd
...
...
@@ -228,7 +228,6 @@ std::vector<const lite::Tensor *> Predictor::GetOutputs() const {
const
cpp
::
ProgramDesc
&
Predictor
::
program_desc
()
const
{
return
program_desc_
;
}
const
RuntimeProgram
&
Predictor
::
runtime_program
()
const
{
return
*
program_
;
}
void
Predictor
::
Build
(
const
lite_api
::
CxxConfig
&
config
,
...
...
@@ -294,7 +293,6 @@ void Predictor::Build(const cpp::ProgramDesc &desc,
inner_places
.
emplace_back
(
TARGET
(
kHost
),
PRECISION
(
kAny
),
DATALAYOUT
(
kAny
));
inner_places
.
emplace_back
(
TARGET
(
kHost
),
PRECISION
(
kFloat
),
DATALAYOUT
(
kNCHW
));
const
std
::
vector
<
std
::
string
>
quant_dequant_op
=
{
"fake_quantize_abs_max"
,
"fake_quantize_range_abs_max"
,
...
...
@@ -321,6 +319,7 @@ void Predictor::Build(const cpp::ProgramDesc &desc,
}
Program
program
(
desc
,
scope_
,
inner_places
);
valid_places_
=
inner_places
;
core
::
KernelPickFactor
factor
;
factor
.
ConsiderTarget
();
...
...
lite/api/cxx_api.h
浏览文件 @
b66052dd
...
...
@@ -46,6 +46,17 @@ class LITE_API Predictor {
// Create a predictor with the weight variable scope set.
explicit
Predictor
(
const
std
::
shared_ptr
<
lite
::
Scope
>&
root_scope
)
:
scope_
(
root_scope
)
{}
Predictor
(
const
cpp
::
ProgramDesc
&
desc
,
const
std
::
shared_ptr
<
Scope
>&
root
,
const
std
::
vector
<
Place
>&
valid_places
)
:
program_desc_
(
desc
),
scope_
(
root
)
{
optimizer_
=
Optimizer
(
new
Program
(
desc
,
scope_
,
valid_places
),
valid_places
);
exec_scope_
=
optimizer_
.
exec_scope
();
GenRuntimeProgram
();
valid_places_
=
valid_places
;
PrepareFeedFetch
();
}
// Build from a model, with places set for hardware config.
void
Build
(
...
...
@@ -67,6 +78,16 @@ class LITE_API Predictor {
const
std
::
vector
<
Place
>&
valid_places
,
const
std
::
vector
<
std
::
string
>&
passes
=
{});
std
::
shared_ptr
<
Predictor
>
Clone
()
const
{
// CHECK(program_desc_) << "Both program and scope of current predicotr
// should be not be nullptr in Clone mode." ;
// CHECK(scope_) << "Both program and scope of current predicotr should
// be not be nullptr in Clone mode.";
auto
predictor
=
std
::
make_shared
<
Predictor
>
(
program_desc_
,
scope_
,
valid_places_
);
return
predictor
;
}
void
GenRuntimeProgram
();
// Run the predictor for a single batch of data.
...
...
@@ -119,11 +140,14 @@ class LITE_API Predictor {
bool
program_generated_
{
false
};
std
::
vector
<
std
::
string
>
input_names_
;
std
::
vector
<
std
::
string
>
output_names_
;
std
::
vector
<
Place
>
valid_places_
;
};
class
CxxPaddleApiImpl
:
public
lite_api
::
PaddlePredictor
{
public:
CxxPaddleApiImpl
()
{}
explicit
CxxPaddleApiImpl
(
const
std
::
shared_ptr
<
Predictor
>&
raw_predictor
)
:
raw_predictor_
(
raw_predictor
)
{}
/// Create a new predictor from a config.
void
Init
(
const
lite_api
::
CxxConfig
&
config
);
...
...
@@ -155,9 +179,10 @@ class CxxPaddleApiImpl : public lite_api::PaddlePredictor {
bool
record_info
=
false
)
override
;
private:
Predictor
raw_predictor_
;
std
::
shared_ptr
<
Predictor
>
raw_predictor_
;
lite_api
::
CxxConfig
config_
;
std
::
mutex
mutex_
;
bool
status_is_cloned_
{
false
};
};
/*
...
...
lite/api/cxx_api_impl.cc
浏览文件 @
b66052dd
...
...
@@ -34,6 +34,7 @@ void CxxPaddleApiImpl::Init(const lite_api::CxxConfig &config) {
#ifdef LITE_WITH_CUDA
Env
<
TARGET
(
kCUDA
)
>::
Init
();
#endif
if
(
!
status_is_cloned_
)
{
auto
places
=
config
.
valid_places
();
std
::
vector
<
std
::
string
>
passes
{};
auto
use_layout_preprocess_pass
=
...
...
@@ -44,7 +45,10 @@ void CxxPaddleApiImpl::Init(const lite_api::CxxConfig &config) {
passes
=
{
"type_layout_cast_preprocess_pass"
};
VLOG
(
1
)
<<
"add pass:"
<<
passes
[
0
];
}
raw_predictor_
.
Build
(
config
,
places
,
passes
);
raw_predictor_
->
Build
(
config
,
places
,
passes
);
}
else
{
CHECK
(
raw_predictor_
)
<<
"The Predictor can not be nullptr in Clone mode."
;
}
mode_
=
config
.
power_mode
();
threads_
=
config
.
threads
();
...
...
@@ -61,34 +65,36 @@ void CxxPaddleApiImpl::Init(const lite_api::CxxConfig &config) {
}
std
::
unique_ptr
<
lite_api
::
Tensor
>
CxxPaddleApiImpl
::
GetInput
(
int
i
)
{
auto
*
x
=
raw_predictor_
.
GetInput
(
i
);
auto
*
x
=
raw_predictor_
->
GetInput
(
i
);
return
std
::
unique_ptr
<
lite_api
::
Tensor
>
(
new
lite_api
::
Tensor
(
x
));
}
std
::
unique_ptr
<
const
lite_api
::
Tensor
>
CxxPaddleApiImpl
::
GetOutput
(
int
i
)
const
{
const
auto
*
x
=
raw_predictor_
.
GetOutput
(
i
);
const
auto
*
x
=
raw_predictor_
->
GetOutput
(
i
);
return
std
::
unique_ptr
<
lite_api
::
Tensor
>
(
new
lite_api
::
Tensor
(
x
));
}
std
::
vector
<
std
::
string
>
CxxPaddleApiImpl
::
GetInputNames
()
{
return
raw_predictor_
.
GetInputNames
();
return
raw_predictor_
->
GetInputNames
();
}
std
::
vector
<
std
::
string
>
CxxPaddleApiImpl
::
GetOutputNames
()
{
return
raw_predictor_
.
GetOutputNames
();
return
raw_predictor_
->
GetOutputNames
();
}
void
CxxPaddleApiImpl
::
Run
()
{
#ifdef LITE_WITH_ARM
lite
::
DeviceInfo
::
Global
().
SetRunMode
(
mode_
,
threads_
);
#endif
raw_predictor_
.
Run
();
raw_predictor_
->
Run
();
}
std
::
shared_ptr
<
lite_api
::
PaddlePredictor
>
CxxPaddleApiImpl
::
Clone
()
{
std
::
lock_guard
<
std
::
mutex
>
lock
(
mutex_
);
auto
predictor
=
std
::
make_shared
<
lite
::
CxxPaddleApiImpl
>
();
auto
predictor
=
std
::
make_shared
<
lite
::
CxxPaddleApiImpl
>
(
raw_predictor_
->
Clone
());
status_is_cloned_
=
true
;
predictor
->
Init
(
config_
);
return
predictor
;
}
...
...
@@ -97,20 +103,20 @@ std::string CxxPaddleApiImpl::GetVersion() const { return version(); }
std
::
unique_ptr
<
const
lite_api
::
Tensor
>
CxxPaddleApiImpl
::
GetTensor
(
const
std
::
string
&
name
)
const
{
auto
*
x
=
raw_predictor_
.
GetTensor
(
name
);
auto
*
x
=
raw_predictor_
->
GetTensor
(
name
);
return
std
::
unique_ptr
<
const
lite_api
::
Tensor
>
(
new
lite_api
::
Tensor
(
x
));
}
std
::
unique_ptr
<
lite_api
::
Tensor
>
CxxPaddleApiImpl
::
GetInputByName
(
const
std
::
string
&
name
)
{
return
std
::
unique_ptr
<
lite_api
::
Tensor
>
(
new
lite_api
::
Tensor
(
raw_predictor_
.
GetInputByName
(
name
)));
new
lite_api
::
Tensor
(
raw_predictor_
->
GetInputByName
(
name
)));
}
void
CxxPaddleApiImpl
::
SaveOptimizedModel
(
const
std
::
string
&
model_dir
,
lite_api
::
LiteModelType
model_type
,
bool
record_info
)
{
raw_predictor_
.
SaveModel
(
model_dir
,
model_type
,
record_info
);
raw_predictor_
->
SaveModel
(
model_dir
,
model_type
,
record_info
);
}
}
// namespace lite
...
...
lite/api/cxx_api_test.cc
浏览文件 @
b66052dd
...
...
@@ -53,6 +53,44 @@ TEST(CXXApi, save_model) {
lite_api
::
LiteModelType
::
kNaiveBuffer
);
}
TEST
(
CXXApi
,
clone_predictor
)
{
lite
::
Predictor
predictor
;
std
::
vector
<
Place
>
valid_places
({
Place
{
TARGET
(
kX86
),
PRECISION
(
kFloat
)}});
predictor
.
Build
(
FLAGS_model_dir
,
""
,
""
,
valid_places
);
auto
cloned_predictor
=
predictor
.
Clone
();
// primary predicotr
auto
*
input_tensor
=
predictor
.
GetInput
(
0
);
input_tensor
->
Resize
(
std
::
vector
<
int64_t
>
({
1
,
100
}));
auto
*
data
=
input_tensor
->
mutable_data
<
float
>
();
for
(
int
i
=
0
;
i
<
100
;
i
++
)
{
data
[
i
]
=
1
;
}
predictor
.
Run
();
auto
*
output_tensor
=
predictor
.
GetOutput
(
0
);
auto
output_shape
=
output_tensor
->
dims
().
Vectorize
();
ASSERT_EQ
(
output_shape
.
size
(),
2
);
ASSERT_EQ
(
output_shape
[
0
],
1
);
ASSERT_EQ
(
output_shape
[
1
],
500
);
// cloned predictor
auto
*
cloned_input_tensor
=
cloned_predictor
->
GetInput
(
0
);
cloned_input_tensor
->
Resize
(
std
::
vector
<
int64_t
>
({
1
,
100
}));
auto
*
cloned_data
=
cloned_input_tensor
->
mutable_data
<
float
>
();
for
(
int
i
=
0
;
i
<
100
;
i
++
)
{
cloned_data
[
i
]
=
1
;
}
cloned_predictor
->
Run
();
auto
*
cloned_output_tensor
=
cloned_predictor
->
GetOutput
(
0
);
int
step
=
50
;
for
(
int
i
=
0
;
i
<
output_tensor
->
data_size
();
i
+=
step
)
{
EXPECT_NEAR
(
output_tensor
->
data
<
float
>
()[
i
],
cloned_output_tensor
->
data
<
float
>
()[
i
],
1e-6
);
}
}
/*TEST(CXXTrainer, train) {
Place place({TARGET(kHost), PRECISION(kFloat), DATALAYOUT(kNCHW)});
std::vector<Place> valid_places({place});
...
...
lite/core/optimizer.h
浏览文件 @
b66052dd
...
...
@@ -37,6 +37,20 @@ namespace lite {
*/
class
Optimizer
{
public:
Optimizer
()
{}
Optimizer
(
Program
*
program
,
const
std
::
vector
<
Place
>&
valid_places
)
{
program_
=
program
;
valid_places_
=
valid_places
;
CHECK
(
!
valid_places
.
empty
())
<<
"At least one valid_place should be set"
;
CHECK
(
!
graph_
)
<<
"duplicate optimize found"
;
graph_
.
reset
(
new
mir
::
SSAGraph
);
graph_
->
Build
(
*
program
,
valid_places
);
graph_
->
SetValidPlaces
(
valid_places
);
exec_scope_
=
program
->
exec_scope
();
}
void
Run
(
Program
&&
program
,
const
std
::
vector
<
Place
>&
valid_places
,
core
::
KernelPickFactor
kernel_pick_factor
,
...
...
lite/core/scope.cc
浏览文件 @
b66052dd
...
...
@@ -13,11 +13,16 @@
// limitations under the License.
#include "lite/core/scope.h"
#define SCOPE_KIDS_READER_LOCK lite::fluid::AutoRDLock auto_lock(&kids_lock_);
#define SCOPE_KIDS_WRITER_LOCK lite::fluid::AutoWRLock auto_lock(&kids_lock_);
#define SCOPE_VARS_READER_LOCK lite::fluid::AutoRDLock auto_lock(&vars_lock_);
#define SCOPE_VARS_WRITER_LOCK lite::fluid::AutoWRLock auto_lock(&vars_lock_);
namespace
paddle
{
namespace
lite
{
Scope
::~
Scope
()
{
SCOPE_KIDS_WRITER_LOCK
for
(
auto
*
x
:
kids_
)
{
if
(
x
)
{
delete
x
;
...
...
@@ -26,12 +31,14 @@ Scope::~Scope() {
}
Scope
&
Scope
::
NewScope
()
const
{
SCOPE_KIDS_WRITER_LOCK
kids_
.
push_back
(
new
Scope
);
kids_
.
back
()
->
parent_
=
this
;
return
*
kids_
.
back
();
}
Variable
*
Scope
::
Var
(
const
std
::
string
&
name
)
{
SCOPE_VARS_WRITER_LOCK
auto
*
var
=
FindVar
(
name
);
if
(
var
)
return
var
;
...
...
@@ -45,6 +52,7 @@ Variable *Scope::FindVar(const std::string &name) const {
var
=
FindLocalVar
(
name
);
const
Scope
*
cur_scope
=
this
;
while
(
!
var
&&
cur_scope
->
parent
())
{
// SCOPE_VARS_READER_LOCK
cur_scope
=
cur_scope
->
parent
();
var
=
cur_scope
->
FindLocalVar
(
name
);
}
...
...
@@ -53,6 +61,7 @@ Variable *Scope::FindVar(const std::string &name) const {
}
Variable
*
Scope
::
FindLocalVar
(
const
std
::
string
&
name
)
const
{
// SCOPE_VARS_READER_LOCK
auto
it
=
vars_
.
find
(
name
);
if
(
it
!=
vars_
.
end
())
{
return
it
->
second
.
get
();
...
...
@@ -62,9 +71,12 @@ Variable *Scope::FindLocalVar(const std::string &name) const {
std
::
vector
<
std
::
string
>
Scope
::
LocalVarNames
()
const
{
std
::
vector
<
std
::
string
>
keys
;
{
// SCOPE_VARS_READER_LOCK
for
(
const
auto
&
item
:
vars_
)
{
keys
.
push_back
(
item
.
first
);
}
}
return
keys
;
}
...
...
lite/core/scope.h
浏览文件 @
b66052dd
...
...
@@ -20,6 +20,7 @@
#include <utility>
#include <vector>
#include "lite/core/variable.h"
#include "lite/fluid/rw_lock.h"
namespace
paddle
{
namespace
lite
{
...
...
@@ -73,6 +74,8 @@ class Scope final {
mutable
std
::
list
<
Scope
*>
kids_
;
const
Scope
*
parent_
{
nullptr
};
std
::
unordered_map
<
std
::
string
,
std
::
unique_ptr
<
Variable
>>
vars_
;
mutable
lite
::
fluid
::
RWLock
kids_lock_
;
mutable
lite
::
fluid
::
RWLock
vars_lock_
;
};
}
// namespace lite
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录