Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
PaddleDetection
提交
e5155713
P
PaddleDetection
项目概览
PaddlePaddle
/
PaddleDetection
大约 1 年 前同步成功
通知
694
Star
11112
Fork
2696
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
184
列表
看板
标记
里程碑
合并请求
40
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
PaddleDetection
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
184
Issue
184
列表
看板
标记
里程碑
合并请求
40
合并请求
40
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
体验新版 GitCode,发现更多精彩内容 >>
提交
e5155713
编写于
10月 09, 2017
作者:
Y
Yang Yang
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
clean up for review
上级
089cc11d
变更
7
隐藏空白更改
内联
并排
Showing
7 changed file
with
56 addition
and
51 deletion
+56
-51
paddle/framework/executor.cc
paddle/framework/executor.cc
+27
-13
paddle/framework/executor.h
paddle/framework/executor.h
+1
-1
paddle/framework/executor_test.cc
paddle/framework/executor_test.cc
+24
-36
paddle/framework/scope.cc
paddle/framework/scope.cc
+1
-0
paddle/operators/feed_op.cc
paddle/operators/feed_op.cc
+1
-0
paddle/operators/fetch_op.cc
paddle/operators/fetch_op.cc
+1
-0
paddle/platform/gpu_info.cc
paddle/platform/gpu_info.cc
+1
-1
未找到文件。
paddle/framework/executor.cc
浏览文件 @
e5155713
...
...
@@ -13,11 +13,13 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/framework/executor.h"
#include <algorithm>
#include <iostream>
#include <memory>
#include <set>
#include <vector>
#include "paddle/framework/lod_tensor.h"
#include "paddle/framework/op_registry.h"
#include "paddle/framework/scope.h"
...
...
@@ -27,7 +29,11 @@ limitations under the License. */
namespace
paddle
{
namespace
framework
{
const
std
::
string
kFeedOpType
=
"feed"
;
const
std
::
string
kFetchOpType
=
"fetch"
;
Executor
::
Executor
(
const
std
::
vector
<
platform
::
Place
>&
places
)
{
PADDLE_ENFORCE_GT
(
places
.
size
(),
0
);
device_contexts_
.
resize
(
places
.
size
());
for
(
size_t
i
=
0
;
i
<
places
.
size
();
i
++
)
{
if
(
platform
::
is_cpu_place
(
places
[
i
]))
{
...
...
@@ -46,9 +52,7 @@ Executor::Executor(const std::vector<platform::Place>& places) {
Executor
::~
Executor
()
{
for
(
auto
&
device_context
:
device_contexts_
)
{
if
(
device_context
)
{
delete
device_context
;
}
delete
device_context
;
}
}
...
...
@@ -56,6 +60,8 @@ void Executor::Run(const ProgramDesc& pdesc, Scope* scope) {
// TODO(tonyyang-svail):
// - only runs the first block (i.e. no RNN support)
// - only runs on the first device (i.e. no interdevice communication)
// - will change to use multiple blocks for RNN op and Cond Op
PADDLE_ENFORCE_GT
(
pdesc
.
blocks_size
(),
0
);
auto
&
block
=
pdesc
.
blocks
(
0
);
auto
&
device
=
device_contexts_
[
0
];
...
...
@@ -66,12 +72,12 @@ void Executor::Run(const ProgramDesc& pdesc, Scope* scope) {
Scope
&
local_scope
=
scope
->
NewScope
();
std
::
vector
<
bool
>
should_run
=
Pr
eprocess
(
pdesc
);
PADDLE_ENFORCE
(
should_run
.
size
()
==
block
.
ops_size
());
std
::
vector
<
bool
>
should_run
=
Pr
une
(
pdesc
);
PADDLE_ENFORCE
_EQ
(
should_run
.
size
(),
block
.
ops_size
());
for
(
size_t
i
=
0
;
i
<
should_run
.
size
();
++
i
)
{
if
(
should_run
[
i
])
{
for
(
auto
var
:
block
.
ops
(
i
).
outputs
())
{
for
(
auto
argu
:
var
.
arguments
())
{
for
(
auto
&
var
:
block
.
ops
(
i
).
outputs
())
{
for
(
auto
&
argu
:
var
.
arguments
())
{
if
(
local_scope
.
FindVar
(
argu
)
==
nullptr
)
{
local_scope
.
NewVar
(
argu
);
}
...
...
@@ -81,28 +87,32 @@ void Executor::Run(const ProgramDesc& pdesc, Scope* scope) {
op
->
Run
(
local_scope
,
*
device
);
}
}
// TODO(tonyyang-svail):
// - Destroy local_scope
}
std
::
vector
<
bool
>
Executor
::
Pr
eprocess
(
const
ProgramDesc
&
pdesc
)
{
std
::
vector
<
bool
>
Executor
::
Pr
une
(
const
ProgramDesc
&
pdesc
)
{
// TODO(tonyyang-svail):
// - only runs the first block
// - will change to use multiple blocks for RNN op and Cond Op
auto
&
block
=
pdesc
.
blocks
(
0
);
auto
&
ops
=
block
.
ops
();
bool
expect_feed
=
true
;
for
(
auto
&
op_desc
:
ops
)
{
PADDLE_ENFORCE
(
op_desc
.
type
()
!=
"feed"
||
expect_feed
,
PADDLE_ENFORCE
(
op_desc
.
type
()
!=
kFeedOpType
||
expect_feed
,
"All FeedOps are at the beginning of the ProgramDesc"
);
expect_feed
=
(
op_desc
.
type
()
==
"feed"
);
expect_feed
=
(
op_desc
.
type
()
==
kFeedOpType
);
}
bool
expect_fetch
=
true
;
for
(
auto
op_iter
=
ops
.
rbegin
();
op_iter
!=
ops
.
rend
();
++
op_iter
)
{
auto
&
op_desc
=
*
op_iter
;
PADDLE_ENFORCE
(
op_desc
.
type
()
!=
"fetch"
||
expect_fetch
,
PADDLE_ENFORCE
(
op_desc
.
type
()
!=
kFetchOpType
||
expect_fetch
,
"All FetchOps must at the end of the ProgramDesc"
);
expect_fetch
=
(
op_desc
.
type
()
==
"fetch"
);
expect_fetch
=
(
op_desc
.
type
()
==
kFetchOpType
);
}
std
::
set
<
std
::
string
>
dependent_vars
;
...
...
@@ -119,7 +129,7 @@ std::vector<bool> Executor::Preprocess(const ProgramDesc& pdesc) {
}
}
if
(
op_desc
.
type
()
==
"fetch"
||
found_dependent_vars
)
{
if
(
op_desc
.
type
()
==
kFetchOpType
||
found_dependent_vars
)
{
// erase its output to the dependency graph
for
(
auto
&
var
:
op_desc
.
outputs
())
{
for
(
auto
&
argu
:
var
.
arguments
())
{
...
...
@@ -140,6 +150,10 @@ std::vector<bool> Executor::Preprocess(const ProgramDesc& pdesc) {
}
}
// TODO(tonyyang-svail):
// - check this after integration of Init
// PADDLE_ENFORCE(dependent_vars.empty());
// since we are traversing the ProgramDesc in reverse order
// we reverse the should_run vector
std
::
reverse
(
should_run
.
begin
(),
should_run
.
end
());
...
...
paddle/framework/executor.h
浏览文件 @
e5155713
...
...
@@ -46,7 +46,7 @@ class Executor {
* @return
* vector<bool> Same size as ops. Indicates whether an op should be run.
*/
std
::
vector
<
bool
>
Pr
eprocess
(
const
ProgramDesc
&
pdesc
);
std
::
vector
<
bool
>
Pr
une
(
const
ProgramDesc
&
pdesc
);
private:
std
::
vector
<
platform
::
DeviceContext
*>
device_contexts_
;
...
...
paddle/framework/executor_test.cc
浏览文件 @
e5155713
...
...
@@ -13,12 +13,14 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/framework/executor.h"
#include <memory>
#include <vector>
#include "gtest/gtest.h"
#include "paddle/framework/attribute.h"
#include "paddle/framework/backward.h"
#include "paddle/framework/block_desc.h"
// #include "paddle/framework/grad_op_builder.h"
#include "paddle/framework/op_desc.h"
#include "paddle/framework/op_registry.h"
#include "paddle/framework/operator.h"
...
...
@@ -34,9 +36,6 @@ using std::string;
using
namespace
paddle
::
platform
;
using
namespace
paddle
::
framework
;
typedef
paddle
::
framework
::
BlockDesc
proto_block
;
typedef
paddle
::
framework
::
OpDesc
proto_op
;
void
AddOp
(
const
std
::
string
&
type
,
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
AttributeMap
attrs
,
paddle
::
framework
::
BlockDescBind
*
block
)
{
...
...
@@ -51,10 +50,10 @@ void AddOp(const std::string& type, const VariableNameMap& inputs,
// insert op
auto
op
=
block
->
AppendOp
();
op
->
SetType
(
type
);
for
(
auto
kv
:
inputs
)
{
for
(
auto
&
kv
:
inputs
)
{
op
->
SetInput
(
kv
.
first
,
kv
.
second
);
}
for
(
auto
kv
:
outputs
)
{
for
(
auto
&
kv
:
outputs
)
{
op
->
SetOutput
(
kv
.
first
,
kv
.
second
);
}
op
->
SetAttrMap
(
attrs
);
...
...
@@ -65,11 +64,11 @@ std::once_flag set_variable_flag;
// Tensors in feed value variable will only be in CPUPlace
// So we can memcpy the data from vector<T> to feed_value
template
<
typename
T
>
void
set_feed_v
ariable
(
const
std
::
vector
<
std
::
vector
<
T
>>&
inputs
)
{
void
SetFeedV
ariable
(
const
std
::
vector
<
std
::
vector
<
T
>>&
inputs
)
{
typedef
std
::
vector
<
paddle
::
framework
::
Tensor
>
FeedInputs
;
Variable
*
g_feed_value
=
GetGlobalScope
()
->
FindVar
(
"feed_value"
);
FeedInputs
&
feed_inputs
=
*
(
g_feed_value
->
GetMutable
<
FeedInputs
>
());
auto
size
=
inputs
.
size
();
size_t
size
=
inputs
.
size
();
feed_inputs
.
resize
(
size
);
for
(
size_t
i
=
0
;
i
<
size
;
i
++
)
{
T
*
dst
=
feed_inputs
[
i
].
mutable_data
<
T
>
(
...
...
@@ -81,12 +80,12 @@ void set_feed_variable(const std::vector<std::vector<T>>& inputs) {
// Tensors in fetch value variable will only be in CPUPlace
// So we can memcpy the data from fetch_value to vector<T>
template
<
typename
T
>
std
::
vector
<
std
::
vector
<
T
>>
get_fetch_v
ariable
()
{
std
::
vector
<
std
::
vector
<
T
>>
GetFetchV
ariable
()
{
typedef
std
::
vector
<
paddle
::
framework
::
Tensor
>
FetchOutputs
;
Variable
*
g_fetch_value
=
GetGlobalScope
()
->
FindVar
(
"fetch_value"
);
FetchOutputs
&
fetch_outputs
=
*
(
g_fetch_value
->
GetMutable
<
FetchOutputs
>
());
auto
size
=
fetch_outputs
.
size
();
size_t
size
=
fetch_outputs
.
size
();
std
::
vector
<
std
::
vector
<
T
>>
result
;
result
.
reserve
(
size
);
for
(
size_t
i
=
0
;
i
<
size
;
i
++
)
{
...
...
@@ -105,7 +104,7 @@ class ExecutorTesterRandom : public ::testing::Test {
virtual
void
SetUp
()
override
{
int
input_dim
=
5
,
batch_size
=
2
,
embed_dim
=
5
;
// init pdesc
-----------------------------------------
// init pdesc
auto
temp_init_root_block
=
init_pdesc_
.
add_blocks
();
temp_init_root_block
->
set_idx
(
0
);
temp_init_root_block
->
set_parent_idx
(
-
1
);
...
...
@@ -128,7 +127,7 @@ class ExecutorTesterRandom : public ::testing::Test {
// flush
init_program
.
Proto
();
// run pdesc
-----------------------------------------
// run pdesc
auto
temp_root_block
=
pdesc_
.
add_blocks
();
temp_root_block
->
set_idx
(
0
);
temp_root_block
->
set_parent_idx
(
-
1
);
...
...
@@ -154,9 +153,6 @@ class ExecutorTesterRandom : public ::testing::Test {
// TODO(tonyyang-svail):
// - Test with Backward
// AddOp("gaussian_random", {}, {{"Out", {"l2_distance@GRAD"}}},
// {{"dims", std::vector<int>{batch_size, 1}}}, root_block);
// AppendBackward(program, {});
}
protected:
...
...
@@ -213,12 +209,11 @@ TEST_F(ExecutorTesterRandom, CPU) {
// "pointer being freed was not allocated" error will appear.
paddle
::
memory
::
Used
(
cpu_place
);
Executor
*
executor
=
new
Executor
(
places
);
std
::
unique_ptr
<
Executor
>
executor
(
new
Executor
(
places
));
executor
->
Run
(
init_pdesc_
,
GetGlobalScope
());
executor
->
Run
(
pdesc_
,
GetGlobalScope
());
std
::
vector
<
std
::
vector
<
float
>>
result
=
get_fetch_variable
<
float
>
();
delete
executor
;
std
::
vector
<
std
::
vector
<
float
>>
result
=
GetFetchVariable
<
float
>
();
}
TEST_F
(
ExecutorTesterFeedAndFetch
,
CPU
)
{
...
...
@@ -232,13 +227,12 @@ TEST_F(ExecutorTesterFeedAndFetch, CPU) {
// "pointer being freed was not allocated" error will appear.
paddle
::
memory
::
Used
(
cpu_place
);
Executor
*
executor
=
new
Executor
(
places
);
std
::
unique_ptr
<
Executor
>
executor
(
new
Executor
(
places
)
);
// 3 mini-batch
for
(
int
i
=
0
;
i
<
3
;
i
++
)
{
set_feed_variable
<
float
>
(
inputs_
);
for
(
int
batch_id
=
0
;
batch_id
<
3
;
batch_id
++
)
{
SetFeedVariable
<
float
>
(
inputs_
);
executor
->
Run
(
pdesc_
,
GetGlobalScope
());
std
::
vector
<
std
::
vector
<
float
>>
result
=
get_fetch_v
ariable
<
float
>
();
std
::
vector
<
std
::
vector
<
float
>>
result
=
GetFetchV
ariable
<
float
>
();
PADDLE_ENFORCE_EQ
(
result
.
size
(),
inputs_
.
size
());
for
(
size_t
i
=
0
;
i
<
result
.
size
();
++
i
)
{
PADDLE_ENFORCE_EQ
(
result
[
i
].
size
(),
inputs_
[
i
].
size
());
...
...
@@ -247,8 +241,6 @@ TEST_F(ExecutorTesterFeedAndFetch, CPU) {
}
}
}
delete
executor
;
}
#else
TEST_F
(
ExecutorTesterRandom
,
GPU
)
{
...
...
@@ -265,13 +257,11 @@ TEST_F(ExecutorTesterRandom, GPU) {
paddle
::
memory
::
Used
(
CPUPlace
());
paddle
::
memory
::
Used
(
gpu_place
);
Executor
*
executor
=
new
Executor
(
places
);
std
::
unique_ptr
<
Executor
>
executor
(
new
Executor
(
places
)
);
executor
->
Run
(
init_pdesc_
,
GetGlobalScope
());
executor
->
Run
(
pdesc_
,
GetGlobalScope
());
std
::
vector
<
std
::
vector
<
float
>>
result
=
get_fetch_variable
<
float
>
();
delete
executor
;
std
::
vector
<
std
::
vector
<
float
>>
result
=
GetFetchVariable
<
float
>
();
}
TEST_F
(
ExecutorTesterFeedAndFetch
,
GPU
)
{
...
...
@@ -287,13 +277,12 @@ TEST_F(ExecutorTesterFeedAndFetch, GPU) {
paddle
::
memory
::
Used
(
CPUPlace
());
paddle
::
memory
::
Used
(
gpu_place
);
Executor
*
executor
=
new
Executor
(
places
);
std
::
unique_ptr
<
Executor
>
executor
(
new
Executor
(
places
)
);
// 3 mini-batch
for
(
int
i
=
0
;
i
<
3
;
i
++
)
{
set_feed_variable
<
float
>
(
inputs_
);
for
(
int
batch_id
=
0
;
batch_id
<
3
;
batch_id
++
)
{
SetFeedVariable
<
float
>
(
inputs_
);
executor
->
Run
(
pdesc_
,
GetGlobalScope
());
std
::
vector
<
std
::
vector
<
float
>>
result
=
get_fetch_v
ariable
<
float
>
();
std
::
vector
<
std
::
vector
<
float
>>
result
=
GetFetchV
ariable
<
float
>
();
PADDLE_ENFORCE_EQ
(
result
.
size
(),
inputs_
.
size
());
for
(
size_t
i
=
0
;
i
<
result
.
size
();
++
i
)
{
PADDLE_ENFORCE_EQ
(
result
[
i
].
size
(),
inputs_
[
i
].
size
());
...
...
@@ -302,6 +291,5 @@ TEST_F(ExecutorTesterFeedAndFetch, GPU) {
}
}
}
delete
executor
;
}
#endif
paddle/framework/scope.cc
浏览文件 @
e5155713
...
...
@@ -13,6 +13,7 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/framework/scope.h"
#include <memory> // for unique_ptr
#include <mutex> // for call_once
#include "paddle/string/printf.h"
...
...
paddle/operators/feed_op.cc
浏览文件 @
e5155713
...
...
@@ -31,6 +31,7 @@ class FeedOp : public framework::OperatorWithKernel {
const
FeedInputs
&
tensors
=
g_feed_variable
->
Get
<
FeedInputs
>
();
PADDLE_ENFORCE_GT
(
tensors
.
size
(),
col
);
auto
in_dim
=
tensors
[
col
].
dims
();
ctx
->
SetOutputDim
(
"Out"
,
in_dim
);
// TODO(qijun): need to handle LodTensor later
...
...
paddle/operators/fetch_op.cc
浏览文件 @
e5155713
...
...
@@ -35,6 +35,7 @@ class FetchOp : public framework::OperatorWithKernel {
}
auto
input_dim
=
ctx
->
GetInputDim
(
"Input"
);
PADDLE_ENFORCE_GT
(
tensors
->
size
(),
col
);
(
*
tensors
)[
col
].
Resize
(
input_dim
);
// TODO(qijun): need to handle LodTensor later
...
...
paddle/platform/gpu_info.cc
浏览文件 @
e5155713
...
...
@@ -44,7 +44,7 @@ int GetCurrentDeviceId() {
void
SetDeviceId
(
int
id
)
{
// TODO(qijun): find a better way to cache the cuda device count
PADDLE_ENFORCE
(
id
<
GetCUDADeviceCount
(),
"id must less than GPU count"
);
PADDLE_ENFORCE
_LT
(
id
,
GetCUDADeviceCount
(),
"id must less than GPU count"
);
PADDLE_ENFORCE
(
cudaSetDevice
(
id
),
"cudaSetDevice failed in paddle::platform::SetDeviceId"
);
}
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录