Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
e5155713
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
e5155713
编写于
10月 09, 2017
作者:
Y
Yang Yang
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
clean up for review
上级
089cc11d
变更
7
隐藏空白更改
内联
并排
Showing
7 changed file
with
56 addition
and
51 deletion
+56
-51
paddle/framework/executor.cc
paddle/framework/executor.cc
+27
-13
paddle/framework/executor.h
paddle/framework/executor.h
+1
-1
paddle/framework/executor_test.cc
paddle/framework/executor_test.cc
+24
-36
paddle/framework/scope.cc
paddle/framework/scope.cc
+1
-0
paddle/operators/feed_op.cc
paddle/operators/feed_op.cc
+1
-0
paddle/operators/fetch_op.cc
paddle/operators/fetch_op.cc
+1
-0
paddle/platform/gpu_info.cc
paddle/platform/gpu_info.cc
+1
-1
未找到文件。
paddle/framework/executor.cc
浏览文件 @
e5155713
...
...
@@ -13,11 +13,13 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/framework/executor.h"
#include <algorithm>
#include <iostream>
#include <memory>
#include <set>
#include <vector>
#include "paddle/framework/lod_tensor.h"
#include "paddle/framework/op_registry.h"
#include "paddle/framework/scope.h"
...
...
@@ -27,7 +29,11 @@ limitations under the License. */
namespace
paddle
{
namespace
framework
{
const
std
::
string
kFeedOpType
=
"feed"
;
const
std
::
string
kFetchOpType
=
"fetch"
;
Executor
::
Executor
(
const
std
::
vector
<
platform
::
Place
>&
places
)
{
PADDLE_ENFORCE_GT
(
places
.
size
(),
0
);
device_contexts_
.
resize
(
places
.
size
());
for
(
size_t
i
=
0
;
i
<
places
.
size
();
i
++
)
{
if
(
platform
::
is_cpu_place
(
places
[
i
]))
{
...
...
@@ -46,9 +52,7 @@ Executor::Executor(const std::vector<platform::Place>& places) {
Executor
::~
Executor
()
{
for
(
auto
&
device_context
:
device_contexts_
)
{
if
(
device_context
)
{
delete
device_context
;
}
delete
device_context
;
}
}
...
...
@@ -56,6 +60,8 @@ void Executor::Run(const ProgramDesc& pdesc, Scope* scope) {
// TODO(tonyyang-svail):
// - only runs the first block (i.e. no RNN support)
// - only runs on the first device (i.e. no interdevice communication)
// - will change to use multiple blocks for RNN op and Cond Op
PADDLE_ENFORCE_GT
(
pdesc
.
blocks_size
(),
0
);
auto
&
block
=
pdesc
.
blocks
(
0
);
auto
&
device
=
device_contexts_
[
0
];
...
...
@@ -66,12 +72,12 @@ void Executor::Run(const ProgramDesc& pdesc, Scope* scope) {
Scope
&
local_scope
=
scope
->
NewScope
();
std
::
vector
<
bool
>
should_run
=
Pr
eprocess
(
pdesc
);
PADDLE_ENFORCE
(
should_run
.
size
()
==
block
.
ops_size
());
std
::
vector
<
bool
>
should_run
=
Pr
une
(
pdesc
);
PADDLE_ENFORCE
_EQ
(
should_run
.
size
(),
block
.
ops_size
());
for
(
size_t
i
=
0
;
i
<
should_run
.
size
();
++
i
)
{
if
(
should_run
[
i
])
{
for
(
auto
var
:
block
.
ops
(
i
).
outputs
())
{
for
(
auto
argu
:
var
.
arguments
())
{
for
(
auto
&
var
:
block
.
ops
(
i
).
outputs
())
{
for
(
auto
&
argu
:
var
.
arguments
())
{
if
(
local_scope
.
FindVar
(
argu
)
==
nullptr
)
{
local_scope
.
NewVar
(
argu
);
}
...
...
@@ -81,28 +87,32 @@ void Executor::Run(const ProgramDesc& pdesc, Scope* scope) {
op
->
Run
(
local_scope
,
*
device
);
}
}
// TODO(tonyyang-svail):
// - Destroy local_scope
}
std
::
vector
<
bool
>
Executor
::
Pr
eprocess
(
const
ProgramDesc
&
pdesc
)
{
std
::
vector
<
bool
>
Executor
::
Pr
une
(
const
ProgramDesc
&
pdesc
)
{
// TODO(tonyyang-svail):
// - only runs the first block
// - will change to use multiple blocks for RNN op and Cond Op
auto
&
block
=
pdesc
.
blocks
(
0
);
auto
&
ops
=
block
.
ops
();
bool
expect_feed
=
true
;
for
(
auto
&
op_desc
:
ops
)
{
PADDLE_ENFORCE
(
op_desc
.
type
()
!=
"feed"
||
expect_feed
,
PADDLE_ENFORCE
(
op_desc
.
type
()
!=
kFeedOpType
||
expect_feed
,
"All FeedOps are at the beginning of the ProgramDesc"
);
expect_feed
=
(
op_desc
.
type
()
==
"feed"
);
expect_feed
=
(
op_desc
.
type
()
==
kFeedOpType
);
}
bool
expect_fetch
=
true
;
for
(
auto
op_iter
=
ops
.
rbegin
();
op_iter
!=
ops
.
rend
();
++
op_iter
)
{
auto
&
op_desc
=
*
op_iter
;
PADDLE_ENFORCE
(
op_desc
.
type
()
!=
"fetch"
||
expect_fetch
,
PADDLE_ENFORCE
(
op_desc
.
type
()
!=
kFetchOpType
||
expect_fetch
,
"All FetchOps must at the end of the ProgramDesc"
);
expect_fetch
=
(
op_desc
.
type
()
==
"fetch"
);
expect_fetch
=
(
op_desc
.
type
()
==
kFetchOpType
);
}
std
::
set
<
std
::
string
>
dependent_vars
;
...
...
@@ -119,7 +129,7 @@ std::vector<bool> Executor::Preprocess(const ProgramDesc& pdesc) {
}
}
if
(
op_desc
.
type
()
==
"fetch"
||
found_dependent_vars
)
{
if
(
op_desc
.
type
()
==
kFetchOpType
||
found_dependent_vars
)
{
// erase its output to the dependency graph
for
(
auto
&
var
:
op_desc
.
outputs
())
{
for
(
auto
&
argu
:
var
.
arguments
())
{
...
...
@@ -140,6 +150,10 @@ std::vector<bool> Executor::Preprocess(const ProgramDesc& pdesc) {
}
}
// TODO(tonyyang-svail):
// - check this after integration of Init
// PADDLE_ENFORCE(dependent_vars.empty());
// since we are traversing the ProgramDesc in reverse order
// we reverse the should_run vector
std
::
reverse
(
should_run
.
begin
(),
should_run
.
end
());
...
...
paddle/framework/executor.h
浏览文件 @
e5155713
...
...
@@ -46,7 +46,7 @@ class Executor {
* @return
* vector<bool> Same size as ops. Indicates whether an op should be run.
*/
std
::
vector
<
bool
>
Pr
eprocess
(
const
ProgramDesc
&
pdesc
);
std
::
vector
<
bool
>
Pr
une
(
const
ProgramDesc
&
pdesc
);
private:
std
::
vector
<
platform
::
DeviceContext
*>
device_contexts_
;
...
...
paddle/framework/executor_test.cc
浏览文件 @
e5155713
...
...
@@ -13,12 +13,14 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/framework/executor.h"
#include <memory>
#include <vector>
#include "gtest/gtest.h"
#include "paddle/framework/attribute.h"
#include "paddle/framework/backward.h"
#include "paddle/framework/block_desc.h"
// #include "paddle/framework/grad_op_builder.h"
#include "paddle/framework/op_desc.h"
#include "paddle/framework/op_registry.h"
#include "paddle/framework/operator.h"
...
...
@@ -34,9 +36,6 @@ using std::string;
using
namespace
paddle
::
platform
;
using
namespace
paddle
::
framework
;
typedef
paddle
::
framework
::
BlockDesc
proto_block
;
typedef
paddle
::
framework
::
OpDesc
proto_op
;
void
AddOp
(
const
std
::
string
&
type
,
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
AttributeMap
attrs
,
paddle
::
framework
::
BlockDescBind
*
block
)
{
...
...
@@ -51,10 +50,10 @@ void AddOp(const std::string& type, const VariableNameMap& inputs,
// insert op
auto
op
=
block
->
AppendOp
();
op
->
SetType
(
type
);
for
(
auto
kv
:
inputs
)
{
for
(
auto
&
kv
:
inputs
)
{
op
->
SetInput
(
kv
.
first
,
kv
.
second
);
}
for
(
auto
kv
:
outputs
)
{
for
(
auto
&
kv
:
outputs
)
{
op
->
SetOutput
(
kv
.
first
,
kv
.
second
);
}
op
->
SetAttrMap
(
attrs
);
...
...
@@ -65,11 +64,11 @@ std::once_flag set_variable_flag;
// Tensors in feed value variable will only be in CPUPlace
// So we can memcpy the data from vector<T> to feed_value
template
<
typename
T
>
void
set_feed_v
ariable
(
const
std
::
vector
<
std
::
vector
<
T
>>&
inputs
)
{
void
SetFeedV
ariable
(
const
std
::
vector
<
std
::
vector
<
T
>>&
inputs
)
{
typedef
std
::
vector
<
paddle
::
framework
::
Tensor
>
FeedInputs
;
Variable
*
g_feed_value
=
GetGlobalScope
()
->
FindVar
(
"feed_value"
);
FeedInputs
&
feed_inputs
=
*
(
g_feed_value
->
GetMutable
<
FeedInputs
>
());
auto
size
=
inputs
.
size
();
size_t
size
=
inputs
.
size
();
feed_inputs
.
resize
(
size
);
for
(
size_t
i
=
0
;
i
<
size
;
i
++
)
{
T
*
dst
=
feed_inputs
[
i
].
mutable_data
<
T
>
(
...
...
@@ -81,12 +80,12 @@ void set_feed_variable(const std::vector<std::vector<T>>& inputs) {
// Tensors in fetch value variable will only be in CPUPlace
// So we can memcpy the data from fetch_value to vector<T>
template
<
typename
T
>
std
::
vector
<
std
::
vector
<
T
>>
get_fetch_v
ariable
()
{
std
::
vector
<
std
::
vector
<
T
>>
GetFetchV
ariable
()
{
typedef
std
::
vector
<
paddle
::
framework
::
Tensor
>
FetchOutputs
;
Variable
*
g_fetch_value
=
GetGlobalScope
()
->
FindVar
(
"fetch_value"
);
FetchOutputs
&
fetch_outputs
=
*
(
g_fetch_value
->
GetMutable
<
FetchOutputs
>
());
auto
size
=
fetch_outputs
.
size
();
size_t
size
=
fetch_outputs
.
size
();
std
::
vector
<
std
::
vector
<
T
>>
result
;
result
.
reserve
(
size
);
for
(
size_t
i
=
0
;
i
<
size
;
i
++
)
{
...
...
@@ -105,7 +104,7 @@ class ExecutorTesterRandom : public ::testing::Test {
virtual
void
SetUp
()
override
{
int
input_dim
=
5
,
batch_size
=
2
,
embed_dim
=
5
;
// init pdesc
-----------------------------------------
// init pdesc
auto
temp_init_root_block
=
init_pdesc_
.
add_blocks
();
temp_init_root_block
->
set_idx
(
0
);
temp_init_root_block
->
set_parent_idx
(
-
1
);
...
...
@@ -128,7 +127,7 @@ class ExecutorTesterRandom : public ::testing::Test {
// flush
init_program
.
Proto
();
// run pdesc
-----------------------------------------
// run pdesc
auto
temp_root_block
=
pdesc_
.
add_blocks
();
temp_root_block
->
set_idx
(
0
);
temp_root_block
->
set_parent_idx
(
-
1
);
...
...
@@ -154,9 +153,6 @@ class ExecutorTesterRandom : public ::testing::Test {
// TODO(tonyyang-svail):
// - Test with Backward
// AddOp("gaussian_random", {}, {{"Out", {"l2_distance@GRAD"}}},
// {{"dims", std::vector<int>{batch_size, 1}}}, root_block);
// AppendBackward(program, {});
}
protected:
...
...
@@ -213,12 +209,11 @@ TEST_F(ExecutorTesterRandom, CPU) {
// "pointer being freed was not allocated" error will appear.
paddle
::
memory
::
Used
(
cpu_place
);
Executor
*
executor
=
new
Executor
(
places
);
std
::
unique_ptr
<
Executor
>
executor
(
new
Executor
(
places
));
executor
->
Run
(
init_pdesc_
,
GetGlobalScope
());
executor
->
Run
(
pdesc_
,
GetGlobalScope
());
std
::
vector
<
std
::
vector
<
float
>>
result
=
get_fetch_variable
<
float
>
();
delete
executor
;
std
::
vector
<
std
::
vector
<
float
>>
result
=
GetFetchVariable
<
float
>
();
}
TEST_F
(
ExecutorTesterFeedAndFetch
,
CPU
)
{
...
...
@@ -232,13 +227,12 @@ TEST_F(ExecutorTesterFeedAndFetch, CPU) {
// "pointer being freed was not allocated" error will appear.
paddle
::
memory
::
Used
(
cpu_place
);
Executor
*
executor
=
new
Executor
(
places
);
std
::
unique_ptr
<
Executor
>
executor
(
new
Executor
(
places
)
);
// 3 mini-batch
for
(
int
i
=
0
;
i
<
3
;
i
++
)
{
set_feed_variable
<
float
>
(
inputs_
);
for
(
int
batch_id
=
0
;
batch_id
<
3
;
batch_id
++
)
{
SetFeedVariable
<
float
>
(
inputs_
);
executor
->
Run
(
pdesc_
,
GetGlobalScope
());
std
::
vector
<
std
::
vector
<
float
>>
result
=
get_fetch_v
ariable
<
float
>
();
std
::
vector
<
std
::
vector
<
float
>>
result
=
GetFetchV
ariable
<
float
>
();
PADDLE_ENFORCE_EQ
(
result
.
size
(),
inputs_
.
size
());
for
(
size_t
i
=
0
;
i
<
result
.
size
();
++
i
)
{
PADDLE_ENFORCE_EQ
(
result
[
i
].
size
(),
inputs_
[
i
].
size
());
...
...
@@ -247,8 +241,6 @@ TEST_F(ExecutorTesterFeedAndFetch, CPU) {
}
}
}
delete
executor
;
}
#else
TEST_F
(
ExecutorTesterRandom
,
GPU
)
{
...
...
@@ -265,13 +257,11 @@ TEST_F(ExecutorTesterRandom, GPU) {
paddle
::
memory
::
Used
(
CPUPlace
());
paddle
::
memory
::
Used
(
gpu_place
);
Executor
*
executor
=
new
Executor
(
places
);
std
::
unique_ptr
<
Executor
>
executor
(
new
Executor
(
places
)
);
executor
->
Run
(
init_pdesc_
,
GetGlobalScope
());
executor
->
Run
(
pdesc_
,
GetGlobalScope
());
std
::
vector
<
std
::
vector
<
float
>>
result
=
get_fetch_variable
<
float
>
();
delete
executor
;
std
::
vector
<
std
::
vector
<
float
>>
result
=
GetFetchVariable
<
float
>
();
}
TEST_F
(
ExecutorTesterFeedAndFetch
,
GPU
)
{
...
...
@@ -287,13 +277,12 @@ TEST_F(ExecutorTesterFeedAndFetch, GPU) {
paddle
::
memory
::
Used
(
CPUPlace
());
paddle
::
memory
::
Used
(
gpu_place
);
Executor
*
executor
=
new
Executor
(
places
);
std
::
unique_ptr
<
Executor
>
executor
(
new
Executor
(
places
)
);
// 3 mini-batch
for
(
int
i
=
0
;
i
<
3
;
i
++
)
{
set_feed_variable
<
float
>
(
inputs_
);
for
(
int
batch_id
=
0
;
batch_id
<
3
;
batch_id
++
)
{
SetFeedVariable
<
float
>
(
inputs_
);
executor
->
Run
(
pdesc_
,
GetGlobalScope
());
std
::
vector
<
std
::
vector
<
float
>>
result
=
get_fetch_v
ariable
<
float
>
();
std
::
vector
<
std
::
vector
<
float
>>
result
=
GetFetchV
ariable
<
float
>
();
PADDLE_ENFORCE_EQ
(
result
.
size
(),
inputs_
.
size
());
for
(
size_t
i
=
0
;
i
<
result
.
size
();
++
i
)
{
PADDLE_ENFORCE_EQ
(
result
[
i
].
size
(),
inputs_
[
i
].
size
());
...
...
@@ -302,6 +291,5 @@ TEST_F(ExecutorTesterFeedAndFetch, GPU) {
}
}
}
delete
executor
;
}
#endif
paddle/framework/scope.cc
浏览文件 @
e5155713
...
...
@@ -13,6 +13,7 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/framework/scope.h"
#include <memory> // for unique_ptr
#include <mutex> // for call_once
#include "paddle/string/printf.h"
...
...
paddle/operators/feed_op.cc
浏览文件 @
e5155713
...
...
@@ -31,6 +31,7 @@ class FeedOp : public framework::OperatorWithKernel {
const
FeedInputs
&
tensors
=
g_feed_variable
->
Get
<
FeedInputs
>
();
PADDLE_ENFORCE_GT
(
tensors
.
size
(),
col
);
auto
in_dim
=
tensors
[
col
].
dims
();
ctx
->
SetOutputDim
(
"Out"
,
in_dim
);
// TODO(qijun): need to handle LodTensor later
...
...
paddle/operators/fetch_op.cc
浏览文件 @
e5155713
...
...
@@ -35,6 +35,7 @@ class FetchOp : public framework::OperatorWithKernel {
}
auto
input_dim
=
ctx
->
GetInputDim
(
"Input"
);
PADDLE_ENFORCE_GT
(
tensors
->
size
(),
col
);
(
*
tensors
)[
col
].
Resize
(
input_dim
);
// TODO(qijun): need to handle LodTensor later
...
...
paddle/platform/gpu_info.cc
浏览文件 @
e5155713
...
...
@@ -44,7 +44,7 @@ int GetCurrentDeviceId() {
void
SetDeviceId
(
int
id
)
{
// TODO(qijun): find a better way to cache the cuda device count
PADDLE_ENFORCE
(
id
<
GetCUDADeviceCount
(),
"id must less than GPU count"
);
PADDLE_ENFORCE
_LT
(
id
,
GetCUDADeviceCount
(),
"id must less than GPU count"
);
PADDLE_ENFORCE
(
cudaSetDevice
(
id
),
"cudaSetDevice failed in paddle::platform::SetDeviceId"
);
}
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录