Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
dc863aac
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
dc863aac
编写于
9月 10, 2018
作者:
M
minqiyang
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Add kids exists detection in Scope
上级
681514e1
变更
16
隐藏空白更改
内联
并排
Showing
16 changed file
with
60 addition
and
79 deletion
+60
-79
paddle/fluid/framework/details/fast_threaded_ssa_graph_executor.cc
...uid/framework/details/fast_threaded_ssa_graph_executor.cc
+1
-2
paddle/fluid/framework/details/fast_threaded_ssa_graph_executor.h
...luid/framework/details/fast_threaded_ssa_graph_executor.h
+5
-6
paddle/fluid/framework/details/fetch_op_handle.cc
paddle/fluid/framework/details/fetch_op_handle.cc
+1
-1
paddle/fluid/framework/details/fetch_op_handle.h
paddle/fluid/framework/details/fetch_op_handle.h
+2
-2
paddle/fluid/framework/details/scope_buffered_ssa_graph_executor.cc
...id/framework/details/scope_buffered_ssa_graph_executor.cc
+1
-2
paddle/fluid/framework/details/scope_buffered_ssa_graph_executor.h
...uid/framework/details/scope_buffered_ssa_graph_executor.h
+2
-3
paddle/fluid/framework/details/threaded_ssa_graph_executor.cc
...le/fluid/framework/details/threaded_ssa_graph_executor.cc
+1
-2
paddle/fluid/framework/details/threaded_ssa_graph_executor.h
paddle/fluid/framework/details/threaded_ssa_graph_executor.h
+5
-6
paddle/fluid/framework/parallel_executor.cc
paddle/fluid/framework/parallel_executor.cc
+14
-20
paddle/fluid/framework/parallel_executor.h
paddle/fluid/framework/parallel_executor.h
+10
-11
paddle/fluid/framework/scope.cc
paddle/fluid/framework/scope.cc
+10
-7
paddle/fluid/framework/scope.h
paddle/fluid/framework/scope.h
+4
-1
python/paddle/fluid/tests/book/high-level-api/image_classification/test_image_classification_resnet.py
.../image_classification/test_image_classification_resnet.py
+1
-4
python/paddle/fluid/tests/book/high-level-api/image_classification/test_image_classification_vgg.py
...api/image_classification/test_image_classification_vgg.py
+1
-4
python/paddle/fluid/tests/book/high-level-api/recognize_digits/test_recognize_digits_conv.py
...-level-api/recognize_digits/test_recognize_digits_conv.py
+1
-4
python/paddle/fluid/tests/book/high-level-api/recognize_digits/test_recognize_digits_mlp.py
...h-level-api/recognize_digits/test_recognize_digits_mlp.py
+1
-4
未找到文件。
paddle/fluid/framework/details/fast_threaded_ssa_graph_executor.cc
浏览文件 @
dc863aac
...
@@ -22,8 +22,7 @@ namespace framework {
...
@@ -22,8 +22,7 @@ namespace framework {
namespace
details
{
namespace
details
{
FastThreadedSSAGraphExecutor
::
FastThreadedSSAGraphExecutor
(
FastThreadedSSAGraphExecutor
::
FastThreadedSSAGraphExecutor
(
const
ExecutionStrategy
&
strategy
,
const
ExecutionStrategy
&
strategy
,
const
std
::
vector
<
Scope
*>
&
local_scopes
,
const
std
::
vector
<
std
::
shared_ptr
<
Scope
>>
&
local_scopes
,
const
std
::
vector
<
platform
::
Place
>
&
places
,
const
std
::
vector
<
platform
::
Place
>
&
places
,
std
::
unique_ptr
<
ir
::
Graph
>
&&
graph
)
std
::
unique_ptr
<
ir
::
Graph
>
&&
graph
)
:
strategy_
(
strategy
),
:
strategy_
(
strategy
),
...
...
paddle/fluid/framework/details/fast_threaded_ssa_graph_executor.h
浏览文件 @
dc863aac
...
@@ -29,17 +29,16 @@ namespace details {
...
@@ -29,17 +29,16 @@ namespace details {
class
OpHandleBase
;
class
OpHandleBase
;
class
FastThreadedSSAGraphExecutor
:
public
SSAGraphExecutor
{
class
FastThreadedSSAGraphExecutor
:
public
SSAGraphExecutor
{
public:
public:
FastThreadedSSAGraphExecutor
(
FastThreadedSSAGraphExecutor
(
const
ExecutionStrategy
&
strategy
,
const
ExecutionStrategy
&
strategy
,
const
std
::
vector
<
Scope
*>
&
local_scopes
,
const
std
::
vector
<
std
::
shared_ptr
<
Scope
>>
&
local_scopes
,
const
std
::
vector
<
platform
::
Place
>
&
places
,
const
std
::
vector
<
platform
::
Place
>
&
places
,
std
::
unique_ptr
<
ir
::
Graph
>
&&
graph
);
std
::
unique_ptr
<
ir
::
Graph
>
&&
graph
);
FeedFetchList
Run
(
const
std
::
vector
<
std
::
string
>
&
fetch_tensors
)
override
;
FeedFetchList
Run
(
const
std
::
vector
<
std
::
string
>
&
fetch_tensors
)
override
;
const
ir
::
Graph
&
Graph
()
const
override
;
const
ir
::
Graph
&
Graph
()
const
override
;
private:
private:
ExecutionStrategy
strategy_
;
ExecutionStrategy
strategy_
;
std
::
vector
<
std
::
shared_ptr
<
Scope
>
>
local_scopes_
;
std
::
vector
<
Scope
*
>
local_scopes_
;
std
::
vector
<
platform
::
Place
>
places_
;
std
::
vector
<
platform
::
Place
>
places_
;
std
::
unique_ptr
<
ir
::
Graph
>
graph_
;
std
::
unique_ptr
<
ir
::
Graph
>
graph_
;
...
...
paddle/fluid/framework/details/fetch_op_handle.cc
浏览文件 @
dc863aac
...
@@ -22,7 +22,7 @@ namespace framework {
...
@@ -22,7 +22,7 @@ namespace framework {
namespace
details
{
namespace
details
{
FetchOpHandle
::
FetchOpHandle
(
ir
::
Node
*
node
,
FeedFetchList
*
data
,
size_t
offset
,
FetchOpHandle
::
FetchOpHandle
(
ir
::
Node
*
node
,
FeedFetchList
*
data
,
size_t
offset
,
std
::
vector
<
std
::
shared_ptr
<
Scope
>
>
*
local_scopes
)
std
::
vector
<
Scope
*
>
*
local_scopes
)
:
OpHandleBase
(
node
),
:
OpHandleBase
(
node
),
data_
(
data
),
data_
(
data
),
offset_
(
offset
),
offset_
(
offset
),
...
...
paddle/fluid/framework/details/fetch_op_handle.h
浏览文件 @
dc863aac
...
@@ -29,7 +29,7 @@ namespace details {
...
@@ -29,7 +29,7 @@ namespace details {
struct
FetchOpHandle
:
public
OpHandleBase
{
struct
FetchOpHandle
:
public
OpHandleBase
{
public:
public:
FetchOpHandle
(
ir
::
Node
*
node
,
FeedFetchList
*
data
,
size_t
offset
,
FetchOpHandle
(
ir
::
Node
*
node
,
FeedFetchList
*
data
,
size_t
offset
,
std
::
vector
<
std
::
shared_ptr
<
Scope
>
>
*
local_scopes
);
std
::
vector
<
Scope
*
>
*
local_scopes
);
~
FetchOpHandle
();
~
FetchOpHandle
();
...
@@ -47,7 +47,7 @@ struct FetchOpHandle : public OpHandleBase {
...
@@ -47,7 +47,7 @@ struct FetchOpHandle : public OpHandleBase {
private:
private:
FeedFetchList
*
data_
;
FeedFetchList
*
data_
;
size_t
offset_
;
size_t
offset_
;
std
::
vector
<
std
::
shared_ptr
<
Scope
>
>
*
local_scopes_
;
std
::
vector
<
Scope
*
>
*
local_scopes_
;
std
::
vector
<
LoDTensor
>
tensors_
;
std
::
vector
<
LoDTensor
>
tensors_
;
};
};
...
...
paddle/fluid/framework/details/scope_buffered_ssa_graph_executor.cc
浏览文件 @
dc863aac
...
@@ -23,8 +23,7 @@ namespace paddle {
...
@@ -23,8 +23,7 @@ namespace paddle {
namespace
framework
{
namespace
framework
{
namespace
details
{
namespace
details
{
ScopeBufferedSSAGraphExecutor
::
ScopeBufferedSSAGraphExecutor
(
ScopeBufferedSSAGraphExecutor
::
ScopeBufferedSSAGraphExecutor
(
ExecutionStrategy
strategy
,
ExecutionStrategy
strategy
,
std
::
vector
<
Scope
*>
local_scopes
,
std
::
vector
<
std
::
shared_ptr
<
Scope
>>
local_scopes
,
std
::
vector
<
VariableInfo
>
var_infos
,
std
::
vector
<
platform
::
Place
>
places
,
std
::
vector
<
VariableInfo
>
var_infos
,
std
::
vector
<
platform
::
Place
>
places
,
std
::
unique_ptr
<
SSAGraphExecutor
>
&&
underlying_executor
)
std
::
unique_ptr
<
SSAGraphExecutor
>
&&
underlying_executor
)
:
strategy_
(
std
::
move
(
strategy
)),
:
strategy_
(
std
::
move
(
strategy
)),
...
...
paddle/fluid/framework/details/scope_buffered_ssa_graph_executor.h
浏览文件 @
dc863aac
...
@@ -37,8 +37,7 @@ struct VariableInfo {
...
@@ -37,8 +37,7 @@ struct VariableInfo {
class
ScopeBufferedSSAGraphExecutor
:
public
SSAGraphExecutor
{
class
ScopeBufferedSSAGraphExecutor
:
public
SSAGraphExecutor
{
public:
public:
ScopeBufferedSSAGraphExecutor
(
ScopeBufferedSSAGraphExecutor
(
ExecutionStrategy
strategy
,
ExecutionStrategy
strategy
,
std
::
vector
<
Scope
*>
local_scopes
,
std
::
vector
<
std
::
shared_ptr
<
Scope
>>
local_scopes
,
std
::
vector
<
VariableInfo
>
var_infos
,
std
::
vector
<
platform
::
Place
>
places
,
std
::
vector
<
VariableInfo
>
var_infos
,
std
::
vector
<
platform
::
Place
>
places
,
std
::
unique_ptr
<
SSAGraphExecutor
>&&
underlying_executor
);
std
::
unique_ptr
<
SSAGraphExecutor
>&&
underlying_executor
);
...
@@ -53,7 +52,7 @@ class ScopeBufferedSSAGraphExecutor : public SSAGraphExecutor {
...
@@ -53,7 +52,7 @@ class ScopeBufferedSSAGraphExecutor : public SSAGraphExecutor {
ExecutionStrategy
strategy_
;
ExecutionStrategy
strategy_
;
std
::
unique_ptr
<
SSAGraphExecutor
>
underlying_executor_
;
std
::
unique_ptr
<
SSAGraphExecutor
>
underlying_executor_
;
std
::
vector
<
std
::
shared_ptr
<
Scope
>
>
local_scopes_
;
std
::
vector
<
Scope
*
>
local_scopes_
;
std
::
vector
<
VariableInfo
>
var_infos_
;
std
::
vector
<
VariableInfo
>
var_infos_
;
std
::
vector
<
platform
::
Place
>
places_
;
std
::
vector
<
platform
::
Place
>
places_
;
};
};
...
...
paddle/fluid/framework/details/threaded_ssa_graph_executor.cc
浏览文件 @
dc863aac
...
@@ -21,8 +21,7 @@ namespace paddle {
...
@@ -21,8 +21,7 @@ namespace paddle {
namespace
framework
{
namespace
framework
{
namespace
details
{
namespace
details
{
ThreadedSSAGraphExecutor
::
ThreadedSSAGraphExecutor
(
ThreadedSSAGraphExecutor
::
ThreadedSSAGraphExecutor
(
const
ExecutionStrategy
&
strategy
,
const
ExecutionStrategy
&
strategy
,
const
std
::
vector
<
Scope
*>
&
local_scopes
,
const
std
::
vector
<
std
::
shared_ptr
<
Scope
>>
&
local_scopes
,
const
std
::
vector
<
platform
::
Place
>
&
places
,
const
std
::
vector
<
platform
::
Place
>
&
places
,
std
::
unique_ptr
<
ir
::
Graph
>
&&
graph
)
std
::
unique_ptr
<
ir
::
Graph
>
&&
graph
)
:
graph_
(
std
::
move
(
graph
)),
:
graph_
(
std
::
move
(
graph
)),
...
...
paddle/fluid/framework/details/threaded_ssa_graph_executor.h
浏览文件 @
dc863aac
...
@@ -38,11 +38,10 @@ namespace details {
...
@@ -38,11 +38,10 @@ namespace details {
class
ThreadedSSAGraphExecutor
:
public
SSAGraphExecutor
{
class
ThreadedSSAGraphExecutor
:
public
SSAGraphExecutor
{
public:
public:
ThreadedSSAGraphExecutor
(
ThreadedSSAGraphExecutor
(
const
ExecutionStrategy
&
strategy
,
const
ExecutionStrategy
&
strategy
,
const
std
::
vector
<
Scope
*>
&
local_scopes
,
const
std
::
vector
<
std
::
shared_ptr
<
Scope
>>
&
local_scopes
,
const
std
::
vector
<
platform
::
Place
>
&
places
,
const
std
::
vector
<
platform
::
Place
>
&
places
,
std
::
unique_ptr
<
ir
::
Graph
>
&&
graph
);
std
::
unique_ptr
<
ir
::
Graph
>
&&
graph
);
const
ir
::
Graph
&
Graph
()
const
override
{
return
*
graph_
;
}
const
ir
::
Graph
&
Graph
()
const
override
{
return
*
graph_
;
}
// Run a SSAGraph by a thread pool
// Run a SSAGraph by a thread pool
...
@@ -58,7 +57,7 @@ class ThreadedSSAGraphExecutor : public SSAGraphExecutor {
...
@@ -58,7 +57,7 @@ class ThreadedSSAGraphExecutor : public SSAGraphExecutor {
private:
private:
std
::
unique_ptr
<
ir
::
Graph
>
graph_
;
std
::
unique_ptr
<
ir
::
Graph
>
graph_
;
std
::
unique_ptr
<::
ThreadPool
>
pool_
;
std
::
unique_ptr
<::
ThreadPool
>
pool_
;
std
::
vector
<
std
::
shared_ptr
<
Scope
>
>
local_scopes_
;
std
::
vector
<
Scope
*
>
local_scopes_
;
std
::
vector
<
platform
::
Place
>
places_
;
std
::
vector
<
platform
::
Place
>
places_
;
platform
::
DeviceContextPool
fetch_ctxs_
;
platform
::
DeviceContextPool
fetch_ctxs_
;
ExceptionHolder
exception_holder_
;
ExceptionHolder
exception_holder_
;
...
...
paddle/fluid/framework/parallel_executor.cc
浏览文件 @
dc863aac
...
@@ -39,8 +39,7 @@ std::unique_ptr<ir::Graph> ApplyParallelExecutorPass(
...
@@ -39,8 +39,7 @@ std::unique_ptr<ir::Graph> ApplyParallelExecutorPass(
const
ProgramDesc
&
main_program
,
const
std
::
vector
<
platform
::
Place
>
&
places
,
const
ProgramDesc
&
main_program
,
const
std
::
vector
<
platform
::
Place
>
&
places
,
const
std
::
string
&
loss_var_name
,
const
std
::
string
&
loss_var_name
,
const
std
::
unordered_set
<
std
::
string
>
&
param_names
,
const
std
::
unordered_set
<
std
::
string
>
&
param_names
,
const
std
::
vector
<
std
::
shared_ptr
<
Scope
>>
&
local_scopes
,
const
std
::
vector
<
Scope
*>
&
local_scopes
,
const
bool
use_cuda
,
const
bool
use_cuda
,
#ifdef PADDLE_WITH_CUDA
#ifdef PADDLE_WITH_CUDA
const
BuildStrategy
&
strategy
,
platform
::
NCCLContextMap
*
nccl_ctxs
)
{
const
BuildStrategy
&
strategy
,
platform
::
NCCLContextMap
*
nccl_ctxs
)
{
#else
#else
...
@@ -67,8 +66,8 @@ std::unique_ptr<ir::Graph> ApplyParallelExecutorPass(
...
@@ -67,8 +66,8 @@ std::unique_ptr<ir::Graph> ApplyParallelExecutorPass(
&
loss_var_name
);
&
loss_var_name
);
multi_devices_pass
->
SetNotOwned
<
const
std
::
unordered_set
<
std
::
string
>>
(
multi_devices_pass
->
SetNotOwned
<
const
std
::
unordered_set
<
std
::
string
>>
(
"params"
,
&
param_names
);
"params"
,
&
param_names
);
multi_devices_pass
->
SetNotOwned
<
const
std
::
vector
<
std
::
shared_ptr
<
Scope
>>>
(
multi_devices_pass
->
SetNotOwned
<
const
std
::
vector
<
Scope
*>>
(
"local_scopes"
,
"local_scopes"
,
&
local_scopes
);
&
local_scopes
);
multi_devices_pass
->
SetNotOwned
<
const
BuildStrategy
>
(
"strategy"
,
&
strategy
);
multi_devices_pass
->
SetNotOwned
<
const
BuildStrategy
>
(
"strategy"
,
&
strategy
);
#ifdef PADDLE_WITH_CUDA
#ifdef PADDLE_WITH_CUDA
...
@@ -101,8 +100,8 @@ class ParallelExecutorPrivate {
...
@@ -101,8 +100,8 @@ class ParallelExecutorPrivate {
:
places_
(
places
)
{}
:
places_
(
places
)
{}
std
::
vector
<
platform
::
Place
>
places_
;
std
::
vector
<
platform
::
Place
>
places_
;
std
::
vector
<
std
::
shared_ptr
<
Scope
>
>
local_scopes_
;
std
::
vector
<
Scope
*
>
local_scopes_
;
std
::
shared_ptr
<
Scope
>
global_scope_
;
Scope
*
global_scope_
;
std
::
unique_ptr
<
details
::
SSAGraphExecutor
>
executor_
;
std
::
unique_ptr
<
details
::
SSAGraphExecutor
>
executor_
;
#ifdef PADDLE_WITH_CUDA
#ifdef PADDLE_WITH_CUDA
...
@@ -113,7 +112,7 @@ class ParallelExecutorPrivate {
...
@@ -113,7 +112,7 @@ class ParallelExecutorPrivate {
bool
use_all_reduce_
;
bool
use_all_reduce_
;
};
};
std
::
vector
<
std
::
shared_ptr
<
Scope
>
>
&
ParallelExecutor
::
GetLocalScopes
()
{
std
::
vector
<
Scope
*
>
&
ParallelExecutor
::
GetLocalScopes
()
{
return
member_
->
local_scopes_
;
return
member_
->
local_scopes_
;
}
}
...
@@ -122,8 +121,7 @@ ParallelExecutor::ParallelExecutor(
...
@@ -122,8 +121,7 @@ ParallelExecutor::ParallelExecutor(
const
std
::
unordered_set
<
std
::
string
>
&
params
,
const
std
::
unordered_set
<
std
::
string
>
&
params
,
const
std
::
unordered_set
<
std
::
string
>
&
bcast_vars
,
const
std
::
unordered_set
<
std
::
string
>
&
bcast_vars
,
const
ProgramDesc
&
main_program
,
const
std
::
string
&
loss_var_name
,
const
ProgramDesc
&
main_program
,
const
std
::
string
&
loss_var_name
,
const
std
::
shared_ptr
<
Scope
>
&
scope
,
Scope
*
scope
,
const
std
::
vector
<
Scope
*>
&
local_scopes
,
const
std
::
vector
<
std
::
shared_ptr
<
Scope
>>
&
local_scopes
,
const
ExecutionStrategy
&
exec_strategy
,
const
BuildStrategy
&
build_strategy
,
const
ExecutionStrategy
&
exec_strategy
,
const
BuildStrategy
&
build_strategy
,
size_t
num_trainers
,
size_t
trainer_id
)
size_t
num_trainers
,
size_t
trainer_id
)
:
member_
(
new
ParallelExecutorPrivate
(
places
))
{
:
member_
(
new
ParallelExecutorPrivate
(
places
))
{
...
@@ -144,13 +142,13 @@ ParallelExecutor::ParallelExecutor(
...
@@ -144,13 +142,13 @@ ParallelExecutor::ParallelExecutor(
member_
->
own_local_scope_
=
true
;
member_
->
own_local_scope_
=
true
;
member_
->
local_scopes_
.
emplace_back
(
member_
->
global_scope_
);
member_
->
local_scopes_
.
emplace_back
(
member_
->
global_scope_
);
for
(
size_t
i
=
1
;
i
<
member_
->
places_
.
size
();
++
i
)
{
for
(
size_t
i
=
1
;
i
<
member_
->
places_
.
size
();
++
i
)
{
member_
->
local_scopes_
.
emplace_back
(
scope
->
NewShared
Scope
());
member_
->
local_scopes_
.
emplace_back
(
&
scope
->
New
Scope
());
}
}
}
else
{
}
else
{
member_
->
own_local_scope_
=
false
;
member_
->
own_local_scope_
=
false
;
PADDLE_ENFORCE_EQ
(
member_
->
places_
.
size
(),
local_scopes
.
size
());
PADDLE_ENFORCE_EQ
(
member_
->
places_
.
size
(),
local_scopes
.
size
());
for
(
size_t
i
=
0
;
i
<
member_
->
places_
.
size
();
++
i
)
{
for
(
size_t
i
=
0
;
i
<
member_
->
places_
.
size
();
++
i
)
{
member_
->
local_scopes_
.
emplace_back
(
local_scopes
[
i
]
->
NewShared
Scope
());
member_
->
local_scopes_
.
emplace_back
(
&
local_scopes
[
i
]
->
New
Scope
());
}
}
}
}
...
@@ -323,7 +321,7 @@ void ParallelExecutor::FeedTensorsIntoLocalScopes(
...
@@ -323,7 +321,7 @@ void ParallelExecutor::FeedTensorsIntoLocalScopes(
for
(
size_t
i
=
0
;
i
<
tensors
.
size
();
++
i
)
{
for
(
size_t
i
=
0
;
i
<
tensors
.
size
();
++
i
)
{
auto
&
map
=
tensors
[
i
];
auto
&
map
=
tensors
[
i
];
auto
&
scope
=
member_
->
local_scopes_
[
i
];
auto
*
scope
=
member_
->
local_scopes_
[
i
];
for
(
auto
&
pair
:
map
)
{
for
(
auto
&
pair
:
map
)
{
auto
*
trg
=
scope
->
Var
(
pair
.
first
)
->
GetMutable
<
LoDTensor
>
();
auto
*
trg
=
scope
->
Var
(
pair
.
first
)
->
GetMutable
<
LoDTensor
>
();
trg
->
ShareDataWith
(
pair
.
second
);
trg
->
ShareDataWith
(
pair
.
second
);
...
@@ -353,15 +351,11 @@ void ParallelExecutor::FeedAndSplitTensorIntoLocalScopes(
...
@@ -353,15 +351,11 @@ void ParallelExecutor::FeedAndSplitTensorIntoLocalScopes(
ParallelExecutor
::~
ParallelExecutor
()
{
ParallelExecutor
::~
ParallelExecutor
()
{
if
(
member_
->
own_local_scope_
)
{
if
(
member_
->
own_local_scope_
)
{
std
::
vector
<
Scope
*>
local_scopes_ptrs
;
local_scopes_ptrs
.
reserve
(
member_
->
local_scopes_
.
size
());
for
(
size_t
i
=
1
;
i
<
member_
->
local_scopes_
.
size
();
++
i
)
{
for
(
size_t
i
=
1
;
i
<
member_
->
local_scopes_
.
size
();
++
i
)
{
local_scopes_ptrs
.
emplace_back
(
member_
->
local_scopes_
[
i
].
get
());
Scope
*
local_scope
=
member_
->
local_scopes_
[
i
];
member_
->
local_scopes_
[
i
].
reset
();
if
(
member_
->
global_scope_
->
HasKid
(
local_scope
))
{
}
member_
->
global_scope_
->
DeleteScope
(
local_scope
);
}
for
(
size_t
i
=
0
;
i
!=
local_scopes_ptrs
.
size
();
++
i
)
{
member_
->
global_scope_
->
DeleteScope
(
local_scopes_ptrs
[
i
]);
}
}
}
}
}
}
...
...
paddle/fluid/framework/parallel_executor.h
浏览文件 @
dc863aac
...
@@ -39,20 +39,19 @@ class ParallelExecutor {
...
@@ -39,20 +39,19 @@ class ParallelExecutor {
DISABLE_COPY_AND_ASSIGN
(
ParallelExecutor
);
DISABLE_COPY_AND_ASSIGN
(
ParallelExecutor
);
public:
public:
explicit
ParallelExecutor
(
explicit
ParallelExecutor
(
const
std
::
vector
<
platform
::
Place
>
&
places
,
const
std
::
vector
<
platform
::
Place
>
&
places
,
const
std
::
unordered_set
<
std
::
string
>
&
params
,
const
std
::
unordered_set
<
std
::
string
>
&
params
,
const
std
::
unordered_set
<
std
::
string
>
&
bcast_vars
,
const
std
::
unordered_set
<
std
::
string
>
&
bcast_vars
,
const
ProgramDesc
&
main_program
,
const
ProgramDesc
&
main_program
,
const
std
::
string
&
loss_var_name
,
const
std
::
string
&
loss_var_name
,
Scope
*
scope
,
const
std
::
shared_ptr
<
Scope
>
&
scope
,
const
std
::
vector
<
Scope
*>
&
local_scopes
,
const
std
::
vector
<
std
::
shared_ptr
<
Scope
>>
&
local_scopes
,
const
ExecutionStrategy
&
exec_strategy
,
const
ExecutionStrategy
&
exec_strategy
,
const
BuildStrategy
&
build_strategy
,
const
BuildStrategy
&
build_strategy
,
size_t
num_trainers
=
1
,
size_t
num_trainers
=
1
,
size_t
trainer_id
=
0
);
size_t
trainer_id
=
0
);
~
ParallelExecutor
();
~
ParallelExecutor
();
std
::
vector
<
std
::
shared_ptr
<
Scope
>
>
&
GetLocalScopes
();
std
::
vector
<
Scope
*
>
&
GetLocalScopes
();
/**
/**
* Feed tensors to local scopes. The size of tensors should be equal to the
* Feed tensors to local scopes. The size of tensors should be equal to the
...
...
paddle/fluid/framework/scope.cc
浏览文件 @
dc863aac
...
@@ -38,8 +38,8 @@ Scope::~Scope() { DropKids(); }
...
@@ -38,8 +38,8 @@ Scope::~Scope() { DropKids(); }
Scope
&
Scope
::
NewScope
()
const
{
Scope
&
Scope
::
NewScope
()
const
{
std
::
unique_lock
<
std
::
mutex
>
lock
(
mutex_
);
std
::
unique_lock
<
std
::
mutex
>
lock
(
mutex_
);
kids_
.
push_back
(
std
::
shared_ptr
<
Scope
>
(
new
Scope
(
this
)
));
kids_
.
push_back
(
new
Scope
(
this
));
return
kids_
.
back
().
get
();
return
*
kids_
.
back
();
}
}
Variable
*
Scope
::
Var
(
const
std
::
string
&
name
)
{
Variable
*
Scope
::
Var
(
const
std
::
string
&
name
)
{
...
@@ -68,9 +68,16 @@ const Scope* Scope::FindScope(const Variable* var) const {
...
@@ -68,9 +68,16 @@ const Scope* Scope::FindScope(const Variable* var) const {
void
Scope
::
DropKids
()
{
void
Scope
::
DropKids
()
{
std
::
unique_lock
<
std
::
mutex
>
lock
(
mutex_
);
std
::
unique_lock
<
std
::
mutex
>
lock
(
mutex_
);
for
(
Scope
*
s
:
kids_
)
delete
s
;
kids_
.
clear
();
kids_
.
clear
();
}
}
bool
Scope
::
HasKid
(
const
Scope
*
scope
)
const
{
std
::
unique_lock
<
std
::
mutex
>
lock
(
mutex_
);
auto
it
=
std
::
find
(
this
->
kids_
.
begin
(),
this
->
kids_
.
end
(),
scope
);
return
it
!=
this
->
kids_
.
end
();
}
std
::
vector
<
std
::
string
>
Scope
::
LocalVarNames
()
const
{
std
::
vector
<
std
::
string
>
Scope
::
LocalVarNames
()
const
{
std
::
unique_lock
<
std
::
mutex
>
lock
(
mutex_
);
std
::
unique_lock
<
std
::
mutex
>
lock
(
mutex_
);
std
::
vector
<
std
::
string
>
known_vars
;
std
::
vector
<
std
::
string
>
known_vars
;
...
@@ -83,12 +90,8 @@ std::vector<std::string> Scope::LocalVarNames() const {
...
@@ -83,12 +90,8 @@ std::vector<std::string> Scope::LocalVarNames() const {
void
Scope
::
DeleteScope
(
Scope
*
scope
)
const
{
void
Scope
::
DeleteScope
(
Scope
*
scope
)
const
{
std
::
unique_lock
<
std
::
mutex
>
lock
(
mutex_
);
std
::
unique_lock
<
std
::
mutex
>
lock
(
mutex_
);
auto
it
=
std
::
find_if
(
this
->
kids_
.
begin
(),
this
->
kids_
.
end
(),
auto
it
=
std
::
find
(
this
->
kids_
.
begin
(),
this
->
kids_
.
end
(),
scope
);
[
&
scope
](
const
std
::
shared_ptr
<
Scope
>&
kid
)
{
return
kid
.
get
()
==
scope
;
});
PADDLE_ENFORCE
(
it
!=
this
->
kids_
.
end
(),
"Cannot find %p as kid scope"
,
scope
);
PADDLE_ENFORCE
(
it
!=
this
->
kids_
.
end
(),
"Cannot find %p as kid scope"
,
scope
);
it
->
reset
();
this
->
kids_
.
erase
(
it
);
this
->
kids_
.
erase
(
it
);
// When making memory benchmark on Fluid, we have to delete scope sync.
// When making memory benchmark on Fluid, we have to delete scope sync.
if
(
FLAGS_benchmark
||
FLAGS_eager_delete_scope
)
{
if
(
FLAGS_benchmark
||
FLAGS_eager_delete_scope
)
{
...
...
paddle/fluid/framework/scope.h
浏览文件 @
dc863aac
...
@@ -71,6 +71,9 @@ class Scope {
...
@@ -71,6 +71,9 @@ class Scope {
/// Drop all kids scopes belonged to this scope.
/// Drop all kids scopes belonged to this scope.
void
DropKids
();
void
DropKids
();
/// Find if a scope exists in the kid scopes
bool
HasKid
(
const
Scope
*
scope
)
const
;
// enumerate all the variables current contains.
// enumerate all the variables current contains.
std
::
vector
<
std
::
string
>
LocalVarNames
()
const
;
std
::
vector
<
std
::
string
>
LocalVarNames
()
const
;
...
@@ -105,7 +108,7 @@ class Scope {
...
@@ -105,7 +108,7 @@ class Scope {
Variable
*
FindVarLocally
(
const
std
::
string
&
name
)
const
;
Variable
*
FindVarLocally
(
const
std
::
string
&
name
)
const
;
// Scope in `kids_` are owned by this class.
// Scope in `kids_` are owned by this class.
mutable
std
::
list
<
std
::
shared_ptr
<
Scope
>
>
kids_
;
mutable
std
::
list
<
Scope
*
>
kids_
;
Scope
const
*
parent_
{
nullptr
};
Scope
const
*
parent_
{
nullptr
};
DISABLE_COPY_AND_ASSIGN
(
Scope
);
DISABLE_COPY_AND_ASSIGN
(
Scope
);
...
...
python/paddle/fluid/tests/book/high-level-api/image_classification/test_image_classification_resnet.py
浏览文件 @
dc863aac
...
@@ -178,7 +178,4 @@ if __name__ == '__main__':
...
@@ -178,7 +178,4 @@ if __name__ == '__main__':
for
parallel
in
(
False
,
True
):
for
parallel
in
(
False
,
True
):
if
use_cuda
and
not
core
.
is_compiled_with_cuda
():
if
use_cuda
and
not
core
.
is_compiled_with_cuda
():
continue
continue
# TODO(minqiyang): remove this line after fixing the deletion
main
(
use_cuda
=
use_cuda
,
parallel
=
parallel
)
# order problem of Scope in ParallelExecutor in manylinux
if
six
.
PY2
:
main
(
use_cuda
=
use_cuda
,
parallel
=
parallel
)
python/paddle/fluid/tests/book/high-level-api/image_classification/test_image_classification_vgg.py
浏览文件 @
dc863aac
...
@@ -152,7 +152,4 @@ if __name__ == '__main__':
...
@@ -152,7 +152,4 @@ if __name__ == '__main__':
for
parallel
in
(
False
,
True
):
for
parallel
in
(
False
,
True
):
if
use_cuda
and
not
core
.
is_compiled_with_cuda
():
if
use_cuda
and
not
core
.
is_compiled_with_cuda
():
continue
continue
# TODO(minqiyang): remove this line after fixing the deletion
main
(
use_cuda
=
use_cuda
,
parallel
=
parallel
)
# order problem of Scope in ParallelExecutor in manylinux
if
six
.
PY2
:
main
(
use_cuda
=
use_cuda
,
parallel
=
parallel
)
python/paddle/fluid/tests/book/high-level-api/recognize_digits/test_recognize_digits_conv.py
浏览文件 @
dc863aac
...
@@ -155,7 +155,4 @@ if __name__ == '__main__':
...
@@ -155,7 +155,4 @@ if __name__ == '__main__':
for
parallel
in
(
False
,
True
):
for
parallel
in
(
False
,
True
):
if
use_cuda
and
not
core
.
is_compiled_with_cuda
():
if
use_cuda
and
not
core
.
is_compiled_with_cuda
():
continue
continue
# TODO(minqiyang): remove this line after fixing the deletion
main
(
use_cuda
=
use_cuda
,
parallel
=
parallel
)
# order problem of Scope in ParallelExecutor in manylinux
if
six
.
PY2
:
main
(
use_cuda
=
use_cuda
,
parallel
=
parallel
)
python/paddle/fluid/tests/book/high-level-api/recognize_digits/test_recognize_digits_mlp.py
浏览文件 @
dc863aac
...
@@ -137,7 +137,4 @@ if __name__ == '__main__':
...
@@ -137,7 +137,4 @@ if __name__ == '__main__':
for
parallel
in
(
False
,
True
):
for
parallel
in
(
False
,
True
):
if
use_cuda
and
not
core
.
is_compiled_with_cuda
():
if
use_cuda
and
not
core
.
is_compiled_with_cuda
():
continue
continue
# TODO(minqiyang): remove this line after fixing the deletion
main
(
use_cuda
=
use_cuda
,
parallel
=
parallel
)
# order problem of Scope in ParallelExecutor in manylinux
if
six
.
PY2
:
main
(
use_cuda
=
use_cuda
,
parallel
=
parallel
)
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录