Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
dfb47986
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
dfb47986
编写于
4月 07, 2022
作者:
L
liutiexing
提交者:
GitHub
4月 07, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Profile Executors (#41100)
* Profile Executors * update * fix ut * fix names * update * update
上级
f87f0656
变更
11
隐藏空白更改
内联
并排
Showing
11 changed file
with
71 addition
and
19 deletion
+71
-19
paddle/fluid/framework/details/fast_threaded_ssa_graph_executor.cc
...uid/framework/details/fast_threaded_ssa_graph_executor.cc
+6
-2
paddle/fluid/framework/executor.cc
paddle/fluid/framework/executor.cc
+8
-0
paddle/fluid/framework/ir/cost_model.cc
paddle/fluid/framework/ir/cost_model.cc
+15
-2
paddle/fluid/framework/new_executor/event_manager.cc
paddle/fluid/framework/new_executor/event_manager.cc
+7
-0
paddle/fluid/framework/new_executor/interpretercore.cc
paddle/fluid/framework/new_executor/interpretercore.cc
+12
-0
paddle/fluid/framework/new_executor/workqueue/nonblocking_threadpool.h
...framework/new_executor/workqueue/nonblocking_threadpool.h
+2
-2
paddle/fluid/framework/new_executor/workqueue/workqueue.cc
paddle/fluid/framework/new_executor/workqueue/workqueue.cc
+6
-4
paddle/fluid/framework/parallel_executor.cc
paddle/fluid/framework/parallel_executor.cc
+2
-0
paddle/fluid/memory/allocation/auto_growth_best_fit_allocator.cc
...fluid/memory/allocation/auto_growth_best_fit_allocator.cc
+6
-4
paddle/fluid/memory/allocation/stream_safe_cuda_allocator.cc
paddle/fluid/memory/allocation/stream_safe_cuda_allocator.cc
+6
-4
paddle/fluid/pybind/pybind.cc
paddle/fluid/pybind/pybind.cc
+1
-1
未找到文件。
paddle/fluid/framework/details/fast_threaded_ssa_graph_executor.cc
浏览文件 @
dfb47986
...
@@ -132,6 +132,9 @@ FetchResultType FastThreadedSSAGraphExecutor::Run(
...
@@ -132,6 +132,9 @@ FetchResultType FastThreadedSSAGraphExecutor::Run(
}
}
// Wait FetchOps.
// Wait FetchOps.
if
(
!
fetch_ops
.
empty
())
{
if
(
!
fetch_ops
.
empty
())
{
platform
::
RecordEvent
record_wait
(
"FastThreadedSSAGraphExecutor::WaitFetchOps"
,
platform
::
TracerEventType
::
Operator
,
1
);
ClearFetchOp
(
graph_
,
&
fetch_ops
);
ClearFetchOp
(
graph_
,
&
fetch_ops
);
for
(
auto
&
place
:
places_
)
{
for
(
auto
&
place
:
places_
)
{
...
@@ -231,8 +234,9 @@ void FastThreadedSSAGraphExecutor::RunOpAsync(
...
@@ -231,8 +234,9 @@ void FastThreadedSSAGraphExecutor::RunOpAsync(
OpHandleBase
*
op
,
OpHandleBase
*
op
,
const
std
::
shared_ptr
<
BlockingQueue
<
size_t
>>
&
complete_q
)
{
const
std
::
shared_ptr
<
BlockingQueue
<
size_t
>>
&
complete_q
)
{
++
remaining_
;
++
remaining_
;
platform
::
RecordEvent
(
"WorkQueue::AddTask"
,
platform
::
RecordEvent
record
(
"WorkQueue::AddTask"
,
platform
::
TracerEventType
::
UserDefined
,
10
/*level*/
);
platform
::
TracerEventType
::
UserDefined
,
10
/*level*/
);
this
->
pool_
->
enqueue
([
=
]
{
this
->
pool_
->
enqueue
([
=
]
{
std
::
deque
<
OpHandleBase
*>
op_queue
;
std
::
deque
<
OpHandleBase
*>
op_queue
;
op_queue
.
push_front
(
op
);
op_queue
.
push_front
(
op
);
...
...
paddle/fluid/framework/executor.cc
浏览文件 @
dfb47986
...
@@ -172,6 +172,8 @@ void Executor::Run(const ProgramDesc& pdesc, Scope* scope, int block_id,
...
@@ -172,6 +172,8 @@ void Executor::Run(const ProgramDesc& pdesc, Scope* scope, int block_id,
bool
create_local_scope
,
bool
create_vars
,
bool
create_local_scope
,
bool
create_vars
,
const
std
::
vector
<
std
::
string
>&
skip_ref_cnt_vars
,
const
std
::
vector
<
std
::
string
>&
skip_ref_cnt_vars
,
bool
force_disable_gc
,
bool
keep_kid_scopes
)
{
bool
force_disable_gc
,
bool
keep_kid_scopes
)
{
platform
::
RecordEvent
record_run
(
"Executor::Run"
,
platform
::
TracerEventType
::
UserDefined
,
1
);
platform
::
RecordBlock
b
(
block_id
);
platform
::
RecordBlock
b
(
block_id
);
if
(
FLAGS_use_mkldnn
)
EnableMKLDNN
(
pdesc
);
if
(
FLAGS_use_mkldnn
)
EnableMKLDNN
(
pdesc
);
auto
ctx
=
Prepare
(
pdesc
,
block_id
,
skip_ref_cnt_vars
,
force_disable_gc
);
auto
ctx
=
Prepare
(
pdesc
,
block_id
,
skip_ref_cnt_vars
,
force_disable_gc
);
...
@@ -301,6 +303,8 @@ void Executor::Run(const ProgramDesc& program, Scope* scope,
...
@@ -301,6 +303,8 @@ void Executor::Run(const ProgramDesc& program, Scope* scope,
bool
create_local_scope
,
bool
create_vars
,
bool
create_local_scope
,
bool
create_vars
,
const
std
::
string
&
feed_holder_name
,
const
std
::
string
&
feed_holder_name
,
const
std
::
string
&
fetch_holder_name
)
{
const
std
::
string
&
fetch_holder_name
)
{
platform
::
RecordEvent
record_run
(
"Executor::Run"
,
platform
::
TracerEventType
::
UserDefined
,
1
);
platform
::
RecordBlock
b
(
kProgramId
);
platform
::
RecordBlock
b
(
kProgramId
);
if
(
FLAGS_use_mkldnn
)
EnableMKLDNN
(
program
);
if
(
FLAGS_use_mkldnn
)
EnableMKLDNN
(
program
);
#ifdef PADDLE_WITH_MKLDNN
#ifdef PADDLE_WITH_MKLDNN
...
@@ -428,6 +432,8 @@ void Executor::RunPartialPreparedContext(ExecutorPrepareContext* ctx,
...
@@ -428,6 +432,8 @@ void Executor::RunPartialPreparedContext(ExecutorPrepareContext* ctx,
int64_t
end_op_index
,
int64_t
end_op_index
,
bool
create_local_scope
,
bool
create_local_scope
,
bool
create_vars
,
bool
keep_kids
)
{
bool
create_vars
,
bool
keep_kids
)
{
platform
::
RecordEvent
record_run
(
"Executor::RunPartialPreparedContext"
,
platform
::
TracerEventType
::
UserDefined
,
1
);
platform
::
RecordBlock
b
(
kProgramId
);
platform
::
RecordBlock
b
(
kProgramId
);
PADDLE_ENFORCE_NOT_NULL
(
PADDLE_ENFORCE_NOT_NULL
(
scope
,
platform
::
errors
::
InvalidArgument
(
"Scope shouldn't be null"
));
scope
,
platform
::
errors
::
InvalidArgument
(
"Scope shouldn't be null"
));
...
@@ -518,6 +524,8 @@ void Executor::RunPartialPreparedContext(ExecutorPrepareContext* ctx,
...
@@ -518,6 +524,8 @@ void Executor::RunPartialPreparedContext(ExecutorPrepareContext* ctx,
auto
&
op
=
ctx
->
ops_
[
i
];
auto
&
op
=
ctx
->
ops_
[
i
];
op
->
Run
(
*
local_scope
,
place_
);
op
->
Run
(
*
local_scope
,
place_
);
if
(
gc
)
{
if
(
gc
)
{
platform
::
RecordEvent
record
(
"CheckGC"
,
platform
::
TracerEventType
::
UserDefined
,
10
);
DeleteUnusedTensors
(
*
local_scope
,
op
.
get
(),
ctx
->
unused_vars_
,
gc
.
get
());
DeleteUnusedTensors
(
*
local_scope
,
op
.
get
(),
ctx
->
unused_vars_
,
gc
.
get
());
}
}
}
}
...
...
paddle/fluid/framework/ir/cost_model.cc
浏览文件 @
dfb47986
...
@@ -44,6 +44,19 @@ double CostData::GetWholeMemoryBytes() const { return whole_memory_bytes_; }
...
@@ -44,6 +44,19 @@ double CostData::GetWholeMemoryBytes() const { return whole_memory_bytes_; }
const
Graph
*
CostData
::
GetGraph
()
const
{
return
graph_
;
}
const
Graph
*
CostData
::
GetGraph
()
const
{
return
graph_
;
}
const
ProgramDesc
*
CostData
::
GetProgram
()
const
{
return
program_
;
}
const
ProgramDesc
*
CostData
::
GetProgram
()
const
{
return
program_
;
}
static
bool
StringHasEnding
(
const
std
::
string
&
full
,
const
std
::
string
&
ending
)
{
if
(
full
.
length
()
<
ending
.
length
())
{
return
false
;
}
if
(
full
.
length
()
==
ending
.
length
())
{
return
full
==
ending
;
}
size_t
prefix_len
=
full
.
length
()
-
ending
.
length
();
return
0
==
full
.
compare
(
prefix_len
,
ending
.
length
(),
ending
)
&&
full
[
prefix_len
-
1
]
==
'/'
;
}
bool
CostData
::
SetCostData
(
const
ProgramDesc
&
program
,
bool
CostData
::
SetCostData
(
const
ProgramDesc
&
program
,
const
std
::
vector
<
std
::
vector
<
Event
>>&
time_events
)
{
const
std
::
vector
<
std
::
vector
<
Event
>>&
time_events
)
{
// TODO(zhhsplendid): Make a copy so that CostData can be available even if
// TODO(zhhsplendid): Make a copy so that CostData can be available even if
...
@@ -77,7 +90,7 @@ bool CostData::SetCostData(const ProgramDesc& program,
...
@@ -77,7 +90,7 @@ bool CostData::SetCostData(const ProgramDesc& program,
std
::
string
op_type
=
op_desc
->
Type
();
std
::
string
op_type
=
op_desc
->
Type
();
while
(
event_index
<
main_thread_events
.
size
())
{
while
(
event_index
<
main_thread_events
.
size
())
{
if
(
main_thread_events
[
event_index
].
name
()
==
op_type
&&
if
(
StringHasEnding
(
main_thread_events
[
event_index
].
name
(),
op_type
)
&&
main_thread_events
[
event_index
].
type
()
==
main_thread_events
[
event_index
].
type
()
==
platform
::
EventType
::
kPushRange
)
{
platform
::
EventType
::
kPushRange
)
{
break
;
break
;
...
@@ -97,7 +110,7 @@ bool CostData::SetCostData(const ProgramDesc& program,
...
@@ -97,7 +110,7 @@ bool CostData::SetCostData(const ProgramDesc& program,
// ControlFlow Op can be like that, but this version only support global
// ControlFlow Op can be like that, but this version only support global
// block
// block
// TODO(zhhsplendid): make a more strict mapping between push and pop
// TODO(zhhsplendid): make a more strict mapping between push and pop
if
(
main_thread_events
[
event_index
].
name
()
==
op_type
&&
if
(
StringHasEnding
(
main_thread_events
[
event_index
].
name
(),
op_type
)
&&
main_thread_events
[
event_index
].
type
()
==
main_thread_events
[
event_index
].
type
()
==
platform
::
EventType
::
kPopRange
)
{
platform
::
EventType
::
kPopRange
)
{
break
;
break
;
...
...
paddle/fluid/framework/new_executor/event_manager.cc
浏览文件 @
dfb47986
...
@@ -13,6 +13,7 @@
...
@@ -13,6 +13,7 @@
// limitations under the License.
// limitations under the License.
#include "paddle/fluid/framework/new_executor/event_manager.h"
#include "paddle/fluid/framework/new_executor/event_manager.h"
#include "paddle/fluid/platform/profiler/event_tracing.h"
namespace
paddle
{
namespace
paddle
{
namespace
framework
{
namespace
framework
{
...
@@ -24,6 +25,8 @@ void WaitEvent(const Instruction& instruction, const platform::Place& place) {
...
@@ -24,6 +25,8 @@ void WaitEvent(const Instruction& instruction, const platform::Place& place) {
VLOG
(
3
)
<<
"Deal StreamWaitEventOrSync for "
<<
instruction
.
OpBase
()
->
Type
();
VLOG
(
3
)
<<
"Deal StreamWaitEventOrSync for "
<<
instruction
.
OpBase
()
->
Type
();
for
(
auto
&
event_iter
:
instruction
.
InputEvents
())
{
for
(
auto
&
event_iter
:
instruction
.
InputEvents
())
{
platform
::
RecordEvent
record
(
"WaitStreamEvent"
,
platform
::
TracerEventType
::
UserDefined
,
10
);
VLOG
(
3
)
<<
"wait var_id: "
<<
event_iter
.
var_id_
VLOG
(
3
)
<<
"wait var_id: "
<<
event_iter
.
var_id_
<<
" 's event with waiter_type: "
<<
event_iter
.
waiter_type_
;
<<
" 's event with waiter_type: "
<<
event_iter
.
waiter_type_
;
event_iter
.
event_
->
Wait
(
event_iter
.
waiter_type_
,
event_iter
.
event_
->
Wait
(
event_iter
.
waiter_type_
,
...
@@ -36,6 +39,8 @@ void RecordEvent(const Instruction& instruction, const platform::Place& place) {
...
@@ -36,6 +39,8 @@ void RecordEvent(const Instruction& instruction, const platform::Place& place) {
if
(
platform
::
is_cpu_place
(
place
))
return
;
if
(
platform
::
is_cpu_place
(
place
))
return
;
for
(
auto
&
event
:
instruction
.
OutputEvents
())
{
for
(
auto
&
event
:
instruction
.
OutputEvents
())
{
platform
::
RecordEvent
record
(
"RecordStreamEvent"
,
platform
::
TracerEventType
::
UserDefined
,
10
);
VLOG
(
3
)
<<
"Record event in out_var_id: "
<<
event
.
var_id_
;
VLOG
(
3
)
<<
"Record event in out_var_id: "
<<
event
.
var_id_
;
event
.
event_
->
Record
(
&
instruction
.
DeviceContext
());
event
.
event_
->
Record
(
&
instruction
.
DeviceContext
());
}
}
...
@@ -46,6 +51,8 @@ void RecordEvent(const Instruction& instruction) {
...
@@ -46,6 +51,8 @@ void RecordEvent(const Instruction& instruction) {
if
(
platform
::
is_cpu_place
(
instruction
.
DeviceContext
().
GetPlace
()))
return
;
if
(
platform
::
is_cpu_place
(
instruction
.
DeviceContext
().
GetPlace
()))
return
;
for
(
auto
&
event
:
instruction
.
OutputEvents
())
{
for
(
auto
&
event
:
instruction
.
OutputEvents
())
{
platform
::
RecordEvent
record
(
"RecordStreamEvent"
,
platform
::
TracerEventType
::
UserDefined
,
10
);
VLOG
(
3
)
<<
"Record event in out_var_id: "
<<
event
.
var_id_
;
VLOG
(
3
)
<<
"Record event in out_var_id: "
<<
event
.
var_id_
;
event
.
event_
->
Record
(
&
instruction
.
DeviceContext
());
event
.
event_
->
Record
(
&
instruction
.
DeviceContext
());
}
}
...
...
paddle/fluid/framework/new_executor/interpretercore.cc
浏览文件 @
dfb47986
...
@@ -489,6 +489,8 @@ void InterpreterCore::RunInstruction(const Instruction& instr_node) {
...
@@ -489,6 +489,8 @@ void InterpreterCore::RunInstruction(const Instruction& instr_node) {
VLOG
(
4
)
<<
"End run "
<<
place
<<
" "
<<
op
->
DebugStringEx
(
global_scope_
);
VLOG
(
4
)
<<
"End run "
<<
place
<<
" "
<<
op
->
DebugStringEx
(
global_scope_
);
if
(
!
instr_node
.
InplaceBackMap
().
empty
())
{
if
(
!
instr_node
.
InplaceBackMap
().
empty
())
{
platform
::
RecordEvent
inplaceback_event
(
"InplaceVarsBack"
,
platform
::
TracerEventType
::
UserDefined
,
10
);
auto
&
m
=
instr_node
.
InplaceBackMap
();
auto
&
m
=
instr_node
.
InplaceBackMap
();
// NOTE(zhiqiu): same logic as TransferInplaceVarsBack() in operator.cc
// NOTE(zhiqiu): same logic as TransferInplaceVarsBack() in operator.cc
for
(
auto
&
p
:
m
)
{
for
(
auto
&
p
:
m
)
{
...
@@ -530,6 +532,8 @@ void InterpreterCore::ExecuteInstructionList(
...
@@ -530,6 +532,8 @@ void InterpreterCore::ExecuteInstructionList(
return
;
return
;
}
}
platform
::
RecordEvent
record_prepare
(
"PrepareAtomic"
,
platform
::
TracerEventType
::
UserDefined
,
1
);
// NOTE(zhiqiu): get the prepared deps from std::future, and async prepare
// NOTE(zhiqiu): get the prepared deps from std::future, and async prepare
// those for the next step
// those for the next step
auto
atomic_deps
=
async_work_queue_
->
AtomicDeps
();
auto
atomic_deps
=
async_work_queue_
->
AtomicDeps
();
...
@@ -537,6 +541,7 @@ void InterpreterCore::ExecuteInstructionList(
...
@@ -537,6 +541,7 @@ void InterpreterCore::ExecuteInstructionList(
async_work_queue_
->
PrepareAtomicDeps
(
dependecy_count_
);
async_work_queue_
->
PrepareAtomicDeps
(
dependecy_count_
);
async_work_queue_
->
PrepareAtomicVarRef
(
global_scope_
->
VecMetaInfo
());
async_work_queue_
->
PrepareAtomicVarRef
(
global_scope_
->
VecMetaInfo
());
record_prepare
.
End
();
exception_holder_
.
Clear
();
exception_holder_
.
Clear
();
...
@@ -573,6 +578,9 @@ void InterpreterCore::RunNextInstructions(
...
@@ -573,6 +578,9 @@ void InterpreterCore::RunNextInstructions(
const
Instruction
&
instr
,
std
::
queue
<
size_t
>*
reserved_next_ops
,
const
Instruction
&
instr
,
std
::
queue
<
size_t
>*
reserved_next_ops
,
std
::
vector
<
std
::
atomic
<
size_t
>>*
atomic_deps
,
std
::
vector
<
std
::
atomic
<
size_t
>>*
atomic_deps
,
std
::
vector
<
std
::
atomic
<
size_t
>>*
atomic_var_ref
)
{
std
::
vector
<
std
::
atomic
<
size_t
>>*
atomic_var_ref
)
{
platform
::
RecordEvent
record
(
"RunNextInstructions"
,
platform
::
TracerEventType
::
UserDefined
,
10
);
VLOG
(
4
)
<<
"atomic 1:"
<<
atomic_deps
;
auto
&
next_instr
=
instr
.
NextInstructions
();
auto
&
next_instr
=
instr
.
NextInstructions
();
auto
IsReady
=
[
atomic_deps
](
size_t
next_id
)
{
auto
IsReady
=
[
atomic_deps
](
size_t
next_id
)
{
...
@@ -708,6 +716,8 @@ void InterpreterCore::RecordStreamForGC(const Instruction& instr) {
...
@@ -708,6 +716,8 @@ void InterpreterCore::RecordStreamForGC(const Instruction& instr) {
instr
.
KernelType
()
!=
OpFuncType
::
kQueueAsync
)
{
instr
.
KernelType
()
!=
OpFuncType
::
kQueueAsync
)
{
return
;
return
;
}
}
platform
::
RecordEvent
record
(
"RecordStreamForGC"
,
platform
::
TracerEventType
::
UserDefined
,
10
);
gpuStream_t
stream
=
reinterpret_cast
<
const
platform
::
CUDADeviceContext
&>
(
gpuStream_t
stream
=
reinterpret_cast
<
const
platform
::
CUDADeviceContext
&>
(
instr
.
DeviceContext
())
instr
.
DeviceContext
())
...
@@ -799,6 +809,8 @@ void InterpreterCore::RecordStreamForGC(const Instruction& instr) {
...
@@ -799,6 +809,8 @@ void InterpreterCore::RecordStreamForGC(const Instruction& instr) {
void
InterpreterCore
::
CheckGC
(
void
InterpreterCore
::
CheckGC
(
const
Instruction
&
instr
,
const
Instruction
&
instr
,
std
::
vector
<
std
::
atomic
<
size_t
>>*
atomic_var_ref
)
{
std
::
vector
<
std
::
atomic
<
size_t
>>*
atomic_var_ref
)
{
platform
::
RecordEvent
record
(
"CheckGC"
,
platform
::
TracerEventType
::
UserDefined
,
10
);
size_t
instr_id
=
instr
.
Id
();
size_t
instr_id
=
instr
.
Id
();
auto
&
var_scope
=
*
global_scope_
;
auto
&
var_scope
=
*
global_scope_
;
...
...
paddle/fluid/framework/new_executor/workqueue/nonblocking_threadpool.h
浏览文件 @
dfb47986
...
@@ -408,8 +408,8 @@ class ThreadPoolTempl {
...
@@ -408,8 +408,8 @@ class ThreadPoolTempl {
ec_
.
Notify
(
true
);
ec_
.
Notify
(
true
);
return
false
;
return
false
;
}
}
platform
::
RecordEvent
(
"Sleep
WaitForWork"
,
platform
::
RecordEvent
record
(
"
WaitForWork"
,
platform
::
TracerEventType
::
UserDefined
,
10
);
platform
::
TracerEventType
::
UserDefined
,
10
);
ec_
.
CommitWait
(
waiter
);
ec_
.
CommitWait
(
waiter
);
blocked_
--
;
blocked_
--
;
return
true
;
return
true
;
...
...
paddle/fluid/framework/new_executor/workqueue/workqueue.cc
浏览文件 @
dfb47986
...
@@ -55,8 +55,9 @@ class WorkQueueImpl : public WorkQueue {
...
@@ -55,8 +55,9 @@ class WorkQueueImpl : public WorkQueue {
}
}
void
AddTask
(
std
::
function
<
void
()
>
fn
)
override
{
void
AddTask
(
std
::
function
<
void
()
>
fn
)
override
{
platform
::
RecordEvent
(
"WorkQueue::AddTask"
,
platform
::
RecordEvent
record
(
"WorkQueue::AddTask"
,
platform
::
TracerEventType
::
UserDefined
,
10
/*level*/
);
platform
::
TracerEventType
::
UserDefined
,
10
/*level*/
);
if
(
tracker_
!=
nullptr
)
{
if
(
tracker_
!=
nullptr
)
{
fn
=
[
fn
=
[
task
=
std
::
move
(
fn
),
raii
=
CounterGuard
<
TaskTracker
>
(
tracker_
)
task
=
std
::
move
(
fn
),
raii
=
CounterGuard
<
TaskTracker
>
(
tracker_
)
...
@@ -146,8 +147,9 @@ WorkQueueGroupImpl::~WorkQueueGroupImpl() {
...
@@ -146,8 +147,9 @@ WorkQueueGroupImpl::~WorkQueueGroupImpl() {
}
}
void
WorkQueueGroupImpl
::
AddTask
(
size_t
queue_idx
,
std
::
function
<
void
()
>
fn
)
{
void
WorkQueueGroupImpl
::
AddTask
(
size_t
queue_idx
,
std
::
function
<
void
()
>
fn
)
{
platform
::
RecordEvent
(
"WorkQueue::AddTask"
,
platform
::
RecordEvent
record
(
"WorkQueue::AddTask"
,
platform
::
TracerEventType
::
UserDefined
,
10
/*level*/
);
platform
::
TracerEventType
::
UserDefined
,
10
/*level*/
);
assert
(
queue_idx
<
queues_
.
size
());
assert
(
queue_idx
<
queues_
.
size
());
if
(
queues_options_
.
at
(
queue_idx
).
track_task
)
{
if
(
queues_options_
.
at
(
queue_idx
).
track_task
)
{
fn
=
[
fn
=
[
...
...
paddle/fluid/framework/parallel_executor.cc
浏览文件 @
dfb47986
...
@@ -916,6 +916,8 @@ void ParallelExecutor::BCastParamsToDevices(
...
@@ -916,6 +916,8 @@ void ParallelExecutor::BCastParamsToDevices(
FetchResultType
ParallelExecutor
::
Run
(
FetchResultType
ParallelExecutor
::
Run
(
const
std
::
vector
<
std
::
string
>
&
fetch_tensors
,
bool
return_merged
)
{
const
std
::
vector
<
std
::
string
>
&
fetch_tensors
,
bool
return_merged
)
{
platform
::
RecordEvent
record_run
(
"ParallelExecutor::Run"
,
platform
::
TracerEventType
::
UserDefined
,
1
);
VLOG
(
3
)
<<
"enter ParallelExecutor Run"
;
VLOG
(
3
)
<<
"enter ParallelExecutor Run"
;
#ifdef PADDLE_WITH_CUDA
#ifdef PADDLE_WITH_CUDA
if
(
platform
::
IsCUDAGraphCapturing
())
{
if
(
platform
::
IsCUDAGraphCapturing
())
{
...
...
paddle/fluid/memory/allocation/auto_growth_best_fit_allocator.cc
浏览文件 @
dfb47986
...
@@ -48,8 +48,9 @@ AutoGrowthBestFitAllocator::AutoGrowthBestFitAllocator(
...
@@ -48,8 +48,9 @@ AutoGrowthBestFitAllocator::AutoGrowthBestFitAllocator(
phi
::
Allocation
*
AutoGrowthBestFitAllocator
::
AllocateImpl
(
phi
::
Allocation
*
AutoGrowthBestFitAllocator
::
AllocateImpl
(
size_t
unaligned_size
)
{
size_t
unaligned_size
)
{
platform
::
RecordEvent
(
"AutoGrowthBestFitAllocator::Allocate"
,
platform
::
RecordEvent
record
(
"AutoGrowthBestFitAllocator::Allocate"
,
platform
::
TracerEventType
::
UserDefined
,
9
/*level*/
);
platform
::
TracerEventType
::
UserDefined
,
9
/*level*/
);
size_t
size
=
AlignedSize
(
unaligned_size
,
alignment_
);
size_t
size
=
AlignedSize
(
unaligned_size
,
alignment_
);
VLOG
(
10
)
<<
"Allocate "
<<
unaligned_size
<<
" bytes, aligned to "
<<
size
;
VLOG
(
10
)
<<
"Allocate "
<<
unaligned_size
<<
" bytes, aligned to "
<<
size
;
...
@@ -111,8 +112,9 @@ phi::Allocation *AutoGrowthBestFitAllocator::AllocateImpl(
...
@@ -111,8 +112,9 @@ phi::Allocation *AutoGrowthBestFitAllocator::AllocateImpl(
}
}
void
AutoGrowthBestFitAllocator
::
FreeImpl
(
phi
::
Allocation
*
allocation
)
{
void
AutoGrowthBestFitAllocator
::
FreeImpl
(
phi
::
Allocation
*
allocation
)
{
platform
::
RecordEvent
(
"AutoGrowthBestFitAllocator::Free"
,
platform
::
RecordEvent
record
(
"AutoGrowthBestFitAllocator::Free"
,
platform
::
TracerEventType
::
UserDefined
,
9
/*level*/
);
platform
::
TracerEventType
::
UserDefined
,
9
/*level*/
);
VLOG
(
10
)
<<
"Free "
<<
allocation
->
size
()
VLOG
(
10
)
<<
"Free "
<<
allocation
->
size
()
<<
" bytes, ptr = "
<<
allocation
->
ptr
();
<<
" bytes, ptr = "
<<
allocation
->
ptr
();
std
::
lock_guard
<
SpinLock
>
guard
(
spinlock_
);
std
::
lock_guard
<
SpinLock
>
guard
(
spinlock_
);
...
...
paddle/fluid/memory/allocation/stream_safe_cuda_allocator.cc
浏览文件 @
dfb47986
...
@@ -163,8 +163,9 @@ void StreamSafeCUDAAllocator::SetDefaultStream(gpuStream_t stream) {
...
@@ -163,8 +163,9 @@ void StreamSafeCUDAAllocator::SetDefaultStream(gpuStream_t stream) {
}
}
phi
::
Allocation
*
StreamSafeCUDAAllocator
::
AllocateImpl
(
size_t
size
)
{
phi
::
Allocation
*
StreamSafeCUDAAllocator
::
AllocateImpl
(
size_t
size
)
{
platform
::
RecordEvent
(
"StreamSafeCUDAAllocator::Allocate"
,
platform
::
RecordEvent
record
(
"StreamSafeCUDAAllocator::Allocate"
,
platform
::
TracerEventType
::
UserDefined
,
9
/*level*/
);
platform
::
TracerEventType
::
UserDefined
,
9
/*level*/
);
ProcessUnfreedAllocations
();
ProcessUnfreedAllocations
();
VLOG
(
8
)
<<
"Try allocate "
<<
size
<<
" bytes"
;
VLOG
(
8
)
<<
"Try allocate "
<<
size
<<
" bytes"
;
AllocationPtr
underlying_allocation
;
AllocationPtr
underlying_allocation
;
...
@@ -192,8 +193,9 @@ phi::Allocation* StreamSafeCUDAAllocator::AllocateImpl(size_t size) {
...
@@ -192,8 +193,9 @@ phi::Allocation* StreamSafeCUDAAllocator::AllocateImpl(size_t size) {
}
}
void
StreamSafeCUDAAllocator
::
FreeImpl
(
phi
::
Allocation
*
allocation
)
{
void
StreamSafeCUDAAllocator
::
FreeImpl
(
phi
::
Allocation
*
allocation
)
{
platform
::
RecordEvent
(
"StreamSafeCUDAAllocator::Free"
,
platform
::
RecordEvent
record
(
"StreamSafeCUDAAllocator::Free"
,
platform
::
TracerEventType
::
UserDefined
,
9
/*level*/
);
platform
::
TracerEventType
::
UserDefined
,
9
/*level*/
);
StreamSafeCUDAAllocation
*
stream_safe_cuda_allocation
=
StreamSafeCUDAAllocation
*
stream_safe_cuda_allocation
=
static_cast
<
StreamSafeCUDAAllocation
*>
(
allocation
);
static_cast
<
StreamSafeCUDAAllocation
*>
(
allocation
);
...
...
paddle/fluid/pybind/pybind.cc
浏览文件 @
dfb47986
...
@@ -2867,7 +2867,7 @@ All parameter, weight, gradient are variables in Paddle.
...
@@ -2867,7 +2867,7 @@ All parameter, weight, gradient are variables in Paddle.
[](
StandaloneExecutor
&
self
,
std
::
vector
<
std
::
string
>
feed_names
,
[](
StandaloneExecutor
&
self
,
std
::
vector
<
std
::
string
>
feed_names
,
std
::
vector
<
std
::
string
>
fetch_names
)
{
std
::
vector
<
std
::
string
>
fetch_names
)
{
platform
::
RecordEvent
record_event
(
platform
::
RecordEvent
record_event
(
"StandaloneExecutor:run"
,
"StandaloneExecutor:
:
run"
,
platform
::
TracerEventType
::
UserDefined
,
1
);
platform
::
TracerEventType
::
UserDefined
,
1
);
paddle
::
framework
::
FetchList
ret
;
paddle
::
framework
::
FetchList
ret
;
{
{
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录