Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
MegEngine 天元
MegEngine
提交
ef04619c
MegEngine
项目概览
MegEngine 天元
/
MegEngine
1 年多 前同步成功
通知
403
Star
4705
Fork
582
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
DevOps
流水线
流水线任务
计划
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
MegEngine
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
DevOps
DevOps
流水线
流水线任务
计划
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
流水线任务
提交
Issue看板
提交
ef04619c
编写于
12月 23, 2022
作者:
M
Megvii Engine Team
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
feat(profiler): record python api scope
GitOrigin-RevId: bd6929f33ffeca8d73f81156aee7bea2600e887c
上级
5bcb8435
变更
14
隐藏空白更改
内联
并排
Showing
14 changed file
with
157 addition
and
21 deletion
+157
-21
imperative/python/megengine/traced_module/traced_module.py
imperative/python/megengine/traced_module/traced_module.py
+1
-0
imperative/python/megengine/utils/profiler.py
imperative/python/megengine/utils/profiler.py
+65
-0
imperative/python/src/tensor.cpp
imperative/python/src/tensor.cpp
+23
-0
imperative/src/impl/interpreter/commands.h
imperative/src/impl/interpreter/commands.h
+5
-4
imperative/src/impl/interpreter/interpreter_impl.cpp
imperative/src/impl/interpreter/interpreter_impl.cpp
+7
-7
imperative/src/impl/interpreter/interpreter_impl.h
imperative/src/impl/interpreter/interpreter_impl.h
+2
-2
imperative/src/impl/profiler/events.h
imperative/src/impl/profiler/events.h
+4
-1
imperative/src/impl/profiler/memory_chunk.cpp
imperative/src/impl/profiler/memory_chunk.cpp
+1
-2
imperative/src/impl/transformations/eval.cpp
imperative/src/impl/transformations/eval.cpp
+6
-0
imperative/src/impl/transformations/grad.cpp
imperative/src/impl/transformations/grad.cpp
+9
-2
imperative/src/impl/transformations/lazy.cpp
imperative/src/impl/transformations/lazy.cpp
+2
-0
imperative/src/include/megbrain/imperative/basic_operators.h
imperative/src/include/megbrain/imperative/basic_operators.h
+19
-0
imperative/src/include/megbrain/imperative/interpreter.h
imperative/src/include/megbrain/imperative/interpreter.h
+5
-3
imperative/src/include/megbrain/imperative/profiler.h
imperative/src/include/megbrain/imperative/profiler.h
+8
-0
未找到文件。
imperative/python/megengine/traced_module/traced_module.py
浏览文件 @
ef04619c
...
@@ -290,6 +290,7 @@ def _wrap_mnode_call(orig_call):
...
@@ -290,6 +290,7 @@ def _wrap_mnode_call(orig_call):
active_module_tracer
().
current_scope
().
_add_input
(
self
)
active_module_tracer
().
current_scope
().
_add_input
(
self
)
rst
=
obj
(
*
args
,
**
kwargs
)
rst
=
obj
(
*
args
,
**
kwargs
)
else
:
else
:
raise
TypeError
(
"'ModuleNode' object is not callable"
)
raise
TypeError
(
"'ModuleNode' object is not callable"
)
return
rst
return
rst
...
...
imperative/python/megengine/utils/profiler.py
浏览文件 @
ef04619c
...
@@ -8,13 +8,16 @@ from typing import List
...
@@ -8,13 +8,16 @@ from typing import List
from
weakref
import
WeakSet
from
weakref
import
WeakSet
from
..
import
_atexit
from
..
import
_atexit
from
..core._imperative_rt.core2
import
Tensor
as
raw_tensor
from
..core._imperative_rt.core2
import
(
from
..core._imperative_rt.core2
import
(
cupti_available
,
cupti_available
,
disable_cupti
,
disable_cupti
,
enable_cupti
,
enable_cupti
,
full_sync
,
full_sync
,
pop_scope
,
pop_scope
,
pop_scope_with_type
,
push_scope
,
push_scope
,
push_scope_with_type
,
set_python_backtrace_enabled
,
set_python_backtrace_enabled
,
start_profile
,
start_profile
,
stop_profile
,
stop_profile
,
...
@@ -33,6 +36,7 @@ class Profiler(ContextDecorator):
...
@@ -33,6 +36,7 @@ class Profiler(ContextDecorator):
Args:
Args:
path: default path prefix for profiler to dump.
path: default path prefix for profiler to dump.
with_backtrace: Whether to record backtrace information for ops.
with_backtrace: Whether to record backtrace information for ops.
with_scopes: Whether to keep more scopes to record record module/functional hierarchy. Enabling this option will slow down your program execution.
Examples:
Examples:
...
@@ -70,6 +74,7 @@ class Profiler(ContextDecorator):
...
@@ -70,6 +74,7 @@ class Profiler(ContextDecorator):
format
:
str
=
"chrome_timeline.json"
,
format
:
str
=
"chrome_timeline.json"
,
formats
:
List
[
str
]
=
None
,
formats
:
List
[
str
]
=
None
,
with_backtrace
:
bool
=
False
,
with_backtrace
:
bool
=
False
,
with_scopes
:
bool
=
False
,
**
kwargs
**
kwargs
)
->
None
:
)
->
None
:
if
not
formats
:
if
not
formats
:
...
@@ -89,6 +94,8 @@ class Profiler(ContextDecorator):
...
@@ -89,6 +94,8 @@ class Profiler(ContextDecorator):
self
.
_options
[
opt
]
=
int
(
kwargs
.
pop
(
opt
,
optval
))
self
.
_options
[
opt
]
=
int
(
kwargs
.
pop
(
opt
,
optval
))
self
.
_pid
=
"<PID>"
self
.
_pid
=
"<PID>"
self
.
_dump_callback
=
None
self
.
_dump_callback
=
None
self
.
_api_patcher
=
None
self
.
_with_scopes
=
with_scopes
if
self
.
_options
.
get
(
"enable_cupti"
,
0
):
if
self
.
_options
.
get
(
"enable_cupti"
,
0
):
if
cupti_available
():
if
cupti_available
():
enable_cupti
()
enable_cupti
()
...
@@ -110,6 +117,59 @@ class Profiler(ContextDecorator):
...
@@ -110,6 +117,59 @@ class Profiler(ContextDecorator):
def
directory
(
self
):
def
directory
(
self
):
return
self
.
_path
return
self
.
_path
@
property
def
_patcher
(
self
):
if
self
.
_api_patcher
!=
None
:
return
self
.
_api_patcher
from
..traced_module.module_tracer
import
Patcher
,
module_tracer
from
..module
import
Module
def
wrap_tensormethod_and_functional
(
origin_fn
):
def
get_tensormeth_name
(
obj
,
func
):
tp
=
obj
if
isinstance
(
obj
,
type
)
else
type
(
obj
)
if
not
issubclass
(
tp
,
raw_tensor
):
return
None
for
cls
in
tp
.
mro
():
for
k
,
v
in
cls
.
__dict__
.
items
():
if
v
==
func
:
return
k
return
None
@
wraps
(
origin_fn
)
def
wrapped_fn
(
*
args
,
**
kwargs
):
methname
=
(
get_tensormeth_name
(
args
[
0
],
wrapped_fn
)
if
len
(
args
)
>
0
else
None
)
name
,
scope_type
=
(
(
"tensor."
+
methname
,
"tensor_method"
)
if
methname
is
not
None
else
(
origin_fn
.
__name__
,
"functional"
)
)
push_scope_with_type
(
name
,
scope_type
)
rst
=
origin_fn
(
*
args
,
**
kwargs
)
pop_scope_with_type
(
name
,
scope_type
)
return
rst
return
wrapped_fn
def
wrap_module_call
(
origin_fn
):
@
wraps
(
origin_fn
)
def
wrapped_fn
(
*
args
,
**
kwargs
):
is_builtin_module
=
module_tracer
.
is_builtin
(
type
(
args
[
0
]))
if
not
is_builtin_module
:
return
origin_fn
(
*
args
,
**
kwargs
)
name
,
scope_type
=
type
(
args
[
0
]).
__name__
,
"module"
push_scope_with_type
(
name
,
scope_type
)
rst
=
origin_fn
(
*
args
,
**
kwargs
)
pop_scope_with_type
(
name
,
scope_type
)
return
rst
return
wrapped_fn
self
.
_api_patcher
=
Patcher
(
wrap_tensormethod_and_functional
)
self
.
_api_patcher
.
patch_method
(
Module
,
"__call__"
,
wrap_module_call
)
return
self
.
_api_patcher
@
property
@
property
def
formats
(
self
):
def
formats
(
self
):
return
list
(
self
.
_formats
)
return
list
(
self
.
_formats
)
...
@@ -171,9 +231,14 @@ class Profiler(ContextDecorator):
...
@@ -171,9 +231,14 @@ class Profiler(ContextDecorator):
def
__enter__
(
self
):
def
__enter__
(
self
):
self
.
start
()
self
.
start
()
if
self
.
_with_scopes
:
self
.
_patcher
.
__enter__
()
def
__exit__
(
self
,
val
,
tp
,
trace
):
def
__exit__
(
self
,
val
,
tp
,
trace
):
self
.
stop
()
self
.
stop
()
if
self
.
_with_scopes
and
self
.
_api_patcher
is
not
None
:
self
.
_api_patcher
.
__exit__
(
val
,
tp
,
trace
)
self
.
_api_patcher
=
None
def
__call__
(
self
,
func
):
def
__call__
(
self
,
func
):
func
=
super
().
__call__
(
func
)
func
=
super
().
__call__
(
func
)
...
...
imperative/python/src/tensor.cpp
浏览文件 @
ef04619c
...
@@ -948,6 +948,29 @@ void init_tensor(py::module m) {
...
@@ -948,6 +948,29 @@ void init_tensor(py::module m) {
channel
->
pop_scope
(
name
);
channel
->
pop_scope
(
name
);
Transformation
::
pop_scope
(
name
);
Transformation
::
pop_scope
(
name
);
});
});
std
::
unordered_map
<
std
::
string
,
ScopeType
>
str2scopetype
=
{
{
"default"
,
ScopeType
::
DEFAULT
},
{
"module"
,
ScopeType
::
MODULE
},
{
"tensor_method"
,
ScopeType
::
TENSOR_METHOD
},
{
"functional"
,
ScopeType
::
FUNCTIONAL
},
{
"backward"
,
ScopeType
::
BACKWARD
}};
m
.
def
(
"push_scope_with_type"
,
[
channel
,
str2scopetype
](
std
::
string
name
,
std
::
string
type
)
{
if
(
str2scopetype
.
find
(
type
)
==
str2scopetype
.
end
())
{
throw
py
::
value_error
(
"unsupport scope type"
);
}
else
{
channel
->
push_scope
(
name
,
str2scopetype
.
find
(
type
)
->
second
);
}
});
m
.
def
(
"pop_scope_with_type"
,
[
channel
,
str2scopetype
](
std
::
string
name
,
std
::
string
type
)
{
if
(
str2scopetype
.
find
(
type
)
==
str2scopetype
.
end
())
{
throw
py
::
value_error
(
"unsupport scope type"
);
}
else
{
channel
->
pop_scope
(
name
,
str2scopetype
.
find
(
type
)
->
second
);
}
});
m
.
def
(
"start_profile"
,
[
channel
](
imperative
::
Profiler
::
options_t
options
)
{
m
.
def
(
"start_profile"
,
[
channel
](
imperative
::
Profiler
::
options_t
options
)
{
channel
->
sync
();
channel
->
sync
();
imperative
::
Profiler
::
load_options
(
std
::
move
(
options
));
imperative
::
Profiler
::
load_options
(
std
::
move
(
options
));
...
...
imperative/src/impl/interpreter/commands.h
浏览文件 @
ef04619c
...
@@ -4,14 +4,14 @@
...
@@ -4,14 +4,14 @@
#include <unordered_set>
#include <unordered_set>
#include <variant>
#include <variant>
#include "./stack_manager.h"
#include "./tensor_info.h"
#include "megbrain/imperative/backtrace.h"
#include "megbrain/imperative/backtrace.h"
#include "megbrain/imperative/op_def.h"
#include "megbrain/imperative/op_def.h"
#include "megbrain/imperative/profiler.h"
#include "megbrain/imperative/utils/to_string.h"
#include "megbrain/imperative/utils/to_string.h"
#include "megbrain/tensor.h"
#include "megbrain/tensor.h"
#include "./stack_manager.h"
#include "./tensor_info.h"
namespace
mgb
::
imperative
{
namespace
mgb
::
imperative
{
namespace
interpreter
::
intl
{
namespace
interpreter
::
intl
{
...
@@ -125,6 +125,7 @@ struct StopStep {
...
@@ -125,6 +125,7 @@ struct StopStep {
struct
PushScope
{
struct
PushScope
{
std
::
string
scope_name
;
std
::
string
scope_name
;
ScopeType
type
;
template
<
typename
TFunctor
>
template
<
typename
TFunctor
>
void
get_props
(
TFunctor
&&
functor
)
const
{
void
get_props
(
TFunctor
&&
functor
)
const
{
...
@@ -136,7 +137,7 @@ struct PushScope {
...
@@ -136,7 +137,7 @@ struct PushScope {
struct
PopScope
{
struct
PopScope
{
std
::
string
scope_name
;
std
::
string
scope_name
;
ScopeType
type
;
template
<
typename
TFunctor
>
template
<
typename
TFunctor
>
void
get_props
(
TFunctor
&&
functor
)
const
{
void
get_props
(
TFunctor
&&
functor
)
const
{
functor
(
"scope_name"
,
scope_name
);
functor
(
"scope_name"
,
scope_name
);
...
...
imperative/src/impl/interpreter/interpreter_impl.cpp
浏览文件 @
ef04619c
...
@@ -1356,7 +1356,7 @@ void ChannelImpl::process_one_task(Command& icmd) {
...
@@ -1356,7 +1356,7 @@ void ChannelImpl::process_one_task(Command& icmd) {
}
else
if
constexpr
(
std
::
is_same_v
<
T
,
StopStep
>
)
{
}
else
if
constexpr
(
std
::
is_same_v
<
T
,
StopStep
>
)
{
MGB_RECORD_EVENT
(
StopStepEvent
);
MGB_RECORD_EVENT
(
StopStepEvent
);
}
else
if
constexpr
(
std
::
is_same_v
<
T
,
PushScope
>
)
{
}
else
if
constexpr
(
std
::
is_same_v
<
T
,
PushScope
>
)
{
MGB_RECORD_EVENT
(
ScopeEvent
,
cmd
.
scope_name
);
MGB_RECORD_EVENT
(
ScopeEvent
,
cmd
.
scope_name
,
cmd
.
type
);
}
else
if
constexpr
(
std
::
is_same_v
<
T
,
PopScope
>
)
{
}
else
if
constexpr
(
std
::
is_same_v
<
T
,
PopScope
>
)
{
MGB_RECORD_EVENT
(
ScopeFinishEvent
,
cmd
.
scope_name
);
MGB_RECORD_EVENT
(
ScopeFinishEvent
,
cmd
.
scope_name
);
}
else
if
constexpr
(
std
::
is_same_v
<
T
,
StartRegen
>
)
{
}
else
if
constexpr
(
std
::
is_same_v
<
T
,
StartRegen
>
)
{
...
@@ -1461,15 +1461,15 @@ void ChannelImpl::stop_step() {
...
@@ -1461,15 +1461,15 @@ void ChannelImpl::stop_step() {
get_channel_state
().
stack_manager
.
dump
()});
get_channel_state
().
stack_manager
.
dump
()});
}
}
void
ChannelImpl
::
push_scope
(
std
::
string
name
)
{
void
ChannelImpl
::
push_scope
(
std
::
string
name
,
ScopeType
type
)
{
MGB_LOCK_GUARD
(
m_spin
);
MGB_LOCK_GUARD
(
m_spin
);
assert_available
();
assert_available
();
auto
&
state
=
get_channel_state
();
auto
&
state
=
get_channel_state
();
state
.
stack_manager
.
enter
(
name
);
state
.
stack_manager
.
enter
(
name
);
MGB_RECORD_EVENT
(
ScopeEvent
,
name
);
MGB_RECORD_EVENT
(
ScopeEvent
,
name
,
type
);
if
(
Profiler
::
is_profiling
())
{
if
(
Profiler
::
is_profiling
())
{
m_worker
.
add_task
(
m_worker
.
add_task
(
{
Profiler
::
next_id
(),
PushScope
{
name
},
{
Profiler
::
next_id
(),
PushScope
{
name
,
type
},
get_channel_state
().
stack_manager
.
dump
()});
get_channel_state
().
stack_manager
.
dump
()});
}
else
{
}
else
{
m_worker
.
add_task
({
m_worker
.
add_task
({
...
@@ -1479,15 +1479,15 @@ void ChannelImpl::push_scope(std::string name) {
...
@@ -1479,15 +1479,15 @@ void ChannelImpl::push_scope(std::string name) {
}
}
}
}
void
ChannelImpl
::
pop_scope
(
std
::
string
name
)
{
void
ChannelImpl
::
pop_scope
(
std
::
string
name
,
ScopeType
type
)
{
MGB_LOCK_GUARD
(
m_spin
);
MGB_LOCK_GUARD
(
m_spin
);
assert_available
();
assert_available
();
auto
&
state
=
get_channel_state
();
auto
&
state
=
get_channel_state
();
state
.
stack_manager
.
exit
(
name
);
state
.
stack_manager
.
exit
(
name
);
MGB_RECORD_EVENT
(
ScopeFinishEvent
,
name
);
MGB_RECORD_EVENT
(
ScopeFinishEvent
,
name
,
type
);
if
(
Profiler
::
is_profiling
())
{
if
(
Profiler
::
is_profiling
())
{
m_worker
.
add_task
(
m_worker
.
add_task
(
{
Profiler
::
next_id
(),
PopScope
{
name
},
{
Profiler
::
next_id
(),
PopScope
{
name
,
type
},
get_channel_state
().
stack_manager
.
dump
()});
get_channel_state
().
stack_manager
.
dump
()});
}
else
{
}
else
{
m_worker
.
add_task
({
m_worker
.
add_task
({
...
...
imperative/src/impl/interpreter/interpreter_impl.h
浏览文件 @
ef04619c
...
@@ -65,8 +65,8 @@ struct ChannelImpl : Interpreter::Channel, NonCopyableObj, NonMoveableObj {
...
@@ -65,8 +65,8 @@ struct ChannelImpl : Interpreter::Channel, NonCopyableObj, NonMoveableObj {
void
stop_profile
()
override
;
void
stop_profile
()
override
;
void
stop_step
()
override
;
void
stop_step
()
override
;
void
push_scope
(
std
::
string
)
override
;
void
push_scope
(
std
::
string
,
ScopeType
type
=
ScopeType
::
DEFAULT
)
override
;
void
pop_scope
(
std
::
string
)
override
;
void
pop_scope
(
std
::
string
,
ScopeType
type
=
ScopeType
::
DEFAULT
)
override
;
BackTraceInfoPtr
&
get_backtrace
()
override
;
BackTraceInfoPtr
&
get_backtrace
()
override
;
void
set_backtrace
(
BackTraceInfoPtr
bt
)
override
;
void
set_backtrace
(
BackTraceInfoPtr
bt
)
override
;
...
...
imperative/src/impl/profiler/events.h
浏览文件 @
ef04619c
...
@@ -139,7 +139,10 @@ DEF_EVENT(WorkerException, {});
...
@@ -139,7 +139,10 @@ DEF_EVENT(WorkerException, {});
DEF_EVENT
(
ShapeInfer
,
{
bool
success
;
});
DEF_EVENT
(
ShapeInfer
,
{
bool
success
;
});
DEF_DUR_EVENT
(
Scope
,
{
std
::
string
name
;
});
DEF_DUR_EVENT
(
Scope
,
{
std
::
string
name
;
ScopeType
type
=
ScopeType
::
DEFAULT
;
});
DEF_DUR_EVENT
(
Sync
,
{
Trace
trace
;
});
DEF_DUR_EVENT
(
Sync
,
{
Trace
trace
;
});
...
...
imperative/src/impl/profiler/memory_chunk.cpp
浏览文件 @
ef04619c
...
@@ -5,11 +5,10 @@
...
@@ -5,11 +5,10 @@
#include "megbrain/imperative/utils/to_string.h"
#include "megbrain/imperative/utils/to_string.h"
#include "megbrain/utils/debug.h"
#include "megbrain/utils/debug.h"
#include "./events.h"
#include "./formats.h"
#include "./formats.h"
#include "./states.h"
#include "./states.h"
#include "./events.h"
namespace
mgb
::
imperative
::
profiler
{
namespace
mgb
::
imperative
::
profiler
{
class
XMLWriter
{
class
XMLWriter
{
...
...
imperative/src/impl/transformations/eval.cpp
浏览文件 @
ef04619c
...
@@ -135,6 +135,12 @@ ValueRefList InterpreterTransformation::apply_transformation(
...
@@ -135,6 +135,12 @@ ValueRefList InterpreterTransformation::apply_transformation(
DeviceTensorND
dev_tensor
;
DeviceTensorND
dev_tensor
;
dev_tensor
.
copy_from
(
m_channel
->
get_dev_tensor
(
input
.
handle
()
->
handle
()));
dev_tensor
.
copy_from
(
m_channel
->
get_dev_tensor
(
input
.
handle
()
->
handle
()));
return
m_value_type
.
make
(
share_handle
(
m_channel
->
put
(
dev_tensor
,
{})));
return
m_value_type
.
make
(
share_handle
(
m_channel
->
put
(
dev_tensor
,
{})));
}
else
if
(
auto
push_scope
=
op
.
as
<
PushScope
>
())
{
m_channel
->
push_scope
(
push_scope
->
name
,
push_scope
->
type
);
return
{};
}
else
if
(
auto
pop_scope
=
op
.
as
<
PopScope
>
())
{
m_channel
->
pop_scope
(
pop_scope
->
name
,
pop_scope
->
type
);
return
{};
}
else
{
}
else
{
return
op
.
fallback
(
inputs
);
return
op
.
fallback
(
inputs
);
}
}
...
...
imperative/src/impl/transformations/grad.cpp
浏览文件 @
ef04619c
...
@@ -2,11 +2,11 @@
...
@@ -2,11 +2,11 @@
#include <variant>
#include <variant>
#include <range/v3/all.hpp>
#include "megbrain/imperative/graph_cache.h"
#include "megbrain/imperative/graph_cache.h"
#include "megbrain/imperative/profiler.h"
#include "megbrain/imperative/resource_manager.h"
#include "megbrain/imperative/resource_manager.h"
#include <range/v3/all.hpp>
namespace
mgb
{
namespace
mgb
{
namespace
imperative
{
namespace
imperative
{
...
@@ -232,7 +232,14 @@ void GradKey::backward() {
...
@@ -232,7 +232,14 @@ void GradKey::backward() {
for
(
auto
&&
slot
:
grad_fn
->
m_slots
)
{
for
(
auto
&&
slot
:
grad_fn
->
m_slots
)
{
*
iter
++
=
slot
.
m_grad
;
*
iter
++
=
slot
.
m_grad
;
}
}
std
::
string
name
=
op
?
op
->
name
()
+
"Backward"
:
"CustomBackward"
;
if
(
Profiler
::
is_profiling
())
{
imperative
::
apply
(
PushScope
(
name
,
ScopeType
::
BACKWARD
),
Span
<
ValueRef
>
(
nullptr
,
nullptr
));
}
backward
(
grads
,
grad_receiver
);
backward
(
grads
,
grad_receiver
);
if
(
Profiler
::
is_profiling
())
{
imperative
::
apply
(
PopScope
(
name
,
ScopeType
::
BACKWARD
),
Span
<
ValueRef
>
(
nullptr
,
nullptr
));
}
}
}
},
grad_fn
->
m_backward
);
},
grad_fn
->
m_backward
);
// clang-format on
// clang-format on
...
...
imperative/src/impl/transformations/lazy.cpp
浏览文件 @
ef04619c
...
@@ -172,6 +172,8 @@ ValueRefList LazyEvalTransformation::apply_transformation(
...
@@ -172,6 +172,8 @@ ValueRefList LazyEvalTransformation::apply_transformation(
}
else
{
}
else
{
return
imperative
::
apply
(
op
,
inputs
);
return
imperative
::
apply
(
op
,
inputs
);
}
}
}
else
if
(
op
.
is
<
PushScope
>
()
||
op
.
is
<
PopScope
>
())
{
return
{};
}
else
{
}
else
{
return
op
.
fallback
(
inputs
);
return
op
.
fallback
(
inputs
);
}
}
...
...
imperative/src/include/megbrain/imperative/basic_operators.h
浏览文件 @
ef04619c
...
@@ -5,6 +5,7 @@
...
@@ -5,6 +5,7 @@
#include "megbrain/imperative/op_def.h"
#include "megbrain/imperative/op_def.h"
#include "megbrain/imperative/operator.h"
#include "megbrain/imperative/operator.h"
#include "megbrain/imperative/profiler.h"
#include "megbrain/imperative/utils/data_format.h"
#include "megbrain/imperative/utils/data_format.h"
#include "megbrain/imperative/utils/helper.h"
#include "megbrain/imperative/utils/helper.h"
#include "megbrain/imperative/utils/value_shape.h"
#include "megbrain/imperative/utils/value_shape.h"
...
@@ -206,6 +207,24 @@ public:
...
@@ -206,6 +207,24 @@ public:
std
::
string
to_string
()
const
override
;
std
::
string
to_string
()
const
override
;
};
};
class
PushScope
final
:
public
OperatorImpl
<
PushScope
>
{
public:
std
::
string
name
;
ScopeType
type
;
PushScope
(
std
::
string
name
,
ScopeType
type
)
:
name
{
std
::
move
(
name
)},
type
{
type
}
{};
std
::
string
raw_type
()
const
{
return
"PushScope"
;
}
std
::
string
to_string
()
const
override
{
return
"PushScope"
;
}
};
class
PopScope
final
:
public
OperatorImpl
<
PopScope
>
{
public:
std
::
string
name
;
ScopeType
type
;
PopScope
(
std
::
string
name
,
ScopeType
type
)
:
name
{
std
::
move
(
name
)},
type
{
type
}
{};
std
::
string
raw_type
()
const
{
return
"PopScope"
;
}
std
::
string
to_string
()
const
override
{
return
"PopScope"
;
}
};
class
DupTensor
final
:
public
OperatorImpl
<
DupTensor
,
Operator
::
IdentityLike
>
{
class
DupTensor
final
:
public
OperatorImpl
<
DupTensor
,
Operator
::
IdentityLike
>
{
public:
public:
std
::
string
to_string
()
const
override
{
return
"DupTensor"
;
}
std
::
string
to_string
()
const
override
{
return
"DupTensor"
;
}
...
...
imperative/src/include/megbrain/imperative/interpreter.h
浏览文件 @
ef04619c
...
@@ -5,6 +5,7 @@
...
@@ -5,6 +5,7 @@
#include "./backtrace.h"
#include "./backtrace.h"
#include "megbrain/imperative/op_def.h"
#include "megbrain/imperative/op_def.h"
#include "megbrain/imperative/profiler.h"
namespace
mgb
::
imperative
::
interpreter
{
namespace
mgb
::
imperative
::
interpreter
{
...
@@ -63,12 +64,13 @@ struct Interpreter {
...
@@ -63,12 +64,13 @@ struct Interpreter {
virtual
void
stop_profile
()
=
0
;
virtual
void
stop_profile
()
=
0
;
virtual
void
stop_step
()
=
0
;
virtual
void
stop_step
()
=
0
;
virtual
void
push_scope
(
std
::
string
name
)
=
0
;
virtual
void
pop_scope
(
std
::
string
name
)
=
0
;
virtual
BackTraceInfoPtr
&
get_backtrace
()
=
0
;
virtual
BackTraceInfoPtr
&
get_backtrace
()
=
0
;
virtual
void
set_backtrace
(
BackTraceInfoPtr
bt
)
=
0
;
virtual
void
set_backtrace
(
BackTraceInfoPtr
bt
)
=
0
;
virtual
void
clear_backtrace
()
=
0
;
virtual
void
clear_backtrace
()
=
0
;
virtual
void
push_scope
(
std
::
string
name
,
ScopeType
type
=
ScopeType
::
DEFAULT
)
=
0
;
virtual
void
pop_scope
(
std
::
string
name
,
ScopeType
type
=
ScopeType
::
DEFAULT
)
=
0
;
};
};
virtual
std
::
unique_ptr
<
Channel
>
create_channel
()
=
0
;
virtual
std
::
unique_ptr
<
Channel
>
create_channel
()
=
0
;
...
...
imperative/src/include/megbrain/imperative/profiler.h
浏览文件 @
ef04619c
...
@@ -41,6 +41,14 @@ public:
...
@@ -41,6 +41,14 @@ public:
static
std
::
shared_ptr
<
CompNode
::
Event
>
record_device
(
CompNode
device
);
static
std
::
shared_ptr
<
CompNode
::
Event
>
record_device
(
CompNode
device
);
};
};
enum
ScopeType
{
DEFAULT
=
0
,
MODULE
=
1
,
TENSOR_METHOD
=
2
,
FUNCTIONAL
=
3
,
BACKWARD
=
4
,
};
class
Profiler
{
class
Profiler
{
public:
public:
struct
Record
{
struct
Record
{
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录