Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
magicwindyyd
mindspore
提交
4bede54f
M
mindspore
项目概览
magicwindyyd
/
mindspore
与 Fork 源项目一致
Fork自
MindSpore / mindspore
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
M
mindspore
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
4bede54f
编写于
6月 02, 2020
作者:
K
kpy
提交者:
kuangpeiyu
6月 02, 2020
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
fix pynative refactor bug
上级
c547c8ca
变更
2
隐藏空白更改
内联
并排
Showing
2 changed file
with
44 addition
and
24 deletion
+44
-24
mindspore/ccsrc/pynative/pynative_execute.cc
mindspore/ccsrc/pynative/pynative_execute.cc
+43
-23
mindspore/ccsrc/pynative/pynative_execute.h
mindspore/ccsrc/pynative/pynative_execute.h
+1
-1
未找到文件。
mindspore/ccsrc/pynative/pynative_execute.cc
浏览文件 @
4bede54f
...
...
@@ -76,6 +76,9 @@ std::string GetId(const py::object &obj) {
std
::
string
prefix
=
""
;
if
(
py
::
isinstance
<
py
::
tuple
>
(
to_process
))
{
auto
p_list
=
py
::
cast
<
py
::
tuple
>
(
to_process
);
if
(
p_list
.
size
()
==
0
)
{
return
"empty"
;
}
to_process
=
p_list
[
0
];
prefix
=
"tuple:"
;
if
(
!
py
::
isinstance
<
tensor
::
Tensor
>
(
to_process
))
{
...
...
@@ -101,14 +104,24 @@ std::string GetId(const py::object &obj) {
return
py
::
cast
<
std
::
string
>
(
ret
);
}
py
::
list
ConvertInputs
(
const
PrimitivePyPtr
&
prim
,
const
py
::
list
&
py_args
)
{
py
::
object
GetTupleObj
(
const
py
::
object
&
obj
)
{
py
::
module
mod
=
parse
::
python_adapter
::
GetPyModule
(
parse
::
PYTHON_MOD_PARSE_MODULE
);
py
::
object
obj_tuple
=
parse
::
python_adapter
::
CallPyModFn
(
mod
,
parse
::
PYTHON_MOD_GET_DEFAULT_INPUT
,
obj
);
return
obj_tuple
;
}
void
ConvertInputs
(
const
PrimitivePyPtr
&
prim
,
const
py
::
list
&
args
,
py
::
tuple
*
out_args
)
{
auto
&
py_args
=
*
out_args
;
for
(
size_t
i
=
0
;
i
<
args
.
size
();
++
i
)
{
py_args
[
i
]
=
GetTupleObj
(
args
[
i
]);
}
auto
signature
=
prim
->
signatures
();
std
::
vector
<
SignatureEnumDType
>
dtypes
;
(
void
)
std
::
transform
(
signature
.
begin
(),
signature
.
end
(),
std
::
back_inserter
(
dtypes
),
[](
const
Signature
&
sig
)
{
return
sig
.
dtype
;
});
int
empty_dtype_count
=
std
::
count
(
dtypes
.
begin
(),
dtypes
.
end
(),
SignatureEnumDType
::
kDTypeEmptyDefaultValue
);
if
(
dtypes
.
size
()
==
0
||
static_cast
<
int
>
(
dtypes
.
size
())
==
empty_dtype_count
)
{
return
py_args
;
return
;
}
std
::
map
<
SignatureEnumDType
,
std
::
vector
<
size_t
>>
type_indexs
;
for
(
size_t
i
=
0
;
i
<
dtypes
.
size
();
++
i
)
{
...
...
@@ -134,22 +147,19 @@ py::list ConvertInputs(const PrimitivePyPtr &prim, const py::list &py_args) {
}
(
void
)
dst_type
.
insert
(
std
::
make_pair
(
type
,
m_index
));
}
py
::
list
py_inputs
(
py_args
.
size
());
for
(
size_t
i
=
0
;
i
<
py_args
.
size
();
++
i
)
{
auto
it
=
dst_type
.
find
(
dtypes
[
i
]);
if
(
it
!=
dst_type
.
end
()
&&
it
->
second
!=
i
&&
(
py
::
isinstance
<
py
::
int_
>
(
py_args
[
i
])
||
py
::
isinstance
<
py
::
float_
>
(
py_args
[
i
])))
{
auto
tensor_ptr
=
py
::
cast
<
tensor
::
TensorPtr
>
(
py_args
[
it
->
second
]);
if
(
py
::
isinstance
<
py
::
int_
>
(
py_args
[
i
]))
{
py_
input
s
[
i
]
=
std
::
make_shared
<
tensor
::
Tensor
>
(
py
::
cast
<
py
::
int_
>
(
py_args
[
i
]),
tensor_ptr
->
Dtype
());
py_
arg
s
[
i
]
=
std
::
make_shared
<
tensor
::
Tensor
>
(
py
::
cast
<
py
::
int_
>
(
py_args
[
i
]),
tensor_ptr
->
Dtype
());
}
else
{
py_
input
s
[
i
]
=
std
::
make_shared
<
tensor
::
Tensor
>
(
py
::
cast
<
py
::
float_
>
(
py_args
[
i
]),
tensor_ptr
->
Dtype
());
py_
arg
s
[
i
]
=
std
::
make_shared
<
tensor
::
Tensor
>
(
py
::
cast
<
py
::
float_
>
(
py_args
[
i
]),
tensor_ptr
->
Dtype
());
}
continue
;
}
py_inputs
[
i
]
=
py_args
[
i
];
}
return
py_inputs
;
}
void
PynativeInfer
(
const
PrimitivePyPtr
&
prim
,
const
py
::
list
&
py_args
,
OpExecInfo
*
const
op_exec_info
)
{
...
...
@@ -167,12 +177,6 @@ void PynativeInfer(const PrimitivePyPtr &prim, const py::list &py_args, OpExecIn
op_exec_info
->
abstract
=
infer_res
;
}
py
::
object
GetTupleObj
(
const
py
::
object
&
obj
)
{
py
::
module
mod
=
parse
::
python_adapter
::
GetPyModule
(
parse
::
PYTHON_MOD_PARSE_MODULE
);
py
::
object
obj_tuple
=
parse
::
python_adapter
::
CallPyModFn
(
mod
,
parse
::
PYTHON_MOD_GET_DEFAULT_INPUT
,
obj
);
return
obj_tuple
;
}
OpExecInfoPtr
GenerateOpExecInfo
(
const
py
::
args
&
args
)
{
if
(
args
.
size
()
!=
PY_ARGS_NUM
)
{
MS_LOG
(
ERROR
)
<<
"Four args are needed by RunOp"
;
...
...
@@ -186,19 +190,18 @@ OpExecInfoPtr GenerateOpExecInfo(const py::args &args) {
if
(
pyobj
==
nullptr
)
{
MS_LOG
(
EXCEPTION
)
<<
"pyobj is empty"
;
}
py
::
list
py_args
=
ConvertInputs
(
prim
,
args
[
PY_INPUTS
]);
py
::
list
a
=
args
[
PY_INPUTS
];
size_t
input_num
=
a
.
size
();
op_exec_info
->
op_inputs
=
py
::
tuple
(
input_num
);
ConvertInputs
(
prim
,
args
[
PY_INPUTS
],
&
op_exec_info
->
op_inputs
);
// use python infer method
if
(
ignore_infer_prim
.
find
(
op_exec_info
->
op_name
)
==
ignore_infer_prim
.
end
())
{
PynativeInfer
(
prim
,
py_arg
s
,
op_exec_info
.
get
());
PynativeInfer
(
prim
,
op_exec_info
->
op_input
s
,
op_exec_info
.
get
());
}
op_exec_info
->
py_primitive
=
prim
;
op_exec_info
->
op_attrs
=
py
::
getattr
(
args
[
PY_PRIM
],
"attrs"
);
size_t
input_num
=
py_args
.
size
();
op_exec_info
->
op_inputs
=
py
::
tuple
(
input_num
);
for
(
size_t
i
=
0
;
i
<
input_num
;
++
i
)
{
auto
obj
=
py_args
[
i
];
op_exec_info
->
op_inputs
[
i
]
=
GetTupleObj
(
obj
);
}
op_exec_info
->
inputs_mask
=
args
[
PY_INPUT_MASK
];
if
(
op_exec_info
->
op_inputs
.
size
()
!=
op_exec_info
->
inputs_mask
.
size
())
{
MS_LOG
(
ERROR
)
<<
"Op:"
<<
op_exec_info
->
op_name
<<
" inputs size not equal op_mask"
;
...
...
@@ -663,8 +666,25 @@ void PynativeExecutor::EndGraph(const py::object &cell, const py::object &out, c
cell_graph_map_
[
cell_id
]
=
curr_g_
;
auto
out_id
=
GetId
(
out
);
if
(
!
graph_info_map_
[
curr_g_
].
obj_node_map
.
count
(
out_id
))
{
MS_LOG
(
ERROR
)
<<
"graph has no this out: "
<<
out_id
;
return
;
// cell construct return x, y
if
(
py
::
isinstance
<
py
::
tuple
>
(
out
))
{
std
::
vector
<
AnfNodePtr
>
args
;
args
.
push_back
(
NewValueNode
(
prim
::
kPrimMakeTuple
));
auto
tuple
=
out
.
cast
<
py
::
tuple
>
();
MS_LOG
(
DEBUG
)
<<
"End graph start tuple size"
<<
tuple
.
size
();
auto
tuple_size
=
static_cast
<
int
>
(
tuple
.
size
());
auto
cnode
=
curr_g_
->
NewCNode
(
args
);
for
(
int
i
=
0
;
i
<
tuple_size
;
i
++
)
{
args
.
push_back
(
GetInput
(
tuple
[
i
],
py
::
object
()));
set_obj_node_map
(
curr_g_
,
GetId
(
tuple
[
i
]),
cnode
,
i
);
}
cnode
->
set_inputs
(
args
);
set_obj_node_map
(
curr_g_
,
out_id
,
cnode
);
}
else
{
MS_LOG
(
ERROR
)
<<
"Graph has no this out: "
<<
out_id
;
return
;
}
}
auto
output_node
=
GetObjNode
(
out
);
...
...
mindspore/ccsrc/pynative/pynative_execute.h
浏览文件 @
4bede54f
...
...
@@ -44,7 +44,7 @@ py::object RunOpInVM(const OpExecInfoPtr &op_exec_info, PynativeStatusCode *stat
py
::
tuple
RunOp
(
const
py
::
args
&
args
);
py
::
list
ConvertInputs
(
const
PrimitivePyPtr
&
prim
,
const
py
::
list
&
py
_args
);
void
ConvertInputs
(
const
PrimitivePyPtr
&
prim
,
const
py
::
list
&
py_args
,
py
::
tuple
*
out
_args
);
void
ClearPyNativeSession
();
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录