Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
Paddle
提交
eef38db1
P
Paddle
项目概览
PaddlePaddle
/
Paddle
大约 1 年 前同步成功
通知
2298
Star
20931
Fork
5422
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1423
列表
看板
标记
里程碑
合并请求
543
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1,423
Issue
1,423
列表
看板
标记
里程碑
合并请求
543
合并请求
543
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
eef38db1
编写于
6月 29, 2023
作者:
H
hong
提交者:
GitHub
6月 29, 2023
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Refactor build attribute (#54968)
* update * refactor build context * fix bug * polish code * change func name
上级
7353e9e9
变更
8
隐藏空白更改
内联
并排
Showing
8 changed file
with
207 addition
and
326 deletion
+207
-326
paddle/fluid/framework/new_executor/interpreter/interpreter_util.cc
...id/framework/new_executor/interpreter/interpreter_util.cc
+15
-6
paddle/fluid/ir/interface/op_yaml_info_parser.cc
paddle/fluid/ir/interface/op_yaml_info_parser.cc
+29
-28
paddle/fluid/ir/interface/op_yaml_info_parser.h
paddle/fluid/ir/interface/op_yaml_info_parser.h
+10
-12
paddle/fluid/ir/phi_kernel_adaptor/phi_kernel_adaptor.h
paddle/fluid/ir/phi_kernel_adaptor/phi_kernel_adaptor.h
+11
-3
paddle/fluid/ir/phi_kernel_adaptor/phi_kernel_util.cc
paddle/fluid/ir/phi_kernel_adaptor/phi_kernel_util.cc
+0
-257
paddle/fluid/ir/phi_kernel_adaptor/phi_kernel_util.h
paddle/fluid/ir/phi_kernel_adaptor/phi_kernel_util.h
+140
-10
paddle/phi/core/meta_tensor.h
paddle/phi/core/meta_tensor.h
+2
-0
test/cpp/ir/core/op_yaml_info_parser_test.cc
test/cpp/ir/core/op_yaml_info_parser_test.cc
+0
-10
未找到文件。
paddle/fluid/framework/new_executor/interpreter/interpreter_util.cc
浏览文件 @
eef38db1
...
@@ -969,11 +969,16 @@ void BuildOpFuncList(
...
@@ -969,11 +969,16 @@ void BuildOpFuncList(
VLOG
(
6
)
<<
"op name"
<<
op_func_node
.
phi_op_name_
;
VLOG
(
6
)
<<
"op name"
<<
op_func_node
.
phi_op_name_
;
dialect
::
OpYamlInfoParser
op_yaml_info_parser
(
impl
->
get_op_info_
());
dialect
::
OpYamlInfoParser
op_yaml_info_parser
(
impl
->
get_op_info_
());
::
ir
::
BuildInferMetaContext
((
*
it
),
::
ir
::
BuildPhiContext
<
value_2_name_map
,
phi
::
InferMetaContext
,
scope
,
phi
::
MetaTensor
,
op_yaml_info_parser
,
phi
::
MetaTensor
,
&
(
op_func_node
.
infer_meta_context_
));
paddle
::
small_vector
<
phi
::
MetaTensor
,
phi
::
kInputSmallVectorSize
>
,
false
>
((
*
it
),
value_2_name_map
,
scope
,
op_yaml_info_parser
,
&
(
op_func_node
.
infer_meta_context_
));
auto
kernel_name
=
auto
kernel_name
=
attr_map
.
at
(
"kernel_name"
).
dyn_cast
<
ir
::
StrAttribute
>
().
data
();
attr_map
.
at
(
"kernel_name"
).
dyn_cast
<
ir
::
StrAttribute
>
().
data
();
...
@@ -990,7 +995,11 @@ void BuildOpFuncList(
...
@@ -990,7 +995,11 @@ void BuildOpFuncList(
true
,
true
,
"not found kernel for [%s]"
,
"not found kernel for [%s]"
,
kernel_name
);
kernel_name
);
::
ir
::
BuildPhiKernelContext
((
*
it
),
::
ir
::
BuildPhiContext
<
phi
::
KernelContext
,
const
phi
::
TensorBase
*
,
phi
::
TensorBase
*
,
paddle
::
small_vector
<
const
phi
::
TensorBase
*>
,
true
>
((
*
it
),
value_2_name_map
,
value_2_name_map
,
scope
,
scope
,
op_yaml_info_parser
,
op_yaml_info_parser
,
...
...
paddle/fluid/ir/interface/op_yaml_info_parser.cc
浏览文件 @
eef38db1
...
@@ -39,21 +39,21 @@ size_t OpYamlInfoParser::InputTensorNumber() const {
...
@@ -39,21 +39,21 @@ size_t OpYamlInfoParser::InputTensorNumber() const {
const
std
::
string
&
OpYamlInfoParser
::
AttrTypeName
(
const
std
::
string
&
OpYamlInfoParser
::
AttrTypeName
(
const
std
::
string
&
name
)
const
{
const
std
::
string
&
name
)
const
{
auto
it
=
map_
attr_info_
.
find
(
name
);
auto
it
=
attr_info_
.
find
(
name
);
PADDLE_ENFORCE_NE
(
PADDLE_ENFORCE_NE
(
it
,
it
,
map_
attr_info_
.
end
(),
attr_info_
.
end
(),
phi
::
errors
::
NotFound
(
"Not found [%s] in attribute map"
,
name
));
phi
::
errors
::
NotFound
(
"Not found [%s] in attribute map"
,
name
));
return
it
->
second
.
type_name
;
return
it
->
second
.
type_name
;
}
}
const
std
::
string
&
OpYamlInfoParser
::
TensorAttrTypeName
(
const
std
::
string
&
OpYamlInfoParser
::
TensorAttrTypeName
(
const
std
::
string
&
name
)
const
{
const
std
::
string
&
name
)
const
{
auto
it
=
map_
input_info_
.
find
(
name
);
auto
it
=
input_info_
.
find
(
name
);
PADDLE_ENFORCE_NE
(
it
,
PADDLE_ENFORCE_NE
(
it
,
map_
input_info_
.
end
(),
input_info_
.
end
(),
phi
::
errors
::
NotFound
(
"Not found [%s] in input map"
,
name
));
phi
::
errors
::
NotFound
(
"Not found [%s] in input map"
,
name
));
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
...
@@ -63,18 +63,21 @@ const std::string& OpYamlInfoParser::TensorAttrTypeName(
...
@@ -63,18 +63,21 @@ const std::string& OpYamlInfoParser::TensorAttrTypeName(
return
it
->
second
.
type_name
;
return
it
->
second
.
type_name
;
}
}
const
std
::
vector
<
std
::
string
>&
OpYamlInfoParser
::
InferMetaTensorParams
()
const
std
::
vector
<
std
::
string
>&
OpYamlInfoParser
::
TensorParams
(
const
{
bool
is_kernel
)
const
{
return
vec_infer_meta_tensor_params_
;
if
(
is_kernel
)
{
}
return
kernel_fn_tensor_params_
;
const
std
::
vector
<
std
::
string
>&
OpYamlInfoParser
::
InferMetaAttrParams
()
const
{
}
else
{
return
vec_infer_meta_attr_params_
;
return
infer_meta_tensor_params_
;
}
}
const
std
::
vector
<
std
::
string
>&
OpYamlInfoParser
::
KernelFnTensorParams
()
const
{
return
vec_kernel_fn_tensor_params_
;
}
}
const
std
::
vector
<
std
::
string
>&
OpYamlInfoParser
::
KernelFnAttrParams
()
const
{
const
std
::
vector
<
std
::
string
>&
OpYamlInfoParser
::
AttrParams
(
return
vec_kernel_fn_attr_params_
;
bool
is_kernel
)
const
{
if
(
is_kernel
)
{
return
kernel_fn_attr_params_
;
}
else
{
return
infer_meta_attr_params_
;
}
}
}
const
OpRunTimeInfo
&
OpYamlInfoParser
::
OpRuntimeInfo
()
const
{
const
OpRunTimeInfo
&
OpYamlInfoParser
::
OpRuntimeInfo
()
const
{
...
@@ -82,7 +85,7 @@ const OpRunTimeInfo& OpYamlInfoParser::OpRuntimeInfo() const {
...
@@ -82,7 +85,7 @@ const OpRunTimeInfo& OpYamlInfoParser::OpRuntimeInfo() const {
}
}
const
std
::
map
<
std
::
string
,
int
>&
OpYamlInfoParser
::
Name2Id
()
const
{
const
std
::
map
<
std
::
string
,
int
>&
OpYamlInfoParser
::
Name2Id
()
const
{
return
map_
name2id_
;
return
name2id_
;
}
}
void
OpYamlInfoParser
::
parse
()
{
void
OpYamlInfoParser
::
parse
()
{
...
@@ -91,43 +94,41 @@ void OpYamlInfoParser::parse() {
...
@@ -91,43 +94,41 @@ void OpYamlInfoParser::parse() {
int
start_index
=
0
;
int
start_index
=
0
;
for
(
size_t
i
=
0
;
i
<
input_info
.
size
();
++
i
)
{
for
(
size_t
i
=
0
;
i
<
input_info
.
size
();
++
i
)
{
map_
name2id_
[
input_info
[
i
].
name
]
=
start_index
++
;
name2id_
[
input_info
[
i
].
name
]
=
start_index
++
;
if
(
!
input_info
[
i
].
is_mutable_attribute
)
{
if
(
!
input_info
[
i
].
is_mutable_attribute
)
{
input_tensor_number_
++
;
input_tensor_number_
++
;
}
}
map_
input_info_
[
input_info
[
i
].
name
]
=
input_info
[
i
];
input_info_
[
input_info
[
i
].
name
]
=
input_info
[
i
];
}
}
auto
attribute_info
=
std
::
get
<
1
>
(
op_info_tuple_
);
auto
attribute_info
=
std
::
get
<
1
>
(
op_info_tuple_
);
for
(
size_t
i
=
0
;
i
<
attribute_info
.
size
();
++
i
)
{
for
(
size_t
i
=
0
;
i
<
attribute_info
.
size
();
++
i
)
{
map_
attr_info_
[
attribute_info
[
i
].
name
]
=
attribute_info
[
i
];
attr_info_
[
attribute_info
[
i
].
name
]
=
attribute_info
[
i
];
}
}
auto
output_info
=
std
::
get
<
2
>
(
op_info_tuple_
);
auto
output_info
=
std
::
get
<
2
>
(
op_info_tuple_
);
for
(
size_t
i
=
0
;
i
<
output_info
.
size
();
++
i
)
{
for
(
size_t
i
=
0
;
i
<
output_info
.
size
();
++
i
)
{
map_
output_info_
[
output_info
[
i
].
name
]
=
output_info
[
i
];
output_info_
[
output_info
[
i
].
name
]
=
output_info
[
i
];
}
}
auto
runtime_info
=
std
::
get
<
3
>
(
op_info_tuple_
);
auto
runtime_info
=
std
::
get
<
3
>
(
op_info_tuple_
);
for
(
auto
&
name
:
runtime_info
.
infer_meta_param
)
{
for
(
auto
&
name
:
runtime_info
.
infer_meta_param
)
{
if
(
map_name2id_
.
count
(
name
)
&&
if
(
name2id_
.
count
(
name
)
&&
!
input_info_
[
name
].
is_mutable_attribute
)
{
!
map_input_info_
[
name
].
is_mutable_attribute
)
{
infer_meta_tensor_params_
.
push_back
(
name
);
vec_infer_meta_tensor_params_
.
push_back
(
name
);
}
else
{
}
else
{
vec_
infer_meta_attr_params_
.
push_back
(
name
);
infer_meta_attr_params_
.
push_back
(
name
);
}
}
}
}
for
(
auto
&
name
:
runtime_info
.
kernel_param
)
{
for
(
auto
&
name
:
runtime_info
.
kernel_param
)
{
if
(
map_name2id_
.
count
(
name
)
&&
if
(
name2id_
.
count
(
name
)
&&
!
input_info_
[
name
].
is_mutable_attribute
)
{
!
map_input_info_
[
name
].
is_mutable_attribute
)
{
kernel_fn_tensor_params_
.
push_back
(
name
);
vec_kernel_fn_tensor_params_
.
push_back
(
name
);
}
else
{
}
else
{
vec_
kernel_fn_attr_params_
.
push_back
(
name
);
kernel_fn_attr_params_
.
push_back
(
name
);
}
}
}
}
}
}
...
...
paddle/fluid/ir/interface/op_yaml_info_parser.h
浏览文件 @
eef38db1
...
@@ -31,10 +31,8 @@ class OpYamlInfoParser {
...
@@ -31,10 +31,8 @@ class OpYamlInfoParser {
const
std
::
string
&
AttrTypeName
(
const
std
::
string
&
name
)
const
;
const
std
::
string
&
AttrTypeName
(
const
std
::
string
&
name
)
const
;
const
std
::
string
&
TensorAttrTypeName
(
const
std
::
string
&
name
)
const
;
const
std
::
string
&
TensorAttrTypeName
(
const
std
::
string
&
name
)
const
;
const
std
::
vector
<
std
::
string
>&
InferMetaTensorParams
()
const
;
const
std
::
vector
<
std
::
string
>&
TensorParams
(
bool
is_kernel
=
false
)
const
;
const
std
::
vector
<
std
::
string
>&
InferMetaAttrParams
()
const
;
const
std
::
vector
<
std
::
string
>&
AttrParams
(
bool
is_kernel
=
false
)
const
;
const
std
::
vector
<
std
::
string
>&
KernelFnTensorParams
()
const
;
const
std
::
vector
<
std
::
string
>&
KernelFnAttrParams
()
const
;
const
OpRunTimeInfo
&
OpRuntimeInfo
()
const
;
const
OpRunTimeInfo
&
OpRuntimeInfo
()
const
;
const
std
::
map
<
std
::
string
,
int
>&
Name2Id
()
const
;
const
std
::
map
<
std
::
string
,
int
>&
Name2Id
()
const
;
...
@@ -46,16 +44,16 @@ class OpYamlInfoParser {
...
@@ -46,16 +44,16 @@ class OpYamlInfoParser {
OpInfoTuple
op_info_tuple_
;
OpInfoTuple
op_info_tuple_
;
std
::
map
<
std
::
string
,
int
>
map_
name2id_
;
std
::
map
<
std
::
string
,
int
>
name2id_
;
std
::
map
<
std
::
string
,
OpInputInfo
>
map_
input_info_
;
std
::
map
<
std
::
string
,
OpInputInfo
>
input_info_
;
std
::
map
<
std
::
string
,
OpAttributeInfo
>
map_
attr_info_
;
std
::
map
<
std
::
string
,
OpAttributeInfo
>
attr_info_
;
std
::
map
<
std
::
string
,
OpOutputInfo
>
map_
output_info_
;
std
::
map
<
std
::
string
,
OpOutputInfo
>
output_info_
;
std
::
vector
<
std
::
string
>
vec_
infer_meta_tensor_params_
;
std
::
vector
<
std
::
string
>
infer_meta_tensor_params_
;
std
::
vector
<
std
::
string
>
vec_
infer_meta_attr_params_
;
std
::
vector
<
std
::
string
>
infer_meta_attr_params_
;
std
::
vector
<
std
::
string
>
vec_
kernel_fn_tensor_params_
;
std
::
vector
<
std
::
string
>
kernel_fn_tensor_params_
;
std
::
vector
<
std
::
string
>
vec_
kernel_fn_attr_params_
;
std
::
vector
<
std
::
string
>
kernel_fn_attr_params_
;
int
input_tensor_number_
{
0
};
int
input_tensor_number_
{
0
};
};
};
...
...
paddle/fluid/ir/phi_kernel_adaptor/phi_kernel_adaptor.h
浏览文件 @
eef38db1
...
@@ -82,8 +82,12 @@ class PhiKernelAdaptor {
...
@@ -82,8 +82,12 @@ class PhiKernelAdaptor {
phi
::
InferMetaContext
ctx
;
phi
::
InferMetaContext
ctx
;
paddle
::
dialect
::
OpYamlInfoParser
op_yaml_info_parser
(
yaml_info
);
paddle
::
dialect
::
OpYamlInfoParser
op_yaml_info_parser
(
yaml_info
);
ir
::
BuildInferMetaContext
(
ir
::
BuildPhiContext
<
(
*
it
),
name_map
,
scope_
,
op_yaml_info_parser
,
&
ctx
);
phi
::
InferMetaContext
,
phi
::
MetaTensor
,
phi
::
MetaTensor
,
paddle
::
small_vector
<
phi
::
MetaTensor
,
phi
::
kInputSmallVectorSize
>
,
false
>
((
*
it
),
name_map
,
scope_
,
op_yaml_info_parser
,
&
ctx
);
infer_meta_impl
->
infer_meta_
(
&
ctx
);
infer_meta_impl
->
infer_meta_
(
&
ctx
);
...
@@ -98,7 +102,11 @@ class PhiKernelAdaptor {
...
@@ -98,7 +102,11 @@ class PhiKernelAdaptor {
phi
::
KernelContext
kernel_ctx
(
dev_ctx
);
phi
::
KernelContext
kernel_ctx
(
dev_ctx
);
ir
::
BuildPhiKernelContext
(
ir
::
BuildPhiContext
<
phi
::
KernelContext
,
const
phi
::
TensorBase
*
,
phi
::
TensorBase
*
,
paddle
::
small_vector
<
const
phi
::
TensorBase
*>
,
true
>
(
(
*
it
),
name_map
,
scope_
,
op_yaml_info_parser
,
&
kernel_ctx
);
(
*
it
),
name_map
,
scope_
,
op_yaml_info_parser
,
&
kernel_ctx
);
kernel_fn
(
&
kernel_ctx
);
kernel_fn
(
&
kernel_ctx
);
...
...
paddle/fluid/ir/phi_kernel_adaptor/phi_kernel_util.cc
浏览文件 @
eef38db1
...
@@ -175,261 +175,4 @@ void BuildScope(ir::Block* block,
...
@@ -175,261 +175,4 @@ void BuildScope(ir::Block* block,
}
}
}
}
void
BuildInferMetaContext
(
ir
::
Operation
*
op
,
const
std
::
unordered_map
<
ir
::
Value
,
std
::
string
>&
name_map
,
paddle
::
framework
::
Scope
*
scope
,
const
paddle
::
dialect
::
OpYamlInfoParser
&
op_yaml_info
,
phi
::
InferMetaContext
*
ctx
)
{
// inputs include input and mutable attributes
auto
attr_map
=
op
->
attributes
();
auto
&
vec_infer_meta_tensor_params
=
op_yaml_info
.
InferMetaTensorParams
();
auto
&
name2id
=
op_yaml_info
.
Name2Id
();
for
(
auto
&
t
:
vec_infer_meta_tensor_params
)
{
PADDLE_ENFORCE_EQ
(
name2id
.
count
(
t
),
true
,
phi
::
errors
::
NotFound
(
"param [%s] MUST in name2id map"
,
t
));
auto
index
=
op_yaml_info
.
Name2Id
().
at
(
t
);
ir
::
Value
ptr
=
op
->
operand
(
index
);
auto
in_var_name
=
name_map
.
at
(
ptr
);
VLOG
(
6
)
<<
"ctx->EmplaceBackInput: "
<<
t
<<
"
\t
"
<<
in_var_name
;
auto
var
=
scope
->
Var
(
in_var_name
);
if
(
var
->
IsType
<
phi
::
DenseTensor
>
())
{
const
phi
::
TensorBase
*
tensor_in
=
&
(
var
->
Get
<
phi
::
DenseTensor
>
());
ctx
->
EmplaceBackInput
(
const_cast
<
phi
::
TensorBase
*>
(
tensor_in
));
}
else
if
(
var
->
IsType
<
paddle
::
framework
::
TensorRefArray
>
())
{
paddle
::
small_vector
<
phi
::
MetaTensor
,
phi
::
kInputSmallVectorSize
>
inputs
;
auto
&
tensor_array
=
var
->
Get
<
paddle
::
framework
::
TensorRefArray
>
();
for
(
size_t
i
=
0
;
i
<
tensor_array
.
size
();
++
i
)
{
inputs
.
emplace_back
(
std
::
move
(
phi
::
MetaTensor
(
*
tensor_array
[
i
])));
}
ctx
->
EmplaceBackInputs
(
std
::
move
(
inputs
));
}
else
{
PADDLE_THROW
(
phi
::
errors
::
Unimplemented
(
"Not support var type [%d] "
,
var
->
Type
()));
}
}
auto
&
vec_infer_meta_attr_params
=
op_yaml_info
.
InferMetaAttrParams
();
for
(
auto
&
t
:
vec_infer_meta_attr_params
)
{
if
(
name2id
.
count
(
t
))
{
// tensor attribute, get information from input
ir
::
Value
ptr
=
op
->
operand
(
name2id
.
at
(
t
));
auto
in_var_name
=
name_map
.
at
(
ptr
);
auto
&
tensor_attr_type
=
op_yaml_info
.
TensorAttrTypeName
(
t
);
VLOG
(
6
)
<<
"ctx->EmplaceBack mutable attr: "
<<
t
<<
"
\t
"
<<
in_var_name
;
if
(
tensor_attr_type
==
"paddle::dialect::IntArrayAttribute"
)
{
phi
::
Attribute
r1
=
phi
::
TensorRef
(
&
(
scope
->
Var
(
in_var_name
)
->
Get
<
phi
::
DenseTensor
>
()));
ctx
->
EmplaceBackAttr
(
r1
);
}
else
if
(
tensor_attr_type
==
"paddle::dialect::ScalarAttribute"
)
{
phi
::
Attribute
r1
=
phi
::
TensorRef
(
&
(
scope
->
Var
(
in_var_name
)
->
Get
<
phi
::
DenseTensor
>
()));
ctx
->
EmplaceBackAttr
(
r1
);
}
else
{
PADDLE_THROW
(
phi
::
errors
::
Unimplemented
(
"attr type not support [%s] "
,
tensor_attr_type
));
}
continue
;
}
auto
&
attr_type_name
=
op_yaml_info
.
AttrTypeName
(
t
);
if
(
attr_type_name
==
"paddle::dialect::IntArrayAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
paddle
::
dialect
::
IntArrayAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"paddle::dialect::DataTypeAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
paddle
::
dialect
::
DataTypeAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"ir::Int32Attribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
ir
::
Int32Attribute
>
().
data
());
}
else
if
(
attr_type_name
==
"ir::FloatAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
ir
::
FloatAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"ir::BoolAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
ir
::
BoolAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"paddle::dialect::PlaceAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
paddle
::
dialect
::
PlaceAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"paddle::dialect::ScalarAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
paddle
::
dialect
::
ScalarAttribute
>
().
data
());
}
else
{
PADDLE_THROW
(
phi
::
errors
::
Unimplemented
(
"attr type not support [%s] "
,
attr_type_name
));
}
VLOG
(
6
)
<<
"ctx->EmplaceBackAttr: "
<<
t
;
}
// TODO(phlrain): use var type instead of op name
if
(
op
->
attributes
().
count
(
"op_name"
)
&&
(
op
->
attributes
().
at
(
"op_name"
).
dyn_cast
<
ir
::
StrAttribute
>
().
data
()
==
"pd.fetch"
))
{
// process fetch op
auto
fetch_var
=
scope
->
Var
(
"fetch"
);
auto
*
fetch_list
=
fetch_var
->
GetMutable
<
paddle
::
framework
::
FetchList
>
();
auto
*
out_tensor
=
&
(
PADDLE_GET
(
phi
::
DenseTensor
,
fetch_list
->
at
(
0
)));
ctx
->
EmplaceBackOutput
(
out_tensor
);
}
else
{
for
(
size_t
i
=
0
;
i
<
op
->
num_results
();
++
i
)
{
ir
::
Value
out_ptr
=
op
->
result
(
i
);
auto
name
=
name_map
.
at
(
out_ptr
);
ctx
->
EmplaceBackOutput
(
scope
->
Var
(
name
)
->
Get
<
phi
::
DenseTensor
>
());
}
}
}
void
BuildPhiKernelContext
(
ir
::
Operation
*
op
,
const
std
::
unordered_map
<
ir
::
Value
,
std
::
string
>&
name_map
,
paddle
::
framework
::
Scope
*
scope
,
const
paddle
::
dialect
::
OpYamlInfoParser
&
op_yaml_info
,
phi
::
KernelContext
*
ctx
,
std
::
map
<
std
::
string
,
std
::
vector
<
int
>>*
input_map
,
std
::
map
<
std
::
string
,
std
::
vector
<
int
>>*
output_map
)
{
// inputs include input and mutable attributes
auto
attr_map
=
op
->
attributes
();
auto
&
vec_kernel_fn_tensor_params
=
op_yaml_info
.
KernelFnTensorParams
();
auto
&
name2id
=
op_yaml_info
.
Name2Id
();
for
(
auto
&
t
:
vec_kernel_fn_tensor_params
)
{
PADDLE_ENFORCE_EQ
(
name2id
.
count
(
t
),
true
,
phi
::
errors
::
NotFound
(
"param [%s] MUST in name2id map"
,
t
));
auto
index
=
op_yaml_info
.
Name2Id
().
at
(
t
);
ir
::
Value
ptr
=
op
->
operand
(
index
);
auto
in_var_name
=
name_map
.
at
(
ptr
);
VLOG
(
6
)
<<
"ctx->EmplaceBackInput: "
<<
t
<<
"
\t
"
<<
in_var_name
;
PADDLE_ENFORCE_NOT_NULL
(
scope
->
FindLocalVar
(
in_var_name
),
phi
::
errors
::
PreconditionNotMet
(
"can not find var[%s] in scope"
,
in_var_name
));
auto
var
=
scope
->
Var
(
in_var_name
);
if
(
var
->
IsType
<
phi
::
DenseTensor
>
())
{
const
phi
::
TensorBase
*
tensor_in
=
&
(
var
->
Get
<
phi
::
DenseTensor
>
());
ctx
->
EmplaceBackInput
(
tensor_in
);
}
else
if
(
var
->
IsType
<
paddle
::
framework
::
TensorRefArray
>
())
{
paddle
::
small_vector
<
const
phi
::
TensorBase
*>
inputs
;
auto
&
tensor_array
=
var
->
Get
<
paddle
::
framework
::
TensorRefArray
>
();
for
(
size_t
i
=
0
;
i
<
tensor_array
.
size
();
++
i
)
{
inputs
.
emplace_back
(
tensor_array
[
i
]);
}
std
::
cerr
<<
"is tensor ref "
<<
std
::
endl
;
ctx
->
EmplaceBackInputs
(
std
::
move
(
inputs
));
}
else
if
(
var
->
IsType
<
paddle
::
framework
::
FeedList
>
())
{
auto
feed_list
=
var
->
Get
<
paddle
::
framework
::
FeedList
>
();
auto
*
in_tensor
=
&
(
PADDLE_GET
(
phi
::
DenseTensor
,
feed_list
.
at
(
0
)));
ctx
->
EmplaceBackOutput
(
in_tensor
);
}
else
{
PADDLE_THROW
(
phi
::
errors
::
Unimplemented
(
"Not support var type [%d] "
,
var
->
Type
()));
}
}
auto
&
vec_kernel_fn_attr_params
=
op_yaml_info
.
KernelFnAttrParams
();
for
(
auto
&
t
:
vec_kernel_fn_attr_params
)
{
if
(
name2id
.
count
(
t
))
{
// tensor attribute, get information from input
ir
::
Value
ptr
=
op
->
operand
(
name2id
.
at
(
t
));
auto
in_var_name
=
name_map
.
at
(
ptr
);
if
(
input_map
!=
nullptr
)
{
// only deal with single input for now, [todo] need support multi input
// like concat
// TODO(phlrain): OpFuncNode need input_index and output_index,
// construct input_index and output_here, should remove input_index and
// output_index from OpFuncNode Each in_var_name named "inner_var_" +
// index, len("inner_var_") = 10
size_t
tmp_id
=
std
::
atol
(
in_var_name
.
substr
(
4
,
100
).
c_str
());
(
*
input_map
)[
std
::
to_string
(
name2id
.
at
(
t
))].
push_back
(
tmp_id
);
}
auto
&
tensor_attr_type
=
op_yaml_info
.
TensorAttrTypeName
(
t
);
VLOG
(
6
)
<<
"ctx->EmplaceBack mutable attr: "
<<
t
<<
"
\t
"
<<
in_var_name
;
if
(
tensor_attr_type
==
"paddle::dialect::IntArrayAttribute"
)
{
phi
::
Attribute
r1
=
phi
::
TensorRef
(
&
(
scope
->
Var
(
in_var_name
)
->
Get
<
phi
::
DenseTensor
>
()));
ctx
->
EmplaceBackAttr
(
r1
);
}
else
if
(
tensor_attr_type
==
"paddle::dialect::ScalarAttribute"
)
{
phi
::
Attribute
r1
=
phi
::
TensorRef
(
&
(
scope
->
Var
(
in_var_name
)
->
Get
<
phi
::
DenseTensor
>
()));
ctx
->
EmplaceBackAttr
(
r1
);
}
else
{
PADDLE_THROW
(
phi
::
errors
::
Unimplemented
(
"attr type not support [%s] "
,
tensor_attr_type
));
}
continue
;
}
auto
&
attr_type_name
=
op_yaml_info
.
AttrTypeName
(
t
);
if
(
attr_type_name
==
"paddle::dialect::IntArrayAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
paddle
::
dialect
::
IntArrayAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"paddle::dialect::DataTypeAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
paddle
::
dialect
::
DataTypeAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"ir::Int32Attribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
ir
::
Int32Attribute
>
().
data
());
}
else
if
(
attr_type_name
==
"ir::FloatAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
ir
::
FloatAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"ir::BoolAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
ir
::
BoolAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"paddle::dialect::PlaceAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
paddle
::
dialect
::
PlaceAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"paddle::dialect::ScalarAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
paddle
::
dialect
::
ScalarAttribute
>
().
data
());
}
else
{
PADDLE_THROW
(
phi
::
errors
::
Unimplemented
(
"attr type not support [%s] "
,
attr_type_name
));
}
VLOG
(
6
)
<<
"ctx->EmplaceBackAttr: "
<<
t
;
}
// TODO(phlrain): use var type instead of op name
if
(
op
->
attributes
().
count
(
"op_name"
)
&&
(
op
->
attributes
().
at
(
"op_name"
).
dyn_cast
<
ir
::
StrAttribute
>
().
data
()
==
"pd.fetch"
))
{
// process fetch op
auto
fetch_var
=
scope
->
Var
(
"fetch"
);
auto
*
fetch_list
=
fetch_var
->
GetMutable
<
paddle
::
framework
::
FetchList
>
();
auto
*
out_tensor
=
&
(
PADDLE_GET
(
phi
::
DenseTensor
,
fetch_list
->
at
(
0
)));
ctx
->
EmplaceBackOutput
(
out_tensor
);
}
else
{
for
(
size_t
i
=
0
;
i
<
op
->
num_results
();
++
i
)
{
ir
::
Value
out_ptr
=
op
->
result
(
i
);
auto
name
=
name_map
.
at
(
out_ptr
);
ctx
->
EmplaceBackOutput
(
const_cast
<
phi
::
DenseTensor
*>
(
&
(
scope
->
Var
(
name
)
->
Get
<
phi
::
DenseTensor
>
())));
if
(
output_map
!=
nullptr
)
{
// only deal with single input for now, [todo] need support multi input
// like concat
// TODO(phlrain): OpFuncNode need input_index and output_index,
// construct input_index and output_here, should remove input_index and
// output_index from OpFuncNode Each in_var_name named "inner_var_" +
// index, len("inner_var_") = 10
size_t
tmp_id
=
std
::
atol
(
name
.
substr
(
4
,
100
).
c_str
());
(
*
output_map
)[
"out"
].
push_back
(
tmp_id
);
}
}
}
}
}
// namespace ir
}
// namespace ir
paddle/fluid/ir/phi_kernel_adaptor/phi_kernel_util.h
浏览文件 @
eef38db1
...
@@ -44,20 +44,150 @@ void BuildScope(ir::Block* block,
...
@@ -44,20 +44,150 @@ void BuildScope(ir::Block* block,
paddle
::
framework
::
Scope
*
scope
,
paddle
::
framework
::
Scope
*
scope
,
std
::
unordered_map
<
ir
::
Value
,
std
::
string
>*
name_map
);
std
::
unordered_map
<
ir
::
Value
,
std
::
string
>*
name_map
);
void
BuildInferMetaContext
(
template
<
typename
Context
,
typename
InType
,
typename
OutType
,
typename
ListType
,
bool
is_kernel
>
void
BuildPhiContext
(
ir
::
Operation
*
op
,
ir
::
Operation
*
op
,
const
std
::
unordered_map
<
ir
::
Value
,
std
::
string
>&
name_map
,
const
std
::
unordered_map
<
ir
::
Value
,
std
::
string
>&
name_map
,
paddle
::
framework
::
Scope
*
scope
,
paddle
::
framework
::
Scope
*
scope
,
const
paddle
::
dialect
::
OpYamlInfoParser
&
op_yaml_info
,
const
paddle
::
dialect
::
OpYamlInfoParser
&
op_yaml_info
,
phi
::
InferMetaContext
*
ctx
);
Context
*
ctx
,
void
BuildPhiKernelContext
(
ir
::
Operation
*
op
,
const
std
::
unordered_map
<
ir
::
Value
,
std
::
string
>&
name_map
,
paddle
::
framework
::
Scope
*
scope
,
const
paddle
::
dialect
::
OpYamlInfoParser
&
op_yaml_info
,
phi
::
KernelContext
*
ctx
,
std
::
map
<
std
::
string
,
std
::
vector
<
int
>>*
input_map
=
nullptr
,
std
::
map
<
std
::
string
,
std
::
vector
<
int
>>*
input_map
=
nullptr
,
std
::
map
<
std
::
string
,
std
::
vector
<
int
>>*
output_map
=
nullptr
);
std
::
map
<
std
::
string
,
std
::
vector
<
int
>>*
output_map
=
nullptr
)
{
// inputs include input and mutable attributes
auto
attr_map
=
op
->
attributes
();
auto
&
vec_kernel_fn_tensor_params
=
op_yaml_info
.
TensorParams
(
is_kernel
);
auto
&
name2id
=
op_yaml_info
.
Name2Id
();
for
(
auto
&
t
:
vec_kernel_fn_tensor_params
)
{
PADDLE_ENFORCE_EQ
(
name2id
.
count
(
t
),
true
,
phi
::
errors
::
NotFound
(
"param [%s] MUST in name2id map"
,
t
));
auto
index
=
op_yaml_info
.
Name2Id
().
at
(
t
);
ir
::
Value
ptr
=
op
->
operand
(
index
);
auto
in_var_name
=
name_map
.
at
(
ptr
);
VLOG
(
6
)
<<
"ctx->EmplaceBackInput: "
<<
t
<<
"
\t
"
<<
in_var_name
;
PADDLE_ENFORCE_NOT_NULL
(
scope
->
FindLocalVar
(
in_var_name
),
phi
::
errors
::
PreconditionNotMet
(
"can not find var[%s] in scope"
,
in_var_name
));
auto
var
=
scope
->
Var
(
in_var_name
);
if
(
var
->
IsType
<
phi
::
DenseTensor
>
())
{
const
phi
::
TensorBase
*
tensor_in
=
&
(
var
->
Get
<
phi
::
DenseTensor
>
());
ctx
->
EmplaceBackInput
(
InType
(
tensor_in
));
}
else
if
(
var
->
IsType
<
paddle
::
framework
::
TensorRefArray
>
())
{
ListType
inputs
;
auto
&
tensor_array
=
var
->
Get
<
paddle
::
framework
::
TensorRefArray
>
();
for
(
size_t
i
=
0
;
i
<
tensor_array
.
size
();
++
i
)
{
inputs
.
emplace_back
(
InType
(
tensor_array
[
i
]));
}
ctx
->
EmplaceBackInputs
(
inputs
);
}
else
{
PADDLE_THROW
(
phi
::
errors
::
Unimplemented
(
"Not support var type [%d] "
,
var
->
Type
()));
}
}
auto
&
vec_kernel_fn_attr_params
=
op_yaml_info
.
AttrParams
(
is_kernel
);
for
(
auto
&
t
:
vec_kernel_fn_attr_params
)
{
if
(
name2id
.
count
(
t
))
{
// tensor attribute, get information from input
ir
::
Value
ptr
=
op
->
operand
(
name2id
.
at
(
t
));
auto
in_var_name
=
name_map
.
at
(
ptr
);
if
(
input_map
!=
nullptr
)
{
// only deal with single input for now, [todo] need support multi input
// like concat
// TODO(phlrain): OpFuncNode need input_index and output_index,
// construct input_index and output_here, should remove input_index and
// output_index from OpFuncNode Each in_var_name named "inner_var_" +
// index, len("inner_var_") = 10
size_t
tmp_id
=
std
::
atol
(
in_var_name
.
substr
(
4
,
100
).
c_str
());
(
*
input_map
)[
std
::
to_string
(
name2id
.
at
(
t
))].
push_back
(
tmp_id
);
}
auto
&
tensor_attr_type
=
op_yaml_info
.
TensorAttrTypeName
(
t
);
VLOG
(
6
)
<<
"ctx->EmplaceBack mutable attr: "
<<
t
<<
"
\t
"
<<
in_var_name
;
if
(
tensor_attr_type
==
"paddle::dialect::IntArrayAttribute"
)
{
phi
::
Attribute
r1
=
phi
::
TensorRef
(
&
(
scope
->
Var
(
in_var_name
)
->
Get
<
phi
::
DenseTensor
>
()));
ctx
->
EmplaceBackAttr
(
r1
);
}
else
if
(
tensor_attr_type
==
"paddle::dialect::ScalarAttribute"
)
{
phi
::
Attribute
r1
=
phi
::
TensorRef
(
&
(
scope
->
Var
(
in_var_name
)
->
Get
<
phi
::
DenseTensor
>
()));
ctx
->
EmplaceBackAttr
(
r1
);
}
else
{
PADDLE_THROW
(
phi
::
errors
::
Unimplemented
(
"attr type not support [%s] "
,
tensor_attr_type
));
}
continue
;
}
auto
&
attr_type_name
=
op_yaml_info
.
AttrTypeName
(
t
);
if
(
attr_type_name
==
"paddle::dialect::IntArrayAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
paddle
::
dialect
::
IntArrayAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"paddle::dialect::DataTypeAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
paddle
::
dialect
::
DataTypeAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"ir::Int32Attribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
ir
::
Int32Attribute
>
().
data
());
}
else
if
(
attr_type_name
==
"ir::FloatAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
ir
::
FloatAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"ir::BoolAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
ir
::
BoolAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"paddle::dialect::PlaceAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
paddle
::
dialect
::
PlaceAttribute
>
().
data
());
}
else
if
(
attr_type_name
==
"paddle::dialect::ScalarAttribute"
)
{
ctx
->
EmplaceBackAttr
(
attr_map
[
t
].
dyn_cast
<
paddle
::
dialect
::
ScalarAttribute
>
().
data
());
}
else
{
PADDLE_THROW
(
phi
::
errors
::
Unimplemented
(
"attr type not support [%s] "
,
attr_type_name
));
}
VLOG
(
6
)
<<
"ctx->EmplaceBackAttr: "
<<
t
;
}
// TODO(phlrain): use var type instead of op name
if
(
op
->
attributes
().
count
(
"op_name"
)
&&
(
op
->
attributes
().
at
(
"op_name"
).
dyn_cast
<
ir
::
StrAttribute
>
().
data
()
==
"pd.fetch"
))
{
// process fetch op
auto
fetch_var
=
scope
->
Var
(
"fetch"
);
auto
*
fetch_list
=
fetch_var
->
GetMutable
<
paddle
::
framework
::
FetchList
>
();
auto
*
out_tensor
=
&
(
PADDLE_GET
(
phi
::
DenseTensor
,
fetch_list
->
at
(
0
)));
ctx
->
EmplaceBackOutput
(
out_tensor
);
}
else
{
for
(
size_t
i
=
0
;
i
<
op
->
num_results
();
++
i
)
{
ir
::
Value
out_ptr
=
op
->
result
(
i
);
auto
name
=
name_map
.
at
(
out_ptr
);
ctx
->
EmplaceBackOutput
(
OutType
(
const_cast
<
phi
::
DenseTensor
*>
(
&
(
scope
->
Var
(
name
)
->
Get
<
phi
::
DenseTensor
>
()))));
if
(
output_map
!=
nullptr
)
{
// only deal with single input for now, [todo] need support multi input
// like concat
// TODO(phlrain): OpFuncNode need input_index and output_index,
// construct input_index and output_here, should remove input_index and
// output_index from OpFuncNode Each in_var_name named "inner_var_" +
// index, len("inner_var_") = 10
size_t
tmp_id
=
std
::
atol
(
name
.
substr
(
4
,
100
).
c_str
());
(
*
output_map
)[
"out"
].
push_back
(
tmp_id
);
}
}
}
}
}
// namespace ir
}
// namespace ir
paddle/phi/core/meta_tensor.h
浏览文件 @
eef38db1
...
@@ -45,6 +45,8 @@ class MetaTensor {
...
@@ -45,6 +45,8 @@ class MetaTensor {
MetaTensor
(
TensorBase
*
tensor
)
:
tensor_
(
tensor
)
{}
// NOLINT
MetaTensor
(
TensorBase
*
tensor
)
:
tensor_
(
tensor
)
{}
// NOLINT
MetaTensor
(
const
TensorBase
&
tensor
)
// NOLINT
MetaTensor
(
const
TensorBase
&
tensor
)
// NOLINT
:
tensor_
(
const_cast
<
TensorBase
*>
(
&
tensor
))
{}
:
tensor_
(
const_cast
<
TensorBase
*>
(
&
tensor
))
{}
MetaTensor
(
const
TensorBase
*
tensor
)
// NOLINT
:
tensor_
(
const_cast
<
TensorBase
*>
(
tensor
))
{}
MetaTensor
(
TensorBase
&
tensor
)
:
tensor_
(
&
tensor
)
{}
// NOLINT
MetaTensor
(
TensorBase
&
tensor
)
:
tensor_
(
&
tensor
)
{}
// NOLINT
MetaTensor
(
MetaTensor
&&
)
=
default
;
MetaTensor
(
MetaTensor
&&
)
=
default
;
...
...
test/cpp/ir/core/op_yaml_info_parser_test.cc
浏览文件 @
eef38db1
...
@@ -55,16 +55,6 @@ TEST(ir_op_info_test, op_op_info_test) {
...
@@ -55,16 +55,6 @@ TEST(ir_op_info_test, op_op_info_test) {
paddle
::
dialect
::
OpYamlInfoParser
op_yaml_info_parser
(
op_info_res
);
paddle
::
dialect
::
OpYamlInfoParser
op_yaml_info_parser
(
op_info_res
);
auto
infer_meta_tensor_param
=
op_yaml_info_parser
.
InferMetaTensorParams
();
auto
infer_meta_attr_param
=
op_yaml_info_parser
.
InferMetaAttrParams
();
auto
kernel_fn_tensor_param
=
op_yaml_info_parser
.
KernelFnTensorParams
();
auto
kernel_fn_attr_param
=
op_yaml_info_parser
.
KernelFnAttrParams
();
EXPECT_EQ
(
infer_meta_tensor_param
.
size
(),
0u
);
EXPECT_EQ
(
infer_meta_attr_param
.
size
(),
2u
);
EXPECT_EQ
(
kernel_fn_tensor_param
.
size
(),
0u
);
EXPECT_EQ
(
kernel_fn_attr_param
.
size
(),
5u
);
EXPECT_EQ
((
op_yaml_info_parser
.
AttrTypeName
(
"seed"
)
==
"ir::Int32Attribute"
),
EXPECT_EQ
((
op_yaml_info_parser
.
AttrTypeName
(
"seed"
)
==
"ir::Int32Attribute"
),
true
);
true
);
EXPECT_EQ
(
op_yaml_info_parser
.
IsTensorAttribute
(
0
),
true
);
EXPECT_EQ
(
op_yaml_info_parser
.
IsTensorAttribute
(
0
),
true
);
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录