Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
Paddle
提交
9bb3744f
P
Paddle
项目概览
PaddlePaddle
/
Paddle
大约 1 年 前同步成功
通知
2298
Star
20931
Fork
5422
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1423
列表
看板
标记
里程碑
合并请求
543
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1,423
Issue
1,423
列表
看板
标记
里程碑
合并请求
543
合并请求
543
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
9bb3744f
编写于
3月 29, 2022
作者:
王
王明冬
提交者:
GitHub
3月 29, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
[Infrt] delete custom_pdop.td and move op to infrt dialect. (#41021)
上级
aeade538
变更
15
隐藏空白更改
内联
并排
Showing
15 changed file
with
74 addition
and
147 deletion
+74
-147
paddle/infrt/dialect/infrt/ir/infrt_base.td
paddle/infrt/dialect/infrt/ir/infrt_base.td
+1
-0
paddle/infrt/dialect/infrt/ir/infrt_dialect.cc
paddle/infrt/dialect/infrt/ir/infrt_dialect.cc
+36
-11
paddle/infrt/dialect/infrt/ir/infrt_dialect.h
paddle/infrt/dialect/infrt/ir/infrt_dialect.h
+1
-0
paddle/infrt/dialect/infrt/ir/infrt_ops.td
paddle/infrt/dialect/infrt/ir/infrt_ops.td
+15
-1
paddle/infrt/dialect/infrt/pass/infrt_op_fuse.td
paddle/infrt/dialect/infrt/pass/infrt_op_fuse.td
+0
-4
paddle/infrt/dialect/infrt/pass/infrt_op_fuse_pass.cc
paddle/infrt/dialect/infrt/pass/infrt_op_fuse_pass.cc
+1
-1
paddle/infrt/dialect/pd/ir/pd_op_base.td
paddle/infrt/dialect/pd/ir/pd_op_base.td
+0
-1
paddle/infrt/dialect/pd/ir/pd_ops.cc
paddle/infrt/dialect/pd/ir/pd_ops.cc
+0
-37
paddle/infrt/dialect/tensorrt/trt_op_teller_pass.cc
paddle/infrt/dialect/tensorrt/trt_op_teller_pass.cc
+0
-7
paddle/infrt/tests/dialect/disabled_rewrite_conv_bn.mlir
paddle/infrt/tests/dialect/disabled_rewrite_conv_bn.mlir
+3
-4
paddle/infrt/tests/dialect/paddle_ops.mlir
paddle/infrt/tests/dialect/paddle_ops.mlir
+0
-9
paddle/infrt/tests/dialect/pd/rewrite.mlir
paddle/infrt/tests/dialect/pd/rewrite.mlir
+11
-19
paddle/infrt/tests/dialect/phi/phi_pass.mlir
paddle/infrt/tests/dialect/phi/phi_pass.mlir
+3
-6
tools/infrt/custom_pdop.td
tools/infrt/custom_pdop.td
+0
-37
tools/infrt/generate_pd_op_dialect_from_paddle_op_maker.py
tools/infrt/generate_pd_op_dialect_from_paddle_op_maker.py
+3
-10
未找到文件。
paddle/infrt/dialect/infrt/ir/infrt_base.td
浏览文件 @
9bb3744f
...
...
@@ -10,6 +10,7 @@ def Infrt_Dialect : Dialect {
let name = "infrt";
let cppNamespace = "::infrt";
let hasConstantMaterializer = 1;
let useDefaultAttributePrinterParser = 1;
}
...
...
paddle/infrt/dialect/infrt/ir/infrt_dialect.cc
浏览文件 @
9bb3744f
...
...
@@ -183,16 +183,41 @@ void InfrtDialect::printType(::mlir::Type type,
llvm_unreachable
(
"unknown infrt type."
);
}
// /// Parse an attribute registered to this dialect.
// ::mlir::Attribute InfrtDialect::parseAttribute(::mlir::DialectAsmParser
// &parser,
// ::mlir::Type type) const {
// return mlir::Attribute();
// }
// /// Print an attribute registered to this dialect.
// void InfrtDialect::printAttribute(::mlir::Attribute attr,
// ::mlir::DialectAsmPrinter &os) const {
// }
mlir
::
Operation
*
InfrtDialect
::
materializeConstant
(
mlir
::
OpBuilder
&
builder
,
mlir
::
Attribute
value
,
mlir
::
Type
type
,
mlir
::
Location
loc
)
{
return
builder
.
create
<
ConstantOp
>
(
loc
,
value
);
}
void
ConstantOp
::
build
(
mlir
::
OpBuilder
&
builder
,
mlir
::
OperationState
&
state
,
mlir
::
Attribute
value
)
{
if
(
auto
elem_attr
=
value
.
dyn_cast
<
mlir
::
ElementsAttr
>
())
{
return
ConstantOp
::
build
(
builder
,
state
,
elem_attr
);
}
else
if
(
value
.
isa
<
mlir
::
BoolAttr
,
mlir
::
FloatAttr
,
mlir
::
IntegerAttr
>
())
{
mlir
::
ShapedType
type
=
mlir
::
RankedTensorType
::
get
(
/*shape=*/
{},
value
.
getType
());
state
.
addAttribute
(
"value"
,
mlir
::
DenseElementsAttr
::
get
(
type
,
value
));
state
.
addTypes
(
type
);
return
;
}
llvm_unreachable
(
"unsupported attribute type for building pd.constant"
);
}
mlir
::
LogicalResult
ConstantOp
::
inferReturnTypes
(
mlir
::
MLIRContext
*
context
,
mlir
::
Optional
<
mlir
::
Location
>
location
,
mlir
::
ValueRange
operands
,
mlir
::
DictionaryAttr
attributes
,
mlir
::
RegionRange
regions
,
llvm
::
SmallVectorImpl
<
mlir
::
Type
>
&
inferredReturnTypes
)
{
inferredReturnTypes
.
push_back
(
attributes
.
get
(
"value"
).
getType
());
return
mlir
::
success
();
}
mlir
::
OpFoldResult
ConstantOp
::
fold
(
::
llvm
::
ArrayRef
<
mlir
::
Attribute
>
operands
)
{
return
value
();
}
}
// namespace infrt
paddle/infrt/dialect/infrt/ir/infrt_dialect.h
浏览文件 @
9bb3744f
...
...
@@ -21,6 +21,7 @@
#include <mlir/IR/BuiltinTypes.h>
#include <mlir/IR/Dialect.h>
#include <mlir/IR/OpDefinition.h>
#include <mlir/Interfaces/InferTypeOpInterface.h>
#include <mlir/Interfaces/SideEffectInterfaces.h>
#include "paddle/infrt/dialect/infrt/common/types.h"
...
...
paddle/infrt/dialect/infrt/ir/infrt_ops.td
浏览文件 @
9bb3744f
include "mlir/Interfaces/InferTypeOpInterface.td"
include "paddle/infrt/dialect/infrt/ir/infrt_base.td"
// Op definition
...
...
@@ -9,7 +10,7 @@ class Infrt_Op<string mnemonic, list<OpTrait> traits = []> : Op<Infrt_Dialect, m
// let parser = [{ return infrt::parse$cppClass(parser, result); }];
}
def
PD
_GraphOp : Infrt_Op<"graph", [SingleBlockImplicitTerminator<"::infrt::ReturnOp">]> {
def
Infrt
_GraphOp : Infrt_Op<"graph", [SingleBlockImplicitTerminator<"::infrt::ReturnOp">]> {
let summary = "paddle graph Op";
let description = [{
Describe a paddle graph or subgraph.
...
...
@@ -69,3 +70,16 @@ def Infrt_TensorCastOp : Infrt_Op<"tensor_cast", [NoSideEffect]> {
let arguments = (ins AnyType:$input);
let results = (outs AnyType:$output);
}
def Infrt_ConstantOp : Infrt_Op<"constant", [NoSideEffect, ConstantLike, DeclareOpInterfaceMethods<InferTypeOpInterface>, AllTypesMatch<["value", "output"]>]> {
let summary = "constant Op";
let description = [{}];
let arguments = (ins ElementsAttr:$value);
let results = (outs AnyType:$output);
let hasFolder = 1;
let builders = [
OpBuilder<(ins "mlir::Attribute":$value)>,
];
}
paddle/infrt/dialect/infrt/pass/infrt_op_fuse.td
浏览文件 @
9bb3744f
...
...
@@ -9,10 +9,6 @@ def FuseTensorCastPattern : Pat<
(Infrt_TensorCastOp (Infrt_TensorCastOp $arg)),
(Infrt_TensorCastOp $arg)>;
def FuseFeedTensorCastPattern : Pat<
(Infrt_TensorCastOp (PD_FeedOp $name)),
(PD_FeedOp $name)>;
def TypesAreIdentical : Constraint<CPred<"$0.getType() == $1.getType()">>;
def RedundantTensorCastOptPattern : Pat<
(Infrt_TensorCastOp:$res $arg), (replaceWithValue $arg),
...
...
paddle/infrt/dialect/infrt/pass/infrt_op_fuse_pass.cc
浏览文件 @
9bb3744f
...
...
@@ -38,7 +38,7 @@ void InfrtOpFusePass::runOnFunction() {
::
mlir
::
RewritePatternSet
patterns
(
&
getContext
());
populateWithGenerated
(
patterns
);
(
void
)
applyPatternsAndFoldGreedily
(
getOperation
(),
std
::
move
(
patterns
));
// Fuse
pd
.return Operation
// Fuse
infrt
.return Operation
auto
terminator_op
=
getFunction
().
front
().
getTerminator
();
if
(
nullptr
==
terminator_op
)
return
;
for
(
auto
operand
:
terminator_op
->
getOperands
())
{
...
...
paddle/infrt/dialect/pd/ir/pd_op_base.td
浏览文件 @
9bb3744f
...
...
@@ -16,7 +16,6 @@ def Paddle_Dialect : Dialect {
This dialect contains the PaddlePaddle operators.
}];
let hasConstantMaterializer = 1;
let cppNamespace = "infrt::pd";
}
...
...
paddle/infrt/dialect/pd/ir/pd_ops.cc
浏览文件 @
9bb3744f
...
...
@@ -35,42 +35,5 @@ void PaddleDialect::initialize() {
#include "paddle/infrt/dialect/pd/ir/pd_extra_ops.cpp.inc" // NOLINT
>
();
}
mlir
::
Operation
*
PaddleDialect
::
materializeConstant
(
mlir
::
OpBuilder
&
builder
,
mlir
::
Attribute
value
,
mlir
::
Type
type
,
mlir
::
Location
loc
)
{
return
builder
.
create
<
ConstantOp
>
(
loc
,
value
);
}
void
ConstantOp
::
build
(
mlir
::
OpBuilder
&
builder
,
mlir
::
OperationState
&
state
,
mlir
::
Attribute
value
)
{
if
(
auto
elem_attr
=
value
.
dyn_cast
<
mlir
::
ElementsAttr
>
())
{
return
ConstantOp
::
build
(
builder
,
state
,
elem_attr
);
}
else
if
(
value
.
isa
<
mlir
::
BoolAttr
,
mlir
::
FloatAttr
,
mlir
::
IntegerAttr
>
())
{
mlir
::
ShapedType
type
=
mlir
::
RankedTensorType
::
get
(
/*shape=*/
{},
value
.
getType
());
state
.
addAttribute
(
"value"
,
mlir
::
DenseElementsAttr
::
get
(
type
,
value
));
state
.
addTypes
(
type
);
return
;
}
llvm_unreachable
(
"unsupported attribute type for building pd.constant"
);
}
mlir
::
LogicalResult
ConstantOp
::
inferReturnTypes
(
mlir
::
MLIRContext
*
context
,
mlir
::
Optional
<
mlir
::
Location
>
location
,
mlir
::
ValueRange
operands
,
mlir
::
DictionaryAttr
attributes
,
mlir
::
RegionRange
regions
,
llvm
::
SmallVectorImpl
<
mlir
::
Type
>
&
inferredReturnTypes
)
{
inferredReturnTypes
.
push_back
(
attributes
.
get
(
"value"
).
getType
());
return
mlir
::
success
();
}
mlir
::
OpFoldResult
ConstantOp
::
fold
(
::
llvm
::
ArrayRef
<
mlir
::
Attribute
>
operands
)
{
return
value
();
}
}
// namespace pd
}
// namespace infrt
paddle/infrt/dialect/tensorrt/trt_op_teller_pass.cc
浏览文件 @
9bb3744f
...
...
@@ -39,13 +39,6 @@ void TRTOpTellerPass::runOnFunction() {
worklist
.
pop_back
();
if
(
op
==
nullptr
)
continue
;
if
(
op
->
getName
().
getStringRef
().
substr
(
0
,
3
)
!=
"pd."
)
continue
;
if
(
::
llvm
::
dyn_cast_or_null
<
infrt
::
pd
::
FeedOp
>
(
op
))
continue
;
if
(
::
llvm
::
dyn_cast_or_null
<
infrt
::
pd
::
FetchOp
>
(
op
))
continue
;
if
(
::
llvm
::
dyn_cast_or_null
<::
infrt
::
GraphOp
>
(
op
))
continue
;
if
(
::
llvm
::
dyn_cast_or_null
<::
infrt
::
ReturnOp
>
(
op
))
continue
;
if
(
::
llvm
::
dyn_cast_or_null
<::
infrt
::
phi
::
TensorMapGetTensorOp
>
(
op
))
continue
;
builder
.
setInsertionPoint
(
op
);
auto
loc
=
getFunction
().
getLoc
();
auto
graph_op
=
builder
.
create
<::
infrt
::
GraphOp
>
(
...
...
paddle/infrt/tests/dialect/disabled_rewrite_conv_bn.mlir
浏览文件 @
9bb3744f
// CHECK-LABEL: @main
func @main() -> tensor<?xf32> {
%a = "pd.feed"() {name="input0"} : () -> tensor<?x3x256x256xf32>
func @main(%a:tensor<?x3x256x256xf32>) -> tensor<?xf32> {
%filter = "pd.constant"(){value = dense<1.000000e+00> : tensor<3x64x3x3xf32>} : () -> tensor<3x64x3x3xf32>
%bias = "pd.constant"(){value = dense<1.000000e+00> : tensor<64xf32>} : () -> tensor<64xf32>
...
...
@@ -11,5 +10,5 @@ func @main() -> tensor<?xf32> {
%c = "pd.conv2d"(%a, %filter, %bias) {} : (tensor<?x3x256x256xf32>, tensor<3x64x3x3xf32>, tensor<64xf32>) -> tensor<?x3x256x256xf32>
%d = "pd.batch_norm"(%c, %scale, %bias2, %mean, %var) {} : (tensor<?x3x256x256xf32>, tensor<64xf32>, tensor<64xf32>, tensor<64xf32>, tensor<64xf32>) -> tensor<?x3x256x256xf32>
"pd.fetch"(%d) {name="output"} :(tensor<?x3x256x256xf32>)->()
}
\ No newline at end of file
infrt.return %d:tensor<?x3x256x256xf32>
}
paddle/infrt/tests/dialect/paddle_ops.mlir
已删除
100644 → 0
浏览文件 @
aeade538
// RUN: infrtopt %s | FileCheck %s
// CHECK-LABEL: @ops
func @ops() {
%a = pd.feed() {name="input0"} : tensor<?xf32>
%b = pd.feed() {name="input1"}: tensor<?xf32>
%d = pd.feed() {name="input3"}: !infrt.lod_tensor<3x4x9xf32, 0>
%c = "pd.matmul"(%a, %b) {transpose_x=true, transpose_y=false} : (tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
infrt.return
}
paddle/infrt/tests/dialect/pd/rewrite.mlir
浏览文件 @
9bb3744f
// RUN: infrtopt --pd-op-fuse %s | FileCheck %s
// CHECK-LABEL: @main
func @main() -> tensor<?xf32> {
%a = "pd.feed"() {name="input0"} : () -> tensor<?xf32>
%b = "pd.feed"() {name="input1"} : () -> tensor<?xf32>
%bias = "pd.feed"() {name="input2"} : () -> tensor<?xf32>
func @main(%arg0: tensor<?xf32>, %arg1: tensor<?xf32>, %arg2:tensor<?xf32>, %arg3:tensor<?xf32>, %arg4:tensor<?xf32>, %arg5:tensor<?xf32>, %arg6:tensor<?xf32>) -> tensor<?xf32> {
%b1 = "pd.feed"() {name="input3"} : () -> tensor<?xf32>
%b2 = "pd.feed"() {name="input4"} : () -> tensor<?xf32>
%bias1 = "pd.feed"() {name="input5"} : () -> tensor<?xf32>
%bias2 = "pd.feed"() {name="input6"} : () -> tensor<?xf32>
// CHECK: %{{[0-9]+}} = "pd.FC"(%{{[0-9]+}}, %{{[0-9]+}}, %{{[0-9]+}}) {in_num_col_dims = 1 : i32} : (tensor<?xf32>, tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%c = "pd.matmul_v2"(%a, %b) {transpose_y=false} : (tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%d = "pd.elementwise_add"(%c, %bias) {axis=1:si32} : (tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
// CHECK: %0 = "pd.FC"(%arg0, %arg1, %arg4) {in_num_col_dims = 1 : i32} : (tensor<?xf32>, tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%c = "pd.matmul_v2"(%arg0, %arg1) {transpose_y=false} : (tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%d = "pd.elementwise_add"(%c, %arg4) {axis=1:si32} : (tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%e = "pd.relu6"(%d) {} : (tensor<?xf32>) -> tensor<?xf32>
// CHECK: %
{{[0-9]+}} = "pd.FC"(%{{[0-9]+}}, %{{[0-9]+}}, %{{[0-9]+}}
) {in_num_col_dims = 1 : i32} : (tensor<?xf32>, tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%c1 = "pd.matmul_v2"(%e, %
b1
) {transpose_x=false, transpose_y=false} : (tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%d1 = "pd.elementwise_add"(%c1, %
bias1
) {axis=1:si32} : (tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
// CHECK: %
2 = "pd.FC"(%1, %arg2, %arg5
) {in_num_col_dims = 1 : i32} : (tensor<?xf32>, tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%c1 = "pd.matmul_v2"(%e, %
arg2
) {transpose_x=false, transpose_y=false} : (tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%d1 = "pd.elementwise_add"(%c1, %
arg5
) {axis=1:si32} : (tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%e1 = "pd.relu"(%d1) {} : (tensor<?xf32>) -> tensor<?xf32>
// CHECK: %
{{[0-9]+}} = "pd.FC"(%{{[0-9]+}}, %{{[0-9]+}}, %{{[0-9]+}}
) {in_num_col_dims = 1 : i32} : (tensor<?xf32>, tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%c2 = "pd.matmul_v2"(%e1, %
b2
) {transpose_x=true, transpose_y=false} : (tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%d2 = "pd.elementwise_add"(%c2, %
bias2
) {axis=1:si32} : (tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
// CHECK: %
4 = "pd.FC"(%3, %arg3, %arg6
) {in_num_col_dims = 1 : i32} : (tensor<?xf32>, tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%c2 = "pd.matmul_v2"(%e1, %
arg3
) {transpose_x=true, transpose_y=false} : (tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%d2 = "pd.elementwise_add"(%c2, %
arg6
) {axis=1:si32} : (tensor<?xf32>, tensor<?xf32>) -> tensor<?xf32>
%e2 = "pd.relu"(%d2) {} : (tensor<?xf32>) -> tensor<?xf32>
"pd.fetch"(%e2) {name="output"} :(tensor<?xf32>)->()
infrt.return %e2:tensor<?xf32>
}
paddle/infrt/tests/dialect/phi/phi_pass.mlir
浏览文件 @
9bb3744f
// RUN: infrtopt -phi-op-convert -infrt-op-fuse %s
// CHECK-LABEL: @ops
func @ops() {
%a = pd.feed() {name="input0"} : !infrt.lod_tensor<?xf32,0>
%b = pd.feed() {name="input1"} : !infrt.lod_tensor<?xf32,0>
%d = pd.feed() {name="input3"} : !infrt.lod_tensor<3x4x9xf32, 0>
func @ops(%a:!infrt.lod_tensor<?xf32,0>, %b:!infrt.lod_tensor<?xf32,0>) {
%g = "pd.elementwise_add"(%a, %b) {axis=1:si32} : (!infrt.lod_tensor<?xf32,0>, !infrt.lod_tensor<?xf32>) -> tensor<?xf32>
%h = "pd.abs"(%g):(tensor<?xf32>) -> tensor<?xf32>
"pd.fetch"(%h) {name="output"} :(tensor<?xf32>)->()
infrt.return %h:tensor<?xf32>
}
// CHECK-LABEL: @op_execute
func @op_execute(%a:!infrt.lod_tensor<?xf32,0>, %b:!infrt.lod_tensor<?xf32,0>, %c:!infrt.lod_tensor<?xf32,0>) -> !infrt.lod_tensor<?xf32,0> {
%g = "pd.elementwise_add"(%a, %b) {axis=1:si32} : (!infrt.lod_tensor<?xf32,0>, !infrt.lod_tensor<?xf32>) -> tensor<?xf32>
%h = "pd.abs"(%g):(tensor<?xf32>) -> tensor<?xf32>
"pd.fetch"(%h) {name="output"} :(tensor<?xf32>)->()
infrt.return %h:tensor<?xf32>
}
tools/infrt/custom_pdop.td
已删除
100644 → 0
浏览文件 @
aeade538
def PD_FeedOp : PD_Op<"feed", [NoSideEffect]> {
let summary = "Feed Op";
let description = [{
Feed a tensor into the model.
}];
let arguments = (ins StrAttr:$name);
let results = (outs PD_Tensor:$out);
let assemblyFormat = [{
`(` `)` attr-dict `:` type($out)
}];
}
def PD_FetchOp : PD_Op<"fetch", [Terminator]> {
let summary = "fetch Op";
let description = [{
Return the output tensor from the subgraph.
}];
let arguments = (ins PD_Tensor :$inputs, StrAttr:$name);
}
def PD_ConstantOp : PD_Op<"constant", [NoSideEffect, ConstantLike, DeclareOpInterfaceMethods<InferTypeOpInterface>, AllTypesMatch<["value", "output"]>]> {
let summary = "constant Op";
let description = [{}];
let arguments = (ins ElementsAttr:$value);
let results = (outs PD_Tensor:$output);
let hasFolder = 1;
let builders = [
OpBuilder<(ins "mlir::Attribute":$value)>,
];
}
tools/infrt/generate_pd_op_dialect_from_paddle_op_maker.py
浏览文件 @
9bb3744f
...
...
@@ -209,7 +209,6 @@ def get_constraint(op_type, op_proto):
# funtion to generate paddle op dialect file
def
convert_op_proto_into_mlir
(
op_descs
):
dst_dialect_file
=
"../../paddle/infrt/dialect/pd/ir/pd_ops.td"
custom_dialect_file
=
"custom_pdop.td"
# 1. Head files
comment_
=
"/*===- TableGen'source file -----------------------------------------------===*
\\\n\
...
...
@@ -372,19 +371,13 @@ def convert_op_proto_into_mlir(op_descs):
ops_mlir_file
.
write
(
RESULTS
)
ops_mlir_file
.
write
(
"}
\n
"
)
with
open
(
dst_dialect_file
,
'a'
)
as
ops_mlir_file
:
ops_mlir_file
.
write
(
"
\n
#endif // PD_OPS"
)
print
(
"Skipped ops num: "
+
str
(
len
(
skipped_op_list
)))
print
(
"Automatically generated op dialects num: "
+
str
(
len
(
automatically_generated_op_dialect
)))
# 3. custom op dialect and end of file
with
open
(
dst_dialect_file
,
'a'
)
as
ops_mlir_file
:
with
open
(
custom_dialect_file
,
'r'
)
as
custom_ops_file
:
custom_ops
=
custom_ops_file
.
readlines
()
ops_mlir_file
.
writelines
(
custom_ops
)
end_
=
"
\n
#endif // PD_OPS"
ops_mlir_file
.
write
(
end_
)
if
__name__
==
"__main__"
:
all_op_protos_dict
=
get_all_ops_desc
()
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录