Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
Paddle-Lite
提交
d6825bf3
P
Paddle-Lite
项目概览
PaddlePaddle
/
Paddle-Lite
通知
338
Star
4
Fork
1
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
271
列表
看板
标记
里程碑
合并请求
78
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle-Lite
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
271
Issue
271
列表
看板
标记
里程碑
合并请求
78
合并请求
78
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
d6825bf3
编写于
5月 20, 2018
作者:
朔-望
提交者:
GitHub
5月 20, 2018
浏览文件
操作
浏览文件
下载
差异文件
Merge pull request #218 from allonli/develop
update code style to Google
上级
7b384fb2
7cc864ea
变更
98
隐藏空白更改
内联
并排
Showing
98 changed file
with
2436 addition
and
2451 deletion
+2436
-2451
.clang-format
.clang-format
+1
-1
src/common/log.h
src/common/log.h
+27
-21
src/common/type_define.h
src/common/type_define.h
+5
-4
src/common/types.h
src/common/types.h
+3
-2
src/common/variant.h
src/common/variant.h
+23
-11
src/framework/attribute.cpp
src/framework/attribute.cpp
+1
-1
src/framework/attribute.h
src/framework/attribute.h
+65
-60
src/framework/block_desc.cpp
src/framework/block_desc.cpp
+2
-2
src/framework/block_desc.h
src/framework/block_desc.h
+7
-6
src/framework/data_layout.h
src/framework/data_layout.h
+11
-11
src/framework/data_transform.cpp
src/framework/data_transform.cpp
+2
-2
src/framework/data_transform.h
src/framework/data_transform.h
+2
-2
src/framework/data_type.h
src/framework/data_type.h
+1
-1
src/framework/ddim.cc
src/framework/ddim.cc
+66
-51
src/framework/ddim.h
src/framework/ddim.h
+10
-6
src/framework/dim.h
src/framework/dim.h
+53
-29
src/framework/executor.cpp
src/framework/executor.cpp
+2
-2
src/framework/executor.h
src/framework/executor.h
+6
-5
src/framework/framework.pb.cpp
src/framework/framework.pb.cpp
+1292
-1332
src/framework/framework.pb.h
src/framework/framework.pb.h
+298
-315
src/framework/lod_tensor.cc
src/framework/lod_tensor.cc
+16
-26
src/framework/lod_tensor.h
src/framework/lod_tensor.h
+8
-9
src/framework/op_desc.cpp
src/framework/op_desc.cpp
+2
-2
src/framework/op_desc.h
src/framework/op_desc.h
+4
-4
src/framework/op_info.h
src/framework/op_info.h
+12
-8
src/framework/op_kernel_type.h
src/framework/op_kernel_type.h
+2
-2
src/framework/op_proto_maker.h
src/framework/op_proto_maker.h
+2
-2
src/framework/operator.cpp
src/framework/operator.cpp
+6
-3
src/framework/operator.h
src/framework/operator.h
+11
-9
src/framework/paddle_mobile_object.h
src/framework/paddle_mobile_object.h
+4
-4
src/framework/program-optimize/node.cpp
src/framework/program-optimize/node.cpp
+2
-2
src/framework/program-optimize/node.h
src/framework/program-optimize/node.h
+4
-4
src/framework/program-optimize/program_optimize.cpp
src/framework/program-optimize/program_optimize.cpp
+4
-4
src/framework/program-optimize/program_optimize.h
src/framework/program-optimize/program_optimize.h
+6
-6
src/framework/program.cpp
src/framework/program.cpp
+1
-1
src/framework/program.h
src/framework/program.h
+4
-4
src/framework/program_desc.cpp
src/framework/program_desc.cpp
+2
-2
src/framework/program_desc.h
src/framework/program_desc.h
+4
-4
src/framework/scope.cc
src/framework/scope.cc
+2
-2
src/framework/scope.h
src/framework/scope.h
+7
-7
src/framework/selected_rows.h
src/framework/selected_rows.h
+4
-4
src/framework/tensor.h
src/framework/tensor.h
+20
-12
src/framework/tensor_util.cc
src/framework/tensor_util.cc
+15
-11
src/framework/tensor_util.h
src/framework/tensor_util.h
+3
-3
src/framework/var_desc.cpp
src/framework/var_desc.cpp
+2
-2
src/framework/var_desc.h
src/framework/var_desc.h
+21
-21
src/framework/var_type.h
src/framework/var_type.h
+2
-2
src/framework/variable.h
src/framework/variable.h
+13
-9
src/io.cpp
src/io.cpp
+101
-167
src/io.h
src/io.h
+3
-3
src/memory/t_malloc.cc
src/memory/t_malloc.cc
+2
-2
src/memory/t_malloc.h
src/memory/t_malloc.h
+8
-6
src/operators/batchnorm_op.cpp
src/operators/batchnorm_op.cpp
+2
-2
src/operators/batchnorm_op.h
src/operators/batchnorm_op.h
+4
-4
src/operators/concat_op.cpp
src/operators/concat_op.cpp
+2
-2
src/operators/concat_op.h
src/operators/concat_op.h
+4
-4
src/operators/conv_op.cpp
src/operators/conv_op.cpp
+2
-2
src/operators/conv_op.h
src/operators/conv_op.h
+4
-4
src/operators/elementwise_add_op.cpp
src/operators/elementwise_add_op.cpp
+2
-2
src/operators/elementwise_add_op.h
src/operators/elementwise_add_op.h
+4
-4
src/operators/kernel/arm/batchnorm_kernel.cpp
src/operators/kernel/arm/batchnorm_kernel.cpp
+2
-2
src/operators/kernel/arm/concat_kernel.cpp
src/operators/kernel/arm/concat_kernel.cpp
+5
-4
src/operators/kernel/arm/conv_kernel.cpp
src/operators/kernel/arm/conv_kernel.cpp
+4
-3
src/operators/kernel/arm/elementwise_add_kernel.cpp
src/operators/kernel/arm/elementwise_add_kernel.cpp
+4
-3
src/operators/kernel/arm/lrn_kernel.cpp
src/operators/kernel/arm/lrn_kernel.cpp
+4
-3
src/operators/kernel/arm/mul_kernel.cpp
src/operators/kernel/arm/mul_kernel.cpp
+4
-3
src/operators/kernel/arm/pool_kernel.cpp
src/operators/kernel/arm/pool_kernel.cpp
+5
-4
src/operators/kernel/batchnorm_kernel.h
src/operators/kernel/batchnorm_kernel.h
+3
-3
src/operators/kernel/concat_kernel.h
src/operators/kernel/concat_kernel.h
+3
-3
src/operators/kernel/conv_kernel.h
src/operators/kernel/conv_kernel.h
+3
-3
src/operators/kernel/elementwise_add_kernel.h
src/operators/kernel/elementwise_add_kernel.h
+3
-3
src/operators/kernel/fpga/conv_kernel.cpp
src/operators/kernel/fpga/conv_kernel.cpp
+1
-1
src/operators/kernel/lrn_kernel.h
src/operators/kernel/lrn_kernel.h
+5
-4
src/operators/kernel/mul_kernel.h
src/operators/kernel/mul_kernel.h
+3
-3
src/operators/kernel/pool_kernel.h
src/operators/kernel/pool_kernel.h
+3
-3
src/operators/lrn_op.cpp
src/operators/lrn_op.cpp
+4
-3
src/operators/lrn_op.h
src/operators/lrn_op.h
+4
-4
src/operators/math/elementwise_op_function.h
src/operators/math/elementwise_op_function.h
+18
-14
src/operators/math/im2col.cc
src/operators/math/im2col.cc
+15
-11
src/operators/math/im2col.h
src/operators/math/im2col.h
+5
-5
src/operators/math/math_function.cc
src/operators/math/math_function.cc
+3
-3
src/operators/math/math_function.h
src/operators/math/math_function.h
+4
-4
src/operators/math/pool3x3.h
src/operators/math/pool3x3.h
+1
-1
src/operators/math/pool_2x2.h
src/operators/math/pool_2x2.h
+1
-1
src/operators/math/pooling.cpp
src/operators/math/pooling.cpp
+4
-5
src/operators/math/pooling.h
src/operators/math/pooling.h
+9
-7
src/operators/math/transform.h
src/operators/math/transform.h
+3
-3
src/operators/math/vol2col.cc
src/operators/math/vol2col.cc
+9
-7
src/operators/math/vol2col.h
src/operators/math/vol2col.h
+9
-7
src/operators/mul_op.cpp
src/operators/mul_op.cpp
+4
-3
src/operators/mul_op.h
src/operators/mul_op.h
+4
-4
src/operators/op_param.cpp
src/operators/op_param.cpp
+2
-2
src/operators/op_param.h
src/operators/op_param.h
+18
-18
src/operators/pool_op.cpp
src/operators/pool_op.cpp
+2
-2
src/operators/pool_op.h
src/operators/pool_op.h
+4
-4
src/platform/data_type.h
src/platform/data_type.h
+56
-56
src/platform/macros.h
src/platform/macros.h
+5
-5
test/operators/test_cov_op.cpp
test/operators/test_cov_op.cpp
+2
-2
未找到文件。
.clang-format
浏览文件 @
d6825bf3
---
Language: Cpp
BasedOnStyle:
LLVM
BasedOnStyle:
Google
Standard: Cpp11
...
src/common/log.h
浏览文件 @
d6825bf3
...
...
@@ -51,12 +51,13 @@ struct Print;
struct
Print
{
friend
struct
ToLog
;
template
<
typename
T
>
Print
&
operator
<<
(
T
const
&
value
)
{
template
<
typename
T
>
Print
&
operator
<<
(
T
const
&
value
)
{
buffer_
<<
value
;
return
*
this
;
}
private:
private:
void
print
(
LogLevel
level
)
{
buffer_
<<
std
::
endl
;
if
(
level
==
kLOG_ERROR
)
{
...
...
@@ -76,14 +77,15 @@ struct ToLog {
printer_
<<
logs
[
level
]
<<
" "
<<
info
<<
":"
<<
std
::
string
(
blanks
,
' '
);
}
template
<
typename
T
>
ToLog
&
operator
<<
(
T
const
&
value
)
{
template
<
typename
T
>
ToLog
&
operator
<<
(
T
const
&
value
)
{
printer_
<<
value
;
return
*
this
;
}
~
ToLog
()
{
printer_
.
print
(
level_
);
}
private:
private:
LogLevel
level_
;
Print
printer_
;
};
...
...
@@ -109,16 +111,16 @@ private:
<< (strrchr(__FILE__, '/') ? (strrchr(__FILE__, '/') + 1) : __FILE__) \
<< "] [line: " << __LINE__ << "] ") \
.str())
}
// namespace paddle_mobile
}
// namespace paddle_mobile
#define LOGF(level, format, ...)
\
if (level > paddle_mobile::log_level) {
\
} else
\
#define LOGF(level, format, ...) \
if (level > paddle_mobile::log_level) { \
} else \
printf(format, ##__VA_ARGS__)
#define DLOGF(format, ...)
\
if (paddle_mobile::kLOG_DEBUG > paddle_mobile::log_level) {
\
} else
\
#define DLOGF(format, ...) \
if (paddle_mobile::kLOG_DEBUG > paddle_mobile::log_level) { \
} else \
printf(format, ##__VA_ARGS__)
#else
...
...
@@ -140,30 +142,34 @@ enum LogLevel {
struct
ToLog
;
struct
Print
{
friend
struct
ToLog
;
template
<
typename
T
>
Print
&
operator
<<
(
T
const
&
value
)
{}
template
<
typename
T
>
Print
&
operator
<<
(
T
const
&
value
)
{}
private:
private:
};
struct
ToLog
{
ToLog
(
LogLevel
level
)
{}
template
<
typename
T
>
ToLog
&
operator
<<
(
T
const
&
value
)
{
return
*
this
;
}
template
<
typename
T
>
ToLog
&
operator
<<
(
T
const
&
value
)
{
return
*
this
;
}
};
#define LOG(level)
\
if (true) {
\
} else
\
#define LOG(level) \
if (true) { \
} else \
paddle_mobile::ToLog(level)
#define DLOG
\
if (true) {
\
} else
\
#define DLOG \
if (true) { \
} else \
paddle_mobile::ToLog(paddle_mobile::kLOG_DEBUG)
#define LOGF(level, format, ...)
#define DLOGF(format, ...)
}
// namespace paddle_mobile
}
// namespace paddle_mobile
#endif
src/common/type_define.h
浏览文件 @
d6825bf3
...
...
@@ -17,18 +17,19 @@ SOFTWARE.
==============================================================================*/
#pragma once;
#include "framework/attribute.h"
#include <map>
#include <string>
#include "framework/attribute.h"
namespace
paddle_mobile
{
namespace
framework
{
template
<
typename
Dtype
>
class
OperatorBase
;
template
<
typename
Dtype
>
class
OperatorBase
;
class
OpDesc
;
class
BlockDesc
;
class
InferShapeContext
;
}
// namespace framework
}
// namespace framework
using
VariableNameMap
=
std
::
map
<
std
::
string
,
std
::
vector
<
std
::
string
>>
;
...
...
@@ -49,4 +50,4 @@ using InferVarTypeFN = std::function<void(const framework::OpDesc & /*op_desc*/,
framework
::
BlockDesc
*
/*block*/
)
>
;
using
InferShapeFN
=
std
::
function
<
void
(
framework
::
InferShapeContext
*
)
>
;
};
// namespace paddle_mobile
};
// namespace paddle_mobile
src/common/types.h
浏览文件 @
d6825bf3
...
...
@@ -24,7 +24,8 @@ enum class Precision : int { FP32 = 0 };
//! device type
enum
DeviceTypeEnum
{
kINVALID
=
-
1
,
kCPU
=
0
,
kFPGA
=
1
,
kGPU_MALI
=
2
};
template
<
DeviceTypeEnum
T
>
struct
DeviceType
{};
template
<
DeviceTypeEnum
T
>
struct
DeviceType
{};
typedef
DeviceType
<
kCPU
>
CPU
;
typedef
DeviceType
<
kFPGA
>
FPGA
;
...
...
@@ -60,4 +61,4 @@ enum PMStatus {
PMUnImplError
=
0x07
,
/*!< Unimplement error. */
PMWrongDevice
=
0x08
/*!< un-correct device. */
};
}
// namespace paddle_mobile
}
// namespace paddle_mobile
src/common/variant.h
浏览文件 @
d6825bf3
...
...
@@ -21,9 +21,13 @@ SOFTWARE.
#pragma once
namespace
paddle_mobile
{
template
<
int
ID
,
typename
Type
>
struct
IDToType
{
typedef
Type
type_t
;
};
template
<
int
ID
,
typename
Type
>
struct
IDToType
{
typedef
Type
type_t
;
};
template
<
typename
F
,
typename
...
Ts
>
struct
VariantHelper
{
template
<
typename
F
,
typename
...
Ts
>
struct
VariantHelper
{
static
const
size_t
size
=
sizeof
(
F
)
>
VariantHelper
<
Ts
...
>::
size
?
sizeof
(
F
)
:
VariantHelper
<
Ts
...
>::
size
;
...
...
@@ -37,7 +41,8 @@ template <typename F, typename... Ts> struct VariantHelper {
}
};
template
<
typename
F
>
struct
VariantHelper
<
F
>
{
template
<
typename
F
>
struct
VariantHelper
<
F
>
{
static
const
size_t
size
=
sizeof
(
F
);
inline
static
void
Destroy
(
size_t
id
,
void
*
data
)
{
if
(
id
==
typeid
(
F
).
hash_code
())
{
...
...
@@ -48,8 +53,9 @@ template <typename F> struct VariantHelper<F> {
}
};
template
<
size_t
size
>
class
RawData
{
public:
template
<
size_t
size
>
class
RawData
{
public:
char
data
[
size
];
RawData
()
{}
RawData
(
const
RawData
&
raw_data
)
{
strcpy
(
data
,
raw_data
.
data
);
}
...
...
@@ -58,7 +64,8 @@ public:
// }
};
template
<
typename
...
Ts
>
struct
Variant
{
template
<
typename
...
Ts
>
struct
Variant
{
Variant
(
const
Variant
&
variant
)
{
// std::cout << " 赋值构造函数 " << std::endl;
type_id
=
variant
.
type_id
;
...
...
@@ -70,13 +77,15 @@ template <typename... Ts> struct Variant {
// helper::Destroy(type_id, &data);
}
template
<
typename
T
,
typename
...
Args
>
void
Set
(
Args
&&
...
args
)
{
template
<
typename
T
,
typename
...
Args
>
void
Set
(
Args
&&
...
args
)
{
helper
::
Destroy
(
type_id
,
&
data
);
new
(
&
data
)
T
(
std
::
forward
<
Args
>
(
args
)...);
type_id
=
typeid
(
T
).
hash_code
();
}
template
<
typename
T
>
T
&
Get
()
const
{
template
<
typename
T
>
T
&
Get
()
const
{
if
(
type_id
==
typeid
(
T
).
hash_code
())
{
return
*
const_cast
<
T
*>
(
reinterpret_cast
<
const
T
*>
(
&
data
));
}
else
{
...
...
@@ -87,13 +96,16 @@ template <typename... Ts> struct Variant {
size_t
TypeId
()
const
{
return
type_id
;
}
private:
private:
static
inline
size_t
invalid_type
()
{
return
typeid
(
void
).
hash_code
();
}
typedef
VariantHelper
<
Ts
...
>
helper
;
size_t
type_id
;
RawData
<
helper
::
size
>
data
;
};
template
<
typename
T
>
struct
Vistor
{
typedef
T
type_t
;
};
template
<
typename
T
>
struct
Vistor
{
typedef
T
type_t
;
};
}
// namespace paddle_mobile
}
// namespace paddle_mobile
src/framework/attribute.cpp
浏览文件 @
d6825bf3
...
...
@@ -18,4 +18,4 @@ SOFTWARE.
namespace
paddle_mobile
{
namespace
framework
{}
}
// namespace paddle_mobile
}
// namespace paddle_mobile
src/framework/attribute.h
浏览文件 @
d6825bf3
...
...
@@ -27,80 +27,84 @@ namespace framework {
class
BlockDesc
;
class
Attribute
{
public:
public:
static
Attribute
GetAttrValue
(
const
proto
::
OpDesc
::
Attr
&
attr_desc
)
{
// std::cout << "begin get attr value" << std::endl;
Attribute
attr
;
switch
(
attr_desc
.
type
())
{
case
proto
::
AttrType
::
BOOLEAN
:
{
attr
.
Set
<
bool
>
(
attr_desc
.
b
());
break
;
}
case
proto
::
AttrType
::
INT
:
{
attr
.
Set
<
int
>
(
attr_desc
.
i
());
break
;
}
case
proto
::
AttrType
::
FLOAT
:
{
attr
.
Set
<
float
>
(
attr_desc
.
f
());
break
;
}
case
proto
::
AttrType
::
STRING
:
{
attr
.
Set
<
std
::
string
>
(
attr_desc
.
s
());
break
;
}
case
proto
::
AttrType
::
BOOLEANS
:
{
std
::
vector
<
bool
>
val
(
attr_desc
.
bools_size
());
for
(
int
i
=
0
;
i
<
attr_desc
.
bools_size
();
++
i
)
{
val
[
i
]
=
attr_desc
.
bools
(
i
);
case
proto
::
AttrType
::
BOOLEAN
:
{
attr
.
Set
<
bool
>
(
attr_desc
.
b
());
break
;
}
attr
.
Set
<
std
::
vector
<
bool
>>
(
val
);
break
;
}
case
proto
::
AttrType
::
INTS
:
{
std
::
vector
<
int
>
val
(
attr_desc
.
ints_size
());
for
(
int
i
=
0
;
i
<
attr_desc
.
ints_size
();
++
i
)
{
val
[
i
]
=
attr_desc
.
ints
(
i
);
case
proto
::
AttrType
::
INT
:
{
attr
.
Set
<
int
>
(
attr_desc
.
i
());
break
;
}
attr
.
Set
<
std
::
vector
<
int
>>
(
val
);
break
;
}
case
proto
::
AttrType
::
FLOATS
:
{
std
::
vector
<
float
>
val
(
attr_desc
.
floats_size
());
for
(
int
i
=
0
;
i
<
attr_desc
.
floats_size
();
++
i
)
{
val
[
i
]
=
attr_desc
.
floats
(
i
);
case
proto
::
AttrType
::
FLOAT
:
{
attr
.
Set
<
float
>
(
attr_desc
.
f
());
break
;
}
attr
.
Set
<
std
::
vector
<
float
>>
(
val
);
break
;
}
case
proto
::
AttrType
::
STRINGS
:
{
std
::
vector
<
std
::
string
>
val
(
attr_desc
.
strings_size
());
for
(
int
i
=
0
;
i
<
attr_desc
.
strings_size
();
++
i
)
{
val
[
i
]
=
attr_desc
.
strings
(
i
);
case
proto
::
AttrType
::
STRING
:
{
attr
.
Set
<
std
::
string
>
(
attr_desc
.
s
());
break
;
}
attr
.
Set
<
std
::
vector
<
std
::
string
>>
(
val
);
break
;
}
case
proto
::
AttrType
::
LONG
:
{
attr
.
Set
<
int64_t
>
(
attr_desc
.
l
());
break
;
}
default:
// std::cout << " not support " << std::endl;
break
;
case
proto
::
AttrType
::
BOOLEANS
:
{
std
::
vector
<
bool
>
val
(
attr_desc
.
bools_size
());
for
(
int
i
=
0
;
i
<
attr_desc
.
bools_size
();
++
i
)
{
val
[
i
]
=
attr_desc
.
bools
(
i
);
}
attr
.
Set
<
std
::
vector
<
bool
>>
(
val
);
break
;
}
case
proto
::
AttrType
::
INTS
:
{
std
::
vector
<
int
>
val
(
attr_desc
.
ints_size
());
for
(
int
i
=
0
;
i
<
attr_desc
.
ints_size
();
++
i
)
{
val
[
i
]
=
attr_desc
.
ints
(
i
);
}
attr
.
Set
<
std
::
vector
<
int
>>
(
val
);
break
;
}
case
proto
::
AttrType
::
FLOATS
:
{
std
::
vector
<
float
>
val
(
attr_desc
.
floats_size
());
for
(
int
i
=
0
;
i
<
attr_desc
.
floats_size
();
++
i
)
{
val
[
i
]
=
attr_desc
.
floats
(
i
);
}
attr
.
Set
<
std
::
vector
<
float
>>
(
val
);
break
;
}
case
proto
::
AttrType
::
STRINGS
:
{
std
::
vector
<
std
::
string
>
val
(
attr_desc
.
strings_size
());
for
(
int
i
=
0
;
i
<
attr_desc
.
strings_size
();
++
i
)
{
val
[
i
]
=
attr_desc
.
strings
(
i
);
}
attr
.
Set
<
std
::
vector
<
std
::
string
>>
(
val
);
break
;
}
case
proto
::
AttrType
::
LONG
:
{
attr
.
Set
<
int64_t
>
(
attr_desc
.
l
());
break
;
}
default:
// std::cout << " not support " << std::endl;
break
;
}
// std::cout << "end get attr value" << std::endl;
return
attr
;
}
Attribute
()
{}
template
<
typename
T
,
typename
...
Args
>
Attribute
&
Set
(
Args
&&
...
args
)
{
template
<
typename
T
,
typename
...
Args
>
Attribute
&
Set
(
Args
&&
...
args
)
{
variant_
.
Set
<
T
>
(
args
...);
return
*
this
;
}
template
<
typename
T
>
T
&
Get
()
const
{
return
variant_
.
Get
<
T
>
();
}
template
<
typename
T
>
T
&
Get
()
const
{
return
variant_
.
Get
<
T
>
();
}
private:
private:
Variant
<
int
,
float
,
std
::
string
,
std
::
vector
<
int
>
,
std
::
vector
<
float
>
,
std
::
vector
<
std
::
string
>
,
bool
,
std
::
vector
<
bool
>
,
BlockDesc
*
,
int64_t
>
...
...
@@ -110,10 +114,11 @@ private:
using
AttributeMap
=
std
::
unordered_map
<
std
::
string
,
Attribute
>
;
class
AttrReader
{
public:
public:
explicit
AttrReader
(
const
AttributeMap
&
attrs
)
:
attrs_
(
attrs
)
{}
template
<
typename
T
>
inline
T
Get
(
const
std
::
string
&
name
)
const
{
template
<
typename
T
>
inline
T
Get
(
const
std
::
string
&
name
)
const
{
// PADDLE_ENFORCE(attrs_.count(name) != 0, "%s should
// be in
// AttributeMap",
...
...
@@ -121,9 +126,9 @@ public:
return
((
Attribute
)
attrs_
.
at
(
name
)).
Get
<
T
>
();
}
private:
private:
const
AttributeMap
&
attrs_
;
};
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/block_desc.cpp
浏览文件 @
d6825bf3
...
...
@@ -46,5 +46,5 @@ BlockDesc::BlockDesc(const proto::BlockDesc &desc) : desc_(desc) {
}
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/block_desc.h
浏览文件 @
d6825bf3
...
...
@@ -27,7 +27,7 @@ namespace paddle_mobile {
namespace
framework
{
class
BlockDesc
:
PaddleMobileObject
{
public:
public:
BlockDesc
(
const
proto
::
BlockDesc
&
desc
);
const
int
&
ID
()
const
{
return
desc_
.
idx
();
}
...
...
@@ -45,18 +45,19 @@ public:
std
::
vector
<
std
::
shared_ptr
<
VarDesc
>>
Vars
()
const
;
std
::
vector
<
std
::
shared_ptr
<
OpDesc
>>
Ops
()
const
;
private:
private:
proto
::
BlockDesc
desc_
;
std
::
vector
<
std
::
shared_ptr
<
OpDesc
>>
ops_
;
std
::
unordered_map
<
std
::
string
,
std
::
shared_ptr
<
VarDesc
>>
vars_
;
};
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
namespace
std
{
template
<
>
struct
hash
<
paddle_mobile
::
framework
::
BlockDesc
>
{
template
<
>
struct
hash
<
paddle_mobile
::
framework
::
BlockDesc
>
{
typedef
paddle_mobile
::
framework
::
BlockDesc
argument_type
;
typedef
std
::
size_t
result_type
;
result_type
operator
()(
argument_type
const
&
s
)
const
noexcept
{
...
...
@@ -66,4 +67,4 @@ template <> struct hash<paddle_mobile::framework::BlockDesc> {
}
};
}
// namespace std
}
// namespace std
src/framework/data_layout.h
浏览文件 @
d6825bf3
...
...
@@ -46,15 +46,15 @@ inline DataLayout StringToDataLayout(const std::string &str) {
inline
std
::
string
DataLayoutToString
(
const
DataLayout
&
data_layout
)
{
switch
(
data_layout
)
{
case
DataLayout
::
kNHWC
:
return
"NHWC"
;
case
DataLayout
::
kNCHW
:
return
"NCHW"
;
case
DataLayout
::
kAnyLayout
:
return
"ANY_LAYOUT"
;
default:
break
;
// std::cout << "unknown DataLayou %d", data_layout;
case
DataLayout
::
kNHWC
:
return
"NHWC"
;
case
DataLayout
::
kNCHW
:
return
"NCHW"
;
case
DataLayout
::
kAnyLayout
:
return
"ANY_LAYOUT"
;
default:
break
;
// std::cout << "unknown DataLayou %d", data_layout;
}
}
...
...
@@ -63,5 +63,5 @@ inline std::ostream &operator<<(std::ostream &out, const DataLayout &l) {
return
out
;
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/data_transform.cpp
浏览文件 @
d6825bf3
...
...
@@ -88,5 +88,5 @@ void CopyVariableWithTensor(const Variable &in_var, const Tensor &tensor,
// }
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/data_transform.h
浏览文件 @
d6825bf3
...
...
@@ -37,5 +37,5 @@ void DataTransform(const OpKernelType &expected_kernel_type,
void
CopyVariableWithTensor
(
const
Variable
&
in_var
,
const
Tensor
&
tensor
,
Variable
&
out_var
);
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/data_type.h
浏览文件 @
d6825bf3
...
...
@@ -40,4 +40,4 @@ namespace framework {
// }
// }
}
}
// namespace paddle_mobile
}
// namespace paddle_mobile
src/framework/ddim.cc
浏览文件 @
d6825bf3
...
...
@@ -19,49 +19,53 @@ namespace framework {
/// @cond HIDDEN
template
<
int
i
>
Dim
<
i
>
make_dim
(
const
int64_t
*
d
)
{
template
<
int
i
>
Dim
<
i
>
make_dim
(
const
int64_t
*
d
)
{
return
Dim
<
i
>
(
*
d
,
make_dim
<
i
-
1
>
(
d
+
1
));
}
template
<
>
Dim
<
0
>
make_dim
<
0
>
(
const
int64_t
*
d
)
{
return
Dim
<
0
>
(
*
d
);
}
template
<
>
Dim
<
0
>
make_dim
<
0
>
(
const
int64_t
*
d
)
{
return
Dim
<
0
>
(
*
d
);
}
void
make_ddim
(
DDim
&
ddim
,
const
int64_t
*
dims
,
int
n
)
{
switch
(
n
)
{
case
0
:
ddim
=
make_dim
<
0
>
(
dims
);
break
;
case
1
:
ddim
=
make_dim
<
1
>
(
dims
);
break
;
case
2
:
ddim
=
make_dim
<
2
>
(
dims
);
break
;
case
3
:
ddim
=
make_dim
<
3
>
(
dims
);
break
;
case
4
:
ddim
=
make_dim
<
4
>
(
dims
);
break
;
case
5
:
ddim
=
make_dim
<
5
>
(
dims
);
break
;
case
6
:
ddim
=
make_dim
<
6
>
(
dims
);
break
;
case
7
:
ddim
=
make_dim
<
7
>
(
dims
);
break
;
case
8
:
ddim
=
make_dim
<
8
>
(
dims
);
break
;
case
9
:
ddim
=
make_dim
<
9
>
(
dims
);
break
;
default:
// std::cout << "Dynamic dimensions must have between [1,
// 9]
// dimensions.";
break
;
case
0
:
ddim
=
make_dim
<
0
>
(
dims
);
break
;
case
1
:
ddim
=
make_dim
<
1
>
(
dims
);
break
;
case
2
:
ddim
=
make_dim
<
2
>
(
dims
);
break
;
case
3
:
ddim
=
make_dim
<
3
>
(
dims
);
break
;
case
4
:
ddim
=
make_dim
<
4
>
(
dims
);
break
;
case
5
:
ddim
=
make_dim
<
5
>
(
dims
);
break
;
case
6
:
ddim
=
make_dim
<
6
>
(
dims
);
break
;
case
7
:
ddim
=
make_dim
<
7
>
(
dims
);
break
;
case
8
:
ddim
=
make_dim
<
8
>
(
dims
);
break
;
case
9
:
ddim
=
make_dim
<
9
>
(
dims
);
break
;
default:
// std::cout << "Dynamic dimensions must have between [1,
// 9]
// dimensions.";
break
;
}
}
...
...
@@ -90,24 +94,28 @@ DDim make_ddim(const std::vector<int> &dims) {
// XXX For some reason, putting this in an anonymous namespace causes
// errors
struct
DynamicMutableIndexer
:
Vistor
<
int64_t
&>
{
public:
public:
explicit
DynamicMutableIndexer
(
int
idx
)
:
idx_
(
idx
)
{}
template
<
int
D
>
int64_t
&
operator
()(
Dim
<
D
>
&
dim
)
const
{
return
dim
[
idx_
];
}
template
<
int
D
>
int64_t
&
operator
()(
Dim
<
D
>
&
dim
)
const
{
return
dim
[
idx_
];
}
private:
private:
int
idx_
;
};
struct
DynamicConstIndexer
:
public
Vistor
<
int64_t
>
{
public:
public:
explicit
DynamicConstIndexer
(
int
idx
)
:
idx_
(
idx
)
{}
template
<
int
D
>
int64_t
operator
()(
const
Dim
<
D
>
&
dim
)
const
{
template
<
int
D
>
int64_t
operator
()(
const
Dim
<
D
>
&
dim
)
const
{
return
dim
[
idx_
];
}
private:
private:
int
idx_
;
};
...
...
@@ -182,7 +190,8 @@ struct VectorizeVisitor : Vistor<void> {
explicit
VectorizeVisitor
(
std
::
vector
<
int64_t
>
&
v
)
:
vector
(
v
)
{}
template
<
typename
T
>
void
operator
()(
const
T
&
t
)
{
template
<
typename
T
>
void
operator
()(
const
T
&
t
)
{
vector
.
push_back
(
t
.
head
);
this
->
operator
()(
t
.
tail
);
}
...
...
@@ -207,7 +216,8 @@ std::vector<int> vectorize2int(const DDim &ddim) {
}
struct
ProductVisitor
:
Vistor
<
int64_t
>
{
template
<
int
D
>
int64_t
operator
()(
const
Dim
<
D
>
&
dim
)
{
template
<
int
D
>
int64_t
operator
()(
const
Dim
<
D
>
&
dim
)
{
return
product
(
dim
);
}
};
...
...
@@ -233,7 +243,8 @@ struct SliceVectorizeVisitor : Vistor<void> {
// ddim slice.");
}
template
<
int
S
>
void
operator
()(
const
Dim
<
S
>
&
dim
)
{
template
<
int
S
>
void
operator
()(
const
Dim
<
S
>
&
dim
)
{
if
(
begin
==
0
)
{
vector
.
push_back
(
dim
.
head
);
}
else
{
...
...
@@ -264,7 +275,10 @@ DDim slice_ddim(const DDim &ddim, int begin, int end) {
/// \cond HIDDEN
struct
ArityVisitor
:
Vistor
<
int
>
{
template
<
int
D
>
int
operator
()(
Dim
<
D
>
)
const
{
return
D
;
}
template
<
int
D
>
int
operator
()(
Dim
<
D
>
)
const
{
return
D
;
}
};
/// \endcond
...
...
@@ -282,11 +296,12 @@ int arity(const DDim &d) {
struct
OSVistor
:
Vistor
<
std
::
ostream
&>
{
OSVistor
(
std
::
ostream
&
os
)
:
os_
(
os
)
{}
template
<
int
D
>
std
::
ostream
&
operator
()(
Dim
<
D
>
dim
)
const
{
template
<
int
D
>
std
::
ostream
&
operator
()(
Dim
<
D
>
dim
)
const
{
return
os_
<<
dim
;
}
private:
private:
std
::
ostream
&
os_
;
};
...
...
@@ -326,5 +341,5 @@ DDim stride_numel(const framework::DDim &ddim) {
return
framework
::
make_ddim
(
strides
);
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/ddim.h
浏览文件 @
d6825bf3
...
...
@@ -14,12 +14,12 @@ limitations under the License. */
#pragma once
#include "common/variant.h"
#include "dim.h"
#include <assert.h>
#include <initializer_list>
#include <stdexcept>
#include <vector>
#include "common/variant.h"
#include "dim.h"
namespace
paddle_mobile
{
namespace
framework
{
...
...
@@ -66,11 +66,15 @@ struct DDim {
DDim
()
{
var
.
Set
<
Dim
<
1
>>
(
Dim
<
1
>
());
}
template
<
int
D
>
explicit
DDim
(
const
Dim
<
D
>
&
in
)
{
var
.
Set
<
Dim
<
D
>>
(
in
);
}
template
<
int
D
>
explicit
DDim
(
const
Dim
<
D
>
&
in
)
{
var
.
Set
<
Dim
<
D
>>
(
in
);
}
/*implicit*/
DDim
(
std
::
initializer_list
<
int64_t
>
init_list
);
template
<
int
D
>
DDim
&
operator
=
(
const
Dim
<
D
>
&
in
)
{
template
<
int
D
>
DDim
&
operator
=
(
const
Dim
<
D
>
&
in
)
{
var
.
Set
<
Dim
<
D
>>
(
in
);
return
*
this
;
}
...
...
@@ -159,5 +163,5 @@ DDim flatten_to_1d(const DDim &src);
DDim
stride
(
const
DDim
&
ddim
);
DDim
stride_numel
(
const
DDim
&
ddim
);
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/dim.h
浏览文件 @
d6825bf3
...
...
@@ -24,7 +24,8 @@ namespace paddle_mobile {
namespace
framework
{
// Statically sized, statically indexed dimension
template
<
int
i
>
struct
Dim
{
template
<
int
i
>
struct
Dim
{
static
constexpr
int
dimensions
=
i
;
template
<
typename
...
Args
>
...
...
@@ -70,7 +71,8 @@ template <int i> struct Dim {
};
// Base case specialization
template
<
>
struct
Dim
<
0
>
{
template
<
>
struct
Dim
<
0
>
{
static
constexpr
int
dimensions
=
0
;
HOSTDEVICE
...
...
@@ -105,28 +107,37 @@ template <> struct Dim<0> {
namespace
{
// Helper for accessing Dim classes
template
<
int
i
>
struct
DimGetter
{
template
<
int
i
>
struct
DimGetter
{
// Return a copy if Dim is const
template
<
typename
D
>
HOSTDEVICE
static
int64_t
impl
(
const
D
&
d
)
{
template
<
typename
D
>
HOSTDEVICE
static
int64_t
impl
(
const
D
&
d
)
{
return
DimGetter
<
i
-
1
>::
impl
(
d
.
tail
);
}
// Return a reference if Dim is mutable
template
<
typename
D
>
HOSTDEVICE
static
int64_t
&
impl
(
D
&
d
)
{
template
<
typename
D
>
HOSTDEVICE
static
int64_t
&
impl
(
D
&
d
)
{
return
DimGetter
<
i
-
1
>::
impl
(
d
.
tail
);
}
};
// Eureka! We found the element!
template
<
>
struct
DimGetter
<
0
>
{
template
<
>
struct
DimGetter
<
0
>
{
// Return a copy if Dim is const
template
<
typename
D
>
HOSTDEVICE
static
int64_t
impl
(
const
D
&
d
)
{
template
<
typename
D
>
HOSTDEVICE
static
int64_t
impl
(
const
D
&
d
)
{
return
d
.
head
;
}
// Return a reference if Dim is mutable
template
<
typename
D
>
HOSTDEVICE
static
int64_t
&
impl
(
D
&
d
)
{
return
d
.
head
;
}
template
<
typename
D
>
HOSTDEVICE
static
int64_t
&
impl
(
D
&
d
)
{
return
d
.
head
;
}
};
template
<
int
D
>
HOSTDEVICE
int64_t
&
indexer
(
Dim
<
D
>
&
dim
,
int
idx
)
{
template
<
int
D
>
HOSTDEVICE
int64_t
&
indexer
(
Dim
<
D
>
&
dim
,
int
idx
)
{
#ifndef __CUDA_ARCH__
if
(
idx
<
0
)
{
throw
std
::
invalid_argument
(
"Tried to access a negative dimension"
);
...
...
@@ -140,7 +151,8 @@ template <int D> HOSTDEVICE int64_t &indexer(Dim<D> &dim, int idx) {
return
indexer
(
dim
.
tail
,
idx
-
1
);
}
template
<
>
HOSTDEVICE
int64_t
&
indexer
<
0
>
(
Dim
<
0
>
&
dim
,
int
idx
)
{
template
<
>
HOSTDEVICE
int64_t
&
indexer
<
0
>
(
Dim
<
0
>
&
dim
,
int
idx
)
{
#ifndef __CUDA_ARCH__
throw
std
::
invalid_argument
(
"Invalid index"
);
#else
...
...
@@ -156,7 +168,8 @@ template <> HOSTDEVICE int64_t &indexer<0>(Dim<0> &dim, int idx) {
#endif
}
template
<
int
D
>
HOSTDEVICE
int64_t
indexer
(
const
Dim
<
D
>
&
dim
,
int
idx
)
{
template
<
int
D
>
HOSTDEVICE
int64_t
indexer
(
const
Dim
<
D
>
&
dim
,
int
idx
)
{
#ifndef __CUDA_ARCH__
if
(
idx
<
0
)
{
throw
std
::
invalid_argument
(
"Tried to access a negative dimension"
);
...
...
@@ -170,7 +183,8 @@ template <int D> HOSTDEVICE int64_t indexer(const Dim<D> &dim, int idx) {
return
indexer
(
dim
.
tail
,
idx
-
1
);
}
template
<
>
HOSTDEVICE
int64_t
indexer
<
0
>
(
const
Dim
<
0
>
&
dim
,
int
idx
)
{
template
<
>
HOSTDEVICE
int64_t
indexer
<
0
>
(
const
Dim
<
0
>
&
dim
,
int
idx
)
{
#ifndef __CUDA_ARCH__
throw
std
::
invalid_argument
(
"Invalid index"
);
#else
...
...
@@ -186,25 +200,29 @@ template <> HOSTDEVICE int64_t indexer<0>(const Dim<0> &dim, int idx) {
#endif
}
}
// namespace
}
// namespace
// Static access to constant Dim
template
<
int
i
,
int
l
>
HOSTDEVICE
int64_t
get
(
const
Dim
<
l
>
&
d
)
{
template
<
int
i
,
int
l
>
HOSTDEVICE
int64_t
get
(
const
Dim
<
l
>
&
d
)
{
return
DimGetter
<
i
>::
impl
(
d
);
}
// Static access to mutable Dim
template
<
int
i
,
int
l
>
HOSTDEVICE
int64_t
&
get
(
Dim
<
l
>
&
d
)
{
template
<
int
i
,
int
l
>
HOSTDEVICE
int64_t
&
get
(
Dim
<
l
>
&
d
)
{
return
DimGetter
<
i
>::
impl
(
d
);
}
// Dynamic access to constant Dim
template
<
int
l
>
HOSTDEVICE
int64_t
Dim
<
l
>::
operator
[](
int
i
)
const
{
template
<
int
l
>
HOSTDEVICE
int64_t
Dim
<
l
>::
operator
[](
int
i
)
const
{
// std::cout << "l: " << l << std::endl;
return
indexer
(
*
this
,
i
);
}
// Dynamic access to mutable Dim
template
<
int
l
>
HOSTDEVICE
int64_t
&
Dim
<
l
>::
operator
[](
int
i
)
{
template
<
int
l
>
HOSTDEVICE
int64_t
&
Dim
<
l
>::
operator
[](
int
i
)
{
return
indexer
(
*
this
,
i
);
}
...
...
@@ -247,13 +265,15 @@ HOSTDEVICE inline int64_t linearize(const Dim<0> &a, const Dim<0> &b) {
}
// Product of a Dim
template
<
int
i
>
HOSTDEVICE
int64_t
product
(
const
Dim
<
i
>
&
a
,
int
prod
=
1
)
{
template
<
int
i
>
HOSTDEVICE
int64_t
product
(
const
Dim
<
i
>
&
a
,
int
prod
=
1
)
{
return
prod
*
a
.
head
*
product
(
a
.
tail
);
}
// Base case product of a Dim
// Notice it is inline because it is no longer a template
template
<
>
HOSTDEVICE
inline
int64_t
product
(
const
Dim
<
0
>
&
a
,
int
prod
)
{
template
<
>
HOSTDEVICE
inline
int64_t
product
(
const
Dim
<
0
>
&
a
,
int
prod
)
{
return
prod
;
}
...
...
@@ -282,7 +302,8 @@ HOSTDEVICE Dim<i> ex_prefix_mul(const Dim<i> &src, int mul = 1) {
///\cond HIDDEN
// Base case of ex_prefix_mul
// Notice it is inline because it is no longer a template
template
<
>
HOSTDEVICE
inline
Dim
<
0
>
ex_prefix_mul
(
const
Dim
<
0
>
&
src
,
int
mul
)
{
template
<
>
HOSTDEVICE
inline
Dim
<
0
>
ex_prefix_mul
(
const
Dim
<
0
>
&
src
,
int
mul
)
{
return
Dim
<
0
>
();
}
///\endcond
...
...
@@ -290,7 +311,8 @@ template <> HOSTDEVICE inline Dim<0> ex_prefix_mul(const Dim<0> &src, int mul) {
/**
* Add two dimensions together
*/
template
<
int
i
>
HOSTDEVICE
Dim
<
i
>
dim_plus
(
const
Dim
<
i
>
&
a
,
const
Dim
<
i
>
&
b
)
{
template
<
int
i
>
HOSTDEVICE
Dim
<
i
>
dim_plus
(
const
Dim
<
i
>
&
a
,
const
Dim
<
i
>
&
b
)
{
return
Dim
<
i
>
(
a
.
head
+
b
.
head
,
dim_plus
(
a
.
tail
,
b
.
tail
));
}
...
...
@@ -308,7 +330,8 @@ HOSTDEVICE Dim<i> operator+(const Dim<i> &lhs, const Dim<i> &rhs) {
/**
* Multiply two dimensions together
*/
template
<
int
i
>
HOSTDEVICE
Dim
<
i
>
dim_mult
(
const
Dim
<
i
>
&
a
,
const
Dim
<
i
>
&
b
)
{
template
<
int
i
>
HOSTDEVICE
Dim
<
i
>
dim_mult
(
const
Dim
<
i
>
&
a
,
const
Dim
<
i
>
&
b
)
{
return
Dim
<
i
>
(
a
.
head
*
b
.
head
,
dim_mult
(
a
.
tail
,
b
.
tail
));
}
...
...
@@ -365,8 +388,8 @@ HOSTDEVICE Dim<sizeof...(Args)> make_dim(Args... idxes) {
// Allows us to output a Dim
// XXX For some reason, overloading fails to resolve this correctly
template
<
int
i
>
typename
std
::
enable_if
<
(
i
>
1
),
std
::
ostream
&>::
type
operator
<<
(
std
::
ostream
&
os
,
const
Dim
<
i
>
&
d
)
{
typename
std
::
enable_if
<
(
i
>
1
),
std
::
ostream
&>::
type
operator
<<
(
std
::
ostream
&
os
,
const
Dim
<
i
>
&
d
)
{
os
<<
d
.
head
<<
", "
<<
d
.
tail
;
return
os
;
}
...
...
@@ -374,8 +397,8 @@ operator<<(std::ostream &os, const Dim<i> &d) {
// Base case that allows us to output a Dim
// XXX I wish this could be an overload instead of a template
template
<
int
i
>
typename
std
::
enable_if
<
(
i
==
1
),
std
::
ostream
&>::
type
operator
<<
(
std
::
ostream
&
os
,
const
Dim
<
i
>
&
d
)
{
typename
std
::
enable_if
<
(
i
==
1
),
std
::
ostream
&>::
type
operator
<<
(
std
::
ostream
&
os
,
const
Dim
<
i
>
&
d
)
{
os
<<
d
.
head
;
return
os
;
}
...
...
@@ -384,7 +407,8 @@ inline std::ostream &operator<<(std::ostream &os, const Dim<0> &d) {
return
os
;
}
template
<
int
i
>
HOST
std
::
string
Dim
<
i
>::
to_string
()
const
{
template
<
int
i
>
HOST
std
::
string
Dim
<
i
>::
to_string
()
const
{
std
::
stringstream
stream
;
stream
<<
*
this
;
...
...
@@ -406,5 +430,5 @@ HOSTDEVICE Dim<D> linear_to_dimension(int linear_index, Dim<D> extents) {
return
result
;
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/executor.cpp
浏览文件 @
d6825bf3
...
...
@@ -90,5 +90,5 @@ void Executor<Dtype>::predict(const Tensor &t, int block_id) {
template
class
Executor
<
CPU
>;
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/executor.h
浏览文件 @
d6825bf3
...
...
@@ -34,15 +34,16 @@ SOFTWARE.
namespace
paddle_mobile
{
namespace
framework
{
template
<
typename
Dtype
>
class
Executor
{
public:
template
<
typename
Dtype
>
class
Executor
{
public:
Executor
();
Executor
(
const
Program
<
Dtype
>
p
);
std
::
shared_ptr
<
Tensor
>
predict
(
Tensor
&
t
);
public:
public:
const
framework
::
Program
<
Dtype
>
program_
;
std
::
shared_ptr
<
ProgramDesc
>
to_predict_program_
;
...
...
@@ -54,5 +55,5 @@ public:
bool
use_optimize_
=
false
;
};
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/framework.pb.cpp
浏览文件 @
d6825bf3
...
...
@@ -18,73 +18,73 @@ namespace paddle_mobile {
namespace
framework
{
namespace
proto
{
class
OpDesc_AttrDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
OpDesc_Attr
>
_instance
;
}
_OpDesc_Attr_default_instance_
;
class
OpDesc_VarDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
OpDesc_Var
>
_instance
;
}
_OpDesc_Var_default_instance_
;
class
OpDescDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
OpDesc
>
_instance
;
}
_OpDesc_default_instance_
;
class
OpProto_VarDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
OpProto_Var
>
_instance
;
}
_OpProto_Var_default_instance_
;
class
OpProto_AttrDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
OpProto_Attr
>
_instance
;
}
_OpProto_Attr_default_instance_
;
class
OpProtoDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
OpProto
>
_instance
;
}
_OpProto_default_instance_
;
class
VarType_TensorDescDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
VarType_TensorDesc
>
_instance
;
}
_VarType_TensorDesc_default_instance_
;
class
VarType_LoDTensorDescDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
VarType_LoDTensorDesc
>
_instance
;
}
_VarType_LoDTensorDesc_default_instance_
;
class
VarType_LoDTensorArrayDescDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
VarType_LoDTensorArrayDesc
>
_instance
;
}
_VarType_LoDTensorArrayDesc_default_instance_
;
class
VarType_ReaderDescDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
VarType_ReaderDesc
>
_instance
;
}
_VarType_ReaderDesc_default_instance_
;
class
VarType_ChannelDescDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
VarType_ChannelDesc
>
_instance
;
}
_VarType_ChannelDesc_default_instance_
;
class
VarType_TupleDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
VarType_Tuple
>
_instance
;
}
_VarType_Tuple_default_instance_
;
class
VarTypeDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
VarType
>
_instance
;
}
_VarType_default_instance_
;
class
VarDescDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
VarDesc
>
_instance
;
}
_VarDesc_default_instance_
;
class
BlockDescDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
BlockDesc
>
_instance
;
}
_BlockDesc_default_instance_
;
class
ProgramDescDefaultTypeInternal
{
public:
public:
::
google
::
protobuf
::
internal
::
ExplicitlyConstructed
<
ProgramDesc
>
_instance
;
}
_ProgramDesc_default_instance_
;
...
...
@@ -217,39 +217,39 @@ void InitDefaults() {
}
namespace
{
void
AddDescriptorsImpl
()
{
InitDefaults
();
}
}
// anonymous namespace
}
// anonymous namespace
void
AddDescriptors
()
{
static
GOOGLE_PROTOBUF_DECLARE_ONCE
(
once
);
::
google
::
protobuf
::
GoogleOnceInit
(
&
once
,
&
AddDescriptorsImpl
);
}
}
// namespace protobuf_framework_2eproto
}
// namespace protobuf_framework_2eproto
bool
VarType_Type_IsValid
(
int
value
)
{
switch
(
value
)
{
case
0
:
case
1
:
case
2
:
case
3
:
case
4
:
case
5
:
case
6
:
case
7
:
case
8
:
case
9
:
case
10
:
case
11
:
case
12
:
case
13
:
case
14
:
case
15
:
case
16
:
case
17
:
case
18
:
return
true
;
default:
return
false
;
case
0
:
case
1
:
case
2
:
case
3
:
case
4
:
case
5
:
case
6
:
case
7
:
case
8
:
case
9
:
case
10
:
case
11
:
case
12
:
case
13
:
case
14
:
case
15
:
case
16
:
case
17
:
case
18
:
return
true
;
default:
return
false
;
}
}
...
...
@@ -276,22 +276,22 @@ const VarType_Type VarType::TUPLE;
const
VarType_Type
VarType
::
Type_MIN
;
const
VarType_Type
VarType
::
Type_MAX
;
const
int
VarType
::
Type_ARRAYSIZE
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
bool
AttrType_IsValid
(
int
value
)
{
switch
(
value
)
{
case
0
:
case
1
:
case
2
:
case
3
:
case
4
:
case
5
:
case
6
:
case
7
:
case
8
:
case
9
:
return
true
;
default:
return
false
;
case
0
:
case
1
:
case
2
:
case
3
:
case
4
:
case
5
:
case
6
:
case
7
:
case
8
:
case
9
:
return
true
;
default:
return
false
;
}
}
...
...
@@ -310,7 +310,7 @@ const int OpDesc_Attr::kBFieldNumber;
const
int
OpDesc_Attr
::
kBoolsFieldNumber
;
const
int
OpDesc_Attr
::
kBlockIdxFieldNumber
;
const
int
OpDesc_Attr
::
kLFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
OpDesc_Attr
::
OpDesc_Attr
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -321,9 +321,14 @@ OpDesc_Attr::OpDesc_Attr()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.OpDesc.Attr)
}
OpDesc_Attr
::
OpDesc_Attr
(
const
OpDesc_Attr
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
ints_
(
from
.
ints_
),
floats_
(
from
.
floats_
),
strings_
(
from
.
strings_
),
bools_
(
from
.
bools_
)
{
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
ints_
(
from
.
ints_
),
floats_
(
from
.
floats_
),
strings_
(
from
.
strings_
),
bools_
(
from
.
bools_
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
name_
.
UnsafeSetDefault
(
&::
google
::
protobuf
::
internal
::
GetEmptyStringAlreadyInited
());
...
...
@@ -422,9 +427,8 @@ void OpDesc_Attr::Clear() {
bool
OpDesc_Attr
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -438,216 +442,223 @@ bool OpDesc_Attr::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// required string name = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_name
()));
}
else
{
goto
handle_unusual
;
// required string name = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_name
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// required .paddle_mobile.framework.proto.AttrType type =
// 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
16u
/* 16 & 0xFF */
))
{
int
value
;
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
int
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_ENUM
>
(
input
,
&
value
)));
if
(
::
paddle_mobile
::
framework
::
proto
::
AttrType_IsValid
(
value
))
{
set_type
(
static_cast
<::
paddle_mobile
::
framework
::
proto
::
AttrType
>
(
value
));
// required .paddle_mobile.framework.proto.AttrType type =
// 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
16u
/* 16 & 0xFF */
))
{
int
value
;
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
int
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_ENUM
>
(
input
,
&
value
)));
if
(
::
paddle_mobile
::
framework
::
proto
::
AttrType_IsValid
(
value
))
{
set_type
(
static_cast
<::
paddle_mobile
::
framework
::
proto
::
AttrType
>
(
value
));
}
else
{
unknown_fields_stream
.
WriteVarint32
(
16u
);
unknown_fields_stream
.
WriteVarint32
(
static_cast
<::
google
::
protobuf
::
uint32
>
(
value
));
}
}
else
{
unknown_fields_stream
.
WriteVarint32
(
16u
);
unknown_fields_stream
.
WriteVarint32
(
static_cast
<::
google
::
protobuf
::
uint32
>
(
value
));
goto
handle_unusual
;
}
}
else
{
goto
handle_unusual
;
break
;
}
break
;
}
// optional int32 i = 3;
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
24u
/* 24 & 0xFF */
))
{
set_has_i
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
&
i_
)));
}
else
{
goto
handle_unusual
;
// optional int32 i = 3;
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
24u
/* 24 & 0xFF */
))
{
set_has_i
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
&
i_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional float f = 4;
case
4
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
37u
/* 37 & 0xFF */
))
{
set_has_f
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
float
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_FLOAT
>
(
input
,
&
f_
)));
}
else
{
goto
handle_unusual
;
// optional float f = 4;
case
4
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
37u
/* 37 & 0xFF */
))
{
set_has_f
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
float
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_FLOAT
>
(
input
,
&
f_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional string s = 5;
case
5
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
42u
/* 42 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_s
()));
}
else
{
goto
handle_unusual
;
// optional string s = 5;
case
5
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
42u
/* 42 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_s
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// repeated int32 ints = 6;
case
6
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
48u
/* 48 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadRepeatedPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
1
,
48u
,
input
,
this
->
mutable_ints
())));
}
else
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
50u
/* 50 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPackedPrimitiveNoInline
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
this
->
mutable_ints
())));
}
else
{
goto
handle_unusual
;
// repeated int32 ints = 6;
case
6
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
48u
/* 48 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadRepeatedPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
1
,
48u
,
input
,
this
->
mutable_ints
())));
}
else
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
50u
/* 50 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPackedPrimitiveNoInline
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
this
->
mutable_ints
())));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// repeated float floats = 7;
case
7
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
61u
/* 61 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadRepeatedPrimitive
<
float
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_FLOAT
>
(
1
,
61u
,
input
,
this
->
mutable_floats
())));
}
else
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
58u
/* 58 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPackedPrimitiveNoInline
<
float
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_FLOAT
>
(
input
,
this
->
mutable_floats
())));
}
else
{
goto
handle_unusual
;
// repeated float floats = 7;
case
7
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
61u
/* 61 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadRepeatedPrimitive
<
float
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_FLOAT
>
(
1
,
61u
,
input
,
this
->
mutable_floats
())));
}
else
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
58u
/* 58 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPackedPrimitiveNoInline
<
float
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_FLOAT
>
(
input
,
this
->
mutable_floats
())));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// repeated string strings = 8;
case
8
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
66u
/* 66 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
add_strings
()));
}
else
{
goto
handle_unusual
;
// repeated string strings = 8;
case
8
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
66u
/* 66 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
add_strings
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional bool b = 10;
case
10
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
80u
/* 80 & 0xFF */
))
{
set_has_b
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
&
b_
)));
}
else
{
goto
handle_unusual
;
// optional bool b = 10;
case
10
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
80u
/* 80 & 0xFF */
))
{
set_has_b
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
&
b_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// repeated bool bools = 11;
case
11
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
88u
/* 88 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadRepeatedPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
1
,
88u
,
input
,
this
->
mutable_bools
())));
}
else
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
90u
/* 90 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPackedPrimitiveNoInline
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
this
->
mutable_bools
())));
}
else
{
goto
handle_unusual
;
// repeated bool bools = 11;
case
11
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
88u
/* 88 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadRepeatedPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
1
,
88u
,
input
,
this
->
mutable_bools
())));
}
else
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
90u
/* 90 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPackedPrimitiveNoInline
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
this
->
mutable_bools
())));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional int32 block_idx = 12;
case
12
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
96u
/* 96 & 0xFF */
))
{
set_has_block_idx
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
&
block_idx_
)));
}
else
{
goto
handle_unusual
;
// optional int32 block_idx = 12;
case
12
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
96u
/* 96 & 0xFF */
))
{
set_has_block_idx
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
&
block_idx_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional int64 l = 13;
case
13
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
104u
/* 104 & 0xFF */
))
{
set_has_l
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int64
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT64
>
(
input
,
&
l_
)));
}
else
{
goto
handle_unusual
;
// optional int64 l = 13;
case
13
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
104u
/* 104 & 0xFF */
))
{
set_has_l
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int64
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT64
>
(
input
,
&
l_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -770,7 +781,7 @@ size_t OpDesc_Attr::ByteSizeLong() const {
total_size
+=
_internal_metadata_
.
unknown_fields
().
size
();
if
(((
_has_bits_
[
0
]
&
0x00000005
)
^
0x00000005
)
==
0
)
{
// All required fields are present.
0
)
{
// All required fields are present.
// required string name = 1;
total_size
+=
1
+
::
google
::
protobuf
::
internal
::
WireFormatLite
::
StringSize
(
this
->
name
());
...
...
@@ -914,21 +925,18 @@ void OpDesc_Attr::MergeFrom(const OpDesc_Attr &from) {
void
OpDesc_Attr
::
CopyFrom
(
const
OpDesc_Attr
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.OpDesc.Attr)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
bool
OpDesc_Attr
::
IsInitialized
()
const
{
if
((
_has_bits_
[
0
]
&
0x00000005
)
!=
0x00000005
)
return
false
;
if
((
_has_bits_
[
0
]
&
0x00000005
)
!=
0x00000005
)
return
false
;
return
true
;
}
void
OpDesc_Attr
::
Swap
(
OpDesc_Attr
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
OpDesc_Attr
::
InternalSwap
(
OpDesc_Attr
*
other
)
{
...
...
@@ -1157,13 +1165,13 @@ void OpDesc_Attr::add_ints(::google::protobuf::int32 value) {
ints_
.
Add
(
value
);
// @@protoc_insertion_point(field_add:paddle_mobile.framework.proto.OpDesc.Attr.ints)
}
const
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int32
>
&
OpDesc_Attr
::
ints
()
const
{
const
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int32
>
&
OpDesc_Attr
::
ints
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpDesc.Attr.ints)
return
ints_
;
}
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int32
>
*
OpDesc_Attr
::
mutable_ints
()
{
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int32
>
*
OpDesc_Attr
::
mutable_ints
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpDesc.Attr.ints)
return
&
ints_
;
}
...
...
@@ -1245,13 +1253,13 @@ void OpDesc_Attr::add_strings(const char *value, size_t size) {
strings_
.
Add
()
->
assign
(
reinterpret_cast
<
const
char
*>
(
value
),
size
);
// @@protoc_insertion_point(field_add_pointer:paddle_mobile.framework.proto.OpDesc.Attr.strings)
}
const
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
&
OpDesc_Attr
::
strings
()
const
{
const
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
&
OpDesc_Attr
::
strings
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpDesc.Attr.strings)
return
strings_
;
}
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
*
OpDesc_Attr
::
mutable_strings
()
{
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
*
OpDesc_Attr
::
mutable_strings
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpDesc.Attr.strings)
return
&
strings_
;
}
...
...
@@ -1336,14 +1344,14 @@ void OpDesc_Attr::set_l(::google::protobuf::int64 value) {
// @@protoc_insertion_point(field_set:paddle_mobile.framework.proto.OpDesc.Attr.l)
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
#if !defined(_MSC_VER) || _MSC_VER >= 1900
const
int
OpDesc_Var
::
kParameterFieldNumber
;
const
int
OpDesc_Var
::
kArgumentsFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
OpDesc_Var
::
OpDesc_Var
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -1354,8 +1362,10 @@ OpDesc_Var::OpDesc_Var()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.OpDesc.Var)
}
OpDesc_Var
::
OpDesc_Var
(
const
OpDesc_Var
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
arguments_
(
from
.
arguments_
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
parameter_
.
UnsafeSetDefault
(
...
...
@@ -1420,9 +1430,8 @@ void OpDesc_Var::Clear() {
bool
OpDesc_Var
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -1436,43 +1445,42 @@ bool OpDesc_Var::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// required string parameter = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_parameter
()));
}
else
{
goto
handle_unusual
;
// required string parameter = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_parameter
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// repeated string arguments = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
18u
/* 18 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
add_arguments
()));
}
else
{
goto
handle_unusual
;
// repeated string arguments = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
18u
/* 18 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
add_arguments
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -1558,21 +1566,18 @@ void OpDesc_Var::MergeFrom(const OpDesc_Var &from) {
void
OpDesc_Var
::
CopyFrom
(
const
OpDesc_Var
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.OpDesc.Var)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
bool
OpDesc_Var
::
IsInitialized
()
const
{
if
((
_has_bits_
[
0
]
&
0x00000001
)
!=
0x00000001
)
return
false
;
if
((
_has_bits_
[
0
]
&
0x00000001
)
!=
0x00000001
)
return
false
;
return
true
;
}
void
OpDesc_Var
::
Swap
(
OpDesc_Var
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
OpDesc_Var
::
InternalSwap
(
OpDesc_Var
*
other
)
{
...
...
@@ -1713,18 +1718,18 @@ void OpDesc_Var::add_arguments(const char *value, size_t size) {
arguments_
.
Add
()
->
assign
(
reinterpret_cast
<
const
char
*>
(
value
),
size
);
// @@protoc_insertion_point(field_add_pointer:paddle_mobile.framework.proto.OpDesc.Var.arguments)
}
const
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
&
OpDesc_Var
::
arguments
()
const
{
const
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
&
OpDesc_Var
::
arguments
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpDesc.Var.arguments)
return
arguments_
;
}
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
*
OpDesc_Var
::
mutable_arguments
()
{
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
*
OpDesc_Var
::
mutable_arguments
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpDesc.Var.arguments)
return
&
arguments_
;
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
...
...
@@ -1734,7 +1739,7 @@ const int OpDesc::kInputsFieldNumber;
const
int
OpDesc
::
kOutputsFieldNumber
;
const
int
OpDesc
::
kAttrsFieldNumber
;
const
int
OpDesc
::
kIsTargetFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
OpDesc
::
OpDesc
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -1745,9 +1750,13 @@ OpDesc::OpDesc()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.OpDesc)
}
OpDesc
::
OpDesc
(
const
OpDesc
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
inputs_
(
from
.
inputs_
),
outputs_
(
from
.
outputs_
),
attrs_
(
from
.
attrs_
)
{
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
inputs_
(
from
.
inputs_
),
outputs_
(
from
.
outputs_
),
attrs_
(
from
.
attrs_
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
type_
.
UnsafeSetDefault
(
&::
google
::
protobuf
::
internal
::
GetEmptyStringAlreadyInited
());
...
...
@@ -1816,9 +1825,8 @@ void OpDesc::Clear() {
bool
OpDesc
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -1832,84 +1840,83 @@ bool OpDesc::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// repeated .paddle_mobile.framework.proto.OpDesc.Var inputs
// = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_inputs
()));
}
else
{
goto
handle_unusual
;
// repeated .paddle_mobile.framework.proto.OpDesc.Var inputs
// = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_inputs
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// repeated .paddle_mobile.framework.proto.OpDesc.Var
// outputs = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
18u
/* 18 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_outputs
()));
}
else
{
goto
handle_unusual
;
// repeated .paddle_mobile.framework.proto.OpDesc.Var
// outputs = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
18u
/* 18 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_outputs
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// required string type = 3;
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
26u
/* 26 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_type
()));
}
else
{
goto
handle_unusual
;
// required string type = 3;
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
26u
/* 26 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_type
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// repeated .paddle_mobile.framework.proto.OpDesc.Attr attrs
// = 4;
case
4
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
34u
/* 34 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_attrs
()));
}
else
{
goto
handle_unusual
;
// repeated .paddle_mobile.framework.proto.OpDesc.Attr attrs
// = 4;
case
4
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
34u
/* 34 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_attrs
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional bool is_target = 5 [default = false];
case
5
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
40u
/* 40 & 0xFF */
))
{
set_has_is_target
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
&
is_target_
)));
}
else
{
goto
handle_unusual
;
// optional bool is_target = 5 [default = false];
case
5
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
40u
/* 40 & 0xFF */
))
{
set_has_is_target
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
&
is_target_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -2061,15 +2068,13 @@ void OpDesc::MergeFrom(const OpDesc &from) {
void
OpDesc
::
CopyFrom
(
const
OpDesc
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.OpDesc)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
bool
OpDesc
::
IsInitialized
()
const
{
if
((
_has_bits_
[
0
]
&
0x00000001
)
!=
0x00000001
)
return
false
;
if
((
_has_bits_
[
0
]
&
0x00000001
)
!=
0x00000001
)
return
false
;
if
(
!::
google
::
protobuf
::
internal
::
AllAreInitialized
(
this
->
inputs
()))
return
false
;
if
(
!::
google
::
protobuf
::
internal
::
AllAreInitialized
(
this
->
outputs
()))
...
...
@@ -2080,8 +2085,7 @@ bool OpDesc::IsInitialized() const {
}
void
OpDesc
::
Swap
(
OpDesc
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
OpDesc
::
InternalSwap
(
OpDesc
*
other
)
{
...
...
@@ -2169,13 +2173,13 @@ void OpDesc::set_allocated_type(::std::string *type) {
// repeated .paddle_mobile.framework.proto.OpDesc.Var inputs = 1;
int
OpDesc
::
inputs_size
()
const
{
return
inputs_
.
size
();
}
void
OpDesc
::
clear_inputs
()
{
inputs_
.
Clear
();
}
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
&
OpDesc
::
inputs
(
int
index
)
const
{
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
&
OpDesc
::
inputs
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.OpDesc.inputs)
return
inputs_
.
Get
(
index
);
}
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
*
OpDesc
::
mutable_inputs
(
int
index
)
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
*
OpDesc
::
mutable_inputs
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.OpDesc.inputs)
return
inputs_
.
Mutable
(
index
);
}
...
...
@@ -2184,14 +2188,14 @@ OpDesc::mutable_inputs(int index) {
return
inputs_
.
Add
();
}
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
*
OpDesc
::
mutable_inputs
()
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
*
OpDesc
::
mutable_inputs
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpDesc.inputs)
return
&
inputs_
;
}
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
&
OpDesc
::
inputs
()
const
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
&
OpDesc
::
inputs
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpDesc.inputs)
return
inputs_
;
}
...
...
@@ -2199,13 +2203,13 @@ OpDesc::inputs() const {
// repeated .paddle_mobile.framework.proto.OpDesc.Var outputs = 2;
int
OpDesc
::
outputs_size
()
const
{
return
outputs_
.
size
();
}
void
OpDesc
::
clear_outputs
()
{
outputs_
.
Clear
();
}
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
&
OpDesc
::
outputs
(
int
index
)
const
{
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
&
OpDesc
::
outputs
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.OpDesc.outputs)
return
outputs_
.
Get
(
index
);
}
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
*
OpDesc
::
mutable_outputs
(
int
index
)
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
*
OpDesc
::
mutable_outputs
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.OpDesc.outputs)
return
outputs_
.
Mutable
(
index
);
}
...
...
@@ -2214,14 +2218,14 @@ OpDesc::mutable_outputs(int index) {
return
outputs_
.
Add
();
}
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
*
OpDesc
::
mutable_outputs
()
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
*
OpDesc
::
mutable_outputs
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpDesc.outputs)
return
&
outputs_
;
}
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
&
OpDesc
::
outputs
()
const
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
&
OpDesc
::
outputs
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpDesc.outputs)
return
outputs_
;
}
...
...
@@ -2229,13 +2233,13 @@ OpDesc::outputs() const {
// repeated .paddle_mobile.framework.proto.OpDesc.Attr attrs = 4;
int
OpDesc
::
attrs_size
()
const
{
return
attrs_
.
size
();
}
void
OpDesc
::
clear_attrs
()
{
attrs_
.
Clear
();
}
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
&
OpDesc
::
attrs
(
int
index
)
const
{
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
&
OpDesc
::
attrs
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.OpDesc.attrs)
return
attrs_
.
Get
(
index
);
}
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
*
OpDesc
::
mutable_attrs
(
int
index
)
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
*
OpDesc
::
mutable_attrs
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.OpDesc.attrs)
return
attrs_
.
Mutable
(
index
);
}
...
...
@@ -2244,14 +2248,14 @@ OpDesc::mutable_attrs(int index) {
return
attrs_
.
Add
();
}
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
>
*
OpDesc
::
mutable_attrs
()
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
>
*
OpDesc
::
mutable_attrs
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpDesc.attrs)
return
&
attrs_
;
}
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
>
&
OpDesc
::
attrs
()
const
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
>
&
OpDesc
::
attrs
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpDesc.attrs)
return
attrs_
;
}
...
...
@@ -2276,7 +2280,7 @@ void OpDesc::set_is_target(bool value) {
// @@protoc_insertion_point(field_set:paddle_mobile.framework.proto.OpDesc.is_target)
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
...
...
@@ -2286,7 +2290,7 @@ const int OpProto_Var::kCommentFieldNumber;
const
int
OpProto_Var
::
kDuplicableFieldNumber
;
const
int
OpProto_Var
::
kIntermediateFieldNumber
;
const
int
OpProto_Var
::
kDispensableFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
OpProto_Var
::
OpProto_Var
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -2297,8 +2301,10 @@ OpProto_Var::OpProto_Var()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.OpProto.Var)
}
OpProto_Var
::
OpProto_Var
(
const
OpProto_Var
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
)
{
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
name_
.
UnsafeSetDefault
(
&::
google
::
protobuf
::
internal
::
GetEmptyStringAlreadyInited
());
...
...
@@ -2394,9 +2400,8 @@ void OpProto_Var::Clear() {
bool
OpProto_Var
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -2410,85 +2415,84 @@ bool OpProto_Var::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// required string name = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_name
()));
}
else
{
goto
handle_unusual
;
// required string name = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_name
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// required string comment = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
18u
/* 18 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_comment
()));
}
else
{
goto
handle_unusual
;
// required string comment = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
18u
/* 18 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_comment
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional bool duplicable = 3 [default = false];
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
24u
/* 24 & 0xFF */
))
{
set_has_duplicable
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
&
duplicable_
)));
}
else
{
goto
handle_unusual
;
// optional bool duplicable = 3 [default = false];
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
24u
/* 24 & 0xFF */
))
{
set_has_duplicable
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
&
duplicable_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional bool intermediate = 4 [default = false];
case
4
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
32u
/* 32 & 0xFF */
))
{
set_has_intermediate
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
&
intermediate_
)));
}
else
{
goto
handle_unusual
;
// optional bool intermediate = 4 [default = false];
case
4
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
32u
/* 32 & 0xFF */
))
{
set_has_intermediate
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
&
intermediate_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional bool dispensable = 5 [default = false];
case
5
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
40u
/* 40 & 0xFF */
))
{
set_has_dispensable
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
&
dispensable_
)));
}
else
{
goto
handle_unusual
;
// optional bool dispensable = 5 [default = false];
case
5
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
40u
/* 40 & 0xFF */
))
{
set_has_dispensable
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
&
dispensable_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -2568,7 +2572,7 @@ size_t OpProto_Var::ByteSizeLong() const {
total_size
+=
_internal_metadata_
.
unknown_fields
().
size
();
if
(((
_has_bits_
[
0
]
&
0x00000003
)
^
0x00000003
)
==
0
)
{
// All required fields are present.
0
)
{
// All required fields are present.
// required string name = 1;
total_size
+=
1
+
::
google
::
protobuf
::
internal
::
WireFormatLite
::
StringSize
(
this
->
name
());
...
...
@@ -2644,21 +2648,18 @@ void OpProto_Var::MergeFrom(const OpProto_Var &from) {
void
OpProto_Var
::
CopyFrom
(
const
OpProto_Var
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.OpProto.Var)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
bool
OpProto_Var
::
IsInitialized
()
const
{
if
((
_has_bits_
[
0
]
&
0x00000003
)
!=
0x00000003
)
return
false
;
if
((
_has_bits_
[
0
]
&
0x00000003
)
!=
0x00000003
)
return
false
;
return
true
;
}
void
OpProto_Var
::
Swap
(
OpProto_Var
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
OpProto_Var
::
InternalSwap
(
OpProto_Var
*
other
)
{
...
...
@@ -2873,7 +2874,7 @@ void OpProto_Var::set_dispensable(bool value) {
// @@protoc_insertion_point(field_set:paddle_mobile.framework.proto.OpProto.Var.dispensable)
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
...
...
@@ -2882,7 +2883,7 @@ const int OpProto_Attr::kNameFieldNumber;
const
int
OpProto_Attr
::
kTypeFieldNumber
;
const
int
OpProto_Attr
::
kCommentFieldNumber
;
const
int
OpProto_Attr
::
kGeneratedFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
OpProto_Attr
::
OpProto_Attr
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -2893,8 +2894,10 @@ OpProto_Attr::OpProto_Attr()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.OpProto.Attr)
}
OpProto_Attr
::
OpProto_Attr
(
const
OpProto_Attr
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
)
{
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
name_
.
UnsafeSetDefault
(
&::
google
::
protobuf
::
internal
::
GetEmptyStringAlreadyInited
());
...
...
@@ -2990,9 +2993,8 @@ void OpProto_Attr::Clear() {
bool
OpProto_Attr
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -3006,80 +3008,79 @@ bool OpProto_Attr::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// required string name = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_name
()));
}
else
{
goto
handle_unusual
;
// required string name = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_name
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// required .paddle_mobile.framework.proto.AttrType type =
// 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
16u
/* 16 & 0xFF */
))
{
int
value
;
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
int
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_ENUM
>
(
input
,
&
value
)));
if
(
::
paddle_mobile
::
framework
::
proto
::
AttrType_IsValid
(
value
))
{
set_type
(
static_cast
<::
paddle_mobile
::
framework
::
proto
::
AttrType
>
(
value
));
// required .paddle_mobile.framework.proto.AttrType type =
// 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
16u
/* 16 & 0xFF */
))
{
int
value
;
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
int
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_ENUM
>
(
input
,
&
value
)));
if
(
::
paddle_mobile
::
framework
::
proto
::
AttrType_IsValid
(
value
))
{
set_type
(
static_cast
<::
paddle_mobile
::
framework
::
proto
::
AttrType
>
(
value
));
}
else
{
unknown_fields_stream
.
WriteVarint32
(
16u
);
unknown_fields_stream
.
WriteVarint32
(
static_cast
<::
google
::
protobuf
::
uint32
>
(
value
));
}
}
else
{
unknown_fields_stream
.
WriteVarint32
(
16u
);
unknown_fields_stream
.
WriteVarint32
(
static_cast
<::
google
::
protobuf
::
uint32
>
(
value
));
goto
handle_unusual
;
}
}
else
{
goto
handle_unusual
;
break
;
}
break
;
}
// required string comment = 3;
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
26u
/* 26 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_comment
()));
}
else
{
goto
handle_unusual
;
// required string comment = 3;
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
26u
/* 26 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_comment
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional bool generated = 4 [default = false];
case
4
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
32u
/* 32 & 0xFF */
))
{
set_has_generated
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
&
generated_
)));
}
else
{
goto
handle_unusual
;
// optional bool generated = 4 [default = false];
case
4
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
32u
/* 32 & 0xFF */
))
{
set_has_generated
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
&
generated_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -3160,7 +3161,7 @@ size_t OpProto_Attr::ByteSizeLong() const {
total_size
+=
_internal_metadata_
.
unknown_fields
().
size
();
if
(((
_has_bits_
[
0
]
&
0x00000007
)
^
0x00000007
)
==
0
)
{
// All required fields are present.
0
)
{
// All required fields are present.
// required string name = 1;
total_size
+=
1
+
::
google
::
protobuf
::
internal
::
WireFormatLite
::
StringSize
(
this
->
name
());
...
...
@@ -3227,21 +3228,18 @@ void OpProto_Attr::MergeFrom(const OpProto_Attr &from) {
void
OpProto_Attr
::
CopyFrom
(
const
OpProto_Attr
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.OpProto.Attr)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
bool
OpProto_Attr
::
IsInitialized
()
const
{
if
((
_has_bits_
[
0
]
&
0x00000007
)
!=
0x00000007
)
return
false
;
if
((
_has_bits_
[
0
]
&
0x00000007
)
!=
0x00000007
)
return
false
;
return
true
;
}
void
OpProto_Attr
::
Swap
(
OpProto_Attr
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
OpProto_Attr
::
InternalSwap
(
OpProto_Attr
*
other
)
{
...
...
@@ -3436,7 +3434,7 @@ void OpProto_Attr::set_generated(bool value) {
// @@protoc_insertion_point(field_set:paddle_mobile.framework.proto.OpProto.Attr.generated)
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
...
...
@@ -3446,7 +3444,7 @@ const int OpProto::kInputsFieldNumber;
const
int
OpProto
::
kOutputsFieldNumber
;
const
int
OpProto
::
kAttrsFieldNumber
;
const
int
OpProto
::
kCommentFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
OpProto
::
OpProto
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -3457,9 +3455,13 @@ OpProto::OpProto()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.OpProto)
}
OpProto
::
OpProto
(
const
OpProto
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
inputs_
(
from
.
inputs_
),
outputs_
(
from
.
outputs_
),
attrs_
(
from
.
attrs_
)
{
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
inputs_
(
from
.
inputs_
),
outputs_
(
from
.
outputs_
),
attrs_
(
from
.
attrs_
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
type_
.
UnsafeSetDefault
(
&::
google
::
protobuf
::
internal
::
GetEmptyStringAlreadyInited
());
...
...
@@ -3544,9 +3546,8 @@ void OpProto::Clear() {
bool
OpProto
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -3560,82 +3561,81 @@ bool OpProto::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// required string type = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_type
()));
}
else
{
goto
handle_unusual
;
// required string type = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_type
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// repeated .paddle_mobile.framework.proto.OpProto.Var
// inputs = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
18u
/* 18 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_inputs
()));
}
else
{
goto
handle_unusual
;
// repeated .paddle_mobile.framework.proto.OpProto.Var
// inputs = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
18u
/* 18 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_inputs
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// repeated .paddle_mobile.framework.proto.OpProto.Var
// outputs = 3;
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
26u
/* 26 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_outputs
()));
}
else
{
goto
handle_unusual
;
// repeated .paddle_mobile.framework.proto.OpProto.Var
// outputs = 3;
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
26u
/* 26 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_outputs
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// repeated .paddle_mobile.framework.proto.OpProto.Attr
// attrs = 4;
case
4
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
34u
/* 34 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_attrs
()));
}
else
{
goto
handle_unusual
;
// repeated .paddle_mobile.framework.proto.OpProto.Attr
// attrs = 4;
case
4
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
34u
/* 34 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_attrs
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// required string comment = 5;
case
5
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
42u
/* 42 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_comment
()));
}
else
{
goto
handle_unusual
;
// required string comment = 5;
case
5
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
42u
/* 42 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_comment
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -3721,7 +3721,7 @@ size_t OpProto::ByteSizeLong() const {
total_size
+=
_internal_metadata_
.
unknown_fields
().
size
();
if
(((
_has_bits_
[
0
]
&
0x00000003
)
^
0x00000003
)
==
0
)
{
// All required fields are present.
0
)
{
// All required fields are present.
// required string type = 1;
total_size
+=
1
+
::
google
::
protobuf
::
internal
::
WireFormatLite
::
StringSize
(
this
->
type
());
...
...
@@ -3810,15 +3810,13 @@ void OpProto::MergeFrom(const OpProto &from) {
void
OpProto
::
CopyFrom
(
const
OpProto
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.OpProto)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
bool
OpProto
::
IsInitialized
()
const
{
if
((
_has_bits_
[
0
]
&
0x00000003
)
!=
0x00000003
)
return
false
;
if
((
_has_bits_
[
0
]
&
0x00000003
)
!=
0x00000003
)
return
false
;
if
(
!::
google
::
protobuf
::
internal
::
AllAreInitialized
(
this
->
inputs
()))
return
false
;
if
(
!::
google
::
protobuf
::
internal
::
AllAreInitialized
(
this
->
outputs
()))
...
...
@@ -3829,8 +3827,7 @@ bool OpProto::IsInitialized() const {
}
void
OpProto
::
Swap
(
OpProto
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
OpProto
::
InternalSwap
(
OpProto
*
other
)
{
...
...
@@ -3918,13 +3915,13 @@ void OpProto::set_allocated_type(::std::string *type) {
// repeated .paddle_mobile.framework.proto.OpProto.Var inputs = 2;
int
OpProto
::
inputs_size
()
const
{
return
inputs_
.
size
();
}
void
OpProto
::
clear_inputs
()
{
inputs_
.
Clear
();
}
const
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
&
OpProto
::
inputs
(
int
index
)
const
{
const
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
&
OpProto
::
inputs
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.OpProto.inputs)
return
inputs_
.
Get
(
index
);
}
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
*
OpProto
::
mutable_inputs
(
int
index
)
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
*
OpProto
::
mutable_inputs
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.OpProto.inputs)
return
inputs_
.
Mutable
(
index
);
}
...
...
@@ -3933,14 +3930,14 @@ OpProto::mutable_inputs(int index) {
return
inputs_
.
Add
();
}
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
*
OpProto
::
mutable_inputs
()
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
*
OpProto
::
mutable_inputs
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpProto.inputs)
return
&
inputs_
;
}
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
&
OpProto
::
inputs
()
const
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
&
OpProto
::
inputs
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpProto.inputs)
return
inputs_
;
}
...
...
@@ -3948,13 +3945,13 @@ OpProto::inputs() const {
// repeated .paddle_mobile.framework.proto.OpProto.Var outputs = 3;
int
OpProto
::
outputs_size
()
const
{
return
outputs_
.
size
();
}
void
OpProto
::
clear_outputs
()
{
outputs_
.
Clear
();
}
const
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
&
OpProto
::
outputs
(
int
index
)
const
{
const
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
&
OpProto
::
outputs
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.OpProto.outputs)
return
outputs_
.
Get
(
index
);
}
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
*
OpProto
::
mutable_outputs
(
int
index
)
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
*
OpProto
::
mutable_outputs
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.OpProto.outputs)
return
outputs_
.
Mutable
(
index
);
}
...
...
@@ -3963,14 +3960,14 @@ OpProto::mutable_outputs(int index) {
return
outputs_
.
Add
();
}
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
*
OpProto
::
mutable_outputs
()
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
*
OpProto
::
mutable_outputs
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpProto.outputs)
return
&
outputs_
;
}
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
&
OpProto
::
outputs
()
const
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
&
OpProto
::
outputs
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpProto.outputs)
return
outputs_
;
}
...
...
@@ -3978,13 +3975,13 @@ OpProto::outputs() const {
// repeated .paddle_mobile.framework.proto.OpProto.Attr attrs = 4;
int
OpProto
::
attrs_size
()
const
{
return
attrs_
.
size
();
}
void
OpProto
::
clear_attrs
()
{
attrs_
.
Clear
();
}
const
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
&
OpProto
::
attrs
(
int
index
)
const
{
const
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
&
OpProto
::
attrs
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.OpProto.attrs)
return
attrs_
.
Get
(
index
);
}
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
*
OpProto
::
mutable_attrs
(
int
index
)
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
*
OpProto
::
mutable_attrs
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.OpProto.attrs)
return
attrs_
.
Mutable
(
index
);
}
...
...
@@ -3993,14 +3990,14 @@ OpProto::mutable_attrs(int index) {
return
attrs_
.
Add
();
}
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
>
*
OpProto
::
mutable_attrs
()
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
>
*
OpProto
::
mutable_attrs
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpProto.attrs)
return
&
attrs_
;
}
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
>
&
OpProto
::
attrs
()
const
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
>
&
OpProto
::
attrs
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpProto.attrs)
return
attrs_
;
}
...
...
@@ -4071,14 +4068,14 @@ void OpProto::set_allocated_comment(::std::string *comment) {
// @@protoc_insertion_point(field_set_allocated:paddle_mobile.framework.proto.OpProto.comment)
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
#if !defined(_MSC_VER) || _MSC_VER >= 1900
const
int
VarType_TensorDesc
::
kDataTypeFieldNumber
;
const
int
VarType_TensorDesc
::
kDimsFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
VarType_TensorDesc
::
VarType_TensorDesc
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -4089,8 +4086,11 @@ VarType_TensorDesc::VarType_TensorDesc()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.VarType.TensorDesc)
}
VarType_TensorDesc
::
VarType_TensorDesc
(
const
VarType_TensorDesc
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
dims_
(
from
.
dims_
)
{
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
dims_
(
from
.
dims_
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
data_type_
=
from
.
data_type_
;
// @@protoc_insertion_point(copy_constructor:paddle_mobile.framework.proto.VarType.TensorDesc)
...
...
@@ -4118,8 +4118,8 @@ const VarType_TensorDesc &VarType_TensorDesc::default_instance() {
return
*
internal_default_instance
();
}
VarType_TensorDesc
*
VarType_TensorDesc
::
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
{
VarType_TensorDesc
*
VarType_TensorDesc
::
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
{
VarType_TensorDesc
*
n
=
new
VarType_TensorDesc
;
if
(
arena
!=
NULL
)
{
arena
->
Own
(
n
);
...
...
@@ -4141,9 +4141,8 @@ void VarType_TensorDesc::Clear() {
bool
VarType_TensorDesc
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -4157,65 +4156,67 @@ bool VarType_TensorDesc::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// required .paddle_mobile.framework.proto.VarType.Type
// data_type = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
8u
/* 8 & 0xFF */
))
{
int
value
;
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
int
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_ENUM
>
(
input
,
&
value
)));
if
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Type_IsValid
(
value
))
{
set_data_type
(
static_cast
<::
paddle_mobile
::
framework
::
proto
::
VarType_Type
>
(
value
));
// required .paddle_mobile.framework.proto.VarType.Type
// data_type = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
8u
/* 8 & 0xFF */
))
{
int
value
;
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
int
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_ENUM
>
(
input
,
&
value
)));
if
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Type_IsValid
(
value
))
{
set_data_type
(
static_cast
<::
paddle_mobile
::
framework
::
proto
::
VarType_Type
>
(
value
));
}
else
{
unknown_fields_stream
.
WriteVarint32
(
8u
);
unknown_fields_stream
.
WriteVarint32
(
static_cast
<::
google
::
protobuf
::
uint32
>
(
value
));
}
}
else
{
unknown_fields_stream
.
WriteVarint32
(
8u
);
unknown_fields_stream
.
WriteVarint32
(
static_cast
<::
google
::
protobuf
::
uint32
>
(
value
));
goto
handle_unusual
;
}
}
else
{
goto
handle_unusual
;
break
;
}
break
;
}
// repeated int64 dims = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
16u
/* 16 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadRepeatedPrimitive
<
::
google
::
protobuf
::
int64
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT64
>
(
1
,
16u
,
input
,
this
->
mutable_dims
())));
}
else
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
18u
/* 18 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPackedPrimitiveNoInline
<
::
google
::
protobuf
::
int64
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT64
>
(
input
,
this
->
mutable_dims
())));
}
else
{
goto
handle_unusual
;
// repeated int64 dims = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
16u
/* 16 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadRepeatedPrimitive
<
::
google
::
protobuf
::
int64
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT64
>
(
1
,
16u
,
input
,
this
->
mutable_dims
())));
}
else
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
18u
/* 18 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPackedPrimitiveNoInline
<
::
google
::
protobuf
::
int64
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT64
>
(
input
,
this
->
mutable_dims
())));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -4301,21 +4302,18 @@ void VarType_TensorDesc::MergeFrom(const VarType_TensorDesc &from) {
void
VarType_TensorDesc
::
CopyFrom
(
const
VarType_TensorDesc
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.VarType.TensorDesc)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
bool
VarType_TensorDesc
::
IsInitialized
()
const
{
if
((
_has_bits_
[
0
]
&
0x00000001
)
!=
0x00000001
)
return
false
;
if
((
_has_bits_
[
0
]
&
0x00000001
)
!=
0x00000001
)
return
false
;
return
true
;
}
void
VarType_TensorDesc
::
Swap
(
VarType_TensorDesc
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
VarType_TensorDesc
::
InternalSwap
(
VarType_TensorDesc
*
other
)
{
...
...
@@ -4347,8 +4345,8 @@ void VarType_TensorDesc::clear_data_type() {
data_type_
=
0
;
clear_has_data_type
();
}
::
paddle_mobile
::
framework
::
proto
::
VarType_Type
VarType_TensorDesc
::
data_type
()
const
{
::
paddle_mobile
::
framework
::
proto
::
VarType_Type
VarType_TensorDesc
::
data_type
()
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.VarType.TensorDesc.data_type)
return
static_cast
<::
paddle_mobile
::
framework
::
proto
::
VarType_Type
>
(
data_type_
);
...
...
@@ -4376,25 +4374,25 @@ void VarType_TensorDesc::add_dims(::google::protobuf::int64 value) {
dims_
.
Add
(
value
);
// @@protoc_insertion_point(field_add:paddle_mobile.framework.proto.VarType.TensorDesc.dims)
}
const
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int64
>
&
VarType_TensorDesc
::
dims
()
const
{
const
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int64
>
&
VarType_TensorDesc
::
dims
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.VarType.TensorDesc.dims)
return
dims_
;
}
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int64
>
*
VarType_TensorDesc
::
mutable_dims
()
{
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int64
>
*
VarType_TensorDesc
::
mutable_dims
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.VarType.TensorDesc.dims)
return
&
dims_
;
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
#if !defined(_MSC_VER) || _MSC_VER >= 1900
const
int
VarType_LoDTensorDesc
::
kTensorFieldNumber
;
const
int
VarType_LoDTensorDesc
::
kLodLevelFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
VarType_LoDTensorDesc
::
VarType_LoDTensorDesc
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -4405,8 +4403,10 @@ VarType_LoDTensorDesc::VarType_LoDTensorDesc()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.VarType.LoDTensorDesc)
}
VarType_LoDTensorDesc
::
VarType_LoDTensorDesc
(
const
VarType_LoDTensorDesc
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
)
{
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
if
(
from
.
has_tensor
())
{
tensor_
=
new
::
paddle_mobile
::
framework
::
proto
::
VarType_TensorDesc
(
...
...
@@ -4432,8 +4432,7 @@ VarType_LoDTensorDesc::~VarType_LoDTensorDesc() {
}
void
VarType_LoDTensorDesc
::
SharedDtor
()
{
if
(
this
!=
internal_default_instance
())
delete
tensor_
;
if
(
this
!=
internal_default_instance
())
delete
tensor_
;
}
void
VarType_LoDTensorDesc
::
SetCachedSize
(
int
size
)
const
{
...
...
@@ -4446,8 +4445,8 @@ const VarType_LoDTensorDesc &VarType_LoDTensorDesc::default_instance() {
return
*
internal_default_instance
();
}
VarType_LoDTensorDesc
*
VarType_LoDTensorDesc
::
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
{
VarType_LoDTensorDesc
*
VarType_LoDTensorDesc
::
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
{
VarType_LoDTensorDesc
*
n
=
new
VarType_LoDTensorDesc
;
if
(
arena
!=
NULL
)
{
arena
->
Own
(
n
);
...
...
@@ -4472,9 +4471,8 @@ void VarType_LoDTensorDesc::Clear() {
bool
VarType_LoDTensorDesc
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -4488,48 +4486,47 @@ bool VarType_LoDTensorDesc::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// required
// .paddle_mobile.framework.proto.VarType.TensorDesc tensor
// = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_tensor
()));
}
else
{
goto
handle_unusual
;
// required
// .paddle_mobile.framework.proto.VarType.TensorDesc tensor
// = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_tensor
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional int32 lod_level = 2 [default = 0];
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
16u
/* 16 & 0xFF */
))
{
set_has_lod_level
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
&
lod_level_
)));
}
else
{
goto
handle_unusual
;
// optional int32 lod_level = 2 [default = 0];
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
16u
/* 16 & 0xFF */
))
{
set_has_lod_level
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
&
lod_level_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -4622,25 +4619,21 @@ void VarType_LoDTensorDesc::MergeFrom(const VarType_LoDTensorDesc &from) {
void
VarType_LoDTensorDesc
::
CopyFrom
(
const
VarType_LoDTensorDesc
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.VarType.LoDTensorDesc)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
bool
VarType_LoDTensorDesc
::
IsInitialized
()
const
{
if
((
_has_bits_
[
0
]
&
0x00000001
)
!=
0x00000001
)
return
false
;
if
((
_has_bits_
[
0
]
&
0x00000001
)
!=
0x00000001
)
return
false
;
if
(
has_tensor
())
{
if
(
!
this
->
tensor_
->
IsInitialized
())
return
false
;
if
(
!
this
->
tensor_
->
IsInitialized
())
return
false
;
}
return
true
;
}
void
VarType_LoDTensorDesc
::
Swap
(
VarType_LoDTensorDesc
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
VarType_LoDTensorDesc
::
InternalSwap
(
VarType_LoDTensorDesc
*
other
)
{
...
...
@@ -4737,14 +4730,14 @@ void VarType_LoDTensorDesc::set_lod_level(::google::protobuf::int32 value) {
// @@protoc_insertion_point(field_set:paddle_mobile.framework.proto.VarType.LoDTensorDesc.lod_level)
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
#if !defined(_MSC_VER) || _MSC_VER >= 1900
const
int
VarType_LoDTensorArrayDesc
::
kTensorFieldNumber
;
const
int
VarType_LoDTensorArrayDesc
::
kLodLevelFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
VarType_LoDTensorArrayDesc
::
VarType_LoDTensorArrayDesc
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -4756,8 +4749,10 @@ VarType_LoDTensorArrayDesc::VarType_LoDTensorArrayDesc()
}
VarType_LoDTensorArrayDesc
::
VarType_LoDTensorArrayDesc
(
const
VarType_LoDTensorArrayDesc
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
)
{
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
if
(
from
.
has_tensor
())
{
tensor_
=
new
::
paddle_mobile
::
framework
::
proto
::
VarType_TensorDesc
(
...
...
@@ -4783,8 +4778,7 @@ VarType_LoDTensorArrayDesc::~VarType_LoDTensorArrayDesc() {
}
void
VarType_LoDTensorArrayDesc
::
SharedDtor
()
{
if
(
this
!=
internal_default_instance
())
delete
tensor_
;
if
(
this
!=
internal_default_instance
())
delete
tensor_
;
}
void
VarType_LoDTensorArrayDesc
::
SetCachedSize
(
int
size
)
const
{
...
...
@@ -4798,8 +4792,8 @@ VarType_LoDTensorArrayDesc::default_instance() {
return
*
internal_default_instance
();
}
VarType_LoDTensorArrayDesc
*
VarType_LoDTensorArrayDesc
::
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
{
VarType_LoDTensorArrayDesc
*
VarType_LoDTensorArrayDesc
::
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
{
VarType_LoDTensorArrayDesc
*
n
=
new
VarType_LoDTensorArrayDesc
;
if
(
arena
!=
NULL
)
{
arena
->
Own
(
n
);
...
...
@@ -4824,9 +4818,8 @@ void VarType_LoDTensorArrayDesc::Clear() {
bool
VarType_LoDTensorArrayDesc
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -4840,48 +4833,47 @@ bool VarType_LoDTensorArrayDesc::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// required
// .paddle_mobile.framework.proto.VarType.TensorDesc tensor
// = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_tensor
()));
}
else
{
goto
handle_unusual
;
// required
// .paddle_mobile.framework.proto.VarType.TensorDesc tensor
// = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_tensor
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional int32 lod_level = 2 [default = 0];
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
16u
/* 16 & 0xFF */
))
{
set_has_lod_level
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
&
lod_level_
)));
}
else
{
goto
handle_unusual
;
// optional int32 lod_level = 2 [default = 0];
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
16u
/* 16 & 0xFF */
))
{
set_has_lod_level
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
&
lod_level_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -4976,25 +4968,21 @@ void VarType_LoDTensorArrayDesc::MergeFrom(
void
VarType_LoDTensorArrayDesc
::
CopyFrom
(
const
VarType_LoDTensorArrayDesc
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.VarType.LoDTensorArrayDesc)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
bool
VarType_LoDTensorArrayDesc
::
IsInitialized
()
const
{
if
((
_has_bits_
[
0
]
&
0x00000001
)
!=
0x00000001
)
return
false
;
if
((
_has_bits_
[
0
]
&
0x00000001
)
!=
0x00000001
)
return
false
;
if
(
has_tensor
())
{
if
(
!
this
->
tensor_
->
IsInitialized
())
return
false
;
if
(
!
this
->
tensor_
->
IsInitialized
())
return
false
;
}
return
true
;
}
void
VarType_LoDTensorArrayDesc
::
Swap
(
VarType_LoDTensorArrayDesc
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
VarType_LoDTensorArrayDesc
::
InternalSwap
(
...
...
@@ -5096,13 +5084,13 @@ void VarType_LoDTensorArrayDesc::set_lod_level(
// @@protoc_insertion_point(field_set:paddle_mobile.framework.proto.VarType.LoDTensorArrayDesc.lod_level)
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
#if !defined(_MSC_VER) || _MSC_VER >= 1900
const
int
VarType_ReaderDesc
::
kLodTensorFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
VarType_ReaderDesc
::
VarType_ReaderDesc
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -5113,8 +5101,10 @@ VarType_ReaderDesc::VarType_ReaderDesc()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.VarType.ReaderDesc)
}
VarType_ReaderDesc
::
VarType_ReaderDesc
(
const
VarType_ReaderDesc
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
lod_tensor_
(
from
.
lod_tensor_
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
// @@protoc_insertion_point(copy_constructor:paddle_mobile.framework.proto.VarType.ReaderDesc)
...
...
@@ -5139,8 +5129,8 @@ const VarType_ReaderDesc &VarType_ReaderDesc::default_instance() {
return
*
internal_default_instance
();
}
VarType_ReaderDesc
*
VarType_ReaderDesc
::
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
{
VarType_ReaderDesc
*
VarType_ReaderDesc
::
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
{
VarType_ReaderDesc
*
n
=
new
VarType_ReaderDesc
;
if
(
arena
!=
NULL
)
{
arena
->
Own
(
n
);
...
...
@@ -5161,9 +5151,8 @@ void VarType_ReaderDesc::Clear() {
bool
VarType_ReaderDesc
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -5177,33 +5166,32 @@ bool VarType_ReaderDesc::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// repeated
// .paddle_mobile.framework.proto.VarType.LoDTensorDesc
// lod_tensor = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_lod_tensor
()));
}
else
{
goto
handle_unusual
;
// repeated
// .paddle_mobile.framework.proto.VarType.LoDTensorDesc
// lod_tensor = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_lod_tensor
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -5280,8 +5268,7 @@ void VarType_ReaderDesc::MergeFrom(const VarType_ReaderDesc &from) {
void
VarType_ReaderDesc
::
CopyFrom
(
const
VarType_ReaderDesc
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.VarType.ReaderDesc)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
...
...
@@ -5293,8 +5280,7 @@ bool VarType_ReaderDesc::IsInitialized() const {
}
void
VarType_ReaderDesc
::
Swap
(
VarType_ReaderDesc
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
VarType_ReaderDesc
::
InternalSwap
(
VarType_ReaderDesc
*
other
)
{
...
...
@@ -5332,26 +5318,26 @@ VarType_ReaderDesc::add_lod_tensor() {
return
lod_tensor_
.
Add
();
}
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
>
*
VarType_ReaderDesc
::
mutable_lod_tensor
()
{
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
>
*
VarType_ReaderDesc
::
mutable_lod_tensor
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.VarType.ReaderDesc.lod_tensor)
return
&
lod_tensor_
;
}
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
>
&
VarType_ReaderDesc
::
lod_tensor
()
const
{
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
>
&
VarType_ReaderDesc
::
lod_tensor
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.VarType.ReaderDesc.lod_tensor)
return
lod_tensor_
;
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
#if !defined(_MSC_VER) || _MSC_VER >= 1900
const
int
VarType_ChannelDesc
::
kDataTypeFieldNumber
;
const
int
VarType_ChannelDesc
::
kCapacityFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
VarType_ChannelDesc
::
VarType_ChannelDesc
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -5362,8 +5348,10 @@ VarType_ChannelDesc::VarType_ChannelDesc()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.VarType.ChannelDesc)
}
VarType_ChannelDesc
::
VarType_ChannelDesc
(
const
VarType_ChannelDesc
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
)
{
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
::
memcpy
(
&
capacity_
,
&
from
.
capacity_
,
static_cast
<
size_t
>
(
reinterpret_cast
<
char
*>
(
&
data_type_
)
-
...
...
@@ -5397,8 +5385,8 @@ const VarType_ChannelDesc &VarType_ChannelDesc::default_instance() {
return
*
internal_default_instance
();
}
VarType_ChannelDesc
*
VarType_ChannelDesc
::
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
{
VarType_ChannelDesc
*
VarType_ChannelDesc
::
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
{
VarType_ChannelDesc
*
n
=
new
VarType_ChannelDesc
;
if
(
arena
!=
NULL
)
{
arena
->
Own
(
n
);
...
...
@@ -5425,9 +5413,8 @@ void VarType_ChannelDesc::Clear() {
bool
VarType_ChannelDesc
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -5441,58 +5428,57 @@ bool VarType_ChannelDesc::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// required .paddle_mobile.framework.proto.VarType.Type
// data_type = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
8u
/* 8 & 0xFF */
))
{
int
value
;
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
int
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_ENUM
>
(
input
,
&
value
)));
if
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Type_IsValid
(
value
))
{
set_data_type
(
static_cast
<::
paddle_mobile
::
framework
::
proto
::
VarType_Type
>
(
value
));
// required .paddle_mobile.framework.proto.VarType.Type
// data_type = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
8u
/* 8 & 0xFF */
))
{
int
value
;
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
int
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_ENUM
>
(
input
,
&
value
)));
if
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Type_IsValid
(
value
))
{
set_data_type
(
static_cast
<::
paddle_mobile
::
framework
::
proto
::
VarType_Type
>
(
value
));
}
else
{
unknown_fields_stream
.
WriteVarint32
(
8u
);
unknown_fields_stream
.
WriteVarint32
(
static_cast
<::
google
::
protobuf
::
uint32
>
(
value
));
}
}
else
{
unknown_fields_stream
.
WriteVarint32
(
8u
);
unknown_fields_stream
.
WriteVarint32
(
static_cast
<::
google
::
protobuf
::
uint32
>
(
value
));
goto
handle_unusual
;
}
}
else
{
goto
handle_unusual
;
break
;
}
break
;
}
// required int64 capacity = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
16u
/* 16 & 0xFF */
))
{
set_has_capacity
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int64
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT64
>
(
input
,
&
capacity_
)));
}
else
{
goto
handle_unusual
;
// required int64 capacity = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
16u
/* 16 & 0xFF */
))
{
set_has_capacity
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int64
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT64
>
(
input
,
&
capacity_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -5556,7 +5542,7 @@ size_t VarType_ChannelDesc::ByteSizeLong() const {
total_size
+=
_internal_metadata_
.
unknown_fields
().
size
();
if
(((
_has_bits_
[
0
]
&
0x00000003
)
^
0x00000003
)
==
0
)
{
// All required fields are present.
0
)
{
// All required fields are present.
// required int64 capacity = 2;
total_size
+=
1
+
::
google
::
protobuf
::
internal
::
WireFormatLite
::
Int64Size
(
this
->
capacity
());
...
...
@@ -5602,21 +5588,18 @@ void VarType_ChannelDesc::MergeFrom(const VarType_ChannelDesc &from) {
void
VarType_ChannelDesc
::
CopyFrom
(
const
VarType_ChannelDesc
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.VarType.ChannelDesc)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
bool
VarType_ChannelDesc
::
IsInitialized
()
const
{
if
((
_has_bits_
[
0
]
&
0x00000003
)
!=
0x00000003
)
return
false
;
if
((
_has_bits_
[
0
]
&
0x00000003
)
!=
0x00000003
)
return
false
;
return
true
;
}
void
VarType_ChannelDesc
::
Swap
(
VarType_ChannelDesc
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
VarType_ChannelDesc
::
InternalSwap
(
VarType_ChannelDesc
*
other
)
{
...
...
@@ -5648,8 +5631,8 @@ void VarType_ChannelDesc::clear_data_type() {
data_type_
=
0
;
clear_has_data_type
();
}
::
paddle_mobile
::
framework
::
proto
::
VarType_Type
VarType_ChannelDesc
::
data_type
()
const
{
::
paddle_mobile
::
framework
::
proto
::
VarType_Type
VarType_ChannelDesc
::
data_type
()
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.VarType.ChannelDesc.data_type)
return
static_cast
<::
paddle_mobile
::
framework
::
proto
::
VarType_Type
>
(
data_type_
);
...
...
@@ -5684,13 +5667,13 @@ void VarType_ChannelDesc::set_capacity(::google::protobuf::int64 value) {
// @@protoc_insertion_point(field_set:paddle_mobile.framework.proto.VarType.ChannelDesc.capacity)
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
#if !defined(_MSC_VER) || _MSC_VER >= 1900
const
int
VarType_Tuple
::
kElementTypeFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
VarType_Tuple
::
VarType_Tuple
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -5701,8 +5684,10 @@ VarType_Tuple::VarType_Tuple()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.VarType.Tuple)
}
VarType_Tuple
::
VarType_Tuple
(
const
VarType_Tuple
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
element_type_
(
from
.
element_type_
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
// @@protoc_insertion_point(copy_constructor:paddle_mobile.framework.proto.VarType.Tuple)
...
...
@@ -5748,9 +5733,8 @@ void VarType_Tuple::Clear() {
bool
VarType_Tuple
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -5764,50 +5748,50 @@ bool VarType_Tuple::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// repeated .paddle_mobile.framework.proto.VarType.Type
// element_type = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
8u
/* 8 & 0xFF */
))
{
int
value
;
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
int
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_ENUM
>
(
input
,
&
value
)));
if
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Type_IsValid
(
value
))
{
add_element_type
(
static_cast
<::
paddle_mobile
::
framework
::
proto
::
VarType_Type
>
(
value
));
// repeated .paddle_mobile.framework.proto.VarType.Type
// element_type = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
8u
/* 8 & 0xFF */
))
{
int
value
;
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
int
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_ENUM
>
(
input
,
&
value
)));
if
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Type_IsValid
(
value
))
{
add_element_type
(
static_cast
<::
paddle_mobile
::
framework
::
proto
::
VarType_Type
>
(
value
));
}
else
{
unknown_fields_stream
.
WriteVarint32
(
tag
);
unknown_fields_stream
.
WriteVarint32
(
static_cast
<::
google
::
protobuf
::
uint32
>
(
value
));
}
}
else
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPackedEnumPreserveUnknowns
(
input
,
1
,
::
paddle_mobile
::
framework
::
proto
::
VarType_Type_IsValid
,
&
unknown_fields_stream
,
this
->
mutable_element_type
())));
}
else
{
unknown_fields_stream
.
WriteVarint32
(
tag
);
unknown_fields_stream
.
WriteVarint32
(
static_cast
<::
google
::
protobuf
::
uint32
>
(
value
));
goto
handle_unusual
;
}
}
else
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPackedEnumPreserveUnknowns
(
input
,
1
,
::
paddle_mobile
::
framework
::
proto
::
VarType_Type_IsValid
,
&
unknown_fields_stream
,
this
->
mutable_element_type
())));
}
else
{
goto
handle_unusual
;
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -5880,8 +5864,7 @@ void VarType_Tuple::MergeFrom(const VarType_Tuple &from) {
void
VarType_Tuple
::
CopyFrom
(
const
VarType_Tuple
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.VarType.Tuple)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
...
...
@@ -5889,8 +5872,7 @@ void VarType_Tuple::CopyFrom(const VarType_Tuple &from) {
bool
VarType_Tuple
::
IsInitialized
()
const
{
return
true
;
}
void
VarType_Tuple
::
Swap
(
VarType_Tuple
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
VarType_Tuple
::
InternalSwap
(
VarType_Tuple
*
other
)
{
...
...
@@ -5912,8 +5894,8 @@ void VarType_Tuple::InternalSwap(VarType_Tuple *other) {
// = 1;
int
VarType_Tuple
::
element_type_size
()
const
{
return
element_type_
.
size
();
}
void
VarType_Tuple
::
clear_element_type
()
{
element_type_
.
Clear
();
}
::
paddle_mobile
::
framework
::
proto
::
VarType_Type
VarType_Tuple
::
element_type
(
int
index
)
const
{
::
paddle_mobile
::
framework
::
proto
::
VarType_Type
VarType_Tuple
::
element_type
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.VarType.Tuple.element_type)
return
static_cast
<::
paddle_mobile
::
framework
::
proto
::
VarType_Type
>
(
element_type_
.
Get
(
index
));
...
...
@@ -5930,8 +5912,8 @@ void VarType_Tuple::add_element_type(
element_type_
.
Add
(
value
);
// @@protoc_insertion_point(field_add:paddle_mobile.framework.proto.VarType.Tuple.element_type)
}
const
::
google
::
protobuf
::
RepeatedField
<
int
>
&
VarType_Tuple
::
element_type
()
const
{
const
::
google
::
protobuf
::
RepeatedField
<
int
>
&
VarType_Tuple
::
element_type
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.VarType.Tuple.element_type)
return
element_type_
;
}
...
...
@@ -5940,7 +5922,7 @@ VarType_Tuple::element_type() const {
return
&
element_type_
;
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
...
...
@@ -5952,7 +5934,7 @@ const int VarType::kTensorArrayFieldNumber;
const
int
VarType
::
kReaderFieldNumber
;
const
int
VarType
::
kChannelFieldNumber
;
const
int
VarType
::
kTupleFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
VarType
::
VarType
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -5963,8 +5945,10 @@ VarType::VarType()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.VarType)
}
VarType
::
VarType
(
const
VarType
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
)
{
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
if
(
from
.
has_selected_rows
())
{
selected_rows_
=
new
::
paddle_mobile
::
framework
::
proto
::
VarType_TensorDesc
(
...
...
@@ -6020,18 +6004,12 @@ VarType::~VarType() {
}
void
VarType
::
SharedDtor
()
{
if
(
this
!=
internal_default_instance
())
delete
selected_rows_
;
if
(
this
!=
internal_default_instance
())
delete
lod_tensor_
;
if
(
this
!=
internal_default_instance
())
delete
tensor_array_
;
if
(
this
!=
internal_default_instance
())
delete
reader_
;
if
(
this
!=
internal_default_instance
())
delete
channel_
;
if
(
this
!=
internal_default_instance
())
delete
tuple_
;
if
(
this
!=
internal_default_instance
())
delete
selected_rows_
;
if
(
this
!=
internal_default_instance
())
delete
lod_tensor_
;
if
(
this
!=
internal_default_instance
())
delete
tensor_array_
;
if
(
this
!=
internal_default_instance
())
delete
reader_
;
if
(
this
!=
internal_default_instance
())
delete
channel_
;
if
(
this
!=
internal_default_instance
())
delete
tuple_
;
}
void
VarType
::
SetCachedSize
(
int
size
)
const
{
...
...
@@ -6095,9 +6073,8 @@ void VarType::Clear() {
bool
VarType
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -6111,126 +6088,126 @@ bool VarType::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// required .paddle_mobile.framework.proto.VarType.Type type
// = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
8u
/* 8 & 0xFF */
))
{
int
value
;
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
int
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_ENUM
>
(
input
,
&
value
)));
if
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Type_IsValid
(
value
))
{
set_type
(
static_cast
<::
paddle_mobile
::
framework
::
proto
::
VarType_Type
>
(
value
));
// required .paddle_mobile.framework.proto.VarType.Type type
// = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
8u
/* 8 & 0xFF */
))
{
int
value
;
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
int
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_ENUM
>
(
input
,
&
value
)));
if
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Type_IsValid
(
value
))
{
set_type
(
static_cast
<::
paddle_mobile
::
framework
::
proto
::
VarType_Type
>
(
value
));
}
else
{
unknown_fields_stream
.
WriteVarint32
(
8u
);
unknown_fields_stream
.
WriteVarint32
(
static_cast
<::
google
::
protobuf
::
uint32
>
(
value
));
}
}
else
{
unknown_fields_stream
.
WriteVarint32
(
8u
);
unknown_fields_stream
.
WriteVarint32
(
static_cast
<::
google
::
protobuf
::
uint32
>
(
value
));
goto
handle_unusual
;
}
}
else
{
goto
handle_unusual
;
break
;
}
break
;
}
// optional
// .paddle_mobile.framework.proto.VarType.TensorDesc
// selected_rows = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
18u
/* 18 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_selected_rows
()));
}
else
{
goto
handle_unusual
;
// optional
// .paddle_mobile.framework.proto.VarType.TensorDesc
// selected_rows = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
18u
/* 18 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_selected_rows
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional
// .paddle_mobile.framework.proto.VarType.LoDTensorDesc
// lod_tensor = 3;
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
26u
/* 26 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_lod_tensor
()));
}
else
{
goto
handle_unusual
;
// optional
// .paddle_mobile.framework.proto.VarType.LoDTensorDesc
// lod_tensor = 3;
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
26u
/* 26 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_lod_tensor
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional
// .paddle_mobile.framework.proto.VarType.LoDTensorArrayDesc
// tensor_array = 4;
case
4
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
34u
/* 34 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_tensor_array
()));
}
else
{
goto
handle_unusual
;
// optional
// .paddle_mobile.framework.proto.VarType.LoDTensorArrayDesc
// tensor_array = 4;
case
4
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
34u
/* 34 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_tensor_array
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional
// .paddle_mobile.framework.proto.VarType.ReaderDesc reader
// = 5;
case
5
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
42u
/* 42 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_reader
()));
}
else
{
goto
handle_unusual
;
// optional
// .paddle_mobile.framework.proto.VarType.ReaderDesc reader
// = 5;
case
5
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
42u
/* 42 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_reader
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional
// .paddle_mobile.framework.proto.VarType.ChannelDesc
// channel =
// 6;
case
6
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
50u
/* 50 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_channel
()));
}
else
{
goto
handle_unusual
;
// optional
// .paddle_mobile.framework.proto.VarType.ChannelDesc
// channel =
// 6;
case
6
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
50u
/* 50 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_channel
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional .paddle_mobile.framework.proto.VarType.Tuple
// tuple = 7;
case
7
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
58u
/* 58 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_tuple
()));
}
else
{
goto
handle_unusual
;
// optional .paddle_mobile.framework.proto.VarType.Tuple
// tuple = 7;
case
7
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
58u
/* 58 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_tuple
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -6440,41 +6417,33 @@ void VarType::MergeFrom(const VarType &from) {
void
VarType
::
CopyFrom
(
const
VarType
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.VarType)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
bool
VarType
::
IsInitialized
()
const
{
if
((
_has_bits_
[
0
]
&
0x00000040
)
!=
0x00000040
)
return
false
;
if
((
_has_bits_
[
0
]
&
0x00000040
)
!=
0x00000040
)
return
false
;
if
(
has_selected_rows
())
{
if
(
!
this
->
selected_rows_
->
IsInitialized
())
return
false
;
if
(
!
this
->
selected_rows_
->
IsInitialized
())
return
false
;
}
if
(
has_lod_tensor
())
{
if
(
!
this
->
lod_tensor_
->
IsInitialized
())
return
false
;
if
(
!
this
->
lod_tensor_
->
IsInitialized
())
return
false
;
}
if
(
has_tensor_array
())
{
if
(
!
this
->
tensor_array_
->
IsInitialized
())
return
false
;
if
(
!
this
->
tensor_array_
->
IsInitialized
())
return
false
;
}
if
(
has_reader
())
{
if
(
!
this
->
reader_
->
IsInitialized
())
return
false
;
if
(
!
this
->
reader_
->
IsInitialized
())
return
false
;
}
if
(
has_channel
())
{
if
(
!
this
->
channel_
->
IsInitialized
())
return
false
;
if
(
!
this
->
channel_
->
IsInitialized
())
return
false
;
}
return
true
;
}
void
VarType
::
Swap
(
VarType
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
VarType
::
InternalSwap
(
VarType
*
other
)
{
...
...
@@ -6691,8 +6660,8 @@ void VarType::clear_reader() {
reader_
->::
paddle_mobile
::
framework
::
proto
::
VarType_ReaderDesc
::
Clear
();
clear_has_reader
();
}
const
::
paddle_mobile
::
framework
::
proto
::
VarType_ReaderDesc
&
VarType
::
reader
()
const
{
const
::
paddle_mobile
::
framework
::
proto
::
VarType_ReaderDesc
&
VarType
::
reader
()
const
{
const
::
paddle_mobile
::
framework
::
proto
::
VarType_ReaderDesc
*
p
=
reader_
;
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.VarType.reader)
return
p
!=
NULL
...
...
@@ -6741,8 +6710,8 @@ void VarType::clear_channel() {
channel_
->::
paddle_mobile
::
framework
::
proto
::
VarType_ChannelDesc
::
Clear
();
clear_has_channel
();
}
const
::
paddle_mobile
::
framework
::
proto
::
VarType_ChannelDesc
&
VarType
::
channel
()
const
{
const
::
paddle_mobile
::
framework
::
proto
::
VarType_ChannelDesc
&
VarType
::
channel
()
const
{
const
::
paddle_mobile
::
framework
::
proto
::
VarType_ChannelDesc
*
p
=
channel_
;
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.VarType.channel)
return
p
!=
NULL
...
...
@@ -6827,7 +6796,7 @@ void VarType::set_allocated_tuple(
// @@protoc_insertion_point(field_set_allocated:paddle_mobile.framework.proto.VarType.tuple)
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
...
...
@@ -6835,7 +6804,7 @@ void VarType::set_allocated_tuple(
const
int
VarDesc
::
kNameFieldNumber
;
const
int
VarDesc
::
kTypeFieldNumber
;
const
int
VarDesc
::
kPersistableFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
VarDesc
::
VarDesc
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -6846,8 +6815,10 @@ VarDesc::VarDesc()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.VarDesc)
}
VarDesc
::
VarDesc
(
const
VarDesc
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
)
{
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
name_
.
UnsafeSetDefault
(
&::
google
::
protobuf
::
internal
::
GetEmptyStringAlreadyInited
());
...
...
@@ -6883,8 +6854,7 @@ VarDesc::~VarDesc() {
void
VarDesc
::
SharedDtor
()
{
name_
.
DestroyNoArena
(
&::
google
::
protobuf
::
internal
::
GetEmptyStringAlreadyInited
());
if
(
this
!=
internal_default_instance
())
delete
type_
;
if
(
this
!=
internal_default_instance
())
delete
type_
;
}
void
VarDesc
::
SetCachedSize
(
int
size
)
const
{
...
...
@@ -6930,9 +6900,8 @@ void VarDesc::Clear() {
bool
VarDesc
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -6946,57 +6915,56 @@ bool VarDesc::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// required string name = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_name
()));
}
else
{
goto
handle_unusual
;
// required string name = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadString
(
input
,
this
->
mutable_name
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// required .paddle_mobile.framework.proto.VarType type = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
18u
/* 18 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_type
()));
}
else
{
goto
handle_unusual
;
// required .paddle_mobile.framework.proto.VarType type = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
18u
/* 18 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
mutable_type
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional bool persistable = 3 [default = false];
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
24u
/* 24 & 0xFF */
))
{
set_has_persistable
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
&
persistable_
)));
}
else
{
goto
handle_unusual
;
// optional bool persistable = 3 [default = false];
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
24u
/* 24 & 0xFF */
))
{
set_has_persistable
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
bool
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_BOOL
>
(
input
,
&
persistable_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -7065,7 +7033,7 @@ size_t VarDesc::ByteSizeLong() const {
total_size
+=
_internal_metadata_
.
unknown_fields
().
size
();
if
(((
_has_bits_
[
0
]
&
0x00000003
)
^
0x00000003
)
==
0
)
{
// All required fields are present.
0
)
{
// All required fields are present.
// required string name = 1;
total_size
+=
1
+
::
google
::
protobuf
::
internal
::
WireFormatLite
::
StringSize
(
this
->
name
());
...
...
@@ -7123,25 +7091,21 @@ void VarDesc::MergeFrom(const VarDesc &from) {
void
VarDesc
::
CopyFrom
(
const
VarDesc
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.VarDesc)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
bool
VarDesc
::
IsInitialized
()
const
{
if
((
_has_bits_
[
0
]
&
0x00000003
)
!=
0x00000003
)
return
false
;
if
((
_has_bits_
[
0
]
&
0x00000003
)
!=
0x00000003
)
return
false
;
if
(
has_type
())
{
if
(
!
this
->
type_
->
IsInitialized
())
return
false
;
if
(
!
this
->
type_
->
IsInitialized
())
return
false
;
}
return
true
;
}
void
VarDesc
::
Swap
(
VarDesc
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
VarDesc
::
InternalSwap
(
VarDesc
*
other
)
{
...
...
@@ -7229,8 +7193,7 @@ bool VarDesc::has_type() const { return (_has_bits_[0] & 0x00000002u) != 0; }
void
VarDesc
::
set_has_type
()
{
_has_bits_
[
0
]
|=
0x00000002u
;
}
void
VarDesc
::
clear_has_type
()
{
_has_bits_
[
0
]
&=
~
0x00000002u
;
}
void
VarDesc
::
clear_type
()
{
if
(
type_
!=
NULL
)
type_
->::
paddle_mobile
::
framework
::
proto
::
VarType
::
Clear
();
if
(
type_
!=
NULL
)
type_
->::
paddle_mobile
::
framework
::
proto
::
VarType
::
Clear
();
clear_has_type
();
}
const
::
paddle_mobile
::
framework
::
proto
::
VarType
&
VarDesc
::
type
()
const
{
...
...
@@ -7289,7 +7252,7 @@ void VarDesc::set_persistable(bool value) {
// @@protoc_insertion_point(field_set:paddle_mobile.framework.proto.VarDesc.persistable)
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
...
...
@@ -7299,7 +7262,7 @@ const int BlockDesc::kParentIdxFieldNumber;
const
int
BlockDesc
::
kVarsFieldNumber
;
const
int
BlockDesc
::
kOpsFieldNumber
;
const
int
BlockDesc
::
kForwardBlockIdxFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
BlockDesc
::
BlockDesc
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -7310,8 +7273,11 @@ BlockDesc::BlockDesc()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.BlockDesc)
}
BlockDesc
::
BlockDesc
(
const
BlockDesc
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
vars_
(
from
.
vars_
),
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
vars_
(
from
.
vars_
),
ops_
(
from
.
ops_
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
::
memcpy
(
&
idx_
,
&
from
.
idx_
,
...
...
@@ -7377,9 +7343,8 @@ void BlockDesc::Clear() {
bool
BlockDesc
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -7393,88 +7358,87 @@ bool BlockDesc::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// required int32 idx = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
8u
/* 8 & 0xFF */
))
{
set_has_idx
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
&
idx_
)));
}
else
{
goto
handle_unusual
;
// required int32 idx = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
8u
/* 8 & 0xFF */
))
{
set_has_idx
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
&
idx_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// required int32 parent_idx = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
16u
/* 16 & 0xFF */
))
{
set_has_parent_idx
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
&
parent_idx_
)));
}
else
{
goto
handle_unusual
;
// required int32 parent_idx = 2;
case
2
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
16u
/* 16 & 0xFF */
))
{
set_has_parent_idx
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
&
parent_idx_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// repeated .paddle_mobile.framework.proto.VarDesc vars = 3;
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
26u
/* 26 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_vars
()));
}
else
{
goto
handle_unusual
;
// repeated .paddle_mobile.framework.proto.VarDesc vars = 3;
case
3
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
26u
/* 26 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_vars
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// repeated .paddle_mobile.framework.proto.OpDesc ops = 4;
case
4
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
34u
/* 34 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_ops
()));
}
else
{
goto
handle_unusual
;
// repeated .paddle_mobile.framework.proto.OpDesc ops = 4;
case
4
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
34u
/* 34 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_ops
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
// optional int32 forward_block_idx = 5 [default = -1];
case
5
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
40u
/* 40 & 0xFF */
))
{
set_has_forward_block_idx
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
&
forward_block_idx_
)));
}
else
{
goto
handle_unusual
;
// optional int32 forward_block_idx = 5 [default = -1];
case
5
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
40u
/* 40 & 0xFF */
))
{
set_has_forward_block_idx
();
DO_
((
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadPrimitive
<
::
google
::
protobuf
::
int32
,
::
google
::
protobuf
::
internal
::
WireFormatLite
::
TYPE_INT32
>
(
input
,
&
forward_block_idx_
)));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -7556,7 +7520,7 @@ size_t BlockDesc::ByteSizeLong() const {
total_size
+=
_internal_metadata_
.
unknown_fields
().
size
();
if
(((
_has_bits_
[
0
]
&
0x00000003
)
^
0x00000003
)
==
0
)
{
// All required fields are present.
0
)
{
// All required fields are present.
// required int32 idx = 1;
total_size
+=
1
+
::
google
::
protobuf
::
internal
::
WireFormatLite
::
Int32Size
(
this
->
idx
());
...
...
@@ -7634,15 +7598,13 @@ void BlockDesc::MergeFrom(const BlockDesc &from) {
void
BlockDesc
::
CopyFrom
(
const
BlockDesc
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.BlockDesc)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
bool
BlockDesc
::
IsInitialized
()
const
{
if
((
_has_bits_
[
0
]
&
0x00000003
)
!=
0x00000003
)
return
false
;
if
((
_has_bits_
[
0
]
&
0x00000003
)
!=
0x00000003
)
return
false
;
if
(
!::
google
::
protobuf
::
internal
::
AllAreInitialized
(
this
->
vars
()))
return
false
;
if
(
!::
google
::
protobuf
::
internal
::
AllAreInitialized
(
this
->
ops
()))
...
...
@@ -7651,8 +7613,7 @@ bool BlockDesc::IsInitialized() const {
}
void
BlockDesc
::
Swap
(
BlockDesc
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
BlockDesc
::
InternalSwap
(
BlockDesc
*
other
)
{
...
...
@@ -7715,8 +7676,8 @@ void BlockDesc::set_parent_idx(::google::protobuf::int32 value) {
// repeated .paddle_mobile.framework.proto.VarDesc vars = 3;
int
BlockDesc
::
vars_size
()
const
{
return
vars_
.
size
();
}
void
BlockDesc
::
clear_vars
()
{
vars_
.
Clear
();
}
const
::
paddle_mobile
::
framework
::
proto
::
VarDesc
&
BlockDesc
::
vars
(
int
index
)
const
{
const
::
paddle_mobile
::
framework
::
proto
::
VarDesc
&
BlockDesc
::
vars
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.BlockDesc.vars)
return
vars_
.
Get
(
index
);
}
...
...
@@ -7734,8 +7695,8 @@ BlockDesc::vars(int index) const {
return
&
vars_
;
}
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
VarDesc
>
&
BlockDesc
::
vars
()
const
{
::
paddle_mobile
::
framework
::
proto
::
VarDesc
>
&
BlockDesc
::
vars
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.BlockDesc.vars)
return
vars_
;
}
...
...
@@ -7743,8 +7704,8 @@ BlockDesc::vars() const {
// repeated .paddle_mobile.framework.proto.OpDesc ops = 4;
int
BlockDesc
::
ops_size
()
const
{
return
ops_
.
size
();
}
void
BlockDesc
::
clear_ops
()
{
ops_
.
Clear
();
}
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc
&
BlockDesc
::
ops
(
int
index
)
const
{
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc
&
BlockDesc
::
ops
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.BlockDesc.ops)
return
ops_
.
Get
(
index
);
}
...
...
@@ -7762,8 +7723,8 @@ BlockDesc::ops(int index) const {
return
&
ops_
;
}
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc
>
&
BlockDesc
::
ops
()
const
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc
>
&
BlockDesc
::
ops
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.BlockDesc.ops)
return
ops_
;
}
...
...
@@ -7788,13 +7749,13 @@ void BlockDesc::set_forward_block_idx(::google::protobuf::int32 value) {
// @@protoc_insertion_point(field_set:paddle_mobile.framework.proto.BlockDesc.forward_block_idx)
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// ===================================================================
#if !defined(_MSC_VER) || _MSC_VER >= 1900
const
int
ProgramDesc
::
kBlocksFieldNumber
;
#endif // !defined(_MSC_VER) || _MSC_VER >= 1900
#endif
// !defined(_MSC_VER) || _MSC_VER >= 1900
ProgramDesc
::
ProgramDesc
()
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
)
{
...
...
@@ -7805,8 +7766,11 @@ ProgramDesc::ProgramDesc()
// @@protoc_insertion_point(constructor:paddle_mobile.framework.proto.ProgramDesc)
}
ProgramDesc
::
ProgramDesc
(
const
ProgramDesc
&
from
)
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
blocks_
(
from
.
blocks_
)
{
:
::
google
::
protobuf
::
MessageLite
(),
_internal_metadata_
(
NULL
),
_has_bits_
(
from
.
_has_bits_
),
_cached_size_
(
0
),
blocks_
(
from
.
blocks_
)
{
_internal_metadata_
.
MergeFrom
(
from
.
_internal_metadata_
);
// @@protoc_insertion_point(copy_constructor:paddle_mobile.framework.proto.ProgramDesc)
}
...
...
@@ -7851,9 +7815,8 @@ void ProgramDesc::Clear() {
bool
ProgramDesc
::
MergePartialFromCodedStream
(
::
google
::
protobuf
::
io
::
CodedInputStream
*
input
)
{
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) \
goto failure
#define DO_(EXPRESSION) \
if (!GOOGLE_PREDICT_TRUE(EXPRESSION)) goto failure
::
google
::
protobuf
::
uint32
tag
;
::
google
::
protobuf
::
io
::
LazyStringOutputStream
unknown_fields_string
(
::
google
::
protobuf
::
NewPermanentCallback
(
...
...
@@ -7867,32 +7830,31 @@ bool ProgramDesc::MergePartialFromCodedStream(
::
std
::
pair
<::
google
::
protobuf
::
uint32
,
bool
>
p
=
input
->
ReadTagWithCutoffNoLastTag
(
127u
);
tag
=
p
.
first
;
if
(
!
p
.
second
)
goto
handle_unusual
;
if
(
!
p
.
second
)
goto
handle_unusual
;
switch
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
GetTagFieldNumber
(
tag
))
{
// repeated .paddle_mobile.framework.proto.BlockDesc blocks
// = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_blocks
()));
}
else
{
goto
handle_unusual
;
// repeated .paddle_mobile.framework.proto.BlockDesc blocks
// = 1;
case
1
:
{
if
(
static_cast
<::
google
::
protobuf
::
uint8
>
(
tag
)
==
static_cast
<::
google
::
protobuf
::
uint8
>
(
10u
/* 10 & 0xFF */
))
{
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
ReadMessageNoVirtual
(
input
,
add_blocks
()));
}
else
{
goto
handle_unusual
;
}
break
;
}
break
;
}
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
default:
{
handle_unusual:
if
(
tag
==
0
)
{
goto
success
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
DO_
(
::
google
::
protobuf
::
internal
::
WireFormatLite
::
SkipField
(
input
,
tag
,
&
unknown_fields_stream
));
break
;
}
}
}
success:
...
...
@@ -7964,8 +7926,7 @@ void ProgramDesc::MergeFrom(const ProgramDesc &from) {
void
ProgramDesc
::
CopyFrom
(
const
ProgramDesc
&
from
)
{
// @@protoc_insertion_point(class_specific_copy_from_start:paddle_mobile.framework.proto.ProgramDesc)
if
(
&
from
==
this
)
return
;
if
(
&
from
==
this
)
return
;
Clear
();
MergeFrom
(
from
);
}
...
...
@@ -7977,8 +7938,7 @@ bool ProgramDesc::IsInitialized() const {
}
void
ProgramDesc
::
Swap
(
ProgramDesc
*
other
)
{
if
(
other
==
this
)
return
;
if
(
other
==
this
)
return
;
InternalSwap
(
other
);
}
void
ProgramDesc
::
InternalSwap
(
ProgramDesc
*
other
)
{
...
...
@@ -7999,13 +7959,13 @@ void ProgramDesc::InternalSwap(ProgramDesc *other) {
// repeated .paddle_mobile.framework.proto.BlockDesc blocks = 1;
int
ProgramDesc
::
blocks_size
()
const
{
return
blocks_
.
size
();
}
void
ProgramDesc
::
clear_blocks
()
{
blocks_
.
Clear
();
}
const
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
&
ProgramDesc
::
blocks
(
int
index
)
const
{
const
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
&
ProgramDesc
::
blocks
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.ProgramDesc.blocks)
return
blocks_
.
Get
(
index
);
}
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
*
ProgramDesc
::
mutable_blocks
(
int
index
)
{
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
*
ProgramDesc
::
mutable_blocks
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.ProgramDesc.blocks)
return
blocks_
.
Mutable
(
index
);
}
...
...
@@ -8014,24 +7974,24 @@ ProgramDesc::mutable_blocks(int index) {
return
blocks_
.
Add
();
}
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
>
*
ProgramDesc
::
mutable_blocks
()
{
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
>
*
ProgramDesc
::
mutable_blocks
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.ProgramDesc.blocks)
return
&
blocks_
;
}
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
>
&
ProgramDesc
::
blocks
()
const
{
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
>
&
ProgramDesc
::
blocks
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.ProgramDesc.blocks)
return
blocks_
;
}
#endif // PROTOBUF_INLINE_NOT_IN_HEADERS
#endif
// PROTOBUF_INLINE_NOT_IN_HEADERS
// @@protoc_insertion_point(namespace_scope)
}
// namespace proto
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace proto
}
// namespace framework
}
// namespace paddle_mobile
// @@protoc_insertion_point(global_scope)
src/framework/framework.pb.h
浏览文件 @
d6825bf3
...
...
@@ -21,14 +21,14 @@
#include <google/protobuf/arena.h>
#include <google/protobuf/arenastring.h>
#include <google/protobuf/extension_set.h> // IWYU pragma: export
#include <google/protobuf/extension_set.h>
// IWYU pragma: export
#include <google/protobuf/generated_enum_util.h>
#include <google/protobuf/generated_message_table_driven.h>
#include <google/protobuf/generated_message_util.h>
#include <google/protobuf/io/coded_stream.h>
#include <google/protobuf/message_lite.h>
#include <google/protobuf/metadata_lite.h>
#include <google/protobuf/repeated_field.h> // IWYU pragma: export
#include <google/protobuf/repeated_field.h>
// IWYU pragma: export
// @@protoc_insertion_point(includes)
namespace
paddle_mobile
{
namespace
framework
{
...
...
@@ -86,9 +86,9 @@ extern VarType_TensorDescDefaultTypeInternal
class
VarType_Tuple
;
class
VarType_TupleDefaultTypeInternal
;
extern
VarType_TupleDefaultTypeInternal
_VarType_Tuple_default_instance_
;
}
// namespace proto
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace proto
}
// namespace framework
}
// namespace paddle_mobile
namespace
paddle_mobile
{
namespace
framework
{
...
...
@@ -108,7 +108,7 @@ struct TableStruct {
};
void
AddDescriptors
();
void
InitDefaults
();
}
// namespace protobuf_framework_2eproto
}
// namespace protobuf_framework_2eproto
enum
VarType_Type
{
VarType_Type_BOOL
=
0
,
...
...
@@ -160,7 +160,7 @@ class OpDesc_Attr
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.OpDesc.Attr)
*/
{
public:
public:
OpDesc_Attr
();
virtual
~
OpDesc_Attr
();
...
...
@@ -177,8 +177,7 @@ public:
inline
OpDesc_Attr
&
operator
=
(
OpDesc_Attr
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -224,17 +223,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
OpDesc_Attr
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -250,8 +249,8 @@ public:
::
google
::
protobuf
::
int32
ints
(
int
index
)
const
;
void
set_ints
(
int
index
,
::
google
::
protobuf
::
int32
value
);
void
add_ints
(
::
google
::
protobuf
::
int32
value
);
const
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int32
>
&
ints
()
const
;
const
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int32
>
&
ints
()
const
;
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int32
>
*
mutable_ints
();
// repeated float floats = 7;
...
...
@@ -369,7 +368,7 @@ public:
void
set_block_idx
(
::
google
::
protobuf
::
int32
value
);
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.OpDesc.Attr)
private:
private:
void
set_has_name
();
void
clear_has_name
();
void
set_has_type
();
...
...
@@ -415,7 +414,7 @@ class OpDesc_Var
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.OpDesc.Var)
*/
{
public:
public:
OpDesc_Var
();
virtual
~
OpDesc_Var
();
...
...
@@ -432,8 +431,7 @@ public:
inline
OpDesc_Var
&
operator
=
(
OpDesc_Var
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -478,17 +476,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
OpDesc_Var
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -535,7 +533,7 @@ public:
void
set_allocated_parameter
(
::
std
::
string
*
parameter
);
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.OpDesc.Var)
private:
private:
void
set_has_parameter
();
void
clear_has_parameter
();
...
...
@@ -554,7 +552,7 @@ class OpDesc
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.OpDesc)
*/
{
public:
public:
OpDesc
();
virtual
~
OpDesc
();
...
...
@@ -569,8 +567,7 @@ public:
inline
OpDesc
&
operator
=
(
OpDesc
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -615,17 +612,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
OpDesc
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -646,11 +643,11 @@ public:
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
*
mutable_inputs
(
int
index
);
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
*
add_inputs
();
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
*
mutable_inputs
();
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
*
mutable_inputs
();
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
&
inputs
()
const
;
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
&
inputs
()
const
;
// repeated .paddle_mobile.framework.proto.OpDesc.Var outputs =
// 2;
...
...
@@ -661,11 +658,11 @@ public:
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
*
mutable_outputs
(
int
index
);
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
*
add_outputs
();
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
*
mutable_outputs
();
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
*
mutable_outputs
();
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
&
outputs
()
const
;
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
&
outputs
()
const
;
// repeated .paddle_mobile.framework.proto.OpDesc.Attr attrs =
// 4;
...
...
@@ -676,11 +673,11 @@ public:
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
*
mutable_attrs
(
int
index
);
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
*
add_attrs
();
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
>
*
mutable_attrs
();
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
>
*
mutable_attrs
();
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
>
&
attrs
()
const
;
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
>
&
attrs
()
const
;
// required string type = 3;
bool
has_type
()
const
;
...
...
@@ -705,7 +702,7 @@ public:
void
set_is_target
(
bool
value
);
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.OpDesc)
private:
private:
void
set_has_type
();
void
clear_has_type
();
void
set_has_is_target
();
...
...
@@ -735,7 +732,7 @@ class OpProto_Var
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.OpProto.Var)
*/
{
public:
public:
OpProto_Var
();
virtual
~
OpProto_Var
();
...
...
@@ -752,8 +749,7 @@ public:
inline
OpProto_Var
&
operator
=
(
OpProto_Var
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -799,17 +795,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
OpProto_Var
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -870,7 +866,7 @@ public:
void
set_dispensable
(
bool
value
);
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.OpProto.Var)
private:
private:
void
set_has_name
();
void
clear_has_name
();
void
set_has_comment
();
...
...
@@ -903,7 +899,7 @@ class OpProto_Attr
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.OpProto.Attr)
*/
{
public:
public:
OpProto_Attr
();
virtual
~
OpProto_Attr
();
...
...
@@ -920,8 +916,7 @@ public:
inline
OpProto_Attr
&
operator
=
(
OpProto_Attr
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -967,17 +962,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
OpProto_Attr
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -1031,7 +1026,7 @@ public:
void
set_generated
(
bool
value
);
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.OpProto.Attr)
private:
private:
void
set_has_name
();
void
clear_has_name
();
void
set_has_type
();
...
...
@@ -1061,7 +1056,7 @@ class OpProto
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.OpProto)
*/
{
public:
public:
OpProto
();
virtual
~
OpProto
();
...
...
@@ -1076,8 +1071,7 @@ public:
inline
OpProto
&
operator
=
(
OpProto
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -1122,17 +1116,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
OpProto
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -1153,27 +1147,27 @@ public:
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
*
mutable_inputs
(
int
index
);
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
*
add_inputs
();
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
*
mutable_inputs
();
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
*
mutable_inputs
();
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
&
inputs
()
const
;
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
&
inputs
()
const
;
// repeated .paddle_mobile.framework.proto.OpProto.Var outputs =
// 3;
int
outputs_size
()
const
;
void
clear_outputs
();
static
const
int
kOutputsFieldNumber
=
3
;
const
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
&
outputs
(
int
index
)
const
;
const
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
&
outputs
(
int
index
)
const
;
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
*
mutable_outputs
(
int
index
);
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
*
add_outputs
();
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
*
mutable_outputs
();
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
*
mutable_outputs
();
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
&
outputs
()
const
;
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
&
outputs
()
const
;
// repeated .paddle_mobile.framework.proto.OpProto.Attr attrs =
// 4;
...
...
@@ -1184,11 +1178,11 @@ public:
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
*
mutable_attrs
(
int
index
);
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
*
add_attrs
();
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
>
*
mutable_attrs
();
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
>
*
mutable_attrs
();
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
>
&
attrs
()
const
;
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
>
&
attrs
()
const
;
// required string type = 1;
bool
has_type
()
const
;
...
...
@@ -1221,7 +1215,7 @@ public:
void
set_allocated_comment
(
::
std
::
string
*
comment
);
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.OpProto)
private:
private:
void
set_has_type
();
void
clear_has_type
();
void
set_has_comment
();
...
...
@@ -1254,7 +1248,7 @@ class VarType_TensorDesc
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.VarType.TensorDesc)
*/
{
public:
public:
VarType_TensorDesc
();
virtual
~
VarType_TensorDesc
();
...
...
@@ -1272,8 +1266,7 @@ public:
inline
VarType_TensorDesc
&
operator
=
(
VarType_TensorDesc
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -1303,8 +1296,8 @@ public:
inline
VarType_TensorDesc
*
New
()
const
PROTOBUF_FINAL
{
return
New
(
NULL
);
}
VarType_TensorDesc
*
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
PROTOBUF_FINAL
;
VarType_TensorDesc
*
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
PROTOBUF_FINAL
;
void
CheckTypeAndMergeFrom
(
const
::
google
::
protobuf
::
MessageLite
&
from
)
PROTOBUF_FINAL
;
void
CopyFrom
(
const
VarType_TensorDesc
&
from
);
...
...
@@ -1320,17 +1313,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
VarType_TensorDesc
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -1346,8 +1339,8 @@ public:
::
google
::
protobuf
::
int64
dims
(
int
index
)
const
;
void
set_dims
(
int
index
,
::
google
::
protobuf
::
int64
value
);
void
add_dims
(
::
google
::
protobuf
::
int64
value
);
const
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int64
>
&
dims
()
const
;
const
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int64
>
&
dims
()
const
;
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int64
>
*
mutable_dims
();
// required .paddle_mobile.framework.proto.VarType.Type
...
...
@@ -1359,7 +1352,7 @@ public:
void
set_data_type
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Type
value
);
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.VarType.TensorDesc)
private:
private:
void
set_has_data_type
();
void
clear_has_data_type
();
...
...
@@ -1378,7 +1371,7 @@ class VarType_LoDTensorDesc
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.VarType.LoDTensorDesc)
*/
{
public:
public:
VarType_LoDTensorDesc
();
virtual
~
VarType_LoDTensorDesc
();
...
...
@@ -1394,11 +1387,10 @@ public:
*
this
=
::
std
::
move
(
from
);
}
inline
VarType_LoDTensorDesc
&
operator
=
(
VarType_LoDTensorDesc
&&
from
)
noexcept
{
inline
VarType_LoDTensorDesc
&
operator
=
(
VarType_LoDTensorDesc
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -1430,8 +1422,8 @@ public:
inline
VarType_LoDTensorDesc
*
New
()
const
PROTOBUF_FINAL
{
return
New
(
NULL
);
}
VarType_LoDTensorDesc
*
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
PROTOBUF_FINAL
;
VarType_LoDTensorDesc
*
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
PROTOBUF_FINAL
;
void
CheckTypeAndMergeFrom
(
const
::
google
::
protobuf
::
MessageLite
&
from
)
PROTOBUF_FINAL
;
void
CopyFrom
(
const
VarType_LoDTensorDesc
&
from
);
...
...
@@ -1447,17 +1439,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
VarType_LoDTensorDesc
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -1485,7 +1477,7 @@ public:
void
set_lod_level
(
::
google
::
protobuf
::
int32
value
);
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.VarType.LoDTensorDesc)
private:
private:
void
set_has_tensor
();
void
clear_has_tensor
();
void
set_has_lod_level
();
...
...
@@ -1506,14 +1498,14 @@ class VarType_LoDTensorArrayDesc
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.VarType.LoDTensorArrayDesc)
*/
{
public:
public:
VarType_LoDTensorArrayDesc
();
virtual
~
VarType_LoDTensorArrayDesc
();
VarType_LoDTensorArrayDesc
(
const
VarType_LoDTensorArrayDesc
&
from
);
inline
VarType_LoDTensorArrayDesc
&
operator
=
(
const
VarType_LoDTensorArrayDesc
&
from
)
{
inline
VarType_LoDTensorArrayDesc
&
operator
=
(
const
VarType_LoDTensorArrayDesc
&
from
)
{
CopyFrom
(
from
);
return
*
this
;
}
...
...
@@ -1523,11 +1515,10 @@ public:
*
this
=
::
std
::
move
(
from
);
}
inline
VarType_LoDTensorArrayDesc
&
operator
=
(
VarType_LoDTensorArrayDesc
&&
from
)
noexcept
{
inline
VarType_LoDTensorArrayDesc
&
operator
=
(
VarType_LoDTensorArrayDesc
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -1562,8 +1553,8 @@ public:
return
New
(
NULL
);
}
VarType_LoDTensorArrayDesc
*
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
PROTOBUF_FINAL
;
VarType_LoDTensorArrayDesc
*
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
PROTOBUF_FINAL
;
void
CheckTypeAndMergeFrom
(
const
::
google
::
protobuf
::
MessageLite
&
from
)
PROTOBUF_FINAL
;
void
CopyFrom
(
const
VarType_LoDTensorArrayDesc
&
from
);
...
...
@@ -1579,17 +1570,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
VarType_LoDTensorArrayDesc
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -1617,7 +1608,7 @@ public:
void
set_lod_level
(
::
google
::
protobuf
::
int32
value
);
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.VarType.LoDTensorArrayDesc)
private:
private:
void
set_has_tensor
();
void
clear_has_tensor
();
void
set_has_lod_level
();
...
...
@@ -1638,7 +1629,7 @@ class VarType_ReaderDesc
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.VarType.ReaderDesc)
*/
{
public:
public:
VarType_ReaderDesc
();
virtual
~
VarType_ReaderDesc
();
...
...
@@ -1656,8 +1647,7 @@ public:
inline
VarType_ReaderDesc
&
operator
=
(
VarType_ReaderDesc
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -1687,8 +1677,8 @@ public:
inline
VarType_ReaderDesc
*
New
()
const
PROTOBUF_FINAL
{
return
New
(
NULL
);
}
VarType_ReaderDesc
*
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
PROTOBUF_FINAL
;
VarType_ReaderDesc
*
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
PROTOBUF_FINAL
;
void
CheckTypeAndMergeFrom
(
const
::
google
::
protobuf
::
MessageLite
&
from
)
PROTOBUF_FINAL
;
void
CopyFrom
(
const
VarType_ReaderDesc
&
from
);
...
...
@@ -1704,17 +1694,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
VarType_ReaderDesc
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -1729,20 +1719,20 @@ public:
int
lod_tensor_size
()
const
;
void
clear_lod_tensor
();
static
const
int
kLodTensorFieldNumber
=
1
;
const
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
&
lod_tensor
(
int
index
)
const
;
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
*
mutable_lod_tensor
(
int
index
);
const
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
&
lod_tensor
(
int
index
)
const
;
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
*
mutable_lod_tensor
(
int
index
);
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
*
add_lod_tensor
();
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
>
*
mutable_lod_tensor
();
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
>
*
mutable_lod_tensor
();
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
>
&
lod_tensor
()
const
;
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
>
&
lod_tensor
()
const
;
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.VarType.ReaderDesc)
private:
private:
::
google
::
protobuf
::
internal
::
InternalMetadataWithArenaLite
_internal_metadata_
;
::
google
::
protobuf
::
internal
::
HasBits
<
1
>
_has_bits_
;
...
...
@@ -1759,7 +1749,7 @@ class VarType_ChannelDesc
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.VarType.ChannelDesc)
*/
{
public:
public:
VarType_ChannelDesc
();
virtual
~
VarType_ChannelDesc
();
...
...
@@ -1777,8 +1767,7 @@ public:
inline
VarType_ChannelDesc
&
operator
=
(
VarType_ChannelDesc
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -1810,8 +1799,8 @@ public:
inline
VarType_ChannelDesc
*
New
()
const
PROTOBUF_FINAL
{
return
New
(
NULL
);
}
VarType_ChannelDesc
*
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
PROTOBUF_FINAL
;
VarType_ChannelDesc
*
New
(
::
google
::
protobuf
::
Arena
*
arena
)
const
PROTOBUF_FINAL
;
void
CheckTypeAndMergeFrom
(
const
::
google
::
protobuf
::
MessageLite
&
from
)
PROTOBUF_FINAL
;
void
CopyFrom
(
const
VarType_ChannelDesc
&
from
);
...
...
@@ -1827,17 +1816,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
VarType_ChannelDesc
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -1862,7 +1851,7 @@ public:
void
set_data_type
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Type
value
);
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.VarType.ChannelDesc)
private:
private:
void
set_has_data_type
();
void
clear_has_data_type
();
void
set_has_capacity
();
...
...
@@ -1886,7 +1875,7 @@ class VarType_Tuple
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.VarType.Tuple)
*/
{
public:
public:
VarType_Tuple
();
virtual
~
VarType_Tuple
();
...
...
@@ -1903,8 +1892,7 @@ public:
inline
VarType_Tuple
&
operator
=
(
VarType_Tuple
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -1950,17 +1938,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
VarType_Tuple
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -1982,7 +1970,7 @@ public:
::
google
::
protobuf
::
RepeatedField
<
int
>
*
mutable_element_type
();
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.VarType.Tuple)
private:
private:
::
google
::
protobuf
::
internal
::
InternalMetadataWithArenaLite
_internal_metadata_
;
::
google
::
protobuf
::
internal
::
HasBits
<
1
>
_has_bits_
;
...
...
@@ -1997,7 +1985,7 @@ class VarType
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.VarType)
*/
{
public:
public:
VarType
();
virtual
~
VarType
();
...
...
@@ -2012,8 +2000,7 @@ public:
inline
VarType
&
operator
=
(
VarType
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -2058,17 +2045,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
VarType
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -2117,8 +2104,8 @@ public:
bool
has_selected_rows
()
const
;
void
clear_selected_rows
();
static
const
int
kSelectedRowsFieldNumber
=
2
;
const
::
paddle_mobile
::
framework
::
proto
::
VarType_TensorDesc
&
selected_rows
()
const
;
const
::
paddle_mobile
::
framework
::
proto
::
VarType_TensorDesc
&
selected_rows
()
const
;
::
paddle_mobile
::
framework
::
proto
::
VarType_TensorDesc
*
mutable_selected_rows
();
::
paddle_mobile
::
framework
::
proto
::
VarType_TensorDesc
*
...
...
@@ -2132,8 +2119,8 @@ public:
bool
has_lod_tensor
()
const
;
void
clear_lod_tensor
();
static
const
int
kLodTensorFieldNumber
=
3
;
const
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
&
lod_tensor
()
const
;
const
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
&
lod_tensor
()
const
;
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
*
mutable_lod_tensor
();
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
*
...
...
@@ -2187,8 +2174,8 @@ public:
const
::
paddle_mobile
::
framework
::
proto
::
VarType_Tuple
&
tuple
()
const
;
::
paddle_mobile
::
framework
::
proto
::
VarType_Tuple
*
mutable_tuple
();
::
paddle_mobile
::
framework
::
proto
::
VarType_Tuple
*
release_tuple
();
void
set_allocated_tuple
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Tuple
*
tuple
);
void
set_allocated_tuple
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Tuple
*
tuple
);
// required .paddle_mobile.framework.proto.VarType.Type type =
// 1;
...
...
@@ -2199,7 +2186,7 @@ public:
void
set_type
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Type
value
);
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.VarType)
private:
private:
void
set_has_type
();
void
clear_has_type
();
void
set_has_selected_rows
();
...
...
@@ -2235,7 +2222,7 @@ class VarDesc
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.VarDesc)
*/
{
public:
public:
VarDesc
();
virtual
~
VarDesc
();
...
...
@@ -2250,8 +2237,7 @@ public:
inline
VarDesc
&
operator
=
(
VarDesc
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -2296,17 +2282,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
VarDesc
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -2347,7 +2333,7 @@ public:
void
set_persistable
(
bool
value
);
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.VarDesc)
private:
private:
void
set_has_name
();
void
clear_has_name
();
void
set_has_type
();
...
...
@@ -2374,7 +2360,7 @@ class BlockDesc
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.BlockDesc)
*/
{
public:
public:
BlockDesc
();
virtual
~
BlockDesc
();
...
...
@@ -2391,8 +2377,7 @@ public:
inline
BlockDesc
&
operator
=
(
BlockDesc
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -2437,17 +2422,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
BlockDesc
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -2464,11 +2449,11 @@ public:
::
paddle_mobile
::
framework
::
proto
::
VarDesc
*
mutable_vars
(
int
index
);
::
paddle_mobile
::
framework
::
proto
::
VarDesc
*
add_vars
();
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
VarDesc
>
*
mutable_vars
();
::
paddle_mobile
::
framework
::
proto
::
VarDesc
>
*
mutable_vars
();
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
VarDesc
>
&
vars
()
const
;
::
paddle_mobile
::
framework
::
proto
::
VarDesc
>
&
vars
()
const
;
// repeated .paddle_mobile.framework.proto.OpDesc ops = 4;
int
ops_size
()
const
;
...
...
@@ -2478,11 +2463,11 @@ public:
::
paddle_mobile
::
framework
::
proto
::
OpDesc
*
mutable_ops
(
int
index
);
::
paddle_mobile
::
framework
::
proto
::
OpDesc
*
add_ops
();
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc
>
*
mutable_ops
();
::
paddle_mobile
::
framework
::
proto
::
OpDesc
>
*
mutable_ops
();
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc
>
&
ops
()
const
;
::
paddle_mobile
::
framework
::
proto
::
OpDesc
>
&
ops
()
const
;
// required int32 idx = 1;
bool
has_idx
()
const
;
...
...
@@ -2506,7 +2491,7 @@ public:
void
set_forward_block_idx
(
::
google
::
protobuf
::
int32
value
);
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.BlockDesc)
private:
private:
void
set_has_idx
();
void
clear_has_idx
();
void
set_has_parent_idx
();
...
...
@@ -2539,7 +2524,7 @@ class ProgramDesc
MessageLite
/* @@protoc_insertion_point(class_definition:paddle_mobile.framework.proto.ProgramDesc)
*/
{
public:
public:
ProgramDesc
();
virtual
~
ProgramDesc
();
...
...
@@ -2556,8 +2541,7 @@ public:
inline
ProgramDesc
&
operator
=
(
ProgramDesc
&&
from
)
noexcept
{
if
(
GetArenaNoVirtual
()
==
from
.
GetArenaNoVirtual
())
{
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
if
(
this
!=
&
from
)
InternalSwap
(
&
from
);
}
else
{
CopyFrom
(
from
);
}
...
...
@@ -2603,17 +2587,17 @@ public:
void
DiscardUnknownFields
();
int
GetCachedSize
()
const
PROTOBUF_FINAL
{
return
_cached_size_
;
}
private:
private:
void
SharedCtor
();
void
SharedDtor
();
void
SetCachedSize
(
int
size
)
const
;
void
InternalSwap
(
ProgramDesc
*
other
);
private:
private:
inline
::
google
::
protobuf
::
Arena
*
GetArenaNoVirtual
()
const
{
return
NULL
;
}
inline
void
*
MaybeArenaPtr
()
const
{
return
NULL
;
}
public:
public:
::
std
::
string
GetTypeName
()
const
PROTOBUF_FINAL
;
// nested types
...
...
@@ -2630,14 +2614,14 @@ public:
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
*
mutable_blocks
(
int
index
);
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
*
add_blocks
();
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
>
*
mutable_blocks
();
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
>
*
mutable_blocks
();
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
>
&
blocks
()
const
;
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
>
&
blocks
()
const
;
// @@protoc_insertion_point(class_scope:paddle_mobile.framework.proto.ProgramDesc)
private:
private:
::
google
::
protobuf
::
internal
::
InternalMetadataWithArenaLite
_internal_metadata_
;
::
google
::
protobuf
::
internal
::
HasBits
<
1
>
_has_bits_
;
...
...
@@ -2655,8 +2639,8 @@ private:
#ifdef __GNUC__
#pragma GCC diagnostic push
#pragma GCC diagnostic ignored "-Wstrict-aliasing"
#endif // __GNUC__
// OpDesc_Attr
#endif
// __GNUC__
// OpDesc_Attr
// required string name = 1;
inline
bool
OpDesc_Attr
::
has_name
()
const
{
...
...
@@ -2737,8 +2721,8 @@ inline ::paddle_mobile::framework::proto::AttrType OpDesc_Attr::type() const {
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.OpDesc.Attr.type)
return
static_cast
<::
paddle_mobile
::
framework
::
proto
::
AttrType
>
(
type_
);
}
inline
void
OpDesc_Attr
::
set_type
(
::
paddle_mobile
::
framework
::
proto
::
AttrType
value
)
{
inline
void
OpDesc_Attr
::
set_type
(
::
paddle_mobile
::
framework
::
proto
::
AttrType
value
)
{
assert
(
::
paddle_mobile
::
framework
::
proto
::
AttrType_IsValid
(
value
));
set_has_type
();
type_
=
value
;
...
...
@@ -2865,13 +2849,13 @@ inline void OpDesc_Attr::add_ints(::google::protobuf::int32 value) {
ints_
.
Add
(
value
);
// @@protoc_insertion_point(field_add:paddle_mobile.framework.proto.OpDesc.Attr.ints)
}
inline
const
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int32
>
&
OpDesc_Attr
::
ints
()
const
{
inline
const
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int32
>
&
OpDesc_Attr
::
ints
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpDesc.Attr.ints)
return
ints_
;
}
inline
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int32
>
*
OpDesc_Attr
::
mutable_ints
()
{
inline
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int32
>
*
OpDesc_Attr
::
mutable_ints
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpDesc.Attr.ints)
return
&
ints_
;
}
...
...
@@ -2891,8 +2875,8 @@ inline void OpDesc_Attr::add_floats(float value) {
floats_
.
Add
(
value
);
// @@protoc_insertion_point(field_add:paddle_mobile.framework.proto.OpDesc.Attr.floats)
}
inline
const
::
google
::
protobuf
::
RepeatedField
<
float
>
&
OpDesc_Attr
::
floats
()
const
{
inline
const
::
google
::
protobuf
::
RepeatedField
<
float
>
&
OpDesc_Attr
::
floats
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpDesc.Attr.floats)
return
floats_
;
}
...
...
@@ -2955,13 +2939,13 @@ inline void OpDesc_Attr::add_strings(const char *value, size_t size) {
strings_
.
Add
()
->
assign
(
reinterpret_cast
<
const
char
*>
(
value
),
size
);
// @@protoc_insertion_point(field_add_pointer:paddle_mobile.framework.proto.OpDesc.Attr.strings)
}
inline
const
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
&
OpDesc_Attr
::
strings
()
const
{
inline
const
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
&
OpDesc_Attr
::
strings
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpDesc.Attr.strings)
return
strings_
;
}
inline
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
*
OpDesc_Attr
::
mutable_strings
()
{
inline
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
*
OpDesc_Attr
::
mutable_strings
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpDesc.Attr.strings)
return
&
strings_
;
}
...
...
@@ -3001,8 +2985,8 @@ inline void OpDesc_Attr::add_bools(bool value) {
bools_
.
Add
(
value
);
// @@protoc_insertion_point(field_add:paddle_mobile.framework.proto.OpDesc.Attr.bools)
}
inline
const
::
google
::
protobuf
::
RepeatedField
<
bool
>
&
OpDesc_Attr
::
bools
()
const
{
inline
const
::
google
::
protobuf
::
RepeatedField
<
bool
>
&
OpDesc_Attr
::
bools
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpDesc.Attr.bools)
return
bools_
;
}
...
...
@@ -3180,13 +3164,13 @@ inline void OpDesc_Var::add_arguments(const char *value, size_t size) {
arguments_
.
Add
()
->
assign
(
reinterpret_cast
<
const
char
*>
(
value
),
size
);
// @@protoc_insertion_point(field_add_pointer:paddle_mobile.framework.proto.OpDesc.Var.arguments)
}
inline
const
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
&
OpDesc_Var
::
arguments
()
const
{
inline
const
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
&
OpDesc_Var
::
arguments
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpDesc.Var.arguments)
return
arguments_
;
}
inline
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
*
OpDesc_Var
::
mutable_arguments
()
{
inline
::
google
::
protobuf
::
RepeatedPtrField
<::
std
::
string
>
*
OpDesc_Var
::
mutable_arguments
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpDesc.Var.arguments)
return
&
arguments_
;
}
...
...
@@ -3263,13 +3247,13 @@ inline void OpDesc::set_allocated_type(::std::string *type) {
// repeated .paddle_mobile.framework.proto.OpDesc.Var inputs = 1;
inline
int
OpDesc
::
inputs_size
()
const
{
return
inputs_
.
size
();
}
inline
void
OpDesc
::
clear_inputs
()
{
inputs_
.
Clear
();
}
inline
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
&
OpDesc
::
inputs
(
int
index
)
const
{
inline
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
&
OpDesc
::
inputs
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.OpDesc.inputs)
return
inputs_
.
Get
(
index
);
}
inline
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
*
OpDesc
::
mutable_inputs
(
int
index
)
{
inline
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
*
OpDesc
::
mutable_inputs
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.OpDesc.inputs)
return
inputs_
.
Mutable
(
index
);
}
...
...
@@ -3278,14 +3262,14 @@ inline ::paddle_mobile::framework::proto::OpDesc_Var *OpDesc::add_inputs() {
return
inputs_
.
Add
();
}
inline
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
*
OpDesc
::
mutable_inputs
()
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
*
OpDesc
::
mutable_inputs
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpDesc.inputs)
return
&
inputs_
;
}
inline
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
&
OpDesc
::
inputs
()
const
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
&
OpDesc
::
inputs
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpDesc.inputs)
return
inputs_
;
}
...
...
@@ -3293,13 +3277,13 @@ OpDesc::inputs() const {
// repeated .paddle_mobile.framework.proto.OpDesc.Var outputs = 2;
inline
int
OpDesc
::
outputs_size
()
const
{
return
outputs_
.
size
();
}
inline
void
OpDesc
::
clear_outputs
()
{
outputs_
.
Clear
();
}
inline
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
&
OpDesc
::
outputs
(
int
index
)
const
{
inline
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
&
OpDesc
::
outputs
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.OpDesc.outputs)
return
outputs_
.
Get
(
index
);
}
inline
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
*
OpDesc
::
mutable_outputs
(
int
index
)
{
inline
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
*
OpDesc
::
mutable_outputs
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.OpDesc.outputs)
return
outputs_
.
Mutable
(
index
);
}
...
...
@@ -3308,14 +3292,14 @@ inline ::paddle_mobile::framework::proto::OpDesc_Var *OpDesc::add_outputs() {
return
outputs_
.
Add
();
}
inline
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
*
OpDesc
::
mutable_outputs
()
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
*
OpDesc
::
mutable_outputs
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpDesc.outputs)
return
&
outputs_
;
}
inline
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
&
OpDesc
::
outputs
()
const
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Var
>
&
OpDesc
::
outputs
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpDesc.outputs)
return
outputs_
;
}
...
...
@@ -3323,13 +3307,13 @@ OpDesc::outputs() const {
// repeated .paddle_mobile.framework.proto.OpDesc.Attr attrs = 4;
inline
int
OpDesc
::
attrs_size
()
const
{
return
attrs_
.
size
();
}
inline
void
OpDesc
::
clear_attrs
()
{
attrs_
.
Clear
();
}
inline
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
&
OpDesc
::
attrs
(
int
index
)
const
{
inline
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
&
OpDesc
::
attrs
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.OpDesc.attrs)
return
attrs_
.
Get
(
index
);
}
inline
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
*
OpDesc
::
mutable_attrs
(
int
index
)
{
inline
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
*
OpDesc
::
mutable_attrs
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.OpDesc.attrs)
return
attrs_
.
Mutable
(
index
);
}
...
...
@@ -3338,14 +3322,14 @@ inline ::paddle_mobile::framework::proto::OpDesc_Attr *OpDesc::add_attrs() {
return
attrs_
.
Add
();
}
inline
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
>
*
OpDesc
::
mutable_attrs
()
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
>
*
OpDesc
::
mutable_attrs
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpDesc.attrs)
return
&
attrs_
;
}
inline
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
>
&
OpDesc
::
attrs
()
const
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc_Attr
>
&
OpDesc
::
attrs
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpDesc.attrs)
return
attrs_
;
}
...
...
@@ -3658,8 +3642,8 @@ inline ::paddle_mobile::framework::proto::AttrType OpProto_Attr::type() const {
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.OpProto.Attr.type)
return
static_cast
<::
paddle_mobile
::
framework
::
proto
::
AttrType
>
(
type_
);
}
inline
void
OpProto_Attr
::
set_type
(
::
paddle_mobile
::
framework
::
proto
::
AttrType
value
)
{
inline
void
OpProto_Attr
::
set_type
(
::
paddle_mobile
::
framework
::
proto
::
AttrType
value
)
{
assert
(
::
paddle_mobile
::
framework
::
proto
::
AttrType_IsValid
(
value
));
set_has_type
();
type_
=
value
;
...
...
@@ -3828,13 +3812,13 @@ inline void OpProto::set_allocated_type(::std::string *type) {
// repeated .paddle_mobile.framework.proto.OpProto.Var inputs = 2;
inline
int
OpProto
::
inputs_size
()
const
{
return
inputs_
.
size
();
}
inline
void
OpProto
::
clear_inputs
()
{
inputs_
.
Clear
();
}
inline
const
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
&
OpProto
::
inputs
(
int
index
)
const
{
inline
const
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
&
OpProto
::
inputs
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.OpProto.inputs)
return
inputs_
.
Get
(
index
);
}
inline
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
*
OpProto
::
mutable_inputs
(
int
index
)
{
inline
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
*
OpProto
::
mutable_inputs
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.OpProto.inputs)
return
inputs_
.
Mutable
(
index
);
}
...
...
@@ -3843,14 +3827,14 @@ inline ::paddle_mobile::framework::proto::OpProto_Var *OpProto::add_inputs() {
return
inputs_
.
Add
();
}
inline
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
*
OpProto
::
mutable_inputs
()
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
*
OpProto
::
mutable_inputs
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpProto.inputs)
return
&
inputs_
;
}
inline
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
&
OpProto
::
inputs
()
const
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
&
OpProto
::
inputs
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpProto.inputs)
return
inputs_
;
}
...
...
@@ -3858,13 +3842,13 @@ OpProto::inputs() const {
// repeated .paddle_mobile.framework.proto.OpProto.Var outputs = 3;
inline
int
OpProto
::
outputs_size
()
const
{
return
outputs_
.
size
();
}
inline
void
OpProto
::
clear_outputs
()
{
outputs_
.
Clear
();
}
inline
const
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
&
OpProto
::
outputs
(
int
index
)
const
{
inline
const
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
&
OpProto
::
outputs
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.OpProto.outputs)
return
outputs_
.
Get
(
index
);
}
inline
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
*
OpProto
::
mutable_outputs
(
int
index
)
{
inline
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
*
OpProto
::
mutable_outputs
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.OpProto.outputs)
return
outputs_
.
Mutable
(
index
);
}
...
...
@@ -3873,14 +3857,14 @@ inline ::paddle_mobile::framework::proto::OpProto_Var *OpProto::add_outputs() {
return
outputs_
.
Add
();
}
inline
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
*
OpProto
::
mutable_outputs
()
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
*
OpProto
::
mutable_outputs
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpProto.outputs)
return
&
outputs_
;
}
inline
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
&
OpProto
::
outputs
()
const
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Var
>
&
OpProto
::
outputs
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpProto.outputs)
return
outputs_
;
}
...
...
@@ -3888,13 +3872,13 @@ OpProto::outputs() const {
// repeated .paddle_mobile.framework.proto.OpProto.Attr attrs = 4;
inline
int
OpProto
::
attrs_size
()
const
{
return
attrs_
.
size
();
}
inline
void
OpProto
::
clear_attrs
()
{
attrs_
.
Clear
();
}
inline
const
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
&
OpProto
::
attrs
(
int
index
)
const
{
inline
const
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
&
OpProto
::
attrs
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.OpProto.attrs)
return
attrs_
.
Get
(
index
);
}
inline
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
*
OpProto
::
mutable_attrs
(
int
index
)
{
inline
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
*
OpProto
::
mutable_attrs
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.OpProto.attrs)
return
attrs_
.
Mutable
(
index
);
}
...
...
@@ -3903,14 +3887,14 @@ inline ::paddle_mobile::framework::proto::OpProto_Attr *OpProto::add_attrs() {
return
attrs_
.
Add
();
}
inline
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
>
*
OpProto
::
mutable_attrs
()
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
>
*
OpProto
::
mutable_attrs
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.OpProto.attrs)
return
&
attrs_
;
}
inline
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
>
&
OpProto
::
attrs
()
const
{
::
paddle_mobile
::
framework
::
proto
::
OpProto_Attr
>
&
OpProto
::
attrs
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.OpProto.attrs)
return
attrs_
;
}
...
...
@@ -4032,13 +4016,13 @@ inline void VarType_TensorDesc::add_dims(::google::protobuf::int64 value) {
dims_
.
Add
(
value
);
// @@protoc_insertion_point(field_add:paddle_mobile.framework.proto.VarType.TensorDesc.dims)
}
inline
const
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int64
>
&
VarType_TensorDesc
::
dims
()
const
{
inline
const
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int64
>
&
VarType_TensorDesc
::
dims
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.VarType.TensorDesc.dims)
return
dims_
;
}
inline
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int64
>
*
VarType_TensorDesc
::
mutable_dims
()
{
inline
::
google
::
protobuf
::
RepeatedField
<::
google
::
protobuf
::
int64
>
*
VarType_TensorDesc
::
mutable_dims
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.VarType.TensorDesc.dims)
return
&
dims_
;
}
...
...
@@ -4121,8 +4105,8 @@ inline ::google::protobuf::int32 VarType_LoDTensorDesc::lod_level() const {
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.VarType.LoDTensorDesc.lod_level)
return
lod_level_
;
}
inline
void
VarType_LoDTensorDesc
::
set_lod_level
(
::
google
::
protobuf
::
int32
value
)
{
inline
void
VarType_LoDTensorDesc
::
set_lod_level
(
::
google
::
protobuf
::
int32
value
)
{
set_has_lod_level
();
lod_level_
=
value
;
// @@protoc_insertion_point(field_set:paddle_mobile.framework.proto.VarType.LoDTensorDesc.lod_level)
...
...
@@ -4206,8 +4190,8 @@ inline ::google::protobuf::int32 VarType_LoDTensorArrayDesc::lod_level() const {
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.VarType.LoDTensorArrayDesc.lod_level)
return
lod_level_
;
}
inline
void
VarType_LoDTensorArrayDesc
::
set_lod_level
(
::
google
::
protobuf
::
int32
value
)
{
inline
void
VarType_LoDTensorArrayDesc
::
set_lod_level
(
::
google
::
protobuf
::
int32
value
)
{
set_has_lod_level
();
lod_level_
=
value
;
// @@protoc_insertion_point(field_set:paddle_mobile.framework.proto.VarType.LoDTensorArrayDesc.lod_level)
...
...
@@ -4239,14 +4223,14 @@ VarType_ReaderDesc::add_lod_tensor() {
return
lod_tensor_
.
Add
();
}
inline
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
>
*
VarType_ReaderDesc
::
mutable_lod_tensor
()
{
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
>
*
VarType_ReaderDesc
::
mutable_lod_tensor
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.VarType.ReaderDesc.lod_tensor)
return
&
lod_tensor_
;
}
inline
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
>
&
VarType_ReaderDesc
::
lod_tensor
()
const
{
::
paddle_mobile
::
framework
::
proto
::
VarType_LoDTensorDesc
>
&
VarType_ReaderDesc
::
lod_tensor
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.VarType.ReaderDesc.lod_tensor)
return
lod_tensor_
;
}
...
...
@@ -4336,13 +4320,13 @@ inline void VarType_Tuple::add_element_type(
element_type_
.
Add
(
value
);
// @@protoc_insertion_point(field_add:paddle_mobile.framework.proto.VarType.Tuple.element_type)
}
inline
const
::
google
::
protobuf
::
RepeatedField
<
int
>
&
VarType_Tuple
::
element_type
()
const
{
inline
const
::
google
::
protobuf
::
RepeatedField
<
int
>
&
VarType_Tuple
::
element_type
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.VarType.Tuple.element_type)
return
element_type_
;
}
inline
::
google
::
protobuf
::
RepeatedField
<
int
>
*
VarType_Tuple
::
mutable_element_type
()
{
inline
::
google
::
protobuf
::
RepeatedField
<
int
>
*
VarType_Tuple
::
mutable_element_type
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.VarType.Tuple.element_type)
return
&
element_type_
;
}
...
...
@@ -4365,8 +4349,8 @@ inline ::paddle_mobile::framework::proto::VarType_Type VarType::type() const {
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.VarType.type)
return
static_cast
<::
paddle_mobile
::
framework
::
proto
::
VarType_Type
>
(
type_
);
}
inline
void
VarType
::
set_type
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Type
value
)
{
inline
void
VarType
::
set_type
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Type
value
)
{
assert
(
::
paddle_mobile
::
framework
::
proto
::
VarType_Type_IsValid
(
value
));
set_has_type
();
type_
=
value
;
...
...
@@ -4654,8 +4638,8 @@ inline void VarType::clear_tuple() {
tuple_
->::
paddle_mobile
::
framework
::
proto
::
VarType_Tuple
::
Clear
();
clear_has_tuple
();
}
inline
const
::
paddle_mobile
::
framework
::
proto
::
VarType_Tuple
&
VarType
::
tuple
()
const
{
inline
const
::
paddle_mobile
::
framework
::
proto
::
VarType_Tuple
&
VarType
::
tuple
()
const
{
const
::
paddle_mobile
::
framework
::
proto
::
VarType_Tuple
*
p
=
tuple_
;
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.VarType.tuple)
return
p
!=
NULL
...
...
@@ -4770,8 +4754,7 @@ inline bool VarDesc::has_type() const {
inline
void
VarDesc
::
set_has_type
()
{
_has_bits_
[
0
]
|=
0x00000002u
;
}
inline
void
VarDesc
::
clear_has_type
()
{
_has_bits_
[
0
]
&=
~
0x00000002u
;
}
inline
void
VarDesc
::
clear_type
()
{
if
(
type_
!=
NULL
)
type_
->::
paddle_mobile
::
framework
::
proto
::
VarType
::
Clear
();
if
(
type_
!=
NULL
)
type_
->::
paddle_mobile
::
framework
::
proto
::
VarType
::
Clear
();
clear_has_type
();
}
inline
const
::
paddle_mobile
::
framework
::
proto
::
VarType
&
VarDesc
::
type
()
const
{
...
...
@@ -4798,8 +4781,8 @@ inline ::paddle_mobile::framework::proto::VarType *VarDesc::release_type() {
type_
=
NULL
;
return
temp
;
}
inline
void
VarDesc
::
set_allocated_type
(
::
paddle_mobile
::
framework
::
proto
::
VarType
*
type
)
{
inline
void
VarDesc
::
set_allocated_type
(
::
paddle_mobile
::
framework
::
proto
::
VarType
*
type
)
{
delete
type_
;
type_
=
type
;
if
(
type
)
{
...
...
@@ -4877,13 +4860,13 @@ inline void BlockDesc::set_parent_idx(::google::protobuf::int32 value) {
// repeated .paddle_mobile.framework.proto.VarDesc vars = 3;
inline
int
BlockDesc
::
vars_size
()
const
{
return
vars_
.
size
();
}
inline
void
BlockDesc
::
clear_vars
()
{
vars_
.
Clear
();
}
inline
const
::
paddle_mobile
::
framework
::
proto
::
VarDesc
&
BlockDesc
::
vars
(
int
index
)
const
{
inline
const
::
paddle_mobile
::
framework
::
proto
::
VarDesc
&
BlockDesc
::
vars
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.BlockDesc.vars)
return
vars_
.
Get
(
index
);
}
inline
::
paddle_mobile
::
framework
::
proto
::
VarDesc
*
BlockDesc
::
mutable_vars
(
int
index
)
{
inline
::
paddle_mobile
::
framework
::
proto
::
VarDesc
*
BlockDesc
::
mutable_vars
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.BlockDesc.vars)
return
vars_
.
Mutable
(
index
);
}
...
...
@@ -4892,14 +4875,14 @@ inline ::paddle_mobile::framework::proto::VarDesc *BlockDesc::add_vars() {
return
vars_
.
Add
();
}
inline
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
VarDesc
>
*
BlockDesc
::
mutable_vars
()
{
::
paddle_mobile
::
framework
::
proto
::
VarDesc
>
*
BlockDesc
::
mutable_vars
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.BlockDesc.vars)
return
&
vars_
;
}
inline
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
VarDesc
>
&
BlockDesc
::
vars
()
const
{
::
paddle_mobile
::
framework
::
proto
::
VarDesc
>
&
BlockDesc
::
vars
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.BlockDesc.vars)
return
vars_
;
}
...
...
@@ -4907,13 +4890,13 @@ BlockDesc::vars() const {
// repeated .paddle_mobile.framework.proto.OpDesc ops = 4;
inline
int
BlockDesc
::
ops_size
()
const
{
return
ops_
.
size
();
}
inline
void
BlockDesc
::
clear_ops
()
{
ops_
.
Clear
();
}
inline
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc
&
BlockDesc
::
ops
(
int
index
)
const
{
inline
const
::
paddle_mobile
::
framework
::
proto
::
OpDesc
&
BlockDesc
::
ops
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.BlockDesc.ops)
return
ops_
.
Get
(
index
);
}
inline
::
paddle_mobile
::
framework
::
proto
::
OpDesc
*
BlockDesc
::
mutable_ops
(
int
index
)
{
inline
::
paddle_mobile
::
framework
::
proto
::
OpDesc
*
BlockDesc
::
mutable_ops
(
int
index
)
{
// @@protoc_insertion_point(field_mutable:paddle_mobile.framework.proto.BlockDesc.ops)
return
ops_
.
Mutable
(
index
);
}
...
...
@@ -4922,14 +4905,14 @@ inline ::paddle_mobile::framework::proto::OpDesc *BlockDesc::add_ops() {
return
ops_
.
Add
();
}
inline
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc
>
*
BlockDesc
::
mutable_ops
()
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc
>
*
BlockDesc
::
mutable_ops
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.BlockDesc.ops)
return
&
ops_
;
}
inline
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
OpDesc
>
&
BlockDesc
::
ops
()
const
{
::
paddle_mobile
::
framework
::
proto
::
OpDesc
>
&
BlockDesc
::
ops
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.BlockDesc.ops)
return
ops_
;
}
...
...
@@ -4965,8 +4948,8 @@ inline void BlockDesc::set_forward_block_idx(::google::protobuf::int32 value) {
// repeated .paddle_mobile.framework.proto.BlockDesc blocks = 1;
inline
int
ProgramDesc
::
blocks_size
()
const
{
return
blocks_
.
size
();
}
inline
void
ProgramDesc
::
clear_blocks
()
{
blocks_
.
Clear
();
}
inline
const
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
&
ProgramDesc
::
blocks
(
int
index
)
const
{
inline
const
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
&
ProgramDesc
::
blocks
(
int
index
)
const
{
// @@protoc_insertion_point(field_get:paddle_mobile.framework.proto.ProgramDesc.blocks)
return
blocks_
.
Get
(
index
);
}
...
...
@@ -4980,23 +4963,23 @@ inline ::paddle_mobile::framework::proto::BlockDesc *ProgramDesc::add_blocks() {
return
blocks_
.
Add
();
}
inline
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
>
*
ProgramDesc
::
mutable_blocks
()
{
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
>
*
ProgramDesc
::
mutable_blocks
()
{
// @@protoc_insertion_point(field_mutable_list:paddle_mobile.framework.proto.ProgramDesc.blocks)
return
&
blocks_
;
}
inline
const
::
google
::
protobuf
::
RepeatedPtrField
<
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
>
&
ProgramDesc
::
blocks
()
const
{
::
paddle_mobile
::
framework
::
proto
::
BlockDesc
>
&
ProgramDesc
::
blocks
()
const
{
// @@protoc_insertion_point(field_list:paddle_mobile.framework.proto.ProgramDesc.blocks)
return
blocks_
;
}
#ifdef __GNUC__
#pragma GCC diagnostic pop
#endif // __GNUC__
#endif // !PROTOBUF_INLINE_NOT_IN_HEADERS
// -------------------------------------------------------------------
#endif
// __GNUC__
#endif
// !PROTOBUF_INLINE_NOT_IN_HEADERS
// -------------------------------------------------------------------
// -------------------------------------------------------------------
...
...
@@ -5028,9 +5011,9 @@ ProgramDesc::blocks() const {
// @@protoc_insertion_point(namespace_scope)
}
// namespace proto
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace proto
}
// namespace framework
}
// namespace paddle_mobile
namespace
google
{
namespace
protobuf
{
...
...
@@ -5042,9 +5025,9 @@ template <>
struct
is_proto_enum
<::
paddle_mobile
::
framework
::
proto
::
AttrType
>
:
::
google
::
protobuf
::
internal
::
true_type
{};
}
// namespace protobuf
}
// namespace google
}
// namespace protobuf
}
// namespace google
// @@protoc_insertion_point(global_scope)
#endif // PROTOBUF_framework_2eproto__INCLUDED
#endif
// PROTOBUF_framework_2eproto__INCLUDED
src/framework/lod_tensor.cc
浏览文件 @
d6825bf3
...
...
@@ -13,10 +13,10 @@ See the License for the specific language governing permissions and
limitations under the License. */
#include "lod_tensor.h"
#include <algorithm>
#include <iterator>
#include <stdint.h>
#include <string.h>
#include <algorithm>
#include <iterator>
namespace
paddle_mobile
{
namespace
framework
{
...
...
@@ -105,8 +105,7 @@ LoD SliceInLevel(const LoD &in, size_t level, size_t elem_begin,
LoD
ToAbsOffset
(
const
LoD
&
in
)
{
// the lowest level stores relative offsets
if
(
in
.
empty
()
||
in
.
size
()
==
1
)
return
in
;
if
(
in
.
empty
()
||
in
.
size
()
==
1
)
return
in
;
LoD
result
=
in
;
for
(
auto
level
=
static_cast
<
int
>
(
in
.
size
()
-
2
);
level
>=
0
;
level
--
)
{
for
(
size_t
i
=
0
;
i
<
in
[
level
].
size
();
++
i
)
{
...
...
@@ -138,23 +137,19 @@ bool operator==(const LoD &a, const LoD &b) {
}
bool
CheckLoD
(
const
LoD
&
in
,
int
tensor_height
)
{
if
(
in
.
empty
())
return
true
;
if
(
in
.
empty
())
return
true
;
for
(
const
auto
&
level
:
in
)
{
// check: there should be more than 2 offsets existing in each
// level.
if
(
level
.
size
()
<
2
)
return
false
;
if
(
level
.
size
()
<
2
)
return
false
;
// check: the first offset(the begin offset) of each level
// should be 0.
if
(
level
.
front
()
!=
0
)
return
false
;
if
(
level
.
front
()
!=
0
)
return
false
;
// check: all the offsets in a level should be ascending(no same
// items
// allows).
if
(
!
std
::
is_sorted
(
level
.
begin
(),
level
.
begin
(),
[](
size_t
a
,
size_t
b
)
{
if
(
a
<
b
)
return
true
;
if
(
a
<
b
)
return
true
;
return
false
;
}))
{
std
::
cout
<<
"ascending error"
;
...
...
@@ -174,22 +169,19 @@ bool CheckLoD(const LoD &in, int tensor_height) {
// goes
// first.
for
(
size_t
level
=
0
;
level
<
in
.
size
()
-
1
;
level
++
)
{
if
(
in
[
level
].
back
()
!=
in
[
level
+
1
].
size
()
-
1
)
return
false
;
if
(
in
[
level
].
back
()
!=
in
[
level
+
1
].
size
()
-
1
)
return
false
;
}
return
true
;
}
bool
CheckAbsLoD
(
const
LoD
&
in
,
int
tensor_height
)
{
if
(
in
.
empty
())
return
true
;
if
(
in
.
empty
())
return
true
;
for
(
const
auto
&
level
:
in
)
{
// check: all the offsets in a level should be ascending(no same
// items
// allows).
if
(
!
std
::
is_sorted
(
level
.
begin
(),
level
.
begin
(),
[](
size_t
a
,
size_t
b
)
{
if
(
a
<
b
)
return
true
;
if
(
a
<
b
)
return
true
;
return
false
;
}))
{
return
false
;
...
...
@@ -197,14 +189,12 @@ bool CheckAbsLoD(const LoD &in, int tensor_height) {
// check: there should be more than 2 offsets existing in each
// level.
if
(
level
.
size
()
<
2
)
return
false
;
if
(
level
.
size
()
<
2
)
return
false
;
// check: the first offset of each level should be 0, and the
// last should be
// the same(the height of underlying tensor).
if
(
level
.
front
()
!=
0
)
return
false
;
if
(
level
.
front
()
!=
0
)
return
false
;
if
(
tensor_height
<
0
)
{
tensor_height
=
level
.
back
();
}
else
if
((
size_t
)
tensor_height
!=
level
.
back
())
{
...
...
@@ -242,7 +232,7 @@ void AppendLoD(LoD *lod, const LoD &lod_length) {
// lod.");
if
(
lod
->
empty
())
{
for
(
size_t
i
=
0
;
i
<
lod_length
.
size
();
++
i
)
{
lod
->
emplace_back
(
1
,
0
);
// size = 1, value = 0;
lod
->
emplace_back
(
1
,
0
);
// size = 1, value = 0;
}
*
lod
=
LoD
(
lod_length
.
size
(),
std
::
vector
<
size_t
>
({
0
}));
}
...
...
@@ -255,7 +245,7 @@ void AppendLoD(LoD *lod, const LoD &lod_length) {
}
void
SerializeToStream
(
std
::
ostream
&
os
,
const
LoDTensor
&
tensor
)
{
{
// the 1st field, uint32_t version for LoDTensor
{
// the 1st field, uint32_t version for LoDTensor
constexpr
uint32_t
version
=
0
;
os
.
write
(
reinterpret_cast
<
const
char
*>
(
&
version
),
sizeof
(
version
));
}
...
...
@@ -307,5 +297,5 @@ void DeserializeFromStream(std::istream &is, LoDTensor *tensor) {
TensorFromStream
(
is
,
static_cast
<
Tensor
*>
(
tensor
));
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/lod_tensor.h
浏览文件 @
d6825bf3
...
...
@@ -14,12 +14,12 @@ limitations under the License. */
#pragma once
#include "tensor.h"
#include "tensor_util.h"
#include <memory>
#include <string>
#include <utility>
#include <vector>
#include "tensor.h"
#include "tensor_util.h"
namespace
paddle_mobile
{
...
...
@@ -102,7 +102,7 @@ bool CheckAbsLoD(const LoD &in, int tensor_height = -1);
* see https://en.wikipedia.org/wiki/Level_of_details for reference.
*/
class
LoDTensor
:
public
Tensor
{
public:
public:
LoDTensor
()
:
Tensor
()
{}
explicit
LoDTensor
(
const
LoD
&
lod
)
:
lod_
(
lod
)
{}
...
...
@@ -139,7 +139,7 @@ public:
return
(
lod_
)[
level
].
size
()
-
1
;
}
private:
private:
LoD
lod_
;
};
...
...
@@ -189,9 +189,8 @@ LoDTensor LodExpand(const LoDTensor &source, const LoD &lod, size_t level) {
// Returns:
// LoD = [[1, 4], [2, 4, 2, 3, 2]]
// pair<size_t, size_t> = {11, 24}
std
::
pair
<
LoD
,
std
::
pair
<
size_t
,
size_t
>>
GetSubLoDAndAbsoluteOffset
(
const
LoD
&
lod
,
size_t
start_idx
,
size_t
end_idx
,
size_t
start_level
);
std
::
pair
<
LoD
,
std
::
pair
<
size_t
,
size_t
>>
GetSubLoDAndAbsoluteOffset
(
const
LoD
&
lod
,
size_t
start_idx
,
size_t
end_idx
,
size_t
start_level
);
void
AppendLoD
(
LoD
*
lod
,
const
LoD
&
lod_length
);
...
...
@@ -204,5 +203,5 @@ void SerializeToStream(std::ostream &os, const LoDTensor &tensor);
void
DeserializeFromStream
(
std
::
istream
&
is
,
LoDTensor
*
tensor
);
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/op_desc.cpp
浏览文件 @
d6825bf3
...
...
@@ -55,5 +55,5 @@ const std::unordered_map<std::string, Attribute> &OpDesc::GetAttrMap() const {
return
attrs_
;
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/op_desc.h
浏览文件 @
d6825bf3
...
...
@@ -26,7 +26,7 @@ namespace paddle_mobile {
namespace
framework
{
class
OpDesc
:
PaddleMobileObject
{
public:
public:
OpDesc
(
const
proto
::
OpDesc
&
desc
);
const
std
::
vector
<
std
::
string
>
&
Input
(
const
std
::
string
&
name
)
const
;
const
std
::
vector
<
std
::
string
>
&
Output
(
const
std
::
string
&
name
)
const
;
...
...
@@ -40,12 +40,12 @@ public:
const
std
::
string
&
Type
()
{
return
desc_
.
type
();
};
private:
private:
proto
::
OpDesc
desc_
;
VariableNameMap
inputs_
;
VariableNameMap
outputs_
;
AttributeMap
attrs_
;
};
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/op_info.h
浏览文件 @
d6825bf3
...
...
@@ -24,7 +24,8 @@ SOFTWARE.
namespace
paddle_mobile
{
namespace
framework
{
template
<
typename
Dtype
>
struct
OpInfo
{
template
<
typename
Dtype
>
struct
OpInfo
{
OpCreator
<
Dtype
>
creator_
;
const
OpCreator
<
Dtype
>
&
Creator
()
const
{
// PADDLE_ENFORCE_NOT_NULL(creator_,
...
...
@@ -34,12 +35,15 @@ template <typename Dtype> struct OpInfo {
}
};
template
<
typename
Dtype
>
class
OpInfoMap
;
template
<
typename
Dtype
>
class
OpInfoMap
;
template
<
typename
Dtype
>
static
OpInfoMap
<
Dtype
>
*
g_op_info_map
=
nullptr
;
template
<
typename
Dtype
>
static
OpInfoMap
<
Dtype
>
*
g_op_info_map
=
nullptr
;
template
<
typename
Dtype
>
class
OpInfoMap
{
public:
template
<
typename
Dtype
>
class
OpInfoMap
{
public:
static
OpInfoMap
&
Instance
()
{
if
(
g_op_info_map
<
Dtype
>
==
nullptr
)
{
g_op_info_map
<
Dtype
>
=
new
OpInfoMap
();
...
...
@@ -83,12 +87,12 @@ public:
return
&
map_
;
}
private:
private:
OpInfoMap
()
=
default
;
std
::
unordered_map
<
std
::
string
,
OpInfo
<
Dtype
>>
map_
;
// DISABLE_COPY_AND_ASSIGN(OpInfoMap);
};
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/op_kernel_type.h
浏览文件 @
d6825bf3
...
...
@@ -60,5 +60,5 @@ inline bool TransFromNeeded(const OpKernelType &l, const OpKernelType &r) {
NeedTransformLayout
(
l
.
data_layout_
,
r
.
data_layout_
);
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/op_proto_maker.h
浏览文件 @
d6825bf3
...
...
@@ -22,5 +22,5 @@ namespace paddle_mobile {
namespace
framework
{
// this class not only make proto but also init attribute checkers.
class
OpProtoAndCheckerMaker
{};
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/operator.cpp
浏览文件 @
d6825bf3
...
...
@@ -28,7 +28,10 @@ OperatorBase<Dtype>::OperatorBase(const std::string &type,
const
VariableNameMap
&
outputs
,
const
AttributeMap
&
attrs
,
std
::
shared_ptr
<
Scope
>
scope
)
:
type_
(
type
),
inputs_
(
inputs
),
outputs_
(
outputs
),
attrs_
(
attrs
),
:
type_
(
type
),
inputs_
(
inputs
),
outputs_
(
outputs
),
attrs_
(
attrs
),
scope_
(
scope
)
{
CheckAllInputOutputSet
();
}
...
...
@@ -38,5 +41,5 @@ void OperatorBase<Dtype>::CheckAllInputOutputSet() const {}
template
class
OperatorBase
<
CPU
>;
template
class
OperatorWithKernel
<
CPU
>;
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/operator.h
浏览文件 @
d6825bf3
...
...
@@ -48,8 +48,9 @@ static std::unordered_map<
{
"feed"
,
{{
"X"
},
{
"Out"
}}},
{
"fetch"
,
{{
"X"
},
{
"Out"
}}}};
template
<
typename
Dtype
>
class
OperatorBase
:
PaddleMobileObject
{
public:
template
<
typename
Dtype
>
class
OperatorBase
:
PaddleMobileObject
{
public:
OperatorBase
(
const
std
::
string
&
type
,
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
AttributeMap
&
attrs
,
std
::
shared_ptr
<
Scope
>
scope
);
...
...
@@ -66,20 +67,20 @@ public:
}
}
protected:
protected:
std
::
shared_ptr
<
Scope
>
scope_
;
std
::
string
type_
;
VariableNameMap
inputs_
;
VariableNameMap
outputs_
;
AttributeMap
attrs_
;
private:
private:
void
CheckAllInputOutputSet
()
const
;
};
template
<
typename
Dtype
>
class
OperatorWithKernel
:
public
OperatorBase
<
Dtype
>
{
public:
public:
OperatorWithKernel
(
const
std
::
string
&
type
,
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
AttributeMap
&
attrs
,
std
::
shared_ptr
<
Scope
>
scope
)
...
...
@@ -88,12 +89,13 @@ public:
virtual
void
Run
()
const
=
0
;
};
template
<
typename
Dtype
,
typename
P
>
class
OpKernelBase
:
PaddleMobileObject
{
public:
template
<
typename
Dtype
,
typename
P
>
class
OpKernelBase
:
PaddleMobileObject
{
public:
virtual
void
Compute
(
const
P
&
para
)
const
=
0
;
virtual
~
OpKernelBase
()
=
default
;
};
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/paddle_mobile_object.h
浏览文件 @
d6825bf3
...
...
@@ -18,19 +18,19 @@ SOFTWARE.
#pragma once
#include "stdio.h"
#include <string>
#include "stdio.h"
namespace
paddle_mobile
{
class
PaddleMobileObject
{
public:
public:
virtual
std
::
string
ToString
()
{
char
address
[
128
]
=
{
0
};
sprintf
(
address
,
"%p"
,
this
);
return
std
::
string
(
address
);
}
private:
private:
};
}
// namespace paddle_mobile
}
// namespace paddle_mobile
src/framework/program-optimize/node.cpp
浏览文件 @
d6825bf3
...
...
@@ -92,5 +92,5 @@ Print &operator<<(Print &printer, const Node &node) {
return
printer
;
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/program-optimize/node.h
浏览文件 @
d6825bf3
...
...
@@ -29,7 +29,7 @@ namespace paddle_mobile {
namespace
framework
{
class
Node
:
PaddleMobileObject
{
public:
public:
Node
(
const
std
::
string
&
type
)
:
type_
(
type
)
{}
Node
(
std
::
shared_ptr
<
OpDesc
>
op_desc
)
:
op_desc_
(
op_desc
),
type_
(
op_desc
->
Type
()){};
...
...
@@ -39,7 +39,7 @@ public:
Node
&
To
(
int
index
);
uint
depth
(
uint
begin
=
0
);
private:
private:
std
::
shared_ptr
<
OpDesc
>
op_desc_
;
std
::
string
ToString
(
std
::
string
blank
,
const
Node
*
node
)
const
;
std
::
vector
<
std
::
shared_ptr
<
Node
>>
outputs_
;
...
...
@@ -48,5 +48,5 @@ private:
};
Print
&
operator
<<
(
Print
&
printer
,
const
Node
&
node
);
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/program-optimize/program_optimize.cpp
浏览文件 @
d6825bf3
...
...
@@ -24,8 +24,8 @@ namespace framework {
std
::
shared_ptr
<
ProgramDesc
>
ProgramOptimize
::
Optimize
()
{}
std
::
shared_ptr
<
ProgramDesc
>
ProgramOptimize
::
FushionOptimize
(
std
::
shared_ptr
<
ProgramDesc
>
ori_des
)
{
std
::
shared_ptr
<
ProgramDesc
>
ProgramOptimize
::
FushionOptimize
(
std
::
shared_ptr
<
ProgramDesc
>
ori_des
)
{
for
(
int
i
=
0
;
i
<
ori_des
->
Blocks
().
size
();
++
i
)
{
std
::
unordered_map
<
std
::
string
,
std
::
shared_ptr
<
Node
>>
output_nodes
;
std
::
shared_ptr
<
Node
>
begin_node
;
...
...
@@ -69,5 +69,5 @@ ProgramOptimize::FushionOptimize(std::shared_ptr<ProgramDesc> ori_des) {
}
return
ori_des
;
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/program-optimize/program_optimize.h
浏览文件 @
d6825bf3
...
...
@@ -26,16 +26,16 @@ namespace paddle_mobile {
namespace
framework
{
class
ProgramOptimize
{
public:
public:
ProgramOptimize
()
{}
std
::
shared_ptr
<
ProgramDesc
>
Optimize
();
std
::
shared_ptr
<
ProgramDesc
>
FushionOptimize
(
std
::
shared_ptr
<
ProgramDesc
>
ori_des
);
std
::
shared_ptr
<
ProgramDesc
>
FushionOptimize
(
std
::
shared_ptr
<
ProgramDesc
>
ori_des
);
private:
private:
// std::shared_ptr<ProgramDesc> ori_desc_;
std
::
vector
<
std
::
unordered_map
<
std
::
string
,
std
::
shared_ptr
<
Node
>>>
outputs_nodes_
;
};
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/program.cpp
浏览文件 @
d6825bf3
...
...
@@ -18,4 +18,4 @@ SOFTWARE.
namespace
paddle_mobile
{
namespace
framework
{}
}
// namespace paddle_mobile
}
// namespace paddle_mobile
src/framework/program.h
浏览文件 @
d6825bf3
...
...
@@ -28,13 +28,13 @@ namespace framework {
template
<
typename
Dtype
,
Precision
P
=
Precision
::
FP32
>
class
Program
:
PaddleMobileObject
{
public:
public:
std
::
shared_ptr
<
ProgramDesc
>
originProgram
;
std
::
shared_ptr
<
ProgramDesc
>
optimizeProgram
;
std
::
shared_ptr
<
Scope
>
scope
;
private:
private:
};
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/program_desc.cpp
浏览文件 @
d6825bf3
...
...
@@ -18,5 +18,5 @@ std::shared_ptr<BlockDesc> ProgramDesc::Block(size_t idx) {
return
blocks_
[
idx
];
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/program_desc.h
浏览文件 @
d6825bf3
...
...
@@ -28,15 +28,15 @@ namespace paddle_mobile {
namespace
framework
{
class
ProgramDesc
:
PaddleMobileObject
{
public:
public:
ProgramDesc
(
const
proto
::
ProgramDesc
&
desc
);
std
::
shared_ptr
<
BlockDesc
>
Block
(
size_t
idx
);
const
std
::
vector
<
std
::
shared_ptr
<
BlockDesc
>>
&
Blocks
()
{
return
blocks_
;
};
private:
private:
std
::
vector
<
std
::
shared_ptr
<
BlockDesc
>>
blocks_
;
proto
::
ProgramDesc
desc_
;
};
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/scope.cc
浏览文件 @
d6825bf3
...
...
@@ -115,5 +115,5 @@ Variable *Scope::FindVarLocally(const std::string &name) const {
return
nullptr
;
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/scope.h
浏览文件 @
d6825bf3
...
...
@@ -18,15 +18,15 @@ SOFTWARE.
==============================================================================*/
#pragma once
#include <list> //std::list
#include <mutex> //std::mutex
#include <unordered_map> //std::unordered_map
#include "variable.h"
#include <list> //std::list
#include <mutex> //std::mutex
#include <unordered_map> //std::unordered_map
namespace
paddle_mobile
{
namespace
framework
{
class
Scope
{
public:
public:
Scope
()
{}
~
Scope
()
{}
...
...
@@ -67,7 +67,7 @@ public:
Variable
*
FindVarLocally
(
const
std
::
string
&
name
)
const
;
private:
private:
// Call Scope::NewScope for a sub-scope.
explicit
Scope
(
Scope
const
*
parent
)
:
parent_
(
parent
)
{}
...
...
@@ -77,5 +77,5 @@ private:
mutable
std
::
mutex
mutex_
;
};
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/selected_rows.h
浏览文件 @
d6825bf3
...
...
@@ -27,7 +27,7 @@ namespace paddle_mobile {
namespace
framework
{
class
SelectedRows
{
public:
public:
SelectedRows
(
const
std
::
vector
<
int64_t
>
&
rows
,
const
int64_t
&
height
)
:
rows_
(
rows
),
height_
(
height
)
{
value_
.
reset
(
new
Tensor
());
...
...
@@ -67,7 +67,7 @@ public:
return
make_ddim
(
dims
);
}
private:
private:
// Notice: rows can be duplicate. We can have {0, 4, 7, 0, 5, 7, 9}
// here.
// SelectedRows are simply concated when adding together. Until a
...
...
@@ -77,5 +77,5 @@ private:
int64_t
height_
;
};
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/tensor.h
浏览文件 @
d6825bf3
...
...
@@ -26,9 +26,11 @@ limitations under the License. */
namespace
paddle_mobile
{
namespace
framework
{
template
<
typename
...
T
>
struct
SizeOfTypeFunctor
;
template
<
typename
...
T
>
struct
SizeOfTypeFunctor
;
template
<
typename
T
>
struct
SizeOfTypeFunctor
<
T
>
{
template
<
typename
T
>
struct
SizeOfTypeFunctor
<
T
>
{
size_t
operator
()(
std
::
type_index
type
)
const
{
if
(
typeid
(
T
).
hash_code
()
==
type
.
hash_code
())
{
return
sizeof
(
T
);
...
...
@@ -38,7 +40,8 @@ template <typename T> struct SizeOfTypeFunctor<T> {
}
};
template
<
>
struct
SizeOfTypeFunctor
<>
{
template
<
>
struct
SizeOfTypeFunctor
<>
{
size_t
operator
()(
std
::
type_index
type
)
const
{
return
0UL
;
}
};
...
...
@@ -66,11 +69,12 @@ static inline size_t SizeOfType(std::type_index type) {
class
LoDTensor
;
class
Tensor
{
public:
public:
Tensor
()
:
offset_
(
0
)
{}
/*! Return a pointer to mutable memory block. */
template
<
typename
T
>
inline
T
*
data
()
{
template
<
typename
T
>
inline
T
*
data
()
{
check_memory_size
();
// PADDLE_ENFORCE(std::is_same<T, void>::value ||
// holder_->type().hash_code() ==
...
...
@@ -82,7 +86,8 @@ public:
}
/*! Return a pointer to constant memory block. */
template
<
typename
T
>
inline
const
T
*
data
()
const
{
template
<
typename
T
>
inline
const
T
*
data
()
const
{
check_memory_size
();
// PADDLE_ENFORCE(std::is_same<T, void>::value ||
// holder_->type().hash_code() ==
...
...
@@ -100,7 +105,8 @@ public:
* @brief Return a pointer to mutable memory block.
* @note If not exist, then allocation.
*/
template
<
typename
T
>
inline
T
*
mutable_data
()
{
template
<
typename
T
>
inline
T
*
mutable_data
()
{
static_assert
(
std
::
is_pod
<
T
>::
value
,
"T must be POD"
);
return
reinterpret_cast
<
T
*>
(
mutable_data
(
typeid
(
T
)));
}
...
...
@@ -141,7 +147,8 @@ public:
*
* @note If not exist, then allocation.
*/
template
<
typename
T
>
inline
T
*
mutable_data
(
DDim
dims
)
{
template
<
typename
T
>
inline
T
*
mutable_data
(
DDim
dims
)
{
static_assert
(
std
::
is_pod
<
T
>::
value
,
"T must be POD"
);
Resize
(
dims
);
return
mutable_data
<
T
>
();
...
...
@@ -235,7 +242,7 @@ public:
inline
void
set_layout
(
const
DataLayout
layout
)
{
layout_
=
layout
;
}
private:
private:
/**
* @note Placeholder hides type T, so it doesn't appear as a
* template
...
...
@@ -257,7 +264,8 @@ private:
PlaceholderImpl
(
size_t
size
,
std
::
type_index
type
)
:
ptr_
(
static_cast
<
uint8_t
*>
(
memory
::
Alloc
(
size
)),
memory
::
PODDeleter
<
uint8_t
>
()),
size_
(
size
),
type_
(
type
)
{
size_
(
size
),
type_
(
type
)
{
// PADDLE_ENFORCE_NOT_NULL(ptr_,
// "Insufficient %s
// memory to allocation.",
...
...
@@ -329,5 +337,5 @@ inline Tensor ReshapeToMatrix(const Tensor &src, int num_col_dims) {
return
res
;
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/tensor_util.cc
浏览文件 @
d6825bf3
...
...
@@ -51,7 +51,8 @@ void TensorCopySync(const Tensor &src, Tensor *dst) {
memory
::
Copy
(
dst_ptr
,
src_ptr
,
size
);
}
template
<
typename
Predicate
>
struct
AnyDTypeVisitor
{
template
<
typename
Predicate
>
struct
AnyDTypeVisitor
{
Predicate
predicate_
;
const
Tensor
&
tensor_
;
Tensor
*
out_
;
...
...
@@ -59,7 +60,8 @@ template <typename Predicate> struct AnyDTypeVisitor {
AnyDTypeVisitor
(
Predicate
predicate
,
const
Tensor
&
tensor
,
Tensor
*
out
)
:
predicate_
(
predicate
),
tensor_
(
tensor
),
out_
(
out
)
{}
template
<
typename
T
>
void
operator
()()
const
{
template
<
typename
T
>
void
operator
()()
const
{
// auto t = EigenVector<T>::Flatten(tensor_);
// auto o = EigenScalar<bool>::From(*out_);
// return any of predicate_(t) is true.
...
...
@@ -74,7 +76,8 @@ inline void AnyImpl(Predicate predicate, const Tensor &tensor,
AnyDTypeVisitor
<
Predicate
>
(
predicate
,
tensor
,
out
));
}
template
<
typename
Predicate
>
struct
AnyVisitor
{
template
<
typename
Predicate
>
struct
AnyVisitor
{
const
framework
::
Tensor
&
tensor_
;
Predicate
predicate_
;
...
...
@@ -130,11 +133,11 @@ bool TensorContainsInf(const framework::Tensor &tensor) {
}
void
TensorToStream
(
std
::
ostream
&
os
,
const
Tensor
&
tensor
)
{
{
// the 1st field, uint32_t version
{
// the 1st field, uint32_t version
constexpr
uint32_t
version
=
0
;
os
.
write
(
reinterpret_cast
<
const
char
*>
(
&
version
),
sizeof
(
version
));
}
{
// the 2nd field, tensor description
{
// the 2nd field, tensor description
// int32_t size
// void* protobuf message
proto
::
VarType
::
TensorDesc
desc
;
...
...
@@ -148,7 +151,7 @@ void TensorToStream(std::ostream &os, const Tensor &tensor) {
auto
out
=
desc
.
SerializeAsString
();
os
.
write
(
out
.
data
(),
size
);
}
{
// the 3rd field, tensor data
{
// the 3rd field, tensor data
uint64_t
size
=
tensor
.
memory_size
();
auto
*
data_ptr
=
tensor
.
data
<
void
>
();
// PADDLE_ENFORCE(size <
...
...
@@ -164,7 +167,8 @@ struct DeserializedDataFunctor {
DeserializedDataFunctor
(
void
**
buf
,
Tensor
*
tensor
)
:
buf_
(
buf
),
tensor_
(
tensor
)
{}
template
<
typename
T
>
void
operator
()()
{
template
<
typename
T
>
void
operator
()()
{
*
buf_
=
tensor_
->
mutable_data
<
T
>
();
}
...
...
@@ -177,7 +181,7 @@ void TensorFromStream(std::istream &is, framework::Tensor *tensor) {
is
.
read
(
reinterpret_cast
<
char
*>
(
&
version
),
sizeof
(
version
));
// PADDLE_ENFORCE_EQ(version, 0U, "Only version 0 is supported");
proto
::
VarType
::
TensorDesc
desc
;
{
// int32_t size
{
// int32_t size
// proto buffer
int32_t
size
;
is
.
read
(
reinterpret_cast
<
char
*>
(
&
size
),
sizeof
(
size
));
...
...
@@ -186,7 +190,7 @@ void TensorFromStream(std::istream &is, framework::Tensor *tensor) {
// PADDLE_ENFORCE(desc.ParseFromArray(buf.get(), size),
// "Cannot parse tensor desc");
}
{
// read tensor
{
// read tensor
std
::
vector
<
int64_t
>
dims
;
dims
.
reserve
(
static_cast
<
size_t
>
(
desc
.
dims
().
size
()));
std
::
copy
(
desc
.
dims
().
begin
(),
desc
.
dims
().
end
(),
std
::
back_inserter
(
dims
));
...
...
@@ -199,5 +203,5 @@ void TensorFromStream(std::istream &is, framework::Tensor *tensor) {
}
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/tensor_util.h
浏览文件 @
d6825bf3
...
...
@@ -13,11 +13,11 @@ See the License for the specific language governing permissions and
limitations under the License. */
#pragma once
#include <vector>
#include "framework.pb.h"
#include "memory/t_malloc.h"
#include "platform/data_type.h"
#include "tensor.h"
#include <vector>
namespace
paddle_mobile
{
namespace
framework
{
...
...
@@ -62,5 +62,5 @@ void TensorToVector(const Tensor &src, std::vector<T> *dst) {
memory
::
Copy
(
dst_ptr
,
src_ptr
,
size
);
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/var_desc.cpp
浏览文件 @
d6825bf3
...
...
@@ -24,5 +24,5 @@ namespace framework {
VarDesc
::
VarDesc
(
const
proto
::
VarDesc
&
desc
)
:
desc_
(
desc
)
{}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/var_desc.h
浏览文件 @
d6825bf3
...
...
@@ -25,7 +25,7 @@ namespace paddle_mobile {
namespace
framework
{
class
VarDesc
{
public:
public:
VarDesc
(
const
proto
::
VarDesc
&
desc
);
std
::
string
Name
()
const
{
return
desc_
.
name
();
}
...
...
@@ -36,33 +36,33 @@ public:
const
proto
::
VarType
::
ChannelDesc
&
channel_desc
()
const
{
switch
(
desc_
.
type
().
type
())
{
case
proto
::
VarType
::
CHANNEL
:
return
desc_
.
type
().
channel
();
default:
break
;
case
proto
::
VarType
::
CHANNEL
:
return
desc_
.
type
().
channel
();
default:
break
;
}
}
const
proto
::
VarType
::
TensorDesc
&
tensor_desc
()
const
{
switch
(
desc_
.
type
().
type
())
{
case
proto
::
VarType
::
SELECTED_ROWS
:
return
desc_
.
type
().
selected_rows
();
case
proto
::
VarType
::
LOD_TENSOR
:
return
desc_
.
type
().
lod_tensor
().
tensor
();
case
proto
::
VarType
::
LOD_TENSOR_ARRAY
:
return
desc_
.
type
().
tensor_array
().
tensor
();
default:
break
;
case
proto
::
VarType
::
SELECTED_ROWS
:
return
desc_
.
type
().
selected_rows
();
case
proto
::
VarType
::
LOD_TENSOR
:
return
desc_
.
type
().
lod_tensor
().
tensor
();
case
proto
::
VarType
::
LOD_TENSOR_ARRAY
:
return
desc_
.
type
().
tensor_array
().
tensor
();
default:
break
;
}
}
proto
::
VarType
::
Type
GetDataType
()
const
{
switch
(
desc_
.
type
().
type
())
{
case
proto
::
VarType
::
CHANNEL
:
return
channel_desc
().
data_type
();
break
;
default:
return
tensor_desc
().
data_type
();
case
proto
::
VarType
::
CHANNEL
:
return
channel_desc
().
data_type
();
break
;
default:
return
tensor_desc
().
data_type
();
}
}
...
...
@@ -80,9 +80,9 @@ public:
return
this
->
RepeatedToVector
(
tensor_desc
().
dims
());
}
private:
private:
proto
::
VarDesc
desc_
;
};
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/var_type.h
浏览文件 @
d6825bf3
...
...
@@ -35,5 +35,5 @@ inline proto::VarType::Type ToVarType(std::type_index type) {
}
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/framework/variable.h
浏览文件 @
d6825bf3
...
...
@@ -18,18 +18,19 @@ SOFTWARE.
==============================================================================*/
#pragma once
#include "paddle_mobile_object.h"
#include <iostream>
#include <memory>
#include <string>
#include <typeindex>
#include <typeinfo>
#include "paddle_mobile_object.h"
namespace
paddle_mobile
{
namespace
framework
{
class
Variable
:
public
PaddleMobileObject
{
public:
template
<
typename
T
>
const
T
*
Get
()
const
{
public:
template
<
typename
T
>
const
T
*
Get
()
const
{
return
static_cast
<
const
T
*>
(
holder_
->
Ptr
());
}
...
...
@@ -37,7 +38,8 @@ public:
const
std
::
string
*
Name
()
{
return
name_
;
}
template
<
typename
T
>
T
*
GetMutable
()
{
template
<
typename
T
>
T
*
GetMutable
()
{
if
(
!
IsType
<
T
>
())
{
if
(
*
Name
()
==
"pixel"
)
{
// std::cout << " reset " << *Name() <<
...
...
@@ -48,7 +50,8 @@ public:
return
static_cast
<
T
*>
(
holder_
->
Ptr
());
}
template
<
typename
T
>
bool
IsType
()
const
{
template
<
typename
T
>
bool
IsType
()
const
{
if
(
holder_
)
{
// printf("not null \n");
printf
(
" holder type : %s, this type %s
\n
"
,
holder_
->
Type
().
name
(),
...
...
@@ -67,7 +70,7 @@ public:
void
SetName
(
const
std
::
string
*
name
)
{
name_
=
name
;
}
private:
private:
struct
Placeholder
{
Placeholder
()
=
default
;
virtual
~
Placeholder
()
=
default
;
...
...
@@ -76,7 +79,8 @@ private:
virtual
void
*
Ptr
()
const
=
0
;
};
template
<
typename
T
>
struct
PlaceholderImp
:
public
Placeholder
{
template
<
typename
T
>
struct
PlaceholderImp
:
public
Placeholder
{
explicit
PlaceholderImp
(
T
*
ptr
)
:
ptr_
(
ptr
),
type_
(
typeid
(
T
))
{}
virtual
const
std
::
type_info
&
Type
()
const
{
return
type_
;
}
...
...
@@ -92,5 +96,5 @@ private:
friend
class
Scope
;
const
std
::
string
*
name_
;
};
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/io.cpp
浏览文件 @
d6825bf3
...
...
@@ -18,13 +18,13 @@ SOFTWARE.
#include <fstream>
#include "../src/io.h"
#include "common/log.h"
#include "framework/framework.pb.h"
#include "framework/lod_tensor.h"
#include "framework/program_desc.h"
#include "framework/scope.h"
#include "framework/tensor.h"
#include "io.h"
namespace
paddle_mobile
{
...
...
@@ -41,26 +41,20 @@ void ReadBinaryFile(const std::string &filename, std::string *contents) {
template
<
typename
Dtype
,
Precision
P
>
void
Loader
<
Dtype
,
P
>::
LoadVar
(
framework
::
LoDTensor
*
tensor
,
const
std
::
string
&
file_path
)
{
// LOG(kLOG_DEBUG) << " to load " << file_path;
// Log(kLOG_DEBUG) << "123";
std
::
ifstream
is
(
file_path
);
std
::
streampos
pos
=
is
.
tellg
();
// save current position
std
::
fpos
<
mbstate_t
>
pos
;
pos
=
is
.
tellg
();
// save current position
is
.
seekg
(
0
,
std
::
ios
::
end
);
// LOG(kLOG_DEBUG) << " file length = " << is.tellg();
is
.
seekg
(
pos
);
// restore saved position
is
.
seekg
(
pos
);
// restore saved position
// 1. version
uint32_t
version
;
is
.
read
(
reinterpret_cast
<
char
*>
(
&
version
),
sizeof
(
version
));
// LOG(kLOG_INFO) << " version: " << version;
// 2 Lod information
uint64_t
lod_level
;
is
.
read
(
reinterpret_cast
<
char
*>
(
&
lod_level
),
sizeof
(
lod_level
));
// LOG(kLOG_DEBUG) << " load level: " << lod_level;
// LOG(kLOG_DEBUG) << " lod info: ";
auto
&
lod
=
*
tensor
->
mutable_lod
();
lod
.
resize
(
lod_level
);
for
(
uint64_t
i
=
0
;
i
<
lod_level
;
++
i
)
{
...
...
@@ -69,8 +63,8 @@ void Loader<Dtype, P>::LoadVar(framework::LoDTensor *tensor,
std
::
vector
<
size_t
>
tmp
(
size
/
sizeof
(
size_t
));
is
.
read
(
reinterpret_cast
<
char
*>
(
tmp
.
data
()),
static_cast
<
std
::
streamsize
>
(
size
));
for
(
int
j
=
0
;
j
<
tmp
.
size
();
++
j
)
{
LOG
(
kLOG_DEBUG1
)
<<
" lod - "
<<
tmp
[
j
]
;
for
(
auto
j
:
tmp
)
{
LOG
(
kLOG_DEBUG1
)
<<
" lod - "
<<
j
;
}
lod
[
i
]
=
tmp
;
}
...
...
@@ -78,26 +72,19 @@ void Loader<Dtype, P>::LoadVar(framework::LoDTensor *tensor,
// 3. tensor version
uint32_t
tensor_version
;
is
.
read
(
reinterpret_cast
<
char
*>
(
&
tensor_version
),
sizeof
(
tensor_version
));
// std::cout << " tensor_version: " << tensor_version << std::endl;
// 4. tensor desc
int32_t
size
;
is
.
read
(
reinterpret_cast
<
char
*>
(
&
size
),
sizeof
(
size
));
// std::cout << " tensor desc size: " << size << std::endl;
std
::
unique_ptr
<
char
[]
>
buf
(
new
char
[
size
]);
is
.
read
(
reinterpret_cast
<
char
*>
(
buf
.
get
()),
size
);
framework
::
proto
::
VarType
::
TensorDesc
desc
;
desc
.
ParseFromArray
(
buf
.
get
(),
size
);
// std::cout << " desc dims size " << desc.dims().size() <<
// std::endl;
int
memory_size
=
1
;
for
(
int
l
=
0
;
l
<
desc
.
dims
().
size
();
++
l
)
{
// std::cout << " dim " << l << " value: " << desc.dims()[l]
// <<
// std::endl;
memory_size
*=
desc
.
dims
()[
l
];
for
(
auto
l
:
desc
.
dims
())
{
memory_size
*=
l
;
}
std
::
vector
<
int64_t
>
dims
;
...
...
@@ -105,50 +92,39 @@ void Loader<Dtype, P>::LoadVar(framework::LoDTensor *tensor,
std
::
copy
(
desc
.
dims
().
begin
(),
desc
.
dims
().
end
(),
std
::
back_inserter
(
dims
));
tensor
->
Resize
(
framework
::
make_ddim
(
dims
));
void
*
memory
;
void
*
memory
=
tensor
;
int
type_size
=
0
;
// std::cout << " desc pre type: ";
switch
(
desc
.
data_type
())
{
case
framework
::
proto
::
VarType
::
FP16
:
// std::cout << "FP16" << std::endl;
type_size
=
2
;
break
;
case
framework
::
proto
::
VarType
::
FP32
:
type_size
=
4
;
memory
=
tensor
->
mutable_data
<
float
>
();
// std::cout << "FP32" << std::endl;
break
;
case
framework
::
proto
::
VarType
::
FP64
:
type_size
=
8
;
// std::cout << "FP64" << std::endl;
break
;
case
framework
::
proto
::
VarType
::
INT32
:
type_size
=
4
;
// std::cout << "INT32" << std::endl;
break
;
case
framework
::
proto
::
VarType
::
INT64
:
type_size
=
8
;
// std::cout << "INT64" << std::endl;
break
;
case
framework
::
proto
::
VarType
::
BOOL
:
type_size
=
1
;
// std::cout << "BOOL" << std::endl;
break
;
default:
break
;
// std::cout << " not support" << std::endl;
case
framework
::
proto
::
VarType
::
FP16
:
type_size
=
2
;
break
;
case
framework
::
proto
::
VarType
::
FP32
:
type_size
=
4
;
memory
=
tensor
->
mutable_data
<
float
>
();
break
;
case
framework
::
proto
::
VarType
::
FP64
:
type_size
=
8
;
break
;
case
framework
::
proto
::
VarType
::
INT32
:
type_size
=
4
;
break
;
case
framework
::
proto
::
VarType
::
INT64
:
type_size
=
8
;
break
;
case
framework
::
proto
::
VarType
::
BOOL
:
type_size
=
1
;
break
;
default:
break
;
}
// std::cout << " malloc size: " << memory_size * type_size <<
// std::endl;
is
.
read
(
static_cast
<
char
*>
(
memory
),
memory_size
*
type_size
);
// std::cout << " memory: " << memory << std::endl;
is
.
close
();
}
;
}
template
<
typename
Dtype
,
Precision
P
>
const
framework
::
Program
<
Dtype
,
P
>
Loader
<
Dtype
,
P
>::
Load
(
const
std
::
string
&
dirname
)
{
const
framework
::
Program
<
Dtype
,
P
>
Loader
<
Dtype
,
P
>::
Load
(
const
std
::
string
&
dirname
)
{
std
::
string
model_filename
=
dirname
+
"/__model__"
;
std
::
string
program_desc_str
;
ReadBinaryFile
(
model_filename
,
&
program_desc_str
);
...
...
@@ -165,10 +141,9 @@ Loader<Dtype, P>::Load(const std::string &dirname) {
std
::
make_shared
<
framework
::
Scope
>
();
program
.
scope
=
scope
;
auto
block
=
originProgramDesc
->
Block
(
0
);
originProgramDesc
->
Block
(
0
);
for
(
auto
block
:
originProgramDesc
->
Blocks
())
{
// std::cout << "for block" << std::endl;
for
(
const
auto
&
block
:
originProgramDesc
->
Blocks
())
{
for
(
int
i
=
0
;
i
<
block
->
Vars
().
size
();
++
i
)
{
std
::
shared_ptr
<
framework
::
VarDesc
>
var_desc
=
block
->
Vars
()[
i
];
auto
var
=
scope
->
Var
(
var_desc
->
Name
());
...
...
@@ -176,20 +151,18 @@ Loader<Dtype, P>::Load(const std::string &dirname) {
if
(
var_desc
->
Persistable
()
&&
var_desc
->
GetType
()
!=
framework
::
proto
::
VarType
::
FEED_MINIBATCH
&&
var_desc
->
GetType
()
!=
framework
::
proto
::
VarType
::
FETCH_LIST
)
{
framework
::
LoDTensor
*
tensor
=
var
->
GetMutable
<
framework
::
LoDTensor
>
();
auto
tensor
=
var
->
GetMutable
<
framework
::
LoDTensor
>
();
// to load
LoadVar
(
tensor
,
dirname
+
"/"
+
var_desc
->
Name
());
}
}
else
{
//
std::cout << "非 lod" << std::endl;
//
TODO by someone
}
}
}
#ifdef PADDLE_MOBILE_DEBUG
for
(
int
i
=
0
;
i
<
program_desc_proto
.
blocks
().
size
();
++
i
)
{
framework
::
proto
::
BlockDesc
block
=
program_desc_proto
.
blocks
()[
i
];
for
(
const
auto
&
block
:
program_desc_proto
.
blocks
())
{
LOG
(
kLOG_DEBUG
)
<<
"block: "
<<
block
.
idx
();
for
(
int
j
=
0
;
j
<
block
.
ops
().
size
();
++
j
)
{
if
(
j
==
2
)
{
...
...
@@ -200,65 +173,64 @@ Loader<Dtype, P>::Load(const std::string &dirname) {
for
(
int
m
=
0
;
m
<
op
.
inputs_size
();
++
m
)
{
const
framework
::
proto
::
OpDesc
::
Var
&
var
=
op
.
inputs
(
m
);
LOG
(
kLOG_DEBUG2
)
<<
"input parameter: "
<<
var
.
parameter
();
for
(
int
n
=
0
;
n
<
var
.
arguments
().
size
();
++
n
)
{
LOG
(
kLOG_DEBUG3
)
<<
"argument - "
<<
var
.
arguments
()[
n
]
;
for
(
const
auto
&
n
:
var
.
arguments
()
)
{
LOG
(
kLOG_DEBUG3
)
<<
"argument - "
<<
n
;
}
}
for
(
int
y
=
0
;
y
<
op
.
outputs_size
();
++
y
)
{
const
framework
::
proto
::
OpDesc
::
Var
&
var
=
op
.
outputs
(
y
);
LOG
(
kLOG_DEBUG2
)
<<
"out parameter: "
<<
var
.
parameter
();
for
(
int
z
=
0
;
z
<
var
.
arguments
().
size
();
++
z
)
{
LOG
(
kLOG_DEBUG3
)
<<
"argument - "
<<
var
.
arguments
()[
z
]
;
for
(
const
auto
&
z
:
var
.
arguments
()
)
{
LOG
(
kLOG_DEBUG3
)
<<
"argument - "
<<
z
;
}
}
for
(
int
x
=
0
;
x
<
op
.
attrs
().
size
();
++
x
)
{
const
framework
::
proto
::
OpDesc_Attr
attr
=
op
.
attrs
()[
x
];
for
(
const
auto
&
attr
:
op
.
attrs
())
{
LOG
(
kLOG_DEBUG2
)
<<
"attr name: "
<<
attr
.
name
();
switch
(
attr
.
type
())
{
case
framework
::
proto
::
AttrType
::
BOOLEAN
:
LOG
(
kLOG_DEBUG3
)
<<
"boolen: "
<<
attr
.
b
();
break
;
case
framework
::
proto
::
AttrType
::
INT
:
LOG
(
kLOG_DEBUG3
)
<<
"int: "
<<
attr
.
i
();
break
;
case
framework
::
proto
::
AttrType
::
FLOAT
:
LOG
(
kLOG_DEBUG3
)
<<
"float: "
<<
attr
.
f
();
case
framework
::
proto
::
AttrType
::
STRING
:
LOG
(
kLOG_DEBUG3
)
<<
"string: "
<<
attr
.
s
();
case
framework
::
proto
::
AttrType
::
BOOLEANS
:
for
(
int
y
=
0
;
y
<
attr
.
bools_size
();
++
y
)
{
LOG
(
kLOG_DEBUG3
)
<<
"bools: "
<<
attr
.
bools
(
y
);
}
case
framework
::
proto
::
AttrType
::
LONG
:
LOG
(
kLOG_DEBUG3
)
<<
"long: "
<<
attr
.
l
();
case
framework
::
proto
::
AttrType
::
FLOATS
:
for
(
int
y
=
0
;
y
<
attr
.
floats_size
();
++
y
)
{
LOG
(
kLOG_DEBUG3
)
<<
"floats: "
<<
attr
.
floats
(
y
);
}
case
framework
::
proto
::
AttrType
::
INTS
:
for
(
int
y
=
0
;
y
<
attr
.
ints_size
();
++
y
)
{
LOG
(
kLOG_DEBUG3
)
<<
"ints: "
<<
attr
.
ints
(
y
);
}
case
framework
::
proto
::
AttrType
::
STRINGS
:
for
(
int
y
=
0
;
y
<
attr
.
strings_size
();
++
y
)
{
LOG
(
kLOG_DEBUG3
)
<<
"strings: "
<<
attr
.
strings
(
y
);
}
case
framework
::
proto
::
AttrType
::
BOOLEAN
:
LOG
(
kLOG_DEBUG3
)
<<
"boolen: "
<<
attr
.
b
();
break
;
case
framework
::
proto
::
AttrType
::
INT
:
LOG
(
kLOG_DEBUG3
)
<<
"int: "
<<
attr
.
i
();
break
;
case
framework
::
proto
::
AttrType
::
FLOAT
:
LOG
(
kLOG_DEBUG3
)
<<
"float: "
<<
attr
.
f
();
case
framework
::
proto
::
AttrType
::
STRING
:
LOG
(
kLOG_DEBUG3
)
<<
"string: "
<<
attr
.
s
();
case
framework
::
proto
::
AttrType
::
BOOLEANS
:
for
(
int
y
=
0
;
y
<
attr
.
bools_size
();
++
y
)
{
LOG
(
kLOG_DEBUG3
)
<<
"bools: "
<<
attr
.
bools
(
y
);
}
case
framework
::
proto
::
AttrType
::
LONG
:
LOG
(
kLOG_DEBUG3
)
<<
"long: "
<<
attr
.
l
();
case
framework
::
proto
::
AttrType
::
FLOATS
:
for
(
int
y
=
0
;
y
<
attr
.
floats_size
();
++
y
)
{
LOG
(
kLOG_DEBUG3
)
<<
"floats: "
<<
attr
.
floats
(
y
);
}
case
framework
::
proto
::
AttrType
::
INTS
:
for
(
int
y
=
0
;
y
<
attr
.
ints_size
();
++
y
)
{
LOG
(
kLOG_DEBUG3
)
<<
"ints: "
<<
attr
.
ints
(
y
);
}
case
framework
::
proto
::
AttrType
::
STRINGS
:
for
(
int
y
=
0
;
y
<
attr
.
strings_size
();
++
y
)
{
LOG
(
kLOG_DEBUG3
)
<<
"strings: "
<<
attr
.
strings
(
y
);
}
case
framework
::
proto
::
BLOCK
:
break
;
}
}
}
for
(
int
k
=
0
;
k
<
block
.
vars
().
size
();
++
k
)
{
framework
::
proto
::
VarDesc
var
=
block
.
vars
()[
k
];
for
(
const
auto
&
var
:
block
.
vars
())
{
if
(
var
.
type
().
type
()
==
framework
::
proto
::
VarType
::
LOD_TENSOR
)
{
LOG
(
kLOG_DEBUG1
)
<<
"var name: "
<<
var
.
name
();
const
framework
::
proto
::
VarType
::
TensorDesc
&
tensor_desc
=
var
.
type
().
lod_tensor
().
tensor
();
LOG
(
kLOG_DEBUG2
)
<<
"in var tensor desc dims size: "
<<
tensor_desc
.
dims
().
size
();
int
memory_size
=
1
;
for
(
int
l
=
0
;
l
<
tensor_desc
.
dims
().
size
();
++
l
)
{
LOG
(
kLOG_DEBUG3
)
<<
"var tensor desc dim "
<<
l
<<
" value: "
<<
tensor_desc
.
dims
()[
l
];
...
...
@@ -268,28 +240,20 @@ Loader<Dtype, P>::Load(const std::string &dirname) {
if
(
var
.
persistable
()
&&
var
.
type
().
type
()
!=
framework
::
proto
::
VarType
::
FEED_MINIBATCH
&&
var
.
type
().
type
()
!=
framework
::
proto
::
VarType
::
FETCH_LIST
)
{
// std::cout << " to load " << var.name() <<
// std::endl;
std
::
string
file_path
=
dirname
+
"/"
+
var
.
name
();
std
::
ifstream
is
(
file_path
);
std
::
streampos
pos
=
is
.
tellg
();
// save current position
std
::
fpos
<
mbstate_t
>
pos
;
pos
=
is
.
tellg
();
// save current position
is
.
seekg
(
0
,
std
::
ios
::
end
);
// std::cout << " file length = " << is.tellg() <<
// std::endl;
is
.
seekg
(
pos
);
// restore saved position
is
.
seekg
(
pos
);
// restore saved position
// 1. version
uint32_t
version
;
is
.
read
(
reinterpret_cast
<
char
*>
(
&
version
),
sizeof
(
version
));
// std::cout << " version: " << version <<
// std::endl;
// 2 Lod information
uint64_t
lod_level
;
is
.
read
(
reinterpret_cast
<
char
*>
(
&
lod_level
),
sizeof
(
lod_level
));
// std::cout << " load level: " << lod_level <<
// std::endl;
// std::cout << " lod info: " << std::endl;
for
(
uint64_t
i
=
0
;
i
<
lod_level
;
++
i
)
{
uint64_t
size
;
is
.
read
(
reinterpret_cast
<
char
*>
(
&
size
),
sizeof
(
size
));
...
...
@@ -297,83 +261,53 @@ Loader<Dtype, P>::Load(const std::string &dirname) {
is
.
read
(
reinterpret_cast
<
char
*>
(
tmp
.
data
()),
static_cast
<
std
::
streamsize
>
(
size
));
for
(
int
j
=
0
;
j
<
tmp
.
size
();
++
j
)
{
// std::cout << " lod - " << tmp[j] <<
// std::endl;
}
}
uint32_t
tensor_version
;
is
.
read
(
reinterpret_cast
<
char
*>
(
&
version
),
sizeof
(
version
));
// std::cout << " tensor_version: " <<
// tensor_version <<
// std::endl;
int32_t
size
;
is
.
read
(
reinterpret_cast
<
char
*>
(
&
size
),
sizeof
(
size
));
// std::cout << " tensor desc size: " << size <<
// std::endl;
std
::
unique_ptr
<
char
[]
>
buf
(
new
char
[
size
]);
is
.
read
(
reinterpret_cast
<
char
*>
(
buf
.
get
()),
size
);
framework
::
proto
::
VarType
::
TensorDesc
desc
;
desc
.
ParseFromArray
(
buf
.
get
(),
size
);
// std::cout << " desc dims size " <<
// desc.dims().size() <<
// std::endl;
int
memory_size
=
1
;
for
(
int
l
=
0
;
l
<
desc
.
dims
().
size
();
++
l
)
{
// std::cout << " dim " << l << " value: "
// <<
// desc.dims()[l]
// << std::endl;
memory_size
*=
desc
.
dims
()[
l
];
for
(
long
long
l
:
desc
.
dims
())
{
memory_size
*=
l
;
}
int
type_size
=
0
;
// std::cout << " desc pre type: ";
switch
(
desc
.
data_type
())
{
case
framework
::
proto
::
VarType
::
FP16
:
// std::cout << "FP16" << std::endl;
type_size
=
2
;
break
;
case
framework
::
proto
::
VarType
::
FP32
:
type_size
=
4
;
// std::cout << "FP32" << std::endl;
break
;
case
framework
::
proto
::
VarType
::
FP64
:
type_size
=
8
;
// std::cout << "FP64" << std::endl;
break
;
case
framework
::
proto
::
VarType
::
INT32
:
type_size
=
4
;
// std::cout << "INT32" << std::endl;
break
;
case
framework
::
proto
::
VarType
::
INT64
:
type_size
=
8
;
// std::cout << "INT64" << std::endl;
break
;
case
framework
::
proto
::
VarType
::
BOOL
:
type_size
=
1
;
// std::cout << "BOOL" << std::endl;
break
;
default:
break
;
// std::cout << " not support" <<
// std::endl;
case
framework
::
proto
::
VarType
::
FP16
:
type_size
=
2
;
break
;
case
framework
::
proto
::
VarType
::
FP32
:
type_size
=
4
;
break
;
case
framework
::
proto
::
VarType
::
FP64
:
type_size
=
8
;
break
;
case
framework
::
proto
::
VarType
::
INT32
:
type_size
=
4
;
break
;
case
framework
::
proto
::
VarType
::
INT64
:
type_size
=
8
;
break
;
case
framework
::
proto
::
VarType
::
BOOL
:
type_size
=
1
;
break
;
default:
break
;
}
// std::cout << " malloc size: " << memory_size *
// type_size
// << std::endl;
void
*
memory
=
malloc
(
memory_size
*
type_size
);
is
.
read
(
static_cast
<
char
*>
(
memory
),
memory_size
*
type_size
);
// std::cout << " memory: " << memory <<
// std::endl;
is
.
close
();
}
else
{
// std::cout << " *not load "
// << " var : " << var.name() << std::endl;
// TODO
}
}
}
...
...
@@ -384,4 +318,4 @@ Loader<Dtype, P>::Load(const std::string &dirname) {
template
class
Loader
<
CPU
,
Precision
::
FP32
>;
}
// namespace paddle_mobile
}
// namespace paddle_mobile
src/io.h
浏览文件 @
d6825bf3
...
...
@@ -29,11 +29,11 @@ namespace paddle_mobile {
template
<
typename
Dtype
,
Precision
P
=
Precision
::
FP32
>
class
Loader
:
PaddleMobileObject
{
public:
public:
const
framework
::
Program
<
Dtype
,
P
>
Load
(
const
std
::
string
&
dirname
);
private:
private:
void
LoadVar
(
framework
::
LoDTensor
*
tensor
,
const
std
::
string
&
file_path
);
};
}
// namespace paddle_mobile
}
// namespace paddle_mobile
src/memory/t_malloc.cc
浏览文件 @
d6825bf3
...
...
@@ -47,5 +47,5 @@ void Free(void *ptr) {
}
}
}
// namespace memory
}
// namespace paddle_mobile
}
// namespace memory
}
// namespace paddle_mobile
src/memory/t_malloc.h
浏览文件 @
d6825bf3
...
...
@@ -37,10 +37,11 @@ void Free(void *ptr);
* std::unique_ptr<T> in tensor.h.
* static_cast
*/
template
<
typename
T
>
class
PODDeleter
{
template
<
typename
T
>
class
PODDeleter
{
static_assert
(
std
::
is_pod
<
T
>::
value
,
"T must be POD"
);
public:
public:
explicit
PODDeleter
(){};
void
operator
()(
T
*
ptr
)
{
Free
(
static_cast
<
void
*>
(
ptr
));
}
...
...
@@ -54,11 +55,12 @@ public:
* std::unique_ptr<T> in tensor.h.
* reinterpret_cast
*/
template
<
typename
T
>
class
PlainDeleter
{
public:
template
<
typename
T
>
class
PlainDeleter
{
public:
explicit
PlainDeleter
(){};
void
operator
()(
T
*
ptr
)
{
Free
(
reinterpret_cast
<
void
*>
(
ptr
));
}
};
}
// namespace memory
}
// namespace paddle_mobile
}
// namespace memory
}
// namespace paddle_mobile
src/operators/batchnorm_op.cpp
浏览文件 @
d6825bf3
...
...
@@ -27,5 +27,5 @@ void BatchNormOp<Dtype, T>::InferShape() const {
param_
.
OutputY
()
->
Resize
(
x_dims
);
}
template
class
BatchNormOp
<
CPU
,
float
>;
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/batchnorm_op.h
浏览文件 @
d6825bf3
...
...
@@ -27,7 +27,7 @@ using namespace framework;
template
<
typename
DeviceType
,
typename
T
>
class
BatchNormOp
:
public
framework
::
OperatorWithKernel
<
DeviceType
>
{
public:
public:
BatchNormOp
(
const
std
::
string
&
type
,
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
attrs
,
...
...
@@ -44,9 +44,9 @@ public:
using
framework
::
OperatorWithKernel
<
DeviceType
>::
OperatorWithKernel
;
void
InferShape
()
const
override
;
protected:
protected:
BatchNormParam
param_
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/concat_op.cpp
浏览文件 @
d6825bf3
...
...
@@ -60,5 +60,5 @@ void ConcatOp<Dtype, T>::InferShape() const {
}
template
class
ConcatOp
<
CPU
,
float
>;
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/concat_op.h
浏览文件 @
d6825bf3
...
...
@@ -26,7 +26,7 @@ using namespace framework;
template
<
typename
DeviceType
,
typename
T
>
class
ConcatOp
:
public
framework
::
OperatorWithKernel
<
DeviceType
>
{
public:
public:
ConcatOp
(
const
std
::
string
&
type
,
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
attrs
,
std
::
shared_ptr
<
framework
::
Scope
>
scope
)
...
...
@@ -42,9 +42,9 @@ public:
using
framework
::
OperatorWithKernel
<
DeviceType
>::
OperatorWithKernel
;
void
InferShape
()
const
override
;
protected:
protected:
ConcatParam
param_
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/conv_op.cpp
浏览文件 @
d6825bf3
...
...
@@ -71,5 +71,5 @@ void ConvOp<Dtype, T>::InferShape() const {
template
class
ConvOp
<
CPU
,
float
>;
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/conv_op.h
浏览文件 @
d6825bf3
...
...
@@ -28,7 +28,7 @@ using namespace framework;
template
<
typename
DeviceType
,
typename
T
>
class
ConvOp
:
public
framework
::
OperatorWithKernel
<
DeviceType
>
{
public:
public:
ConvOp
(
const
std
::
string
&
type
,
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
,
std
::
shared_ptr
<
framework
::
Scope
>
scope
)
...
...
@@ -45,9 +45,9 @@ public:
this
->
ClearVariables
({
"Filter"
,
"Input"
});
}
private:
private:
ConvParam
param_
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/elementwise_add_op.cpp
浏览文件 @
d6825bf3
...
...
@@ -27,5 +27,5 @@ void ElementwiseAddOp<Dtype, T>::InferShape() const {
param_
.
Out
()
->
Resize
(
x_dim
);
}
template
class
ElementwiseAddOp
<
CPU
,
float
>;
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/elementwise_add_op.h
浏览文件 @
d6825bf3
...
...
@@ -27,7 +27,7 @@ using namespace framework;
template
<
typename
DeviceType
,
typename
T
>
class
ElementwiseAddOp
:
public
framework
::
OperatorWithKernel
<
DeviceType
>
{
public:
public:
ElementwiseAddOp
(
const
std
::
string
&
type
,
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
attrs
,
...
...
@@ -44,8 +44,8 @@ public:
using
framework
::
OperatorWithKernel
<
DeviceType
>::
OperatorWithKernel
;
void
InferShape
()
const
override
;
protected:
protected:
ElementwiseAddParam
param_
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/kernel/arm/batchnorm_kernel.cpp
浏览文件 @
d6825bf3
...
...
@@ -88,5 +88,5 @@ void BatchNormKernel<CPU, float>::Compute(const BatchNormParam ¶m) const {
DLOG
<<
"new_bias_ptr : "
<<
new_bias_ptr
[
5
];
DLOG
<<
"out_ptr : "
<<
out_ptr
[
102
];
}
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/kernel/arm/concat_kernel.cpp
浏览文件 @
d6825bf3
...
...
@@ -18,8 +18,9 @@ limitations under the License. */
namespace
paddle_mobile
{
namespace
operators
{
template
<
typename
T
>
class
ConcatFunctor
{
public:
template
<
typename
T
>
class
ConcatFunctor
{
public:
void
operator
()(
const
std
::
vector
<
framework
::
Tensor
>
&
input
,
const
int
axis
,
framework
::
Tensor
*
output
)
{
size_t
num
=
input
.
size
();
...
...
@@ -112,5 +113,5 @@ void ConcatKernel<CPU, float>::Compute(const ConcatParam ¶m) const {
}
}
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/kernel/arm/conv_kernel.cpp
浏览文件 @
d6825bf3
...
...
@@ -34,7 +34,8 @@ bool IsExpand(const std::vector<int64_t> &filter_dim,
return
!
(
filter_1
&&
strides_1
&&
padding_0
&&
dilation_1
);
}
template
<
>
void
ConvKernel
<
CPU
,
float
>::
Compute
(
const
ConvParam
&
param
)
const
{
template
<
>
void
ConvKernel
<
CPU
,
float
>::
Compute
(
const
ConvParam
&
param
)
const
{
LOG
(
kLOG_DEBUG
)
<<
param
;
const
Tensor
*
input
=
param
.
Input
();
...
...
@@ -148,5 +149,5 @@ template <> void ConvKernel<CPU, float>::Compute(const ConvParam ¶m) const {
template
class
ConvKernel
<
CPU
,
float
>;
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/kernel/arm/elementwise_add_kernel.cpp
浏览文件 @
d6825bf3
...
...
@@ -19,7 +19,8 @@ limitations under the License. */
namespace
paddle_mobile
{
namespace
operators
{
template
<
typename
T
>
struct
AddFunctor
{
template
<
typename
T
>
struct
AddFunctor
{
inline
T
operator
()(
T
a
,
T
b
)
const
{
return
a
+
b
;
}
};
...
...
@@ -37,5 +38,5 @@ void ElementwiseAddKernel<CPU, float>::Compute(
template
class
ElementwiseAddKernel
<
CPU
,
float
>;
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/kernel/arm/lrn_kernel.cpp
浏览文件 @
d6825bf3
...
...
@@ -23,7 +23,8 @@ SOFTWARE.
namespace
paddle_mobile
{
namespace
operators
{
template
<
>
void
LrnKernel
<
CPU
,
float
>::
Compute
(
const
LrnParam
&
param
)
const
{
template
<
>
void
LrnKernel
<
CPU
,
float
>::
Compute
(
const
LrnParam
&
param
)
const
{
const
Tensor
*
input_x
=
param
.
InputX
();
auto
x_dims
=
input_x
->
dims
();
/// data_format = NCHW
...
...
@@ -43,5 +44,5 @@ template <> void LrnKernel<CPU, float>::Compute(const LrnParam ¶m) const {
template
class
LrnKernel
<
CPU
,
float
>;
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/kernel/arm/mul_kernel.cpp
浏览文件 @
d6825bf3
...
...
@@ -23,7 +23,8 @@ SOFTWARE.
namespace
paddle_mobile
{
namespace
operators
{
template
<
>
void
MulKernel
<
CPU
,
float
>::
Compute
(
const
MulParam
&
param
)
const
{
template
<
>
void
MulKernel
<
CPU
,
float
>::
Compute
(
const
MulParam
&
param
)
const
{
const
Tensor
*
input_x
=
param
.
InputX
();
const
Tensor
*
input_y
=
param
.
InputY
();
Tensor
*
out
=
param
.
Out
();
...
...
@@ -49,5 +50,5 @@ template <> void MulKernel<CPU, float>::Compute(const MulParam ¶m) const {
template
class
MulKernel
<
CPU
,
float
>;
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/kernel/arm/pool_kernel.cpp
浏览文件 @
d6825bf3
...
...
@@ -15,8 +15,8 @@ LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
SOFTWARE.
==============================================================================*/
#include "common/log.h"
#include <operators/kernel/pool_kernel.h>
#include "common/log.h"
namespace
paddle_mobile
{
namespace
operators
{
...
...
@@ -36,7 +36,8 @@ inline void PoolBasic(std::string pooling_type, std::vector<int> ksize,
}
}
template
<
>
void
PoolKernel
<
CPU
,
float
>::
Compute
(
const
PoolParam
&
param
)
const
{
template
<
>
void
PoolKernel
<
CPU
,
float
>::
Compute
(
const
PoolParam
&
param
)
const
{
const
Tensor
*
in_x
=
param
.
Input
();
Tensor
*
out
=
param
.
Output
();
std
::
string
pooling_type
=
param
.
PoolingType
();
...
...
@@ -73,5 +74,5 @@ template <> void PoolKernel<CPU, float>::Compute(const PoolParam ¶m) const {
// PoolBasic(pooling_type, ksize, strides, paddings, in_x, out);
// }
}
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/kernel/batchnorm_kernel.h
浏览文件 @
d6825bf3
...
...
@@ -28,9 +28,9 @@ using namespace framework;
template
<
typename
DeviceType
,
typename
T
>
class
BatchNormKernel
:
public
framework
::
OpKernelBase
<
DeviceType
,
BatchNormParam
>
{
public:
public:
void
Compute
(
const
BatchNormParam
&
param
)
const
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/kernel/concat_kernel.h
浏览文件 @
d6825bf3
...
...
@@ -26,9 +26,9 @@ using namespace framework;
template
<
typename
DeviceType
,
typename
T
>
class
ConcatKernel
:
public
framework
::
OpKernelBase
<
DeviceType
,
ConcatParam
>
{
public:
public:
void
Compute
(
const
ConcatParam
&
param
)
const
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/kernel/conv_kernel.h
浏览文件 @
d6825bf3
...
...
@@ -31,8 +31,8 @@ using namespace framework;
template
<
typename
DeviceType
,
typename
T
>
class
ConvKernel
:
public
framework
::
OpKernelBase
<
DeviceType
,
ConvParam
>
{
public:
public:
void
Compute
(
const
ConvParam
&
param
)
const
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/kernel/elementwise_add_kernel.h
浏览文件 @
d6825bf3
...
...
@@ -29,8 +29,8 @@ using namespace framework;
template
<
typename
DeviceType
,
typename
T
>
class
ElementwiseAddKernel
:
public
framework
::
OpKernelBase
<
DeviceType
,
ElementwiseAddParam
>
{
public:
public:
void
Compute
(
const
ElementwiseAddParam
&
param
)
const
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/kernel/fpga/conv_kernel.cpp
浏览文件 @
d6825bf3
...
...
@@ -25,4 +25,4 @@ namespace operators {
//
// template class ConvKernel<FPGA, float>;
}
}
// namespace paddle_mobile
}
// namespace paddle_mobile
src/operators/kernel/lrn_kernel.h
浏览文件 @
d6825bf3
...
...
@@ -25,7 +25,8 @@ namespace operators {
using
namespace
framework
;
template
<
typename
T
>
struct
LRNFunctor
{
template
<
typename
T
>
struct
LRNFunctor
{
void
operator
()(
const
framework
::
Tensor
&
input
,
framework
::
Tensor
*
out
,
int
N
,
int
C
,
int
H
,
int
W
,
int
n
,
T
k
,
T
alpha
,
T
beta
)
{
auto
input_ptr
=
input
.
data
<
T
>
();
...
...
@@ -67,8 +68,8 @@ template <typename T> struct LRNFunctor {
template
<
typename
DeviceType
,
typename
T
>
class
LrnKernel
:
public
framework
::
OpKernelBase
<
DeviceType
,
LrnParam
>
{
public:
public:
void
Compute
(
const
LrnParam
&
param
)
const
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/kernel/mul_kernel.h
浏览文件 @
d6825bf3
...
...
@@ -28,8 +28,8 @@ using namespace framework;
template
<
typename
DeviceType
,
typename
T
>
class
MulKernel
:
public
framework
::
OpKernelBase
<
DeviceType
,
MulParam
>
{
public:
public:
void
Compute
(
const
MulParam
&
param
)
const
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/kernel/pool_kernel.h
浏览文件 @
d6825bf3
...
...
@@ -28,8 +28,8 @@ using namespace framework;
template
<
typename
DeviceType
,
typename
T
>
class
PoolKernel
:
public
framework
::
OpKernelBase
<
DeviceType
,
PoolParam
>
{
public:
public:
void
Compute
(
const
PoolParam
&
param
)
const
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/lrn_op.cpp
浏览文件 @
d6825bf3
...
...
@@ -21,10 +21,11 @@ SOFTWARE.
namespace
paddle_mobile
{
namespace
operators
{
template
<
typename
Dtype
,
typename
T
>
void
LrnOp
<
Dtype
,
T
>::
InferShape
()
const
{
template
<
typename
Dtype
,
typename
T
>
void
LrnOp
<
Dtype
,
T
>::
InferShape
()
const
{
auto
x_dims
=
param_
.
InputX
()
->
dims
();
param_
.
Out
()
->
Resize
(
x_dims
);
}
template
class
LrnOp
<
CPU
,
float
>;
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/lrn_op.h
浏览文件 @
d6825bf3
...
...
@@ -27,7 +27,7 @@ using namespace framework;
template
<
typename
DeviceType
,
typename
T
>
class
LrnOp
:
public
framework
::
OperatorWithKernel
<
DeviceType
>
{
public:
public:
LrnOp
(
const
std
::
string
&
type
,
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
attrs
,
std
::
shared_ptr
<
framework
::
Scope
>
scope
)
...
...
@@ -43,9 +43,9 @@ public:
using
framework
::
OperatorWithKernel
<
DeviceType
>::
OperatorWithKernel
;
void
InferShape
()
const
override
;
protected:
protected:
LrnParam
param_
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/math/elementwise_op_function.h
浏览文件 @
d6825bf3
...
...
@@ -58,8 +58,7 @@ inline void trim_trailing_singular_dims(framework::DDim *dims) {
// Remove trailing dimensions of size 1 for y
auto
actual_dims_size
=
dims
->
size
();
for
(;
actual_dims_size
!=
0
;
--
actual_dims_size
)
{
if
((
*
dims
)[
actual_dims_size
-
1
]
!=
1
)
break
;
if
((
*
dims
)[
actual_dims_size
-
1
]
!=
1
)
break
;
}
if
(
actual_dims_size
!=
dims
->
size
())
{
auto
actual_dims
=
framework
::
vectorize
(
*
dims
);
...
...
@@ -68,8 +67,9 @@ inline void trim_trailing_singular_dims(framework::DDim *dims) {
}
}
template
<
typename
T
>
class
RowwiseTransformIterator
{
public:
template
<
typename
T
>
class
RowwiseTransformIterator
{
public:
RowwiseTransformIterator
(
const
T
*
ptr
,
int
n
)
:
ptr_
(
ptr
),
i_
(
0
),
n_
(
n
)
{}
RowwiseTransformIterator
<
T
>
&
operator
++
()
{
...
...
@@ -90,7 +90,7 @@ public:
const
T
&
operator
*
()
{
return
ptr_
[
i_
];
}
private:
private:
const
T
*
ptr_
;
int
i_
;
int64_t
n_
;
...
...
@@ -100,8 +100,9 @@ private:
/// dimension
/// in (4,20,2) is 2 ,
/// (20,1) move 1 stride , to fill(add) 2 element with the same number.
template
<
typename
T
>
class
MidWiseTransformIterator
{
public:
template
<
typename
T
>
class
MidWiseTransformIterator
{
public:
MidWiseTransformIterator
(
const
T
*
ptr
,
int
n
,
int
post
)
:
ptr_
(
ptr
),
i_
(
0
),
j_
(
0
),
n_
(
n
),
post_
(
post
)
{}
...
...
@@ -127,7 +128,7 @@ public:
const
T
&
operator
*
()
{
return
ptr_
[
i_
];
}
private:
private:
const
T
*
ptr_
;
int64_t
i_
;
int64_t
j_
;
...
...
@@ -137,11 +138,14 @@ private:
template
<
typename
Functor
,
typename
T
,
typename
OutType
=
T
>
class
TransformFunctor
{
public:
public:
TransformFunctor
(
const
framework
::
Tensor
*
x
,
const
framework
::
Tensor
*
y
,
framework
::
Tensor
*
z
,
Functor
func
)
:
x_
(
x
->
data
<
T
>
()),
y_
(
y
->
data
<
T
>
()),
z_
(
z
->
mutable_data
<
OutType
>
()),
nx_
(
x
->
numel
()),
func_
(
func
)
{}
:
x_
(
x
->
data
<
T
>
()),
y_
(
y
->
data
<
T
>
()),
z_
(
z
->
mutable_data
<
OutType
>
()),
nx_
(
x
->
numel
()),
func_
(
func
)
{}
inline
void
Run
()
const
{
math
::
Transform
trans
;
...
...
@@ -159,7 +163,7 @@ public:
trans
(
x_
,
x_
+
nx_
,
MidWiseTransformIterator
<
T
>
(
y_
,
n
,
post
),
z_
,
func_
);
}
private:
private:
const
T
*
x_
;
const
T
*
y_
;
OutType
*
z_
;
...
...
@@ -202,5 +206,5 @@ void ElementwiseComputeEx(const framework::Tensor *x,
}
}
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/math/im2col.cc
浏览文件 @
d6825bf3
...
...
@@ -25,8 +25,9 @@ namespace math {
* [input_channels, filter_height, filter_width, output_height,
* output_width]
*/
template
<
class
T
>
class
Im2ColFunctor
<
ColFormat
::
kCFO
,
CPU
,
T
>
{
public:
template
<
class
T
>
class
Im2ColFunctor
<
ColFormat
::
kCFO
,
CPU
,
T
>
{
public:
void
operator
()(
const
framework
::
Tensor
&
im
,
const
std
::
vector
<
int
>
&
dilation
,
const
std
::
vector
<
int
>
&
stride
,
const
std
::
vector
<
int
>
&
padding
,
framework
::
Tensor
*
col
)
{
...
...
@@ -93,8 +94,9 @@ public:
* [input_channels, filter_height, filter_width, output_height,
* output_width]
*/
template
<
class
T
>
class
Col2ImFunctor
<
ColFormat
::
kCFO
,
CPU
,
T
>
{
public:
template
<
class
T
>
class
Col2ImFunctor
<
ColFormat
::
kCFO
,
CPU
,
T
>
{
public:
void
operator
()(
const
framework
::
Tensor
&
col
,
const
std
::
vector
<
int
>
&
dilation
,
const
std
::
vector
<
int
>
&
stride
,
...
...
@@ -165,8 +167,9 @@ template class Col2ImFunctor<ColFormat::kCFO, CPU, double>;
* [output_height, output_width, input_channels, filter_height,
* filter_width]
*/
template
<
class
T
>
class
Im2ColFunctor
<
ColFormat
::
kOCF
,
CPU
,
T
>
{
public:
template
<
class
T
>
class
Im2ColFunctor
<
ColFormat
::
kOCF
,
CPU
,
T
>
{
public:
void
operator
()(
const
framework
::
Tensor
&
im
,
const
std
::
vector
<
int
>
&
dilation
,
const
std
::
vector
<
int
>
&
stride
,
const
std
::
vector
<
int
>
&
padding
,
framework
::
Tensor
*
col
)
{
...
...
@@ -237,8 +240,9 @@ public:
* [output_height, output_width, input_channels, filter_height,
* filter_width]
*/
template
<
class
T
>
class
Col2ImFunctor
<
ColFormat
::
kOCF
,
CPU
,
T
>
{
public:
template
<
class
T
>
class
Col2ImFunctor
<
ColFormat
::
kOCF
,
CPU
,
T
>
{
public:
void
operator
()(
const
framework
::
Tensor
&
col
,
const
std
::
vector
<
int
>
&
dilation
,
const
std
::
vector
<
int
>
&
stride
,
...
...
@@ -309,6 +313,6 @@ template class Im2ColFunctor<ColFormat::kOCF, CPU, double>;
template
class
Col2ImFunctor
<
ColFormat
::
kOCF
,
CPU
,
float
>;
template
class
Col2ImFunctor
<
ColFormat
::
kOCF
,
CPU
,
double
>;
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
src/operators/math/im2col.h
浏览文件 @
d6825bf3
...
...
@@ -89,7 +89,7 @@ enum class ColFormat { kCFO = 0, kOCF = 1 };
*/
template
<
ColFormat
Format
,
typename
DeviceType
,
typename
T
>
class
Im2ColFunctor
{
public:
public:
void
operator
()(
const
framework
::
Tensor
&
im
,
const
std
::
vector
<
int
>
&
dilation
,
const
std
::
vector
<
int
>
&
stride
,
const
std
::
vector
<
int
>
&
padding
,
framework
::
Tensor
*
col
);
...
...
@@ -97,13 +97,13 @@ public:
template
<
ColFormat
Format
,
typename
DeviceType
,
typename
T
>
class
Col2ImFunctor
{
public:
public:
void
operator
()(
const
framework
::
Tensor
&
col
,
const
std
::
vector
<
int
>
&
dilation
,
const
std
::
vector
<
int
>
&
stride
,
const
std
::
vector
<
int
>
&
padding
,
framework
::
Tensor
*
im
);
};
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
src/operators/math/math_function.cc
浏览文件 @
d6825bf3
...
...
@@ -119,6 +119,6 @@ void matmul<double>(const framework::Tensor &matrix_a, bool trans_a,
matrix_b
.
data
<
double
>
(),
beta
,
matrix_out
->
data
<
double
>
());
}
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
src/operators/math/math_function.h
浏览文件 @
d6825bf3
...
...
@@ -14,9 +14,9 @@ limitations under the License. */
#pragma once
#include "framework/tensor.h"
#include <cblas.h>
#include <cmath>
#include "framework/tensor.h"
namespace
paddle_mobile
{
namespace
operators
{
...
...
@@ -37,6 +37,6 @@ template <typename T>
void
matmul
(
const
framework
::
Tensor
&
matrix_a
,
bool
trans_a
,
const
framework
::
Tensor
&
matrix_b
,
bool
trans_b
,
T
alpha
,
framework
::
Tensor
*
matrix_out
,
T
beta
);
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
src/operators/math/pool3x3.h
浏览文件 @
d6825bf3
...
...
@@ -19,7 +19,7 @@ SOFTWARE.
#if __ARM_NEON
#include <arm_neon.h>
#endif // __ARM_NEON
#endif
// __ARM_NEON
static
void
Pool3x3Max
()
{
// todo impl with neon
...
...
src/operators/math/pool_2x2.h
浏览文件 @
d6825bf3
...
...
@@ -19,7 +19,7 @@ SOFTWARE.
#if __ARM_NEON
#include <arm_neon.h>
#endif // __ARM_NEON
#endif
// __ARM_NEON
static
void
Pool2x2Max
()
{
// todo impl with neon
...
...
src/operators/math/pooling.cpp
浏览文件 @
d6825bf3
...
...
@@ -30,12 +30,11 @@ namespace math {
*/
template
<
typename
PoolProcess
,
typename
T
>
class
PoolFunctor
<
CPU
,
PoolProcess
,
T
>
{
public:
public:
void
operator
()(
const
framework
::
Tensor
&
input
,
const
std
::
vector
<
int
>
&
ksize
,
const
std
::
vector
<
int
>
&
strides
,
const
std
::
vector
<
int
>
&
paddings
,
PoolProcess
pool_process
,
framework
::
Tensor
*
output
)
{
const
int
batch_size
=
input
.
dims
()[
0
];
const
int
input_height
=
input
.
dims
()[
2
];
...
...
@@ -93,6 +92,6 @@ public:
template
class
PoolFunctor
<
CPU
,
math
::
AvgPool
<
float
>,
float
>
;
template
class
PoolFunctor
<
CPU
,
math
::
MaxPool
<
float
>,
float
>
;
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
src/operators/math/pooling.h
浏览文件 @
d6825bf3
...
...
@@ -37,8 +37,9 @@ namespace math {
* in pool pooling, and finally takes the average.
* MaxPoolGrad and AvgPoolGrad are gradient operations respectively.
*/
template
<
class
T
>
class
MaxPool
{
public:
template
<
class
T
>
class
MaxPool
{
public:
inline
T
initial
()
{
return
static_cast
<
T
>
(
-
FLT_MAX
);
}
inline
void
compute
(
const
T
&
x
,
T
*
y
)
{
*
y
=
*
y
>
x
?
*
y
:
x
;
}
...
...
@@ -46,8 +47,9 @@ public:
inline
void
finalize
(
const
T
&
pool_field
,
T
*
y
)
{}
};
template
<
class
T
>
class
AvgPool
{
public:
template
<
class
T
>
class
AvgPool
{
public:
inline
T
initial
()
{
return
static_cast
<
T
>
(
0
);
}
inline
void
compute
(
const
T
&
x
,
T
*
y
)
{
*
y
+=
x
;
}
...
...
@@ -57,12 +59,12 @@ public:
template
<
typename
DeviceType
,
typename
PoolProcess
,
typename
T
>
class
PoolFunctor
{
public:
public:
void
operator
()(
const
framework
::
Tensor
&
input
,
const
std
::
vector
<
int
>
&
ksize
,
const
std
::
vector
<
int
>
&
strides
,
const
std
::
vector
<
int
>
&
paddings
,
PoolProcess
pool_compute
,
framework
::
Tensor
*
output
);
};
}
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/math/transform.h
浏览文件 @
d6825bf3
...
...
@@ -50,6 +50,6 @@ struct Transform {
std
::
transform
(
first1
,
last1
,
first2
,
result
,
op
);
}
};
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
src/operators/math/vol2col.cc
浏览文件 @
d6825bf3
...
...
@@ -25,8 +25,9 @@ using Tensor = paddle_mobile::framework::Tensor;
* [input_channels, filter_depth, filter_height, filter_width,
* output_depth, output_height, output_width]
*/
template
<
typename
T
>
class
Vol2ColFunctor
<
CPU
,
T
>
{
public:
template
<
typename
T
>
class
Vol2ColFunctor
<
CPU
,
T
>
{
public:
void
operator
()(
const
Tensor
&
vol
,
const
std
::
vector
<
int
>
&
dilations
,
const
std
::
vector
<
int
>
&
strides
,
const
std
::
vector
<
int
>
&
paddings
,
Tensor
*
col
)
const
{
...
...
@@ -111,8 +112,9 @@ public:
* [input_channels, filter_depth, filter_height, filter_width,
* output_depth, output_height, output_width]
*/
template
<
typename
T
>
class
Col2VolFunctor
<
CPU
,
T
>
{
public:
template
<
typename
T
>
class
Col2VolFunctor
<
CPU
,
T
>
{
public:
void
operator
()(
const
Tensor
&
col
,
const
std
::
vector
<
int
>
&
dilations
,
const
std
::
vector
<
int
>
&
strides
,
const
std
::
vector
<
int
>
&
paddings
,
Tensor
*
vol
)
const
{
...
...
@@ -196,6 +198,6 @@ template class Vol2ColFunctor<CPU, double>;
template
class
Col2VolFunctor
<
CPU
,
float
>;
template
class
Col2VolFunctor
<
CPU
,
double
>;
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
src/operators/math/vol2col.h
浏览文件 @
d6825bf3
...
...
@@ -72,20 +72,22 @@ namespace math {
*/
using
Tensor
=
paddle_mobile
::
framework
::
Tensor
;
template
<
typename
DeviceType
,
typename
T
>
class
Vol2ColFunctor
{
public:
template
<
typename
DeviceType
,
typename
T
>
class
Vol2ColFunctor
{
public:
void
operator
()(
const
Tensor
&
vol
,
const
std
::
vector
<
int
>
&
dilations
,
const
std
::
vector
<
int
>
&
strides
,
const
std
::
vector
<
int
>
&
paddings
,
Tensor
*
col
)
const
;
};
template
<
typename
DeviceType
,
typename
T
>
class
Col2VolFunctor
{
public:
template
<
typename
DeviceType
,
typename
T
>
class
Col2VolFunctor
{
public:
void
operator
()(
const
Tensor
&
col
,
const
std
::
vector
<
int
>
&
dilations
,
const
std
::
vector
<
int
>
&
strides
,
const
std
::
vector
<
int
>
&
paddings
,
Tensor
*
vol
)
const
;
};
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace math
}
// namespace operators
}
// namespace paddle_mobile
src/operators/mul_op.cpp
浏览文件 @
d6825bf3
...
...
@@ -21,7 +21,8 @@ SOFTWARE.
namespace
paddle_mobile
{
namespace
operators
{
template
<
typename
Dtype
,
typename
T
>
void
MulOp
<
Dtype
,
T
>::
InferShape
()
const
{
template
<
typename
Dtype
,
typename
T
>
void
MulOp
<
Dtype
,
T
>::
InferShape
()
const
{
auto
x_dims
=
param_
.
InputX
()
->
dims
();
auto
y_dims
=
param_
.
InputY
()
->
dims
();
int
x_num_col_dims
=
param_
.
XNumColDims
();
...
...
@@ -52,5 +53,5 @@ template <typename Dtype, typename T> void MulOp<Dtype, T>::InferShape() const {
param_
.
Out
()
->
Resize
(
ddim
);
}
template
class
MulOp
<
CPU
,
float
>;
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/mul_op.h
浏览文件 @
d6825bf3
...
...
@@ -27,7 +27,7 @@ using namespace framework;
template
<
typename
DeviceType
,
typename
T
>
class
MulOp
:
public
framework
::
OperatorWithKernel
<
DeviceType
>
{
public:
public:
MulOp
(
const
std
::
string
&
type
,
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
attrs
,
std
::
shared_ptr
<
framework
::
Scope
>
scope
)
...
...
@@ -43,9 +43,9 @@ public:
using
framework
::
OperatorWithKernel
<
DeviceType
>::
OperatorWithKernel
;
void
InferShape
()
const
override
;
protected:
protected:
MulParam
param_
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/op_param.cpp
浏览文件 @
d6825bf3
...
...
@@ -40,5 +40,5 @@ Print &operator<<(Print &printer, const ConvParam &conv_param) {
printer
<<
" output dims: "
<<
conv_param
.
Output
()
->
dims
();
return
printer
;
}
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/op_param.h
浏览文件 @
d6825bf3
...
...
@@ -31,8 +31,8 @@ namespace operators {
using
namespace
framework
;
class
OpParam
:
PaddleMobileObject
{
public:
protected:
public:
protected:
template
<
typename
T
>
static
T
*
InputFrom
(
const
VariableNameMap
&
inputs
,
const
Scope
&
scope
)
{
return
GetVarValue
<
T
>
(
"Input"
,
inputs
,
scope
);
...
...
@@ -132,7 +132,7 @@ protected:
};
class
ConvParam
:
OpParam
{
public:
public:
ConvParam
(
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
,
const
framework
::
Scope
&
scope
)
{
...
...
@@ -159,7 +159,7 @@ public:
const
int
&
Groups
()
const
{
return
groups
;
}
private:
private:
Tensor
*
input_
;
Tensor
*
output_
;
LoDTensor
*
filter_
;
...
...
@@ -172,7 +172,7 @@ private:
Print
&
operator
<<
(
Print
&
printer
,
const
ConvParam
&
conv_param
);
class
ElementwiseAddParam
:
OpParam
{
public:
public:
ElementwiseAddParam
(
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
,
...
...
@@ -191,7 +191,7 @@ public:
const
int
&
Axis
()
const
{
return
axis_
;
}
private:
private:
Tensor
*
input_x_
;
Tensor
*
input_y_
;
Tensor
*
out_
;
...
...
@@ -199,7 +199,7 @@ private:
};
class
MulParam
:
OpParam
{
public:
public:
MulParam
(
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
,
const
framework
::
Scope
&
scope
)
{
...
...
@@ -220,7 +220,7 @@ public:
const
int
&
YNumColDims
()
const
{
return
y_num_col_dims_
;
}
private:
private:
Tensor
*
input_x_
;
Tensor
*
input_y_
;
Tensor
*
out_
;
...
...
@@ -229,7 +229,7 @@ private:
};
class
ConcatParam
:
public
OpParam
{
public:
public:
ConcatParam
(
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
,
const
framework
::
Scope
&
scope
)
{
...
...
@@ -244,14 +244,14 @@ public:
const
int
&
Axis
()
const
{
return
axis_
;
}
private:
private:
std
::
vector
<
Tensor
*>
inputs_
;
Tensor
*
out_
;
int
axis_
;
};
class
LrnParam
:
public
OpParam
{
public:
public:
LrnParam
(
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
,
const
framework
::
Scope
&
scope
)
{
...
...
@@ -281,7 +281,7 @@ public:
const
std
::
string
&
DataFormat
()
const
{
return
data_format_
;
}
private:
private:
Tensor
*
input_x_
;
Tensor
*
out_
;
Tensor
*
mid_out_
;
...
...
@@ -292,7 +292,7 @@ private:
std
::
string
data_format_
;
};
class
BatchNormParam
:
OpParam
{
public:
public:
BatchNormParam
(
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
,
const
framework
::
Scope
&
scope
)
{
...
...
@@ -327,7 +327,7 @@ public:
const
std
::
string
&
DataFormat
()
const
{
return
data_format_
;
}
private:
private:
Tensor
*
input_x_
;
Tensor
*
output_y_
;
Tensor
*
input_bias_
;
...
...
@@ -340,7 +340,7 @@ private:
std
::
string
data_format_
;
};
class
PoolParam
:
public
OpParam
{
public:
public:
PoolParam
(
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
,
const
framework
::
Scope
&
scope
)
{
...
...
@@ -371,7 +371,7 @@ public:
bool
isGlobalPooling
()
const
{
return
gloabal_pooling_
;
}
private:
private:
Tensor
*
input_
;
Tensor
*
output_
;
std
::
string
pooling_type_
;
...
...
@@ -382,5 +382,5 @@ private:
bool
gloabal_pooling_
=
false
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/pool_op.cpp
浏览文件 @
d6825bf3
...
...
@@ -56,5 +56,5 @@ void PoolOp<DeviceType, T>::InferShape() const {
DLOG
<<
"infer shape out size ="
<<
param_
.
Output
()
->
numel
();
}
template
class
PoolOp
<
CPU
,
float
>;
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/operators/pool_op.h
浏览文件 @
d6825bf3
...
...
@@ -28,7 +28,7 @@ using namespace framework;
template
<
typename
DeviceType
,
typename
T
>
class
PoolOp
:
public
framework
::
OperatorWithKernel
<
DeviceType
>
{
public:
public:
PoolOp
(
const
std
::
string
&
type
,
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
,
std
::
shared_ptr
<
framework
::
Scope
>
scope
)
...
...
@@ -45,8 +45,8 @@ public:
this
->
ClearVariables
({
"X"
});
}
private:
private:
PoolParam
param_
;
};
}
// namespace operators
}
// namespace paddle_mobile
}
// namespace operators
}
// namespace paddle_mobile
src/platform/data_type.h
浏览文件 @
d6825bf3
...
...
@@ -14,9 +14,9 @@ limitations under the License. */
#pragma once
#include "framework/framework.pb.h"
#include <string>
#include <typeindex>
#include "framework/framework.pb.h"
namespace
paddle_mobile
{
namespace
framework
{
...
...
@@ -48,70 +48,70 @@ inline proto::VarType::Type ToDataType(std::type_index type) {
inline
std
::
type_index
ToTypeIndex
(
proto
::
VarType
::
Type
type
)
{
switch
(
type
)
{
// case proto::VarType::FP16:
// return typeid(platform::float16);
case
proto
::
VarType
::
FP32
:
return
typeid
(
float
);
case
proto
::
VarType
::
FP64
:
return
typeid
(
double
);
case
proto
::
VarType
::
INT32
:
return
typeid
(
int
);
case
proto
::
VarType
::
INT64
:
return
typeid
(
int64_t
);
case
proto
::
VarType
::
BOOL
:
return
typeid
(
bool
);
default:
// PADDLE_THROW("Not support type %d", type);
printf
(
"Not support type %d"
,
type
);
// case proto::VarType::FP16:
// return typeid(platform::float16);
case
proto
::
VarType
::
FP32
:
return
typeid
(
float
);
case
proto
::
VarType
::
FP64
:
return
typeid
(
double
);
case
proto
::
VarType
::
INT32
:
return
typeid
(
int
);
case
proto
::
VarType
::
INT64
:
return
typeid
(
int64_t
);
case
proto
::
VarType
::
BOOL
:
return
typeid
(
bool
);
default:
// PADDLE_THROW("Not support type %d", type);
printf
(
"Not support type %d"
,
type
);
}
}
template
<
typename
Visitor
>
inline
void
VisitDataType
(
proto
::
VarType
::
Type
type
,
Visitor
visitor
)
{
switch
(
type
)
{
// case proto::VarType::FP16:
// visitor.template operator()<platform::float16>();
// break;
case
proto
::
VarType
::
FP32
:
visitor
.
template
operator
()
<
float
>();
break
;
case
proto
::
VarType
::
FP64
:
visitor
.
template
operator
()
<
double
>();
break
;
case
proto
::
VarType
::
INT32
:
visitor
.
template
operator
()
<
int
>();
break
;
case
proto
::
VarType
::
INT64
:
visitor
.
template
operator
()
<
int64_t
>();
break
;
case
proto
::
VarType
::
BOOL
:
visitor
.
template
operator
()
<
bool
>();
break
;
default:
// PADDLE_THROW("Not supported");
printf
(
"Not supported"
);
// case proto::VarType::FP16:
// visitor.template operator()<platform::float16>();
// break;
case
proto
::
VarType
::
FP32
:
visitor
.
template
operator
()
<
float
>();
break
;
case
proto
::
VarType
::
FP64
:
visitor
.
template
operator
()
<
double
>();
break
;
case
proto
::
VarType
::
INT32
:
visitor
.
template
operator
()
<
int
>();
break
;
case
proto
::
VarType
::
INT64
:
visitor
.
template
operator
()
<
int64_t
>();
break
;
case
proto
::
VarType
::
BOOL
:
visitor
.
template
operator
()
<
bool
>();
break
;
default:
// PADDLE_THROW("Not supported");
printf
(
"Not supported"
);
}
}
inline
std
::
string
DataTypeToString
(
const
proto
::
VarType
::
Type
type
)
{
switch
(
type
)
{
case
proto
::
VarType
::
FP16
:
return
"float16"
;
case
proto
::
VarType
::
FP32
:
return
"float32"
;
case
proto
::
VarType
::
FP64
:
return
"float64"
;
case
proto
::
VarType
::
INT16
:
return
"int16"
;
case
proto
::
VarType
::
INT32
:
return
"int32"
;
case
proto
::
VarType
::
INT64
:
return
"int64"
;
case
proto
::
VarType
::
BOOL
:
return
"bool"
;
default:
// PADDLE_THROW("Not support type %d", type);
printf
(
"Not support type %d"
,
type
);
case
proto
::
VarType
::
FP16
:
return
"float16"
;
case
proto
::
VarType
::
FP32
:
return
"float32"
;
case
proto
::
VarType
::
FP64
:
return
"float64"
;
case
proto
::
VarType
::
INT16
:
return
"int16"
;
case
proto
::
VarType
::
INT32
:
return
"int32"
;
case
proto
::
VarType
::
INT64
:
return
"int64"
;
case
proto
::
VarType
::
BOOL
:
return
"bool"
;
default:
// PADDLE_THROW("Not support type %d", type);
printf
(
"Not support type %d"
,
type
);
}
}
...
...
@@ -121,5 +121,5 @@ inline std::ostream &operator<<(std::ostream &out,
return
out
;
}
}
// namespace framework
}
// namespace paddle_mobile
}
// namespace framework
}
// namespace paddle_mobile
src/platform/macros.h
浏览文件 @
d6825bf3
...
...
@@ -16,10 +16,10 @@ limitations under the License. */
// Disable the copy and assignment operator for a class.
#ifndef DISABLE_COPY_AND_ASSIGN
#define DISABLE_COPY_AND_ASSIGN(classname)
\
private:
\
classname(const classname &) = delete;
\
classname(classname &&) = delete;
\
classname &operator=(const classname &) = delete;
\
#define DISABLE_COPY_AND_ASSIGN(classname) \
private:
\
classname(const classname &) = delete; \
classname(classname &&) = delete; \
classname &operator=(const classname &) = delete; \
classname &operator=(classname &&) = delete
#endif
test/operators/test_cov_op.cpp
浏览文件 @
d6825bf3
...
...
@@ -18,7 +18,7 @@ SOFTWARE.
#include "../framework/executor_for_test.h"
#include "../test_helper.h"
#include "io.h"
#include "
./
io.h"
int
main
()
{
paddle_mobile
::
Loader
<
paddle_mobile
::
CPU
>
loader
;
...
...
@@ -38,7 +38,7 @@ int main() {
auto
output
=
executor
.
predict
(
input
,
"data"
,
"conv2d_0.tmp_0"
,
{
1
,
64
,
56
,
56
});
float
*
output_ptr
=
output
->
data
<
float
>
();
auto
output_ptr
=
output
->
data
<
float
>
();
for
(
int
j
=
0
;
j
<
output
->
numel
();
++
j
)
{
DLOG
<<
" value of output: "
<<
output_ptr
[
j
];
}
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录