Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
06056fe2
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
06056fe2
编写于
2月 27, 2017
作者:
L
Luo Tao
浏览文件
操作
浏览文件
下载
差异文件
Merge branch 'develop' into layer
上级
6f63eb9f
d425a5ca
变更
3
显示空白变更内容
内联
并排
Showing
3 changed file
with
286 addition
and
29 deletion
+286
-29
python/paddle/trainer_config_helpers/layers.py
python/paddle/trainer_config_helpers/layers.py
+85
-22
python/paddle/v2/layer.py
python/paddle/v2/layer.py
+113
-7
python/paddle/v2/tests/test_layer.py
python/paddle/v2/tests/test_layer.py
+88
-0
未找到文件。
python/paddle/trainer_config_helpers/layers.py
浏览文件 @
06056fe2
...
@@ -30,28 +30,90 @@ except ImportError:
...
@@ -30,28 +30,90 @@ except ImportError:
import
copy
import
copy
__all__
=
[
__all__
=
[
"full_matrix_projection"
,
"AggregateLevel"
,
"ExpandLevel"
,
"full_matrix_projection"
,
"identity_projection"
,
"dotmul_projection"
,
"dotmul_operator"
,
"AggregateLevel"
,
"repeat_layer"
,
"seq_reshape_layer"
,
"table_projection"
,
"mixed_layer"
,
"ExpandLevel"
,
"data_layer"
,
"embedding_layer"
,
"fc_layer"
,
"grumemory"
,
"pooling_layer"
,
"identity_projection"
,
"lstmemory"
,
"last_seq"
,
"first_seq"
,
"cos_sim"
,
"hsigmoid"
,
"dotmul_projection"
,
"conv_projection"
,
"regression_cost"
,
'classification_cost'
,
"LayerOutput"
,
"dotmul_operator"
,
'img_conv_layer'
,
'img_pool_layer'
,
'batch_norm_layer'
,
'img_cmrnorm_layer'
,
"repeat_layer"
,
'addto_layer'
,
'concat_layer'
,
'seq_concat_layer'
,
'lstm_step_layer'
,
"seq_reshape_layer"
,
'recurrent_group'
,
'memory'
,
'StaticInput'
,
'expand_layer'
,
'scaling_layer'
,
"table_projection"
,
'scaling_projection'
,
'power_layer'
,
'interpolation_layer'
,
"mixed_layer"
,
'bilinear_interp_layer'
,
'trans_layer'
,
'rotate_layer'
,
"data_layer"
,
'sum_to_one_norm_layer'
,
'get_output_layer'
,
'LayerType'
,
"embedding_layer"
,
'context_projection'
,
'beam_search'
,
'maxid_layer'
,
'GeneratedInput'
,
"fc_layer"
,
'SubsequenceInput'
,
'gru_step_layer'
,
'recurrent_layer'
,
"grumemory"
,
'BaseGeneratedInput'
,
'conv_operator'
,
'conv_shift_layer'
,
'tensor_layer'
,
"pooling_layer"
,
'selective_fc_layer'
,
'sampling_id_layer'
,
'slope_intercept_layer'
,
"lstmemory"
,
'trans_full_matrix_projection'
,
'linear_comb_layer'
,
'convex_comb_layer'
,
"last_seq"
,
'ctc_layer'
,
'warp_ctc_layer'
,
'crf_layer'
,
'crf_decoding_layer'
,
"first_seq"
,
'nce_layer'
,
'cross_entropy_with_selfnorm'
,
'cross_entropy'
,
"cos_sim"
,
'multi_binary_label_cross_entropy'
,
'sum_cost'
,
'rank_cost'
,
'lambda_cost'
,
"hsigmoid"
,
'huber_cost'
,
'block_expand_layer'
,
'maxout_layer'
,
'out_prod_layer'
,
"conv_projection"
,
'print_layer'
,
'priorbox_layer'
,
'spp_layer'
,
'pad_layer'
,
'eos_layer'
"regression_cost"
,
'classification_cost'
,
"LayerOutput"
,
'img_conv_layer'
,
'img_pool_layer'
,
'batch_norm_layer'
,
'img_cmrnorm_layer'
,
'addto_layer'
,
'concat_layer'
,
'seq_concat_layer'
,
'lstm_step_layer'
,
'recurrent_group'
,
'memory'
,
'StaticInput'
,
'expand_layer'
,
'scaling_layer'
,
'scaling_projection'
,
'power_layer'
,
'interpolation_layer'
,
'bilinear_interp_layer'
,
'trans_layer'
,
'rotate_layer'
,
'sum_to_one_norm_layer'
,
'get_output_layer'
,
'LayerType'
,
'context_projection'
,
'beam_search'
,
'maxid_layer'
,
'GeneratedInput'
,
'SubsequenceInput'
,
'gru_step_layer'
,
'recurrent_layer'
,
'BaseGeneratedInput'
,
'conv_operator'
,
'conv_shift_layer'
,
'tensor_layer'
,
'selective_fc_layer'
,
'sampling_id_layer'
,
'slope_intercept_layer'
,
'trans_full_matrix_projection'
,
'linear_comb_layer'
,
'convex_comb_layer'
,
'ctc_layer'
,
'warp_ctc_layer'
,
'crf_layer'
,
'crf_decoding_layer'
,
'nce_layer'
,
'cross_entropy_with_selfnorm'
,
'cross_entropy'
,
'multi_binary_label_cross_entropy'
,
'sum_cost'
,
'rank_cost'
,
'lambda_cost'
,
'huber_cost'
,
'block_expand_layer'
,
'maxout_layer'
,
'out_prod_layer'
,
'print_layer'
,
'priorbox_layer'
,
'spp_layer'
,
'pad_layer'
,
'eos_layer'
,
'layer_support'
,
]
]
...
@@ -648,6 +710,7 @@ class MixedLayerType(LayerOutput):
...
@@ -648,6 +710,7 @@ class MixedLayerType(LayerOutput):
# update the size which might be computed inside MixedLayer
# update the size which might be computed inside MixedLayer
# according to the operator's output size
# according to the operator's output size
self
.
size
=
ml
.
config
.
size
self
.
size
=
ml
.
config
.
size
self
.
finalized
=
True
@
wrap_name_default
(
"mixed"
)
@
wrap_name_default
(
"mixed"
)
...
...
python/paddle/v2/layer.py
浏览文件 @
06056fe2
...
@@ -71,9 +71,15 @@ import collections
...
@@ -71,9 +71,15 @@ import collections
import
paddle.trainer_config_helpers
as
conf_helps
import
paddle.trainer_config_helpers
as
conf_helps
from
paddle.trainer_config_helpers.config_parser_utils
import
\
from
paddle.trainer_config_helpers.config_parser_utils
import
\
parse_network_config
as
__parse__
parse_network_config
as
__parse__
from
paddle.trainer_config_helpers.default_decorators
import
wrap_name_default
from
paddle.trainer_config_helpers.default_decorators
import
wrap_name_default
from
paddle.trainer_config_helpers.default_decorators
import
wrap_act_default
from
paddle.trainer_config_helpers.default_decorators
import
wrap_bias_attr_default
from
paddle.trainer_config_helpers.layers
import
layer_support
import
data_type
import
data_type
import
activation
import
attr
__all__
=
[
__all__
=
[
'parse_network'
,
'data'
,
'fc'
,
'conv_shift'
,
'img_conv'
,
'img_pool'
,
'spp'
,
'parse_network'
,
'data'
,
'fc'
,
'conv_shift'
,
'img_conv'
,
'img_pool'
,
'spp'
,
...
@@ -89,6 +95,13 @@ __all__ = [
...
@@ -89,6 +95,13 @@ __all__ = [
'hsigmoid'
,
'eos'
'hsigmoid'
,
'eos'
]
]
__projection_names__
=
filter
(
lambda
x
:
x
.
endswith
(
'_projection'
),
dir
(
conf_helps
))
__all__
+=
__projection_names__
__operator_names__
=
filter
(
lambda
x
:
x
.
endswith
(
'_operator'
),
dir
(
conf_helps
))
__all__
+=
__operator_names__
def
parse_network
(
*
outputs
):
def
parse_network
(
*
outputs
):
"""
"""
...
@@ -106,9 +119,8 @@ def parse_network(*outputs):
...
@@ -106,9 +119,8 @@ def parse_network(*outputs):
class
Layer
(
object
):
class
Layer
(
object
):
def
__init__
(
self
,
name
,
parent_layers
):
def
__init__
(
self
,
name
=
None
,
parent_layers
=
None
):
assert
isinstance
(
parent_layers
,
dict
)
assert
isinstance
(
parent_layers
,
dict
)
assert
isinstance
(
name
,
basestring
)
self
.
name
=
name
self
.
name
=
name
self
.
__parent_layers__
=
parent_layers
self
.
__parent_layers__
=
parent_layers
...
@@ -127,19 +139,25 @@ class Layer(object):
...
@@ -127,19 +139,25 @@ class Layer(object):
self
.
__parent_layers__
[
layer_name
])
self
.
__parent_layers__
[
layer_name
])
kwargs
[
layer_name
]
=
v1_layer
kwargs
[
layer_name
]
=
v1_layer
if
self
.
name
not
in
context
:
if
self
.
name
is
None
:
return
self
.
to_proto_impl
(
**
kwargs
)
elif
self
.
name
not
in
context
:
context
[
self
.
name
]
=
self
.
to_proto_impl
(
**
kwargs
)
context
[
self
.
name
]
=
self
.
to_proto_impl
(
**
kwargs
)
return
context
[
self
.
name
]
return
context
[
self
.
name
]
def
to_proto_impl
(
self
,
**
kwargs
):
def
to_proto_impl
(
self
,
**
kwargs
):
raise
NotImplementedError
()
raise
NotImplementedError
()
def
__convert_to_v2__
(
method_name
,
parent_names
):
def
__convert_to_v2__
(
method_name
,
parent_names
,
is_default_name
=
True
):
if
is_default_name
:
wrapper
=
wrap_name_default
(
name_prefix
=
method_name
)
wrapper
=
wrap_name_default
(
name_prefix
=
method_name
)
else
:
wrapper
=
None
class
V2LayerImpl
(
Layer
):
class
V2LayerImpl
(
Layer
):
def
__init__
(
self
,
name
=
None
,
**
kwargs
):
def
__init__
(
self
,
**
kwargs
):
parent_layers
=
dict
()
parent_layers
=
dict
()
other_kwargs
=
dict
()
other_kwargs
=
dict
()
for
pname
in
parent_names
:
for
pname
in
parent_names
:
...
@@ -150,6 +168,7 @@ def __convert_to_v2__(method_name, parent_names):
...
@@ -150,6 +168,7 @@ def __convert_to_v2__(method_name, parent_names):
if
key
not
in
parent_names
:
if
key
not
in
parent_names
:
other_kwargs
[
key
]
=
kwargs
[
key
]
other_kwargs
[
key
]
=
kwargs
[
key
]
name
=
kwargs
.
get
(
'name'
,
None
)
super
(
V2LayerImpl
,
self
).
__init__
(
name
,
parent_layers
)
super
(
V2LayerImpl
,
self
).
__init__
(
name
,
parent_layers
)
self
.
__other_kwargs__
=
other_kwargs
self
.
__other_kwargs__
=
other_kwargs
...
@@ -162,7 +181,7 @@ def __convert_to_v2__(method_name, parent_names):
...
@@ -162,7 +181,7 @@ def __convert_to_v2__(method_name, parent_names):
args
[
each
]
=
kwargs
[
each
]
args
[
each
]
=
kwargs
[
each
]
for
each
in
self
.
__other_kwargs__
:
for
each
in
self
.
__other_kwargs__
:
args
[
each
]
=
self
.
__other_kwargs__
[
each
]
args
[
each
]
=
self
.
__other_kwargs__
[
each
]
return
getattr
(
conf_helps
,
method_name
)(
name
=
self
.
name
,
**
args
)
return
getattr
(
conf_helps
,
method_name
)(
**
args
)
return
V2LayerImpl
return
V2LayerImpl
...
@@ -193,6 +212,78 @@ class DataLayerV2(Layer):
...
@@ -193,6 +212,78 @@ class DataLayerV2(Layer):
return
getattr
(
conf_helps
,
self
.
__method_name__
)(
name
=
self
.
name
,
**
args
)
return
getattr
(
conf_helps
,
self
.
__method_name__
)(
name
=
self
.
name
,
**
args
)
class
MixedLayerV2
(
Layer
):
"""
This class is use to support `with` grammar. If not, the following code
could convert mixed_layer simply.
mixed = __convert_to_v2__(
'mixed_layer', name_prefix='mixed', parent_names=['input'])
"""
class
AddToSealedMixedLayerExceptionV2
(
Exception
):
pass
def
__init__
(
self
,
size
=
0
,
input
=
None
,
name
=
None
,
act
=
None
,
bias_attr
=
None
,
layer_attr
=
None
):
self
.
__method_name__
=
'mixed_layer'
self
.
finalized
=
False
self
.
__inputs__
=
[]
if
input
is
not
None
:
self
.
__inputs__
=
input
other_kwargs
=
dict
()
other_kwargs
[
'name'
]
=
name
other_kwargs
[
'size'
]
=
size
other_kwargs
[
'act'
]
=
act
other_kwargs
[
'bias_attr'
]
=
bias_attr
other_kwargs
[
'layer_attr'
]
=
layer_attr
parent_layers
=
{
"input"
:
self
.
__inputs__
}
super
(
MixedLayerV2
,
self
).
__init__
(
name
,
parent_layers
)
self
.
__other_kwargs__
=
other_kwargs
def
__iadd__
(
self
,
other
):
if
not
self
.
finalized
:
self
.
__inputs__
.
append
(
other
)
return
self
else
:
raise
MixedLayerTypeV2
.
AddToSealedMixedLayerExceptionV2
()
def
__enter__
(
self
):
assert
len
(
self
.
__inputs__
)
==
0
return
self
def
__exit__
(
self
,
*
args
,
**
kwargs
):
self
.
finalized
=
True
def
to_proto_impl
(
self
,
**
kwargs
):
args
=
dict
()
for
each
in
kwargs
:
args
[
each
]
=
kwargs
[
each
]
for
each
in
self
.
__other_kwargs__
:
args
[
each
]
=
self
.
__other_kwargs__
[
each
]
return
getattr
(
conf_helps
,
self
.
__method_name__
)(
**
args
)
@
wrap_name_default
(
"mixed"
)
@
wrap_act_default
(
act
=
activation
.
Linear
())
@
wrap_bias_attr_default
(
has_bias
=
False
)
@
layer_support
(
conf_helps
.
layers
.
ERROR_CLIPPING
,
conf_helps
.
layers
.
DROPOUT
)
def
mixed
(
size
=
0
,
name
=
None
,
input
=
None
,
act
=
None
,
bias_attr
=
False
,
layer_attr
=
None
):
return
MixedLayerV2
(
size
,
input
,
name
,
act
,
bias_attr
,
layer_attr
)
data
=
DataLayerV2
data
=
DataLayerV2
AggregateLevel
=
conf_helps
.
layers
.
AggregateLevel
AggregateLevel
=
conf_helps
.
layers
.
AggregateLevel
ExpandLevel
=
conf_helps
.
layers
.
ExpandLevel
ExpandLevel
=
conf_helps
.
layers
.
ExpandLevel
...
@@ -274,3 +365,18 @@ layer_list = [
...
@@ -274,3 +365,18 @@ layer_list = [
]
]
for
l
in
layer_list
:
for
l
in
layer_list
:
globals
()[
l
[
0
]]
=
__convert_to_v2__
(
l
[
1
],
l
[
2
])
globals
()[
l
[
0
]]
=
__convert_to_v2__
(
l
[
1
],
l
[
2
])
# convert projection
for
prj
in
__projection_names__
:
globals
()[
prj
]
=
__convert_to_v2__
(
prj
,
parent_names
=
[
'input'
],
is_default_name
=
False
)
# convert operator
operator_list
=
[
# [V1_method_name, parent_names],
[
'dotmul_operator'
,
[
'a'
,
'b'
]],
[
'conv_operator'
,
[
'img'
,
'filter'
]]
]
for
op
in
operator_list
:
globals
()[
op
[
0
]]
=
__convert_to_v2__
(
op
[
0
],
parent_names
=
op
[
1
],
is_default_name
=
False
)
python/paddle/v2/tests/test_layer.py
浏览文件 @
06056fe2
...
@@ -167,5 +167,93 @@ class OtherLayerTest(unittest.TestCase):
...
@@ -167,5 +167,93 @@ class OtherLayerTest(unittest.TestCase):
print
layer
.
parse_network
(
pad
)
print
layer
.
parse_network
(
pad
)
class
ProjOpTest
(
unittest
.
TestCase
):
def
test_projection
(
self
):
input
=
layer
.
data
(
name
=
'data'
,
type
=
data_type
.
dense_vector
(
784
))
word
=
layer
.
data
(
name
=
'word'
,
type
=
data_type
.
integer_value_sequence
(
10000
))
fc0
=
layer
.
fc
(
input
=
input
,
size
=
100
,
act
=
activation
.
Sigmoid
())
fc1
=
layer
.
fc
(
input
=
input
,
size
=
200
,
act
=
activation
.
Sigmoid
())
mixed0
=
layer
.
mixed
(
size
=
256
,
input
=
[
layer
.
full_matrix_projection
(
input
=
fc0
),
layer
.
full_matrix_projection
(
input
=
fc1
)
])
with
layer
.
mixed
(
size
=
200
)
as
mixed1
:
mixed1
+=
layer
.
full_matrix_projection
(
input
=
fc0
)
mixed1
+=
layer
.
identity_projection
(
input
=
fc1
)
table
=
layer
.
table_projection
(
input
=
word
)
emb0
=
layer
.
mixed
(
size
=
512
,
input
=
table
)
with
layer
.
mixed
(
size
=
512
)
as
emb1
:
emb1
+=
table
scale
=
layer
.
scaling_projection
(
input
=
fc0
)
scale0
=
layer
.
mixed
(
size
=
100
,
input
=
scale
)
with
layer
.
mixed
(
size
=
100
)
as
scale1
:
scale1
+=
scale
dotmul
=
layer
.
dotmul_projection
(
input
=
fc0
)
dotmul0
=
layer
.
mixed
(
size
=
100
,
input
=
dotmul
)
with
layer
.
mixed
(
size
=
100
)
as
dotmul1
:
dotmul1
+=
dotmul
context
=
layer
.
context_projection
(
input
=
fc0
,
context_len
=
5
)
context0
=
layer
.
mixed
(
size
=
100
,
input
=
context
)
with
layer
.
mixed
(
size
=
100
)
as
context1
:
context1
+=
context
conv
=
layer
.
conv_projection
(
input
=
input
,
filter_size
=
1
,
num_channels
=
1
,
num_filters
=
128
,
stride
=
1
,
padding
=
0
)
conv0
=
layer
.
mixed
(
input
=
conv
,
bias_attr
=
True
)
with
layer
.
mixed
(
bias_attr
=
True
)
as
conv1
:
conv1
+=
conv
print
layer
.
parse_network
(
mixed0
)
print
layer
.
parse_network
(
mixed1
)
print
layer
.
parse_network
(
emb0
)
print
layer
.
parse_network
(
emb1
)
print
layer
.
parse_network
(
scale0
)
print
layer
.
parse_network
(
scale1
)
print
layer
.
parse_network
(
dotmul0
)
print
layer
.
parse_network
(
dotmul1
)
print
layer
.
parse_network
(
conv0
)
print
layer
.
parse_network
(
conv1
)
def
test_operator
(
self
):
ipt0
=
layer
.
data
(
name
=
'data'
,
type
=
data_type
.
dense_vector
(
784
))
ipt1
=
layer
.
data
(
name
=
'word'
,
type
=
data_type
.
dense_vector
(
128
))
fc0
=
layer
.
fc
(
input
=
ipt0
,
size
=
100
,
act
=
activation
.
Sigmoid
())
fc1
=
layer
.
fc
(
input
=
ipt0
,
size
=
100
,
act
=
activation
.
Sigmoid
())
dotmul_op
=
layer
.
dotmul_operator
(
a
=
fc0
,
b
=
fc1
)
dotmul0
=
layer
.
mixed
(
input
=
dotmul_op
)
with
layer
.
mixed
()
as
dotmul1
:
dotmul1
+=
dotmul_op
conv
=
layer
.
conv_operator
(
img
=
ipt0
,
filter
=
ipt1
,
filter_size
=
1
,
num_channels
=
1
,
num_filters
=
128
,
stride
=
1
,
padding
=
0
)
conv0
=
layer
.
mixed
(
input
=
conv
)
with
layer
.
mixed
()
as
conv1
:
conv1
+=
conv
print
layer
.
parse_network
(
dotmul0
)
print
layer
.
parse_network
(
dotmul1
)
print
layer
.
parse_network
(
conv0
)
print
layer
.
parse_network
(
conv1
)
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
unittest
.
main
()
unittest
.
main
()
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录