提交 b9f8cc06 编写于 作者: Y Yu Yang

Merge branch 'develop' of github.com:baidu/Paddle into feature/tester

...@@ -52,6 +52,10 @@ def wrap_param_default(param_names=None, ...@@ -52,6 +52,10 @@ def wrap_param_default(param_names=None,
kwargs[name] = default_factory(func) kwargs[name] = default_factory(func)
return func(*args, **kwargs) return func(*args, **kwargs)
if hasattr(func, 'argspec'):
__wrapper__.argspec = func.argspec
else:
__wrapper__.argspec = inspect.getargspec(func)
return __wrapper__ return __wrapper__
return __impl__ return __impl__
......
...@@ -14,6 +14,7 @@ ...@@ -14,6 +14,7 @@
import functools import functools
import collections import collections
import inspect
from paddle.trainer.config_parser import * from paddle.trainer.config_parser import *
from .activations import LinearActivation, SigmoidActivation, TanhActivation, \ from .activations import LinearActivation, SigmoidActivation, TanhActivation, \
...@@ -316,6 +317,11 @@ def layer_support(*attrs): ...@@ -316,6 +317,11 @@ def layer_support(*attrs):
val.check(method.__name__) val.check(method.__name__)
return method(*args, **kwargs) return method(*args, **kwargs)
if hasattr(method, 'argspec'):
wrapper.argspec = method.argspec
else:
wrapper.argspec = inspect.getargspec(method)
return wrapper return wrapper
return decorator return decorator
......
...@@ -67,6 +67,7 @@ paddle.v2.parameters.create, no longer exposed to users. ...@@ -67,6 +67,7 @@ paddle.v2.parameters.create, no longer exposed to users.
""" """
import collections import collections
import inspect
import paddle.trainer_config_helpers as conf_helps import paddle.trainer_config_helpers as conf_helps
from paddle.trainer_config_helpers.config_parser_utils import \ from paddle.trainer_config_helpers.config_parser_utils import \
...@@ -74,26 +75,14 @@ from paddle.trainer_config_helpers.config_parser_utils import \ ...@@ -74,26 +75,14 @@ from paddle.trainer_config_helpers.config_parser_utils import \
from paddle.trainer_config_helpers.default_decorators import wrap_name_default from paddle.trainer_config_helpers.default_decorators import wrap_name_default
from paddle.trainer_config_helpers.default_decorators import wrap_act_default from paddle.trainer_config_helpers.default_decorators import wrap_act_default
from paddle.trainer_config_helpers.default_decorators import wrap_bias_attr_default from paddle.trainer_config_helpers.default_decorators import \
wrap_bias_attr_default
from paddle.trainer_config_helpers.layers import layer_support from paddle.trainer_config_helpers.layers import layer_support
import data_type import data_type
import activation import activation
import attr
__all__ = ['parse_network', 'data']
__all__ = [
'parse_network', 'data', 'fc', 'conv_shift', 'img_conv', 'img_pool', 'spp',
'maxout', 'img_cmrnorm', 'batch_norm', 'sum_to_one_norm', 'recurrent',
'lstmemory', 'grumemory', 'pool', 'last_seq', 'first_seq', 'concat',
'seq_concat', 'block_expand', 'expand', 'repeat', 'seq_reshape', 'addto',
'linear_comb', 'interpolation', 'bilinear_interp', 'power', 'scaling',
'slope_intercept', 'tensor', 'cos_sim', 'trans', 'max_id', 'sampling_id',
'pad', 'classification_cost', 'cross_entropy_cost',
'cross_entropy_with_selfnorm_cost', 'regression_cost',
'multi_binary_label_cross_entropy_cost', 'rank_cost', 'lambda_cost',
'sum_cost', 'huber_cost', 'crf', 'crf_decoding', 'ctc', 'warp_ctc', 'nce',
'hsigmoid', 'eos'
]
__projection_names__ = filter(lambda x: x.endswith('_projection'), __projection_names__ = filter(lambda x: x.endswith('_projection'),
dir(conf_helps)) dir(conf_helps))
...@@ -289,83 +278,51 @@ data = DataLayerV2 ...@@ -289,83 +278,51 @@ data = DataLayerV2
AggregateLevel = conf_helps.layers.AggregateLevel AggregateLevel = conf_helps.layers.AggregateLevel
ExpandLevel = conf_helps.layers.ExpandLevel ExpandLevel = conf_helps.layers.ExpandLevel
layer_list = [
# [V2LayerImpl, V1_method_name, parent_names] def __layer_name_mapping__(inname):
# fully connected layers if inname in ['data_layer', 'memory', 'mixed_layer']:
['fc', 'fc_layer', ['input']], # Do Not handle these layers
# conv layers return
['conv_shift', 'conv_shift_layer', ['a', 'b']], elif inname == 'maxid_layer':
['img_conv', 'img_conv_layer', ['input']], return 'max_id'
# image pooling layers elif inname.endswith('memory') or inname.endswith(
['img_pool', 'img_pool_layer', ['input']], '_seq') or inname.endswith('_sim') or inname == 'hsigmoid':
['spp', 'spp_layer', ['input']], return inname
['maxout', 'maxout_layer', ['input']], elif inname in [
# norm layers 'cross_entropy', 'multi_binary_label_cross_entropy',
['img_cmrnorm', 'img_cmrnorm_layer', ['input']], 'cross_entropy_with_selfnorm'
['batch_norm', 'batch_norm_layer', ['input']], ]:
['sum_to_one_norm', 'sum_to_one_norm_layer', ['input']], return inname + "_cost"
# recurrent layers elif inname.endswith('_cost'):
['recurrent', 'recurrent_layer', ['input']], return inname
['lstmemory', 'lstmemory', ['input']], elif inname.endswith("_layer"):
['grumemory', 'grumemory', ['input']], return inname[:-len("_layer")]
# aggregate layers
['pool', 'pooling_layer', ['input']],
['last_seq', 'last_seq', ['input']], def __layer_name_mapping_parent_names__(inname):
['first_seq', 'first_seq', ['input']], all_args = getattr(conf_helps, inname).argspec.args
['concat', 'concat_layer', ['input']], return filter(
['seq_concat', 'seq_concat_layer', ['a', 'b']], lambda x: x in ['input1', 'input2','label', 'input', 'a', 'b', 'expand_as',
# reshaping layers 'weights', 'vectors', 'weight', 'score', 'left', 'right'],
['block_expand', 'block_expand_layer', ['input']], all_args)
['expand', 'expand_layer', ['input', 'expand_as']],
['repeat', 'repeat_layer', ['input']],
['rotate', 'rotate_layer', ['input']], def __convert_layer__(_new_name_, _old_name_, _parent_names_):
['seq_reshape', 'seq_reshape_layer', ['input']], global __all__
# math layers __all__.append(_new_name_)
['addto', 'addto_layer', ['input']], globals()[new_name] = __convert_to_v2__(_old_name_, _parent_names_)
['linear_comb', 'linear_comb_layer', ['weights', 'vectors']],
['interpolation', 'interpolation_layer', ['input', 'weight']],
['bilinear_interp', 'bilinear_interp_layer', ['input']], for each_layer_name in dir(conf_helps):
['power', 'power_layer', ['input', 'weight']], new_name = __layer_name_mapping__(each_layer_name)
['scaling', 'scaling_layer', ['input', 'weight']], if new_name is not None:
['slope_intercept', 'slope_intercept_layer', ['input']], parent_names = __layer_name_mapping_parent_names__(each_layer_name)
['tensor', 'tensor_layer', ['a', 'b']], assert len(parent_names) != 0, each_layer_name
['cos_sim', 'cos_sim', ['a', 'b']], __convert_layer__(new_name, each_layer_name, parent_names)
['trans', 'trans_layer', ['input']],
# sampling layers del parent_names
['max_id', 'maxid_layer', ['input']], del new_name
['sampling_id', 'sampling_id_layer', ['input']], del each_layer_name
# slicing and joining layers
['pad', 'pad_layer', ['input']],
# cost layers
[
'classification_cost', 'classification_cost',
['input', 'label', 'weight']
],
['regression_cost', 'regression_cost', ['input', 'label', 'weight']],
['cross_entropy_cost', 'cross_entropy', ['input', 'label']],
[
'cross_entropy_with_selfnorm_cost', 'cross_entropy_with_selfnorm',
['input', 'label']
],
[
'multi_binary_label_cross_entropy_cost',
'multi_binary_label_cross_entropy', ['input', 'label']
],
['rank_cost', 'rank_cost', ['left', 'right', 'label', 'weight']],
['lambda_cost', 'lambda_cost', ['input', 'score']],
['sum_cost', 'sum_cost', ['input']],
['huber_cost', 'huber_cost', ['input', 'label']],
['crf', 'crf_layer', ['input', 'label']],
['crf_decoding', 'crf_decoding_layer', ['input']],
['ctc', 'ctc_layer', ['input', 'label']],
['warp_ctc', 'warp_ctc_layer', ['input', 'label']],
['nce', 'nce_layer', ['input', 'label']],
['hsigmoid', 'hsigmoid', ['input', 'label']],
# check layers
['eos', 'eos_layer', ['input']]
]
for l in layer_list:
globals()[l[0]] = __convert_to_v2__(l[1], l[2])
# convert projection # convert projection
for prj in __projection_names__: for prj in __projection_names__:
......
...@@ -11,17 +11,13 @@ ...@@ -11,17 +11,13 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and # See the License for the specific language governing permissions and
# limitations under the License. # limitations under the License.
import difflib
import unittest import unittest
import paddle.trainer_config_helpers as conf_helps
import paddle.v2.activation as activation import paddle.v2.activation as activation
import paddle.v2.attr as attr import paddle.v2.attr as attr
import paddle.v2.data_type as data_type import paddle.v2.data_type as data_type
import paddle.v2.layer as layer import paddle.v2.layer as layer
import paddle.v2.pooling as pooling import paddle.v2.pooling as pooling
from paddle.trainer_config_helpers.config_parser_utils import \
parse_network_config as parse_network
pixel = layer.data(name='pixel', type=data_type.dense_vector(128)) pixel = layer.data(name='pixel', type=data_type.dense_vector(128))
label = layer.data(name='label', type=data_type.integer_value(10)) label = layer.data(name='label', type=data_type.integer_value(10))
...@@ -70,7 +66,7 @@ class ImageLayerTest(unittest.TestCase): ...@@ -70,7 +66,7 @@ class ImageLayerTest(unittest.TestCase):
class AggregateLayerTest(unittest.TestCase): class AggregateLayerTest(unittest.TestCase):
def test_aggregate_layer(self): def test_aggregate_layer(self):
pool = layer.pool( pool = layer.pooling(
input=pixel, input=pixel,
pooling_type=pooling.Avg(), pooling_type=pooling.Avg(),
agg_level=layer.AggregateLevel.EACH_SEQUENCE) agg_level=layer.AggregateLevel.EACH_SEQUENCE)
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册