Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
fe84517b
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
fe84517b
编写于
9月 20, 2017
作者:
C
Cao Ying
提交者:
GitHub
9月 20, 2017
浏览文件
操作
浏览文件
下载
差异文件
Merge pull request #4158 from ranqiu92/r-doc
update F&Q.
上级
8deda323
c1e3550e
变更
2
显示空白变更内容
内联
并排
Showing
2 changed file
with
254 addition
and
153 deletion
+254
-153
doc/faq/index_cn.rst
doc/faq/index_cn.rst
+52
-0
python/paddle/trainer_config_helpers/layers.py
python/paddle/trainer_config_helpers/layers.py
+202
-153
未找到文件。
doc/faq/index_cn.rst
浏览文件 @
fe84517b
...
@@ -321,3 +321,55 @@ pip uninstall py_paddle paddle
...
@@ -321,3 +321,55 @@ pip uninstall py_paddle paddle
然后安装paddle的python环境, 在build目录下执行
然后安装paddle的python环境, 在build目录下执行
pip install python/dist/paddle*.whl && pip install ../paddle/dist/py_paddle*.whl
pip install python/dist/paddle*.whl && pip install ../paddle/dist/py_paddle*.whl
16. PaddlePaddle存储的参数格式是什么,如何和明文进行相互转化
---------------------------------------------------------
PaddlePaddle保存的模型参数文件内容由16字节头信息和网络参数两部分组成。头信息中,1~4字节表示PaddlePaddle版本信息,请直接填充0;5~8字节表示每个参数占用的字节数,当保存的网络参数为float类型时为4,double类型时为8;9~16字节表示保存的参数总个数。
将PaddlePaddle保存的模型参数还原回明文时,可以使用相应数据类型的 :code:`numpy.array` 加载具体网络参数,此时可以跳过PaddlePaddle模型参数文件的头信息。若在PaddlePaddle编译时,未指定按照double精度编译,默认情况下按照float精度计算,保存的参数也是float类型。这时在使用 :code:`numpy.array` 时,一般设置 :code:`dtype=float32` 。示例如下:
.. code-block:: python
def read_parameter(fname, width):
s = open(fname).read()
# skip header
vec = np.fromstring(s[16:], dtype=np.float32)
# width is the size of the corresponding layer
np.savetxt(fname + ".csv", vec.reshape(width, -1),
fmt="%.6f", delimiter=",")
将明文参数转化为PaddlePaddle可加载的模型参数时,首先构造头信息,再写入网络参数。下面的代码将随机生成的矩阵转化为可以被PaddlePaddle加载的模型参数。
.. code-block:: python
def gen_rand_param(param_file, width, height, need_trans):
np.random.seed()
header = struct.pack("iil", 0, 4, height * width)
param = np.float32(np.random.rand(height, width))
with open(param_file, "w") as fparam:
fparam.write(header + param.tostring())
17. 如何加载预训练参数
------------------------------
* 对加载预训练参数的层,设置其参数属性 :code:`is_static=True`,使该层的参数在训练过程中保持不变。以embedding层为例,代码如下:
.. code-block:: python
emb_para = paddle.attr.Param(name='emb', is_static=True)
paddle.layer.embedding(size=word_dim, input=x, param_attr=emb_para)
* 从模型文件将预训练参数载入 :code:`numpy.array`,在创建parameters后,使用 :code:`parameters.set()` 加载预训练参数。PaddlePaddle保存的模型参数文件前16字节为头信息,用户将参数载入 :code:`numpy.array` 时须从第17字节开始。以embedding层为例,代码如下:
.. code-block:: python
def load_parameter(file_name, h, w):
with open(file_name, 'rb') as f:
f.read(16) # skip header.
return np.fromfile(f, dtype=np.float32).reshape(h, w)
parameters = paddle.parameters.create(my_cost)
parameters.set('emb', load_parameter(emb_param_file, 30000, 256))
python/paddle/trainer_config_helpers/layers.py
浏览文件 @
fe84517b
...
@@ -781,11 +781,11 @@ class MixedLayerType(LayerOutput):
...
@@ -781,11 +781,11 @@ class MixedLayerType(LayerOutput):
:type size: int
:type size: int
:param act: activation type.
:param act: activation type.
:type act: BaseActivation
:type act: BaseActivation
:param bias_attr: The Bias Attribute. If
no bias, then pass False or
:param bias_attr: The Bias Attribute. If
the parameter is set to
something not type of ParameterAttribute. None will
False or something not type of ParameterAttribute,
get a default Bias.
no bias is defined. If the parameter is set to
:type bias_attr: ParameterAttribute or None means has bias. Any other
True, the bias is initialized to zero.
type means no bias.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param layer_attr: Extra Layer Attribute.
:param layer_attr: Extra Layer Attribute.
:type layer_attr: ExtraLayerAttribute or None
:type layer_attr: ExtraLayerAttribute or None
"""
"""
...
@@ -881,10 +881,11 @@ def mixed_layer(size=0,
...
@@ -881,10 +881,11 @@ def mixed_layer(size=0,
then this function will just return layer's name.
then this function will just return layer's name.
:param act: Activation Type.
:param act: Activation Type.
:type act: BaseActivation
:type act: BaseActivation
:param bias_attr: The Bias Attribute. If no bias, then pass False or
:param bias_attr: The Bias Attribute. If the parameter is set to
something not type of ParameterAttribute. None will get a
False or something not type of ParameterAttribute,
default Bias.
no bias is defined. If the parameter is set to
:type bias_attr: ParameterAttribute or None or bool
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param layer_attr: The extra layer config. Default is None.
:param layer_attr: The extra layer config. Default is None.
:type layer_attr: ExtraLayerAttribute
:type layer_attr: ExtraLayerAttribute
:return: MixedLayerType object can add inputs or layer name.
:return: MixedLayerType object can add inputs or layer name.
...
@@ -920,7 +921,7 @@ def data_layer(name, size, depth=None, height=None, width=None,
...
@@ -920,7 +921,7 @@ def data_layer(name, size, depth=None, height=None, width=None,
data = data_layer(name="input", size=1000)
data = data_layer(name="input", size=1000)
:param name:
Name of this data layer
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param size: Size of this data layer.
:param size: Size of this data layer.
:type size: int
:type size: int
...
@@ -960,7 +961,7 @@ def embedding_layer(input, size, name=None, param_attr=None, layer_attr=None):
...
@@ -960,7 +961,7 @@ def embedding_layer(input, size, name=None, param_attr=None, layer_attr=None):
"""
"""
Define a embedding Layer.
Define a embedding Layer.
:param name:
Name of this embedding layer
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: The input layer for this embedding. NOTE: must be Index Data.
:param input: The input layer for this embedding. NOTE: must be Index Data.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -1015,7 +1016,7 @@ def fc_layer(input,
...
@@ -1015,7 +1016,7 @@ def fc_layer(input,
with mixed_layer(size=1024) as fc:
with mixed_layer(size=1024) as fc:
fc += full_matrix_projection(input=layer)
fc += full_matrix_projection(input=layer)
:param name: The
Layer Name
.
:param name: The
name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: The input layer. Could be a list/tuple of input layer.
:param input: The input layer. Could be a list/tuple of input layer.
:type input: LayerOutput|list|tuple
:type input: LayerOutput|list|tuple
...
@@ -1025,10 +1026,11 @@ def fc_layer(input,
...
@@ -1025,10 +1026,11 @@ def fc_layer(input,
:type act: BaseActivation
:type act: BaseActivation
:param param_attr: The Parameter Attribute|list.
:param param_attr: The Parameter Attribute|list.
:type param_attr: ParameterAttribute
:type param_attr: ParameterAttribute
:param bias_attr: The Bias Attribute. If no bias, then pass False or
:param bias_attr: The Bias Attribute. If the parameter is set to
something not type of ParameterAttribute. None will get a
False or something not type of ParameterAttribute,
default Bias.
no bias is defined. If the parameter is set to
:type bias_attr: ParameterAttribute|None|Any
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param layer_attr: Extra Layer config.
:param layer_attr: Extra Layer config.
:type layer_attr: ExtraLayerAttribute|None
:type layer_attr: ExtraLayerAttribute|None
:return: LayerOutput object.
:return: LayerOutput object.
...
@@ -1065,7 +1067,7 @@ def printer_layer(input, format=None, name=None):
...
@@ -1065,7 +1067,7 @@ def printer_layer(input, format=None, name=None):
"""
"""
Print the output value of input layers. This layer is useful for debugging.
Print the output value of input layers. This layer is useful for debugging.
:param name: The
Layer Name
.
:param name: The
name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: The input layer. Could be a list/tuple of input layer.
:param input: The input layer. Could be a list/tuple of input layer.
:type input: LayerOutput|list|tuple
:type input: LayerOutput|list|tuple
...
@@ -1103,7 +1105,7 @@ def priorbox_layer(input,
...
@@ -1103,7 +1105,7 @@ def priorbox_layer(input,
"""
"""
Compute the priorbox and set the variance. This layer is necessary for ssd.
Compute the priorbox and set the variance. This layer is necessary for ssd.
:param name: The
Layer Name
.
:param name: The
name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: The input layer.
:param input: The input layer.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -1152,7 +1154,7 @@ def multibox_loss_layer(input_loc,
...
@@ -1152,7 +1154,7 @@ def multibox_loss_layer(input_loc,
"""
"""
Compute the location loss and the confidence loss for ssd.
Compute the location loss and the confidence loss for ssd.
:param name: The
Layer Name
.
:param name: The
name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input_loc: The input predict locations.
:param input_loc: The input predict locations.
:type input_loc: LayerOutput | List of LayerOutput
:type input_loc: LayerOutput | List of LayerOutput
...
@@ -1227,7 +1229,7 @@ def detection_output_layer(input_loc,
...
@@ -1227,7 +1229,7 @@ def detection_output_layer(input_loc,
box location. The output's shape of this layer could be zero if there is
box location. The output's shape of this layer could be zero if there is
no valid bounding box.
no valid bounding box.
:param name: The
Layer Name
.
:param name: The
name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input_loc: The input predict locations.
:param input_loc: The input predict locations.
:type input_loc: LayerOutput | List of LayerOutput.
:type input_loc: LayerOutput | List of LayerOutput.
...
@@ -1299,7 +1301,7 @@ def cross_channel_norm_layer(input, name=None, param_attr=None):
...
@@ -1299,7 +1301,7 @@ def cross_channel_norm_layer(input, name=None, param_attr=None):
a conv layer's output and scale the output by a group of trainable
a conv layer's output and scale the output by a group of trainable
factors which dimensions equal to the channel's number.
factors which dimensions equal to the channel's number.
:param name: The
Layer Name
.
:param name: The
name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: The input layer.
:param input: The input layer.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -1364,7 +1366,7 @@ def pooling_layer(input,
...
@@ -1364,7 +1366,7 @@ def pooling_layer(input,
:param agg_level: AggregateLevel.TO_NO_SEQUENCE or
:param agg_level: AggregateLevel.TO_NO_SEQUENCE or
AggregateLevel.TO_SEQUENCE
AggregateLevel.TO_SEQUENCE
:type agg_level: AggregateLevel
:type agg_level: AggregateLevel
:param name:
layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: input layer name.
:param input: input layer name.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -1373,8 +1375,11 @@ def pooling_layer(input,
...
@@ -1373,8 +1375,11 @@ def pooling_layer(input,
:type pooling_type: BasePoolingType|None
:type pooling_type: BasePoolingType|None
:param stride: The step size between successive pooling regions.
:param stride: The step size between successive pooling regions.
:type stride: Int
:type stride: Int
:param bias_attr: Bias parameter attribute. False if no bias.
:param bias_attr: The Bias Attribute. If the parameter is set to
:type bias_attr: ParameterAttribute|None|False
False or something not type of ParameterAttribute,
no bias is defined. If the parameter is set to
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param layer_attr: The Extra Attributes for layer, such as dropout.
:param layer_attr: The Extra Attributes for layer, such as dropout.
:type layer_attr: ExtraLayerAttribute|None
:type layer_attr: ExtraLayerAttribute|None
:return: LayerOutput object.
:return: LayerOutput object.
...
@@ -1471,10 +1476,11 @@ def lstmemory(input,
...
@@ -1471,10 +1476,11 @@ def lstmemory(input,
:type gate_act: BaseActivation
:type gate_act: BaseActivation
:param state_act: state activation type, TanhActivation by default.
:param state_act: state activation type, TanhActivation by default.
:type state_act: BaseActivation
:type state_act: BaseActivation
:param bias_attr: The Bias Attribute. If the parameter is set to
:param bias_attr: Bias attribute. None means default bias. False means no
False or something not type of ParameterAttribute,
bias.
no bias is defined. If the parameter is set to
:type bias_attr: ParameterAttribute|None|False
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param param_attr: Parameter Attribute.
:param param_attr: Parameter Attribute.
:type param_attr: ParameterAttribute|None|False
:type param_attr: ParameterAttribute|None|False
:param layer_attr: Extra Layer attribute
:param layer_attr: Extra Layer attribute
...
@@ -1596,9 +1602,11 @@ def grumemory(input,
...
@@ -1596,9 +1602,11 @@ def grumemory(input,
This activation affects the :math:`z_t` and :math:`r_t`. It is the
This activation affects the :math:`z_t` and :math:`r_t`. It is the
:math:`
\\
sigma` in the above formula.
:math:`
\\
sigma` in the above formula.
:type gate_act: BaseActivation
:type gate_act: BaseActivation
:param bias_attr: Bias attribute. None means default bias. False means no
:param bias_attr: The Bias Attribute. If the parameter is set to
bias.
False or something not type of ParameterAttribute,
:type bias_attr: ParameterAttribute|None|False
no bias is defined. If the parameter is set to
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param param_attr: Parameter Attribute.
:param param_attr: Parameter Attribute.
:type param_attr: ParameterAttribute|None|False
:type param_attr: ParameterAttribute|None|False
:param layer_attr: Extra Layer attribute
:param layer_attr: Extra Layer attribute
...
@@ -1657,7 +1665,7 @@ def last_seq(input,
...
@@ -1657,7 +1665,7 @@ def last_seq(input,
seq = last_seq(input=layer)
seq = last_seq(input=layer)
:param agg_level: Aggregated level
:param agg_level: Aggregated level
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: Input layer name.
:param input: Input layer name.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -1713,7 +1721,7 @@ def first_seq(input,
...
@@ -1713,7 +1721,7 @@ def first_seq(input,
seq = first_seq(input=layer)
seq = first_seq(input=layer)
:param agg_level: aggregation level
:param agg_level: aggregation level
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: Input layer name.
:param input: Input layer name.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -1792,11 +1800,13 @@ def expand_layer(input,
...
@@ -1792,11 +1800,13 @@ def expand_layer(input,
:type input: LayerOutput
:type input: LayerOutput
:param expand_as: Expand as this layer's sequence info.
:param expand_as: Expand as this layer's sequence info.
:type expand_as: LayerOutput
:type expand_as: LayerOutput
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param bias_attr: Bias attribute. None means default bias. False means no
:param bias_attr: The Bias Attribute. If the parameter is set to
bias.
False or something not type of ParameterAttribute,
:type bias_attr: ParameterAttribute|None|False
no bias is defined. If the parameter is set to
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param expand_level: whether input layer is timestep(default) or sequence.
:param expand_level: whether input layer is timestep(default) or sequence.
:type expand_level: ExpandLevel
:type expand_level: ExpandLevel
:param layer_attr: extra layer attributes.
:param layer_attr: extra layer attributes.
...
@@ -1849,7 +1859,7 @@ def repeat_layer(input,
...
@@ -1849,7 +1859,7 @@ def repeat_layer(input,
:type input: LayerOutput
:type input: LayerOutput
:param num_repeats: Repeat the input so many times
:param num_repeats: Repeat the input so many times
:type num_repeats: int
:type num_repeats: int
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:param as_row_vector: True for treating input as row vector and repeating
:param as_row_vector: True for treating input as row vector and repeating
in the column direction. This is equivalent to apply
in the column direction. This is equivalent to apply
concat_layer() with num_repeats same input.
concat_layer() with num_repeats same input.
...
@@ -1908,16 +1918,17 @@ def seq_reshape_layer(input,
...
@@ -1908,16 +1918,17 @@ def seq_reshape_layer(input,
:type input: LayerOutput
:type input: LayerOutput
:param reshape_size: the size of reshaped sequence.
:param reshape_size: the size of reshaped sequence.
:type reshape_size: int
:type reshape_size: int
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param act: Activation type.
:param act: Activation type.
:type act: BaseActivation
:type act: BaseActivation
:param layer_attr: extra layer attributes.
:param layer_attr: extra layer attributes.
:type layer_attr: ExtraLayerAttribute.
:type layer_attr: ExtraLayerAttribute.
:param bias_attr: The Bias Attribute. If no bias, then pass False or
:param bias_attr: The Bias Attribute. If the parameter is set to
something not type of ParameterAttribute. None will get a
False or something not type of ParameterAttribute,
default Bias.
no bias is defined. If the parameter is set to
:type bias_attr: ParameterAttribute or None or bool
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:return: LayerOutput object.
:return: LayerOutput object.
:rtype: LayerOutput
:rtype: LayerOutput
"""
"""
...
@@ -1960,7 +1971,7 @@ def interpolation_layer(input, weight, name=None, layer_attr=None):
...
@@ -1960,7 +1971,7 @@ def interpolation_layer(input, weight, name=None, layer_attr=None):
:type input: list|tuple
:type input: list|tuple
:param weight: Weight layer.
:param weight: Weight layer.
:type weight: LayerOutput
:type weight: LayerOutput
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param layer_attr: extra layer attributes.
:param layer_attr: extra layer attributes.
:type layer_attr: ExtraLayerAttribute.
:type layer_attr: ExtraLayerAttribute.
...
@@ -2065,7 +2076,7 @@ def power_layer(input, weight, name=None, layer_attr=None):
...
@@ -2065,7 +2076,7 @@ def power_layer(input, weight, name=None, layer_attr=None):
:type input: LayerOutput
:type input: LayerOutput
:param weight: Weight layer.
:param weight: Weight layer.
:type weight: LayerOutput
:type weight: LayerOutput
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param layer_attr: extra layer attributes.
:param layer_attr: extra layer attributes.
:type layer_attr: ExtraLayerAttribute.
:type layer_attr: ExtraLayerAttribute.
...
@@ -2109,7 +2120,7 @@ def scaling_layer(input, weight, name=None, layer_attr=None):
...
@@ -2109,7 +2120,7 @@ def scaling_layer(input, weight, name=None, layer_attr=None):
:type input: LayerOutput
:type input: LayerOutput
:param weight: Weight layer.
:param weight: Weight layer.
:type weight: LayerOutput
:type weight: LayerOutput
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param layer_attr: extra layer attributes.
:param layer_attr: extra layer attributes.
:type layer_attr: ExtraLayerAttribute.
:type layer_attr: ExtraLayerAttribute.
...
@@ -2147,7 +2158,7 @@ def trans_layer(input, name=None, layer_attr=None):
...
@@ -2147,7 +2158,7 @@ def trans_layer(input, name=None, layer_attr=None):
:param input: Input layer.
:param input: Input layer.
:type input: LayerOutput
:type input: LayerOutput
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param layer_attr: extra layer attributes.
:param layer_attr: extra layer attributes.
:type layer_attr: ExtraLayerAttribute.
:type layer_attr: ExtraLayerAttribute.
...
@@ -2187,7 +2198,7 @@ def rotate_layer(input, height, width, name=None, layer_attr=None):
...
@@ -2187,7 +2198,7 @@ def rotate_layer(input, height, width, name=None, layer_attr=None):
:type input: LayerOutput
:type input: LayerOutput
:param height: The height of the sample matrix
:param height: The height of the sample matrix
:type height: int
:type height: int
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param layer_attr: extra layer attributes.
:param layer_attr: extra layer attributes.
:type layer_attr: ExtraLayerAttribute.
:type layer_attr: ExtraLayerAttribute.
...
@@ -2232,7 +2243,7 @@ def cos_sim(a, b, scale=1, size=1, name=None, layer_attr=None):
...
@@ -2232,7 +2243,7 @@ def cos_sim(a, b, scale=1, size=1, name=None, layer_attr=None):
cos = cos_sim(a=layer1, b=layer2, size=3)
cos = cos_sim(a=layer1, b=layer2, size=3)
:param name:
layer name
:param name:
The name of this layer. It is optional.
:type name: basestring
:type name: basestring
:param a: input layer a
:param a: input layer a
:type a: LayerOutput
:type a: LayerOutput
...
@@ -2299,11 +2310,13 @@ def hsigmoid(input,
...
@@ -2299,11 +2310,13 @@ def hsigmoid(input,
:type label: LayerOutput
:type label: LayerOutput
:param num_classes: number of classes.
:param num_classes: number of classes.
:type num_classes: int|None
:type num_classes: int|None
:param name:
layer name
:param name:
The name of this layer. It is optional.
:type name: basestring
:type name: basestring
:param bias_attr: Bias attribute. None means default bias.
:param bias_attr: The Bias Attribute. If the parameter is set to
False means no bias.
False or something not type of ParameterAttribute,
:type bias_attr: ParameterAttribute|False
no bias is defined. If the parameter is set to
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param param_attr: Parameter Attribute. None means default parameter.
:param param_attr: Parameter Attribute. None means default parameter.
:type param_attr: ParameterAttribute|None
:type param_attr: ParameterAttribute|None
:param layer_attr: Extra Layer Attribute.
:param layer_attr: Extra Layer Attribute.
...
@@ -2411,7 +2424,7 @@ def img_conv_layer(input,
...
@@ -2411,7 +2424,7 @@ def img_conv_layer(input,
bias_attr=False,
bias_attr=False,
act=ReluActivation())
act=ReluActivation())
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: Layer Input.
:param input: Layer Input.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -2442,9 +2455,11 @@ def img_conv_layer(input,
...
@@ -2442,9 +2455,11 @@ def img_conv_layer(input,
:type dilation: int|tuple|list
:type dilation: int|tuple|list
:param dilation_y: The y dimension of the dilation.
:param dilation_y: The y dimension of the dilation.
:type dilation_y: int
:type dilation_y: int
:param bias_attr: Convolution bias attribute. None means default bias.
:param bias_attr: The Bias Attribute. If the parameter is set to
False means no bias.
False or something not type of ParameterAttribute,
:type bias_attr: ParameterAttribute|False
no bias is defined. If the parameter is set to
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param num_channels: number of input channels. If None will be set
:param num_channels: number of input channels. If None will be set
automatically from previous output.
automatically from previous output.
:type num_channels: int
:type num_channels: int
...
@@ -2835,7 +2850,7 @@ def spp_layer(input,
...
@@ -2835,7 +2850,7 @@ def spp_layer(input,
num_channels=16,
num_channels=16,
pool_type=MaxPooling())
pool_type=MaxPooling())
:param name:
layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: layer's input.
:param input: layer's input.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -2929,7 +2944,7 @@ def img_cmrnorm_layer(input,
...
@@ -2929,7 +2944,7 @@ def img_cmrnorm_layer(input,
norm = img_cmrnorm_layer(input=net, size=5)
norm = img_cmrnorm_layer(input=net, size=5)
:param name:
layer name
.
:param name:
The name of this layer. It is optional
.
:type name: None|basestring
:type name: None|basestring
:param input: layer's input.
:param input: layer's input.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -2992,7 +3007,7 @@ def batch_norm_layer(input,
...
@@ -2992,7 +3007,7 @@ def batch_norm_layer(input,
norm = batch_norm_layer(input=net, act=ReluActivation())
norm = batch_norm_layer(input=net, act=ReluActivation())
:param name:
layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: batch normalization input. Better be linear activation.
:param input: batch normalization input. Better be linear activation.
Because there is an activation inside batch_normalization.
Because there is an activation inside batch_normalization.
...
@@ -3016,7 +3031,7 @@ def batch_norm_layer(input,
...
@@ -3016,7 +3031,7 @@ def batch_norm_layer(input,
:type num_channels: int
:type num_channels: int
:param bias_attr: :math:`
\\
beta`, better be zero when initialize. So the
:param bias_attr: :math:`
\\
beta`, better be zero when initialize. So the
initial_std=0, initial_mean=1 is best practice.
initial_std=0, initial_mean=1 is best practice.
:type bias_attr: ParameterAttribute
:type bias_attr: ParameterAttribute
|None|Bool|Any
:param param_attr: :math:`
\\
gamma`, better be one when initialize. So the
:param param_attr: :math:`
\\
gamma`, better be one when initialize. So the
initial_std=0, initial_mean=1 is best practice.
initial_std=0, initial_mean=1 is best practice.
:type param_attr: ParameterAttribute
:type param_attr: ParameterAttribute
...
@@ -3091,7 +3106,7 @@ def sum_to_one_norm_layer(input, name=None, layer_attr=None):
...
@@ -3091,7 +3106,7 @@ def sum_to_one_norm_layer(input, name=None, layer_attr=None):
:param input: Input layer.
:param input: Input layer.
:type input: LayerOutput
:type input: LayerOutput
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param layer_attr: extra layer attributes.
:param layer_attr: extra layer attributes.
:type layer_attr: ExtraLayerAttribute.
:type layer_attr: ExtraLayerAttribute.
...
@@ -3127,7 +3142,7 @@ def row_l2_norm_layer(input, name=None, layer_attr=None):
...
@@ -3127,7 +3142,7 @@ def row_l2_norm_layer(input, name=None, layer_attr=None):
:param input: Input layer.
:param input: Input layer.
:type input: LayerOutput
:type input: LayerOutput
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param layer_attr: extra layer attributes.
:param layer_attr: extra layer attributes.
:type layer_attr: ExtraLayerAttribute.
:type layer_attr: ExtraLayerAttribute.
...
@@ -3179,16 +3194,18 @@ def addto_layer(input, act=None, name=None, bias_attr=None, layer_attr=None):
...
@@ -3179,16 +3194,18 @@ def addto_layer(input, act=None, name=None, bias_attr=None, layer_attr=None):
dropout here.
dropout here.
Please refer to dropout_layer for details.
Please refer to dropout_layer for details.
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: Input layers. It could be a LayerOutput or list/tuple of
:param input: Input layers. It could be a LayerOutput or list/tuple of
LayerOutput.
LayerOutput.
:type input: LayerOutput|list|tuple
:type input: LayerOutput|list|tuple
:param act: Activation Type, default is tanh.
:param act: Activation Type, default is tanh.
:type act: BaseActivation
:type act: BaseActivation
:param bias_attr: Bias attribute. If False, means no bias. None is default
:param bias_attr: The Bias Attribute. If the parameter is set to
bias.
False or something not type of ParameterAttribute,
:type bias_attr: ParameterAttribute|bool
no bias is defined. If the parameter is set to
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param layer_attr: Extra Layer attribute.
:param layer_attr: Extra Layer attribute.
:type layer_attr: ExtraLayerAttribute
:type layer_attr: ExtraLayerAttribute
:return: LayerOutput object.
:return: LayerOutput object.
...
@@ -3237,7 +3254,7 @@ def concat_layer(input, act=None, name=None, layer_attr=None, bias_attr=None):
...
@@ -3237,7 +3254,7 @@ def concat_layer(input, act=None, name=None, layer_attr=None, bias_attr=None):
concat = concat_layer(input=[layer1, layer2])
concat = concat_layer(input=[layer1, layer2])
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: input layers or projections
:param input: input layers or projections
:type input: list|tuple|collections.Sequence
:type input: list|tuple|collections.Sequence
...
@@ -3330,7 +3347,7 @@ def seq_concat_layer(a, b, act=None, name=None, layer_attr=None,
...
@@ -3330,7 +3347,7 @@ def seq_concat_layer(a, b, act=None, name=None, layer_attr=None,
concat = seq_concat_layer(a=layer1, b=layer2)
concat = seq_concat_layer(a=layer1, b=layer2)
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param a: input sequence layer
:param a: input sequence layer
:type a: LayerOutput
:type a: LayerOutput
...
@@ -3340,10 +3357,11 @@ def seq_concat_layer(a, b, act=None, name=None, layer_attr=None,
...
@@ -3340,10 +3357,11 @@ def seq_concat_layer(a, b, act=None, name=None, layer_attr=None,
:type act: BaseActivation
:type act: BaseActivation
:param layer_attr: Extra Layer Attribute.
:param layer_attr: Extra Layer Attribute.
:type layer_attr: ExtraLayerAttribute
:type layer_attr: ExtraLayerAttribute
:param bias_attr: The Bias Attribute. If no bias, then pass False or
:param bias_attr: The Bias Attribute. If the parameter is set to
something not type of ParameterAttribute. None will get a
False or something not type of ParameterAttribute,
default Bias.
no bias is defined. If the parameter is set to
:type bias_attr: ParameterAttribute or None or bool
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:return: LayerOutput object.
:return: LayerOutput object.
:rtype: LayerOutput
:rtype: LayerOutput
"""
"""
...
@@ -3506,7 +3524,7 @@ def lstm_step_layer(input,
...
@@ -3506,7 +3524,7 @@ def lstm_step_layer(input,
output is :math:`o_t`, whose name is 'state' and can use
output is :math:`o_t`, whose name is 'state' and can use
:code:`get_output_layer` to extract this output.
:code:`get_output_layer` to extract this output.
:param name:
Layer's name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param size: Layer's size. NOTE: lstm layer's size, should be equal to
:param size: Layer's size. NOTE: lstm layer's size, should be equal to
:code:`input.size/4`, and should be equal to
:code:`input.size/4`, and should be equal to
...
@@ -3524,8 +3542,11 @@ def lstm_step_layer(input,
...
@@ -3524,8 +3542,11 @@ def lstm_step_layer(input,
:param state_act: State Activation Type. Default is sigmoid, and should
:param state_act: State Activation Type. Default is sigmoid, and should
be sigmoid only.
be sigmoid only.
:type state_act: BaseActivation
:type state_act: BaseActivation
:param bias_attr: Bias Attribute.
:param bias_attr: The Bias Attribute. If the parameter is set to
:type bias_attr: ParameterAttribute
False or something not type of ParameterAttribute,
no bias is defined. If the parameter is set to
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param layer_attr: layer's extra attribute.
:param layer_attr: layer's extra attribute.
:type layer_attr: ExtraLayerAttribute
:type layer_attr: ExtraLayerAttribute
:return: LayerOutput object.
:return: LayerOutput object.
...
@@ -3576,9 +3597,13 @@ def gru_step_layer(input,
...
@@ -3576,9 +3597,13 @@ def gru_step_layer(input,
:param output_mem:
:param output_mem:
:param size:
:param size:
:param act:
:param act:
:param name:
:param name:
The name of this layer. It is optional.
:param gate_act:
:param gate_act:
:param bias_attr:
:param bias_attr: The Bias Attribute. If the parameter is set to
False or something not type of ParameterAttribute,
no bias is defined. If the parameter is set to
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param param_attr: the parameter_attribute for transforming the output_mem
:param param_attr: the parameter_attribute for transforming the output_mem
from previous step.
from previous step.
:param layer_attr:
:param layer_attr:
...
@@ -3632,10 +3657,14 @@ def gru_step_naive_layer(input,
...
@@ -3632,10 +3657,14 @@ def gru_step_naive_layer(input,
:param input:
:param input:
:param output_mem:
:param output_mem:
:param size:
:param size:
:param name:
:param name:
The name of this layer. It is optional.
:param act:
:param act:
:param gate_act:
:param gate_act:
:param bias_attr:
:param bias_attr: The Bias Attribute. If the parameter is set to
False or something not type of ParameterAttribute,
no bias is defined. If the parameter is set to
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param param_attr:
:param param_attr:
:param layer_attr:
:param layer_attr:
:return:
:return:
...
@@ -3691,7 +3720,7 @@ def get_output_layer(input, arg_name, name=None, layer_attr=None):
...
@@ -3691,7 +3720,7 @@ def get_output_layer(input, arg_name, name=None, layer_attr=None):
output besides the default one, please use get_output_layer first to get
output besides the default one, please use get_output_layer first to get
the output from input.
the output from input.
:param name:
Layer's name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: get output layer's input. And this layer should contains
:param input: get output layer's input. And this layer should contains
multiple outputs.
multiple outputs.
...
@@ -3757,11 +3786,14 @@ def recurrent_layer(input,
...
@@ -3757,11 +3786,14 @@ def recurrent_layer(input,
:type input: LayerOutput
:type input: LayerOutput
:param act: activation.
:param act: activation.
:type act: BaseActivation
:type act: BaseActivation
:param bias_attr: bias attribute.
:param bias_attr: The Bias Attribute. If the parameter is set to
:type bias_attr: ParameterAttribute
False or something not type of ParameterAttribute,
no bias is defined. If the parameter is set to
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param param_attr: parameter attribute.
:param param_attr: parameter attribute.
:type param_attr: ParameterAttribute
:type param_attr: ParameterAttribute
:param name:
name of the layer
:param name:
The name of this layer. It is optional.
:type name: basestring
:type name: basestring
:param layer_attr: Layer Attribute.
:param layer_attr: Layer Attribute.
:type layer_attr: ExtraLayerAttribute
:type layer_attr: ExtraLayerAttribute
...
@@ -4000,7 +4032,7 @@ def maxid_layer(input, name=None, layer_attr=None):
...
@@ -4000,7 +4032,7 @@ def maxid_layer(input, name=None, layer_attr=None):
:param input: Input layer name.
:param input: Input layer name.
:type input: LayerOutput
:type input: LayerOutput
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param layer_attr: extra layer attributes.
:param layer_attr: extra layer attributes.
:type layer_attr: ExtraLayerAttribute.
:type layer_attr: ExtraLayerAttribute.
...
@@ -4033,7 +4065,7 @@ def out_prod_layer(input1, input2, name=None, layer_attr=None):
...
@@ -4033,7 +4065,7 @@ def out_prod_layer(input1, input2, name=None, layer_attr=None):
out_prod = out_prod_layer(input1=vec1, input2=vec2)
out_prod = out_prod_layer(input1=vec1, input2=vec2)
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input1: The first input layer name.
:param input1: The first input layer name.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -4074,7 +4106,7 @@ def eos_layer(input, eos_id, name=None, layer_attr=None):
...
@@ -4074,7 +4106,7 @@ def eos_layer(input, eos_id, name=None, layer_attr=None):
eos = eos_layer(input=layer, eos_id=id)
eos = eos_layer(input=layer, eos_id=id)
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: Input layer name.
:param input: Input layer name.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -4265,7 +4297,7 @@ def square_error_cost(input,
...
@@ -4265,7 +4297,7 @@ def square_error_cost(input,
cost =
\\
sum_{i=1}^N(t_i-y_i)^2
cost =
\\
sum_{i=1}^N(t_i-y_i)^2
:param name:
layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: Network prediction.
:param input: Network prediction.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -4307,7 +4339,7 @@ def classification_cost(input,
...
@@ -4307,7 +4339,7 @@ def classification_cost(input,
"""
"""
classification cost Layer.
classification cost Layer.
:param name:
layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: input layer name. network output.
:param input: input layer name. network output.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -4611,7 +4643,7 @@ def pad_layer(input,
...
@@ -4611,7 +4643,7 @@ def pad_layer(input,
:type pad_w: list|None
:type pad_w: list|None
:param layer_attr: Extra Layer Attribute.
:param layer_attr: Extra Layer Attribute.
:type layer_attr: ExtraLayerAttribute
:type layer_attr: ExtraLayerAttribute
:param name:
layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:return: LayerOutput object.
:return: LayerOutput object.
:rtype: LayerOutput
:rtype: LayerOutput
...
@@ -4679,7 +4711,7 @@ def conv_shift_layer(a, b, name=None, layer_attr=None):
...
@@ -4679,7 +4711,7 @@ def conv_shift_layer(a, b, name=None, layer_attr=None):
conv_shift = conv_shift_layer(a=layer1, b=layer2)
conv_shift = conv_shift_layer(a=layer1, b=layer2)
:param name:
layer name
:param name:
The name of this layer. It is optional.
:type name: basestring
:type name: basestring
:param a: Input layer a.
:param a: Input layer a.
:type a: LayerOutput
:type a: LayerOutput
...
@@ -4735,7 +4767,7 @@ def tensor_layer(a,
...
@@ -4735,7 +4767,7 @@ def tensor_layer(a,
tensor = tensor_layer(a=layer1, b=layer2, size=1000)
tensor = tensor_layer(a=layer1, b=layer2, size=1000)
:param name:
layer name
:param name:
The name of this layer. It is optional.
:type name: basestring
:type name: basestring
:param a: Input layer a.
:param a: Input layer a.
:type a: LayerOutput
:type a: LayerOutput
...
@@ -4747,10 +4779,11 @@ def tensor_layer(a,
...
@@ -4747,10 +4779,11 @@ def tensor_layer(a,
:type act: BaseActivation
:type act: BaseActivation
:param param_attr: The Parameter Attribute.
:param param_attr: The Parameter Attribute.
:type param_attr: ParameterAttribute
:type param_attr: ParameterAttribute
:param bias_attr: The Bias Attribute. If no bias, then pass False or
:param bias_attr: The Bias Attribute. If the parameter is set to
something not type of ParameterAttribute. None will get a
False or something not type of ParameterAttribute,
default Bias.
no bias is defined. If the parameter is set to
:type bias_attr: ParameterAttribute|None|Any
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param layer_attr: Extra Layer config.
:param layer_attr: Extra Layer config.
:type layer_attr: ExtraLayerAttribute|None
:type layer_attr: ExtraLayerAttribute|None
:return: LayerOutput object.
:return: LayerOutput object.
...
@@ -4797,7 +4830,7 @@ def selective_fc_layer(input,
...
@@ -4797,7 +4830,7 @@ def selective_fc_layer(input,
sel_fc = selective_fc_layer(input=input, size=128, act=TanhActivation())
sel_fc = selective_fc_layer(input=input, size=128, act=TanhActivation())
:param name: The
Layer Name
.
:param name: The
name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: The input layer.
:param input: The input layer.
:type input: LayerOutput|list|tuple
:type input: LayerOutput|list|tuple
...
@@ -4811,10 +4844,11 @@ def selective_fc_layer(input,
...
@@ -4811,10 +4844,11 @@ def selective_fc_layer(input,
:type act: BaseActivation
:type act: BaseActivation
:param param_attr: The Parameter Attribute.
:param param_attr: The Parameter Attribute.
:type param_attr: ParameterAttribute
:type param_attr: ParameterAttribute
:param bias_attr: The Bias Attribute. If no bias, then pass False or
:param bias_attr: The Bias Attribute. If the parameter is set to
something not type of ParameterAttribute. None will get a
False or something not type of ParameterAttribute,
default Bias.
no bias is defined. If the parameter is set to
:type bias_attr: ParameterAttribute|None|Any
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param layer_attr: Extra Layer config.
:param layer_attr: Extra Layer config.
:type layer_attr: ExtraLayerAttribute|None
:type layer_attr: ExtraLayerAttribute|None
:return: LayerOutput object.
:return: LayerOutput object.
...
@@ -4870,7 +4904,7 @@ def sampling_id_layer(input, name=None, layer_attr=None):
...
@@ -4870,7 +4904,7 @@ def sampling_id_layer(input, name=None, layer_attr=None):
:param input: The input layer.
:param input: The input layer.
:type input: LayerOutput
:type input: LayerOutput
:param name: The
Layer Name
.
:param name: The
name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param layer_attr: Extra Layer config.
:param layer_attr: Extra Layer config.
:type layer_attr: ExtraLayerAttribute|None
:type layer_attr: ExtraLayerAttribute|None
...
@@ -4908,7 +4942,7 @@ def slope_intercept_layer(input,
...
@@ -4908,7 +4942,7 @@ def slope_intercept_layer(input,
:param input: The input layer.
:param input: The input layer.
:type input: LayerOutput
:type input: LayerOutput
:param name: The
Layer Name
.
:param name: The
name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param slope: the scale factor.
:param slope: the scale factor.
:type slope: float.
:type slope: float.
...
@@ -4972,7 +5006,7 @@ def linear_comb_layer(weights, vectors, size=None, name=None, layer_attr=None):
...
@@ -4972,7 +5006,7 @@ def linear_comb_layer(weights, vectors, size=None, name=None, layer_attr=None):
:type vectors: LayerOutput
:type vectors: LayerOutput
:param size: the dimension of this layer.
:param size: the dimension of this layer.
:type size: int
:type size: int
:param name: The
Layer Name
.
:param name: The
name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param layer_attr: Extra Layer config.
:param layer_attr: Extra Layer config.
:type layer_attr: ExtraLayerAttribute|None
:type layer_attr: ExtraLayerAttribute|None
...
@@ -5055,7 +5089,7 @@ def block_expand_layer(input,
...
@@ -5055,7 +5089,7 @@ def block_expand_layer(input,
:type padding_x: int
:type padding_x: int
:param padding_y: The padding size in vertical direction.
:param padding_y: The padding size in vertical direction.
:type padding_y: int
:type padding_y: int
:param name: The name of this layer
, which can not specify
.
:param name: The name of this layer
. It is optional
.
:type name: None|basestring.
:type name: None|basestring.
:param layer_attr: Extra Layer config.
:param layer_attr: Extra Layer config.
:type layer_attr: ExtraLayerAttribute|None
:type layer_attr: ExtraLayerAttribute|None
...
@@ -5124,7 +5158,7 @@ def maxout_layer(input, groups, num_channels=None, name=None, layer_attr=None):
...
@@ -5124,7 +5158,7 @@ def maxout_layer(input, groups, num_channels=None, name=None, layer_attr=None):
:type num_channels: int|None
:type num_channels: int|None
:param groups: The group number of input layer.
:param groups: The group number of input layer.
:type groups: int
:type groups: int
:param name: The name of this layer
, which can not specify
.
:param name: The name of this layer
. It is optional
.
:type name: None|basestring.
:type name: None|basestring.
:param layer_attr: Extra Layer attribute.
:param layer_attr: Extra Layer attribute.
:type layer_attr: ExtraLayerAttribute
:type layer_attr: ExtraLayerAttribute
...
@@ -5188,7 +5222,7 @@ def ctc_layer(input,
...
@@ -5188,7 +5222,7 @@ def ctc_layer(input,
:type label: LayerOutput
:type label: LayerOutput
:param size: category numbers + 1.
:param size: category numbers + 1.
:type size: int
:type size: int
:param name: The name of this layer
:param name: The name of this layer
. It is optional.
:type name: basestring|None
:type name: basestring|None
:param norm_by_times: Whether to normalization by times. False by default.
:param norm_by_times: Whether to normalization by times. False by default.
:type norm_by_times: bool
:type norm_by_times: bool
...
@@ -5265,7 +5299,7 @@ def warp_ctc_layer(input,
...
@@ -5265,7 +5299,7 @@ def warp_ctc_layer(input,
:type label: LayerOutput
:type label: LayerOutput
:param size: category numbers + 1.
:param size: category numbers + 1.
:type size: int
:type size: int
:param name: The name of this layer
, which can not specify
.
:param name: The name of this layer
. It is optional
.
:type name: basestring|None
:type name: basestring|None
:param blank: the 'blank' label used in ctc
:param blank: the 'blank' label used in ctc
:type blank: int
:type blank: int
...
@@ -5329,7 +5363,7 @@ def crf_layer(input,
...
@@ -5329,7 +5363,7 @@ def crf_layer(input,
:type weight: LayerOutput
:type weight: LayerOutput
:param param_attr: Parameter attribute. None means default attribute
:param param_attr: Parameter attribute. None means default attribute
:type param_attr: ParameterAttribute
:type param_attr: ParameterAttribute
:param name: The name of this layer
s. It is not necessary
.
:param name: The name of this layer
. It is optional
.
:type name: None|basestring
:type name: None|basestring
:param coeff: The coefficient affects the gradient in the backward.
:param coeff: The coefficient affects the gradient in the backward.
:type coeff: float
:type coeff: float
...
@@ -5399,7 +5433,7 @@ def crf_decoding_layer(input,
...
@@ -5399,7 +5433,7 @@ def crf_decoding_layer(input,
:type label: LayerOutput or None
:type label: LayerOutput or None
:param param_attr: Parameter attribute. None means default attribute
:param param_attr: Parameter attribute. None means default attribute
:type param_attr: ParameterAttribute
:type param_attr: ParameterAttribute
:param name: The name of this layer
s. It is not necessary
.
:param name: The name of this layer
. It is optional
.
:type name: None|basestring
:type name: None|basestring
:param layer_attr: Extra Layer config.
:param layer_attr: Extra Layer config.
:type layer_attr: ExtraLayerAttribute|None
:type layer_attr: ExtraLayerAttribute|None
...
@@ -5458,9 +5492,9 @@ def nce_layer(input,
...
@@ -5458,9 +5492,9 @@ def nce_layer(input,
param_attr=[attr1, attr2], weight=layer3,
param_attr=[attr1, attr2], weight=layer3,
num_classes=3, neg_distribution=[0.1,0.3,0.6])
num_classes=3, neg_distribution=[0.1,0.3,0.6])
:param name:
layer name
:param name:
The name of this layer. It is optional.
:type name: basestring
:type name: basestring
:param input: input layers. It could be a LayerOutput of list/tuple of LayerOutput.
:param input:
The
input layers. It could be a LayerOutput of list/tuple of LayerOutput.
:type input: LayerOutput|list|tuple|collections.Sequence
:type input: LayerOutput|list|tuple|collections.Sequence
:param label: label layer
:param label: label layer
:type label: LayerOutput
:type label: LayerOutput
...
@@ -5478,8 +5512,11 @@ def nce_layer(input,
...
@@ -5478,8 +5512,11 @@ def nce_layer(input,
A uniform distribution will be used if not provided.
A uniform distribution will be used if not provided.
If not None, its length must be equal to num_classes.
If not None, its length must be equal to num_classes.
:type neg_distribution: list|tuple|collections.Sequence|None
:type neg_distribution: list|tuple|collections.Sequence|None
:param bias_attr: Bias parameter attribute. True if no bias.
:param bias_attr: The Bias Attribute. If the parameter is set to
:type bias_attr: ParameterAttribute|None|False
False or something not type of ParameterAttribute,
no bias is defined. If the parameter is set to
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:param layer_attr: Extra Layer Attribute.
:param layer_attr: Extra Layer Attribute.
:type layer_attr: ExtraLayerAttribute
:type layer_attr: ExtraLayerAttribute
:return: layer name.
:return: layer name.
...
@@ -5594,7 +5631,7 @@ def rank_cost(left,
...
@@ -5594,7 +5631,7 @@ def rank_cost(left,
:param weight: The weight affects the cost, namely the scale of cost.
:param weight: The weight affects the cost, namely the scale of cost.
It is an optional argument.
It is an optional argument.
:type weight: LayerOutput
:type weight: LayerOutput
:param name: The name of this layer
s. It is not necessary
.
:param name: The name of this layer
. It is optional
.
:type name: None|basestring
:type name: None|basestring
:param coeff: The coefficient affects the gradient in the backward.
:param coeff: The coefficient affects the gradient in the backward.
:type coeff: float
:type coeff: float
...
@@ -5648,7 +5685,7 @@ def lambda_cost(input,
...
@@ -5648,7 +5685,7 @@ def lambda_cost(input,
:param score: The 2nd input. Score of each sample.
:param score: The 2nd input. Score of each sample.
:type input: LayerOutput
:type input: LayerOutput
:param NDCG_num: The size of NDCG (Normalized Discounted Cumulative Gain),
:param NDCG_num: The size of NDCG (Normalized Discounted Cumulative Gain),
e.g., 5 for NDCG@5. It must be less than
f
or equal to the
e.g., 5 for NDCG@5. It must be less than or equal to the
minimum size of lists.
minimum size of lists.
:type NDCG_num: int
:type NDCG_num: int
:param max_sort_size: The size of partial sorting in calculating gradient.
:param max_sort_size: The size of partial sorting in calculating gradient.
...
@@ -5659,7 +5696,7 @@ def lambda_cost(input,
...
@@ -5659,7 +5696,7 @@ def lambda_cost(input,
than the size of a list, the algorithm will sort the
than the size of a list, the algorithm will sort the
entire list of get gradient.
entire list of get gradient.
:type max_sort_size: int
:type max_sort_size: int
:param name: The name of this layer
s. It is not necessary
.
:param name: The name of this layer
. It is optional
.
:type name: None|basestring
:type name: None|basestring
:param layer_attr: Extra Layer Attribute.
:param layer_attr: Extra Layer Attribute.
:type layer_attr: ExtraLayerAttribute
:type layer_attr: ExtraLayerAttribute
...
@@ -5703,7 +5740,7 @@ def cross_entropy(input,
...
@@ -5703,7 +5740,7 @@ def cross_entropy(input,
:type input: LayerOutput.
:type input: LayerOutput.
:param label: The input label.
:param label: The input label.
:type input: LayerOutput.
:type input: LayerOutput.
:param name: The name of this layer
s. It is not necessary
.
:param name: The name of this layer
. It is optional
.
:type name: None|basestring.
:type name: None|basestring.
:param coeff: The cost is multiplied with coeff.
:param coeff: The cost is multiplied with coeff.
The coefficient affects the gradient in the backward.
The coefficient affects the gradient in the backward.
...
@@ -5751,7 +5788,7 @@ def cross_entropy_with_selfnorm(input,
...
@@ -5751,7 +5788,7 @@ def cross_entropy_with_selfnorm(input,
:type input: LayerOutput.
:type input: LayerOutput.
:param label: The input label.
:param label: The input label.
:type input: LayerOutput.
:type input: LayerOutput.
:param name: The name of this layer
s. It is not necessary
.
:param name: The name of this layer
. It is optional
.
:type name: None|basestring.
:type name: None|basestring.
:param coeff: The coefficient affects the gradient in the backward.
:param coeff: The coefficient affects the gradient in the backward.
:type coeff: float.
:type coeff: float.
...
@@ -5791,7 +5828,7 @@ def sum_cost(input, name=None, layer_attr=None):
...
@@ -5791,7 +5828,7 @@ def sum_cost(input, name=None, layer_attr=None):
:param input: The first input layer.
:param input: The first input layer.
:type input: LayerOutput.
:type input: LayerOutput.
:param name: The name of this layer
s. It is not necessary
.
:param name: The name of this layer
. It is optional
.
:type name: None|basestring.
:type name: None|basestring.
:param layer_attr: Extra Layer Attribute.
:param layer_attr: Extra Layer Attribute.
:type layer_attr: ExtraLayerAttribute
:type layer_attr: ExtraLayerAttribute
...
@@ -5836,7 +5873,7 @@ def huber_regression_cost(input,
...
@@ -5836,7 +5873,7 @@ def huber_regression_cost(input,
:type input: LayerOutput.
:type input: LayerOutput.
:param label: The input label.
:param label: The input label.
:type input: LayerOutput.
:type input: LayerOutput.
:param name: The name of this layer
s. It is not necessary
.
:param name: The name of this layer
. It is optional
.
:type name: None|basestring.
:type name: None|basestring.
:param delta: The difference between the observed and predicted values.
:param delta: The difference between the observed and predicted values.
:type delta: float.
:type delta: float.
...
@@ -5886,7 +5923,7 @@ def huber_classification_cost(input,
...
@@ -5886,7 +5923,7 @@ def huber_classification_cost(input,
:type input: LayerOutput.
:type input: LayerOutput.
:param label: The input label.
:param label: The input label.
:type input: LayerOutput.
:type input: LayerOutput.
:param name: The name of this layer
s. It is not necessary
.
:param name: The name of this layer
. It is optional
.
:type name: None|basestring.
:type name: None|basestring.
:param coeff: The coefficient affects the gradient in the backward.
:param coeff: The coefficient affects the gradient in the backward.
:type coeff: float.
:type coeff: float.
...
@@ -5929,7 +5966,7 @@ def multi_binary_label_cross_entropy(input,
...
@@ -5929,7 +5966,7 @@ def multi_binary_label_cross_entropy(input,
:type input: LayerOutput
:type input: LayerOutput
:param label: The input label.
:param label: The input label.
:type input: LayerOutput
:type input: LayerOutput
:param name: The name of this layer
s. It is not necessary
.
:param name: The name of this layer
. It is optional
.
:type name: None|basestring
:type name: None|basestring
:param coeff: The coefficient affects the gradient in the backward.
:param coeff: The coefficient affects the gradient in the backward.
:type coeff: float
:type coeff: float
...
@@ -6034,9 +6071,9 @@ def cross_entropy_over_beam(input, name=None):
...
@@ -6034,9 +6071,9 @@ def cross_entropy_over_beam(input, name=None):
])
])
:param input:
i
nput beams for this layer.
:param input:
I
nput beams for this layer.
:type input: BeamInput
:type input: BeamInput
:param name:
input beams for
this layer.
:param name:
The name of
this layer.
:type name: basestring
:type name: basestring
:return: LayerOutput object.
:return: LayerOutput object.
:rtype: LayerOutput
:rtype: LayerOutput
...
@@ -6097,7 +6134,7 @@ def smooth_l1_cost(input, label, name=None, coeff=1.0, layer_attr=None):
...
@@ -6097,7 +6134,7 @@ def smooth_l1_cost(input, label, name=None, coeff=1.0, layer_attr=None):
:type input: LayerOutput
:type input: LayerOutput
:param label: The input label.
:param label: The input label.
:type input: LayerOutput
:type input: LayerOutput
:param name: The name of this layer
s. It is not necessary
.
:param name: The name of this layer
. It is optional
.
:type name: None|basestring
:type name: None|basestring
:param coeff: The coefficient affects the gradient in the backward.
:param coeff: The coefficient affects the gradient in the backward.
:type coeff: float
:type coeff: float
...
@@ -6145,7 +6182,7 @@ def multiplex_layer(input, name=None, layer_attr=None):
...
@@ -6145,7 +6182,7 @@ def multiplex_layer(input, name=None, layer_attr=None):
:param input: Input layers.
:param input: Input layers.
:type input: list of LayerOutput
:type input: list of LayerOutput
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param layer_attr: extra layer attributes.
:param layer_attr: extra layer attributes.
:type layer_attr: ExtraLayerAttribute.
:type layer_attr: ExtraLayerAttribute.
...
@@ -6176,12 +6213,21 @@ def multiplex_layer(input, name=None, layer_attr=None):
...
@@ -6176,12 +6213,21 @@ def multiplex_layer(input, name=None, layer_attr=None):
@
wrap_name_default
(
"dropout"
)
@
wrap_name_default
(
"dropout"
)
def
dropout_layer
(
input
,
dropout_rate
,
name
=
None
):
def
dropout_layer
(
input
,
dropout_rate
,
name
=
None
):
"""
"""
@TODO(yuyang18): Add comments.
:param name:
The example usage is:
:param input:
:param dropout_rate:
.. code-block:: python
:return:
dropout = dropout_layer(input=input_layer, dropout_rate=0.5)
:param name: The name of this layer. It is optional.
:type name: basestring
:param input: The input layer.
:type input: LayerOutput
:param dropout_rate: The probability of dropout.
:type dropout_rate: float
:return: LayerOutput object.
:rtype: LayerOutput
"""
"""
return
addto_layer
(
return
addto_layer
(
name
=
name
,
name
=
name
,
...
@@ -6204,7 +6250,7 @@ def row_conv_layer(input,
...
@@ -6204,7 +6250,7 @@ def row_conv_layer(input,
"""
"""
The row convolution is called lookahead convolution. It is firstly
The row convolution is called lookahead convolution. It is firstly
introduced in paper of `Deep Speech 2: End-toEnd Speech Recognition
introduced in paper of `Deep Speech 2: End-to
-
End Speech Recognition
in English and Mandarin <https://arxiv.org/pdf/1512.02595v1.pdf>`_ .
in English and Mandarin <https://arxiv.org/pdf/1512.02595v1.pdf>`_ .
The bidirectional RNN that learns representation for a sequence by
The bidirectional RNN that learns representation for a sequence by
...
@@ -6212,9 +6258,9 @@ def row_conv_layer(input,
...
@@ -6212,9 +6258,9 @@ def row_conv_layer(input,
However, unlike unidirectional RNNs, bidirectional RNNs are challenging
However, unlike unidirectional RNNs, bidirectional RNNs are challenging
to deploy in an online and low-latency setting. The lookahead convolution
to deploy in an online and low-latency setting. The lookahead convolution
incorporates information from future subsequences in a computationally
incorporates information from future subsequences in a computationally
efficient manner to improve unidirectional
recurrent neural network
s.
efficient manner to improve unidirectional
RNN
s.
The connection of row convolution is different f
or
m the 1D sequence
The connection of row convolution is different f
ro
m the 1D sequence
convolution. Assumed that, the future context-length is k, that is to say,
convolution. Assumed that, the future context-length is k, that is to say,
it can get the output at timestep t by using the the input feature from t-th
it can get the output at timestep t by using the the input feature from t-th
timestep to (t+k+1)-th timestep. Assumed that the hidden dim of input
timestep to (t+k+1)-th timestep. Assumed that the hidden dim of input
...
@@ -6243,7 +6289,7 @@ def row_conv_layer(input,
...
@@ -6243,7 +6289,7 @@ def row_conv_layer(input,
:param act: Activation Type. Default is linear activation.
:param act: Activation Type. Default is linear activation.
:type act: BaseActivation
:type act: BaseActivation
:param param_attr: The Parameter Attribute. If None, the parameter will be
:param param_attr: The Parameter Attribute. If None, the parameter will be
initialized smartly. It's better set it by yourself.
initialized smartly. It's better
to
set it by yourself.
:type param_attr: ParameterAttribute
:type param_attr: ParameterAttribute
:param layer_attr: Extra Layer config.
:param layer_attr: Extra Layer config.
:type layer_attr: ExtraLayerAttribute|None
:type layer_attr: ExtraLayerAttribute|None
...
@@ -6290,7 +6336,7 @@ def prelu_layer(input,
...
@@ -6290,7 +6336,7 @@ def prelu_layer(input,
prelu = prelu_layer(input=layers, partial_sum=1)
prelu = prelu_layer(input=layers, partial_sum=1)
:param name:
Name of this layer
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: The input layer.
:param input: The input layer.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -6343,7 +6389,7 @@ def gated_unit_layer(input,
...
@@ -6343,7 +6389,7 @@ def gated_unit_layer(input,
The gated unit layer implements a simple gating mechanism over the input.
The gated unit layer implements a simple gating mechanism over the input.
The input :math:`X` is first projected into a new space :math:`X'`, and
The input :math:`X` is first projected into a new space :math:`X'`, and
it is also used to produce a gate weight :math:`\sigma`. Element-wise
it is also used to produce a gate weight :math:`\sigma`. Element-wise
prod
i
ct between :match:`X'` and :math:`\sigma` is finally returned.
prod
u
ct between :match:`X'` and :math:`\sigma` is finally returned.
Reference:
Reference:
Language Modeling with Gated Convolutional Networks
Language Modeling with Gated Convolutional Networks
...
@@ -6363,7 +6409,7 @@ def gated_unit_layer(input,
...
@@ -6363,7 +6409,7 @@ def gated_unit_layer(input,
:type size: int
:type size: int
:param act: activation type of the projected input.
:param act: activation type of the projected input.
:type act: BaseActivation
:type act: BaseActivation
:param name:
name of this layer
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param gate_attr: Attributes to tune the gate output, for example, error
:param gate_attr: Attributes to tune the gate output, for example, error
clipping threshold, dropout and so on. See ExtraLayerAttribute for
clipping threshold, dropout and so on. See ExtraLayerAttribute for
...
@@ -6439,10 +6485,10 @@ def switch_order_layer(input,
...
@@ -6439,10 +6485,10 @@ def switch_order_layer(input,
:param input: The input layer.
:param input: The input layer.
:type input: LayerOutput
:type input: LayerOutput
:param name:
Name of this layer
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param reshape
: reshape matrix by axises
.
:param reshape
_axis: Specify the axises of 'height'. Its value should be positive and less than 4
.
:type reshape
: Dic
t
:type reshape
_axis: in
t
:return: LayerOutput object.
:return: LayerOutput object.
:rtype: LayerOutput
:rtype: LayerOutput
"""
"""
...
@@ -6492,7 +6538,7 @@ def crop_layer(input, offset, axis=2, shape=None, name=None, layer_attr=None):
...
@@ -6492,7 +6538,7 @@ def crop_layer(input, offset, axis=2, shape=None, name=None, layer_attr=None):
:type partial_sum: int
:type partial_sum: int
:param shape: The shape to be cropped. Default is None.
:param shape: The shape to be cropped. Default is None.
:type shape: Sequence | None
:type shape: Sequence | None
:param name:
Name of this layer
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:return: LayerOutput object.
:return: LayerOutput object.
:rtype: LayerOutput
:rtype: LayerOutput
...
@@ -6538,7 +6584,7 @@ def sub_nested_seq_layer(input, selected_indices, name=None):
...
@@ -6538,7 +6584,7 @@ def sub_nested_seq_layer(input, selected_indices, name=None):
:type input: LayerOutput
:type input: LayerOutput
:param selected_indices: a set of sequence indices in the nested sequence.
:param selected_indices: a set of sequence indices in the nested sequence.
:type input: LayerOutput
:type input: LayerOutput
:param name:
name of this layer
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:return: LayerOutput object.
:return: LayerOutput object.
:rtype: LayerOutput
:rtype: LayerOutput
...
@@ -6576,7 +6622,7 @@ def clip_layer(input, min, max, name=None):
...
@@ -6576,7 +6622,7 @@ def clip_layer(input, min, max, name=None):
clip = clip_layer(input=input_layer, min=-10, max=10)
clip = clip_layer(input=input_layer, min=-10, max=10)
:param name: The
Layer Name
.
:param name: The
name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: The input layer.
:param input: The input layer.
:type input: LayerOutput.
:type input: LayerOutput.
...
@@ -6621,7 +6667,7 @@ def seq_slice_layer(input, starts, ends, name=None):
...
@@ -6621,7 +6667,7 @@ def seq_slice_layer(input, starts, ends, name=None):
seq_silce = seq_slice_layer(input=input_seq,
seq_silce = seq_slice_layer(input=input_seq,
starts=start_pos, ends=end_pos)
starts=start_pos, ends=end_pos)
:param name:
name of this layer
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: input for this layer, it should be a sequence.
:param input: input for this layer, it should be a sequence.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -6675,12 +6721,12 @@ def kmax_seq_score_layer(input, name=None, beam_size=1):
...
@@ -6675,12 +6721,12 @@ def kmax_seq_score_layer(input, name=None, beam_size=1):
kmax_indices = kmax_seq_score_layer(input=input_layer, beam_size)
kmax_indices = kmax_seq_score_layer(input=input_layer, beam_size)
:param name: The
Layer Name
.
:param name: The
name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: The input layer. It stores scores over a sequence or a nested
:param input: The input layer. It stores scores over a sequence or a nested
sequence and its size must be 1.
sequence and its size must be 1.
:type input: LayerOutput.
:type input: LayerOutput.
:param beam_size: squence indices with top beam_size scores are returned.
:param beam_size: s
e
quence indices with top beam_size scores are returned.
:type beam_size: double
:type beam_size: double
:return: LayerOutput object.
:return: LayerOutput object.
:rtype: LayerOutput
:rtype: LayerOutput
...
@@ -6733,7 +6779,7 @@ def img_conv3d_layer(input,
...
@@ -6733,7 +6779,7 @@ def img_conv3d_layer(input,
bias_attr=False,
bias_attr=False,
act=ReluActivation())
act=ReluActivation())
:param name:
Layer name
.
:param name:
The name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: Layer Input.
:param input: Layer Input.
:type input: LayerOutput
:type input: LayerOutput
...
@@ -6752,7 +6798,7 @@ def img_conv3d_layer(input,
...
@@ -6752,7 +6798,7 @@ def img_conv3d_layer(input,
:type padding: int|tuple|list
:type padding: int|tuple|list
:param bias_attr: Convolution bias attribute. None means default bias.
:param bias_attr: Convolution bias attribute. None means default bias.
False means no bias.
False means no bias.
:type bias_attr: ParameterAttribute|
False
:type bias_attr: ParameterAttribute|
None|Bool|Any
:param num_channels: number of input channels. If None will be set
:param num_channels: number of input channels. If None will be set
automatically from previous output.
automatically from previous output.
:type num_channels: int
:type num_channels: int
...
@@ -6864,14 +6910,17 @@ def scale_shift_layer(input, name=None, param_attr=None, bias_attr=None):
...
@@ -6864,14 +6910,17 @@ def scale_shift_layer(input, name=None, param_attr=None, bias_attr=None):
scale_shift = scale_shift_layer(input=input_layer, bias_attr=False)
scale_shift = scale_shift_layer(input=input_layer, bias_attr=False)
:param name: The
Layer Name
.
:param name: The
name of this layer. It is optional
.
:type name: basestring
:type name: basestring
:param input: The input layer.
:param input: The input layer.
:type input: LayerOutput.
:type input: LayerOutput.
:param param_attr: The parameter attribute of scaling.
:param param_attr: The parameter attribute of scaling.
:type param_attr: ParameterAttribute
:type param_attr: ParameterAttribute
:param bias_attr: The parameter attribute of shifting.
:param bias_attr: The Bias Attribute. If the parameter is set to
:type bias_attr: ParameterAttribute
False or something not type of ParameterAttribute,
no bias is defined. If the parameter is set to
True, the bias is initialized to zero.
:type bias_attr: ParameterAttribute|None|Bool|Any
:return: LayerOutput object.
:return: LayerOutput object.
:rtype: LayerOutput
:rtype: LayerOutput
"""
"""
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录