提交 4acfa557 编写于 作者: T Travis CI

Deploy to GitHub Pages: 68b958c3

上级 9b1461a4
......@@ -11,8 +11,7 @@ Data layer
data
----
.. automodule:: paddle.v2.layer
:members: data
.. autoclass:: paddle.v2.layer.data
:noindex:
Fully Connected Layers
......@@ -22,14 +21,12 @@ Fully Connected Layers
fc
--
.. automodule:: paddle.v2.layer
:members: fc
.. autoclass:: paddle.v2.layer.fc
:noindex:
selective_fc
------------
.. automodule:: paddle.v2.layer
:members: selective_fc
.. autoclass:: paddle.v2.layer.selective_fc
:noindex:
Conv Layers
......@@ -37,34 +34,29 @@ Conv Layers
conv_operator
-------------
.. automodule:: paddle.v2.layer
:members: conv_operator
.. autoclass:: paddle.v2.layer.conv_operator
:noindex:
conv_projection
---------------
.. automodule:: paddle.v2.layer
:members: conv_projection
.. autoclass:: paddle.v2.layer.conv_projection
:noindex:
conv_shift
----------
.. automodule:: paddle.v2.layer
:members: conv_shift
.. autoclass:: paddle.v2.layer.conv_shift
:noindex:
img_conv
--------
.. automodule:: paddle.v2.layer
:members: img_conv
.. autoclass:: paddle.v2.layer.img_conv
:noindex:
.. _api_v2.layer_context_projection:
context_projection
------------------
.. automodule:: paddle.v2.layer
:members: context_projection
.. autoclass:: paddle.v2.layer.context_projection
:noindex:
Image Pooling Layer
......@@ -72,20 +64,17 @@ Image Pooling Layer
img_pool
--------
.. automodule:: paddle.v2.layer
:members: img_pool
.. autoclass:: paddle.v2.layer.img_pool
:noindex:
spp
---
.. automodule:: paddle.v2.layer
:members: spp
.. autoclass:: paddle.v2.layer.spp
:noindex:
maxout
------
.. automodule:: paddle.v2.layer
:members: maxout
.. autoclass:: paddle.v2.layer.maxout
:noindex:
Norm Layer
......@@ -93,26 +82,22 @@ Norm Layer
img_cmrnorm
-----------
.. automodule:: paddle.v2.layer
:members: img_cmrnorm
.. autoclass:: paddle.v2.layer.img_cmrnorm
:noindex:
batch_norm
----------
.. automodule:: paddle.v2.layer
:members: batch_norm
.. autoclass:: paddle.v2.layer.batch_norm
:noindex:
sum_to_one_norm
---------------
.. automodule:: paddle.v2.layer
:members: sum_to_one_norm
.. autoclass:: paddle.v2.layer.sum_to_one_norm
:noindex:
cross_channel_norm
------------------
.. automodule:: paddle.v2.layer
:members: cross_channel_norm
.. autoclass:: paddle.v2.layer.cross_channel_norm
:noindex:
Recurrent Layers
......@@ -120,20 +105,17 @@ Recurrent Layers
recurrent
---------
.. automodule:: paddle.v2.layer
:members: recurrent
.. autoclass:: paddle.v2.layer.recurrent
:noindex:
lstmemory
---------
.. automodule:: paddle.v2.layer
:members: lstmemory
.. autoclass:: paddle.v2.layer.lstmemory
:noindex:
grumemory
---------
.. automodule:: paddle.v2.layer
:members: grumemory
.. autoclass:: paddle.v2.layer.grumemory
:noindex:
Recurrent Layer Group
......@@ -141,38 +123,32 @@ Recurrent Layer Group
memory
------
.. automodule:: paddle.v2.layer
:members: memory
.. autoclass:: paddle.v2.layer.memory
:noindex:
recurrent_group
---------------
.. automodule:: paddle.v2.layer
:members: recurrent_group
.. autoclass:: paddle.v2.layer.recurrent_group
:noindex:
lstm_step
---------
.. automodule:: paddle.v2.layer
:members: lstm_step
.. autoclass:: paddle.v2.layer.lstm_step
:noindex:
gru_step
--------
.. automodule:: paddle.v2.layer
:members: gru_step
.. autoclass:: paddle.v2.layer.gru_step
:noindex:
beam_search
------------
.. automodule:: paddle.v2.layer
:members: beam_search
.. autoclass:: paddle.v2.layer.beam_search
:noindex:
get_output
----------
.. automodule:: paddle.v2.layer
:members: get_output
.. autoclass:: paddle.v2.layer.get_output
:noindex:
Mixed Layer
......@@ -182,59 +158,50 @@ Mixed Layer
mixed
-----
.. automodule:: paddle.v2.layer
:members: mixed
.. autoclass:: paddle.v2.layer.mixed
:noindex:
.. _api_v2.layer_embedding:
embedding
---------
.. automodule:: paddle.v2.layer
:members: embedding
.. autoclass:: paddle.v2.layer.embedding
:noindex:
scaling_projection
------------------
.. automodule:: paddle.v2.layer
:members: scaling_projection
.. autoclass:: paddle.v2.layer.scaling_projection
:noindex:
dotmul_projection
-----------------
.. automodule:: paddle.v2.layer
:members: dotmul_projection
.. autoclass:: paddle.v2.layer.dotmul_projection
:noindex:
dotmul_operator
---------------
.. automodule:: paddle.v2.layer
:members: dotmul_operator
.. autoclass:: paddle.v2.layer.dotmul_operator
:noindex:
full_matrix_projection
----------------------
.. automodule:: paddle.v2.layer
:members: full_matrix_projection
.. autoclass:: paddle.v2.layer.full_matrix_projection
:noindex:
identity_projection
-------------------
.. automodule:: paddle.v2.layer
:members: identity_projection
.. autoclass:: paddle.v2.layer.identity_projection
:noindex:
table_projection
----------------
.. automodule:: paddle.v2.layer
:members: table_projection
.. autoclass:: paddle.v2.layer.table_projection
:noindex:
trans_full_matrix_projection
----------------------------
.. automodule:: paddle.v2.layer
:members: trans_full_matrix_projection
.. autoclass:: paddle.v2.layer.trans_full_matrix_projection
:noindex:
Aggregate Layers
......@@ -244,36 +211,31 @@ Aggregate Layers
pooling
-------
.. automodule:: paddle.v2.layer
:members: pooling
.. autoclass:: paddle.v2.layer.pooling
:noindex:
.. _api_v2.layer_last_seq:
last_seq
--------
.. automodule:: paddle.v2.layer
:members: last_seq
.. autoclass:: paddle.v2.layer.last_seq
:noindex:
.. _api_v2.layer_first_seq:
first_seq
---------
.. automodule:: paddle.v2.layer
:members: first_seq
.. autoclass:: paddle.v2.layer.first_seq
:noindex:
concat
------
.. automodule:: paddle.v2.layer
:members: concat
.. autoclass:: paddle.v2.layer.concat
:noindex:
seq_concat
----------
.. automodule:: paddle.v2.layer
:members: seq_concat
.. autoclass:: paddle.v2.layer.seq_concat
:noindex:
Reshaping Layers
......@@ -281,34 +243,29 @@ Reshaping Layers
block_expand
------------
.. automodule:: paddle.v2.layer
:members: block_expand
.. autoclass:: paddle.v2.layer.block_expand
:noindex:
.. _api_v2.layer_expand:
expand
------
.. automodule:: paddle.v2.layer
:members: expand
.. autoclass:: paddle.v2.layer.expand
:noindex:
repeat
------
.. automodule:: paddle.v2.layer
:members: repeat
.. autoclass:: paddle.v2.layer.repeat
:noindex:
rotate
------
.. automodule:: paddle.v2.layer
:members: rotate
.. autoclass:: paddle.v2.layer.rotate
:noindex:
seq_reshape
-----------
.. automodule:: paddle.v2.layer
:members: seq_reshape
.. autoclass:: paddle.v2.layer.seq_reshape
:noindex:
Math Layers
......@@ -316,64 +273,54 @@ Math Layers
addto
-----
.. automodule:: paddle.v2.layer
:members: addto
.. autoclass:: paddle.v2.layer.addto
:noindex:
linear_comb
-----------
.. automodule:: paddle.v2.layer
:members: linear_comb
.. autoclass:: paddle.v2.layer.linear_comb
:noindex:
interpolation
-------------
.. automodule:: paddle.v2.layer
:members: interpolation
.. autoclass:: paddle.v2.layer.interpolation
:noindex:
bilinear_interp
---------------
.. automodule:: paddle.v2.layer
:members: bilinear_interp
.. autoclass:: paddle.v2.layer.bilinear_interp
:noindex:
power
-----
.. automodule:: paddle.v2.layer
:members: power
.. autoclass:: paddle.v2.layer.power
:noindex:
scaling
-------
.. automodule:: paddle.v2.layer
:members: scaling
.. autoclass:: paddle.v2.layer.scaling
:noindex:
slope_intercept
---------------
.. automodule:: paddle.v2.layer
:members: slope_intercept
.. autoclass:: paddle.v2.layer.slope_intercept
:noindex:
tensor
------
.. automodule:: paddle.v2.layer
:members: tensor
.. autoclass:: paddle.v2.layer.tensor
:noindex:
.. _api_v2.layer_cos_sim:
cos_sim
-------
.. automodule:: paddle.v2.layer
:members: cos_sim
.. autoclass:: paddle.v2.layer.cos_sim
:noindex:
trans
-----
.. automodule:: paddle.v2.layer
:members: trans
.. autoclass:: paddle.v2.layer.trans
:noindex:
Sampling Layers
......@@ -381,14 +328,12 @@ Sampling Layers
maxid
-----
.. automodule:: paddle.v2.layer
:members: maxid
.. autoclass:: paddle.v2.layer.max_id
:noindex:
sampling_id
-----------
.. automodule:: paddle.v2.layer
:members: sampling_id
.. autoclass:: paddle.v2.layer.sampling_id
:noindex:
Slicing and Joining Layers
......@@ -396,8 +341,7 @@ Slicing and Joining Layers
pad
----
.. automodule:: paddle.v2.layer
:members: pad
.. autoclass:: paddle.v2.layer.pad
:noindex:
.. _api_v2.layer_costs:
......@@ -407,80 +351,72 @@ Cost Layers
cross_entropy_cost
------------------
.. automodule:: paddle.v2.layer
:members: cross_entropy_cost
.. autoclass:: paddle.v2.layer.cross_entropy_cost
:noindex:
cross_entropy_with_selfnorm_cost
--------------------------------
.. automodule:: paddle.v2.layer
:members: cross_entropy_with_selfnorm_cost
.. autoclass:: paddle.v2.layer.cross_entropy_with_selfnorm_cost
:noindex:
multi_binary_label_cross_entropy_cost
-------------------------------------
.. automodule:: paddle.v2.layer
:members: multi_binary_label_cross_entropy_cost
.. autoclass:: paddle.v2.layer.multi_binary_label_cross_entropy_cost
:noindex:
huber_cost
----------
.. automodule:: paddle.v2.layer
:members: huber_cost
.. autoclass:: paddle.v2.layer.huber_cost
:noindex:
lambda_cost
-----------
.. automodule:: paddle.v2.layer
:members: lambda_cost
.. autoclass:: paddle.v2.layer.lambda_cost
:noindex:
mse_cost
--------
.. autoclass:: paddle.v2.layer.mse_cost
:noindex:
rank_cost
---------
.. automodule:: paddle.v2.layer
:members: rank_cost
.. autoclass:: paddle.v2.layer.rank_cost
:noindex:
sum_cost
---------
.. automodule:: paddle.v2.layer
:members: sum_cost
.. autoclass:: paddle.v2.layer.sum_cost
:noindex:
crf
---
.. automodule:: paddle.v2.layer
:members: crf
.. autoclass:: paddle.v2.layer.crf
:noindex:
crf_decoding
------------
.. automodule:: paddle.v2.layer
:members: crf_decoding
.. autoclass:: paddle.v2.layer.crf_decoding
:noindex:
ctc
---
.. automodule:: paddle.v2.layer
:members: ctc
.. autoclass:: paddle.v2.layer.ctc
:noindex:
warp_ctc
--------
.. automodule:: paddle.v2.layer
:members: warp_ctc
.. autoclass:: paddle.v2.layer.warp_ctc
:noindex:
nce
---
.. automodule:: paddle.v2.layer
:members: nce
.. autoclass:: paddle.v2.layer.nce
:noindex:
hsigmoid
---------
.. automodule:: paddle.v2.layer
:members: hsigmoid
.. autoclass:: paddle.v2.layer.hsigmoid
:noindex:
Check Layer
......@@ -488,6 +424,5 @@ Check Layer
eos
---
.. automodule:: paddle.v2.layer
:members: eos
.. autoclass:: paddle.v2.layer.eos
:noindex:
......@@ -270,7 +270,7 @@ layer that not support this attribute, paddle will print an error and core.</p>
<li><strong>drop_rate</strong> (<em>float</em>) &#8211; Dropout rate. Dropout will create a mask on layer output.
The dropout rate is the zero rate of this mask. The
details of what dropout is please refer to <a class="reference external" href="https://www.cs.toronto.edu/~hinton/absps/JMLRdropout.pdf">here</a>.</li>
<li><strong>device</strong> (<em>int</em>) &#8211; <p>device ID of layer. device=-1, use CPU. device&gt;0, use GPU.
<li><strong>device</strong> (<em>int</em>) &#8211; <p>device ID of layer. device=-1, use CPU. device&gt;=0, use GPU.
The details allocation in parallel_nn please refer to <a class="reference external" href="http://www.paddlepaddle.org/doc/ui/cmd_argument/use_case.html#case-2-specify-layers-in-different-devices">here</a>.</p>
</li>
</ul>
......
......@@ -1757,11 +1757,11 @@ It performs element-wise multiplication with weight.</p>
<code class="descclassname">paddle.trainer_config_helpers.layers.</code><code class="descname">dotmul_operator</code><span class="sig-paren">(</span><em>a=None</em>, <em>b=None</em>, <em>scale=1</em>, <em>**kwargs</em><span class="sig-paren">)</span></dt>
<dd><p>DotMulOperator takes two inputs and performs element-wise multiplication:</p>
<div class="math">
\[out.row[i] += scale * (x.row[i] .* y.row[i])\]</div>
\[out.row[i] += scale * (a.row[i] .* b.row[i])\]</div>
<p>where <span class="math">\(.*\)</span> means element-wise multiplication, and
scale is a config scalar, its default value is one.</p>
<p>The example usage is:</p>
<div class="highlight-python"><div class="highlight"><pre><span></span><span class="n">op</span> <span class="o">=</span> <span class="n">dotmul_operator</span><span class="p">(</span><span class="n">x</span><span class="o">=</span><span class="n">layer1</span><span class="p">,</span> <span class="n">y</span><span class="o">=</span><span class="n">layer2</span><span class="p">,</span> <span class="n">scale</span><span class="o">=</span><span class="mf">0.5</span><span class="p">)</span>
<div class="highlight-python"><div class="highlight"><pre><span></span><span class="n">op</span> <span class="o">=</span> <span class="n">dotmul_operator</span><span class="p">(</span><span class="n">a</span><span class="o">=</span><span class="n">layer1</span><span class="p">,</span> <span class="n">b</span><span class="o">=</span><span class="n">layer2</span><span class="p">,</span> <span class="n">scale</span><span class="o">=</span><span class="mf">0.5</span><span class="p">)</span>
</pre></div>
</div>
<table class="docutils field-list" frame="void" rules="none">
......@@ -3077,9 +3077,9 @@ Input should be a vector of positive numbers, without normalization.</p>
<dd><blockquote>
<div><p>mean squared error cost:</p>
<div class="math">
\[$\]</div>
\[\]</div>
</div></blockquote>
<p>rac{1}{N}sum_{i=1}^N(t _i- y_i)^2$</p>
<p>rac{1}{N}sum_{i=1}^N(t_i-y_i)^2</p>
<blockquote>
<div><table class="docutils field-list" frame="void" rules="none">
<col class="field-name" />
......
......@@ -301,7 +301,7 @@ layer that not support this attribute, paddle will print an error and core.</p>
<li><strong>drop_rate</strong> (<em>float</em>) &#8211; Dropout rate. Dropout will create a mask on layer output.
The dropout rate is the zero rate of this mask. The
details of what dropout is please refer to <a class="reference external" href="https://www.cs.toronto.edu/~hinton/absps/JMLRdropout.pdf">here</a>.</li>
<li><strong>device</strong> (<em>int</em>) &#8211; <p>device ID of layer. device=-1, use CPU. device&gt;0, use GPU.
<li><strong>device</strong> (<em>int</em>) &#8211; <p>device ID of layer. device=-1, use CPU. device&gt;=0, use GPU.
The details allocation in parallel_nn please refer to <a class="reference external" href="http://www.paddlepaddle.org/doc/ui/cmd_argument/use_case.html#case-2-specify-layers-in-different-devices">here</a>.</p>
</li>
</ul>
......
因为 它太大了无法显示 source diff 。你可以改为 查看blob
因为 它太大了无法显示 source diff 。你可以改为 查看blob
......@@ -11,8 +11,7 @@ Data layer
data
----
.. automodule:: paddle.v2.layer
:members: data
.. autoclass:: paddle.v2.layer.data
:noindex:
Fully Connected Layers
......@@ -22,14 +21,12 @@ Fully Connected Layers
fc
--
.. automodule:: paddle.v2.layer
:members: fc
.. autoclass:: paddle.v2.layer.fc
:noindex:
selective_fc
------------
.. automodule:: paddle.v2.layer
:members: selective_fc
.. autoclass:: paddle.v2.layer.selective_fc
:noindex:
Conv Layers
......@@ -37,34 +34,29 @@ Conv Layers
conv_operator
-------------
.. automodule:: paddle.v2.layer
:members: conv_operator
.. autoclass:: paddle.v2.layer.conv_operator
:noindex:
conv_projection
---------------
.. automodule:: paddle.v2.layer
:members: conv_projection
.. autoclass:: paddle.v2.layer.conv_projection
:noindex:
conv_shift
----------
.. automodule:: paddle.v2.layer
:members: conv_shift
.. autoclass:: paddle.v2.layer.conv_shift
:noindex:
img_conv
--------
.. automodule:: paddle.v2.layer
:members: img_conv
.. autoclass:: paddle.v2.layer.img_conv
:noindex:
.. _api_v2.layer_context_projection:
context_projection
------------------
.. automodule:: paddle.v2.layer
:members: context_projection
.. autoclass:: paddle.v2.layer.context_projection
:noindex:
Image Pooling Layer
......@@ -72,20 +64,17 @@ Image Pooling Layer
img_pool
--------
.. automodule:: paddle.v2.layer
:members: img_pool
.. autoclass:: paddle.v2.layer.img_pool
:noindex:
spp
---
.. automodule:: paddle.v2.layer
:members: spp
.. autoclass:: paddle.v2.layer.spp
:noindex:
maxout
------
.. automodule:: paddle.v2.layer
:members: maxout
.. autoclass:: paddle.v2.layer.maxout
:noindex:
Norm Layer
......@@ -93,26 +82,22 @@ Norm Layer
img_cmrnorm
-----------
.. automodule:: paddle.v2.layer
:members: img_cmrnorm
.. autoclass:: paddle.v2.layer.img_cmrnorm
:noindex:
batch_norm
----------
.. automodule:: paddle.v2.layer
:members: batch_norm
.. autoclass:: paddle.v2.layer.batch_norm
:noindex:
sum_to_one_norm
---------------
.. automodule:: paddle.v2.layer
:members: sum_to_one_norm
.. autoclass:: paddle.v2.layer.sum_to_one_norm
:noindex:
cross_channel_norm
------------------
.. automodule:: paddle.v2.layer
:members: cross_channel_norm
.. autoclass:: paddle.v2.layer.cross_channel_norm
:noindex:
Recurrent Layers
......@@ -120,20 +105,17 @@ Recurrent Layers
recurrent
---------
.. automodule:: paddle.v2.layer
:members: recurrent
.. autoclass:: paddle.v2.layer.recurrent
:noindex:
lstmemory
---------
.. automodule:: paddle.v2.layer
:members: lstmemory
.. autoclass:: paddle.v2.layer.lstmemory
:noindex:
grumemory
---------
.. automodule:: paddle.v2.layer
:members: grumemory
.. autoclass:: paddle.v2.layer.grumemory
:noindex:
Recurrent Layer Group
......@@ -141,38 +123,32 @@ Recurrent Layer Group
memory
------
.. automodule:: paddle.v2.layer
:members: memory
.. autoclass:: paddle.v2.layer.memory
:noindex:
recurrent_group
---------------
.. automodule:: paddle.v2.layer
:members: recurrent_group
.. autoclass:: paddle.v2.layer.recurrent_group
:noindex:
lstm_step
---------
.. automodule:: paddle.v2.layer
:members: lstm_step
.. autoclass:: paddle.v2.layer.lstm_step
:noindex:
gru_step
--------
.. automodule:: paddle.v2.layer
:members: gru_step
.. autoclass:: paddle.v2.layer.gru_step
:noindex:
beam_search
------------
.. automodule:: paddle.v2.layer
:members: beam_search
.. autoclass:: paddle.v2.layer.beam_search
:noindex:
get_output
----------
.. automodule:: paddle.v2.layer
:members: get_output
.. autoclass:: paddle.v2.layer.get_output
:noindex:
Mixed Layer
......@@ -182,59 +158,50 @@ Mixed Layer
mixed
-----
.. automodule:: paddle.v2.layer
:members: mixed
.. autoclass:: paddle.v2.layer.mixed
:noindex:
.. _api_v2.layer_embedding:
embedding
---------
.. automodule:: paddle.v2.layer
:members: embedding
.. autoclass:: paddle.v2.layer.embedding
:noindex:
scaling_projection
------------------
.. automodule:: paddle.v2.layer
:members: scaling_projection
.. autoclass:: paddle.v2.layer.scaling_projection
:noindex:
dotmul_projection
-----------------
.. automodule:: paddle.v2.layer
:members: dotmul_projection
.. autoclass:: paddle.v2.layer.dotmul_projection
:noindex:
dotmul_operator
---------------
.. automodule:: paddle.v2.layer
:members: dotmul_operator
.. autoclass:: paddle.v2.layer.dotmul_operator
:noindex:
full_matrix_projection
----------------------
.. automodule:: paddle.v2.layer
:members: full_matrix_projection
.. autoclass:: paddle.v2.layer.full_matrix_projection
:noindex:
identity_projection
-------------------
.. automodule:: paddle.v2.layer
:members: identity_projection
.. autoclass:: paddle.v2.layer.identity_projection
:noindex:
table_projection
----------------
.. automodule:: paddle.v2.layer
:members: table_projection
.. autoclass:: paddle.v2.layer.table_projection
:noindex:
trans_full_matrix_projection
----------------------------
.. automodule:: paddle.v2.layer
:members: trans_full_matrix_projection
.. autoclass:: paddle.v2.layer.trans_full_matrix_projection
:noindex:
Aggregate Layers
......@@ -244,36 +211,31 @@ Aggregate Layers
pooling
-------
.. automodule:: paddle.v2.layer
:members: pooling
.. autoclass:: paddle.v2.layer.pooling
:noindex:
.. _api_v2.layer_last_seq:
last_seq
--------
.. automodule:: paddle.v2.layer
:members: last_seq
.. autoclass:: paddle.v2.layer.last_seq
:noindex:
.. _api_v2.layer_first_seq:
first_seq
---------
.. automodule:: paddle.v2.layer
:members: first_seq
.. autoclass:: paddle.v2.layer.first_seq
:noindex:
concat
------
.. automodule:: paddle.v2.layer
:members: concat
.. autoclass:: paddle.v2.layer.concat
:noindex:
seq_concat
----------
.. automodule:: paddle.v2.layer
:members: seq_concat
.. autoclass:: paddle.v2.layer.seq_concat
:noindex:
Reshaping Layers
......@@ -281,34 +243,29 @@ Reshaping Layers
block_expand
------------
.. automodule:: paddle.v2.layer
:members: block_expand
.. autoclass:: paddle.v2.layer.block_expand
:noindex:
.. _api_v2.layer_expand:
expand
------
.. automodule:: paddle.v2.layer
:members: expand
.. autoclass:: paddle.v2.layer.expand
:noindex:
repeat
------
.. automodule:: paddle.v2.layer
:members: repeat
.. autoclass:: paddle.v2.layer.repeat
:noindex:
rotate
------
.. automodule:: paddle.v2.layer
:members: rotate
.. autoclass:: paddle.v2.layer.rotate
:noindex:
seq_reshape
-----------
.. automodule:: paddle.v2.layer
:members: seq_reshape
.. autoclass:: paddle.v2.layer.seq_reshape
:noindex:
Math Layers
......@@ -316,64 +273,54 @@ Math Layers
addto
-----
.. automodule:: paddle.v2.layer
:members: addto
.. autoclass:: paddle.v2.layer.addto
:noindex:
linear_comb
-----------
.. automodule:: paddle.v2.layer
:members: linear_comb
.. autoclass:: paddle.v2.layer.linear_comb
:noindex:
interpolation
-------------
.. automodule:: paddle.v2.layer
:members: interpolation
.. autoclass:: paddle.v2.layer.interpolation
:noindex:
bilinear_interp
---------------
.. automodule:: paddle.v2.layer
:members: bilinear_interp
.. autoclass:: paddle.v2.layer.bilinear_interp
:noindex:
power
-----
.. automodule:: paddle.v2.layer
:members: power
.. autoclass:: paddle.v2.layer.power
:noindex:
scaling
-------
.. automodule:: paddle.v2.layer
:members: scaling
.. autoclass:: paddle.v2.layer.scaling
:noindex:
slope_intercept
---------------
.. automodule:: paddle.v2.layer
:members: slope_intercept
.. autoclass:: paddle.v2.layer.slope_intercept
:noindex:
tensor
------
.. automodule:: paddle.v2.layer
:members: tensor
.. autoclass:: paddle.v2.layer.tensor
:noindex:
.. _api_v2.layer_cos_sim:
cos_sim
-------
.. automodule:: paddle.v2.layer
:members: cos_sim
.. autoclass:: paddle.v2.layer.cos_sim
:noindex:
trans
-----
.. automodule:: paddle.v2.layer
:members: trans
.. autoclass:: paddle.v2.layer.trans
:noindex:
Sampling Layers
......@@ -381,14 +328,12 @@ Sampling Layers
maxid
-----
.. automodule:: paddle.v2.layer
:members: maxid
.. autoclass:: paddle.v2.layer.max_id
:noindex:
sampling_id
-----------
.. automodule:: paddle.v2.layer
:members: sampling_id
.. autoclass:: paddle.v2.layer.sampling_id
:noindex:
Slicing and Joining Layers
......@@ -396,8 +341,7 @@ Slicing and Joining Layers
pad
----
.. automodule:: paddle.v2.layer
:members: pad
.. autoclass:: paddle.v2.layer.pad
:noindex:
.. _api_v2.layer_costs:
......@@ -407,80 +351,72 @@ Cost Layers
cross_entropy_cost
------------------
.. automodule:: paddle.v2.layer
:members: cross_entropy_cost
.. autoclass:: paddle.v2.layer.cross_entropy_cost
:noindex:
cross_entropy_with_selfnorm_cost
--------------------------------
.. automodule:: paddle.v2.layer
:members: cross_entropy_with_selfnorm_cost
.. autoclass:: paddle.v2.layer.cross_entropy_with_selfnorm_cost
:noindex:
multi_binary_label_cross_entropy_cost
-------------------------------------
.. automodule:: paddle.v2.layer
:members: multi_binary_label_cross_entropy_cost
.. autoclass:: paddle.v2.layer.multi_binary_label_cross_entropy_cost
:noindex:
huber_cost
----------
.. automodule:: paddle.v2.layer
:members: huber_cost
.. autoclass:: paddle.v2.layer.huber_cost
:noindex:
lambda_cost
-----------
.. automodule:: paddle.v2.layer
:members: lambda_cost
.. autoclass:: paddle.v2.layer.lambda_cost
:noindex:
mse_cost
--------
.. autoclass:: paddle.v2.layer.mse_cost
:noindex:
rank_cost
---------
.. automodule:: paddle.v2.layer
:members: rank_cost
.. autoclass:: paddle.v2.layer.rank_cost
:noindex:
sum_cost
---------
.. automodule:: paddle.v2.layer
:members: sum_cost
.. autoclass:: paddle.v2.layer.sum_cost
:noindex:
crf
---
.. automodule:: paddle.v2.layer
:members: crf
.. autoclass:: paddle.v2.layer.crf
:noindex:
crf_decoding
------------
.. automodule:: paddle.v2.layer
:members: crf_decoding
.. autoclass:: paddle.v2.layer.crf_decoding
:noindex:
ctc
---
.. automodule:: paddle.v2.layer
:members: ctc
.. autoclass:: paddle.v2.layer.ctc
:noindex:
warp_ctc
--------
.. automodule:: paddle.v2.layer
:members: warp_ctc
.. autoclass:: paddle.v2.layer.warp_ctc
:noindex:
nce
---
.. automodule:: paddle.v2.layer
:members: nce
.. autoclass:: paddle.v2.layer.nce
:noindex:
hsigmoid
---------
.. automodule:: paddle.v2.layer
:members: hsigmoid
.. autoclass:: paddle.v2.layer.hsigmoid
:noindex:
Check Layer
......@@ -488,6 +424,5 @@ Check Layer
eos
---
.. automodule:: paddle.v2.layer
:members: eos
.. autoclass:: paddle.v2.layer.eos
:noindex:
......@@ -277,7 +277,7 @@ layer that not support this attribute, paddle will print an error and core.</p>
<li><strong>drop_rate</strong> (<em>float</em>) &#8211; Dropout rate. Dropout will create a mask on layer output.
The dropout rate is the zero rate of this mask. The
details of what dropout is please refer to <a class="reference external" href="https://www.cs.toronto.edu/~hinton/absps/JMLRdropout.pdf">here</a>.</li>
<li><strong>device</strong> (<em>int</em>) &#8211; <p>device ID of layer. device=-1, use CPU. device&gt;0, use GPU.
<li><strong>device</strong> (<em>int</em>) &#8211; <p>device ID of layer. device=-1, use CPU. device&gt;=0, use GPU.
The details allocation in parallel_nn please refer to <a class="reference external" href="http://www.paddlepaddle.org/doc/ui/cmd_argument/use_case.html#case-2-specify-layers-in-different-devices">here</a>.</p>
</li>
</ul>
......
......@@ -1764,11 +1764,11 @@ It performs element-wise multiplication with weight.</p>
<code class="descclassname">paddle.trainer_config_helpers.layers.</code><code class="descname">dotmul_operator</code><span class="sig-paren">(</span><em>a=None</em>, <em>b=None</em>, <em>scale=1</em>, <em>**kwargs</em><span class="sig-paren">)</span></dt>
<dd><p>DotMulOperator takes two inputs and performs element-wise multiplication:</p>
<div class="math">
\[out.row[i] += scale * (x.row[i] .* y.row[i])\]</div>
\[out.row[i] += scale * (a.row[i] .* b.row[i])\]</div>
<p>where <span class="math">\(.*\)</span> means element-wise multiplication, and
scale is a config scalar, its default value is one.</p>
<p>The example usage is:</p>
<div class="highlight-python"><div class="highlight"><pre><span></span><span class="n">op</span> <span class="o">=</span> <span class="n">dotmul_operator</span><span class="p">(</span><span class="n">x</span><span class="o">=</span><span class="n">layer1</span><span class="p">,</span> <span class="n">y</span><span class="o">=</span><span class="n">layer2</span><span class="p">,</span> <span class="n">scale</span><span class="o">=</span><span class="mf">0.5</span><span class="p">)</span>
<div class="highlight-python"><div class="highlight"><pre><span></span><span class="n">op</span> <span class="o">=</span> <span class="n">dotmul_operator</span><span class="p">(</span><span class="n">a</span><span class="o">=</span><span class="n">layer1</span><span class="p">,</span> <span class="n">b</span><span class="o">=</span><span class="n">layer2</span><span class="p">,</span> <span class="n">scale</span><span class="o">=</span><span class="mf">0.5</span><span class="p">)</span>
</pre></div>
</div>
<table class="docutils field-list" frame="void" rules="none">
......@@ -3084,9 +3084,9 @@ Input should be a vector of positive numbers, without normalization.</p>
<dd><blockquote>
<div><p>mean squared error cost:</p>
<div class="math">
\[$\]</div>
\[\]</div>
</div></blockquote>
<p>rac{1}{N}sum_{i=1}^N(t _i- y_i)^2$</p>
<p>rac{1}{N}sum_{i=1}^N(t_i-y_i)^2</p>
<blockquote>
<div><table class="docutils field-list" frame="void" rules="none">
<col class="field-name" />
......
......@@ -308,7 +308,7 @@ layer that not support this attribute, paddle will print an error and core.</p>
<li><strong>drop_rate</strong> (<em>float</em>) &#8211; Dropout rate. Dropout will create a mask on layer output.
The dropout rate is the zero rate of this mask. The
details of what dropout is please refer to <a class="reference external" href="https://www.cs.toronto.edu/~hinton/absps/JMLRdropout.pdf">here</a>.</li>
<li><strong>device</strong> (<em>int</em>) &#8211; <p>device ID of layer. device=-1, use CPU. device&gt;0, use GPU.
<li><strong>device</strong> (<em>int</em>) &#8211; <p>device ID of layer. device=-1, use CPU. device&gt;=0, use GPU.
The details allocation in parallel_nn please refer to <a class="reference external" href="http://www.paddlepaddle.org/doc/ui/cmd_argument/use_case.html#case-2-specify-layers-in-different-devices">here</a>.</p>
</li>
</ul>
......
因为 它太大了无法显示 source diff 。你可以改为 查看blob
此差异已折叠。
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册