Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
cbc1b7f1
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
cbc1b7f1
编写于
6月 14, 2018
作者:
Y
yuyang18
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Polish documentation
上级
674327a4
变更
8
隐藏空白更改
内联
并排
Showing
8 changed file
with
144 addition
and
64 deletion
+144
-64
paddle/fluid/operators/activation_op.cc
paddle/fluid/operators/activation_op.cc
+6
-7
paddle/fluid/operators/compare_op.cc
paddle/fluid/operators/compare_op.cc
+5
-6
paddle/fluid/operators/multiplex_op.cc
paddle/fluid/operators/multiplex_op.cc
+3
-1
paddle/fluid/operators/row_conv_op.cc
paddle/fluid/operators/row_conv_op.cc
+1
-1
python/paddle/fluid/layers/control_flow.py
python/paddle/fluid/layers/control_flow.py
+25
-6
python/paddle/fluid/layers/detection.py
python/paddle/fluid/layers/detection.py
+48
-33
python/paddle/fluid/layers/io.py
python/paddle/fluid/layers/io.py
+34
-9
python/paddle/fluid/layers/ops.py
python/paddle/fluid/layers/ops.py
+22
-1
未找到文件。
paddle/fluid/operators/activation_op.cc
浏览文件 @
cbc1b7f1
...
@@ -275,7 +275,7 @@ class HardShrinkOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -275,7 +275,7 @@ class HardShrinkOpMaker : public framework::OpProtoAndCheckerMaker {
"The value of threshold for HardShrink. [default: 0.5]"
)
"The value of threshold for HardShrink. [default: 0.5]"
)
.
SetDefault
(
0.5
f
);
.
SetDefault
(
0.5
f
);
AddComment
(
R"DOC(
AddComment
(
R"DOC(
HardShrink Activation Operator.
** HardShrink activation operator **
.. math::
.. math::
out = \begin{cases}
out = \begin{cases}
...
@@ -399,13 +399,12 @@ class ThresholdedReluOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -399,13 +399,12 @@ class ThresholdedReluOpMaker : public framework::OpProtoAndCheckerMaker {
AddComment
(
R"DOC(
AddComment
(
R"DOC(
ThresholdedRelu Activation Operator.
ThresholdedRelu Activation Operator.
$$
.. math::
out = \begin{cases}
x, \text{if } x > threshold \\
0, \text{otherwise}
\end{cases}
$$
out = \begin{cases}
x, \text{if } x > threshold \\
0, \text{otherwise}
\end{cases}
)DOC"
);
)DOC"
);
}
}
};
};
...
...
paddle/fluid/operators/compare_op.cc
浏览文件 @
cbc1b7f1
...
@@ -34,16 +34,15 @@ class CompareOpProtoMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -34,16 +34,15 @@ class CompareOpProtoMaker : public framework::OpProtoAndCheckerMaker {
.
SetDefault
(
true
);
.
SetDefault
(
true
);
AddOutput
(
"Out"
,
string
::
Sprintf
(
"n-dim bool tensor. Each element is %s"
,
AddOutput
(
"Out"
,
string
::
Sprintf
(
"n-dim bool tensor. Each element is %s"
,
comment
.
equation
));
comment
.
equation
));
AddComment
(
string
::
Sprintf
(
R"DOC(%s Operator
AddComment
(
string
::
Sprintf
(
R"DOC(
It operates element-wise on X and Y, and returns the Out. Each of them is a
It operates element-wise on X and Y, and returns the Out. Each of them is a
N-dim tensor. X and Y could be any type. The each element of the Out tensor is
N-dim tensor. X and Y could be any type. The each element of the Out tensor is
calculated by $%s$
calculated by $%s$
)DOC"
,
)DOC"
,
comment
.
type
,
comment
.
equation
));
comment
.
equation
));
AddAttr
<
int
>
(
"axis"
,
AddAttr
<
int
>
(
"(int, default -1). The start dimension index "
"axis"
,
"for broadcasting Y onto X.
"
)
"The start dimension index for broadcasting Y onto X. [default -1]
"
)
.
SetDefault
(
-
1
)
.
SetDefault
(
-
1
)
.
EqualGreaterThan
(
-
1
);
.
EqualGreaterThan
(
-
1
);
}
}
...
...
paddle/fluid/operators/multiplex_op.cc
浏览文件 @
cbc1b7f1
...
@@ -96,7 +96,9 @@ the (Ids[i])-th tensor.
...
@@ -96,7 +96,9 @@ the (Ids[i])-th tensor.
For i-th row of the output tensor:
For i-th row of the output tensor:
$ y[i] = x_{k}[i] $
$$
y[i] = x_{k}[i]
$$
where $y$ is the output tensor, $x_{k}$ is the k-th input tensor,
where $y$ is the output tensor, $x_{k}$ is the k-th input tensor,
and $k = Ids[i]$.
and $k = Ids[i]$.
...
...
paddle/fluid/operators/row_conv_op.cc
浏览文件 @
cbc1b7f1
...
@@ -94,7 +94,7 @@ class RowConvOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -94,7 +94,7 @@ class RowConvOpMaker : public framework::OpProtoAndCheckerMaker {
"in this LodTensor is a matrix with shape T x N, i.e., the "
"in this LodTensor is a matrix with shape T x N, i.e., the "
"same shape as X."
);
"same shape as X."
);
AddComment
(
R"DOC(
AddComment
(
R"DOC(
Row-convolution Operator.
** Row-convolution operator **
The row convolution is called lookahead convolution. This operator was
The row convolution is called lookahead convolution. This operator was
introduced in the following paper for DeepSpeech2:
introduced in the following paper for DeepSpeech2:
...
...
python/paddle/fluid/layers/control_flow.py
浏览文件 @
cbc1b7f1
...
@@ -1008,8 +1008,28 @@ def array_read(array, i):
...
@@ -1008,8 +1008,28 @@ def array_read(array, i):
def
shrink_memory
(
x
,
i
,
table
):
def
shrink_memory
(
x
,
i
,
table
):
"""
"""
This function creates an operator to shrink
_rnn_
memory using the RankTable
This function creates an operator to shrink
rnn
memory using the RankTable
as mentioned in the input parameter.
as mentioned in the input parameter.
NOTE: This API is very low-level API. It is used by DynamicRNN only.
Since the Dynamic RNN uses no-padding way to implement RNN. The sequence
will be sorted by order, and the length of valid memory will be shrink after
each time step.
Args:
x(Variable): The memory object in the previous time step.
i(Variable): The step count variable. A int scalar as LoDTensor.
table(Variable): The RNNRankTable object.
Returns:
the memory variable after shrink.
Examples:
Since this API is very low level API. The example is not provided.
Please reference the implementation of class DynamicRNN for detail
usage.
"""
"""
helper
=
LayerHelper
(
'shrink_memory'
,
**
locals
())
helper
=
LayerHelper
(
'shrink_memory'
,
**
locals
())
out
=
helper
.
create_tmp_variable
(
dtype
=
x
.
dtype
)
out
=
helper
.
create_tmp_variable
(
dtype
=
x
.
dtype
)
...
@@ -1316,10 +1336,9 @@ class IfElse(object):
...
@@ -1316,10 +1336,9 @@ class IfElse(object):
class
DynamicRNN
(
object
):
class
DynamicRNN
(
object
):
"""
"""
Dynamic RNN.
The dynamic RNN can process a batch of sequence data. The length of each
sample sequence can be different. This API automatically process them in
This RNN can process a batch of sequence data. The length of each sample
batch.
sequence can be different. This API automatically process them in batch.
The input lod must be set. Please reference `lod_tensor`
The input lod must be set. Please reference `lod_tensor`
...
@@ -1500,7 +1519,7 @@ class DynamicRNN(object):
...
@@ -1500,7 +1519,7 @@ class DynamicRNN(object):
need_reorder
=
False
,
need_reorder
=
False
,
dtype
=
'float32'
):
dtype
=
'float32'
):
"""
"""
Create a memory variable.
Create a memory variable
for dynamic rnn
.
If the :code:`init` is not None, :code:`memory` will be initialized by
If the :code:`init` is not None, :code:`memory` will be initialized by
this variable. The :code:`need_reorder` is used to reorder the memory as
this variable. The :code:`need_reorder` is used to reorder the memory as
...
...
python/paddle/fluid/layers/detection.py
浏览文件 @
cbc1b7f1
...
@@ -210,53 +210,68 @@ def bipartite_match(dist_matrix,
...
@@ -210,53 +210,68 @@ def bipartite_match(dist_matrix,
dist_threshold
=
None
,
dist_threshold
=
None
,
name
=
None
):
name
=
None
):
"""
"""
**Bipartite matchint operator**
This operator implements a greedy bipartite matching algorithm, which is
used to obtain the matching with the maximum distance based on the input
This operator is a greedy bipartite matching algorithm, which is used to
obtain the matching with the maximum distance based on the input
distance matrix. For input 2D matrix, the bipartite matching algorithm can
distance matrix. For input 2D matrix, the bipartite matching algorithm can
find the matched column for each row, also can find the matched row for
find the matched column for each row (matched means the largest distance),
each column. And this operator only calculate matched indices from column
also can find the matched row for each column. And this operator only
to row. For each instance, the number of matched indices is the number of
calculate matched indices from column to row. For each instance,
of columns of the input ditance matrix.
the number of matched indices is the column number of the input distance
matrix.
There are two outputs to save matched indices and distance.
A simple description, this algothrim matched the best (maximum distance)
There are two outputs, matched indices and distance.
A simple description, this algorithm matched the best (maximum distance)
row entity to the column entity and the matched indices are not duplicated
row entity to the column entity and the matched indices are not duplicated
in each row of ColToRowMatchIndices. If the column entity is not matched
in each row of ColToRowMatchIndices. If the column entity is not matched
any row entity, set -1 in ColToRowMatchIndices.
any row entity, set -1 in ColToRowMatchIndices.
Please note that
the input DistMat can be LoDTensor (with LoD) or Tensor.
NOTE:
the input DistMat can be LoDTensor (with LoD) or Tensor.
If LoDTensor with LoD, the height of ColToRowMatchIndices is batch size.
If LoDTensor with LoD, the height of ColToRowMatchIndices is batch size.
If Tensor, the height of ColToRowMatchIndices is 1.
If Tensor, the height of ColToRowMatchIndices is 1.
NOTE: This API is a very low level API. It is used by :code:`ssd_loss`
layer. Please consider to use :code:`ssd_loss` instead.
Args:
Args:
dist_matrix(Variable): This input is a 2-D LoDTensor with shape
dist_matrix(Variable): This input is a 2-D LoDTensor with shape
[K, M]. It is pair-wise distance matrix between the entities
[K, M]. It is pair-wise distance matrix between the entities
represented by each row and each column. For example, assumed one
represented by each row and each column. For example, assumed one
entity is A with shape [K], another entity is B with shape [M]. The
entity is A with shape [K], another entity is B with shape [M]. The
dist_matirx[i][j] is the distance between A[i] and B[j]. The bigger
dist_matrix[i][j] is the distance between A[i] and B[j]. The bigger
the distance is, the better macthing the pairs are. Please note,
the distance is, the better matching the pairs are.
This tensor can contain LoD information to represent a batch of
inputs. One instance of this batch can contain different numbers of
NOTE: This tensor can contain LoD information to represent a batch
entities.
of inputs. One instance of this batch can contain different numbers
of entities.
match_type(string|None): The type of matching method, should be
match_type(string|None): The type of matching method, should be
'bipartite' or 'per_prediction'
, 'bipartite' by defalut
.
'bipartite' or 'per_prediction'
. [default 'bipartite']
.
dist_threshold(float|None): If `match_type` is 'per_prediction',
dist_threshold(float|None): If `match_type` is 'per_prediction',
this threshold is to determine the extra matching bboxes based
this threshold is to determine the extra matching bboxes based
on the maximum distance, 0.5 by defa
lu
t.
on the maximum distance, 0.5 by defa
ul
t.
Returns:
Returns:
match_indices(Variable): A 2-D Tensor with shape [N, M] in int type.
tuple: a tuple with two elements is returned. The first is
N is the batch size. If match_indices[i][j] is -1, it
matched_indices, the second is matched_distance.
means B[j] does not match any entity in i-th instance.
Otherwise, it means B[j] is matched to row
The matched_indices is a 2-D Tensor with shape [N, M] in int type.
match_indices[i][j] in i-th instance. The row number of
N is the batch size. If match_indices[i][j] is -1, it
i-th instance is saved in match_indices[i][j].
means B[j] does not match any entity in i-th instance.
match_distance(Variable): A 2-D Tensor with shape [N, M] in float type.
Otherwise, it means B[j] is matched to row
N is batch size. If match_indices[i][j] is -1,
match_indices[i][j] in i-th instance. The row number of
match_distance[i][j] is also -1.0. Otherwise, assumed
i-th instance is saved in match_indices[i][j].
match_distance[i][j] = d, and the row offsets of each instance
are called LoD. Then match_distance[i][j] = dist_matrix[d+LoD[i]][j].
The matched_distance is a 2-D Tensor with shape [N, M] in float type
. N is batch size. If match_indices[i][j] is -1,
match_distance[i][j] is also -1.0. Otherwise, assumed
match_distance[i][j] = d, and the row offsets of each instance
are called LoD. Then match_distance[i][j] =
dist_matrix[d+LoD[i]][j].
Examples:
>>> x = fluid.layers.data(name='x', shape=[4], dtype='float32')
>>> y = fluid.layers.data(name='y', shape=[4], dtype='float32')
>>> iou = fluid.layers.iou_similarity(x=x, y=y)
>>> matched_indices, matched_dist = fluid.layers.bipartite_match(iou)
"""
"""
helper
=
LayerHelper
(
'bipartite_match'
,
**
locals
())
helper
=
LayerHelper
(
'bipartite_match'
,
**
locals
())
match_indices
=
helper
.
create_tmp_variable
(
dtype
=
'int32'
)
match_indices
=
helper
.
create_tmp_variable
(
dtype
=
'int32'
)
...
@@ -364,7 +379,7 @@ def ssd_loss(location,
...
@@ -364,7 +379,7 @@ def ssd_loss(location,
normalize
=
True
,
normalize
=
True
,
sample_size
=
None
):
sample_size
=
None
):
"""
"""
**Multi-box loss layer for object dection algorithm of SSD**
**Multi-box loss layer for object de
te
ction algorithm of SSD**
This layer is to compute dection loss for SSD given the location offset
This layer is to compute dection loss for SSD given the location offset
predictions, confidence predictions, prior boxes and ground-truth boudding
predictions, confidence predictions, prior boxes and ground-truth boudding
...
@@ -372,7 +387,7 @@ def ssd_loss(location,
...
@@ -372,7 +387,7 @@ def ssd_loss(location,
is a weighted sum of the localization loss (or regression loss) and
is a weighted sum of the localization loss (or regression loss) and
confidence loss (or classification loss) by performing the following steps:
confidence loss (or classification loss) by performing the following steps:
1. Find matched bound
d
ing box by bipartite matching algorithm.
1. Find matched bounding box by bipartite matching algorithm.
1.1 Compute IOU similarity between ground-truth boxes and prior boxes.
1.1 Compute IOU similarity between ground-truth boxes and prior boxes.
...
@@ -435,7 +450,7 @@ def ssd_loss(location,
...
@@ -435,7 +450,7 @@ def ssd_loss(location,
mining_type (str): The hard example mining type, should be 'hard_example'
mining_type (str): The hard example mining type, should be 'hard_example'
or 'max_negative', now only support `max_negative`.
or 'max_negative', now only support `max_negative`.
normalize (bool): Whether to normalize the SSD loss by the total number
normalize (bool): Whether to normalize the SSD loss by the total number
of output locations, True by defa
lu
t.
of output locations, True by defa
ul
t.
sample_size (int): The max sample size of negative box, used only when
sample_size (int): The max sample size of negative box, used only when
mining_type is 'hard_example'.
mining_type is 'hard_example'.
...
...
python/paddle/fluid/layers/io.py
浏览文件 @
cbc1b7f1
...
@@ -302,15 +302,6 @@ def open_recordio_file(filename,
...
@@ -302,15 +302,6 @@ def open_recordio_file(filename,
"""
"""
${comment}
${comment}
>>> import paddle.fluid as fluid
>>> reader = fluid.layers.io.open_recordio_file(
>>> filename='./data.recordio',
>>> shapes=[(3,224,224), (1)],
>>> lod_levels=[0, 0],
>>> dtypes=['float32', 'int64'])
>>> # Via the reader, we can use 'read_file' layer to get data:
>>> image, label = fluid.layers.io.read_file(reader)
Args:
Args:
filename(${filename_type}): ${filename_comment}.
filename(${filename_type}): ${filename_comment}.
shapes(list): List of tuples which declaring data shapes.
shapes(list): List of tuples which declaring data shapes.
...
@@ -322,6 +313,17 @@ def open_recordio_file(filename,
...
@@ -322,6 +313,17 @@ def open_recordio_file(filename,
Returns:
Returns:
${out_comment}.
${out_comment}.
Examples:
>>> import paddle.fluid as fluid
>>> reader = fluid.layers.io.open_recordio_file(
>>> filename='./data.recordio',
>>> shapes=[(3,224,224), (1)],
>>> lod_levels=[0, 0],
>>> dtypes=['float32', 'int64'])
>>> # Via the reader, we can use 'read_file' layer to get data:
>>> image, label = fluid.layers.io.read_file(reader)
"""
"""
dtypes
=
[
convert_np_dtype_to_dtype_
(
dt
)
for
dt
in
dtypes
]
dtypes
=
[
convert_np_dtype_to_dtype_
(
dt
)
for
dt
in
dtypes
]
shape_concat
=
[]
shape_concat
=
[]
...
@@ -549,6 +551,29 @@ def batch(reader, batch_size):
...
@@ -549,6 +551,29 @@ def batch(reader, batch_size):
def
double_buffer
(
reader
,
place
=
None
,
name
=
None
):
def
double_buffer
(
reader
,
place
=
None
,
name
=
None
):
"""
Wrap a double buffer reader. The data will copy to target place with a
double buffer queue. If the target place is None, the place that executor
perform on will be used.
Args:
reader(Variable): the reader variable need to be wrapped.
place(Place): the place of target data. Default is the sample place of
executor perform.
name(str): Variable name. None if the user does not care.
Returns:
wrapped reader with double buffer.
Examples:
>>> reader = fluid.layers.open_files(filenames=['somefile'],
>>> shapes=[[-1, 784], [-1, 1]],
>>> dtypes=['float32', 'int64'])
>>> reader = fluid.layers.double_buffer(reader)
>>> img, label = fluid.layers.read_file(reader)
"""
attrs
=
dict
()
attrs
=
dict
()
if
place
is
not
None
:
if
place
is
not
None
:
attrs
[
'place'
]
=
str
(
place
).
upper
()
attrs
[
'place'
]
=
str
(
place
).
upper
()
...
...
python/paddle/fluid/layers/ops.py
浏览文件 @
cbc1b7f1
...
@@ -66,7 +66,6 @@ __all__ = [
...
@@ -66,7 +66,6 @@ __all__ = [
'uniform_random_batch_size_like'
,
'uniform_random_batch_size_like'
,
'gaussian_random'
,
'gaussian_random'
,
'gaussian_random_batch_size_like'
,
'gaussian_random_batch_size_like'
,
'cumsum'
,
'scatter'
,
'scatter'
,
'sum'
,
'sum'
,
'slice'
,
'slice'
,
...
@@ -120,3 +119,25 @@ Examples:
...
@@ -120,3 +119,25 @@ Examples:
>>> data = fluid.layers.data(name="input", shape=[784])
>>> data = fluid.layers.data(name="input", shape=[784])
>>> result = fluid.layers.hard_shrink(x=data, threshold=0.3)
>>> result = fluid.layers.hard_shrink(x=data, threshold=0.3)
"""
"""
__all__
+=
[
'cumsum'
]
_cum_sum_
=
generate_layer_fn
(
'cumsum'
)
def
cumsum
(
x
,
axis
=
None
,
exclusive
=
None
,
reverse
=
None
):
kwargs
=
dict
()
for
name
in
locals
():
val
=
locals
()[
name
]
if
val
is
not
None
:
kwargs
[
name
]
=
val
return
_cum_sum_
(
**
kwargs
)
cumsum
.
__doc__
=
_cum_sum_
.
__doc__
+
"""
Examples:
>>> data = fluid.layers.data(name="input", shape=[32, 784])
>>> result = fluid.layers.cumsum(data, axis=0)
"""
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录