Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
b4f28ccc
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
b4f28ccc
编写于
6月 25, 2018
作者:
F
fengjiayi
提交者:
GitHub
6月 25, 2018
浏览文件
操作
浏览文件
下载
差异文件
Merge pull request #11632 from JiayiFeng/some_small_fixes
Some small fixes
上级
f0cf70ec
e1a46bba
变更
8
隐藏空白更改
内联
并排
Showing
8 changed file
with
53 addition
and
44 deletion
+53
-44
paddle/fluid/operators/assign_value_op.cc
paddle/fluid/operators/assign_value_op.cc
+2
-1
paddle/fluid/operators/random_crop_op.cc
paddle/fluid/operators/random_crop_op.cc
+5
-3
paddle/fluid/operators/random_crop_op.h
paddle/fluid/operators/random_crop_op.h
+15
-9
paddle/fluid/operators/reader/create_custom_reader_op.cc
paddle/fluid/operators/reader/create_custom_reader_op.cc
+6
-4
paddle/fluid/operators/reader/create_double_buffer_reader_op.cc
.../fluid/operators/reader/create_double_buffer_reader_op.cc
+2
-2
python/paddle/fluid/layers/io.py
python/paddle/fluid/layers/io.py
+7
-4
python/paddle/fluid/layers/nn.py
python/paddle/fluid/layers/nn.py
+10
-17
python/paddle/fluid/layers/tensor.py
python/paddle/fluid/layers/tensor.py
+6
-4
未找到文件。
paddle/fluid/operators/assign_value_op.cc
浏览文件 @
b4f28ccc
...
@@ -70,6 +70,7 @@ $$Out = values$$
...
@@ -70,6 +70,7 @@ $$Out = values$$
namespace
ops
=
paddle
::
operators
;
namespace
ops
=
paddle
::
operators
;
REGISTER_OPERATOR
(
assign_value
,
ops
::
AssignValueOp
,
ops
::
AssignValueOpMaker
);
REGISTER_OPERATOR
(
assign_value
,
ops
::
AssignValueOp
,
ops
::
AssignValueOpMaker
,
paddle
::
framework
::
EmptyGradOpMaker
);
REGISTER_OP_CPU_KERNEL
(
assign_value
,
ops
::
AssignValueKernel
<
int
>
,
REGISTER_OP_CPU_KERNEL
(
assign_value
,
ops
::
AssignValueKernel
<
int
>
,
ops
::
AssignValueKernel
<
float
>
);
ops
::
AssignValueKernel
<
float
>
);
paddle/fluid/operators/random_crop_op.cc
浏览文件 @
b4f28ccc
...
@@ -37,6 +37,11 @@ class RandomCropOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -37,6 +37,11 @@ class RandomCropOpMaker : public framework::OpProtoAndCheckerMaker {
AddOutput
(
"SeedOut"
,
"The random seed after random cropping."
)
AddOutput
(
"SeedOut"
,
"The random seed after random cropping."
)
.
AsIntermediate
();
.
AsIntermediate
();
AddAttr
<
std
::
vector
<
int
>>
(
"shape"
,
"The shape of a cropped instance."
);
AddAttr
<
std
::
vector
<
int
>>
(
"shape"
,
"The shape of a cropped instance."
);
AddAttr
<
int
>
(
"startup_seed"
,
"If the input 'Seed' is not initialized, the 'startup_seed' "
"will be used to replace it. Even so, the seed after random "
"crop will also be outputed to the 'SeedOut'."
)
.
SetDefault
(
0
);
AddComment
(
R"DOC(
AddComment
(
R"DOC(
This operator takes a batch of instance, and do random cropping on each instance.
This operator takes a batch of instance, and do random cropping on each instance.
It means that cropping positions differs on each instance, which is determined
It means that cropping positions differs on each instance, which is determined
...
@@ -49,8 +54,6 @@ class RandomCropOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -49,8 +54,6 @@ class RandomCropOpMaker : public framework::OpProtoAndCheckerMaker {
class
RandomCropOpInferShape
:
public
framework
::
InferShapeBase
{
class
RandomCropOpInferShape
:
public
framework
::
InferShapeBase
{
public:
public:
void
operator
()(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
operator
()(
framework
::
InferShapeContext
*
ctx
)
const
override
{
auto
seed_dim
=
ctx
->
GetInputDim
(
"Seed"
);
PADDLE_ENFORCE
(
seed_dim
.
size
()
==
1
&&
seed_dim
[
0
]
==
1
);
auto
shape
=
ctx
->
Attrs
().
Get
<
std
::
vector
<
int
>>
(
"shape"
);
auto
shape
=
ctx
->
Attrs
().
Get
<
std
::
vector
<
int
>>
(
"shape"
);
auto
x_dim
=
ctx
->
GetInputDim
(
"X"
);
auto
x_dim
=
ctx
->
GetInputDim
(
"X"
);
PADDLE_ENFORCE_GT
(
x_dim
.
size
(),
static_cast
<
int64_t
>
(
shape
.
size
()));
PADDLE_ENFORCE_GT
(
x_dim
.
size
(),
static_cast
<
int64_t
>
(
shape
.
size
()));
...
@@ -62,7 +65,6 @@ class RandomCropOpInferShape : public framework::InferShapeBase {
...
@@ -62,7 +65,6 @@ class RandomCropOpInferShape : public framework::InferShapeBase {
out_dim
[
x_i
]
=
shape
[
shape_i
];
out_dim
[
x_i
]
=
shape
[
shape_i
];
}
}
ctx
->
SetOutputDim
(
"Out"
,
framework
::
make_ddim
(
out_dim
));
ctx
->
SetOutputDim
(
"Out"
,
framework
::
make_ddim
(
out_dim
));
ctx
->
SetOutputDim
(
"SeedOut"
,
framework
::
make_ddim
({
1
}));
}
}
};
};
...
...
paddle/fluid/operators/random_crop_op.h
浏览文件 @
b4f28ccc
...
@@ -142,16 +142,22 @@ template <typename DeviceContext, typename T>
...
@@ -142,16 +142,22 @@ template <typename DeviceContext, typename T>
class
RandomCropKernel
:
public
framework
::
OpKernel
<
T
>
{
class
RandomCropKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
public:
virtual
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
virtual
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
auto
&
seed_tensor
=
detail
::
Ref
(
ctx
.
Input
<
framework
::
LoDTensor
>
(
"Seed"
));
int64_t
seed
=
0
;
int64_t
seed
=
0
;
if
(
platform
::
is_cpu_place
(
seed_tensor
.
place
()))
{
auto
&
seed_tensor
=
detail
::
Ref
(
ctx
.
Input
<
framework
::
LoDTensor
>
(
"Seed"
));
seed
=
*
seed_tensor
.
data
<
int64_t
>
();
if
(
seed_tensor
.
IsInitialized
())
{
if
(
platform
::
is_cpu_place
(
seed_tensor
.
place
()))
{
seed
=
*
seed_tensor
.
data
<
int64_t
>
();
}
else
{
LOG
(
WARNING
)
<<
"It is slow to place seed in GPU memory. Please verify "
"your program"
;
framework
::
LoDTensor
cpu_seed
;
framework
::
TensorCopySync
(
seed_tensor
,
platform
::
CPUPlace
(),
&
cpu_seed
);
seed
=
*
cpu_seed
.
data
<
int64_t
>
();
}
}
else
{
}
else
{
LOG
(
WARNING
)
<<
"It is slow to place seed in GPU memory. Please verify "
VLOG
(
5
)
<<
"WARNING: The input 'Seed' is not initialized, use attribute "
"your program"
;
"'startup_seed' instead."
;
framework
::
LoDTensor
cpu_seed
;
seed
=
ctx
.
Attr
<
int
>
(
"startup_seed"
);
framework
::
TensorCopySync
(
seed_tensor
,
platform
::
CPUPlace
(),
&
cpu_seed
);
seed
=
*
cpu_seed
.
data
<
int64_t
>
();
}
}
auto
shape
=
ctx
.
Attr
<
std
::
vector
<
int
>>
(
"shape"
);
auto
shape
=
ctx
.
Attr
<
std
::
vector
<
int
>>
(
"shape"
);
auto
&
x
=
detail
::
Ref
(
ctx
.
Input
<
framework
::
LoDTensor
>
(
"X"
));
auto
&
x
=
detail
::
Ref
(
ctx
.
Input
<
framework
::
LoDTensor
>
(
"X"
));
...
@@ -171,7 +177,7 @@ class RandomCropKernel : public framework::OpKernel<T> {
...
@@ -171,7 +177,7 @@ class RandomCropKernel : public framework::OpKernel<T> {
engine
.
discard
(
functor
.
prod_batchsize_dims_
*
engine
.
discard
(
functor
.
prod_batchsize_dims_
*
(
functor
.
rank_
-
functor
.
num_batchsize_dims_
));
(
functor
.
rank_
-
functor
.
num_batchsize_dims_
));
*
ctx
.
Output
<
framework
::
LoDTensor
>
(
"SeedOut"
)
->
mutable_data
<
int64_t
>
(
*
ctx
.
Output
<
framework
::
LoDTensor
>
(
"SeedOut"
)
->
mutable_data
<
int64_t
>
(
platform
::
CPUPlace
())
=
engine
();
framework
::
make_ddim
({
1
}),
platform
::
CPUPlace
())
=
engine
();
}
}
};
};
...
...
paddle/fluid/operators/reader/create_custom_reader_op.cc
浏览文件 @
b4f28ccc
...
@@ -39,6 +39,7 @@ class CustomReader : public framework::DecoratedReader {
...
@@ -39,6 +39,7 @@ class CustomReader : public framework::DecoratedReader {
const
framework
::
ProgramDesc
program_
;
const
framework
::
ProgramDesc
program_
;
int
sub_block_id_
;
int
sub_block_id_
;
framework
::
Executor
exe_
;
framework
::
Executor
exe_
;
framework
::
Scope
scope_
;
std
::
vector
<
std
::
string
>
source_var_names_
;
std
::
vector
<
std
::
string
>
source_var_names_
;
std
::
vector
<
std
::
string
>
sink_var_names_
;
std
::
vector
<
std
::
string
>
sink_var_names_
;
...
@@ -158,23 +159,24 @@ void CustomReader::ReadNext(std::vector<framework::LoDTensor>* out) {
...
@@ -158,23 +159,24 @@ void CustomReader::ReadNext(std::vector<framework::LoDTensor>* out) {
// The scope for CustomReader's sub-block should be independent and shouldn't
// The scope for CustomReader's sub-block should be independent and shouldn't
// be any other computation scope's child. Otherwise, data preprocessing and
// be any other computation scope's child. Otherwise, data preprocessing and
// compution cannot be concurrent.
// compution cannot be concurrent.
framework
::
Scope
scope
;
framework
::
Scope
*
exe_scope
=
&
scope_
.
NewScope
()
;
// 1. Copy LoDTensors from underlying reader's output to source variables.
// 1. Copy LoDTensors from underlying reader's output to source variables.
for
(
size_t
i
=
0
;
i
<
source_var_names_
.
size
();
++
i
)
{
for
(
size_t
i
=
0
;
i
<
source_var_names_
.
size
();
++
i
)
{
framework
::
Variable
*
var
=
scope
.
Var
(
source_var_names_
[
i
]);
framework
::
Variable
*
var
=
exe_scope
->
Var
(
source_var_names_
[
i
]);
framework
::
LoDTensor
*
tensor
=
var
->
GetMutable
<
framework
::
LoDTensor
>
();
framework
::
LoDTensor
*
tensor
=
var
->
GetMutable
<
framework
::
LoDTensor
>
();
tensor
->
ShareDataWith
(
underlying_outs
[
i
]);
tensor
->
ShareDataWith
(
underlying_outs
[
i
]);
tensor
->
set_lod
(
underlying_outs
[
i
].
lod
());
tensor
->
set_lod
(
underlying_outs
[
i
].
lod
());
}
}
// 2. Run the sub-block.
// 2. Run the sub-block.
exe_
.
Run
(
program_
,
&
scope
,
sub_block_id_
,
false
,
true
);
exe_
.
Run
(
program_
,
exe_
scope
,
sub_block_id_
,
false
,
true
);
// 3. Copy LoDTensors from sink variables to out.
// 3. Copy LoDTensors from sink variables to out.
out
->
resize
(
sink_var_names_
.
size
());
out
->
resize
(
sink_var_names_
.
size
());
for
(
size_t
i
=
0
;
i
<
sink_var_names_
.
size
();
++
i
)
{
for
(
size_t
i
=
0
;
i
<
sink_var_names_
.
size
();
++
i
)
{
const
auto
&
tensor
=
detail
::
Ref
(
scope
.
FindVar
(
sink_var_names_
[
i
]))
const
auto
&
tensor
=
detail
::
Ref
(
exe_scope
->
FindVar
(
sink_var_names_
[
i
]))
.
Get
<
framework
::
LoDTensor
>
();
.
Get
<
framework
::
LoDTensor
>
();
framework
::
TensorCopySync
(
tensor
,
platform
::
CPUPlace
(),
&
(
*
out
)[
i
]);
framework
::
TensorCopySync
(
tensor
,
platform
::
CPUPlace
(),
&
(
*
out
)[
i
]);
}
}
scope_
.
DeleteScope
(
exe_scope
);
}
}
}
// namespace reader
}
// namespace reader
...
...
paddle/fluid/operators/reader/create_double_buffer_reader_op.cc
浏览文件 @
b4f28ccc
...
@@ -23,13 +23,13 @@ namespace reader {
...
@@ -23,13 +23,13 @@ namespace reader {
// 'Double buffer' means we shall maintain two batches of input data at the same
// 'Double buffer' means we shall maintain two batches of input data at the same
// time. So the kCacheSize shoul be at least 2.
// time. So the kCacheSize shoul be at least 2.
static
constexpr
size_t
kCacheSize
=
3
;
static
constexpr
size_t
kCacheSize
=
5
;
// There will be two bacthes out of the channel during training:
// There will be two bacthes out of the channel during training:
// 1. the one waiting to be sent to the channel
// 1. the one waiting to be sent to the channel
// 2. the one just be received from the channel, which is also being used by
// 2. the one just be received from the channel, which is also being used by
// subsequent operators.
// subsequent operators.
// So the channel size should be kChacheSize - 2
// So the channel size should be kChacheSize - 2
static
constexpr
size_t
kChannelSize
=
1
;
// kCacheSize - 2
static
constexpr
size_t
kChannelSize
=
3
;
// kCacheSize - 2
class
DoubleBufferReader
:
public
framework
::
DecoratedReader
{
class
DoubleBufferReader
:
public
framework
::
DecoratedReader
{
public:
public:
...
...
python/paddle/fluid/layers/io.py
浏览文件 @
b4f28ccc
...
@@ -110,7 +110,7 @@ class BlockGuardServ(BlockGuard):
...
@@ -110,7 +110,7 @@ class BlockGuardServ(BlockGuard):
class
ListenAndServ
(
object
):
class
ListenAndServ
(
object
):
"""
"""
**ListenAndServ Layer**
**ListenAndServ Layer**
ListenAndServ is used to create a rpc server bind and listen
ListenAndServ is used to create a rpc server bind and listen
on specific TCP port, this server will run the sub-block when
on specific TCP port, this server will run the sub-block when
received variables from clients.
received variables from clients.
...
@@ -212,7 +212,7 @@ def Send(endpoints, send_vars, sync=True):
...
@@ -212,7 +212,7 @@ def Send(endpoints, send_vars, sync=True):
of send_vars to send
of send_vars to send
send_vars (list): variables to send to server
send_vars (list): variables to send to server
sync (bool): whether to wait the request finish
sync (bool): whether to wait the request finish
"""
"""
assert
(
type
(
send_vars
)
==
list
)
assert
(
type
(
send_vars
)
==
list
)
...
@@ -469,10 +469,13 @@ def open_files(filenames,
...
@@ -469,10 +469,13 @@ def open_files(filenames,
lod_levels(list): List of ints which declaring data lod_level.
lod_levels(list): List of ints which declaring data lod_level.
dtypes(list): List of strs which declaring data type.
dtypes(list): List of strs which declaring data type.
thread_num(int): The maximal concurrent prefetch thread number.
thread_num(int): The maximal concurrent prefetch thread number.
buffer_size(int): The size of prefetch buffer.
buffer_size(int|None): The size of prefetch buffer. If it is setted None,
buffer size will be thread_num * 3.
Default: None
pass_num(int): Number of passes to run.
pass_num(int): Number of passes to run.
for_parallel(Bool): Set it as True if you are going to run
for_parallel(Bool): Set it as True if you are going to run
subsequent operators in parallel.
subsequent operators in parallel.
Default: True
Returns:
Returns:
Variable: A Reader Variable via which we can get file data.
Variable: A Reader Variable via which we can get file data.
...
@@ -492,7 +495,7 @@ def open_files(filenames,
...
@@ -492,7 +495,7 @@ def open_files(filenames,
image, label = fluid.layers.io.read_file(reader)
image, label = fluid.layers.io.read_file(reader)
"""
"""
if
buffer_size
is
None
:
if
buffer_size
is
None
:
buffer_size
=
thread_num
buffer_size
=
thread_num
*
3
if
isinstance
(
filenames
,
basestring
):
if
isinstance
(
filenames
,
basestring
):
filenames
=
[
filenames
]
filenames
=
[
filenames
]
dtypes
=
[
convert_np_dtype_to_dtype_
(
dt
)
for
dt
in
dtypes
]
dtypes
=
[
convert_np_dtype_to_dtype_
(
dt
)
for
dt
in
dtypes
]
...
...
python/paddle/fluid/layers/nn.py
浏览文件 @
b4f28ccc
...
@@ -23,6 +23,7 @@ from layer_function_generator import autodoc, templatedoc
...
@@ -23,6 +23,7 @@ from layer_function_generator import autodoc, templatedoc
from
tensor
import
concat
from
tensor
import
concat
import
utils
import
utils
import
random
import
random
from
..
import
unique_name
__all__
=
[
__all__
=
[
'fc'
,
'fc'
,
...
@@ -4896,34 +4897,26 @@ def random_crop(x, shape, seed=None):
...
@@ -4896,34 +4897,26 @@ def random_crop(x, shape, seed=None):
>>> cropped_img = fluid.layers.random_crop(img, shape=[3, 224, 224])
>>> cropped_img = fluid.layers.random_crop(img, shape=[3, 224, 224])
"""
"""
helper
=
LayerHelper
(
"random_crop"
,
**
locals
())
helper
=
LayerHelper
(
"random_crop"
,
**
locals
())
dtype
=
helper
.
input_dtype
()
dtype
=
x
.
dtype
out
=
helper
.
create_tmp_variable
(
dtype
)
out
=
helper
.
create_tmp_variable
(
dtype
)
if
seed
is
None
:
if
seed
is
None
:
seed
=
random
.
randint
(
-
65536
,
65535
)
seed
=
random
.
randint
(
-
65536
,
65535
)
op_attrs
=
{
"shape"
:
shape
}
if
isinstance
(
seed
,
int
):
if
isinstance
(
seed
,
int
):
seed_value
=
seed
op_attrs
[
"startup_seed"
]
=
seed
seed
=
helper
.
create_tmp_variable
(
dtype
=
"int64"
)
seed
=
helper
.
create_variable
(
helper
.
append_op
(
name
=
unique_name
.
generate
(
"random_crop_seed"
),
type
=
"fill_constant"
,
dtype
=
"int64"
,
inputs
=
{},
persistable
=
True
)
outputs
=
{
"Out"
:
seed
},
attrs
=
{
"dtype"
:
seed
.
dtype
,
"shape"
:
[
1
],
"value"
:
float
(
seed_value
),
"force_cpu"
:
True
})
elif
not
isinstance
(
seed
,
Variable
):
elif
not
isinstance
(
seed
,
Variable
):
raise
ValueError
(
"'seed' must be a Variable or an int."
)
raise
ValueError
(
"'seed' must be a Variable or an int."
)
seed_out
=
helper
.
create_tmp_variable
(
dtype
=
"int64"
)
helper
.
append_op
(
helper
.
append_op
(
type
=
"random_crop"
,
type
=
"random_crop"
,
inputs
=
{
"X"
:
x
,
inputs
=
{
"X"
:
x
,
"Seed"
:
seed
},
"Seed"
:
seed
},
outputs
=
{
"Out"
:
out
,
outputs
=
{
"Out"
:
out
,
"SeedOut"
:
seed
_out
},
"SeedOut"
:
seed
},
attrs
=
{
"shape"
:
shape
}
)
attrs
=
op_attrs
)
return
out
return
out
...
...
python/paddle/fluid/layers/tensor.py
浏览文件 @
b4f28ccc
...
@@ -155,7 +155,7 @@ def cast(x, dtype):
...
@@ -155,7 +155,7 @@ def cast(x, dtype):
Examples:
Examples:
.. code-block:: python
.. code-block:: python
data = fluid.layers.data(name='x', shape=[13], dtype='float32')
data = fluid.layers.data(name='x', shape=[13], dtype='float32')
result = fluid.layers.cast(x=data, dtype='float64')
result = fluid.layers.cast(x=data, dtype='float64')
"""
"""
...
@@ -188,7 +188,7 @@ def concat(input, axis=0, name=None):
...
@@ -188,7 +188,7 @@ def concat(input, axis=0, name=None):
Examples:
Examples:
.. code-block:: python
.. code-block:: python
out = fluid.layers.concat(input=[Efirst, Esecond, Ethird, Efourth])
out = fluid.layers.concat(input=[Efirst, Esecond, Ethird, Efourth])
"""
"""
helper
=
LayerHelper
(
'concat'
,
**
locals
())
helper
=
LayerHelper
(
'concat'
,
**
locals
())
...
@@ -238,7 +238,7 @@ def sums(input, out=None):
...
@@ -238,7 +238,7 @@ def sums(input, out=None):
return
out
return
out
def
assign
(
input
,
output
):
def
assign
(
input
,
output
=
None
):
"""
"""
**Assign**
**Assign**
...
@@ -246,7 +246,7 @@ def assign(input, output):
...
@@ -246,7 +246,7 @@ def assign(input, output):
Args:
Args:
input(Variable|numpy.ndarray): The source variable
input(Variable|numpy.ndarray): The source variable
output(Variable): The destination variable
output(Variable
|None
): The destination variable
Returns:
Returns:
Variable: The destination variable that was supplied as the *output*.
Variable: The destination variable that was supplied as the *output*.
...
@@ -259,6 +259,8 @@ def assign(input, output):
...
@@ -259,6 +259,8 @@ def assign(input, output):
fluid.layers.assign(hidden, out)
fluid.layers.assign(hidden, out)
"""
"""
helper
=
LayerHelper
(
'assign'
,
**
locals
())
helper
=
LayerHelper
(
'assign'
,
**
locals
())
if
output
is
None
:
output
=
helper
.
create_tmp_variable
(
dtype
=
input
.
dtype
)
if
isinstance
(
input
,
Variable
):
if
isinstance
(
input
,
Variable
):
helper
.
append_op
(
helper
.
append_op
(
type
=
'assign'
,
inputs
=
{
'X'
:
[
input
]},
outputs
=
{
'Out'
:
[
output
]})
type
=
'assign'
,
inputs
=
{
'X'
:
[
input
]},
outputs
=
{
'Out'
:
[
output
]})
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录