Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
94b94e5b
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
94b94e5b
编写于
10月 03, 2017
作者:
Z
zchen0211
浏览文件
操作
浏览文件
下载
差异文件
Merge branch 'develop' of
https://github.com/PaddlePaddle/Paddle
into develop
上级
2d876b86
42e7fe05
变更
10
隐藏空白更改
内联
并排
Showing
10 changed file
with
43 addition
and
68 deletion
+43
-68
paddle/framework/grad_op_builder_test.cc
paddle/framework/grad_op_builder_test.cc
+0
-22
paddle/operators/CMakeLists.txt
paddle/operators/CMakeLists.txt
+5
-1
paddle/operators/sgd_op.cc
paddle/operators/sgd_op.cc
+3
-1
paddle/operators/sgd_op.h
paddle/operators/sgd_op.h
+1
-1
paddle/operators/sum_op.cc
paddle/operators/sum_op.cc
+15
-14
paddle/operators/sum_op.cu
paddle/operators/sum_op.cu
+0
-2
paddle/operators/sum_op.h
paddle/operators/sum_op.h
+0
-19
paddle/pybind/pybind.cc
paddle/pybind/pybind.cc
+7
-0
python/paddle/v2/framework/tests/op_test.py
python/paddle/v2/framework/tests/op_test.py
+11
-6
python/paddle/v2/framework/tests/test_sgd_op.py
python/paddle/v2/framework/tests/test_sgd_op.py
+1
-2
未找到文件。
paddle/framework/grad_op_builder_test.cc
浏览文件 @
94b94e5b
...
@@ -39,28 +39,6 @@ class IOIgnoredOpMaker : public OpProtoAndCheckerMaker {
...
@@ -39,28 +39,6 @@ class IOIgnoredOpMaker : public OpProtoAndCheckerMaker {
namespace
f
=
paddle
::
framework
;
namespace
f
=
paddle
::
framework
;
TEST
(
GradOpBuilder
,
AddTwo
)
{
std
::
shared_ptr
<
f
::
OperatorBase
>
add_op
(
f
::
OpRegistry
::
CreateOp
(
"sum"
,
{{
"X"
,
{
"x"
,
"y"
}}},
{{
"Out"
,
{
"out"
}}},
{}));
std
::
shared_ptr
<
f
::
OperatorBase
>
grad_add_op
=
f
::
OpRegistry
::
CreateGradOp
(
*
add_op
);
EXPECT_EQ
(
grad_add_op
->
Inputs
().
size
(),
1UL
);
EXPECT_EQ
(
grad_add_op
->
Outputs
().
size
(),
1UL
);
EXPECT_EQ
(
grad_add_op
->
Input
(
f
::
GradVarName
(
"Out"
)),
f
::
GradVarName
(
"out"
));
auto
&
outputs
=
grad_add_op
->
Outputs
(
f
::
GradVarName
(
"X"
));
EXPECT_EQ
(
2UL
,
outputs
.
size
());
auto
in_output
=
[
&
outputs
](
const
std
::
string
&
name
)
{
for
(
auto
&
output_name
:
outputs
)
{
if
(
output_name
==
name
)
return
true
;
}
return
false
;
};
EXPECT_TRUE
(
in_output
(
f
::
GradVarName
(
"x"
)));
EXPECT_TRUE
(
in_output
(
f
::
GradVarName
(
"y"
)));
}
REGISTER_OP
(
mult_io
,
f
::
NOP
,
f
::
MutiInOutOpMaker
,
mult_io_grad
,
f
::
NOP
);
REGISTER_OP
(
mult_io
,
f
::
NOP
,
f
::
MutiInOutOpMaker
,
mult_io_grad
,
f
::
NOP
);
REGISTER_OP
(
io_ignored
,
f
::
NOP
,
f
::
IOIgnoredOpMaker
,
io_ignored_grad
,
f
::
NOP
);
REGISTER_OP
(
io_ignored
,
f
::
NOP
,
f
::
IOIgnoredOpMaker
,
io_ignored_grad
,
f
::
NOP
);
...
...
paddle/operators/CMakeLists.txt
浏览文件 @
94b94e5b
...
@@ -103,12 +103,16 @@ set(DEPS_OPS
...
@@ -103,12 +103,16 @@ set(DEPS_OPS
recurrent_op
recurrent_op
cond_op
cond_op
cross_entropy_op
cross_entropy_op
softmax_with_cross_entropy_op
)
softmax_with_cross_entropy_op
sum_op
)
op_library
(
recurrent_op SRCS recurrent_op.cc rnn/recurrent_op_utils.cc
op_library
(
recurrent_op SRCS recurrent_op.cc rnn/recurrent_op_utils.cc
DEPS framework_proto tensor net_op
)
DEPS framework_proto tensor net_op
)
op_library
(
cond_op SRCS cond_op.cc DEPS framework_proto tensor operator net_op
)
op_library
(
cond_op SRCS cond_op.cc DEPS framework_proto tensor operator net_op
)
op_library
(
cross_entropy_op DEPS cross_entropy
)
op_library
(
cross_entropy_op DEPS cross_entropy
)
op_library
(
softmax_with_cross_entropy_op DEPS cross_entropy softmax
)
op_library
(
softmax_with_cross_entropy_op DEPS cross_entropy softmax
)
op_library
(
sum_op DEPS net_op
)
list
(
REMOVE_ITEM GENERAL_OPS
${
DEPS_OPS
}
)
list
(
REMOVE_ITEM GENERAL_OPS
${
DEPS_OPS
}
)
foreach
(
src
${
GENERAL_OPS
}
)
foreach
(
src
${
GENERAL_OPS
}
)
...
...
paddle/operators/sgd_op.cc
浏览文件 @
94b94e5b
...
@@ -27,6 +27,8 @@ class SGDOp : public framework::OperatorWithKernel {
...
@@ -27,6 +27,8 @@ class SGDOp : public framework::OperatorWithKernel {
"Input(param) of SGDOp should not be null."
);
"Input(param) of SGDOp should not be null."
);
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"grad"
),
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"grad"
),
"Input(grad) of SGDOp should not be null."
);
"Input(grad) of SGDOp should not be null."
);
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"learning_rate"
),
"Input(learning_rate) of SGDOp should not be null."
);
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"param_out"
),
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"param_out"
),
"Output(param_out) of SGDOp should not be null."
);
"Output(param_out) of SGDOp should not be null."
);
...
@@ -42,9 +44,9 @@ class SGDOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -42,9 +44,9 @@ class SGDOpMaker : public framework::OpProtoAndCheckerMaker {
SGDOpMaker
(
framework
::
OpProto
*
proto
,
framework
::
OpAttrChecker
*
op_checker
)
SGDOpMaker
(
framework
::
OpProto
*
proto
,
framework
::
OpAttrChecker
*
op_checker
)
:
OpProtoAndCheckerMaker
(
proto
,
op_checker
)
{
:
OpProtoAndCheckerMaker
(
proto
,
op_checker
)
{
AddInput
(
"param"
,
"input parameter"
);
AddInput
(
"param"
,
"input parameter"
);
AddInput
(
"learning_rate"
,
"learning rate of sgd"
);
AddInput
(
"grad"
,
"input gradient"
);
AddInput
(
"grad"
,
"input gradient"
);
AddOutput
(
"param_out"
,
"output parameter"
);
AddOutput
(
"param_out"
,
"output parameter"
);
AddAttr
<
float
>
(
"learning_rate"
,
"learning rate of sgd"
);
AddComment
(
R"DOC(
AddComment
(
R"DOC(
Simplest sgd algorithm.
Simplest sgd algorithm.
...
...
paddle/operators/sgd_op.h
浏览文件 @
94b94e5b
...
@@ -31,7 +31,7 @@ class SGDOpKernel : public framework::OpKernel<T> {
...
@@ -31,7 +31,7 @@ class SGDOpKernel : public framework::OpKernel<T> {
auto
param
=
ctx
.
Input
<
Tensor
>
(
"param"
);
auto
param
=
ctx
.
Input
<
Tensor
>
(
"param"
);
auto
grad
=
ctx
.
Input
<
Tensor
>
(
"grad"
);
auto
grad
=
ctx
.
Input
<
Tensor
>
(
"grad"
);
auto
param_out
=
ctx
.
Output
<
Tensor
>
(
"param_out"
);
auto
param_out
=
ctx
.
Output
<
Tensor
>
(
"param_out"
);
float
lr
=
ctx
.
Attr
<
float
>
(
"learning_rate"
);
float
lr
=
*
ctx
.
Input
<
float
>
(
"learning_rate"
);
param_out
->
mutable_data
<
T
>
(
ctx
.
GetPlace
());
param_out
->
mutable_data
<
T
>
(
ctx
.
GetPlace
());
...
...
paddle/operators/sum_op.cc
浏览文件 @
94b94e5b
...
@@ -11,6 +11,7 @@ limitations under the License. */
...
@@ -11,6 +11,7 @@ limitations under the License. */
#include "paddle/operators/sum_op.h"
#include "paddle/operators/sum_op.h"
#include <vector>
#include <vector>
#include "paddle/operators/net_op.h"
namespace
paddle
{
namespace
paddle
{
namespace
operators
{
namespace
operators
{
...
@@ -57,21 +58,23 @@ or not. But the output only shares the LoD with the first input.
...
@@ -57,21 +58,23 @@ or not. But the output only shares the LoD with the first input.
}
}
};
};
class
SumGradOp
:
public
framework
::
OperatorWithKernel
{
class
SumGradOp
:
public
NetOp
{
public:
public:
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
SumGradOp
(
const
std
::
string
&
type
,
const
framework
::
VariableNameMap
&
inputs
,
const
framework
::
VariableNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
)
:
NetOp
(
type
,
inputs
,
outputs
,
attrs
)
{
auto
&
x_grad_names
=
Outputs
(
framework
::
GradVarName
(
"X"
));
auto
out_grad_name
=
this
->
Input
(
framework
::
GradVarName
(
"Out"
));
protected:
framework
::
AttributeMap
grad_attrs
;
void
InferShape
(
framework
::
InferShapeContextBase
*
ctx
)
const
override
{
grad_attrs
[
"scale"
]
=
1.0
f
;
auto
out_grad_dims
=
ctx
->
GetInputDim
(
framework
::
GradVarName
(
"Out"
));
for
(
auto
&
x_grad_name
:
x_grad_names
)
{
auto
x_grad_names
=
ctx
->
Outputs
(
framework
::
GradVarName
(
"X"
));
AppendOp
(
framework
::
OpRegistry
::
CreateOp
(
size_t
x_length
=
x_grad_names
.
size
();
"scale"
,
{{
"X"
,
{
out_grad_name
}}},
{{
"Out"
,
{
x_grad_name
}}},
std
::
vector
<
framework
::
DDim
>
x_grad_dims
;
grad_attrs
));
x_grad_dims
.
reserve
(
x_length
);
for
(
size_t
i
=
0
;
i
<
x_length
;
++
i
)
{
x_grad_dims
.
push_back
(
out_grad_dims
);
}
}
ctx
->
SetOutputsDim
(
framework
::
GradVarName
(
"X"
),
x_grad_dims
);
CompleteAddOp
(
false
);
}
}
};
};
...
@@ -81,5 +84,3 @@ class SumGradOp : public framework::OperatorWithKernel {
...
@@ -81,5 +84,3 @@ class SumGradOp : public framework::OperatorWithKernel {
namespace
ops
=
paddle
::
operators
;
namespace
ops
=
paddle
::
operators
;
REGISTER_OP
(
sum
,
ops
::
SumOp
,
ops
::
SumOpMaker
,
sum_grad
,
ops
::
SumGradOp
);
REGISTER_OP
(
sum
,
ops
::
SumOp
,
ops
::
SumOpMaker
,
sum_grad
,
ops
::
SumGradOp
);
REGISTER_OP_CPU_KERNEL
(
sum
,
ops
::
SumKernel
<
paddle
::
platform
::
CPUPlace
,
float
>
);
REGISTER_OP_CPU_KERNEL
(
sum
,
ops
::
SumKernel
<
paddle
::
platform
::
CPUPlace
,
float
>
);
REGISTER_OP_CPU_KERNEL
(
sum_grad
,
ops
::
SumGradKernel
<
paddle
::
platform
::
CPUPlace
,
float
>
);
paddle/operators/sum_op.cu
浏览文件 @
94b94e5b
...
@@ -14,5 +14,3 @@ limitations under the License. */
...
@@ -14,5 +14,3 @@ limitations under the License. */
namespace
ops
=
paddle
::
operators
;
namespace
ops
=
paddle
::
operators
;
REGISTER_OP_GPU_KERNEL
(
sum
,
ops
::
SumKernel
<
paddle
::
platform
::
GPUPlace
,
float
>
);
REGISTER_OP_GPU_KERNEL
(
sum
,
ops
::
SumKernel
<
paddle
::
platform
::
GPUPlace
,
float
>
);
REGISTER_OP_GPU_KERNEL
(
sum_grad
,
ops
::
SumGradKernel
<
paddle
::
platform
::
GPUPlace
,
float
>
);
paddle/operators/sum_op.h
浏览文件 @
94b94e5b
...
@@ -42,24 +42,5 @@ class SumKernel : public framework::OpKernel<T> {
...
@@ -42,24 +42,5 @@ class SumKernel : public framework::OpKernel<T> {
}
}
};
};
template
<
typename
Place
,
typename
T
>
class
SumGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
input
=
context
.
Input
<
Tensor
>
(
framework
::
GradVarName
(
"Out"
));
auto
outs
=
context
.
MultiOutput
<
Tensor
>
(
framework
::
GradVarName
(
"X"
));
for
(
auto
out
:
outs
)
{
out
->
mutable_data
<
T
>
(
context
.
GetPlace
());
}
auto
place
=
context
.
GetEigenDevice
<
Place
>
();
auto
in
=
EigenVector
<
T
>::
Flatten
(
*
input
);
for
(
auto
out
:
outs
)
{
auto
result
=
EigenVector
<
T
>::
Flatten
(
*
out
);
result
.
device
(
place
)
=
in
;
}
}
};
}
// namespace operators
}
// namespace operators
}
// namespace paddle
}
// namespace paddle
paddle/pybind/pybind.cc
浏览文件 @
94b94e5b
...
@@ -143,6 +143,13 @@ All parameter, weight, gradient are variables in Paddle.
...
@@ -143,6 +143,13 @@ All parameter, weight, gradient are variables in Paddle.
.
def
(
"set_int"
,
.
def
(
"set_int"
,
[](
Variable
&
var
,
int
val
)
->
void
{
*
var
.
GetMutable
<
int
>
()
=
val
;
})
[](
Variable
&
var
,
int
val
)
->
void
{
*
var
.
GetMutable
<
int
>
()
=
val
;
})
.
def
(
"get_int"
,
[](
const
Variable
&
var
)
->
int
{
return
var
.
Get
<
int
>
();
})
.
def
(
"get_int"
,
[](
const
Variable
&
var
)
->
int
{
return
var
.
Get
<
int
>
();
})
.
def
(
"is_float"
,
[](
const
Variable
&
var
)
{
return
var
.
IsType
<
float
>
();
})
.
def
(
"set_float"
,
[](
Variable
&
var
,
float
val
)
->
void
{
*
var
.
GetMutable
<
float
>
()
=
val
;
})
.
def
(
"get_float"
,
[](
const
Variable
&
var
)
->
float
{
return
var
.
Get
<
float
>
();
})
.
def
(
"get_tensor"
,
.
def
(
"get_tensor"
,
[](
Variable
&
self
)
->
LoDTensor
*
{
[](
Variable
&
self
)
->
LoDTensor
*
{
return
self
.
GetMutable
<
LoDTensor
>
();
return
self
.
GetMutable
<
LoDTensor
>
();
...
...
python/paddle/v2/framework/tests/op_test.py
浏览文件 @
94b94e5b
...
@@ -46,12 +46,17 @@ def create_op(scope, op_type, inputs, outputs, attrs):
...
@@ -46,12 +46,17 @@ def create_op(scope, op_type, inputs, outputs, attrs):
def
set_input
(
scope
,
op
,
inputs
,
place
):
def
set_input
(
scope
,
op
,
inputs
,
place
):
def
__set_input__
(
var_name
,
var
):
def
__set_input__
(
var_name
,
var
):
tensor
=
scope
.
find_var
(
var_name
).
get_tensor
()
if
isinstance
(
var
,
tuple
)
or
isinstance
(
var
,
np
.
ndarray
):
if
isinstance
(
var
,
tuple
):
tensor
=
scope
.
find_var
(
var_name
).
get_tensor
()
tensor
.
set_lod
(
var
[
1
])
if
isinstance
(
var
,
tuple
):
var
=
var
[
0
]
tensor
.
set_lod
(
var
[
1
])
tensor
.
set_dims
(
var
.
shape
)
var
=
var
[
0
]
tensor
.
set
(
var
,
place
)
tensor
.
set_dims
(
var
.
shape
)
tensor
.
set
(
var
,
place
)
elif
isinstance
(
var
,
float
):
scope
.
find_var
(
var_name
).
set_float
(
var
)
elif
isinstance
(
var
,
int
):
scope
.
find_var
(
var_name
).
set_int
(
var
)
for
in_name
,
in_dup
in
Operator
.
get_op_inputs
(
op
.
type
()):
for
in_name
,
in_dup
in
Operator
.
get_op_inputs
(
op
.
type
()):
if
in_name
in
inputs
:
if
in_name
in
inputs
:
...
...
python/paddle/v2/framework/tests/test_sgd_op.py
浏览文件 @
94b94e5b
...
@@ -10,8 +10,7 @@ class TestSGDOp(OpTest):
...
@@ -10,8 +10,7 @@ class TestSGDOp(OpTest):
g
=
np
.
random
.
random
((
102
,
105
)).
astype
(
"float32"
)
g
=
np
.
random
.
random
((
102
,
105
)).
astype
(
"float32"
)
lr
=
0.1
lr
=
0.1
self
.
inputs
=
{
'param'
:
w
,
'grad'
:
g
}
self
.
inputs
=
{
'param'
:
w
,
'grad'
:
g
,
'learning_rate'
:
lr
}
self
.
attrs
=
{
'learning_rate'
:
lr
}
self
.
outputs
=
{
'param_out'
:
w
-
lr
*
g
}
self
.
outputs
=
{
'param_out'
:
w
-
lr
*
g
}
def
test_check_output
(
self
):
def
test_check_output
(
self
):
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录