Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Crayon鑫
Paddle
提交
d3f219aa
P
Paddle
项目概览
Crayon鑫
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
d3f219aa
编写于
8月 21, 2017
作者:
Y
Yu Yang
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Change IdentityOp to ScaleOp
上级
c108d610
变更
11
隐藏空白更改
内联
并排
Showing
11 changed file
with
158 addition
and
14 deletion
+158
-14
paddle/framework/CMakeLists.txt
paddle/framework/CMakeLists.txt
+1
-1
paddle/framework/pybind.cc
paddle/framework/pybind.cc
+2
-1
paddle/framework/tensor.h
paddle/framework/tensor.h
+4
-1
paddle/operators/CMakeLists.txt
paddle/operators/CMakeLists.txt
+1
-1
paddle/operators/net_op.cc
paddle/operators/net_op.cc
+7
-2
paddle/operators/scale_op.cc
paddle/operators/scale_op.cc
+102
-0
paddle/operators/scale_op.cu
paddle/operators/scale_op.cu
+3
-2
paddle/operators/scale_op.h
paddle/operators/scale_op.h
+12
-4
python/paddle/v2/framework/tests/CMakeLists.txt
python/paddle/v2/framework/tests/CMakeLists.txt
+1
-1
python/paddle/v2/framework/tests/gradient_checker.py
python/paddle/v2/framework/tests/gradient_checker.py
+6
-1
python/paddle/v2/framework/tests/test_scale_and_identity_op.py
...n/paddle/v2/framework/tests/test_scale_and_identity_op.py
+19
-0
未找到文件。
paddle/framework/CMakeLists.txt
浏览文件 @
d3f219aa
...
...
@@ -56,5 +56,5 @@ cc_library(paddle_pybind SHARED
uniform_random_op
gaussian_random_op
fill_zeros_like_op
identity
_op
)
scale
_op
)
endif
(
WITH_PYTHON
)
paddle/framework/pybind.cc
浏览文件 @
d3f219aa
...
...
@@ -42,7 +42,8 @@ USE_OP(fill_zeros_like);
USE_OP_ITSELF
(
recurrent_op
);
USE_OP
(
gaussian_random
);
USE_OP
(
uniform_random
);
USE_OP
(
identity
);
USE_OP
(
scale
);
USE_OP_ITSELF
(
identity
);
namespace
paddle
{
namespace
framework
{
...
...
paddle/framework/tensor.h
浏览文件 @
d3f219aa
...
...
@@ -105,7 +105,10 @@ class Tensor {
template
<
typename
T
>
inline
Tensor
Slice
(
const
int
&
begin_idx
,
const
int
&
end_idx
)
const
;
platform
::
Place
place
()
const
{
return
holder_
->
place
();
}
platform
::
Place
place
()
const
{
PADDLE_ENFORCE_NOT_NULL
(
holder_
,
"Tensor get place() must contains holder"
);
return
holder_
->
place
();
}
private:
template
<
typename
T
>
...
...
paddle/operators/CMakeLists.txt
浏览文件 @
d3f219aa
...
...
@@ -68,4 +68,4 @@ op_library(recurrent_op SRCS recurrent_op.cc rnn/recurrent_op_utils.cc
DEPS framework_proto tensor op_registry operator net_op
)
op_library
(
uniform_random_op
SRCS uniform_random_op.cc uniform_random_op.cu
)
op_library
(
identity_op SRCS identity_op.cc identity
_op.cu DEPS net_op
)
op_library
(
scale_op SRCS scale_op.cc scale
_op.cu DEPS net_op
)
paddle/operators/net_op.cc
浏览文件 @
d3f219aa
...
...
@@ -68,10 +68,15 @@ std::string NetOp::DebugString() const {
bool
NetOp
::
IsNetOp
()
const
{
return
true
;
}
std
::
vector
<
std
::
string
>
NetOp
::
OutputVars
(
bool
has_intermediate
)
const
{
std
::
vector
<
std
::
string
>
all
;
for
(
auto
&
pair
:
this
->
outputs_
)
{
for
(
auto
&
var_name
:
pair
.
second
)
{
all
.
push_back
(
var_name
);
}
}
if
(
has_intermediate
)
{
return
this
->
outputs_
.
at
(
kAll
)
;
return
all
;
}
auto
&
all
=
this
->
outputs_
.
at
(
kAll
);
std
::
vector
<
std
::
string
>
ret_val
;
for
(
auto
&
each
:
all
)
{
if
(
!
Contains
(
intermediate_outputs_
,
each
))
{
...
...
paddle/operators/
identity
_op.cc
→
paddle/operators/
scale
_op.cc
浏览文件 @
d3f219aa
...
...
@@ -12,16 +12,16 @@
See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/operators/
identity
_op.h"
#include "paddle/operators/
scale
_op.h"
#include "paddle/operators/net_op.h"
namespace
paddle
{
namespace
operators
{
class
Identity
Op
:
public
framework
::
OperatorWithKernel
{
class
Scale
Op
:
public
framework
::
OperatorWithKernel
{
public:
Identity
Op
(
const
std
::
string
&
type
,
const
VarNameMap
&
inputs
,
const
VarNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
)
Scale
Op
(
const
std
::
string
&
type
,
const
VarNameMap
&
inputs
,
const
VarNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
)
:
OperatorWithKernel
(
type
,
inputs
,
outputs
,
attrs
)
{}
protected:
...
...
@@ -32,40 +32,71 @@ class IdentityOp : public framework::OperatorWithKernel {
}
};
class
IdentityOpMaker
:
public
framework
::
OpProtoAndCheckerMaker
{
template
<
typename
AttrType
>
class
ScaleOpMaker
:
public
framework
::
OpProtoAndCheckerMaker
{
public:
IdentityOpMaker
(
framework
::
OpProto
*
proto
,
framework
::
OpAttrChecker
*
op_checker
)
ScaleOpMaker
(
framework
::
OpProto
*
proto
,
framework
::
OpAttrChecker
*
op_checker
)
:
OpProtoAndCheckerMaker
(
proto
,
op_checker
)
{
AddInput
(
"X"
,
"The input tensor of
identity
operator."
).
NotInGradient
();
AddOutput
(
"Out"
,
"The output tensor of
identity
operator."
).
NotInGradient
();
AddComment
(
R"DOC(
Identity
operator
AddInput
(
"X"
,
"The input tensor of
scale
operator."
).
NotInGradient
();
AddOutput
(
"Out"
,
"The output tensor of
scale
operator."
).
NotInGradient
();
AddComment
(
R"DOC(
Scale
operator
The equation is: Out = X
The equation is: Out =
scale*
X
)DOC"
);
AddAttr
<
AttrType
>
(
"scale"
,
"scale of scale operator."
).
SetDefault
(
1.0
);
}
};
// Identity Op's gradient is identity op, too.
// Grad(Out=identity_op(X)) => Grad(X) = identity_op(Grad(Out))
class
IdentityGradOp
:
public
NetOp
{
// Grad(Out=scale(X)) => Grad(X) = scale(Grad(Out))
template
<
typename
AttrType
>
class
ScaleGradOp
:
public
NetOp
{
public:
IdentityGradOp
(
const
std
::
string
&
type
,
const
VarNameMap
&
inputs
,
const
VarNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
)
ScaleGradOp
(
const
std
::
string
&
type
,
const
VarNameMap
&
inputs
,
const
VarNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
)
:
NetOp
(
type
,
inputs
,
outputs
,
attrs
)
{
AddOp
(
framework
::
OpRegistry
::
CreateOp
(
"identity"
,
{{
"X"
,
{
Input
(
framework
::
GradVarName
(
"Out"
))}}},
{{
"Out"
,
{
Output
(
framework
::
GradVarName
(
"X"
))}}},
{}));
"scale"
,
{{
"X"
,
{
Input
(
framework
::
GradVarName
(
"Out"
))}}},
{{
"Out"
,
{
Output
(
framework
::
GradVarName
(
"X"
))}}},
{{
"scale"
,
GetAttr
<
AttrType
>
(
"scale"
)}}));
CompleteAddOp
(
false
);
}
};
// identity is a alias of scale op. This is also a example for creating a alias
// operator.
template
<
typename
AttrType
>
class
IdentityOpMaker
:
public
framework
::
OpProtoAndCheckerMaker
{
public:
IdentityOpMaker
(
framework
::
OpProto
*
proto
,
framework
::
OpAttrChecker
*
op_checker
)
:
OpProtoAndCheckerMaker
(
proto
,
op_checker
)
{
AddInput
(
"X"
,
"input tensor of identity op"
);
AddOutput
(
"Out"
,
"output tensor of identity op"
);
AddComment
(
"identity operator. Just a alias of scale op which scale = 1.0"
);
}
};
template
<
typename
AttrType
>
class
IdentityOp
:
public
NetOp
{
public:
IdentityOp
(
const
std
::
string
&
type
,
const
VarNameMap
&
inputs
,
const
VarNameMap
&
outputs
,
const
framework
::
AttributeMap
&
attrs
)
:
NetOp
(
type
,
inputs
,
outputs
,
attrs
)
{
AddOp
(
framework
::
OpRegistry
::
CreateOp
(
"scale"
,
{{
"X"
,
{
Input
(
"X"
)}}},
{{
"Out"
,
{
Output
(
"Out"
)}}},
{{
"scale"
,
static_cast
<
AttrType
>
(
1
)}}));
}
};
}
// namespace operators
}
// namespace paddle
namespace
ops
=
paddle
::
operators
;
REGISTER_OP
(
identity
,
ops
::
IdentityOp
,
ops
::
IdentityOpMaker
,
identity_grad
,
ops
::
IdentityGradOp
);
REGISTER_OP_CPU_KERNEL
(
identity
,
ops
::
IdentityKernel
<
float
>
);
REGISTER_OP
(
scale
,
ops
::
ScaleOp
,
ops
::
ScaleOpMaker
<
float
>
,
scale_grad
,
ops
::
ScaleGradOp
<
float
>
);
REGISTER_OP_CPU_KERNEL
(
scale
,
ops
::
ScaleKernel
<
paddle
::
platform
::
CPUPlace
,
float
>
);
REGISTER_OP_WITHOUT_GRADIENT
(
identity
,
ops
::
IdentityOp
<
float
>
,
ops
::
IdentityOpMaker
<
float
>
);
paddle/operators/
identity
_op.cu
→
paddle/operators/
scale
_op.cu
浏览文件 @
d3f219aa
...
...
@@ -12,6 +12,7 @@
See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/operators/
identity
_op.h"
#include "paddle/operators/
scale
_op.h"
REGISTER_OP_GPU_KERNEL
(
identity
,
paddle
::
operators
::
IdentityKernel
<
float
>
);
REGISTER_OP_GPU_KERNEL
(
scale
,
paddle
::
operators
::
ScaleKernel
<
paddle
::
platform
::
GPUPlace
,
float
>
);
paddle/operators/
identity
_op.h
→
paddle/operators/
scale
_op.h
浏览文件 @
d3f219aa
...
...
@@ -14,17 +14,25 @@
#pragma once
#include "paddle/framework/eigen.h"
#include "paddle/framework/op_registry.h"
#include "paddle/memory/memcpy.h"
namespace
paddle
{
namespace
operators
{
template
<
typename
T
>
class
Identity
Kernel
:
public
framework
::
OpKernel
{
template
<
typename
Place
,
typename
T
,
typename
AttrType
=
T
>
class
Scale
Kernel
:
public
framework
::
OpKernel
{
public:
virtual
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
{
auto
*
tensor
=
context
.
Output
<
framework
::
Tensor
>
(
"Out"
);
auto
*
in
=
context
.
Input
<
framework
::
Tensor
>
(
"X"
);
tensor
->
CopyFrom
<
T
>
(
*
in
,
in
->
place
());
tensor
->
mutable_data
<
T
>
(
in
->
place
());
auto
scale
=
static_cast
<
T
>
(
context
.
op_
.
GetAttr
<
AttrType
>
(
"scale"
));
auto
eigen_out
=
framework
::
EigenVector
<
T
>::
Flatten
(
*
tensor
);
auto
eigen_in
=
framework
::
EigenVector
<
T
>::
Flatten
(
*
in
);
auto
&
dev
=
context
.
GetEigenDevice
<
Place
>
();
eigen_out
.
device
(
dev
)
=
scale
*
eigen_in
;
}
};
...
...
python/paddle/v2/framework/tests/CMakeLists.txt
浏览文件 @
d3f219aa
...
...
@@ -27,4 +27,4 @@ py_test(test_uniform_random_op SRCS test_uniform_random_op.py)
py_test
(
test_recurrent_op SRCS test_recurrent_op.py
)
py_test
(
test_sgd_op SRCS test_sgd_op.py
)
py_test
(
test_gradient_checker SRCS test_gradient_checker.py
)
py_test
(
test_
identity_op SRCS test
_identity_op.py
)
py_test
(
test_
scale_and_identity_op SRCS test_scale_and
_identity_op.py
)
python/paddle/v2/framework/tests/gradient_checker.py
浏览文件 @
d3f219aa
...
...
@@ -160,8 +160,13 @@ class GradientChecker(unittest.TestCase):
grad_tensor
.
set
(
data
,
place
)
# run backward op
for
name
in
backward_op
.
outputs
():
backward_outs
=
backward_op
.
outputs
()
backward_names
=
[
item
for
key
in
backward_outs
for
item
in
backward_outs
[
key
]
]
for
name
in
backward_names
:
scope
.
new_var
(
name
)
backward_op
.
infer_shape
(
scope
)
backward_op
.
run
(
scope
,
ctx
)
...
...
python/paddle/v2/framework/tests/test_identity_op.py
→
python/paddle/v2/framework/tests/test_
scale_and_
identity_op.py
浏览文件 @
d3f219aa
...
...
@@ -2,6 +2,7 @@ import unittest
from
op_test_util
import
OpTestMeta
from
gradient_checker
import
GradientChecker
,
create_op
import
numpy
as
np
from
paddle.v2.framework.op
import
Operator
class
IdentityTest
(
unittest
.
TestCase
):
...
...
@@ -20,5 +21,23 @@ class IdentityGradOpTest(GradientChecker):
self
.
check_grad
(
op
,
inputs
,
set
(
"X"
),
"Out"
)
class
ScaleTest
(
unittest
.
TestCase
):
__metaclass__
=
OpTestMeta
def
setUp
(
self
):
self
.
type
=
"scale"
self
.
inputs
=
{
'X'
:
np
.
random
.
random
((
32
,
784
)).
astype
(
"float32"
)}
self
.
attrs
=
{
'scale'
:
-
2.3
}
self
.
outputs
=
{
'Out'
:
self
.
inputs
[
'X'
]
*
self
.
attrs
[
'scale'
]}
class
ScaleGradTest
(
GradientChecker
):
def
test_normal
(
self
):
op
=
Operator
(
"scale"
,
X
=
"X"
,
Out
=
"Out"
,
scale
=
3.2
)
self
.
check_grad
(
op
,
{
"X"
:
np
.
random
.
random
((
10
,
10
)).
astype
(
"float32"
)},
set
(
"X"
),
"Out"
)
if
__name__
==
'__main__'
:
unittest
.
main
()
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录