Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
938717ba
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
体验新版 GitCode,发现更多精彩内容 >>
提交
938717ba
编写于
12月 26, 2017
作者:
Y
Yang Yu
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Stash
上级
2a36e8ad
变更
2
隐藏空白更改
内联
并排
Showing
2 changed file
with
79 addition
and
8 deletion
+79
-8
paddle/framework/executor.cc
paddle/framework/executor.cc
+8
-0
python/paddle/v2/fluid/tests/test_dynrnn_gradient_check.py
python/paddle/v2/fluid/tests/test_dynrnn_gradient_check.py
+71
-8
未找到文件。
paddle/framework/executor.cc
浏览文件 @
938717ba
...
@@ -66,6 +66,14 @@ void Executor::Run(const ProgramDesc& pdesc, Scope* scope, int block_id,
...
@@ -66,6 +66,14 @@ void Executor::Run(const ProgramDesc& pdesc, Scope* scope, int block_id,
PADDLE_ENFORCE_LT
(
static_cast
<
size_t
>
(
block_id
),
pdesc
.
Size
());
PADDLE_ENFORCE_LT
(
static_cast
<
size_t
>
(
block_id
),
pdesc
.
Size
());
auto
&
block
=
pdesc
.
Block
(
block_id
);
auto
&
block
=
pdesc
.
Block
(
block_id
);
if
(
VLOG_IS_ON
(
100
))
{
std
::
ostringstream
sout
;
for
(
auto
&
name
:
scope
->
GetAllNames
(
false
))
{
sout
<<
name
<<
", "
;
}
VLOG
(
100
)
<<
"Scope has variable "
<<
sout
.
str
();
}
Scope
*
local_scope
=
scope
;
Scope
*
local_scope
=
scope
;
if
(
create_vars
)
{
if
(
create_vars
)
{
if
(
create_local_scope
)
{
if
(
create_local_scope
)
{
...
...
python/paddle/v2/fluid/tests/test_dynrnn_gradient_check.py
浏览文件 @
938717ba
...
@@ -159,6 +159,39 @@ class BaseRNN(object):
...
@@ -159,6 +159,39 @@ class BaseRNN(object):
g
[
i
][
j
]
=
(
pos
-
neg
)
/
(
delta
*
2
)
g
[
i
][
j
]
=
(
pos
-
neg
)
/
(
delta
*
2
)
return
g
return
g
def
get_numeric_gradient_of_input
(
self
,
input_name
,
delta
=
0.001
,
return_one_tensor
=
True
):
ipt
=
self
.
inputs
[
input_name
]
grad
=
[]
for
seq
in
ipt
:
seq_grad
=
[]
for
item
in
seq
:
item_grad
=
numpy
.
zeros
(
shape
=
item
.
shape
,
dtype
=
item
.
dtype
)
if
len
(
item
.
shape
)
!=
1
:
raise
ValueError
(
"Not support"
)
for
i
in
xrange
(
len
(
item
)):
o
=
item
[
i
]
item
[
i
]
+=
delta
pos
=
self
.
_exe_mean_out_
()
item
[
i
]
-=
2
*
delta
neg
=
self
.
_exe_mean_out_
()
item
[
i
]
=
o
item_grad
[
i
]
=
(
pos
-
neg
)
/
(
delta
*
2
)
seq_grad
.
append
(
item_grad
)
grad
.
append
(
seq_grad
)
if
not
return_one_tensor
:
return
grad
for
i
in
xrange
(
len
(
grad
)):
grad
[
i
]
=
numpy
.
concatenate
(
grad
[
i
])
grad
=
numpy
.
concatenate
(
grad
)
return
grad
def
_exe_mean_out_
(
self
):
def
_exe_mean_out_
(
self
):
outs
=
self
.
exe
()
outs
=
self
.
exe
()
return
numpy
.
array
([
o
.
mean
()
for
o
in
outs
.
itervalues
()]).
mean
()
return
numpy
.
array
([
o
.
mean
()
for
o
in
outs
.
itervalues
()]).
mean
()
...
@@ -191,9 +224,10 @@ class TestSimpleMul(unittest.TestCase):
...
@@ -191,9 +224,10 @@ class TestSimpleMul(unittest.TestCase):
# @many_times(10)
# @many_times(10)
@
prog_scope
()
@
prog_scope
()
def
test_forward_backward
(
self
):
def
test_forward_backward
(
self
):
py
thon_impl
=
TestSimpleMul
.
SimpleMul
()
py
_rnn
=
TestSimpleMul
.
SimpleMul
()
dat
=
fluid
.
layers
.
data
(
dat
=
fluid
.
layers
.
data
(
name
=
self
.
DATA_NAME
,
shape
=
[
self
.
DATA_WIDTH
],
lod_level
=
1
)
name
=
self
.
DATA_NAME
,
shape
=
[
self
.
DATA_WIDTH
],
lod_level
=
1
)
dat
.
stop_gradient
=
False
rnn
=
fluid
.
layers
.
DynamicRNN
()
rnn
=
fluid
.
layers
.
DynamicRNN
()
with
rnn
.
block
():
with
rnn
.
block
():
...
@@ -212,17 +246,26 @@ class TestSimpleMul(unittest.TestCase):
...
@@ -212,17 +246,26 @@ class TestSimpleMul(unittest.TestCase):
cpu
=
fluid
.
CPUPlace
()
cpu
=
fluid
.
CPUPlace
()
exe
=
fluid
.
Executor
(
cpu
)
exe
=
fluid
.
Executor
(
cpu
)
out
,
w_g
=
exe
.
run
(
feed
=
python_impl
.
to_feed
(
cpu
),
out
,
w_g
,
i_g
=
map
(
numpy
.
array
,
fetch_list
=
[
out
,
self
.
PARAM_NAME
+
"@GRAD"
])
exe
.
run
(
feed
=
py_rnn
.
to_feed
(
cpu
),
out_by_python
=
python_impl
.
exe
()[
self
.
OUT_NAME
]
fetch_list
=
[
out
,
self
.
PARAM_NAME
+
"@GRAD"
,
self
.
DATA_NAME
+
"@GRAD"
],
return_numpy
=
False
))
out_by_python
=
py_rnn
.
exe
()[
self
.
OUT_NAME
]
self
.
assertTrue
(
numpy
.
allclose
(
out
,
out_by_python
))
self
.
assertTrue
(
numpy
.
allclose
(
out
,
out_by_python
))
w_g_num
=
py
thon_impl
.
get_numeric_gradient_of_param
(
self
.
PARAM_NAME
)
w_g_num
=
py
_rnn
.
get_numeric_gradient_of_param
(
self
.
PARAM_NAME
)
self
.
assertTrue
(
numpy
.
allclose
(
w_g_num
,
w_g
,
rtol
=
0.05
))
self
.
assertTrue
(
numpy
.
allclose
(
w_g_num
,
w_g
,
rtol
=
0.05
))
i_g_num
=
py_rnn
.
get_numeric_gradient_of_input
(
input_name
=
self
.
DATA_NAME
)
i_g_num
=
i_g_num
.
reshape
(
i_g
.
shape
)
self
.
assertTrue
(
numpy
.
allclose
(
i_g_num
,
i_g
,
rtol
=
0.05
))
class
TestSimpleMulWithMemory
(
unittest
.
TestCase
):
class
TestSimpleMulWithMemory
(
unittest
.
TestCase
):
DATA_WIDTH
=
32
DATA_WIDTH
=
32
HIDDEN_WIDTH
=
1
0
HIDDEN_WIDTH
=
2
0
DATA_NAME
=
'X'
DATA_NAME
=
'X'
PARAM_NAME
=
'W'
PARAM_NAME
=
'W'
...
@@ -251,12 +294,14 @@ class TestSimpleMulWithMemory(unittest.TestCase):
...
@@ -251,12 +294,14 @@ class TestSimpleMulWithMemory(unittest.TestCase):
assert
isinstance
(
Out
,
Output
)
assert
isinstance
(
Out
,
Output
)
Out
.
out
(
o
)
Out
.
out
(
o
)
# @many_times(10)
@
prog_scope
()
@
prog_scope
()
def
test_forward_backward
(
self
):
def
test_forward_backward
(
self
):
py_rnn
=
TestSimpleMulWithMemory
.
SimpleMulWithMemory
()
py_rnn
=
TestSimpleMulWithMemory
.
SimpleMulWithMemory
()
data
=
fluid
.
layers
.
data
(
data
=
fluid
.
layers
.
data
(
name
=
self
.
DATA_NAME
,
shape
=
[
self
.
DATA_WIDTH
],
lod_level
=
1
)
name
=
self
.
DATA_NAME
,
shape
=
[
self
.
DATA_WIDTH
],
lod_level
=
1
)
data
.
stop_gradient
=
False
rnn
=
fluid
.
layers
.
DynamicRNN
()
rnn
=
fluid
.
layers
.
DynamicRNN
()
with
rnn
.
block
():
with
rnn
.
block
():
d
=
rnn
.
step_input
(
data
)
d
=
rnn
.
step_input
(
data
)
...
@@ -272,14 +317,32 @@ class TestSimpleMulWithMemory(unittest.TestCase):
...
@@ -272,14 +317,32 @@ class TestSimpleMulWithMemory(unittest.TestCase):
out
=
rnn
()
out
=
rnn
()
last
=
fluid
.
layers
.
sequence_pool
(
input
=
out
,
pool_type
=
'last'
)
last
=
fluid
.
layers
.
sequence_pool
(
input
=
out
,
pool_type
=
'last'
)
loss
=
fluid
.
layers
.
mean
(
x
=
last
)
fluid
.
backward
.
append_backward_ops
(
loss
)
cpu
=
fluid
.
CPUPlace
()
cpu
=
fluid
.
CPUPlace
()
exe
=
fluid
.
Executor
(
cpu
)
exe
=
fluid
.
Executor
(
cpu
)
feed
=
py_rnn
.
to_feed
(
cpu
)
last_np
,
=
exe
.
run
(
feed
=
py_rnn
.
to_feed
(
cpu
),
fetch_list
=
[
last
])
for
_
in
xrange
(
2
):
last_np
,
w_g
,
i_g
=
map
(
numpy
.
array
,
exe
.
run
(
feed
=
feed
,
fetch_list
=
[
last
,
self
.
PARAM_NAME
+
"@GRAD"
,
self
.
DATA_NAME
+
"@GRAD"
],
return_numpy
=
False
))
last_by_py
,
=
py_rnn
.
exe
().
values
()
last_by_py
,
=
py_rnn
.
exe
().
values
()
self
.
assertTrue
(
numpy
.
allclose
(
last_np
,
last_by_py
))
self
.
assertTrue
(
numpy
.
allclose
(
last_np
,
last_by_py
))
w_g_num
=
py_rnn
.
get_numeric_gradient_of_param
(
self
.
PARAM_NAME
)
print
w_g
[
0
],
w_g_num
[
0
]
self
.
assertTrue
(
numpy
.
allclose
(
w_g_num
,
w_g
,
rtol
=
0.1
))
i_g_num
=
py_rnn
.
get_numeric_gradient_of_input
(
self
.
DATA_NAME
)
i_g_num
=
i_g_num
.
reshape
(
i_g
.
shape
)
# Since this RNN has many float add. The number could be not stable.
# rtol = 0.1
self
.
assertTrue
(
numpy
.
allclose
(
i_g_num
,
i_g
,
rtol
=
0.1
))
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录