Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
Paddle
提交
96180fff
P
Paddle
项目概览
PaddlePaddle
/
Paddle
大约 1 年 前同步成功
通知
2298
Star
20931
Fork
5422
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1423
列表
看板
标记
里程碑
合并请求
543
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1,423
Issue
1,423
列表
看板
标记
里程碑
合并请求
543
合并请求
543
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
96180fff
编写于
4月 28, 2023
作者:
M
Meteor Liu
提交者:
GitHub
4月 28, 2023
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
replace varbase to relevant name or notes as per the context (#53431)
上级
396fe483
变更
9
隐藏空白更改
内联
并排
Showing
9 changed file
with
90 addition
and
96 deletion
+90
-96
python/paddle/distributed/fleet/meta_parallel/sharding/group_sharded_stage3.py
...uted/fleet/meta_parallel/sharding/group_sharded_stage3.py
+6
-6
python/paddle/distributed/parallel.py
python/paddle/distributed/parallel.py
+13
-17
python/paddle/fluid/dygraph/base.py
python/paddle/fluid/dygraph/base.py
+1
-1
python/paddle/fluid/tests/unittests/test_eager_run_program.py
...on/paddle/fluid/tests/unittests/test_eager_run_program.py
+2
-3
python/paddle/fluid/tests/unittests/test_imperative_group.py
python/paddle/fluid/tests/unittests/test_imperative_group.py
+53
-53
python/paddle/fluid/tests/unittests/test_view_op_reuse_allocation.py
...le/fluid/tests/unittests/test_view_op_reuse_allocation.py
+2
-2
python/paddle/jit/dy2static/partial_program.py
python/paddle/jit/dy2static/partial_program.py
+7
-8
python/paddle/jit/translated_layer.py
python/paddle/jit/translated_layer.py
+3
-3
test/dygraph_to_static/test_list.py
test/dygraph_to_static/test_list.py
+3
-3
未找到文件。
python/paddle/distributed/fleet/meta_parallel/sharding/group_sharded_stage3.py
浏览文件 @
96180fff
...
@@ -584,7 +584,7 @@ class GroupShardedStage3(nn.Layer):
...
@@ -584,7 +584,7 @@ class GroupShardedStage3(nn.Layer):
param
,
"fw_storage"
param
,
"fw_storage"
),
f
"Find
{
param
.
name
}
don't have fw_storage attribute"
),
f
"Find
{
param
.
name
}
don't have fw_storage attribute"
param
.
fw_storage
=
_
VarBase
Wrapper
(
param
)
param
.
fw_storage
=
_
Tensor
Wrapper
(
param
)
assert
param
.
fw_storage
.
grad
is
None
assert
param
.
fw_storage
.
grad
is
None
param
.
fw_storage
.
_copy_gradient_from
(
param
.
bw_storage
)
param
.
fw_storage
.
_copy_gradient_from
(
param
.
bw_storage
)
update_list
.
append
(
param
)
update_list
.
append
(
param
)
...
@@ -1062,17 +1062,17 @@ def _UnsliceParam(param):
...
@@ -1062,17 +1062,17 @@ def _UnsliceParam(param):
return
param
return
param
def
_
VarBase
Wrapper
(
param
):
def
_
Tensor
Wrapper
(
param
):
var
base
=
param
.
fw_storage
var
=
param
.
fw_storage
tmp_param
=
EagerParamBase
(
tmp_param
=
EagerParamBase
(
shape
=
var
base
.
shape
,
dtype
=
varbase
.
dtype
,
name
=
"slice@"
+
param
.
name
shape
=
var
.
shape
,
dtype
=
var
.
dtype
,
name
=
"slice@"
+
param
.
name
)
)
var
base
.
_share_buffer_to
(
tmp_param
)
var
.
_share_buffer_to
(
tmp_param
)
tmp_param
.
regularizer
=
param
.
regularizer
tmp_param
.
regularizer
=
param
.
regularizer
tmp_param
.
optimize_attr
[
'learning_rate'
]
=
param
.
optimize_attr
[
tmp_param
.
optimize_attr
[
'learning_rate'
]
=
param
.
optimize_attr
[
'learning_rate'
'learning_rate'
]
]
var
base
.
_clear
()
var
.
_clear
()
return
tmp_param
return
tmp_param
...
...
python/paddle/distributed/parallel.py
浏览文件 @
96180fff
...
@@ -146,20 +146,18 @@ def sync_params_buffers(
...
@@ -146,20 +146,18 @@ def sync_params_buffers(
for
_
,
param
in
model
.
_obtain_parameters_buffers
().
items
():
for
_
,
param
in
model
.
_obtain_parameters_buffers
().
items
():
if
not
isinstance
(
param
,
core
.
eager
.
Tensor
):
if
not
isinstance
(
param
,
core
.
eager
.
Tensor
):
raise
TypeError
(
raise
TypeError
(
"The data type of '%s' must be Varbase or eager.Tensor"
"The data type of '%s' must be core.eager.Tensor"
%
param
.
name
%
param
.
name
)
)
# is_distributed param not need to sync when in mp mode
if
is_model_parallel
:
if
isinstance
(
param
,
core
.
eager
.
Tensor
):
if
hasattr
(
param
,
"is_distributed"
)
and
param
.
is_distributed
:
if
is_model_parallel
:
if
hasattr
(
param
,
"is_distributed"
)
and
param
.
is_distributed
:
continue
# NOTE(shenliang03): Support situations that do not require synchronization parameters,
# such as moe's expert parameters
if
getattr
(
param
,
"no_sync"
,
False
):
continue
continue
# NOTE(shenliang03): Support situations that do not require synchronization parameters,
# such as moe's expert parameters
if
getattr
(
param
,
"no_sync"
,
False
):
continue
if
param
.
type
==
core
.
VarDesc
.
VarType
.
VOCAB
:
if
param
.
type
==
core
.
VarDesc
.
VarType
.
VOCAB
:
continue
continue
...
@@ -474,14 +472,14 @@ class DataParallel(layers.Layer):
...
@@ -474,14 +472,14 @@ class DataParallel(layers.Layer):
self
.
find_unused_parameters
,
self
.
find_unused_parameters
,
)
)
def
_find_
varbase
(
self
,
obj
):
def
_find_
tensor
(
self
,
obj
):
var_type
=
core
.
eager
.
Tensor
var_type
=
core
.
eager
.
Tensor
if
isinstance
(
obj
,
var_type
):
if
isinstance
(
obj
,
var_type
):
return
[
obj
]
return
[
obj
]
if
isinstance
(
obj
,
(
list
,
tuple
)):
if
isinstance
(
obj
,
(
list
,
tuple
)):
return
itertools
.
chain
(
*
map
(
self
.
_find_
varbase
,
obj
))
return
itertools
.
chain
(
*
map
(
self
.
_find_
tensor
,
obj
))
if
isinstance
(
obj
,
dict
):
if
isinstance
(
obj
,
dict
):
return
itertools
.
chain
(
*
map
(
self
.
_find_
varbase
,
obj
.
values
()))
return
itertools
.
chain
(
*
map
(
self
.
_find_
tensor
,
obj
.
values
()))
return
[]
return
[]
@
contextmanager
@
contextmanager
...
@@ -536,9 +534,7 @@ class DataParallel(layers.Layer):
...
@@ -536,9 +534,7 @@ class DataParallel(layers.Layer):
and
framework
.
_dygraph_tracer
().
_has_grad
and
framework
.
_dygraph_tracer
().
_has_grad
and
self
.
grad_need_sync
and
self
.
grad_need_sync
):
):
self
.
_reducer
.
prepare_for_backward
(
self
.
_reducer
.
prepare_for_backward
(
list
(
self
.
_find_tensor
(
outputs
)))
list
(
self
.
_find_varbase
(
outputs
))
)
return
outputs
return
outputs
@
deprecated
(
@
deprecated
(
...
...
python/paddle/fluid/dygraph/base.py
浏览文件 @
96180fff
...
@@ -125,7 +125,7 @@ def param_guard(parameters):
...
@@ -125,7 +125,7 @@ def param_guard(parameters):
def
_convert_into_variable
(
tensor
):
def
_convert_into_variable
(
tensor
):
"""
"""
Convert
Varbase
into Variable.
Convert
Tensor
into Variable.
"""
"""
if
isinstance
(
tensor
,
core
.
eager
.
Tensor
):
if
isinstance
(
tensor
,
core
.
eager
.
Tensor
):
# Check whether has been created before.
# Check whether has been created before.
...
...
python/paddle/fluid/tests/unittests/test_eager_run_program.py
浏览文件 @
96180fff
...
@@ -57,15 +57,14 @@ def _append_backward_desc(main_program, outs):
...
@@ -57,15 +57,14 @@ def _append_backward_desc(main_program, outs):
def
_create_out
(
var
):
def
_create_out
(
var
):
assert
isinstance
(
var
,
Variable
)
assert
isinstance
(
var
,
Variable
)
var_desc
=
var
.
desc
var_desc
=
var
.
desc
varbase
=
None
out
=
core
.
eager
.
Tensor
(
var_base
=
core
.
eager
.
Tensor
(
var_desc
.
dtype
(),
var_desc
.
dtype
(),
var_desc
.
shape
(),
var_desc
.
shape
(),
var_desc
.
name
(),
var_desc
.
name
(),
var_desc
.
type
(),
var_desc
.
type
(),
False
,
False
,
)
)
return
var_base
return
out
@
switch_to_static_graph
@
switch_to_static_graph
...
...
python/paddle/fluid/tests/unittests/test_imperative_group.py
浏览文件 @
96180fff
...
@@ -20,7 +20,7 @@ from paddle.fluid.framework import in_dygraph_mode
...
@@ -20,7 +20,7 @@ from paddle.fluid.framework import in_dygraph_mode
class
TestDataParallelGroup
(
unittest
.
TestCase
):
class
TestDataParallelGroup
(
unittest
.
TestCase
):
def
create_varbase
(
self
,
dtype
,
shape
):
def
_create_var
(
self
,
dtype
,
shape
):
return
paddle
.
rand
(
shape
=
shape
,
dtype
=
dtype
)
return
paddle
.
rand
(
shape
=
shape
,
dtype
=
dtype
)
def
assign_group_by_size
(
self
,
*
args
):
def
assign_group_by_size
(
self
,
*
args
):
...
@@ -30,10 +30,10 @@ class TestDataParallelGroup(unittest.TestCase):
...
@@ -30,10 +30,10 @@ class TestDataParallelGroup(unittest.TestCase):
def
test_construct_group0
(
self
):
def
test_construct_group0
(
self
):
# one dtype & one limit capability
# one dtype & one limit capability
var_list
=
[]
var_list
=
[]
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
100
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
100
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
25
]))
res
=
self
.
assign_group_by_size
(
res
=
self
.
assign_group_by_size
(
var_list
,
[
False
,
False
,
False
,
False
],
[
400
]
var_list
,
[
False
,
False
,
False
,
False
],
[
400
]
)
)
...
@@ -42,12 +42,12 @@ class TestDataParallelGroup(unittest.TestCase):
...
@@ -42,12 +42,12 @@ class TestDataParallelGroup(unittest.TestCase):
def
test_construct_group1
(
self
):
def
test_construct_group1
(
self
):
# multi dtype & one limit capability
# multi dtype & one limit capability
var_list
=
[]
var_list
=
[]
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
res
=
self
.
assign_group_by_size
(
res
=
self
.
assign_group_by_size
(
var_list
,
[
False
,
False
,
False
,
False
,
False
,
False
],
[
400
]
var_list
,
[
False
,
False
,
False
,
False
,
False
,
False
],
[
400
]
)
)
...
@@ -56,10 +56,10 @@ class TestDataParallelGroup(unittest.TestCase):
...
@@ -56,10 +56,10 @@ class TestDataParallelGroup(unittest.TestCase):
def
test_construct_group2
(
self
):
def
test_construct_group2
(
self
):
# one dtype & multi limit capability
# one dtype & multi limit capability
var_list
=
[]
var_list
=
[]
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
50
]))
res
=
self
.
assign_group_by_size
(
res
=
self
.
assign_group_by_size
(
var_list
,
[
False
,
False
,
False
,
False
],
[
400
,
800
]
var_list
,
[
False
,
False
,
False
,
False
],
[
400
,
800
]
)
)
...
@@ -68,12 +68,12 @@ class TestDataParallelGroup(unittest.TestCase):
...
@@ -68,12 +68,12 @@ class TestDataParallelGroup(unittest.TestCase):
def
test_construct_group3
(
self
):
def
test_construct_group3
(
self
):
# multi dtype & multi limit capability
# multi dtype & multi limit capability
var_list
=
[]
var_list
=
[]
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
res
=
self
.
assign_group_by_size
(
res
=
self
.
assign_group_by_size
(
var_list
,
[
False
,
False
,
False
,
False
,
False
,
False
],
[
200
,
400
]
var_list
,
[
False
,
False
,
False
,
False
,
False
,
False
],
[
200
,
400
]
)
)
...
@@ -82,12 +82,12 @@ class TestDataParallelGroup(unittest.TestCase):
...
@@ -82,12 +82,12 @@ class TestDataParallelGroup(unittest.TestCase):
def
test_construct_group4
(
self
):
def
test_construct_group4
(
self
):
# multi dtype & zero limit capability
# multi dtype & zero limit capability
var_list
=
[]
var_list
=
[]
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
res
=
self
.
assign_group_by_size
(
res
=
self
.
assign_group_by_size
(
var_list
,
[
False
,
False
,
False
,
False
,
False
,
False
],
[
0
]
var_list
,
[
False
,
False
,
False
,
False
,
False
,
False
],
[
0
]
)
)
...
@@ -96,12 +96,12 @@ class TestDataParallelGroup(unittest.TestCase):
...
@@ -96,12 +96,12 @@ class TestDataParallelGroup(unittest.TestCase):
def
test_construct_group5
(
self
):
def
test_construct_group5
(
self
):
# multi dtype & infinite capability
# multi dtype & infinite capability
var_list
=
[]
var_list
=
[]
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
res
=
self
.
assign_group_by_size
(
res
=
self
.
assign_group_by_size
(
var_list
,
[
False
,
False
,
False
,
False
,
False
,
False
],
[
10000
]
var_list
,
[
False
,
False
,
False
,
False
,
False
,
False
],
[
10000
]
)
)
...
@@ -111,16 +111,16 @@ class TestDataParallelGroup(unittest.TestCase):
...
@@ -111,16 +111,16 @@ class TestDataParallelGroup(unittest.TestCase):
# multi dtype & limit capability & multi tensor type
# multi dtype & limit capability & multi tensor type
var_list
=
[]
var_list
=
[]
var_list
.
append
(
var_list
.
append
(
self
.
create_varbase
(
self
.
_create_var
(
"float32"
,
"float32"
,
[
1
,
50
],
[
1
,
50
],
)
)
)
)
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
res
=
self
.
assign_group_by_size
(
res
=
self
.
assign_group_by_size
(
var_list
,
[
True
,
False
,
False
,
False
,
False
,
True
],
[
400
]
var_list
,
[
True
,
False
,
False
,
False
,
False
,
True
],
[
400
]
)
)
...
@@ -129,12 +129,12 @@ class TestDataParallelGroup(unittest.TestCase):
...
@@ -129,12 +129,12 @@ class TestDataParallelGroup(unittest.TestCase):
def
test_construct_group7
(
self
):
def
test_construct_group7
(
self
):
# multi dtype & multi limit capability & multi tensor type
# multi dtype & multi limit capability & multi tensor type
var_list
=
[]
var_list
=
[]
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
1
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float64"
,
[
1
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float64"
,
[
1
,
25
]))
res
=
self
.
assign_group_by_size
(
res
=
self
.
assign_group_by_size
(
var_list
,
[
True
,
False
,
False
,
False
,
False
,
True
],
[
200
,
400
]
var_list
,
[
True
,
False
,
False
,
False
,
False
,
True
],
[
200
,
400
]
)
)
...
@@ -143,10 +143,10 @@ class TestDataParallelGroup(unittest.TestCase):
...
@@ -143,10 +143,10 @@ class TestDataParallelGroup(unittest.TestCase):
def
test_construct_group8
(
self
):
def
test_construct_group8
(
self
):
# one dtype & one limit capability & have tensor_indices
# one dtype & one limit capability & have tensor_indices
var_list
=
[]
var_list
=
[]
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
100
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
100
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
50
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
50
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
25
]))
res
=
self
.
assign_group_by_size
(
res
=
self
.
assign_group_by_size
(
var_list
,
[
False
,
False
,
False
,
False
],
[
400
],
[
3
,
0
,
1
,
2
]
var_list
,
[
False
,
False
,
False
,
False
],
[
400
],
[
3
,
0
,
1
,
2
]
)
)
...
@@ -155,10 +155,10 @@ class TestDataParallelGroup(unittest.TestCase):
...
@@ -155,10 +155,10 @@ class TestDataParallelGroup(unittest.TestCase):
def
test_construct_group9
(
self
):
def
test_construct_group9
(
self
):
# one dtype & one limit capability & have tensor_indices
# one dtype & one limit capability & have tensor_indices
var_list
=
[]
var_list
=
[]
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
25
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
25
]))
var_list
.
append
(
self
.
create_varbase
(
"float32"
,
[
2
,
1000
]))
var_list
.
append
(
self
.
_create_var
(
"float32"
,
[
2
,
1000
]))
res
=
self
.
assign_group_by_size
(
res
=
self
.
assign_group_by_size
(
var_list
,
[
False
,
False
,
False
,
True
],
[
300
],
[
1
,
0
,
2
,
3
]
var_list
,
[
False
,
False
,
False
,
True
],
[
300
],
[
1
,
0
,
2
,
3
]
)
)
...
...
python/paddle/fluid/tests/unittests/test_view_op_reuse_allocation.py
浏览文件 @
96180fff
...
@@ -21,8 +21,8 @@ import paddle
...
@@ -21,8 +21,8 @@ import paddle
# NOTE(pangyoki): Tensor View Strategy.
# NOTE(pangyoki): Tensor View Strategy.
# Refer to `op_function_generator.py`.
# Refer to `op_function_generator.py`.
# For view op, a new output
varbase will be created, and this varbase
will
# For view op, a new output
tensor will be created, and this tensor
will
# reuse the input
varbase
's allocation.
# reuse the input
tensor
's allocation.
# View APIs include: `squeeze`, `unsqueeze`, `reshape`, `flatten`, `detach`
# View APIs include: `squeeze`, `unsqueeze`, `reshape`, `flatten`, `detach`
class
TestDygraphViewReuseAllocation
(
unittest
.
TestCase
):
class
TestDygraphViewReuseAllocation
(
unittest
.
TestCase
):
def
setUp
(
self
):
def
setUp
(
self
):
...
...
python/paddle/jit/dy2static/partial_program.py
浏览文件 @
96180fff
...
@@ -913,27 +913,26 @@ class PartialProgramLayer:
...
@@ -913,27 +913,26 @@ class PartialProgramLayer:
input_vars
.
append
(
var
)
input_vars
.
append
(
var
)
# mapping from name(string) -> Tensor
# mapping from name(string) -> Tensor
out_
varbase
_map
=
{}
out_
tensor
_map
=
{}
def
create_out
(
var_id
):
def
create_out
(
var_id
):
var
=
self
.
_outputs
[
var_id
]
var
=
self
.
_outputs
[
var_id
]
assert
isinstance
(
var
,
framework
.
Variable
)
assert
isinstance
(
var
,
framework
.
Variable
)
var_desc
=
var
.
desc
var_desc
=
var
.
desc
varbase
=
None
if
var_desc
.
name
()
in
out_
varbase
_map
:
if
var_desc
.
name
()
in
out_
tensor
_map
:
return
out_
varbase
_map
[
var_desc
.
name
()]
return
out_
tensor
_map
[
var_desc
.
name
()]
var_base
=
core
.
eager
.
Tensor
(
out
=
core
.
eager
.
Tensor
(
var_desc
.
dtype
(),
var_desc
.
dtype
(),
var_desc
.
shape
(),
var_desc
.
shape
(),
var_desc
.
name
(),
var_desc
.
name
(),
var_desc
.
type
(),
var_desc
.
type
(),
False
,
False
,
)
)
var_base
.
stop_gradient
=
var
.
stop_gradient
out
.
stop_gradient
=
var
.
stop_gradient
out_
varbase_map
[
var_desc
.
name
()]
=
var_base
out_
tensor_map
[
var_desc
.
name
()]
=
out
return
var_base
return
out
# Create Tensor to receive output data.
# Create Tensor to receive output data.
out_vars
=
list
(
map
(
create_out
,
self
.
_outputs
.
var_ids
))
out_vars
=
list
(
map
(
create_out
,
self
.
_outputs
.
var_ids
))
...
...
python/paddle/jit/translated_layer.py
浏览文件 @
96180fff
...
@@ -644,7 +644,7 @@ class _ProgramHolder:
...
@@ -644,7 +644,7 @@ class _ProgramHolder:
# and executor, executes this program. Key points:
# and executor, executes this program. Key points:
#
#
# 1. Data Sharing:
# 1. Data Sharing:
# The var
Base
of the dynamic graph is not in the scope, so before the op
# The var
iable/parameter
of the dynamic graph is not in the scope, so before the op
# executes the program internally, create persistent variables with the
# executes the program internally, create persistent variables with the
# same name as feed, parameters, and fetch in the scope, and share the
# same name as feed, parameters, and fetch in the scope, and share the
# LoDTensor of the op input.
# LoDTensor of the op input.
...
@@ -668,7 +668,7 @@ def _load_persistable_vars_by_program(
...
@@ -668,7 +668,7 @@ def _load_persistable_vars_by_program(
for
each_var
in
persistable_vars
:
for
each_var
in
persistable_vars
:
orig_each_name
=
program_holder
.
_suffix_varname_dict
[
each_var
.
name
()]
orig_each_name
=
program_holder
.
_suffix_varname_dict
[
each_var
.
name
()]
if
_is_parameter
(
each_var
,
program_holder
.
infer_program
):
if
_is_parameter
(
each_var
,
program_holder
.
infer_program
):
# create output
varbase
# create output
param
new_var
=
framework
.
EagerParamBase
(
new_var
=
framework
.
EagerParamBase
(
shape
=
each_var
.
shape
(),
shape
=
each_var
.
shape
(),
dtype
=
each_var
.
dtype
(),
dtype
=
each_var
.
dtype
(),
...
@@ -755,7 +755,7 @@ def _load_persistable_vars(
...
@@ -755,7 +755,7 @@ def _load_persistable_vars(
)
)
# get suffix var name, see [why need to append suffix to persistable vars]
# get suffix var name, see [why need to append suffix to persistable vars]
new_name
=
inv_suffix_varname_dict
[
name
]
new_name
=
inv_suffix_varname_dict
[
name
]
# create output var
base
# create output var
or param
if
extra_var_info
[
name
].
get
(
'trainable'
,
None
)
is
not
None
:
if
extra_var_info
[
name
].
get
(
'trainable'
,
None
)
is
not
None
:
# use default shape and dtype
# use default shape and dtype
new_var
=
framework
.
EagerParamBase
(
new_var
=
framework
.
EagerParamBase
(
...
...
test/dygraph_to_static/test_list.py
浏览文件 @
96180fff
...
@@ -228,7 +228,7 @@ class TestListWithoutControlFlow(unittest.TestCase):
...
@@ -228,7 +228,7 @@ class TestListWithoutControlFlow(unittest.TestCase):
test_list_pop_without_control_flow_2
,
test_list_pop_without_control_flow_2
,
]
]
def
varbase
_to_numpy
(
self
,
res
):
def
result
_to_numpy
(
self
,
res
):
if
isinstance
(
res
,
(
list
,
tuple
)):
if
isinstance
(
res
,
(
list
,
tuple
)):
res
=
paddle
.
utils
.
map_structure
(
lambda
x
:
x
.
numpy
(),
res
)
res
=
paddle
.
utils
.
map_structure
(
lambda
x
:
x
.
numpy
(),
res
)
else
:
else
:
...
@@ -248,7 +248,7 @@ class TestListWithoutControlFlow(unittest.TestCase):
...
@@ -248,7 +248,7 @@ class TestListWithoutControlFlow(unittest.TestCase):
res
=
paddle
.
jit
.
to_static
(
self
.
dygraph_func
)(
self
.
input
)
res
=
paddle
.
jit
.
to_static
(
self
.
dygraph_func
)(
self
.
input
)
else
:
else
:
res
=
self
.
dygraph_func
(
self
.
input
)
res
=
self
.
dygraph_func
(
self
.
input
)
return
self
.
varbase
_to_numpy
(
res
)
return
self
.
result
_to_numpy
(
res
)
def
test_transformed_static_result
(
self
):
def
test_transformed_static_result
(
self
):
for
dyfunc
in
self
.
all_dygraph_funcs
:
for
dyfunc
in
self
.
all_dygraph_funcs
:
...
@@ -294,7 +294,7 @@ class TestListInWhileLoop(TestListWithoutControlFlow):
...
@@ -294,7 +294,7 @@ class TestListInWhileLoop(TestListWithoutControlFlow):
)
)
else
:
else
:
res
=
self
.
dygraph_func
(
self
.
input
,
self
.
iter_num
)
res
=
self
.
dygraph_func
(
self
.
input
,
self
.
iter_num
)
return
self
.
varbase
_to_numpy
(
res
)
return
self
.
result
_to_numpy
(
res
)
class
TestListInWhileLoopWithStack
(
TestListInWhileLoop
):
class
TestListInWhileLoopWithStack
(
TestListInWhileLoop
):
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录