未验证 提交 aab713ea 编写于 作者: W wangzhen38 提交者: GitHub

[clear ps] del some all list (#51289)

* [clear ps] del some all list

* [clear ps] for ci

* [clear ps] rm fluid emb
上级 f66a2040
...@@ -40,6 +40,9 @@ from paddle.incubate.distributed.fleet.base import Mode ...@@ -40,6 +40,9 @@ from paddle.incubate.distributed.fleet.base import Mode
from paddle.incubate.distributed.fleet.role_maker import MPISymetricRoleMaker from paddle.incubate.distributed.fleet.role_maker import MPISymetricRoleMaker
from paddle.incubate.distributed.fleet.parameter_server import version from paddle.incubate.distributed.fleet.parameter_server import version
from paddle.incubate.distributed.fleet.parameter_server.pslib.optimizer_factory import (
DistributedAdam,
)
from paddle.incubate.distributed.fleet.parameter_server.ir.public import ( from paddle.incubate.distributed.fleet.parameter_server.ir.public import (
get_sparse_tablenames, get_sparse_tablenames,
) )
......
...@@ -14,15 +14,16 @@ ...@@ -14,15 +14,16 @@
import os import os
import sys import sys
from .optimizer_factory import * # noqa: F403 from .optimizer_factory import FLEET_GLOBAL_DICT # noqa: F403
from .optimizer_factory import DistributedAdam # noqa: F403
from google.protobuf import text_format from google.protobuf import text_format
from paddle.framework import core from paddle.framework import core
from paddle.incubate.distributed.fleet.base import Fleet from paddle.incubate.distributed.fleet.base import Fleet
from paddle.incubate.distributed.fleet.base import Mode from paddle.incubate.distributed.fleet.base import Mode
from paddle.incubate.distributed.fleet.base import DistributedOptimizer from paddle.incubate.distributed.fleet.base import DistributedOptimizer
from paddle.incubate.distributed.fleet.role_maker import MPISymetricRoleMaker from paddle.incubate.distributed.fleet.role_maker import MPISymetricRoleMaker
from paddle.incubate.distributed.fleet.role_maker import HeterRoleMaker from paddle.incubate.distributed.fleet.role_maker import HeterRoleMaker
import paddle
class PSLib(Fleet): class PSLib(Fleet):
...@@ -534,7 +535,7 @@ class PSLib(Fleet): ...@@ -534,7 +535,7 @@ class PSLib(Fleet):
>>> fleet.shrink_dense_table(0.98, 11, myscope2, 3) >>> fleet.shrink_dense_table(0.98, 11, myscope2, 3)
""" """
if scope is None: if scope is None:
scope = fluid.global_scope() scope = paddle.static.global_scope()
self._role_maker._barrier_worker() self._role_maker._barrier_worker()
if self._role_maker.is_first_worker(): if self._role_maker.is_first_worker():
for tp in self._opt_info["fleet_desc"].trainer_param: for tp in self._opt_info["fleet_desc"].trainer_param:
...@@ -971,7 +972,7 @@ def _fleet_embedding( ...@@ -971,7 +972,7 @@ def _fleet_embedding(
if padding_idx is None: if padding_idx is None:
padding_idx = 0 padding_idx = 0
global FLEET_GLOBAL_DICT global FLEET_GLOBAL_DICT
return fluid.layers.nn._pull_sparse( return paddle.static.nn._pull_sparse(
input=input, input=input,
size=size, size=size,
table_id=FLEET_GLOBAL_DICT["emb_to_table"][name], table_id=FLEET_GLOBAL_DICT["emb_to_table"][name],
...@@ -1013,7 +1014,7 @@ def _fleet_embedding_v2( ...@@ -1013,7 +1014,7 @@ def _fleet_embedding_v2(
if padding_idx is None: if padding_idx is None:
padding_idx = 0 padding_idx = 0
return fluid.layers.nn._pull_sparse_v2( return paddle.static.nn._pull_sparse_v2(
input=input, input=input,
size=size, size=size,
table_id=FLEET_GLOBAL_DICT["emb_to_table"][name], table_id=FLEET_GLOBAL_DICT["emb_to_table"][name],
...@@ -1042,8 +1043,8 @@ class fleet_embedding: ...@@ -1042,8 +1043,8 @@ class fleet_embedding:
def __init__(self, click_name, scale_sparse_grad=True): def __init__(self, click_name, scale_sparse_grad=True):
"""Init.""" """Init."""
self.origin_emb = fluid.layers.embedding # self.origin_emb = fluid.layers.embedding
self.origin_emb_v2 = fluid.embedding self.origin_emb_v2 = paddle.static.nn.embedding
# if user uses cvm layer after embedding, click_name can be None # if user uses cvm layer after embedding, click_name can be None
self.click_name = "" if click_name is None else click_name self.click_name = "" if click_name is None else click_name
self.scale_sparse_grad = scale_sparse_grad self.scale_sparse_grad = scale_sparse_grad
...@@ -1052,16 +1053,16 @@ class fleet_embedding: ...@@ -1052,16 +1053,16 @@ class fleet_embedding:
def __enter__(self): def __enter__(self):
"""Enter.""" """Enter."""
fluid.layers.embedding = _fleet_embedding # fluid.layers.embedding = _fleet_embedding
fluid.embedding = _fleet_embedding_v2 paddle.static.nn.embedding = _fleet_embedding_v2
FLEET_GLOBAL_DICT["cur_accessor"] = self.accessor FLEET_GLOBAL_DICT["cur_accessor"] = self.accessor
FLEET_GLOBAL_DICT["click_name"] = self.click_name FLEET_GLOBAL_DICT["click_name"] = self.click_name
FLEET_GLOBAL_DICT["scale_sparse_grad"] = self.scale_sparse_grad FLEET_GLOBAL_DICT["scale_sparse_grad"] = self.scale_sparse_grad
def __exit__(self, exc_type, exc_val, exc_tb): def __exit__(self, exc_type, exc_val, exc_tb):
"""Exit.""" """Exit."""
fluid.layers.embedding = self.origin_emb # fluid.layers.embedding = self.origin_emb
fluid.embedding = self.origin_emb_v2 paddle.static.nn.embedding = self.origin_emb_v2
FLEET_GLOBAL_DICT["cur_accessor"] = "" FLEET_GLOBAL_DICT["cur_accessor"] = ""
FLEET_GLOBAL_DICT["click_name"] = "" FLEET_GLOBAL_DICT["click_name"] = ""
FLEET_GLOBAL_DICT["scale_sparse_grad"] = None FLEET_GLOBAL_DICT["scale_sparse_grad"] = None
...@@ -1220,7 +1221,7 @@ class DownpourOptimizer(DistributedOptimizer): ...@@ -1220,7 +1221,7 @@ class DownpourOptimizer(DistributedOptimizer):
programs = [loss.block.program for loss in losses] programs = [loss.block.program for loss in losses]
if scopes is None: if scopes is None:
scopes = [fluid.global_scope()] * len(programs) scopes = [paddle.static.global_scope()] * len(programs)
if len(scopes) != len(programs): if len(scopes) != len(programs):
raise ValueError( raise ValueError(
......
...@@ -13,7 +13,7 @@ ...@@ -13,7 +13,7 @@
# limitations under the License. # limitations under the License.
"""Optimizer Factory.""" """Optimizer Factory."""
__all__ = ["DistributedAdam", "FLEET_GLOBAL_DICT"] __all__ = []
import copy import copy
import logging import logging
import os import os
......
...@@ -41,6 +41,8 @@ from .common import layer_norm # noqa: F401 ...@@ -41,6 +41,8 @@ from .common import layer_norm # noqa: F401
from .common import embedding # noqa: F401 from .common import embedding # noqa: F401
from ...fluid.contrib.layers import sparse_embedding # noqa: F401 from ...fluid.contrib.layers import sparse_embedding # noqa: F401
from ...fluid.layers import StaticRNN # noqa: F401 from ...fluid.layers import StaticRNN # noqa: F401
from ...fluid.layers.nn import _pull_sparse # noqa: F401
from ...fluid.layers.nn import _pull_sparse_v2 # noqa: F401
from .sequence_lod import sequence_conv # noqa: F401 from .sequence_lod import sequence_conv # noqa: F401
from .sequence_lod import sequence_softmax # noqa: F401 from .sequence_lod import sequence_softmax # noqa: F401
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册