未验证 提交 994217ea 编写于 作者: D Dong Daxiang 提交者: GitHub

【paddle.fleet】fix api documents (#26777)

* fix api document
上级 db68e085
...@@ -30,8 +30,11 @@ __all__ = ["spawn"] ...@@ -30,8 +30,11 @@ __all__ = ["spawn"]
# dygraph parallel apis # dygraph parallel apis
__all__ += [ __all__ += [
"init_parallel_env", "get_rank", "get_world_size", "prepare_context", "init_parallel_env",
"ParallelEnv" "get_rank",
"get_world_size",
"prepare_context",
"ParallelEnv",
] ]
# collective apis # collective apis
......
...@@ -18,16 +18,15 @@ from .base.distributed_strategy import DistributedStrategy ...@@ -18,16 +18,15 @@ from .base.distributed_strategy import DistributedStrategy
from .base.fleet_base import Fleet from .base.fleet_base import Fleet
from .base.util_factory import UtilBase from .base.util_factory import UtilBase
from .dataset import * from .dataset import *
#from . import metrics
__all__ = [ __all__ = [
"DistributedStrategy", "DistributedStrategy",
"UtilBase", "UtilBase",
"DatasetFactory", "DatasetFactory",
"DatasetBase",
"InMemoryDataset",
"QueueDataset",
"UserDefinedRoleMaker", "UserDefinedRoleMaker",
"PaddleCloudRoleMaker", "PaddleCloudRoleMaker",
"Fleet",
] ]
fleet = Fleet() fleet = Fleet()
......
...@@ -17,6 +17,8 @@ from paddle.distributed.fleet.proto import distributed_strategy_pb2 ...@@ -17,6 +17,8 @@ from paddle.distributed.fleet.proto import distributed_strategy_pb2
from paddle.fluid.framework import Variable, set_flags, core from paddle.fluid.framework import Variable, set_flags, core
import google.protobuf.text_format import google.protobuf.text_format
__all__ = ["DistributedStrategy"]
def get_msg_dict(msg): def get_msg_dict(msg):
res_dict = {} res_dict = {}
......
...@@ -22,7 +22,7 @@ from .runtime_factory import RuntimeFactory ...@@ -22,7 +22,7 @@ from .runtime_factory import RuntimeFactory
from .util_factory import UtilFactory from .util_factory import UtilFactory
from paddle.fluid.wrapped_decorator import wrap_decorator from paddle.fluid.wrapped_decorator import wrap_decorator
__all__ = ['Fleet'] #__all__ = ['Fleet']
def _inited_runtime_handler_(func): def _inited_runtime_handler_(func):
......
...@@ -12,8 +12,6 @@ ...@@ -12,8 +12,6 @@
# See the License for the specific language governing permissions and # See the License for the specific language governing permissions and
# limitations under the License. # limitations under the License.
__all__ = ["MetaOptimizerFactory"]
from ..meta_optimizers import * from ..meta_optimizers import *
meta_optimizer_names = list( meta_optimizer_names = list(
......
...@@ -17,7 +17,7 @@ import numpy as np ...@@ -17,7 +17,7 @@ import numpy as np
from multiprocessing import Process, Manager from multiprocessing import Process, Manager
import paddle.fluid as fluid import paddle.fluid as fluid
__all__ = ['RoleMakerBase', 'UserDefinedRoleMaker', 'PaddleCloudRoleMaker'] #__all__ = ['UserDefinedRoleMaker', 'PaddleCloudRoleMaker']
class Role: class Role:
......
...@@ -22,17 +22,3 @@ from .lars_optimizer import LarsOptimizer ...@@ -22,17 +22,3 @@ from .lars_optimizer import LarsOptimizer
from .async_graph_execution_optimizer import AsyncGraphExecutionOptimizer from .async_graph_execution_optimizer import AsyncGraphExecutionOptimizer
from .dgc_optimizer import DGCOptimizer from .dgc_optimizer import DGCOptimizer
from .lamb_optimizer import LambOptimizer from .lamb_optimizer import LambOptimizer
__all__ = [
'AMPOptimizer',
'RecomputeOptimizer',
'GradientMergeOptimizer',
'AsyncMetaOptimizer',
'GraphExecutionOptimizer',
'PipelineOptimizer',
'LocalSGDOptimizer',
'LarsOptimizer',
'AsyncGraphExecutionOptimizer',
'DGCOptimizer',
'LambOptimizer',
]
...@@ -14,8 +14,6 @@ ...@@ -14,8 +14,6 @@
import paddle.fluid.contrib.mixed_precision as mixed_precision import paddle.fluid.contrib.mixed_precision as mixed_precision
from .meta_optimizer_base import MetaOptimizerBase from .meta_optimizer_base import MetaOptimizerBase
__all__ = ["AMPOptimizer"]
class AMPOptimizer(MetaOptimizerBase): class AMPOptimizer(MetaOptimizerBase):
def __init__(self, optimizer): def __init__(self, optimizer):
......
...@@ -15,8 +15,6 @@ from paddle.fluid.optimizer import Momentum, DGCMomentumOptimizer ...@@ -15,8 +15,6 @@ from paddle.fluid.optimizer import Momentum, DGCMomentumOptimizer
from .meta_optimizer_base import MetaOptimizerBase from .meta_optimizer_base import MetaOptimizerBase
import logging import logging
__all__ = ["DGCOptimizer"]
class DGCOptimizer(MetaOptimizerBase): class DGCOptimizer(MetaOptimizerBase):
def __init__(self, optimizer): def __init__(self, optimizer):
......
...@@ -14,10 +14,6 @@ ...@@ -14,10 +14,6 @@
from paddle.fluid.optimizer import GradientMergeOptimizer as GM from paddle.fluid.optimizer import GradientMergeOptimizer as GM
from .meta_optimizer_base import MetaOptimizerBase from .meta_optimizer_base import MetaOptimizerBase
__all__ = ["GradientMergeOptimizer"]
# amp + gradient merge + lamb
class GradientMergeOptimizer(MetaOptimizerBase): class GradientMergeOptimizer(MetaOptimizerBase):
def __init__(self, optimizer): def __init__(self, optimizer):
......
...@@ -16,8 +16,6 @@ from paddle.fluid.optimizer import LambOptimizer as LAMB ...@@ -16,8 +16,6 @@ from paddle.fluid.optimizer import LambOptimizer as LAMB
from .meta_optimizer_base import MetaOptimizerBase from .meta_optimizer_base import MetaOptimizerBase
import logging import logging
__all__ = ["LambOptimizer"]
class LambOptimizer(MetaOptimizerBase): class LambOptimizer(MetaOptimizerBase):
def __init__(self, optimizer): def __init__(self, optimizer):
......
...@@ -15,8 +15,6 @@ from paddle.fluid.optimizer import Momentum, LarsMomentumOptimizer ...@@ -15,8 +15,6 @@ from paddle.fluid.optimizer import Momentum, LarsMomentumOptimizer
from .meta_optimizer_base import MetaOptimizerBase from .meta_optimizer_base import MetaOptimizerBase
import logging import logging
__all__ = ["LarsOptimizer"]
class LarsOptimizer(MetaOptimizerBase): class LarsOptimizer(MetaOptimizerBase):
def __init__(self, optimizer): def __init__(self, optimizer):
......
...@@ -12,8 +12,6 @@ ...@@ -12,8 +12,6 @@
# See the License for the specific language governing permissions and # See the License for the specific language governing permissions and
# limitations under the License. # limitations under the License.
__all__ = ["MetaOptimizerBase"]
from paddle.fluid.optimizer import Optimizer from paddle.fluid.optimizer import Optimizer
......
...@@ -20,8 +20,6 @@ from paddle.fluid.optimizer import PipelineOptimizer as PO ...@@ -20,8 +20,6 @@ from paddle.fluid.optimizer import PipelineOptimizer as PO
from .meta_optimizer_base import MetaOptimizerBase from .meta_optimizer_base import MetaOptimizerBase
from .common import OpRole, OP_ROLE_KEY, OP_ROLE_VAR_KEY, CollectiveHelper, is_update_op, is_loss_grad_op, is_backward_op, is_optimizer_op from .common import OpRole, OP_ROLE_KEY, OP_ROLE_VAR_KEY, CollectiveHelper, is_update_op, is_loss_grad_op, is_backward_op, is_optimizer_op
__all__ = ["PipelineOptimizer"]
class PipelineHelper(CollectiveHelper): class PipelineHelper(CollectiveHelper):
def __init__(self, role_maker, nrings=1, wait_port='6174'): def __init__(self, role_maker, nrings=1, wait_port='6174'):
......
...@@ -14,8 +14,6 @@ ...@@ -14,8 +14,6 @@
from paddle.fluid.optimizer import RecomputeOptimizer as RO from paddle.fluid.optimizer import RecomputeOptimizer as RO
from .meta_optimizer_base import MetaOptimizerBase from .meta_optimizer_base import MetaOptimizerBase
__all__ = ["RecomputeOptimizer"]
class RecomputeOptimizer(MetaOptimizerBase): class RecomputeOptimizer(MetaOptimizerBase):
def __init__(self, optimizer): def __init__(self, optimizer):
......
...@@ -11,3 +11,16 @@ ...@@ -11,3 +11,16 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and # See the License for the specific language governing permissions and
# limitations under the License. # limitations under the License.
from .metric import *
__all__ = [
"sum",
"max",
"min",
"auc",
"mae",
"rmse",
"mse",
"acc",
]
...@@ -14,5 +14,3 @@ ...@@ -14,5 +14,3 @@
from .collective_runtime import CollectiveRuntime from .collective_runtime import CollectiveRuntime
from .parameter_server_runtime import ParameterServerRuntime from .parameter_server_runtime import ParameterServerRuntime
__all__ = ["CollectiveRuntime," "ParameterServerRuntime", ]
...@@ -15,4 +15,4 @@ ...@@ -15,4 +15,4 @@
from .fs import * from .fs import *
from .http_server import KVHandler, KVHTTPServer, KVServer from .http_server import KVHandler, KVHTTPServer, KVServer
__all__ = ['KVHandler', 'KVHTTPServer', 'KVServer'] + fs.__all__ #__all__ = ['KVHandler', 'KVHTTPServer', 'KVServer'] + fs.__all__
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册