1. 01 2月, 2023 1 次提交
    • Z
      remove fluid.initializer.UniformInitializer, ConstantInitializer,... · 6edc7bba
      zqw_1997 提交于
      remove fluid.initializer.UniformInitializer, ConstantInitializer, NormalInitializer, TruncatedNormalInitializer, XavierInitializer, BilinearInitializer, MSRAInitializer, NumpyArrayInitializer and calculate_gain.. (#49498)
      
      * move UniformInitializer and ConstantInitializer
      
      * more modify
      
      * circular import resolved
      
      * another circular import resolved?
      
      * more circular import 2
      
      * circular import 3
      
      * change import paddle in metric.py
      
      * BuildStrategy import from fluid
      
      * modify the framework import path in common.py
      
      * change rnn.py import, from static to original framework
      
      * change import static in the nn folder
      
      * default_main_program should import from common_ops_import
      
      * add import paddle in param_attr.py
      
      * use core not paddle module for using VarDesc
      
      * another old uniform
      
      * mistake that use Uniform instead of UniformInitializer
      
      * modify UniformInitializer doc
      
      * move fluid.NormalInitializer to nn.initializer.NormalInitializer
      
      * remove import of Normal in fluid.layers.nn.py
      
      * remove more import of old Normal
      
      * remove more import of old Normal
      
      * sample code modify and tests modify import
      
      * is_listen_failed passing arg should be log file
      
      * problem solved
      
      * a mistake solved
      
      * comments resoleved and remove paddle.fluid.initializer.TruncatedNormalInitializer
      
      * remove paddle.fluid.initializer.XavierInitializer and paddle.fluid.initializer.MSRAInitializer
      
      * remove paddle.fluid.initializer.BilinearInitializer NumpyArrayInitializer and set_global_initializer
      
      * change fluid to static
      
      * change static to fluid to avoid circular import in distributed_strategy.py
      
      * fix example code and test_initializer
      
      * ValueType
      
      * sample code fix
      
      * change set_global_initializer back to fluid
      
      * put paddle.static.BuildStrategy.ReduceStrategy into the fuction to avoid circular import
      
      * remove calculate_gain, delete BilinearInitializer and revert set_global_initializer
      
      * change the time of using UniformInitializer, ConstantInitializer, NormalInitializer, TruncatedNormalInitializer, XavierInitializer, MSRAInitializer, NumpyArrayInitializer as few as possible
      
      * fix argument incampatible
      
      * fix more arg incompatible
      
      * fix test_prelu_op_xpu.py Constant
      
      * fix inaccurate doc
      
      * more doc fix: default value
      6edc7bba
  2. 25 11月, 2022 1 次提交
  3. 23 10月, 2022 1 次提交
  4. 29 9月, 2022 1 次提交
  5. 27 9月, 2022 1 次提交
  6. 18 8月, 2022 1 次提交
  7. 05 6月, 2022 1 次提交
    • S
      【code format check upgrade】 step2:yapf (#42944) · a072fca8
      Sing_chan 提交于
      * use yapf to format all python file
      
      * yapf exclude two unittests file for they rely on writing and reading file, and format will break them
      
      * disable diff_py_file because too many diff files cause command following failed
      a072fca8
  8. 09 6月, 2021 1 次提交
  9. 31 12月, 2020 1 次提交
  10. 22 9月, 2020 1 次提交
    • P
      Use dygraph mode by default (#27443) · 827ac36f
      pangyoki 提交于
      * default open dygraph mode
      
      * fix CI-Mac
      
      * fix Mac-CI other unittest file
      
      * fix CI-Py3
      
      * fix test_communicator_geo and test_buffer_shared_memory_reuse_pass
      
      * add enable_static to fix CI-Py3
      
      * add enable_static to fix CI-coverage
      
      * delete try except
      827ac36f
  11. 27 8月, 2020 1 次提交
  12. 03 12月, 2019 1 次提交
  13. 27 6月, 2019 1 次提交
    • H
      supports collective communicated training (#18175) · b7128bac
      HaoRen 提交于
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * fix comment
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * fix comment
      test=develop
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * test=develop
      add collective op unittest standard
      
      * test=develop
      remove the test_collective directory
      
      * test=develop
      remove the test_collective directory
      
      * remove slicegather test
      
      * code format for reducescatter
      
      * update attr of shard_index_op
      
      * Modify macro nccl_helper
      
      * remove test without distribute
      
      * macro collective_helper
      
      * marcro update
      
      * test=develop
      update support python3.5
      
      * test=develop change gpu memory use to 0.1 when test
      
      * test=develop
      update ut equal func
      
      * test=develop
      set flags to 1.5
      
      * test=develop fix pickle dumple  py35
      
      * test=develop
      fix divide in slice and add sync_comm_stream
      update atol and rtol to 1e-05
      rm shard_index op and test
      modify read input from file to read from memory
      remove origin_program in framework and add i/o in c_sync_calc_stream
      
      * test=develop update unittest sync operator I/O
      b7128bac