- 22 11月, 2022 1 次提交
-
-
由 wangzhen38 提交于
* [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz * update * [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz * [remove fluid] under fleet meta_optimizers_wz
-
- 03 11月, 2022 1 次提交
-
-
由 Nyakku Shigure 提交于
* [CodeStyle][py2][U008] remove unnecessary args in `super()` * remove remained args * revert changes in test_pylayer_op * Revert "revert changes in test_pylayer_op" This reverts commit ff185a9ae738afac3b0264f61bde6c6b7f72e7c4. * revert some changes in example code
-
- 01 11月, 2022 2 次提交
-
-
由 Nyakku Shigure 提交于
* [CodeStyle][E711] use `is`/`is not` for comparison with `None` * `self.assertTrue($A is None)` -> `self.assertIsNone($A)` * `self.assertTrue($A is not None)` -> `self.assertIsNotNone($A)` * `self.assertFalse($A is None)` -> `self.assertIsNotNone($A)` * `self.assertEqual($A, None)` -> `self.assertIsNone($A)` * `self.assertNotEqual($A, None)` -> `self.assertIsNotNone($A)`
-
由 Nyakku Shigure 提交于
* [CodeStyle][E712] use `if cond`/`if cond is True` for comparison with `True` * revert changes in fluid * revert unrelated file * revert changes in norm * revert changes in auto_parallel_amp * fix norm and auto_parallel_amp * revert a typo fix due to fixed at #47477
-
- 23 10月, 2022 1 次提交
-
-
由 Nyakku Shigure 提交于
* update config * re-blacken python code * temporarily disable date and diff_py_file * skip a format
-
- 09 6月, 2022 1 次提交
-
-
由 sneaxiy 提交于
* add nproc_per_node for DistributedFusedLamb * fix nproc_per_node communicator bug * fix ring_id = 1 init bug * fix ci * fix test_parallel_executor_mnist.py
-
- 05 6月, 2022 1 次提交
-
-
由 Sing_chan 提交于
* use yapf to format all python file * yapf exclude two unittests file for they rely on writing and reading file, and format will break them * disable diff_py_file because too many diff files cause command following failed
-
- 08 9月, 2021 1 次提交
-
-
由 Zeng Jinle 提交于
* add fleet api for program pass * turn on apply pass for CI test * fix disable fuse_all_optimizer bug * try to test ci * fix CI * fill unspecified op role * fix fuse_allreduce * add ut to improve coverage * remove useless change * improve c++ coverage * follow some comments * test ir pass pipeline * update doc * reduce ut time again
-
- 01 7月, 2021 1 次提交
-
-
由 Yuang Liu 提交于
-
- 13 5月, 2021 1 次提交
-
-
由 WangXi 提交于
-
- 06 5月, 2021 1 次提交
-
-
由 zhiboniu 提交于
-
- 07 4月, 2021 1 次提交
-
-
由 zhang wenhui 提交于
* Ascend rc (#30483) * Fix compilcation on CANN20.1 and older (#30494) Fix compilcation on CANN20.1 and older * Add distribution supported (#30578) Add distribution supported * Build praser for Hcom* operators (#30627) Build praser for Hcom* operators * Pass device_ids info from launch to trainer. (#30632) Pass device_ids info from launch to trainer * Add Hccl program group (#30642) Add Hccl program group * Add startup bash files of test_ascend_group. (#30645) Add startup bash files of test_ascend_group * cleanup (#30646) cleanup test_ascend_group.py * [Feature] Build parser to support distributed training (#30658) [Feature] Build parser to support distributed training * fix compilation on ascend-20.1 (#30722) fix compilation on ascend-20.1 * Dev/fix ascend string (#30749) Dev/fix ascend string * code style (#30781) code style * Merge ascend_optimizer and ascend_parser. (#30776) Merge ascend_optimizer and ascend_parser. * Ascendrc add converted op : [range/equal/range/uniform_random/expand/squeeze], fix cast op bug (#30797) Ascendrc add converted op : [range/equal/range/uniform_random/expand/squeeze], fix cast op bug * Add paddle ascend distribution training supported (#30796) Add paddle ascend distribution training supported * pass cxx_flags to gloo cmake (#30857) * Destroy session first. (#30954) Destroy session first. * merge * fix, test=develop * fix, test=develop * fix style, test=develop * fix, test=develop * fix * fix log fatal, test=develop * fix enforce style, test=develop * fix, test=develop * fix, test=develop * fix rccl, test=develop * fix test, test=develop * fix, test=develop * fix, test=develop * fix, test=develop * fix node_num, test=develop * fix ids str, test=develop * fix ids str, test=develop * fix ids str, test=develop * fix, test=develop * fix, test=develop * fix, test=develop * fix, test=develop * fix, test=develop * fix, test=develop * fix, test=develop * fix, test=develop * fix style code, test=develop * fix style code, test=develop * fix style code, test=develop * fix style code, test=develop Co-authored-by: Nhutuxian <hutuxian2011@sina.cn> Co-authored-by: Ngongweibao <weibao.gong@gmail.com> Co-authored-by: NVoid Main <voidmain1313113@gmail.com> Co-authored-by: NLeo Chen <chenqiuliang@baidu.com> Co-authored-by: Ndingsiyu <18369187719@163.com> Co-authored-by: NOleNet <olenet@126.com>
-
- 05 2月, 2021 1 次提交
-
-
由 liuyuhui 提交于
-
- 01 2月, 2021 1 次提交
-
-
由 WangXi 提交于
-
- 17 12月, 2020 1 次提交
-
-
由 WangXi 提交于
-
- 26 11月, 2020 1 次提交
-
-
由 WangXi 提交于
-
- 20 9月, 2020 1 次提交
-
-
由 tangwei12 提交于
* fix fleet util and gloo * fix worker endpoints * fix * fix UT * fix gloo * fix gloo * update gloo * update gloo * update gloo * update gloo * update gloo * fix gloo wrapper for hdfs * add file gloo and UT * fix UT * fix UT * fix UT * hide public method of RoleMaker * fix UT * GPU fleetrun support gloo * parameterserver fleetrun support gloo * add UT * add UT * fix UT * fix get server endpoint * fix get server endpoint * fix UT * hide public method of rolemaker * hide public method of rolemaker * hide public method of rolemaker * Update test_fleet_rolemaker_new.py * hide public method of rolemaker * hide public method of rolemaker
-
- 10 9月, 2020 1 次提交
-
-
由 123malin 提交于
* parameter_server_optimizer support auto_strategy
-
- 07 9月, 2020 1 次提交
-
-
由 Dong Daxiang 提交于
* add auto parallel L1 implementation test=develop
-
- 21 8月, 2020 1 次提交
-
-
由 Dong Daxiang 提交于
* consider the combination of different strategies to work together
-
- 18 8月, 2020 1 次提交
-
-
由 mapingshuo 提交于
* add feature to fleet2.0 role_maker, distribute_strategy, test=develop
-
- 13 8月, 2020 1 次提交
-
-
由 Dong Daxiang 提交于
* move paddle.fleet to paddle.distributed.fleet
-
- 11 8月, 2020 1 次提交
-
-
由 pangyoki 提交于
* Directory migration, test=develop * Change imperative from paddle init to paddle framework, test=develop * Fixed jit bug, test=develop * default static mode, test=develop * fixed format and create parameter belongs to framework, test=develop * Fixed import package, test=develop * fix __init__ format, test=develop * fixed alias problem * fixed paddle.enable_imperative problems, test=develop * Add unittest * delete install_check comment * Fixed unittest timeout * fixed unittest error * move Program default_xx_program to static package * optimize unittest method * fixed framework __init__ format * fixed jit path * delete alias * move jit to paddle * Fixed unittest format * fixed paddle.default_main_program * Fixed save load API in paddle __init__.py * fixed ci paddle.imperative.to_variable
-
- 03 8月, 2020 1 次提交
-
-
由 Dong Daxiang 提交于
* split meta optimizer files * add graph execution in execution, update two properties in DistributedStrategy, unit tests for these features
-
- 30 7月, 2020 1 次提交
-
-
由 tangwei12 提交于
Integrated Trainer of Parameter Server (API add `fluid.contrib.layers.sparse_embedding` only) (#22957) * Integrated Trainer of Parameter Server
-
- 29 7月, 2020 1 次提交
-
-
由 Dong Daxiang 提交于
* refine strategy compiler and meta optimizers make async as a_sync
-
- 28 7月, 2020 1 次提交
-
-
由 Dong Daxiang 提交于
* add more settings for distributed strategy Basically, DistributedStrategy has several parts of configurations: - BuildStrategy: the same as paddle.fluid.BuildStrategy, but the distributed arguments are moved out of BuildStrategy - ExecutionStrategy: the same as paddle.fluid.ExecutionStrategy - collective communication configs: nccl_comm_num, hierarchical allreduce and so on - distributed algorithms: async_update(mainly used in PS), lars, lamb and so on
-
- 23 7月, 2020 1 次提交
-
-
由 Dong Daxiang 提交于
* fix gen nccl id bug
-
- 20 7月, 2020 1 次提交
-
-
由 Dong Daxiang 提交于
refactor fleet api under paddle.fleet update DistributedStrategy
-