1. 14 4月, 2023 1 次提交
    • F
      1. modify set_value op, use Scalars to represent attr `values`, instead of a... · dd2a749a
      Feiyu Chan 提交于
      1. modify set_value op, use Scalars to represent attr `values`, instead of a bunch of attributs of various types; (#52408)
      
      2. add program converter and set_value op as an example, which provides the functionality to convert `paddle::framework::ProgramDesc` between old and new formats(the differences are mainly some operators with incompatible updates in the definition);
      3. program version and operator version map now are always saved when serializing `paddle::framework::ProgramDesc` to identify the version;
      3. provide an option `legacy_format=false` in  serialization of `paddle::framework::ProgramDesc`, it decided whether to convert ProgramDesc back to a legacy format, which is compatible for paddle 2.4.2 or earlier versions to load and execute;
      4. deserialization of `paddle::framework::ProgramDesc` is now automatically detecting whether the bytes it receives is in legacy format(contains any of the operators that has been incompatibly updated and have any attribute of type `Scalar`) and convert it to new format. But if you want a faithful deserialization without the automatic conversion, you can use protobuf's deserialization instead. Though it is not recommended, it can be used for the purpose of testing.
      dd2a749a
  2. 31 3月, 2023 1 次提交
  3. 25 3月, 2023 1 次提交
  4. 22 3月, 2023 1 次提交
  5. 08 3月, 2023 1 次提交
  6. 01 3月, 2023 1 次提交
  7. 22 2月, 2023 1 次提交
  8. 12 1月, 2023 1 次提交
  9. 16 11月, 2022 1 次提交
    • W
      [remove fluid] under fleet meta_optimizers (#47864) · a2a97cbb
      wangzhen38 提交于
      * [remove fluid] under fleet meta_optimizers
      
      * [remove fluid] under fleet meta_optimizers
      
      * [remove fluid] under fleet meta_optimizers
      
      * [remove fluid] under fleet meta_optimizers
      
      * [remove fluid] under fleet meta_optimizers
      
      * [remove fluid] under fleet meta_optimizers
      
      * [remove fluid] under fleet meta_optimizers
      
      * [remove fluid] under fleet meta_optimizers
      
      * [remove fluid] under fleet meta_optimizers
      
      * [remove fluid] under fleet meta_optimizers
      
      * [remove fluid] under fleet meta_optimizers
      
      * [remove fluid] under fleet meta_optimizers
      a2a97cbb
  10. 08 11月, 2022 1 次提交
  11. 03 11月, 2022 1 次提交
  12. 01 11月, 2022 1 次提交
    • N
      [CodeStyle][E711] use `is`/`is not` for comparison with `None` (#47452) · a35a4a53
      Nyakku Shigure 提交于
      * [CodeStyle][E711] use `is`/`is not` for comparison with `None`
      
      * `self.assertTrue($A is None)` -> `self.assertIsNone($A)`
      
      * `self.assertTrue($A is not None)` -> `self.assertIsNotNone($A)`
      
      * `self.assertFalse($A is None)` -> `self.assertIsNotNone($A)`
      
      * `self.assertEqual($A, None)` -> `self.assertIsNone($A)`
      
      * `self.assertNotEqual($A, None)` -> `self.assertIsNotNone($A)`
      a35a4a53
  13. 23 10月, 2022 1 次提交
  14. 19 10月, 2022 1 次提交
  15. 05 6月, 2022 1 次提交
    • S
      【code format check upgrade】 step2:yapf (#42944) · a072fca8
      Sing_chan 提交于
      * use yapf to format all python file
      
      * yapf exclude two unittests file for they rely on writing and reading file, and format will break them
      
      * disable diff_py_file because too many diff files cause command following failed
      a072fca8
  16. 19 10月, 2020 1 次提交
  17. 10 8月, 2020 1 次提交
  18. 08 8月, 2020 1 次提交
  19. 07 7月, 2020 1 次提交
  20. 15 4月, 2020 1 次提交
  21. 03 4月, 2020 1 次提交
  22. 23 2月, 2020 1 次提交
  23. 31 12月, 2019 1 次提交
  24. 05 12月, 2019 1 次提交
  25. 12 11月, 2019 1 次提交
  26. 15 10月, 2019 2 次提交
  27. 23 9月, 2019 1 次提交
    • M
      Forward recompute3 (#19913) · 9901f696
      mapingshuo 提交于
      * add recompute based checkpoints methods for large batch training
      test=develop
      
      * add append_backward_with_forward_recomputation
      test=develop
      
      * refine optimizer
      test=develop
      
      * update backward and optimizer
      test=develop
      
      * make Variable usable
      test=develop
      
      * add recompute code
      
      * refine optimizer
      test=develop
      
      * refine addup _append_backward_ops_with_checkpoints_
      1) for recompute part, just cache the grad_op_desc without appending to block
      2) before appending grad_op_desc to backward part, addup_repetitive_vars, remove unused branch
      test=develop
      
      * make method private
      
      * add recompute strategy into DistributedStrategy
      test=develop
      
      * checkpoint version3
      test=develop
      
      * remove some print information
      test=develop
      
      * remove unused sumop
      test=develop
      
      * try to fix recompute with graph building modules
      
      * add input names to vars should be held
      
      * add memory debug tool
      
      * backup backward
      
      * Fix bugs
      
      * add backward desc for op not in any segments
      
      * add exception info for sub_block
      
      test=develop
      
      * modify code style
      
      test=develop
      
      * modify code style
      
      test=develop
      
      * remove print functions
      
      test=develop
      
      * add API spec
      
      test=develop
      test=document_preview
      
      * make Recompute a child class of Optimizer
      
      test=develop
      test=document_preview
      
      * add API spec
      
      test=develop
      test=document_preview
      
      * modify API spec
      
      test=develop
      test=document_preview
      
      * add document for Recompute
      
      test=develop
      test=document_preview
      
      * change API doc of Rcompute
      
      test=develop
      test=document_preview
      
      * code cleaning
      
      test=develop
      test=document_preview
      
      * modify API spec
      
      * fix bugs when segments hold no element
      
      * add testcase for Recompute Optimizer
      
      test=develop
      test=document_preview
      
      * add test for apply_gradient, and code cleaning
      
      test=develop
      test=document_preview
      
      * add test case for load function
      
      * enable CI
      
      test=develop
      test=document
      
      * add test case
      
      test=develop
      test=document_preview
      
      * add sample code for 4 function of recompute optimizer
      
      test=develop
      test=document_preview
      9901f696
  28. 19 9月, 2019 1 次提交
  29. 10 9月, 2019 1 次提交
  30. 28 8月, 2019 1 次提交
  31. 16 8月, 2019 1 次提交
  32. 12 8月, 2019 1 次提交
  33. 10 7月, 2019 1 次提交
  34. 27 6月, 2019 1 次提交
    • H
      supports collective communicated training (#18175) · b7128bac
      HaoRen 提交于
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * fix comment
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * fix comment
      test=develop
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * test=develop
      add collective op unittest standard
      
      * test=develop
      remove the test_collective directory
      
      * test=develop
      remove the test_collective directory
      
      * remove slicegather test
      
      * code format for reducescatter
      
      * update attr of shard_index_op
      
      * Modify macro nccl_helper
      
      * remove test without distribute
      
      * macro collective_helper
      
      * marcro update
      
      * test=develop
      update support python3.5
      
      * test=develop change gpu memory use to 0.1 when test
      
      * test=develop
      update ut equal func
      
      * test=develop
      set flags to 1.5
      
      * test=develop fix pickle dumple  py35
      
      * test=develop
      fix divide in slice and add sync_comm_stream
      update atol and rtol to 1e-05
      rm shard_index op and test
      modify read input from file to read from memory
      remove origin_program in framework and add i/o in c_sync_calc_stream
      
      * test=develop update unittest sync operator I/O
      b7128bac
  35. 12 6月, 2019 1 次提交
  36. 23 5月, 2019 1 次提交
  37. 09 5月, 2019 1 次提交
  38. 25 4月, 2019 1 次提交