1. 10 11月, 2022 1 次提交
  2. 03 11月, 2022 1 次提交
  3. 26 10月, 2022 1 次提交
  4. 23 10月, 2022 1 次提交
  5. 10 10月, 2022 1 次提交
  6. 30 9月, 2022 1 次提交
    • S
      support pure bfloat16 for more ops (#46364) · b7b231a6
      sneaxiy 提交于
      * support pure bfloat16
      
      * support bf16 linear
      
      * update PR to pass CI
      
      * tiny fix where_grad_kernel.cu
      
      * add bfloat16 to selu_grad to pass CI
      
      * fix selu grad compilation error
      b7b231a6
  7. 23 9月, 2022 1 次提交
  8. 14 9月, 2022 1 次提交
  9. 29 8月, 2022 1 次提交
    • H
      [phi] Transfer merged_adam yaml to phi (#45367) · b4f74eed
      HongyuJia 提交于
      * add legacy_api.yaml
      
      * set merged_momentum inplace only
      
      * support inplace optional<vector<tensor>>
      
      * add dygraph_mode api
      
      * add merged_adam yaml
      
      * add merged_adam python api
      
      * change testcase of merged_adam and adam
      
      * fix import of test_merged_adam_op
      b4f74eed
  10. 26 8月, 2022 1 次提交
  11. 29 6月, 2022 1 次提交
  12. 05 6月, 2022 1 次提交
    • S
      【code format check upgrade】 step2:yapf (#42944) · a072fca8
      Sing_chan 提交于
      * use yapf to format all python file
      
      * yapf exclude two unittests file for they rely on writing and reading file, and format will break them
      
      * disable diff_py_file because too many diff files cause command following failed
      a072fca8
  13. 13 4月, 2022 1 次提交
  14. 31 3月, 2022 1 次提交
  15. 25 3月, 2022 1 次提交
    • J
      Refactor Dygraph Flags (#40786) · 3085d5e4
      Jiabin Yang 提交于
      * refactor eager flags
      
      * fix flags error when we switch from eager to dygraph
      
      * fix ci problem
      
      * fix ci
      
      * fix ci
      
      * merge develop and fix code style
      
      * merge develop and fix code style
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * merge develop
      3085d5e4
  16. 20 1月, 2022 1 次提交
  17. 07 1月, 2022 1 次提交
  18. 17 9月, 2021 1 次提交
    • Z
      [AMP] Support pure fp16 training mode for dygraph (#35521) · adaeee4d
      zhangbo9674 提交于
      * add pure fp16 major function in auto_cast & tracer
      
      * support master weight in dygraph for pure fp16
      
      * check mix dtype of fp16&fp32 for check_finite_and_unscale op
      
      * change pure fp16 funtion name
      
      * refine some bug in auto_cast
      
      * refine auto_cast interface logic
      
      * add param _casted_by_pure_fp16 for class Layer
      
      * support state_dict hook for save model by user appointed dtype in pure_fp16_decorator
      
      * refine pure_fp16_decorator as decorator
      
      * add unittest
      
      * add comment
      
      * add comment
      
      * support recompute
      
      * add comment for auto_cast and decorator
      
      * support to_static_state_dict for paddle.jit.save
      
      * unlimite models num and optimizers num
      
      * add lookup_table in black_list
      
      * fix momentum and layer state_dict
      
      * fix bug in layer state_dict
      
      * fix bug in layer state_dict_helper
      
      * refine unittest
      
      * refine test_momentun_op
      
      * refine interface and some code
      
      * refine amp_decorator interface
      
      * refine pure fp16 interface
      
      * refine master weight interface
      adaeee4d
  19. 02 8月, 2021 1 次提交
  20. 15 7月, 2021 1 次提交
  21. 31 5月, 2021 1 次提交
  22. 29 4月, 2021 1 次提交
  23. 28 4月, 2021 1 次提交
  24. 27 4月, 2021 1 次提交
  25. 23 4月, 2021 1 次提交
  26. 24 3月, 2021 1 次提交
  27. 01 2月, 2021 1 次提交
  28. 08 1月, 2021 1 次提交
    • Z
      Support pure fp16 training for AMP API. (#29544) · 7f7dfccf
      Zhen Wang 提交于
      * add cast ops before and after unsupported fp16 ops.
      
      * Keep partial net in FP32 pattern.
      
      * Support check_finite_and_unscale and update_loss_scaling for FP16 calculation mode.
      
      * Add fp16 support for adam op.
      
      * add multi precision attr for adam.
      
      * Fix the bug of test_multi_precision_fp16_train UT.
      
      * Code format for CI.
      
      * Fix the redefine error about MPTypeTrait on windows.
      
      * fix bugs of the _create_accumulators func in Momentum.
      
      * fix bug when inserting post cast op.
      
      * Add the update_loss_scaling op in allow_set of UnusedVarCheck.
      
      * Update for ci coverage.
      
      * Add some doc for OptimizerWithMixedPrecision.
      
      * Fix the code style.
      
      * Imporve the doc of `amp_init`.
      
      * Change for fp16 testing if users have the infer program defined in separate way.
      7f7dfccf
  29. 07 1月, 2021 1 次提交
  30. 28 12月, 2020 1 次提交
  31. 27 11月, 2020 1 次提交
    • S
      Support dynamic graph distributed (#28997) · e2d01eb6
      ShenLiang 提交于
      * add reducer
      
      * refine envent for memorycopy
      
      * add concat&split for allreduce
      
      * apply concat & split for fuse tensor
      
      * fix nccl dep
      
      * fix the untest, compile problem and ddp initialize problem
      
      * fix untest for mac & add some comments & solve the repeated param in sublayers
      
      * fix untest for windows & fix document
      e2d01eb6
  32. 24 11月, 2020 1 次提交
    • L
      Upgrade string literals to raw string (#28989) · 3815d7aa
      Leo Chen 提交于
      * upgrade comment string to raw string
      
      * fix string in
      
      * fix string with ' '
      
      * revert update on comments
      
      * upgrade only necessary
      
      * fix sample code checker
      
      * fix comments with '''
      3815d7aa
  33. 13 10月, 2020 1 次提交
  34. 12 10月, 2020 1 次提交
    • M
      refine adam/strided_slice && fix doc for rmsprop/unstack (#27740) · 84d8e49d
      MRXLT 提交于
      * refine parameters order && doc
      
      * update rmsprop doc
      
      * refine adam/transpose/unstack/stride_slice
      
      * fix bug && doc
      
      * fix doc
      
      * bug fix
      
      * bug fix
      
      * fix doc
      
      * fix doc
      
      * fix doc
      
      * fix doc
      
      * depercate old strided_slice
      
      * update doc
      
      * set default value for name
      
      * update doc
      84d8e49d
  35. 29 9月, 2020 1 次提交
  36. 21 9月, 2020 1 次提交
  37. 14 9月, 2020 1 次提交
  38. 01 9月, 2020 1 次提交
    • M
      update optimizer (#26711) · 1f36d3cd
      MRXLT 提交于
      * update doc
      
      * update doc
      
      * fix optimizer sample code
      
      * add default value for adamw weight_decay
      
      * fix adamw
      
      * change LearningRateDecay to _LRScheduler
      
      * fix adamw;notest
      
      * fix load;notest
      
      * remove file
      
      * bug fix
      
      * fix code style
      
      * bug fix
      
      * add ut
      
      * adamw support weight_decay=0
      
      * fix ut
      
      * fix set_lr doc
      
      * fix doc
      
      * change parameters place
      1f36d3cd
  39. 23 8月, 2020 1 次提交
    • M
      [WIP] update optimizer for 2.0 (#26288) · eeda90d6
      MRXLT 提交于
      refine Optimizer/Adam/Admax/RMSProp && add Admw
      
      * buf fix
      
      * update comment
      
      * unify arguments place; notest
      
      * fix ut, test=develop
      
      * bug fix
      
      * fix conflicts, test=develop
      
      * add examples code
      
      * bug fix
      
      * fix comments
      
      * fix sample code
      
      * add sample code for Optimizer
      
      * add adamax ut, test=develop
      
      * fix rmsprop ut, test=develop
      
      * add ut for optimizer.py and adamw.py
      
      * remove TestAdamOptimizerBetaVariable
      
      * update api && add ut
      
      * update doc && fix ut
      
      * add ut
      Co-authored-by: Nmapingshuo <mps2012@yeah.net>
      eeda90d6