1. 25 3月, 2022 1 次提交
    • J
      Refactor Dygraph Flags (#40786) · 3085d5e4
      Jiabin Yang 提交于
      * refactor eager flags
      
      * fix flags error when we switch from eager to dygraph
      
      * fix ci problem
      
      * fix ci
      
      * fix ci
      
      * merge develop and fix code style
      
      * merge develop and fix code style
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * merge develop
      3085d5e4
  2. 24 3月, 2022 1 次提交
    • Z
      [AMP] Support amp for Intermediate_dygraph (#40623) · c12f7d48
      zhangbo9674 提交于
      * approve amp for intermediate_dygraph
      
      * add amp_utils for intermediate_dygraph
      
      * add amp needcast check for mlu & npu
      
      * test unittest
      
      * add SetGradNode for set_stop_gradient && add checktensor for GradientHooks
      
      * refine code
      
      * refien unittest of imperative_amp for new dygraph
      
      * inplace api skip amp
      
      * add test_imperative_qat_amp for intermediate amp
      
      * refine code
      
      * refine test_amp ci strategy
      
      * refine unittest code
      
      * refine amp_utils code
      
      * refine amp getpromotetype for some special op
      
      * refine unittest code
      c12f7d48
  3. 16 3月, 2022 3 次提交
  4. 15 3月, 2022 1 次提交
  5. 11 3月, 2022 1 次提交
  6. 04 3月, 2022 1 次提交
  7. 03 3月, 2022 2 次提交
    • B
      change_ASP_sharding_option (#40028) · 815f7a67
      Baibaifan 提交于
      815f7a67
    • J
      Support slim eager (#39874) · da47544c
      Jiabin Yang 提交于
      * eager, test=develop
      
      * fix bug, test=develop
      
      * eager, test=develop
      
      * merge legacy to fluid
      
      * eager, test=develop
      
      * eager, test=develop
      
      * Refactor TensorAdd func by template and remove gradient_accumulation in eager
      
      * Remove needless target name
      
      * eager, test=develop
      
      * eager, test=develop
      
      * Use overload instead of template
      
      * Remove legacy code
      
      * Remove legacy code
      
      * selectedrows, test=develop
      
      * Remove DataType test
      
      * eager, test=develop
      
      * eager, test=develop
      
      * support gan, test=develop
      
      * Using Tensor directly instead of using EagerTensor
      
      * support gradient_accumulation
      
      * make test_imperative_lod_tensor_to_selected_rows longer
      
      * make test_imperative_lod_tensor_to_selected_rows longer
      
      * refine code
      
      * ptb, test=develop
      
      * Rename all EagerTensor to Tensor
      
      * Rename some EagerTensor to Tensor
      
      * rename EagerTensor to EagerVariable
      
      * eager, test=develop
      
      * eager, test=develop
      
      * eager, test=develop
      
      * eager, test=develop
      
      * add more test
      
      * eager, test=develop
      
      * Support copiable selected rows and merge develop
      
      * save load, eager, test=develop
      
      * save load, eager, test=develop
      
      * refine, test=develop
      
      * remove useless _set_value method
      
      * refine, test=develop
      
      * refine, test=develop
      
      * revert static_runner, test=develop
      
      * EagerTensor to Tensor, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * clear grad, test=develop
      
      * merge, develop
      
      * merge, develop
      
      * merge, test=develop
      
      * merge, test=develop
      
      * Support quant and part of slice
      
      * support legacy static save
      
      * extend slim tests time
      
      * remove imperative on inference
      
      * remove imperative on inference
      
      * merge develop
      
      * fix typo
      
      * fix typo
      
      * split slice related code into 2 part for imperative and eager
      
      * split slice from inference
      
      * split slice from inference
      
      * fix test_tensor_register_hook
      Co-authored-by: NWang Huan <wanghuan29@baidu.com>
      Co-authored-by: NWeilong Wu <veyron_wu@163.com>
      Co-authored-by: Nwanghuancoder <wanghuancoder@163.com>
      da47544c
  8. 01 3月, 2022 2 次提交
  9. 19 2月, 2022 1 次提交
    • S
      Add the DistributedFusedLamb optimizer (#39148) · 5df3cd61
      sneaxiy 提交于
      * add DistributedFusedLamb op
      
      * polish code
      
      * fix compile error
      
      * compatible with pten changement
      
      * fix rocm compile error
      
      * improve converage
      
      * update upstream/develop
      
      * fix cast_with_ptr.h
      
      * add FLAGS_distributed_lamb_divide_nranks_when_allreduce=1
      
      * fix clip before allreduce
      
      * add use_master_param_norm
      
      * code polish
      
      * fix bug
      
      * fix ROCM ci
      5df3cd61
  10. 14 2月, 2022 1 次提交
  11. 09 2月, 2022 1 次提交
  12. 07 2月, 2022 1 次提交
  13. 27 1月, 2022 1 次提交
  14. 21 1月, 2022 1 次提交
  15. 13 1月, 2022 1 次提交
    • J
      Added mul BF16/FP32 FWD/BWD oneDNN kernel (#38552) · fc6eed5b
      jakpiase 提交于
      * base changes for mul reimplementation
      
      * empty commit
      
      * tmp save
      
      * full implementation of mul bf16/fp32 fwd bwd
      
      * CI fix
      
      * CI rerun
      
      * changed unity build cmake to avoid gpu issues
      
      * removed mul mkldnn from unity build
      
      * added skipping tests if not cpu_bf16
      
      * CI fix
      
      * CI fix
      
      * CI fix
      fc6eed5b
  16. 12 1月, 2022 1 次提交
  17. 06 1月, 2022 1 次提交
  18. 05 1月, 2022 2 次提交
  19. 28 12月, 2021 1 次提交
  20. 22 12月, 2021 1 次提交
  21. 20 12月, 2021 1 次提交
    • S
      Support FP16 for more ops (#38123) · 1f445bf3
      sneaxiy 提交于
      * support FP16 for more ops
      
      * add amp list tests
      
      * refine reduce_mean_grad
      
      * fix OP benchmark ci
      
      * fix fp16 reduce_mean
      
      * updat ut, but still have some problems
      
      * remove mean/reduce_mean fp16 kernel
      1f445bf3
  22. 17 12月, 2021 1 次提交
    • S
      Refine some AMP operators for BERT (#37923) · d80fe268
      sneaxiy 提交于
      * support multi precision update for LAMB
      
      * hide some api
      
      * fix ci uts
      
      * fix lamb output of dygraph
      
      * remove some changes to some PR
      
      * try to fix Py3 CI compile error
      
      * fix test_imperative_optimizer, add lars ut, add layer_norm ut
      
      * fix ut, fix format
      
      * fix ut
      
      * fix windows ci
      d80fe268
  23. 14 12月, 2021 3 次提交
  24. 13 12月, 2021 1 次提交
  25. 10 12月, 2021 2 次提交
  26. 07 12月, 2021 1 次提交
  27. 01 12月, 2021 2 次提交
  28. 30 11月, 2021 1 次提交
  29. 26 11月, 2021 1 次提交
  30. 04 11月, 2021 1 次提交
  31. 29 10月, 2021 1 次提交