1. 20 1月, 2021 1 次提交
    • C
      make abs op support complex types (#30375) · 358106fc
      chentianyu03 提交于
      * rewrite abs op
      
      * rewrite abs op and remove abs in activation
      
      * remove abs register in old codes
      
      * fix abs_grad type error
      
      * fix abs double_grad output name error
      
      * modify abs_grad, abs_grad_grad functor for windows building
      
      * format code style
      
      * fix the bug of result is nan when the divisor is zero
      
      * add missing abs attr and add abs for float16
      358106fc
  2. 19 1月, 2021 3 次提交
  3. 18 1月, 2021 4 次提交
  4. 17 1月, 2021 1 次提交
  5. 15 1月, 2021 5 次提交
  6. 14 1月, 2021 2 次提交
  7. 13 1月, 2021 1 次提交
  8. 12 1月, 2021 8 次提交
  9. 11 1月, 2021 9 次提交
  10. 10 1月, 2021 2 次提交
  11. 09 1月, 2021 2 次提交
  12. 08 1月, 2021 2 次提交
    • Z
      Support pure fp16 training for AMP API. (#29544) · 7f7dfccf
      Zhen Wang 提交于
      * add cast ops before and after unsupported fp16 ops.
      
      * Keep partial net in FP32 pattern.
      
      * Support check_finite_and_unscale and update_loss_scaling for FP16 calculation mode.
      
      * Add fp16 support for adam op.
      
      * add multi precision attr for adam.
      
      * Fix the bug of test_multi_precision_fp16_train UT.
      
      * Code format for CI.
      
      * Fix the redefine error about MPTypeTrait on windows.
      
      * fix bugs of the _create_accumulators func in Momentum.
      
      * fix bug when inserting post cast op.
      
      * Add the update_loss_scaling op in allow_set of UnusedVarCheck.
      
      * Update for ci coverage.
      
      * Add some doc for OptimizerWithMixedPrecision.
      
      * Fix the code style.
      
      * Imporve the doc of `amp_init`.
      
      * Change for fp16 testing if users have the infer program defined in separate way.
      7f7dfccf
    • L
      use cuda generator in bernoulli cuda kernel (#30199) · 789743e1
      Leo Chen 提交于
      789743e1