1. 06 9月, 2023 1 次提交
  2. 02 8月, 2023 1 次提交
  3. 11 7月, 2023 1 次提交
    • L
      replace the AdagradOptimizer... · 94365855
      LoneRanger 提交于
      replace the AdagradOptimizer 、adamaxOptimizer、AdadeltaOptimizer、RMSPropOptimizer、LambOptimizer and Momentum (#54152)
      
      * replace the AdadeltaOptimizer with Adadelta
      
      * replace the RMSPropOptimizer with RMSProp
      
      * replace the LambOptimizer with lamb
      
      * replace the momentum in contrib/optimizer.py with Momentum in python/paddle/optimizer/momentum.py
      
      * fix bug
      
      * fix bug
      
      * fix bug
      
      * fix bug of Lamp
      
      * fix bug of Lamp
      
      * fix bug of import
      
      * replace the AdamaxOptimizer with Admax and change the optimizer base for AdagradOptimizer
      
      * fix bug
      
      * fix bug
      
      * Update optimizer.py
      
      * fix bug
      
      * fix bug
      94365855
  4. 23 5月, 2023 1 次提交
  5. 22 5月, 2023 1 次提交
  6. 18 5月, 2023 1 次提交
    • shaojie_wang's avatar
      [AMP]Master grad in static graph (#53362) · 972581d8
      shaojie_wang 提交于
      * add master gradients on static graph
      
      * add unit test for bf16 master grad static graph
      
      * use float16 as v100 test dtype
      
      * only skip GPU which do not support bf16
      
      * use linear layer to test master grad
      
      * 1.push master grad creation before all optimizer ops; 2.remove useless unittest; 3.use a function to create master grad states
      972581d8
  7. 16 5月, 2023 2 次提交
  8. 27 4月, 2023 1 次提交