1. 13 5月, 2019 4 次提交
  2. 12 5月, 2019 1 次提交
  3. 10 5月, 2019 4 次提交
  4. 08 5月, 2019 1 次提交
  5. 07 5月, 2019 5 次提交
    • Z
      Enhance inplace/mem-opt pass and enhance softmax_with_cross_entropy op inplace (#17225) · 4f859408
      Zeng Jinle 提交于
      * add use_cuda to inplace pass,test=develop
      
      * add test softmax_with_xe_inplace test,test=develop
      
      * fix potential inplace bug
      test=develop
      
      * add more skip vars in mem opt pass,test=develop
      
      * follow comment,test=develop
      
      * follow comments,move duplicate out arg check to program->graph,test=develop
      4f859408
    • B
      update sofmax with axis arg test=develop (#17190) · e782b54b
      baojun 提交于
      e782b54b
    • T
      remove unused FLAGS_warpctc_dir (#17162) · ff1661f1
      Tao Luo 提交于
      * remove unused FLAGS_warpctc_dir
      
      test=develop
      
      * remove FLAGS_warpctc_dir
      
      test=develop
      ff1661f1
    • K
      Softmax_cross_entropy op add axis (#16806) · a71d8fdb
      Kaipeng Deng 提交于
      * add attr axis infershape. test=develop
      
      * add CUDA kernel. test=develop
      
      * fix unittest. test=develop
      
      * fix unittest for soft_label. test=develop
      
      * fix fp16 unittest. test=develop
      
      * remove comment code. test=develop
      
      * refine test for axis. test=develop
      
      * add python api. test=develop
      
      * fix doc. test=develop
      
      * fix fp16 unittest. test=develop
      
      * fix ngraph test. test=develop
      
      * fix ENFORCE for test_imperative_transformer. test=develop
      
      * fit for ngraph test. test=develop
      
      * fix after rebase develop. test=develop
      
      * fix doc. test=develop
      
      * fix API.spec. test=develop
      
      * fix test_layers. test=develop
      
      * fix format. test=develop
      a71d8fdb
    • Z
      Quant output scale (#17215) · a914d9b1
      Zhen Wang 提交于
      * Add MovingAverageAbsMaxScale operator which is only used for calculating the quantization scale.
      
      * test=develop
      
      * change the output into inplace. test=develop
      
      * Revert "test=develop"
      
      This reverts commit 696cf62699ba1e1c98f61f7345ac7060010eb29a.
      
      * Revert "change the output into inplace. test=develop"
      
      This reverts commit a19acd20f07eee82622701a3015e6e9c073a5e0b.
      
      * test=develop.
      
      * update the MovingAverageAbsMaxScaleOp test. test=develop
      a914d9b1
  6. 06 5月, 2019 2 次提交
  7. 05 5月, 2019 3 次提交
  8. 01 5月, 2019 1 次提交
  9. 30 4月, 2019 3 次提交
  10. 29 4月, 2019 2 次提交
  11. 28 4月, 2019 1 次提交
    • Z
      Refine dropout gpu memory (#17095) · 28d69d71
      Zeng Jinle 提交于
      * refine_dropout_mem,test=develop
      
      * # This is a combination of 14 commits.
      # The first commit's message is:
      remove ut test_dist_word2vec in mac ci, will fix it in private, test=develop (#17066)
      
      # This is the 2nd commit message:
      
      Fleet unify distributed training (#16791)
      
      * implement distributed transpiler with fleet
      # This is the 3rd commit message:
      
      ParallelDyGraph with GPU collective mode (#16827)
      
      implement dygraph.parallel.DataParallel to hook reduce op.
      
      # This is the 4th commit message:
      
      Init mixed precision training interface (#16856)
      
      * Init mixed precision training interface
      
      * Add fp16 test script
      
      test=develop
      
      * All initializers support float16
      
      test=develop
      
      * Code cleanup & add more code annotations
      
      test=develop
      
      * Update API spec
      
      test=develop
      
      * Add usage example in doc
      
      test=develop
      
      # This is the 5th commit message:
      
      fix reference_count_pass,test=develop (#17060)
      
      test=develop
      # This is the 6th commit message:
      
      Speedup roi_perspective_transform op by caching the information of linear interpolation in forward (#17090)
      
      * Cache the information of linear interpolation in forward and use it in backward.
      test=develop
      
      * Fix cuda kernel.
      test=develop
      
      # This is the 7th commit message:
      
      remove unnecessary prepare_data (#17080)
      
      test=develop
      # This is the 8th commit message:
      
      fix interpolate cu. test=develop (#17101)
      
      # This is the 9th commit message:
      
      test=develop, double backward leaky_relu (#17067)
      
      backward of backward: leaky_relu
      # This is the 10th commit message:
      
      fix fuse optimizer ops (#17102)
      
      test=develop
      # This is the 11th commit message:
      
      truncated_gaussian_random supported in distributed training, test=develop (#17091)
      
      # This is the 12th commit message:
      
       Detailed coordinate description for yolov3 loss (#17007)
      
      * Detailed coordinate description for yolov3 loss
      
      test=develop
      
      * modified api.spec
      
      test=develop
      
      * modified loss name
      
      * fix api.spec
      
      test=develop
      
      * polish description
      
      test=develop
      
      * modified api.spec
      
      test=develop
      
      # This is the 13th commit message:
      
      fix test_weight_decay (#17109)
      
      test=develop
      # This is the 14th commit message:
      
      Path flag (#17105)
      
      * fix python/paddle/fluid/__init__.py detecting problems
      28d69d71
  12. 27 4月, 2019 1 次提交
  13. 26 4月, 2019 2 次提交
  14. 25 4月, 2019 5 次提交
  15. 24 4月, 2019 1 次提交
  16. 23 4月, 2019 3 次提交
  17. 22 4月, 2019 1 次提交