1. 11 4月, 2022 1 次提交
  2. 10 4月, 2022 1 次提交
  3. 09 4月, 2022 2 次提交
  4. 08 4月, 2022 1 次提交
  5. 07 4月, 2022 8 次提交
  6. 06 4月, 2022 4 次提交
  7. 05 4月, 2022 4 次提交
    • Z
      Fix bug of data transform in inference executor (#41349) · 91212104
      zyfncg 提交于
      * fix bug of data transform in inference executor
      
      * fix bug
      91212104
    • Z
      [DoubleGrad PR #8] Enabled triple grads for sigmoid and matmul (#41387) · d8a10977
      Zhanlue Yang 提交于
      * [Refactor] refactored eager_gen.py PR #2
      
      * [DoubleGrad PR #1] Decoupled code generation logics for Dygraph ForwardFunctions and GradNodes
      
      * Fixed minor issue
      
      * Adjusted logics of GenerateNodeCreationCodes and GenerateForwardDefinition
      
      * Fixed issues
      
      * Supported higher-order grad node generation
      
      * [DoubleGrad PR #4] Supported higher-order GradNode generation
      
      * [DoubleGrad #4] Bug Fixes to Double Grad Node Generation
      
      * Fixed yaml typo
      
      * Fixed yaml typo
      
      * fixed minor issues
      
      * [DoubleGrad PR #5] Enabled gradient computations for grad_tensors passed to paddle.grad()
      
      * Fixed minor issue
      
      * Fixed CI-Inference issue
      
      * Fixed CI-inference issues
      
      * [DoubleGrad PR #7] paddle.grad() to copy backward graph before backward run
      
      * Fixed minor issues
      
      * Fixed issue with backward graph construction logic
      
      * Fixed implementation issues with backward graph reconstruction
      
      * Fixed unittest issue
      
      * Fixed issues
      
      * [DoubleGrad PR #8] Enabled triple grads for sigmoid and matmul
      
      * Fixed issues with phi kernel
      
      * Added triple grad test case
      
      * Fixed minor issue
      d8a10977
    • G
      add new format of quantization (#41041) · b72a7ebb
      Guanghua Yu 提交于
      b72a7ebb
    • Z
      Implement AutoTuneStatus class for Kernel Auto Tune (#41218) · b0f8000e
      Zhang Ting 提交于
      * switch autotune
      
      * implement AutoTuneCache
      
      * implement AutoTuneCache class
      
      * add pybind api
      
      * add dygraph test
      
      * support static mode and eager mode and improve unittests
      
      * rename the SwitchAutoTune Class and improve tests
      
      * improve AutoTuneStatus and reduce the cost of tests
      b0f8000e
  8. 04 4月, 2022 4 次提交
  9. 03 4月, 2022 4 次提交
    • C
      [Phi]Concat grad (#41112) · 3f57ef7a
      chentianyu03 提交于
      * add concat_grad kernel
      
      * fix error
      
      * remove comment code
      
      * fix outs nullptr error
      
      * change to phi header
      
      * add concat_grad declare for standalone_executor_test
      3f57ef7a
    • F
      add maximum limit for grid of index_select (#41127) · af8d2482
      FlyingQianMM 提交于
      * limit grid dim for index select
      
      * mv LimitGridDim into gpu_launch_config.h
      
      * fix conflicts
      
      * fix conflicts
      
      * fix code style
      
      * set block to 256
      
      * fix grid setting
      
      * set dtype of block_dim to unsigned int
      af8d2482
    • Z
      Add randperm and range yaml (#41265) · fd1ecfc5
      zyfncg 提交于
      * add randperm and range yaml
      
      * add eager test for randperm
      fd1ecfc5
    • F
      Add some yaml config (#41053) · e4914734
      From00 提交于
      * Add yaml config
      
      * Add yaml for flatten_contiguous_range_op
      
      * Remove h_sigmoid yaml
      
      * Fix CI errors
      
      * Fix code format
      
      * Fix flatten OP errors
      
      * Fix conflicts
      
      * Fix CI errors
      
      * Remove flatten_contiguous_range OP
      
      * Remove redundant code
      
      * Fix typos
      e4914734
  10. 02 4月, 2022 9 次提交
  11. 01 4月, 2022 2 次提交
    • H
      update (#41245) · 99029dc9
      hong 提交于
      99029dc9
    • W
      [Eager] Support pinned (#41035) · f3270fc8
      wanghuancoder 提交于
      * support pinned, test=develop
      
      * support async_write, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine,test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      f3270fc8