1. 18 1月, 2023 4 次提交
    • R
      [PHI] remove bitwise and, or, xor (#49916) · 9056cc8b
      RuohengMa 提交于
      * add reduce_sum_int64 and reduce_sum_int8 xpu kernels
      
      * [PHI] add clip grad kernel with support type float32 and int32
      
      * [PHI unittest] add clip_grad unit test
      
      * adapt code to clang-format
      
      * update xpu api output with clip_grad api
      
      * remove int8 support of reduce_sum xpu kernel since it can not pass unit tests
      
      * adapt license date, add code for XPUDataType convertion
      
      * add int8 support of reduce_sum
      
      * add reduce_sum unit tests for dtype int64, int8, and add more test cases
      
      * update license date
      
      * remove buggy bitwise and, or and xor xpu kernels, refine bitwise not xpu kernel
      
      * change license date
      9056cc8b
    • H
      [XPU] add logical_not op. (#49911) · 60d1199a
      houj04 提交于
      60d1199a
    • W
      [0 Tensor support] support the 0d tensor for the cumsum (#49518) · 5fca45ea
      wawltor 提交于
      * Add the cumsum 0d tensor
      
      * xpu and cpu judge the 0d  tensor
      
      * change to 2022 to 2023 in new commit
      
      * fix the reverse logic
      5fca45ea
    • Z
      [Zero-Dim] Fix bug in masked_select for XPU (#49904) · 1a8be158
      Zhang Zheng 提交于
      1a8be158
  2. 13 1月, 2023 4 次提交
  3. 12 1月, 2023 3 次提交
  4. 09 1月, 2023 2 次提交
  5. 06 1月, 2023 1 次提交
    • R
      Dev (#49591) · 07db4a9f
      RuohengMa 提交于
      * add bitwise and, bitwise not, bitwise or and bitwise xor
      
      * correct typo
      07db4a9f
  6. 05 1月, 2023 1 次提交
  7. 29 12月, 2022 1 次提交
  8. 27 12月, 2022 1 次提交
  9. 26 12月, 2022 1 次提交
    • Y
      fix dlrm qpsproblem (#49171) · c8f76337
      ykkk2333 提交于
      * migrate shaple sgd, split,sign xpu kernels to phi, test=kunlun
      
      * fix dlrm throughput problem, test=kunlun
      c8f76337
  10. 23 12月, 2022 2 次提交
  11. 22 12月, 2022 1 次提交
  12. 20 12月, 2022 1 次提交
  13. 19 12月, 2022 1 次提交
  14. 14 12月, 2022 1 次提交
  15. 08 12月, 2022 1 次提交
  16. 07 12月, 2022 1 次提交
  17. 06 12月, 2022 2 次提交
  18. 05 12月, 2022 1 次提交
  19. 03 12月, 2022 1 次提交
  20. 02 12月, 2022 3 次提交
    • J
      [XPU ]Fix xpu compile error (#48621) · 2af82190
      Jiabin Yang 提交于
      * [Eager] Fix paddle.grad interface
      
      * [Eager] Support minimum SubGraph for GeneralGrad
      
      * Add needed_nodes to prune grad graph more thoroughly
      
      * [Eager] Add grad_node_trans_mapping_ to record which grad_node has been transformed to AccumulationNode
      
      * [Eager] Fix paddle.grad interface
      
      * Polish code
      
      * remove potential_stop_node
      
      * Add endding_nodes to enhance genSugraph logic
      
      * clear endding_nodes_
      
      * polish code
      
      * rename endding_nodes to endding_nades_
      
      * Refactor grad interface
      
      * Add register_hook case to fix coverage-ci
      
      * Fix code format
      
      * Refactor general_grad
      
      * Add more code comments
      
      * call clear directly to release GradSlotMeta
      
      * fix a mistake
      
      * fix matmul/ multiply kernel logic and optional input in yaml, fill zeros logic and so on.
      
      * fix batch_norm_double_grad yaml optional config
      
      * fix tanh_triple_grad yaml and kernels
      
      * fix MultiplyTripleGradKernel optional logic
      
      * fix merge mistake
      
      * fix compile error
      
      * remove legacy attr for bn
      
      * polish code
      
      * fix some kernel
      
      * merge develop
      
      * fix error
      
      * remote log
      
      * fix kernel with full like
      
      * hide value log behind
      
      * hide value log behind
      
      * fix matmul_triple grad
      
      * fix xpu compile error
      
      * fix xpu compile error
      
      * fix xpu ut
      
      * fix xpu ut
      
      * fix_xpu_compile_error
      Co-authored-by: NWeilong Wu <veyron_wu@163.com>
      2af82190
    • J
      [Eager] Optimize Grad by prune useless branch (#47827) · d1e93be1
      Jiabin Yang 提交于
      * [Eager] Fix paddle.grad interface
      
      * [Eager] Support minimum SubGraph for GeneralGrad
      
      * Add needed_nodes to prune grad graph more thoroughly
      
      * [Eager] Add grad_node_trans_mapping_ to record which grad_node has been transformed to AccumulationNode
      
      * [Eager] Fix paddle.grad interface
      
      * Polish code
      
      * remove potential_stop_node
      
      * Add endding_nodes to enhance genSugraph logic
      
      * clear endding_nodes_
      
      * polish code
      
      * rename endding_nodes to endding_nades_
      
      * Refactor grad interface
      
      * Add register_hook case to fix coverage-ci
      
      * Fix code format
      
      * Refactor general_grad
      
      * Add more code comments
      
      * call clear directly to release GradSlotMeta
      
      * fix a mistake
      
      * fix matmul/ multiply kernel logic and optional input in yaml, fill zeros logic and so on.
      
      * fix batch_norm_double_grad yaml optional config
      
      * fix tanh_triple_grad yaml and kernels
      
      * fix MultiplyTripleGradKernel optional logic
      
      * fix merge mistake
      
      * fix compile error
      
      * remove legacy attr for bn
      
      * polish code
      
      * fix some kernel
      
      * merge develop
      
      * fix error
      
      * remote log
      
      * fix kernel with full like
      
      * hide value log behind
      
      * hide value log behind
      
      * fix matmul_triple grad
      Co-authored-by: NWeilong Wu <veyron_wu@163.com>
      d1e93be1
    • Y
      add silu, silu_grad, unfold and unfold_grad xpu kernels (#48325) · f71de378
      ykkk2333 提交于
      * add stat tool
      
      * add roll and roll_grad kernels and strided_slice and strided_slice_grad kernels, test=kunlun
      
      * add silu, unfold and their grads,test=kunlun
      f71de378
  21. 01 12月, 2022 2 次提交
  22. 30 11月, 2022 1 次提交
  23. 29 11月, 2022 1 次提交
  24. 28 11月, 2022 3 次提交