1. 09 8月, 2023 1 次提交
    • X
      【New IR】backward code of new ir (#55957) · 0399b39f
      xiaoguoguo626807 提交于
      * [prim][newir] add basic framework for primitive
      
      * support desctensor in new ir
      
      * add vjp interface
      
      * support vjp in new ir
      
      * support vjp in new ir
      
      * polish vjp interface
      
      * fix stop_gradients set
      
      * fix vjp dispatch
      
      * add comment
      
      * add vjp test for new ir
      
      * add test for tanh vjp
      
      * [prim][newir] add basic framework for primitive
      
      * support desctensor in new ir
      
      * support vjp in new ir
      
      * support vjp in new ir
      
      * polish vjp interface
      
      * fix stop_gradients set
      
      * fix vjp dispatch
      
      * add comment
      
      * add vjp test for new ir
      
      * add test for tanh vjp
      
      * add eager and static backend for warp lower level api
      
      * support call_vjp pybind
      
      * polish code and add test for vjp
      
      * remove useless code
      
      * polish code
      
      * remove useless code
      
      * support mean vjp
      
      * backward origin code
      
      * add test for mean vjp and support has_vjp function
      
      * fix call_vjp
      
      * polish code
      
      * add attrs and dtype interface
      
      * add primitive ops set for backend
      
      * fix compile bugs
      
      * fix some bugs
      
      * fix windows bugs
      
      * add vjp test for tanh_
      
      * fix inference CI
      
      * fix inference ci
      
      * modify fluid cmake
      
      * origin test of tanh and mean passed
      
      * fix conflict
      
      * modify stop_gradient
      
      * modify block.ops
      
      * modify test
      
      * fix conflict
      
      * reply review comments
      
      * reply review comments
      
      * pulish code
      
      ---------
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      Co-authored-by: NCharles-hit <wanghao107@baidu.com>
      Co-authored-by: Nzhangbo9674 <zhangbo54@baidu.com>
      Co-authored-by: NYuanRisheng <yuanrisheng@baidu.com>
      0399b39f
  2. 02 8月, 2023 1 次提交
    • O
      [xdoctest] reformat example code with google style in No.21-30 (#55849) · 5d26d79f
      ooooo-create 提交于
      * [Doctest]fix No.21, test=docs_preview
      
      * Revert "[Doctest]fix No.21, test=docs_preview"
      
      This reverts commit 76bcdb280e254d682be6fc6f85588f1940bb1ade.
      
      * [Doctest]fix No.21, test=docs_preview
      
      * fix bugs,test=docs_preview
      
      * [Doctest]fix No.22-24,26,27, test=docs_preview
      
      * update fix
      
      * with pre-commit, test=docs_preview
      
      * fix seed, test=docs_preview
      
      * fix error, test=docs_preview
      
      * fix seed, test=docs_preview
      
      * fix seed, test=docs_preview
      5d26d79f
  3. 06 7月, 2023 1 次提交
  4. 23 5月, 2023 1 次提交
  5. 22 5月, 2023 1 次提交
    • M
      [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode() (#53856) · 3794d171
      Meteor Liu 提交于
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * fixed cyclic reference that caused patial import
      
      * fixed bad change
      
      * fix bad import
      
      * fix bad import
      
      * fix bad import
      
      * fix ut failed caused by change in_dynamic_mode
      
      * fix ut failed caused by change in_dynamic_mode
      
      * fixed usage of in_dynamic_mode() or in_dygraph_mode()
      
      * revert python3 to python in .pre-commit-config.yaml
      
      * fix merge conflicts
      3794d171
  6. 16 5月, 2023 1 次提交
  7. 12 5月, 2023 1 次提交
  8. 27 4月, 2023 1 次提交
    • H
      Add jacobian and hessian (#53331) · e8d296ef
      HydrogenSulfate 提交于
      * add jacobian and hessian in paddle.autograd
      
      * disable unitest 'func_multi_input' for bug in high-order gradient of multiply
      
      * add dimension checks
      
      * add support for 0-D tensor
      
      * change return type from Jacobian to Hessian in hessian function
      
      * refine Jacobian _flatten function for single xs
      
      * refine support for 0-D tensor
      
      * 1. add 'func_multi_input' unitest for multiply_grad_kernel bug fixed
      already.
      2. support non-inplace math operation via magical method overwriting.
      
      * add unitest for math operation and raise error when 0-D tensor is indexed
      
      * add ndim check on ys and xs according to is_batched, and add one unitest
      
      * refine docstring of jacobian and hessian
      
      * move paddle.incubate.autograd.Jacobian/Hessian to paddle.incubate.autograd.functional.Jacobian/Hessian
      
      * remove single_input unitest case because numerical differentiation is wrong
      
      * remove 3 unitest for numerical result(reference result) is wrong
      
      * 1. rename autodiff.py to autograd.py
      2. increase TIMEOUT to 100
      
      * cancel modification for functional Jacobian/Hessian
      
      * 1. use tuple as return type instead of list
      2. refine docstring
      
      * add more unitest case to improve coverage
      
      * remove 2 unitest of Hessian for numerical result is wrong
      
      * remove 1 unitest of Hessian for numerical result is wrong
      
      * remove 1 unitest of Hessian for numerical result is wrong
      
      * change unit test to shape check
      
      * correct doc and replace incubate API to stable API in _grad
      e8d296ef
  9. 31 3月, 2023 1 次提交
  10. 28 3月, 2023 1 次提交
  11. 07 3月, 2023 1 次提交
  12. 27 2月, 2023 1 次提交
  13. 22 2月, 2023 1 次提交
  14. 17 1月, 2023 1 次提交
  15. 28 12月, 2022 1 次提交
  16. 20 12月, 2022 1 次提交
  17. 05 12月, 2022 1 次提交
  18. 29 11月, 2022 1 次提交
  19. 08 11月, 2022 1 次提交
  20. 03 11月, 2022 1 次提交
  21. 23 10月, 2022 1 次提交
  22. 10 10月, 2022 1 次提交
  23. 22 9月, 2022 1 次提交
    • Fix the En docs (delete some expression like 'This OP') (#46165) · 3a928a8c
      张春乔 提交于
      * 1. Delete some expression like 'This Op'
      2. remove import numpy as np
      
      * test=document_fix
      
      * fix eg; test=document_fix
      
      * fix 'import numpy' cases; test=document_fix
      
      * fix 'import numpy' cases; test=document_fix
      
      * fix some docs; test=document_fix
      
      * delete raise; test=document_fix
      
      * add some introduction; test=document_fix
      
      * add some introduction; test=document_fix
      
      * test=document_fix
      
      * Fix ’note‘ format; test=document_fix
      
      * Fix Returns of cholesky; test=document_fix
      
      * Fix Example format; test=document_fix
      
      * Fix det; test=document_fix
      
      * Fix eig; test=document_fix
      
      * Fix eigh; test=document_fix
      
      * Fix eigh; test=document_fix
      
      * Apply suggestions from code review;test = document_fix
      Co-authored-by: NNyakku Shigure <sigure.qaq@gmail.com>
      
      * Apply suggestions from code review;test = document_fix
      Co-authored-by: NNyakku Shigure <sigure.qaq@gmail.com>
      
      * Apply suggestions from code review;test = document_fix
      Co-authored-by: NNyakku Shigure <sigure.qaq@gmail.com>
      
      * test=document_fix
      
      * test=document_fix
      
      * KLDiv;test=document_fix
      
      * norm example code; test=document_fix
      
      * revert python/paddle/fluid/**/*
      
      * revert python/paddle/distributed/spawn.py
      
      * revert python/paddle/fluid/*
      
      * fix a `Note` format
      
      * Fix inv; test=document_fix
      
      * Fix lu; test=document_fix
      
      * Fix lu_unpack; test=document_fix
      
      * Fix matrix_power; test=document_fix
      
      * Fix multi_dot; test=document_fix
      
      * Fix solve; test=document_fix
      Co-authored-by: NNyakku Shigure <sigure.qaq@gmail.com>
      3a928a8c
  24. 19 9月, 2022 1 次提交
  25. 15 9月, 2022 1 次提交
  26. 14 9月, 2022 1 次提交
  27. 20 8月, 2022 1 次提交
  28. 11 7月, 2022 1 次提交
    • X
      reorganize the higher order autodiff api (#44119) · dd63e5b4
      Xiaoxu Chen 提交于
      * move _gradients to primapi and rename to grad
      
      * modify jvp to call forward_grad in primitive mode
      
      * add primapi unittest and remove some unused test cases.
      
      * fix  circular import problem
      
      * move paddle/autograd/functional into paddle/incubate.autograd/functional
      
      * remove unused JacobianBatchLast class
      dd63e5b4
  29. 28 6月, 2022 1 次提交
  30. 27 6月, 2022 1 次提交
  31. 05 6月, 2022 1 次提交
    • S
      【code format check upgrade】 step2:yapf (#42944) · a072fca8
      Sing_chan 提交于
      * use yapf to format all python file
      
      * yapf exclude two unittests file for they rely on writing and reading file, and format will break them
      
      * disable diff_py_file because too many diff files cause command following failed
      a072fca8
  32. 18 5月, 2022 1 次提交
    • W
      Add support for forward and reverse high-order automatic differentiation mechanism (#41919) · f6ee202f
      WangZhen 提交于
      * Updated triple_grad_check func
      
      * add todo for gradient checker and refine some comments
      
      * remove additional code
      
      * add test for warnging in backward.py
      
      * format python code
      
      * support multi input in triple gradient checker
      
      * Add matmul triple grad kernel
      
      * Updated comments of TODO
      
      * Supported some special tests
      
      * Change code-format to follow CI std
      
      * Updated gradient_checker.py
      
      * Fix conflicts
      
      * Removed unnecessary printing log
      
      * Change code style to follow CI std
      
      * merge upstream
      
      * add priops.py
      
      * add_p
      
      * rm useless files
      
      * add sub_p mul_p div_p
      
      * add sqrt_p and tanh_p
      
      * add reshape_p
      
      * add broadcast_p
      
      * Add python primitive wrappers.
      
      * Jvp rules updated.
      
      * JVP rules done for all the 17 primops.
      
      * quick check and fixes.
      
      * add jvp(op, *args)
      
      * add broadcast_p fill_constant_p matmul_p reduce_p reshape_p transpose_p
      
      * add split_p and concat_p
      
      * add gather_p and scatter_add_p
      
      * add slice_select_p and slice_assign_p
      
      * Add transpose rules.
      
      * add multi input check for add_p, sub_p, mul_p, div_p
      
      * update concat_p
      
      * Linearize and transpose in progress..
      
      * refine gather_p and scatter_add_p
      
      * updated.
      
      * update transpose.
      
      * refine slice_assign_p and slice_select_p
      
      * init commit for lower
      
      * Merged with primitive ops.
      
      * small update
      
      * add rules for orig2prim and prim2orig
      
      * add 9 test for prim ops
      
      * add more test and fix some bug
      
      * add more test
      
      * register proto
      
      * Adding primops test.
      
      * add shape valid check for broadcast_p op, and add keepdim attr into reduce_p op proto
      
      * support multi input and multi output for split_p and concat_p
      
      * Test updated.
      
      * update
      
      * fix slice bug for slice_select_p and slice_assign_p
      
      * updated.
      
      * Ops updated.
      
      * Refactor and bug fixes.
      
      * updated.
      
      * finish orig2prim and prim2orig rules
      
      * dtype for axis attr should be long int
      
      * update dtype for axis attr int64_t
      
      * update for iscan CI
      
      * Update primx.
      
      * Refactor vars in primx.
      
      * update for lower transform
      
      * add more shape and dtype check
      
      * update primx.py
      
      * change IndexTensor into int32 dtype
      
      * update
      
      * Fix linearize and transpose.
      
      * Update is_dot
      
      * Update is_dot
      
      * Update is_dot
      
      * add gradient aggregation, fix add_transpose.
      
      * pass first linearize+transpose test.
      
      * update test
      
      * refactor op registration and primx.
      
      * update rule for slice_assign
      
      * try test lower
      
      * update orig2prim and prim2orig
      
      * pass simple lower pass
      
      * update
      
      * Update input types in the unit test.
      
      * orig2prim segfault.
      
      * 50% for adam.minimize
      
      * test updated.
      
      * temp fix erros in removing vars.
      
      * primx updated.
      
      * update for matmul_v2 and reshape2 orig2prim
      
      * update for minimize
      
      * Refine primrules
      
      * Remove some code
      
      * supporting unused and unreachable vars.
      
      * update for use prim2orig in minimize
      
      * fix gather and scatter_add transpose.
      
      * Add rules UT
      
      * update scatter_add
      
      * Refine UT code
      
      * fix nonetype check in topo
      
      * Update gather_p pywrapper.
      
      * remove useless print
      
      * Merge tongxin PR and refine code
      
      * readd some test
      
      * rm useless print
      
      * polish code.
      
      * fix bug in minimize
      
      * add get_input_var_list and get_output_var_list and use it in lower
      
      * Fix scatter_add_p prim2orig
      
      * Update code and fix orig2prim/prim2orig UT
      
      * delete vars after block.desc._remove
      
      * Improve ops and vars clean up logics.
      
      * fix some bug in linearize and lower
      
      * update tanh transpose.
      
      * use set instead of list for var2remove
      
      * test updated.
      
      * polish code.
      
      * fix dot2bar delete.
      
      * merge tx/ad
      
      * add indextensor_dot for gather and scatter_add
      
      * add sorted for set
      
      * Fix scale_orig2prim params
      
      * fix some syntax bug
      
      * add golbal_lower_update list
      
      * Better handling of unused vars.
      
      * update tests.
      
      * Fix elementwise_sub orig2prim
      
      * support none for transpose rule
      
      * Merge and add transform UT
      
      * fix a bug in transpose
      
      * Fix transpose and UT
      
      * a hacky fix for cancat op
      
      * Fix exector place
      
      * Refine variable name
      
      * Add elementwise_mul orig2prim and support p_norm when p=1
      
      * Add sqrt orig2prim rule and UT
      
      * merge wz test
      
      * rename files, add enable_prim, disable_prim, prim_enabled, delete global_lower_update
      
      * fix a bug in test_ad_transform_trans
      
      * revert modify in framework.py
      
      * add paddle.fluid.incubate.ad_transform to  python/setup.py.in
      
      * Fix remove vars error
      
      * Fix p_norm_orig2prim
      
      * merge wz
      
      * Modify the code directory
      
      * Add utils.py and remove get_input/output_vars functions
      
      * Update maolin code
      
      * Rename UT and refine test_ad_transform_primops
      
      * Fix div_p jvp rule
      
      * Add higher derivatives UT
      
      * Remove UT to autograd dir
      
      * Fix comments
      
      * import paddle in primops.py
      
      * Add some error message for assert
      
      * Refine UT class name and refine some comments in primreg.py
      
      * update minimize of paddle/optimizer for supporting new autograd
      
      * resolve cicular importing between backward.py and optimizer.py
      
      * fill gradients and minimize unittest
      
      * Replace `assert isinstance` with `raise TypeError`
      
      * Add some assert message for primx.py
      
      * Polish variable name
      
      * Add some assert message
      
      * add some docstring
      
      * refine some name
      
      * update the format of english documents
      
      * Split test_transform.py to two files to avoid ci error
      
      * fix the document format of enable_prim/disable_prim/prim2orig/prim_enabled
      
      * polish test_gradients_and_minimize
      
      * add default value for prim_enabled api doc
      
      * Remove some UT to avoid windows ci error
      
      * Enlarge test_gradients_and_minimize limit time
      
      * Fix ut limit time
      Co-authored-by: Nveyron95 <veyron_wu@163.com>
      Co-authored-by: NJiabin Yang <360788950@qq.com>
      Co-authored-by: Nlevi131 <limaolin01@baidu.com>
      Co-authored-by: NTongxin Bai <waffle.bai@gmail.com>
      Co-authored-by: NXiaoxu Chen <chenxx_id@163.com>
      Co-authored-by: Nlevi131 <83750468+levi131@users.noreply.github.com>
      f6ee202f
  33. 16 4月, 2022 1 次提交
    • L
      Lml/prim op pywrapper (#41813) · ebf4fe6e
      levi131 提交于
      * native commit for triple grad of sigmod
      
      * Updated unittests files
      
      * init functional jacobian api
      
      * Updated trible_test func
      
      * Updated gradient_checker & test_script
      
      * finish test with dtype float32
      
      * add float64 test case
      
      * polish code
      
      * use atol=1e-5 with dtype float64
      
      * fix for ci
      
      * set timeout for test_jacobian
      
      * fix dygraph grad to support high differential
      
      * polish API docstring
      
      * Updated gradient checker and some related files
      
      * fix double grad strip error for high differential
      
      * fix double grad strip error for high differential
      
      * Add Sigmoid triple grad tests
      
      * fix dygraph double grad dtype error when calling for high differential senario
      
      * Updated triple grad teses func
      
      * Use np.random to initialize ddx
      
      * Updated triple_grad_check func
      
      * add todo for gradient checker and refine some comments
      
      * remove additional code
      
      * add test for warnging in backward.py
      
      * format python code
      
      * support multi input in triple gradient checker
      
      * Add matmul triple grad kernel
      
      * Updated comments of TODO
      
      * Supported some special tests
      
      * Change code-format to follow CI std
      
      * Updated gradient_checker.py
      
      * Fix conflicts
      
      * Removed unnecessary printing log
      
      * Change code style to follow CI std
      
      * merge upstream
      
      * add priops.py
      
      * add_p
      
      * rm useless files
      
      * add sub_p mul_p div_p
      
      * add sqrt_p and tanh_p
      
      * add reshape_p
      
      * add broadcast_p
      
      * Add python primitive wrappers.
      
      * Jvp rules updated.
      
      * JVP rules done for all the 17 primops.
      
      * quick check and fixes.
      
      * add jvp(op, *args)
      
      * add broadcast_p fill_constant_p matmul_p reduce_p reshape_p transpose_p
      
      * add split_p and concat_p
      
      * add gather_p and scatter_add_p
      
      * add slice_select_p and slice_assign_p
      
      * Add transpose rules.
      
      * add multi input check for add_p, sub_p, mul_p, div_p
      
      * update concat_p
      
      * Linearize and transpose in progress..
      
      * refine gather_p and scatter_add_p
      
      * updated.
      
      * update transpose.
      
      * refine slice_assign_p and slice_select_p
      
      * init commit for lower
      
      * Merged with primitive ops.
      
      * small update
      
      * add rules for orig2prim and prim2orig
      
      * add 9 test for prim ops
      
      * add more test and fix some bug
      
      * add more test
      
      * register proto
      
      * Adding primops test.
      
      * add shape valid check for broadcast_p op, and add keepdim attr into reduce_p op proto
      
      * support multi input and multi output for split_p and concat_p
      
      * Test updated.
      
      * update
      
      * fix slice bug for slice_select_p and slice_assign_p
      
      * updated.
      
      * Ops updated.
      
      * Refactor and bug fixes.
      
      * updated.
      
      * finish orig2prim and prim2orig rules
      
      * dtype for axis attr should be long int
      
      * update dtype for axis attr int64_t
      
      * update for iscan CI
      
      * Update primx.
      
      * Refactor vars in primx.
      
      * update for lower transform
      
      * update primx.py
      
      * update
      
      * Fix linearize and transpose.
      
      * Update is_dot
      
      * Update is_dot
      
      * Update is_dot
      
      * add gradient aggregation, fix add_transpose.
      
      * pass first linearize+transpose test.
      
      * update test
      
      * add_prim_op_pywrapper
      
      * Add primops UT
      
      * Fix set_value and update
      
      * Fix code format and PR-CI-Coverage
      Co-authored-by: Nveyron95 <veyron_wu@163.com>
      Co-authored-by: NJiabin Yang <360788950@qq.com>
      Co-authored-by: NTongxin Bai <waffle.bai@gmail.com>
      Co-authored-by: N0x45f <wangzhen45@baidu.com>
      ebf4fe6e
  34. 14 4月, 2022 1 次提交
    • Z
      [DoubleGrad] Enabled test_autograd_functional_dynamic.py under eager mode (#41668) · ad9585b6
      Zhanlue Yang 提交于
      * [DoubleGrad] Enabled double grad test cases in eager_mode for test_imperative_double_grad
      
      * Fixed elementwise issue
      
      * Addressed CI failures
      
      * [DoubleGrad] Enabled test_imperative_triple_grad test cases under eager_mode
      
      * [DoubleGrad] Enabled test_autograd_functional_dynamic.py under eager mode
      
      * Enabled more test cases
      
      * Fixed performance issues
      
      * Fixed minor issue
      ad9585b6
  35. 02 4月, 2022 1 次提交
    • X
      Enhance vjp/jvp/Jacobian/Hessian API for supporting dynamic, static graph and... · 9e764d82
      Xiaoxu Chen 提交于
      Enhance vjp/jvp/Jacobian/Hessian API for supporting dynamic, static graph and batched, unbatched mode (#40692)
      
      * modify vjp/jvp for both dynamic and static graph
      
      * enforce jacobian class for supporting first/last batch
      
      * add unittest for jvp, jacobian withlast batch, jacobian with first batch
      
      * fix the incorrect shape when multi-index Jacobian
      
      * enforce Hessian class for supporting dynamic graph
      
      * add Hessian class unittest
      
      * bugfix, jvp double_backward_trick zeros_like return stop_gradient=True in static graph
      
      * add API beta warnnings
      
      * add white_list for cuda11.x ci windows.
      
      * optimize some code snippets and documments
      
      * set unittest timeout to 100 seconds
      
      * move vjp,jvp,Jacobian,Hessian to incubate
      
      * fix vjp,vjp import path of sample code
      
      * fix code style error of augtograd/__init__ file
      9e764d82
  36. 30 3月, 2022 1 次提交
    • W
      [Eager] Pylayer (#39989) · 157c1a28
      wanghuancoder 提交于
      * Supported Complex2Real Conversion for Eager Dygraph
      
      * Supported Complex2Real Conversion for Eager Dygraph
      
      * Enabled complex type promotion test for matmul_v2
      
      * pylayer, test=develop
      
      * Fix CI issues
      
      * Support initializing specific grad tensors to zero for selected operators
      
      * finish forward, test=develop
      
      * create grad node finish, test=develop
      
      * Merged adj_edges_ with GradSlotMeta
      
      * Fixed monir issue
      
      * backward finish, start dbg, test=develop
      
      * Adjusted num runs
      
      * Recovered Eager performance tests configurations
      
      * Recovered Eager performance tests configurations
      
      * finish, test=develop
      
      * polish, test=develop
      
      * polish, test=develop
      
      * refine, test=develop
      
      * eager, test=develop
      
      * Adjusted performance tests configurations
      
      * Fixed Minor Issues with performance tests
      
      * [Phi] Fix macro name typo
      
      * support set_materialize_grads, test=develop
      
      * suppotr mark_non_differentiable, test=develop
      
      * support once_differentiable, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * Moved out Edge from GradSlotMeta
      
      * Fixed issues from merge
      
      * Fixed typo
      
      * Addressed review comments
      
      * Fixed merge issues
      
      * Fixed minor issues
      
      * Fixed minor issue
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * Fixed major issues and enabled auto_prune test cases
      
      * Fixed issues from merge
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      Co-authored-by: Njim19930609 <jim19930609@gmail.com>
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      157c1a28
  37. 25 3月, 2022 1 次提交
    • J
      Refactor Dygraph Flags (#40786) · 3085d5e4
      Jiabin Yang 提交于
      * refactor eager flags
      
      * fix flags error when we switch from eager to dygraph
      
      * fix ci problem
      
      * fix ci
      
      * fix ci
      
      * merge develop and fix code style
      
      * merge develop and fix code style
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * merge develop
      3085d5e4
  38. 14 3月, 2022 1 次提交
    • J
      Support custom op and paddle.autograd.bacward in eager (#40423) · 227fa408
      Jiabin Yang 提交于
      * eager, test=develop
      
      * fix bug, test=develop
      
      * eager, test=develop
      
      * merge legacy to fluid
      
      * eager, test=develop
      
      * eager, test=develop
      
      * Refactor TensorAdd func by template and remove gradient_accumulation in eager
      
      * Remove needless target name
      
      * eager, test=develop
      
      * eager, test=develop
      
      * Use overload instead of template
      
      * Remove legacy code
      
      * Remove legacy code
      
      * selectedrows, test=develop
      
      * Remove DataType test
      
      * eager, test=develop
      
      * eager, test=develop
      
      * support gan, test=develop
      
      * Using Tensor directly instead of using EagerTensor
      
      * support gradient_accumulation
      
      * make test_imperative_lod_tensor_to_selected_rows longer
      
      * make test_imperative_lod_tensor_to_selected_rows longer
      
      * refine code
      
      * ptb, test=develop
      
      * Rename all EagerTensor to Tensor
      
      * Rename some EagerTensor to Tensor
      
      * rename EagerTensor to EagerVariable
      
      * eager, test=develop
      
      * eager, test=develop
      
      * eager, test=develop
      
      * eager, test=develop
      
      * add more test
      
      * eager, test=develop
      
      * Support copiable selected rows and merge develop
      
      * save load, eager, test=develop
      
      * save load, eager, test=develop
      
      * refine, test=develop
      
      * remove useless _set_value method
      
      * refine, test=develop
      
      * refine, test=develop
      
      * revert static_runner, test=develop
      
      * EagerTensor to Tensor, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * clear grad, test=develop
      
      * merge, develop
      
      * merge, develop
      
      * merge, test=develop
      
      * merge, test=develop
      
      * Support quant and part of slice
      
      * support legacy static save
      
      * extend slim tests time
      
      * remove imperative on inference
      
      * remove imperative on inference
      
      * merge develop
      
      * fix typo
      
      * fix typo
      
      * split slice related code into 2 part for imperative and eager
      
      * split slice from inference
      
      * split slice from inference
      
      * fix test_tensor_register_hook
      
      * support custom op in eager mode
      
      * fix inference deps error
      
      * split eager utils from custom operator
      
      * fix type match
      
      * fix typo
      Co-authored-by: NWang Huan <wanghuan29@baidu.com>
      Co-authored-by: NWeilong Wu <veyron_wu@163.com>
      Co-authored-by: Nwanghuancoder <wanghuancoder@163.com>
      227fa408
  39. 27 2月, 2022 1 次提交
  40. 29 1月, 2022 1 次提交
    • T
      Symbolic Hessian (#39221) · 64e7c715
      Tongxin Bai 提交于
      * [autograd] static Jacobian pass tests.
      
      * [autograd] apply CR suggested changes.
      
      * [autograd] more tests.
      
      * [autograd] add CPUPlace in tests.
      
      * [autograd] bug fixes.
      
      * [autograd] reformatted.
      
      * [autograd] adding Hessian, in progress.
      
      * [autograd] Hessian passes. A double grad bug fixed.
      
      * [autograd] fix renaming conflict in double backward pass.
      
      * [autograd] polish test.s
      
      * fix a bug when using brackets
      
      * debug for ci
      
      * [autograd] fixing Hessian test.
      
      * polish format.
      Co-authored-by: Nlevi131 <83750468+levi131@users.noreply.github.com>
      Co-authored-by: Nlevi131 <limaolin01@baidu.com>
      64e7c715