1. 19 10月, 2022 1 次提交
  2. 27 9月, 2022 1 次提交
  3. 26 9月, 2022 1 次提交
  4. 23 9月, 2022 1 次提交
  5. 19 9月, 2022 1 次提交
  6. 15 9月, 2022 1 次提交
  7. 14 9月, 2022 1 次提交
  8. 06 9月, 2022 1 次提交
  9. 27 7月, 2022 1 次提交
  10. 25 7月, 2022 1 次提交
  11. 05 7月, 2022 1 次提交
  12. 01 7月, 2022 1 次提交
  13. 30 6月, 2022 1 次提交
  14. 05 6月, 2022 1 次提交
    • S
      【code format check upgrade】 step2:yapf (#42944) · a072fca8
      Sing_chan 提交于
      * use yapf to format all python file
      
      * yapf exclude two unittests file for they rely on writing and reading file, and format will break them
      
      * disable diff_py_file because too many diff files cause command following failed
      a072fca8
  15. 29 4月, 2022 1 次提交
  16. 30 3月, 2022 1 次提交
  17. 25 3月, 2022 1 次提交
    • J
      Refactor Dygraph Flags (#40786) · 3085d5e4
      Jiabin Yang 提交于
      * refactor eager flags
      
      * fix flags error when we switch from eager to dygraph
      
      * fix ci problem
      
      * fix ci
      
      * fix ci
      
      * merge develop and fix code style
      
      * merge develop and fix code style
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * merge develop
      3085d5e4
  18. 17 3月, 2022 1 次提交
    • 0
      Refine io for test_mnist.py (#40496) · 1e045cae
      0x45f 提交于
      * for test_mnist.py
      
      * remove comments
      
      * using type() replace isinstance()
      
      * valid vars for run program OP in io.py
      
      * open test_mnist in eager_gurad for coverage
      1e045cae
  19. 03 3月, 2022 1 次提交
    • J
      Support slim eager (#39874) · da47544c
      Jiabin Yang 提交于
      * eager, test=develop
      
      * fix bug, test=develop
      
      * eager, test=develop
      
      * merge legacy to fluid
      
      * eager, test=develop
      
      * eager, test=develop
      
      * Refactor TensorAdd func by template and remove gradient_accumulation in eager
      
      * Remove needless target name
      
      * eager, test=develop
      
      * eager, test=develop
      
      * Use overload instead of template
      
      * Remove legacy code
      
      * Remove legacy code
      
      * selectedrows, test=develop
      
      * Remove DataType test
      
      * eager, test=develop
      
      * eager, test=develop
      
      * support gan, test=develop
      
      * Using Tensor directly instead of using EagerTensor
      
      * support gradient_accumulation
      
      * make test_imperative_lod_tensor_to_selected_rows longer
      
      * make test_imperative_lod_tensor_to_selected_rows longer
      
      * refine code
      
      * ptb, test=develop
      
      * Rename all EagerTensor to Tensor
      
      * Rename some EagerTensor to Tensor
      
      * rename EagerTensor to EagerVariable
      
      * eager, test=develop
      
      * eager, test=develop
      
      * eager, test=develop
      
      * eager, test=develop
      
      * add more test
      
      * eager, test=develop
      
      * Support copiable selected rows and merge develop
      
      * save load, eager, test=develop
      
      * save load, eager, test=develop
      
      * refine, test=develop
      
      * remove useless _set_value method
      
      * refine, test=develop
      
      * refine, test=develop
      
      * revert static_runner, test=develop
      
      * EagerTensor to Tensor, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * clear grad, test=develop
      
      * merge, develop
      
      * merge, develop
      
      * merge, test=develop
      
      * merge, test=develop
      
      * Support quant and part of slice
      
      * support legacy static save
      
      * extend slim tests time
      
      * remove imperative on inference
      
      * remove imperative on inference
      
      * merge develop
      
      * fix typo
      
      * fix typo
      
      * split slice related code into 2 part for imperative and eager
      
      * split slice from inference
      
      * split slice from inference
      
      * fix test_tensor_register_hook
      Co-authored-by: NWang Huan <wanghuan29@baidu.com>
      Co-authored-by: NWeilong Wu <veyron_wu@163.com>
      Co-authored-by: Nwanghuancoder <wanghuancoder@163.com>
      da47544c
  20. 11 1月, 2022 1 次提交
    • M
      Jit pre save hook (#38186) · e91f7c02
      Ming-Xu Huang 提交于
      * Pre-save hooks of jit.save
      
      1. Added pre_save_hooks features to jit.save.
      2. Added related unittests
      
      * Added jit pre_save_hooks functions's alias to paddle.jit and copyright.
      
      * Make jit.save_pre_hook style be consisent with Paddle's rule.
      
      * Fixed arguments passing bug in run_save_pre_hooks
      
      * Added API Documents
      
      * Move clear and run_pre_save_hooks as internal methonds only.
      
      * Made register_save_pre_hook as an internal function.
      e91f7c02
  21. 20 10月, 2021 1 次提交
    • S
      Add FasterTokenizer Operator (#34491) · 3f2d6a3f
      Steffy-zxf 提交于
      Add Tokenizer related functionalities for Transformer model in order that the process of training and predicting is consistent.
      
      * support the text string as an input Tensor
      * support the "VOCAB"unordered_map<wstring, int> as an input Tensor to lookup tokens
      * Tokenizer used for BERT. This tokenizer applies an end-to-end, text string to wordpiece tokenization.
      * It first applies basic tokenization, followed by wordpiece tokenization.
      3f2d6a3f
  22. 17 9月, 2021 2 次提交
    • Z
      [AMP] Support pure fp16 training mode for dygraph (#35521) · adaeee4d
      zhangbo9674 提交于
      * add pure fp16 major function in auto_cast & tracer
      
      * support master weight in dygraph for pure fp16
      
      * check mix dtype of fp16&fp32 for check_finite_and_unscale op
      
      * change pure fp16 funtion name
      
      * refine some bug in auto_cast
      
      * refine auto_cast interface logic
      
      * add param _casted_by_pure_fp16 for class Layer
      
      * support state_dict hook for save model by user appointed dtype in pure_fp16_decorator
      
      * refine pure_fp16_decorator as decorator
      
      * add unittest
      
      * add comment
      
      * add comment
      
      * support recompute
      
      * add comment for auto_cast and decorator
      
      * support to_static_state_dict for paddle.jit.save
      
      * unlimite models num and optimizers num
      
      * add lookup_table in black_list
      
      * fix momentum and layer state_dict
      
      * fix bug in layer state_dict
      
      * fix bug in layer state_dict_helper
      
      * refine unittest
      
      * refine test_momentun_op
      
      * refine interface and some code
      
      * refine amp_decorator interface
      
      * refine pure fp16 interface
      
      * refine master weight interface
      adaeee4d
    • W
      polish code. (#35783) · 61010bb8
      WeiXin 提交于
      61010bb8
  23. 15 9月, 2021 1 次提交
  24. 28 7月, 2021 1 次提交
  25. 14 7月, 2021 1 次提交
  26. 17 6月, 2021 1 次提交
  27. 27 4月, 2021 1 次提交
  28. 26 4月, 2021 1 次提交
  29. 07 4月, 2021 1 次提交
  30. 11 3月, 2021 1 次提交
  31. 20 1月, 2021 1 次提交
  32. 18 12月, 2020 1 次提交
  33. 03 12月, 2020 1 次提交
    • L
      [Dy2stat] Add a decorator paddle.jit.not_to_static to support that not to... · b9a8ebd5
      liym27 提交于
      [Dy2stat] Add a decorator paddle.jit.not_to_static to support that not to convert a function in Dynamic-to-Static. (#29253)
      
      Usage scenarios:A function could have run successfully in static mode,  you can use it to decorate a function in the following cases:
        1. An unknown error occurs in the dynamic-to-static conversion process of the function;
        2. In the internal implementation of the function, it has two branches: dynamic branch and static branch;
        3. Users don't want to convert the function in the process of dynamic to static.
      b9a8ebd5
  34. 27 11月, 2020 1 次提交
  35. 19 11月, 2020 1 次提交
  36. 11 11月, 2020 1 次提交
  37. 06 11月, 2020 2 次提交
  38. 26 10月, 2020 1 次提交