- 10 3月, 2023 1 次提交
-
-
由 sneaxiy 提交于
* add attn_bias.py * add Python interface * add license * add test_attn_bias.py * fix CPU test error * fix ci error
-
- 22 2月, 2023 1 次提交
-
-
由 Shuangchi He 提交于
* Fix some typos. Signed-off-by: Yulv-git <yulvchi@qq.com> * pre-commit Signed-off-by: Yulv-git <yulvchi@qq.com> --------- Signed-off-by: Yulv-git <yulvchi@qq.com>
-
- 15 2月, 2023 1 次提交
-
-
由 lzy 提交于
* make FusedMultiTransformer supports variable-lengths. * modify ffn2 when cuda_version >= 11.6 because of #49392. * code style * delete remove_padding
-
- 01 2月, 2023 1 次提交
-
-
由 RedContritio 提交于
* add shape check for fused_multi_head_attention * use raise for coverage test * add unittest * remove unnecessary pass * add unittest
-
- 05 1月, 2023 2 次提交
- 23 12月, 2022 1 次提交
-
-
由 lzy 提交于
-
- 22 12月, 2022 1 次提交
-
-
由 xiaoxiaohehe001 提交于
-
- 07 12月, 2022 1 次提交
-
-
由 Kevin吴嘉文 提交于
-
- 29 11月, 2022 1 次提交
-
-
由 Nyakku Shigure 提交于
* isort all files * revert conflicting files * revert conflicting files * revert conflicting files
-
- 28 11月, 2022 1 次提交
-
-
由 yuehuayingxueluo 提交于
* clear fluid api: warpctc, nce, identity_loss * fix test_layers.py __init__.py * fix loss.py * change __init__.py and api calling method * fix nce * fix nce * fix fluid.data * delete warpctc api document * fix loss.py * fix ctc_loss * fix test_warpctc_op.py * fix test_layers.py * fix some bug * fix conflict * fix ci bug * Empty Commit test=allcase * fix ci bug
-
- 22 11月, 2022 1 次提交
-
-
由 ustiniankw 提交于
* list112-122, test=document_fix * precommitfix, test=document_fix * list112-127, test=document_fix * fix_ResNetBasicBlock, test=document_fix * pre-commit_resnet, test=document_fix * refix, test=document * refix, test=document_fix
-
- 03 11月, 2022 1 次提交
-
-
由 Nyakku Shigure 提交于
* [CodeStyle][py2][U008] remove unnecessary args in `super()` * remove remained args * revert changes in test_pylayer_op * Revert "revert changes in test_pylayer_op" This reverts commit ff185a9ae738afac3b0264f61bde6c6b7f72e7c4. * revert some changes in example code
-
- 02 11月, 2022 1 次提交
-
-
由 Kevin吴嘉文 提交于
-
- 23 10月, 2022 1 次提交
-
-
由 Nyakku Shigure 提交于
* update config * re-blacken python code * temporarily disable date and diff_py_file * skip a format
-
- 20 10月, 2022 1 次提交
-
-
由 Tony Cao 提交于
* Fix W605 in tools folder by adding escape symbols * Fix W605 in incubate and some other folders * Fix W605 in /fluid/test folders * Update tools/analysisPyXml.py Co-authored-by: NNyakku Shigure <sigure.qaq@gmail.com> * Add some changes to manual and auto escape symbols * revert changes in transformer.py * Fix new code with W605 error: add escape symbols * revert changes in transformer.py * revert changes in transformer.py Co-authored-by: NNyakku Shigure <sigure.qaq@gmail.com>
-
- 12 10月, 2022 1 次提交
-
-
由 Shuangchi He 提交于
[CodeStyle][F401] remove unused imports in python_paddle/inference_device_profiler_text_metric_incubate_quantization_libs_audio_amp_jit. (#46762)
-
- 10 10月, 2022 1 次提交
-
-
由 carryyu 提交于
make fused_multi_transformer support dynamically set the cache_kvs' shape and support input prefix_caches. (#46777) * make fused_multi_transformer support dynamically set the cache_kvs' shape and support input prefix_caches.
-
- 23 9月, 2022 1 次提交
-
-
由 Nyakku Shigure 提交于
-
- 14 9月, 2022 1 次提交
-
-
由 Nyakku Shigure 提交于
* trim trailing whitespace * fix `.cmake-format.py` * revert npu ut changes, avoid npu ci error
-
- 26 8月, 2022 1 次提交
-
-
由 wanghuancoder 提交于
-
- 30 6月, 2022 1 次提交
-
-
由 Zhang Zheng 提交于
* Add new attr of fused_multi_transformer * fix format * add note * add in layer * fixfixfixfix
-
- 28 6月, 2022 1 次提交
-
-
由 Yuang Liu 提交于
-
- 21 6月, 2022 1 次提交
-
-
由 Yiqun Liu 提交于
-
- 17 6月, 2022 1 次提交
-
-
由 Yiqun Liu 提交于
* Support optional residual add in fused_attention and fused_feedforward. * Add checkpoint and add the check of add_residual when pre_layer_norm is false. * Add TODO and change the python api to add add_residual argument.
-
- 14 6月, 2022 1 次提交
-
-
由 Li Min 提交于
-
- 13 6月, 2022 1 次提交
-
-
由 WangXi 提交于
-
- 05 6月, 2022 1 次提交
-
-
由 Sing_chan 提交于
* use yapf to format all python file * yapf exclude two unittests file for they rely on writing and reading file, and format will break them * disable diff_py_file because too many diff files cause command following failed
-
- 01 6月, 2022 1 次提交
-
-
由 sneaxiy 提交于
* support weight transpose * add ut * add template * fix transpose error * fix transpose_comment * add api tests * add skipif * add doc
-
- 31 5月, 2022 1 次提交
-
-
由 Li Min 提交于
* replace dropout_is_test with is_test. * improve atol on a100.
-
- 30 5月, 2022 1 次提交
-
-
由 Li Min 提交于
* add fused_bias_dropout_residual_ln op and layer.
-
- 12 5月, 2022 1 次提交
-
-
由 Shuangchi He 提交于
-
- 26 4月, 2022 1 次提交
-
-
由 WangXi 提交于
-
- 25 3月, 2022 1 次提交
-
-
由 Jiabin Yang 提交于
* refactor eager flags * fix flags error when we switch from eager to dygraph * fix ci problem * fix ci * fix ci * merge develop and fix code style * merge develop and fix code style * fix op test error * fix op test error * fix op test error * fix op test error * fix op test error * merge develop
-
- 11 3月, 2022 1 次提交
-
-
由 Yuang Liu 提交于
-
- 24 2月, 2022 1 次提交
-
-
由 Leo Chen 提交于
* fix 'invalid escape sequence' * fix assert error
-
- 28 1月, 2022 1 次提交
-
-
由 zhangkaihuo 提交于
-
- 27 1月, 2022 1 次提交
-
-
由 zhangkaihuo 提交于
* fix bug: 1. atten: set the default value of attn_dropout_rate to None 2. ffn: add activation parameter * for pure fp16 * Add a SparseCsrTensor * remove unused functional * remove const * remove SetMemoberTensor * remove non_zero_nums_, the number of non zero elements of each batch can be obtained from the crows * SparseCooTensor * add SetMember * merge upstream; add SetMember * merge upstream * merge upstream; add newline at end of file * add newline at end of file * remove newline at end of file * remove newline at end of file * stash * user pten::framework::make_ddim * user pten::framework::make_ddim * merge upstream; use the latest mutable_data * merge upstream; use the latest mutable_data * return mutable dense tensor
-
- 26 11月, 2021 1 次提交
-
-
由 Li Min 提交于
* Fix bugs when bias is none for static graph for fused_attention op.
-
- 23 11月, 2021 1 次提交
-
-
由 Li Min 提交于
Add support for bias is none for fused_attention op.
-