1. 31 3月, 2021 1 次提交
    • F
      [ROCM] Add ROCm support for warpctc op (#31817) (#31971) · 3560e680
      furnace 提交于
      * bugfix for warpctc
      
      * fix warpctc commit id
      
      * fix warpctc commit id
      
      * fix warpctc commit id
      
      * fix warpctc commit id
      
      * fix warpctc commit id
      
      * fix WARPCTC_WITH_HIP invalid
      
      * Add logs to find out why can not dlopen libwarpctc.so
      
      * fix warpctc commit id
      
      * fix unit test test_warpctc_op
      
      * Optime failed log for dlopen
      
      * Optime failed log for dlopen
      
      * Delete extra changes
      
      * fix warpctc commit id
      
      * fix warpctc commit id
      
      * Add is_compiled_with_rocm for test_warpctc_op
      
      * fix warpctc commit id
      
      * Cancel optimize dlopen failed reason, move to next pr, due to it makes windows ci failed
      
      * Cancel optimize dlopen failed reason, move to next pr, due to it makes windows ci failed
      
      * Cancel optimize dlopen failed reason, move to next pr, due to it makes windows ci failed
      
      * fix code style problems
      3560e680
  2. 04 3月, 2021 1 次提交
  3. 27 9月, 2020 1 次提交
    • L
      add support to float64 input of warpctc op. (#27399) · 1501a80f
      Li Fuchen 提交于
      * add float64 input to ctc_loss
      
      * modified error message of  warpctc
      
      * update repo and tag of warpctc
      
      * add test for warpctc with float64 input
      
      * modified warpctc.cmake to make sure build always
      
      * resolved sample code bug of warpctc
      
      * add core.ops in warpctc dygraph
      
      * fix a bug of test
      1501a80f
  4. 30 4月, 2020 1 次提交
  5. 02 12月, 2019 1 次提交
  6. 14 11月, 2019 1 次提交
  7. 23 10月, 2019 1 次提交
  8. 27 9月, 2019 1 次提交
  9. 27 8月, 2019 1 次提交
    • V
      Support Tensor input with padding for warpctc op (#19322) · 482ce818
      vincentXiyu 提交于
      * support tensor input with padding for warpctc op
      
      * merge with develop
      
      * test=develop
      
      * modified python API examples test=develop
      
      * nn.py is modified for code coverage test=develop
      
      * update documents info about warpctc op in API.spec test=develop
      
      * add test_warpctc_with_padding in test_layers test=develop
      
      * add warning log for cuda_version back to warpctc_op.cc
      
      * modify API.spec for warpctc op test=develop
      
      * modify API.spec
      
      * update warpctc test to new CompiledProgram API test=develop
      
      * modify code examples for warpctc op test=develop
      
      * modify API.spec for warpctc op test=develop
      
      * modify API.spec for warpctc op test=develop
      482ce818
  10. 20 8月, 2018 1 次提交
  11. 16 8月, 2018 1 次提交
  12. 24 5月, 2018 1 次提交
  13. 11 5月, 2018 1 次提交
  14. 10 5月, 2018 1 次提交
  15. 02 5月, 2018 1 次提交
  16. 28 4月, 2018 1 次提交
  17. 09 4月, 2018 1 次提交
  18. 15 2月, 2018 1 次提交
    • Y
      Update tensor_util.h (#8422) · cfffb1a3
      Yi Wang 提交于
      * Update tensor_util.h
      
      * Update with moved TensorDesc
      
      * Fix tensur_utils.cu
      
      * Update
      
      * Update
      
      * Update
      
      * Update
      
      * Make tensor_util.cu a symbolic link
      cfffb1a3
  19. 12 2月, 2018 1 次提交
  20. 10 2月, 2018 2 次提交
  21. 13 1月, 2018 1 次提交
  22. 11 1月, 2018 2 次提交
  23. 09 1月, 2018 1 次提交
    • Y
      Port WarpCTC Operator (#5107) · b5fda272
      Yiqun Liu 提交于
      * Add Seq2BatchFunctor, which will be used in WarpCTCOp.
      
      * Implement WrapCTCFunctor and WrapCTCKernel.
      
      * Add unittest of warpctc_op.
      
      * Modify the check_output inferface in python unittest framework to allow check a subset of outputs.
      
      * Use absolute offset lod in warpctc_op and related functors.
      
      * Refine the comments of warpctc_op.
      
      * The new python unittest supports checking a subset of the outputs, so revoke the previous change.
      
      * Rename the transform from LoDTensor to Tensor with shape [max_sequence_length, num_sequences, sequence_width] to PaddingSequenceFunctor.
      
      * Update to the newest codes.
      
      * Rename the PaddingSequenceFunctor to PaddingLoDTensorFunctor and remove the computation of dimensions out of the functos.
      b5fda272