1. 24 10月, 2021 1 次提交
    • J
      Add viterbi decode (#35778) (#36615) · 1906c746
      Jack Zhou 提交于
      * add viterbi decode cpu kernel
      
      * add viterbi decoder api in paddle.text
      
      * add a data buffer once to avoid create many small pieces of data buffer frequently
      
      * fix viterbi max_seq_length bug
      
      * fix seq_len=1 bug
      
      * fix device context
      
      * move split out of for loop
      
      * remove INVERSE_SUB
      
      * remove 2 GET_CAST_MASK
      
      * remove 1 loop
      
      * remove Functor
      
      * add to_static deploy code
      
      * use MAX_FUNC instead of ELE_MAX
      
      * add MaxFunctor
      
      * impl max_func
      
      * remove MaxFunctor
      
      * remove cast op
      
      * use REGISTER_OP_WITHOUT_GRADIENT
      
      * add viterbi cuda kernel
      
      * add FIX_BLOCKDIM_CASE macro
      
      * add MKL add, mul; add get data mask
      
      * add arange mkl impl
      
      * add CPU Argmax
      
      * add cpu gather
      
      * use EXECUTE_MKL_ELEMENT_BINARY_OP instead of some ADD, MUL
      
      * use SameDimsBinaryOP instead of EXECUTE_MKL_ELEMENT_BINARY_OP
      
      * use SAME_DIMS_ELEMENT_BINARY_OP
      
      * add SimpleBroadcastBinaryOP
      
      * use int instead of int64_t to accelerate
      
      * optimize SimpleBroadcastBinaryOP
      
      * optimize SimpleBroadcastBinaryOP
      
      * optimize performance in both single thread and multithread situation
      
      * remove useless line
      
      * remove useless code
      
      * add CREATE_TENSOR_BUFFER macro
      
      * add INIT_REQUIRED_TENSOR macro
      
      * add comment
      
      * fix windows ci
      
      * add viterbi unittest
      
      * remove cuda add functor
      
      * remove cuda equal
      
      * remove a template function
      
      * fix windows ci
      
      * fix windows dtype
      
      * remove some template instance
      
      * remove useless header file
      
      * remove some blockdim
      
      * remove transpose impl
      
      * accelerate cpu performance on single thread situation
      
      * viterbi_decode->crf_decode
      
      * rename crf params name
      
      * add viterbi api test
      
      * remove useless import
      
      * add enable_static
      
      * use viterbi decoder
      
      * fix viterbi len=1
      
      * fix  viterbi unittest
      
      * remove useless comments
      
      * reconstruct viterbi decode
      
      * remove ADD,SUB,MUL structure
      
      * fix coverage
      
      * remove CREATE_TENSOR
      
      * add name args
      
      * crf.py->ops.py; with_start_stop_tag->include_start_end_tag
      
      * update crf_decode en docs
      
      * fix viterbi decode en docs
      
      * fix some review comments
      
      * add FIXED_BLOCK_DIM_CASE in cuda
      
      * push_back->emplace_back
      
      * crf_decode->viterbi_decode; include_start_end_tag->include_bos_eos_tag
      
      * paddle.text.ops.viterbi_decode->paddle.text.viterbi_decode
      
      * fix viterbi_decode en docs
      1906c746
  2. 22 10月, 2021 1 次提交
  3. 30 9月, 2021 1 次提交
  4. 22 9月, 2021 1 次提交
  5. 18 9月, 2021 1 次提交
    • J
      [oneDNN] Disable caching of Reorder operation (#35664) · e4c2a854
      Jacek Czaja 提交于
      * - REorder disabling caching
      
      * - compilation fix
      
      * - another compilation fix
      
      * - another compilation fix
      
      * - compilation fix
      
      * - Fix
      
      * - yet another compilation fix
      
      * - suppresingly another compilation fix
      
      * - lint
      
      * - fix after review
      
      * - fix
      e4c2a854
  6. 15 9月, 2021 1 次提交
  7. 14 9月, 2021 2 次提交
  8. 13 9月, 2021 2 次提交
  9. 08 9月, 2021 1 次提交
  10. 07 9月, 2021 1 次提交
  11. 06 9月, 2021 1 次提交
  12. 03 9月, 2021 2 次提交
  13. 02 9月, 2021 1 次提交
  14. 31 8月, 2021 1 次提交
  15. 27 8月, 2021 1 次提交
    • B
      add elementwise max grad op for npu (#34862) · 5310ceab
      baoachun 提交于
      * add elementwise max grad op for npu
      
      * add elementwise max grad op for npu
      
      * add elementwise max grad op for npu
      
      * add elementwise max grad op for npu
      
      * add elementwise max grad op for npu
      5310ceab
  16. 26 8月, 2021 1 次提交
    • J
      [oneDNN] disable caching oneDNN primitives in matmul v2, Reduce grad and... · 31f0221f
      Jacek Czaja 提交于
      [oneDNN] disable caching oneDNN primitives in  matmul v2, Reduce grad and elementwise_add grad, expand_v2 (#35132)
      
      * - grad caching disabled of matmul_v1
      
      - compilation fix
      
      - compilation fix
      
      * - reduction removed
      
      * - Matmul v2 disabled caching
      
      * Draft of further changes
      
      * - workaround for reducegrad
      
      * - fixes to UT
      
      * - fix to compilation
      
      * - another fix
      
      * - fix
      31f0221f
  17. 25 8月, 2021 2 次提交
  18. 22 8月, 2021 1 次提交
  19. 16 8月, 2021 1 次提交
    • J
      [oneDNN] Fix to 34554 (same as previous PR but should build with GPU) (#34859) · 9cb65653
      Jacek Czaja 提交于
      * - Added softmax without caching
      
      * - Binary is no longer manually cached
      
      * - Activation onednn caching removed
      
      * - Removed manual caching of activation
      
      * - modified UT
      
      * - fix
      
      * - fix
      
      * - fixes to building
      
      * - fix
      
      * - fix
      
      * - fix to UT
      
      * - Faulty UT workaround
      
      * - approval workaround
      
      * - Fixes after review
      
      * - compilation fixes
      
      * - more lint fixes
      
      * - more fixes after review
      
      * - fixes after another round of review
      
      * - hopefully compilation fix
      
      - compilation fix
      9cb65653
  20. 12 8月, 2021 1 次提交
  21. 11 8月, 2021 2 次提交
    • J
      [oneDNN] Fix to issue #34554 (#34623) · 0a5c99e8
      Jacek Czaja 提交于
      * - Added softmax without caching
      
      * - Binary is no longer manually cached
      
      * - Activation onednn caching removed
      
      * - Removed manual caching of activation
      
      * - modified UT
      
      * - fix
      
      * - fix
      
      * - fixes to building
      
      * - fix
      
      * - fix
      
      * - fix to UT
      
      * - Faulty UT workaround
      
      * - approval workaround
      
      * - Fixes after review
      
      * - compilation fixes
      
      * - more lint fixes
      
      * - more fixes after review
      
      * - fixes after another round of review
      0a5c99e8
    • A
      45af4f2a
  22. 09 8月, 2021 1 次提交
  23. 05 8月, 2021 1 次提交
  24. 07 7月, 2021 1 次提交
  25. 05 7月, 2021 2 次提交
  26. 24 6月, 2021 1 次提交
  27. 23 6月, 2021 1 次提交
  28. 12 6月, 2021 1 次提交
  29. 04 6月, 2021 1 次提交
  30. 02 6月, 2021 2 次提交
  31. 26 5月, 2021 1 次提交
    • L
      [NPU] refine NpuOpRunner (#32869) · 8259d9bf
      Leo Chen 提交于
      * refine ~npuOpRunner
      
      * implement destructor and forbid copy
      
      * use reference to avoid copy
      
      * use const reference
      
      * relax adam precision
      
      * fix top_k
      8259d9bf
  32. 25 5月, 2021 1 次提交
    • C
      modify complex template for elementwise ops (#33071) · dbc08d69
      chentianyu03 提交于
      * modify complex template for elementwise ops
      
      * modify mul, div grad struct
      
      * add complex template for CudaShuffleDownSync CudaShuffleXorSync funcs and fix the bug when delete cuda<9000
      
      * fix shuffle func args bug
      
      * fix shuffle func args bug
      
      * fix shuffle func args bug
      dbc08d69
  33. 24 5月, 2021 1 次提交