1. 24 4月, 2020 1 次提交
  2. 21 4月, 2020 1 次提交
  3. 10 4月, 2020 1 次提交
  4. 26 3月, 2020 1 次提交
    • Z
      [Paddle-TRT]: Ernie Dynamic shape support. (#23138) · 430b0099
      Zhaolong Xing 提交于
      * add dynamic plugin support.
      test=develop
      
      * change emb eltwise layernorm to math function
      test=develop
      
      * add emb eltwise layernorm
      test=develop
      
      * can run dynamic shape ernie
      test=develop
      
      * fix ci
      test=develop
      
      * add ut for trt ernie dynamic
      
      test=develop
      
      * refine dynamic shape c++ interface.
      test=develop
      
      * fix comments
      test=develop
      
      * fix comments
      test=develop
      430b0099
  5. 11 3月, 2020 1 次提交
  6. 28 2月, 2020 1 次提交
  7. 23 2月, 2020 1 次提交
  8. 10 2月, 2020 1 次提交
  9. 07 2月, 2020 1 次提交
  10. 06 2月, 2020 1 次提交
    • Y
      Correct the use of DeviceContext in unittest sequence_pooling_test and... · 44b45b9f
      Yiqun Liu 提交于
      Correct the use of DeviceContext in unittest sequence_pooling_test and sequence_padding_test (#22456)
      
      * Add log in memory::Copy for debug purpose.
      
      * Change to use context in DeviceContextPool directly in sequence_pooling_test, instead to new one.
      
      * Change to use context in DeviceContextPool directly in sequence_padding_test, instead to new one.
      test=develop
      
      * Change the type of second_dim from size_t to int64_t.
      test=develop
      44b45b9f
  11. 19 1月, 2020 1 次提交
  12. 07 1月, 2020 1 次提交
  13. 04 1月, 2020 1 次提交
  14. 23 12月, 2019 1 次提交
  15. 11 12月, 2019 1 次提交
  16. 02 12月, 2019 1 次提交
  17. 28 11月, 2019 1 次提交
  18. 27 11月, 2019 1 次提交
  19. 26 11月, 2019 1 次提交
    • G
      Add fc padding to improve mkl GEMM's performance when N and K are multiple of 128. (#20972) · 234060f8
      GaoWei8 提交于
      * Add fc padding to solve mkl performance
      test=develop
      
      * fix gpu pass and error information
      test=develop
      
      * fix fc_fuse_pass_test
      test=develop
      
      * fix error information
      test=develop
      
      * fix error information
      test=develop
      
      * fix name and add fc op padding test
      test=develop
      
      * fix attributes
      test=develop
      
      * optimize fc padding
      test=develop
      
      * fix test
      test=develop
      234060f8
  20. 22 11月, 2019 1 次提交
    • L
      add dequantize_abs_max op and modify lookup_table op (#20899) · f0b15184
      Liufang Sang 提交于
      * add int8 kernel to lookup_table op and add dequantize op test=develop
      
      * change paddle_enforce to paddle_enforce_eq test=develop
      
      * change copyright and change some not suitable code test=develop
      
      * remove debug log test=develop
      
      * replace GetInputType with IndicateVarDataType test=develop
      
      * fix EmptyGradMaker test=develop
      
      * fix diff between cpu and gpu test=develop
      
      * use memcopy when int8_t test=develop
      f0b15184
  21. 14 11月, 2019 1 次提交
  22. 12 11月, 2019 1 次提交
    • L
      fix the computation for dx (grad for x) for prelu operation. (#20949) · e249d9a3
      lilong12 提交于
      * set the default value of alpha for prelu to 0.25, test=develop
      
      * add the call to __syncthreads(), test=develop
      
      * fix the implementation of cpu prelu, test=develop
      
      * repair the implementation of element mode prelu, test=develop
      
      * modify test_prelu_op.py, test=develop
      e249d9a3
  23. 08 11月, 2019 1 次提交
    • C
      Add dependency for error_codes.proto (#21084) · 2f27b103
      Chen Weihang 提交于
      * fix activation_functions deps, test=develop, test=document_fix
      
      * add error_codes_proto deps, test=develop, test=document_fix
      
      * try delete enforce.h, test=develop, test=document_fix
      2f27b103
  24. 05 11月, 2019 2 次提交
  25. 01 11月, 2019 1 次提交
  26. 31 10月, 2019 2 次提交
  27. 30 10月, 2019 1 次提交
  28. 28 10月, 2019 1 次提交
  29. 23 10月, 2019 1 次提交
  30. 16 10月, 2019 1 次提交
  31. 13 10月, 2019 1 次提交
  32. 09 10月, 2019 1 次提交
  33. 07 10月, 2019 1 次提交
  34. 30 9月, 2019 1 次提交
  35. 29 9月, 2019 1 次提交
    • L
      fix conv2d and conv3d: (#20042) · 3aa331d9
      liym27 提交于
      1.support asymmetric padding;
          2.support padding algorithm:"SAME" and "VALID";
          3.support channel_last: data_format NHWC and NDHWC;
          4.change doc of python API and c++;
      
          test=develop, test=document_preview
      3aa331d9
  36. 28 9月, 2019 1 次提交
    • L
      fix pool2d pool3d,support asymmetric padding and channel_last (#19739) · 24010472
      liym27 提交于
      * fix pool2d pool3d:
      1. support asymmetric padding;
      2. support padding algorithm:"SAME" and "VALID";
      3. support channel_last: data_format NHWC and NDHWC;
      4. support inferring shape when input with negative dims in compile time;
      5. change doc of python API and c++;
      6. fix bug in cuda kernel when Attr(adaptive) is true.
      
      test=develop,test=document_preview
      
      * fix 'tensors' to 'Tensors'. test=develop,test=document_preview
      
      * add test for converage ValueError.test=develop,test=document_preview
      
      * resolve conflict in test_pool2d. test=develop
      24010472
  37. 27 9月, 2019 1 次提交
  38. 25 9月, 2019 1 次提交
    • B
      add support of matmul with multiple head even different width and height (#19708) · c670058a
      Bob Zhu 提交于
      * add support of matmul with multiple head even different width and height
      
      Original matmul with multiple head supports only the mat_a.width == mat_b.height,
      in that case, mat_b will be horizontally split. In this patch, we extend the
      support when mat_a.width != mat_b.height but mat_a.width/head_number == mat_b.height,
      in this case, mab_b will be vertically split.
      
      One example is A is [3, 8], B is [2, 16], head_number is 4. In this
      case, A will be split as [3, 2], B will be (vertically) split as
      [2, 4]. The final result will be 4 matrix of 4 matrix of [3,4], i.e. [3, 16]
      
      test=develop
      
      * add support of matmul with multiple head even different width and height
      
      Original matmul with multiple head supports only the mat_a.width == mat_b.height,
      in that case, mat_b will be horizontally split. In this patch, we extend the
      support when mat_a.width != mat_b.height but mat_a.width/head_number == mat_b.height,
      in this case, mab_b will be vertically split.
      
      One example is A is [3, 8], B is [2, 16], head_number is 4. In this
      case, A will be split as [3, 2], B will be (vertically) split as
      [2, 4]. The final result will be 4 matrix of 4 matrix of [3,4], i.e. [3, 16]
      
      test=develop
      
      * refactor the code of matmul with multiple head even different width and height
      
      test=develop
      c670058a