1. 19 10月, 2020 2 次提交
    • P
      Add uniform_random XPU kernel (#27846) · 74ce0397
      pangyoki 提交于
      * support uniform_random op on Baidu Kunlun
      
      * change dtype of attr shape from int to int64_t
      
      * kunlun ci, test=kunlun
      
      * new version, test=kunlun
      
      * change boost_get to BOOST_GET_CONST
      
      * change boost_get to BOOST_GET_CONST, test=kunlun
      
      * use Generator to generate random number and optimize format
      
      * run Kunlun CI, test=kunlun
      
      * add TODO, test=kunlun
      74ce0397
    • X
      Polish kunlun error (#27974) · abf4d52a
      xiaoting 提交于
      * polish error message,test=kunlun
      
      * polish error,test=kunlun
      
      * polish error,test=kunlun
      
      * polish error,test=kunlun
      abf4d52a
  2. 18 10月, 2020 1 次提交
    • L
      add cast/concat/assign xpu op (#27911) · 3e956865
      liuyuhui 提交于
      * addd
      
      * add cast_op_xpu, test=kunlun
      
      * fix bug for cast_op_xpu,test=kunlun
      
      * add concat_op_xpu, test=kunlun
      
      * slove conflicts, test=kunlun
      
      * fix bug,test=kunlun
      
      * add assign_op_xpu, test=kunlun
      
      * fix bug,test=kunlun
      
      * test=kunlun;test=develop
      
      * fix concat bug,test=kunlun
      
      * fix check_dygraph set in test_concat_op_xpu.py,test=kunlun
      
      * fix error message,test=kunlun
      Co-authored-by: Nmapingshuo <mps2012@yeah.net>
      3e956865
  3. 16 10月, 2020 6 次提交
    • G
      Incorporate cudnn_lstm into LSTM api (#27217) · fa9d3fa5
      Guo Sheng 提交于
      * Incorporate cudnn_lstm into LSTM api.
      test=develop
      
      * Make coalesce_tensor support alignment optionally.
      test=develop
      
      * Reorganize RNN apis. test=develop
      
      * Fix cudnn rnn layout conversion.
      test=develop
      
      * Add sequence_length support for RNN cudnn implement.
      Add optional init_h and init_c gradient for cudnn_lstm_op.
      test=develop
      
      * Use create_parameter for rnn cudnn impl.
      test=develop
      
      * Move `self._flat_weight = self.create_parameter()` in RNNBase to main_program.
      test=develop
      
      * Update RNN api unittest to use set_device.
      test=develop
      
      * Fix set_place for unit tests of RNN apis.
      test=develop
      
      * Fix use_align in coalesce_tensor_op.
      test=develop
      
      * Adjust RNN apis arguments according to comments.
      test=develop
      
      * Polish documents for SimpleRNN apis.
      test=develop
      
      * Refine random seed in cudnn_lstm_op.
      Expose rnn params from sublayers to RNN.
      test=develop
      
      * Fix RNN saving for jit.save.
      Refine cudnn_lstm dropout behavior.
      test=develop
      
      * Fix doc of GRU. test=develop
      
      * Use ShareDataWith to avoid copying for cudnn_lstm_op test.
      test=develop
      
      * Remove updates on cudnn_lstm temporarily.
      test=develop
      
      * Use ShareDataWith to avoid copying for cudnn_lstm_op test.
      test=develop
      
      * Refine random seed in cudnn_lstm_op.
      test=develop
      
      * Fix test_lstm by adjust ConcreteProgram buffer getter.
      test=develop
      
      * Use create_parameter instead of create_var for rnn._flat_weight for static graph usage.
      test=develop
      
      * Remove W input for cudnn_lstm to pass unused_var_check.
      test=develop
      
      * Add test_predict for RNN unit tests coverage.
      test=develop
      
      * Fix code style of rnn.
      test=develop
      
      * Fix F.rnn usage in rnn.py.
      test=develop
      fa9d3fa5
    • C
      change paddle.fluid.layers.reduce_sum to paddle.sum in sample codes (#27998) · 05fd49e9
      chentianyu03 提交于
      * change paddle.fluid.layers.reduce_sum to paddle.sum in sample codes
      
      * format codes
      05fd49e9
    • G
    • J
      Fix xpu enforce (#27978) · d330cf66
      Jack Zhou 提交于
      * test=kunlun;
      
      Add elementwise XPU OP kernel for KUNLUN core, including (but still cannot process common broadcast):
      
          * elementwise_div op
          * elementwise_max op
          * elementwise_mul op (with grad op)
          * elementwise_sub op (with grad op)
      
      * 0.05->0.01
      
      * add xpu error message description;test=kunlun
      d330cf66
    • L
      [oneDNN] Conv dilation support (#27914) · 7cb4a8b8
      lidanqing 提交于
      * conv dilated mkldnn support: forward and backward pass
      
      * add mkldnn conv_transpose dilation UT
      test=develop
      
      * remove unnecessary PADDLE_ENFORCE
      
      * add int8 and bf16 dilated conv UT
      
      * update according to reviews
      7cb4a8b8
    • M
      fix kunlun kernel of reshape op (#27988) · 64c26349
      mapingshuo 提交于
      64c26349
  4. 15 10月, 2020 8 次提交
  5. 14 10月, 2020 21 次提交
    • G
      Implement the function of OutScaleForTraining/OutScaleForInference in dygraph (#26601) · 6bbb6e7f
      guofei 提交于
      * Implement the function of OueScaleForTraining/OutScaleForInference in dygraph
      
      test=develop
      6bbb6e7f
    • C
      Remove and reorganize the alias of APIs (#27717) · d05058d2
      chentianyu03 提交于
      * modify cond while_loop to paddle.static.nn.cond
      
      * modify crop_tensor to paddle.crop
      
      * modify Variable to paddle.static.Variable
      
      * remove nn.beam_search, nn.beam_search_decode, nn.gather_tree
      
      * remove bpr_loss, center_loss, rank_loss, smooth_l1, teacher_student_sigmoid_loss, edit_distance, sampled_softmax_with_cross_entropy in nn.functional
      
      * remove apis in nn.functional.learn_rate.py
      
      * remove pool2d, pool3d, adaptive_pool2d, adaptive_pool3d in nn.functional
      
      * remove apis in nn.functional.vision
      
      * remove erf, soft_relu in nn.functional.activation
      
      * remove apis in nn.functional.extension
      
      * remove nn.functional.rnn
      
      * remove hash from nn.functional.lod
      
      * remove row_conv from nn.functional.extension
      
      * remove one_hot, pad2d, pad_constant_like from nn.functional.common
      
      * remove nn.gather_tree, nn.BilinearTensorProduct, nn.Pool2D, nn.Pad2D
      
      * remove apis from optimizer.__init
      
      * remove tensor.creation.fill_constant
      
      * remove elementwise_mul in nn.functional.common and  modify to paddle.multiply
      
      * remove  tensor.stat.reduce_mean
      
      * remove reduce_all, reduce_any in tensor.logic
      
      * remove apis in tensor.math
      
      * remove apis in tensor.__init__
      
      * remove has_inf, has_nan in tensor.search
      
      * remove apis in framework.__init__
      
      * remove apis in paddle.__init__
      
      * remove apis in nn.functional.__init__
      
      * modify removed alias apis to raw api in doc and unittests
      
      * fix remove grid_sample bug
      
      * modify removed alias apis to raw api in doc and unittests
      
      * modify removed alias apis to raw api in doc and unittests
      
      * modify removed alias apis to raw api in doc and unittests
      
      * modify removed alias apis to raw api in doc and unittests
      
      * modify removed alias apis to raw api in doc and unittests
      
      * modify removed alias apis to raw api in doc and unittests
      
      * delete alias api relastions in doc
      
      * reserve paddle.compat, paddle.sysconfig
      
      * remove unittest for paddle.reduce_all, paddle.reduce_any
      
      * modify removed alias apis to raw api in doc and unittests
      
      * recover paddle.save and paddle.load
      
      * resolve conflicts
      
      * fix sample code missing paddle.enable_static() bug
      
      * fix sample code missing paddle.enable_static() bug
      
      * fix to_string sample code error
      d05058d2
    • L
      Support setting xpu place in dygraph mode (#27909) · 9a2a4b5f
      Leo Chen 提交于
      * support setting xpu place
      
      * add ut, test=kunlun
      9a2a4b5f
    • T
      solve bug in pull_dense_worker (#27918) · 3ee6ad6e
      Thunderbrook 提交于
      * op error info
      
      * style
      
      * code format
      
      * create pin var bug
      3ee6ad6e
    • M
      Fix adam (#27778) · 263a9e97
      MRXLT 提交于
      * fix adam
      
      * fix gpu adam
      
      * fix code style
      
      * fix ut
      
      * update ut add cuda code
      263a9e97
    • D
      kunlun add op (#27890) · b0edda4d
      Double_V 提交于
      * add stack pool2d roi_align xpu op,test=kunlun
      
      * error message opt, test=kunlun
      
      * add xpu unittest,test=kunlun
      
      * skip check grad,test=kunlun
      
      * fix boostget , test=kunlun
      b0edda4d
    • J
      Add elementwise XPU OP kernel for KUNLUN core, including (but still cannot process common broadcast · c791df09
      Jack Zhou 提交于
      Add elementwise XPU OP kernel for KUNLUN core, including (but still cannot process common broadcast
      c791df09
    • W
      xpu support for fill_constant Op (#27675) · c5fcc96d
      wangchaochaohu 提交于
      c5fcc96d
    • C
      【paddle.fleet】fix sparse load (#27680) · 328cb289
      Chengmo 提交于
      * add sparse tensor load method
      328cb289
    • T
      fix paddle error informations (#27889) · cf70d5b3
      tangwei12 提交于
      cf70d5b3
    • W
      update the code for the topk message optimize · 95aa5342
      wawltor 提交于
      update the code for the topk message optimize 
      95aa5342
    • C
      Polish some error message in opeators (#27876) · 4ba977c7
      Chen Weihang 提交于
      * polish some error message
      
      * add white list
      
      * revert shell script change
      4ba977c7
    • 1
      【paddle.fleet】bug fix for parameter_recv (#27838) · a4f85074
      123malin 提交于
      * test=develop, bug fix for parameter_recv
      * test=develop, for unittest, test_fleet_rolemaker_new
      a4f85074
    • Q
      support kunlun matmul_v2 (#27910) · 2712d076
      QingshuChen 提交于
      *test=kunlun
      2712d076
    • Z
      Multi task (#26002) · 5a83496c
      zhang wenhui 提交于
      * add multitask
      
      * add multitask, test=develop
      
      * fix code style, test=develop
      
      * add partail push dense, test=develop
      
      * fix has_kay in py3, test=develop
      
      * fix, test=develop
      
      * fix, test=develop
      
      * fix, test=develop
      5a83496c
    • Z
      fix norm api doc, test=develop (#27652) · 7a58431c
      zhang wenhui 提交于
      * fix norm api doc, test=develop
      
      * fix error message, test=develop
      
      * fix api norm, test=develop
      
      * add adagrad, test=develop
      
      * fix bug, test=develop
      
      * fix bug, test=develop
      
      * add spetral_norm, test=develop
      
      * fix adagrad, test=develop
      
      * merge , test=develop
      7a58431c
    • Y
      Lookup table v2 xpu (#27888) · 3eb106da
      yinhaofeng 提交于
      * add lookup_table_v2_op_xpu, test=kunlun
      
      * add lookup_table_v2_op_xpu, test=kunlun
      
      * change some Tips ,test=kunlun
      3eb106da
    • Z
      tune backward filter algorithm for float16 (#27529) · d5cc144c
      Zhang Ting 提交于
      * use exhaustive_search for float16
      
      * tune algo only when dtype is float16
      d5cc144c
    • W
      41aad9bf
    • H
      fix error msg (#27887) · 3f2a6ab6
      hutuxian 提交于
      3f2a6ab6
    • X
      Add dropout and log_loss for kunlun (#27790) · ae01801f
      xiaoting 提交于
      * add dropout,log_loss, test=kunlun
      * fix dropout, test=kunlun
      * polish error message, test=kunlun
      * change boost::get to BOOST_GET_CONST, test=kunlun
      * fix copyright, test=kunlun
      ae01801f
  6. 13 10月, 2020 2 次提交