1. 27 9月, 2019 7 次提交
  2. 26 9月, 2019 15 次提交
  3. 25 9月, 2019 13 次提交
    • Z
      Add AdadeltaOptimizer doc (#19875) · 4a5ce4fe
      Zeng Jinle 提交于
      * add AdadeltaOptimizer doc, test=develop
      
      * refine doc,test=develop
      
      * folllow lanxiang's comments, test=develop, test=document_fix
      4a5ce4fe
    • Z
      Expose set_gradient_clip API (#19869) · 7912e6ca
      Zeng Jinle 提交于
      * expose set_gradient_clip, test=develop, test=document_preview, test=preview
      
      * expose gradient clip, test=develop, test=document_fix
      
      * refine doc, test=develop
      
      * follow lanxiang's comments, test=develop, test=document_fix
      7912e6ca
    • C
      refine deformable roi pooling doc (#19944) · 0099e549
      chengjuntao 提交于
      * refine doc, test=develop, test=document_preview
      0099e549
    • Z
      add kernel for fill_op, test=develop (#19719) · b1bb2384
      zhongpu 提交于
      * add kernel for fill_op, test=develop
      
      * modify PADDLE_ENFORCE to PADDLE_ENFORCE_EQ, test=develop
      
      * add op test for fill_op, test=develop
      
      * REGISTER COP CUDA KERNEL, test=develop
      
      * update test_fill_op.py, test=develop
      
      * change FillConstantOpVarTypeInference to FillOpVarTypeInference, test=develop
      
      * fix op test, test=develop
      
      * add head file, test=develop
      b1bb2384
    • W
      add support tensor and tensorlist for strided_slice OP (#19929) · 382d099d
      wangchaochaohu 提交于
      * add support tensor and tensorlist for strided_slice OP test=develop
      
      * fix the commnet test=develop
      
      * fix test=develop
      
      * fix the bug test=develop
      
      * delete log test=develop
      
      * fix API.spec test=develop
      
      * fix test=develop
      382d099d
    • L
      Fix ssdloss num and batch norm format and conv2d (#19754) · fe218df3
      lvmengsi 提交于
      * update API.spec
      fe218df3
    • L
      Fix OpTest of bn (#19062) · 619a241b
      lvmengsi 提交于
      * fix bn
      619a241b
    • S
      Avoid treating broadcast as initialization operation (#19857) · 5920d69d
      ShenLiang 提交于
      * treat broadcast as non-initial, test=develop
      
      * rename the class name
      
      * rename the class name, test=develop
      5920d69d
    • B
      add support of matmul with multiple head even different width and height (#19708) · c670058a
      Bob Zhu 提交于
      * add support of matmul with multiple head even different width and height
      
      Original matmul with multiple head supports only the mat_a.width == mat_b.height,
      in that case, mat_b will be horizontally split. In this patch, we extend the
      support when mat_a.width != mat_b.height but mat_a.width/head_number == mat_b.height,
      in this case, mab_b will be vertically split.
      
      One example is A is [3, 8], B is [2, 16], head_number is 4. In this
      case, A will be split as [3, 2], B will be (vertically) split as
      [2, 4]. The final result will be 4 matrix of 4 matrix of [3,4], i.e. [3, 16]
      
      test=develop
      
      * add support of matmul with multiple head even different width and height
      
      Original matmul with multiple head supports only the mat_a.width == mat_b.height,
      in that case, mat_b will be horizontally split. In this patch, we extend the
      support when mat_a.width != mat_b.height but mat_a.width/head_number == mat_b.height,
      in this case, mab_b will be vertically split.
      
      One example is A is [3, 8], B is [2, 16], head_number is 4. In this
      case, A will be split as [3, 2], B will be (vertically) split as
      [2, 4]. The final result will be 4 matrix of 4 matrix of [3,4], i.e. [3, 16]
      
      test=develop
      
      * refactor the code of matmul with multiple head even different width and height
      
      test=develop
      c670058a
    • L
      refine ctc align op with padding (#19926) · 6884dc80
      Liufang Sang 提交于
      * refine ctc align op with padding 
      * refine api sample code
      6884dc80
    • T
      add input type and dtype check for softmax_op (#19975) · 65a02fc1
      Tao Luo 提交于
      * add input type and dtype check for softmax_op
      
      test=develop
      
      * refine error message
      
      test=develop
      65a02fc1
    • W
      Add support for new QAT models (#18970) · 4286a627
      Wojciech Uss 提交于
      * Add support for new QAT models
      
      test=develop
      Co-Authored-By: NMichał Gallus <michal.gallus@intel.com>
      Co-Authored-By: NWojciech Uss <wojciech.uss@intel.com>
      
      * fixed fps results
      
      test=develop
      
      * fix top5 accuracy drop problem
      
      * updated for new QAT models
      
      * skip quantizing average pooling - dirty but working
      
      * add missing pass
      
      * added missing conv+brelu fuse pass
      
      * removed a call to non-existent pass
      
      test=develop
      
      * renamed pass
      
      test=develop
      
      * Adjust finding pooling scale to newest QAT models
      
      * Remove unnecessary code from quantization_mkldnn_pass
      
      * Copy Pooling input scale to output scale in QAT
      
      * Refactor & remove unused code in QAT
      
      * Incorporate fp32 FC into QAT
      
      test=develop
      
      * Enable graph drawing with debug flag
      
      test=develop
      
      * Add tests for QATv2
      
      * Fix paths for QATv2 models
      
      test=develop
      
      * Add option to save transformed int8 qat model
      
      test=develop
      
      * Remove redundant lines from qat mkldnn pass
      
      test=develop
      
      * Delegate disablement of avg pooling to qat
      
      test=develop
      
      * fix CI bug, test=develop
      
      * Follow Wangzhen's Review, test=develop
      
      * Update API.spec
      
      test=develop
      
      * Name False in (is_unsigned, TensorScale) tuple
      
      test=develop
      4286a627
    • A
      Removing length dims constraints of seq_pad and seq_unpad (#19497) · 99a9615a
      Aurelius84 提交于
      * Removing last dims constraints of seq_pad and seq_unpad test=develop
      
      * fix test_layer api code test=develop
      
      * fix sequence_pad_op.cc conflict test=develop
      
      * remove test_analyzer_mm_dnn test=develop
      
      * fix vectorize bug test=develop
      
      * fix vectorize<int> test=develop
      99a9615a
  4. 24 9月, 2019 5 次提交
    • Y
      update en document of shard_index_op (#19963) · 2efdf0ef
      Yi Liu 提交于
      test=develop
      test=document_fix
      2efdf0ef
    • J
      add optimizer:dpsgd,test=develop (#19915) · 766bd529
      jhjiangcs 提交于
      766bd529
    • Y
      Add float16 support to `sync_batch_norm_op` (#19681) · ebff68fa
      Yang Zhang 提交于
      * Add float16 support to `sync_batch_norm_op`
      
      test=develop
      
      * Add test for sync_bn with FP16 input
      
      test=develop
      ebff68fa
    • A
      Remove constraint that last dimension is forced to be 1 by adding lookup_table_v2 (#19735) · 039b9710
      Aurelius84 提交于
      * Remove constraint that last dimension is forced to be 1 by add
      lookup_table_v2 test=develop
      
      * modify into PADDLE_ENFORCE_CUDA_SUCCESS test=develop
      
      * Revert "modify into PADDLE_ENFORCE_CUDA_SUCCESS test=develop"
      
      This reverts commit 8a960bfc61e51aa27c3c529df8fb90b93ebd19f9.
      
      * move api into fluid.embedding test=develop
      
      * fix example code test=develop
      
      * move one_hot into fluid.one_hot
      
      * modify api.spec test=develop
      
      * fix loss shape test=develop
      039b9710
    • W
      [PaddleSlim] Enhence compressor api in PaddleSlim (#19894) · bdb3e376
      whs 提交于
      
      1. Support customize eval function instead of eval program.
      2. Fix loading checkpoint in quantization strategy.
      3. Support saving eval model when saving a checkpoint.
      4. Fix decoder of loading context in PaddleSlim.
      5. Fix restoring from the checkpoint of uniform prune strategy.
      6. Support saving eval model and infer model during training.
      7. Add ‘unitest’ for saving eval model, saving infer model and uniform pruning restoring from the checkpoint.
      8. Fix pruning of depthwise_conv_grad op by updating the groups.
      bdb3e376