- 13 5月, 2019 1 次提交
-
-
由 Kaipeng Deng 提交于
* add double grad for square. test=develop * formax code. test=develop * fix for grad sum. test=develop * refine shape. test=develop * refine extract. test=develop
-
- 12 5月, 2019 1 次提交
-
-
由 chengduo 提交于
* reset drop local scope counter test=develop
-
- 10 5月, 2019 4 次提交
-
-
由 zhoukunsheng 提交于
-
由 zhoukunsheng 提交于
-
由 qingqing01 提交于
* Add conv2d_grad_grad_op * Extracte the cuDNN conv algo searching code in conv_cudnn_helper.h. - Now use it in conv2d_grad_grad. - Will simply the searching code in conv2d and conv2d_grad in next PR. * Enhance and fix bug in unit testing of gradient_checker. * Support to fetch empty variables,return None in Python.
-
由 wopeizl 提交于
* rename the default version from '0.0.0' to 'latest'
-
- 08 5月, 2019 1 次提交
-
-
由 baojun 提交于
* added lrn op test=develop * Added CreateConstant method test=develop * avoid duplicates test=develop
-
- 07 5月, 2019 5 次提交
-
-
由 Zeng Jinle 提交于
* add use_cuda to inplace pass,test=develop * add test softmax_with_xe_inplace test,test=develop * fix potential inplace bug test=develop * add more skip vars in mem opt pass,test=develop * follow comment,test=develop * follow comments,move duplicate out arg check to program->graph,test=develop
-
由 baojun 提交于
-
由 Tao Luo 提交于
* remove unused FLAGS_warpctc_dir test=develop * remove FLAGS_warpctc_dir test=develop
-
由 Kaipeng Deng 提交于
* add attr axis infershape. test=develop * add CUDA kernel. test=develop * fix unittest. test=develop * fix unittest for soft_label. test=develop * fix fp16 unittest. test=develop * remove comment code. test=develop * refine test for axis. test=develop * add python api. test=develop * fix doc. test=develop * fix fp16 unittest. test=develop * fix ngraph test. test=develop * fix ENFORCE for test_imperative_transformer. test=develop * fit for ngraph test. test=develop * fix after rebase develop. test=develop * fix doc. test=develop * fix API.spec. test=develop * fix test_layers. test=develop * fix format. test=develop
-
由 Zhen Wang 提交于
* Add MovingAverageAbsMaxScale operator which is only used for calculating the quantization scale. * test=develop * change the output into inplace. test=develop * Revert "test=develop" This reverts commit 696cf62699ba1e1c98f61f7345ac7060010eb29a. * Revert "change the output into inplace. test=develop" This reverts commit a19acd20f07eee82622701a3015e6e9c073a5e0b. * test=develop. * update the MovingAverageAbsMaxScaleOp test. test=develop
-
- 06 5月, 2019 2 次提交
-
-
由 jerrywgz 提交于
* fix distribute fpn proposals, test=develop
-
由 Zeng Jinle 提交于
* add use_cuda to inplace pass,test=develop * add test softmax_with_xe_inplace test,test=develop
-
- 05 5月, 2019 3 次提交
-
-
由 jerrywgz 提交于
* enhance_concat, test=develop
-
由 wopeizl 提交于
-
由 tianshuo78520a 提交于
* test=develop * test=deelop
-
- 01 5月, 2019 1 次提交
-
-
由 guru4elephant 提交于
* remove async executor python api test=develop * remove test_async_executor.py add executor train_from_dataset demo test=develop * fix import bug test=develop
-
- 30 4月, 2019 3 次提交
-
-
由 Zeng Jinle 提交于
* fix mem leak when converting Tensor to numpy array test=develop * remove unused unittest,test=develop * follow comments, test=develop * fix dygraph bug,test=develop
-
由 Zeng Jinle 提交于
* fix op graph view test=develop * rewrite inplace pass and fix reference count pass bug test=develop * fix unittest failed test=develop * follow comments, test=develop
-
由 xiaoting 提交于
* polish the label_smooth test=develop * polish code test=develop
-
- 29 4月, 2019 2 次提交
-
-
由 tangwei12 提交于
cvm without LoD.
-
由 Jiancheng Li 提交于
-
- 28 4月, 2019 1 次提交
-
-
由 Zeng Jinle 提交于
* refine_dropout_mem,test=develop * # This is a combination of 14 commits. # The first commit's message is: remove ut test_dist_word2vec in mac ci, will fix it in private, test=develop (#17066) # This is the 2nd commit message: Fleet unify distributed training (#16791) * implement distributed transpiler with fleet # This is the 3rd commit message: ParallelDyGraph with GPU collective mode (#16827) implement dygraph.parallel.DataParallel to hook reduce op. # This is the 4th commit message: Init mixed precision training interface (#16856) * Init mixed precision training interface * Add fp16 test script test=develop * All initializers support float16 test=develop * Code cleanup & add more code annotations test=develop * Update API spec test=develop * Add usage example in doc test=develop # This is the 5th commit message: fix reference_count_pass,test=develop (#17060) test=develop # This is the 6th commit message: Speedup roi_perspective_transform op by caching the information of linear interpolation in forward (#17090) * Cache the information of linear interpolation in forward and use it in backward. test=develop * Fix cuda kernel. test=develop # This is the 7th commit message: remove unnecessary prepare_data (#17080) test=develop # This is the 8th commit message: fix interpolate cu. test=develop (#17101) # This is the 9th commit message: test=develop, double backward leaky_relu (#17067) backward of backward: leaky_relu # This is the 10th commit message: fix fuse optimizer ops (#17102) test=develop # This is the 11th commit message: truncated_gaussian_random supported in distributed training, test=develop (#17091) # This is the 12th commit message: Detailed coordinate description for yolov3 loss (#17007) * Detailed coordinate description for yolov3 loss test=develop * modified api.spec test=develop * modified loss name * fix api.spec test=develop * polish description test=develop * modified api.spec test=develop # This is the 13th commit message: fix test_weight_decay (#17109) test=develop # This is the 14th commit message: Path flag (#17105) * fix python/paddle/fluid/__init__.py detecting problems
-
- 27 4月, 2019 1 次提交
-
-
由 chengduo 提交于
test=develop
-
- 26 4月, 2019 2 次提交
-
-
由 ceci3 提交于
backward of backward: leaky_relu
-
由 Kaipeng Deng 提交于
-
- 25 4月, 2019 5 次提交
-
-
由 whs 提交于
Speedup roi_perspective_transform op by caching the information of linear interpolation in forward (#17090) * Cache the information of linear interpolation in forward and use it in backward. test=develop * Fix cuda kernel. test=develop
-
由 Yibing Liu 提交于
* Init mixed precision training interface * Add fp16 test script test=develop * All initializers support float16 test=develop * Code cleanup & add more code annotations test=develop * Update API spec test=develop * Add usage example in doc test=develop
-
由 Yan Xu 提交于
implement dygraph.parallel.DataParallel to hook reduce op.
-
由 tangwei12 提交于
* implement distributed transpiler with fleet
-
由 tangwei12 提交于
-
- 24 4月, 2019 1 次提交
-
-
由 guomingz 提交于
* resolve #17057 Fixed the bug that fuse_relu/fuse_residual option couldn't be passed to class TestConv2dInt8Op. test=develop * Fix the bug of test_conv2d_int8_mkldnn case which raised by improper parameter passing. test=develop
-
- 23 4月, 2019 3 次提交
-
-
由 chengduo 提交于
* Add fuse momenutum ops
-
由 chengduo 提交于
test=develop
-
由 qingqing01 提交于
Support backward of backward for Relu and add a new gradient checker by comparing theoretical and numerical Jacobian. (#16862) * Support backward of backward and a new gradient checker * Rename decorators.py to decorator_helper.py, since Python on Windows CI has decorators package. 1. Add ReluDoubleGradMaker when register relu_grad. 2. Add a new gradient checker by comparing theoretical and numerical Jacobian. Check double gradients by double_grad_check.
-
- 22 4月, 2019 4 次提交
-
-
由 Zeng Jinle 提交于
* move gc test to op_test test=develop * Revert "move gc test to op_test" This reverts commit cf15da65. * enable gc test in some ops test=develop
-
由 chengduo 提交于
* fix ramdom fail test=develop
-
由 Tao Luo 提交于
test=develop
-
由 wopeizl 提交于
* add parallel build script to ci test=develop * 1. classify the test case as single card/two cards/multiple cards type 2. run test case according to the run type
-