- 28 9月, 2019 5 次提交
-
-
由 Youwei Song 提交于
* opt FC * opt rest of dygraph.nn * new param shape check and unittest * add kwargs for Layer * add new set_value api * use property decorator * update API.spec, test=develop * use UserList, separate gettersetters, test=develop * update test_custom_layer_with_kwargs, test=develop * fix UserList compatibility, test=develop * fix UserList compatibility, test=develop * keep FC._w, test=develop * add unittests, Conv3D bug fix, test=develop * clean code, test=develop * fix dygraph guard in unittest, test=develop * add property setters, remove unused param in tracer, test=develop * tracer none check, test=develop * merge, test=develop * refine, test=develop * bug fix in prelu and conv3d_transpose, test=develop * rm __set__, test=develop * set tensor value instead of assign op * fix property setter call, test=develop * fix api.spec, test=develop * fix doc sample, test=develop
-
由 bingyanghuang 提交于
* Follow Wangzhen's comment in PR 18970, test=develop * Review comments, test=develop * Leave fake quantization around mul test=develop * Replace Fake with Real Quantized Mul test=develop * Fix bug in quantize placement pass Nodes in the graph now have checked type instead of node name when they are to be marked for quantization test=develop
-
由 hong 提交于
* add parameter shape check when load parameter from file; test=develop * fix test error; test=develop * add wrong shape check; test=develop * remove useless code; test=develop * add testcase setup * add teardown, remove temp model_path; test=develop * add clean process; test=develop
-
由 gongweibao 提交于
-
由 Bai Yifan 提交于
-
- 27 9月, 2019 11 次提交
-
-
由 石晓伟 提交于
* update operator compatible info, test=develop * revert cmake/version.cmake, test=develop * add unit_tests and fix bugs, test=develop * update ../paddle/fluid/framework/framework.proto, test=develop * fix bug of paddle/fluid/inference/api/analysis_predictor.cc, test=develop * update paddle/fluid/framework/version_test.cc, test=develop * add comments and rename interfaces, test=develop
-
由 Zeng Jinle 提交于
* fix en docs of apis, test=develop, test=document_fix * follow chunwei's comments, test=develop
-
由 Jiabin Yang 提交于
* test=develop, fix docker with paddle nccl problem * test=develop, Add Variable api and refine dygraph related API * test=develop, Add Variable api and refine dygraph related API * test=develop, refine test for new api and error info * test=develop, refine error info and test_layers * test=develop, add API.spec * test=devleop, fix to_string python2 and python3 compat error and refien doc * test=devleop, add API spec * test=devleop, update API spec * test=devleop, update API spec * test=develop, invoke ci * test=develop, fix example code * test=develop, update API spec * test=develop, add compat test and fix inplace campat dict error
-
由 Kaipeng Deng 提交于
* fix pool infershape. test=develop * fix unittest converage. test=develop * fix format. test=develop
-
由 chengduo 提交于
* make pad and split support fp16 test=develop
-
由 tangwei12 提交于
* add a base class for the Communicator * add AsyncCommunicator Impl for async distributed training
-
由 WangXi 提交于
Refine document of DGCMomentumOptimizer
-
由 danleifeng 提交于
Polish English docs of elementwise_add/sub/mul/div
-
由 Chen Weihang 提交于
* shape and optimize paddle error message stack, test=develop * limit exception type & add unittest, test=develop * fix multi-platform problem, test=develop * fix related unnitest failed, test=develop * add doc & fix unittest errors, test=develop * fix function name error, test=develop * update tensor test exception msg compare, test=develop * remove unittest on win32, the dir format is different, test=develop * remove useless package, test=develop * add paddle enforce handler unittest, test=develop * add exception checkout, test=develop * fix coverage failed, test=develop * fix op registry test failed, test=develop * refactor whole pr, test=develop * remove test in CMakelist, test=develop * fix coverage, test=develop
-
由 zhaoyuchen2018 提交于
test=develop Signed-off-by: Nzhaoyuchen <zhaoyuchen01@baidu.com>
-
由 Huihuang Zheng 提交于
Set output type LoDTensor only After code experiment, I found data doens't support other type
-
- 26 9月, 2019 15 次提交
-
-
由 gongweibao 提交于
-
由 123malin 提交于
* fix DistributeTranspilerConfig document, test=develop
-
由 wangchaochaohu 提交于
-
由 whs 提交于
* Make PaddleSlim support PyReader. * Fix unittest of sensitive pruning. * Add some assert.
-
由 liuwei1031 提交于
* impove error message when passing ndarray with object dtype * imporve message format * change assert to raise TypeError * remind user how to locate the irregular data instead of printing * add unittest for input array type check
-
由 joanna.wozna.intel 提交于
* Fix conv2d+dequantize squash for residual fusion test=develop * Correct int8 input test=develop * Add if exclude or include padding in pool2d mkldnn test=develop
-
由 Aurelius84 提交于
* x.dims == y.dims test=develop * refine comment
-
由 Yang Zhang 提交于
* Expose `mutable_data` as python binding test=develop * Add test for device pointer binding test=develop * Make test compatible with python 2
-
由 Aurelius84 提交于
* fix input shape check test=develop * move PADDLE_ENFORCE test=develop
-
由 chengduo 提交于
Add dtype for coalesce_tensor_op
-
由 gongweibao 提交于
Polish elementwise max min pow document to add more examples
-
由 mapingshuo 提交于
* fix doc of apply_optimize test=document_fix test=document_preview * modify doc of backward test=develop test=document_fix * modify document hash test=develop test=document_preview
-
由 Huihuang Zheng 提交于
The new "fluid.data" changes old "fluid.layers.data": 1. Add shape and dtype check. 2. Remove "append_batch_size" parameter. We won't offer this in the new data layer because other deep learning platforms don't have this kind of data layer pre-processing. It may confuse users. 3. Remove "stop gradient" parameter because the data layer doesn't do back-propagation TODO: Now data layer feeded by executor is checked, will we want to check the feed data of readers in the future?
-
由 Zeng Jinle 提交于
-
由 qingqing01 提交于
-
- 25 9月, 2019 9 次提交
-
-
由 Zeng Jinle 提交于
* add AdadeltaOptimizer doc, test=develop * refine doc,test=develop * folllow lanxiang's comments, test=develop, test=document_fix
-
由 Zeng Jinle 提交于
* expose set_gradient_clip, test=develop, test=document_preview, test=preview * expose gradient clip, test=develop, test=document_fix * refine doc, test=develop * follow lanxiang's comments, test=develop, test=document_fix
-
由 chengjuntao 提交于
* refine doc, test=develop, test=document_preview
-
由 zhongpu 提交于
* add kernel for fill_op, test=develop * modify PADDLE_ENFORCE to PADDLE_ENFORCE_EQ, test=develop * add op test for fill_op, test=develop * REGISTER COP CUDA KERNEL, test=develop * update test_fill_op.py, test=develop * change FillConstantOpVarTypeInference to FillOpVarTypeInference, test=develop * fix op test, test=develop * add head file, test=develop
-
由 wangchaochaohu 提交于
* add support tensor and tensorlist for strided_slice OP test=develop * fix the commnet test=develop * fix test=develop * fix the bug test=develop * delete log test=develop * fix API.spec test=develop * fix test=develop
-
由 lvmengsi 提交于
* update API.spec
-
由 lvmengsi 提交于
* fix bn
-
由 ShenLiang 提交于
* treat broadcast as non-initial, test=develop * rename the class name * rename the class name, test=develop
-
由 Bob Zhu 提交于
* add support of matmul with multiple head even different width and height Original matmul with multiple head supports only the mat_a.width == mat_b.height, in that case, mat_b will be horizontally split. In this patch, we extend the support when mat_a.width != mat_b.height but mat_a.width/head_number == mat_b.height, in this case, mab_b will be vertically split. One example is A is [3, 8], B is [2, 16], head_number is 4. In this case, A will be split as [3, 2], B will be (vertically) split as [2, 4]. The final result will be 4 matrix of 4 matrix of [3,4], i.e. [3, 16] test=develop * add support of matmul with multiple head even different width and height Original matmul with multiple head supports only the mat_a.width == mat_b.height, in that case, mat_b will be horizontally split. In this patch, we extend the support when mat_a.width != mat_b.height but mat_a.width/head_number == mat_b.height, in this case, mab_b will be vertically split. One example is A is [3, 8], B is [2, 16], head_number is 4. In this case, A will be split as [3, 2], B will be (vertically) split as [2, 4]. The final result will be 4 matrix of 4 matrix of [3,4], i.e. [3, 16] test=develop * refactor the code of matmul with multiple head even different width and height test=develop
-