- 20 3月, 2018 2 次提交
-
-
由 Kexin Zhao 提交于
-
由 Kexin Zhao 提交于
-
- 24 2月, 2018 1 次提交
-
-
由 Luo Tao 提交于
-
- 13 2月, 2018 1 次提交
-
-
由 Xin Pan 提交于
Currently, our tests run with 2 GPUs, the init time is absurdly long: about 4s for each process. Currently, we run each OP test on different processes. This PR: 1. create cmake function py_test_modules which will generate the Makefile that runs a list of Python unittest module in a single Python process. 2. move all "python unittest compatible" (e.g., used the unittest package, not just a regular python file). from fluid/tests to fluid/tests/unittests. 3. cmake now will run all OP tests in fluid/tests/unittests in a single process, except the time-consuming tests, they are separated into different processes to utilize parallelism. Please make sure to use the unittest package if you put the python test file in fluid/tests/unittests 4. remove all exit(0) from fluid/tests/unittests/*.py, exit(0) is used to disable unittest, we can not do it when running all tests in a single process since it will terminate the process without running the other tests. Instead, the test is disabled in fluid/tests/unittests/CMakeLists.txt. FIXME is added for each disabled item. Please disable the unittest from fluid/tests/unittests/CMakeLists.txt, instead of adding exit(0) to the Python file, for all Python file in fluid/tests/unittests/. 5. add an option WITH_FAST_BUNDLE_TEST. When OFF, will run the unit tests in separate process so that they can be tested individually.
-
- 12 2月, 2018 1 次提交
-
-
由 qingqing01 提交于
-
- 30 1月, 2018 1 次提交
-
-
由 caoying03 提交于
-
- 21 1月, 2018 1 次提交
-
-
由 dzhwinter 提交于
* "fix decode bug" * "follow commnet" * "fix error" * "fix hook bug" * fix based comment * fix copyright * fix based on comment
-
- 15 1月, 2018 1 次提交
-
-
由 dzhwinter 提交于
* add copyright hook * add copyright hook * refine copyright hook * "test copyright hook" * fix check style * fix ci
-
- 21 12月, 2017 1 次提交
-
-
由 Yibing Liu 提交于
-
- 24 11月, 2017 1 次提交
-
-
由 QI JUN 提交于
* is_training to is_test in dropout op * handle dropout and batch_norm operator when prune pdesc in testing mode * handle dropout and batch_norm operator when prune pdesc in testing mode * add get_inference_program method * fix dropout op * fix ci * test data after each batch training * refine code * refine test_book3 * fix ci * follow comments
-
- 14 11月, 2017 1 次提交
-
-
由 Qiao Longfei 提交于
* init commit * change some dir name
-
- 27 10月, 2017 1 次提交
-
-
由 Yu Yang 提交于
* Simplize Gradient Check * Stash * Extract apply_backward_pass to backward.py Rename apply_backward_pass to append_backward_ops * Use graph API to check gradient * Fix ci * Fix CI * Fix backward for double precision * Stash * Fix CI * Fix ci * Ignore GRU test * Ignore xe op * Fix CI * Fix softmax with xe gradient The correct equation should be IG = OG * (d_softmax_with_xe()) * Fix typo * Fix merge error * Disable LRN
-
- 20 9月, 2017 1 次提交
-
-
由 dangqingqing 提交于
-
- 19 9月, 2017 2 次提交
-
-
由 Xinghai Sun 提交于
-
由 Xinghai Sun 提交于
Change type of dropout_prob to template typename.
-
- 03 9月, 2017 2 次提交
-
-
由 Xinghai Sun 提交于
-
由 Xinghai Sun 提交于
-