- 21 12月, 2017 1 次提交
-
-
由 Yibing Liu 提交于
-
- 24 11月, 2017 1 次提交
-
-
由 QI JUN 提交于
* is_training to is_test in dropout op * handle dropout and batch_norm operator when prune pdesc in testing mode * handle dropout and batch_norm operator when prune pdesc in testing mode * add get_inference_program method * fix dropout op * fix ci * test data after each batch training * refine code * refine test_book3 * fix ci * follow comments
-
- 14 11月, 2017 1 次提交
-
-
由 Qiao Longfei 提交于
* init commit * change some dir name
-
- 27 10月, 2017 1 次提交
-
-
由 Yu Yang 提交于
* Simplize Gradient Check * Stash * Extract apply_backward_pass to backward.py Rename apply_backward_pass to append_backward_ops * Use graph API to check gradient * Fix ci * Fix CI * Fix backward for double precision * Stash * Fix CI * Fix ci * Ignore GRU test * Ignore xe op * Fix CI * Fix softmax with xe gradient The correct equation should be IG = OG * (d_softmax_with_xe()) * Fix typo * Fix merge error * Disable LRN
-
- 20 9月, 2017 1 次提交
-
-
由 dangqingqing 提交于
-
- 19 9月, 2017 2 次提交
-
-
由 Xinghai Sun 提交于
-
由 Xinghai Sun 提交于
Change type of dropout_prob to template typename.
-
- 03 9月, 2017 2 次提交
-
-
由 Xinghai Sun 提交于
-
由 Xinghai Sun 提交于
-