1. 05 11月, 2019 1 次提交
  2. 10 6月, 2019 1 次提交
  3. 06 6月, 2019 1 次提交
    • INT8 MKL-DNN v2 integrate to slim (#17634) · 993c703b
      翟飞跃 提交于
      * refactor PR 16865
      
      * delete mergetool files
      
      * test=develop
      
      * test=develop
      
      * test=develop
      
      * test=develop
      
      * create dir for int8 model before call SaveOptimModel
      
      * test=develop
      
      * mkldnn int8 only support linux; test=develop
      
      * refine code; test=develop
      
      * remove comment; test=develop
      
      * refine code; test=develop
      
      * fix bug; test=develop
      
      * add exception for mkldnn_post_training_strategy
      
      * reuse int8v2 CAPI dataset; test=develop
      
      * fix accuracy check bug; test=develop
      
      * remove tab
      
      * convert files to unix format
      
      * test=develop
      
      * reduce CI time;test=develop
      
      * reduce CI time and refine code;test=develop
      
      * refine comment; test=develop
      
      * add cmake FLAGS;test=develop
      
      * remove predict_num;test=develop
      993c703b
  4. 26 3月, 2019 1 次提交
    • W
      [slim] Add quantization strategy and distillation strategy. (#16408) · e9bec936
      whs 提交于
      * Add fsp operator.
      1 Add unitest.
      2. Add python API.
      3. Add layer test.
      
      * Add quantization strategy.
      1. Add API.
      2. Add unitest.
      
      * Add distillatoin strategy.
      
      * Add unitest config file for quantization
      
      * Fix Copyright
      test=develop
      
      * Fix setup.py
      
      * Fix document of layers.py.
      test=develop
      
      * Fix unitest in python3.
      test=develop
      
      * Fix documents.
      test=develop
      
      * 1. refine fsp op by batched gemm
      2. remove unused import
      test=develop
      
      * Fix test_dist_se_resnext.
      1. disable test distillation.
      2. reset framework.py
      test=develop
      
      * Enable unitest of distillation after fixing Block._clone_variable
      test=develop
      
      * Fix cdn issue.
      test=develop
      e9bec936
  5. 21 1月, 2019 1 次提交
  6. 20 1月, 2019 1 次提交
  7. 18 9月, 2018 1 次提交
  8. 13 9月, 2018 1 次提交
  9. 15 8月, 2018 1 次提交
  10. 02 8月, 2018 1 次提交
  11. 18 7月, 2018 1 次提交
  12. 24 2月, 2018 1 次提交
  13. 13 2月, 2018 1 次提交
    • X
      Run Python OP tests in a single Python process to improve test time. (#8362) · cde6241a
      Xin Pan 提交于
      Currently, our tests run with 2 GPUs, the init time is absurdly long:
      about 4s for each process.  Currently, we run each OP test on
      different processes. This PR:
      
      1. create cmake function py_test_modules which will generate the
      Makefile that runs a list of Python unittest module in a single Python
      process.
      
      2. move all "python unittest compatible" (e.g., used the unittest
      package, not just a regular python file). from fluid/tests to
      fluid/tests/unittests.
      
      3. cmake now will run all OP tests in fluid/tests/unittests in a
      single process, except the time-consuming tests, they are separated
      into different processes to utilize parallelism. Please make sure to
      use the unittest package if you put the python test file in
      fluid/tests/unittests
      
      4. remove all exit(0) from fluid/tests/unittests/*.py, exit(0) is used
      to disable unittest, we can not do it when running all tests in a
      single process since it will terminate the process without running the
      other tests. Instead, the test is disabled in
      fluid/tests/unittests/CMakeLists.txt. FIXME is added for each disabled
      item. Please disable the unittest from
      fluid/tests/unittests/CMakeLists.txt, instead of adding exit(0) to the
      Python file, for all Python file in fluid/tests/unittests/.
      
      5. add an option WITH_FAST_BUNDLE_TEST. When OFF, will run the unit
      tests in separate process so that they can be tested individually.
      cde6241a
  14. 09 12月, 2016 1 次提交
  15. 12 11月, 2016 1 次提交
  16. 29 8月, 2016 1 次提交