1. 23 7月, 2019 1 次提交
  2. 18 7月, 2019 1 次提交
    • Z
      Feature/auto_growth_allocator (#18561) · ae58afc5
      Zeng Jinle 提交于
      * feature/auto_growth_allocator, test=develop
      
      * add unittest of AlignedAllocator, test=develop
      
      * try to turn on auto_growth to test on CI, test=develop
      
      * fix segmentation fault in mixed_vector.h, test=develop
      
      * add unittests, test=develop
      ae58afc5
  3. 17 7月, 2019 1 次提交
  4. 12 7月, 2019 1 次提交
  5. 11 7月, 2019 2 次提交
    • G
    • Z
      Feature/buffer_shared_inplace (#17911) · d3003a16
      Zeng Jinle 提交于
      * feature/buffer_shared_inplace, test=develop
      
      * refine code, test=develop
      
      * fix elementwise_add op cpu inplace and sum inplace bug, test=develop
      
      * add unittest and debug log, test=develop
      
      * fix parallel_executor scope bug, polish code, test=develop
      
      * fix sum op, activation op, single_in_place_inference bug, test=develop
      
      * remove kLocalExecScopeName, test=develop
      
      * fix unittest,test=develop
      
      * fix out_var first version bug, test=develop
      
      * follow comments,test=develop
      d3003a16
  6. 08 7月, 2019 1 次提交
  7. 02 7月, 2019 1 次提交
    • Y
      supports collective training with programs (#18392) · a873fa84
      Yi Liu 提交于
      1. Since allreduce op has 4 reduce types, We split these four reduce types into four ops
      2. We also refined the collective op code, e.g. we separated the collective op kernel into CPUKernel and CUDAKernel, and remove the device specified DeviceContext parameter in template as we already knew the target DeviceContext
      3. We remove the newly added Collective op role to reduce the complexity of program and graph analysis
      a873fa84
  8. 01 7月, 2019 1 次提交
  9. 27 6月, 2019 3 次提交
    • L
      Fix dygraph show style (#18297) · fd6631ef
      lujun 提交于
      Fix dygraph show style for FluidDoc.
      fd6631ef
    • T
      fix communicator with pyreader (#18350) · 999d9a59
      tangwei12 提交于
      * add is_runnning in communicator, test=develop
      999d9a59
    • H
      supports collective communicated training (#18175) · b7128bac
      HaoRen 提交于
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * fix comment
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * fix comment
      test=develop
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * test=develop
      add collective op unittest standard
      
      * test=develop
      remove the test_collective directory
      
      * test=develop
      remove the test_collective directory
      
      * remove slicegather test
      
      * code format for reducescatter
      
      * update attr of shard_index_op
      
      * Modify macro nccl_helper
      
      * remove test without distribute
      
      * macro collective_helper
      
      * marcro update
      
      * test=develop
      update support python3.5
      
      * test=develop change gpu memory use to 0.1 when test
      
      * test=develop
      update ut equal func
      
      * test=develop
      set flags to 1.5
      
      * test=develop fix pickle dumple  py35
      
      * test=develop
      fix divide in slice and add sync_comm_stream
      update atol and rtol to 1e-05
      rm shard_index op and test
      modify read input from file to read from memory
      remove origin_program in framework and add i/o in c_sync_calc_stream
      
      * test=develop update unittest sync operator I/O
      b7128bac
  10. 26 6月, 2019 1 次提交
  11. 21 6月, 2019 1 次提交
    • J
      dataset (#17973) · 3f8031e2
      jiaqi 提交于
      (1) use channel instead of vector/BlockingQueue in Dataset,to keep same with existing implementation, and make code more readable and flexible (dataset single output channel or multi output channel). one previous memory out of limit problem is cause by not release memory after training.
      (2) add Record because MultiSlotType costs too much memory (80B),fix memory out of limit problem.
      (3) add Channel, Archive in paddle/fluid/framework
      (4) change dataset from shared_ptr to unique_ptr in pybind
      (5) move create/destroy readers from trainer to dataset
      (6) move shuffle from datafeed to dataset. dataset holds memory, datafeed is only for load data and feed data to network.
      (7) fix thread num bug of Dataset when filelist size < thread num
      (8) support set_queue_num in InMemoryDataset
      3f8031e2
  12. 19 6月, 2019 1 次提交
  13. 18 6月, 2019 1 次提交
  14. 15 6月, 2019 1 次提交
  15. 12 6月, 2019 1 次提交
    • T
      combine noavx and avx package (#17889) · 5c06bff2
      tensor-tang 提交于
      * support avx and noavx core
      
      * add catch and give some log
      
      test=develop
      
      * fix build
      
      test=develop
      
      * add missing package
      
      test=develop
      
      * fix pybind name
      
      test=develop
      
      * fix import error
      
      test=develop
      
      * conbime noavx core
      
      test=develop
      
      * add requirements
      
      test=develop
      
      * fix unkown message
      
      test=develop
      
      * fix api spec
      
      test=develop
      
      * refine and clean
      
      test=develop
      
      * update
      
      * pass dist ut
      
      * follow comments
      
      test=develop
      
      * refine scripts
      
      test=develop
      5c06bff2
  16. 10 6月, 2019 1 次提交
  17. 06 6月, 2019 4 次提交
    • G
      fbbdc9cc
    • W
      Make ParallelExecutor support Windows GPU (#17787) · 453a49b1
      wopeizl 提交于
      * fix the ParallelExecutor on Windows
      test=develop
      * restrict to use one GPU only under windows
      453a49b1
    • INT8 MKL-DNN v2 integrate to slim (#17634) · 993c703b
      翟飞跃 提交于
      * refactor PR 16865
      
      * delete mergetool files
      
      * test=develop
      
      * test=develop
      
      * test=develop
      
      * test=develop
      
      * create dir for int8 model before call SaveOptimModel
      
      * test=develop
      
      * mkldnn int8 only support linux; test=develop
      
      * refine code; test=develop
      
      * remove comment; test=develop
      
      * refine code; test=develop
      
      * fix bug; test=develop
      
      * add exception for mkldnn_post_training_strategy
      
      * reuse int8v2 CAPI dataset; test=develop
      
      * fix accuracy check bug; test=develop
      
      * remove tab
      
      * convert files to unix format
      
      * test=develop
      
      * reduce CI time;test=develop
      
      * reduce CI time and refine code;test=develop
      
      * refine comment; test=develop
      
      * add cmake FLAGS;test=develop
      
      * remove predict_num;test=develop
      993c703b
    • W
      use pyreader to read data in dygraph mode (#17314) · 841553e1
      wopeizl 提交于
      * use pyreader to read data
      
      * add return_list to PyReader to support return value represented as list
      841553e1
  18. 05 6月, 2019 1 次提交
  19. 04 6月, 2019 1 次提交
  20. 31 5月, 2019 1 次提交
  21. 27 5月, 2019 2 次提交
  22. 25 5月, 2019 1 次提交
    • Z
      TRT: Support set dynamic range in int8 mode. (#17524) · 61221ebc
      Zhaolong Xing 提交于
      * fluid int8 train and trt int8 predict align.
      trt int8 predict init
      op converter
      
      * 2. align fluid int8 train and trt int8 inference.
      enhance quant dequant fuse pass
      enhance op converter, trt engine, trt engine op, trt subgraph pass.
      
      * 3. add delete_quant_dequant_pass for trt
      
      test=develop
      
      * 4. add the missing file
      test=develop
      
      * 5. i modify the c++ interface, but forget to modify the pybind code
      fix the IS_TRT_VERSION_GE bug, and fix elementwise op converter
      test=develop
      61221ebc
  23. 24 5月, 2019 4 次提交
  24. 23 5月, 2019 3 次提交
  25. 20 5月, 2019 1 次提交
  26. 17 5月, 2019 2 次提交
    • Y
      polish parallel dygraph code (#17164) · 02175555
      Yan Xu 提交于
      * add var grad hook test=develop
      02175555
    • J
      Fix/Fix memory leak in dygraph (#17394) · d7df4e5e
      Jiabin Yang 提交于
      * test=develop, add gradient sort backward strategy
      
      * test=develop, fix test by add FLAGS_cudnn_deterministic on new tests
      
      * test=develop, fix memory leak in dygraph mode
      
      * test=develop, fix memory leak in dygraph mode
      
      * test=develop, polish code
      
      * test=develop, polish code
      
      * test=develop, polish code
      d7df4e5e
  27. 16 5月, 2019 1 次提交