1. 09 9月, 2021 6 次提交
  2. 08 9月, 2021 31 次提交
    • L
      add backward inplace for dygraph (#35412) · 0cb413d3
      Leo Chen 提交于
      * add backward inplace for dygraph
      
      * fix bug
      
      * support gradient accumulation
      0cb413d3
    • X
    • L
      ef61da86
    • zhouweiwei2014's avatar
      1159f753
    • zhouweiwei2014's avatar
      2133f3dd
    • Y
      [Auto Parallel] Integrate all modules (#35483) · 12155358
      Yulong Ao 提交于
      * add auto_parallel dir
      
      * mv to paddle.distributed
      
      * add shard_xx api
      
      * add distributed attrs for var
      
      * add ut, test=develop
      
      * add dist
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update, test=develop
      
      * update, test=develop
      
      * update
      
      * update
      
      * delete unused proto
      
      * resotre op_desc
      
      * restore type_defs
      
      * update var_desc
      
      * remove dimss_mapping for proto_pybind
      
      * update interface.py
      
      * update framework.py
      
      * update
      
      * update
      
      * add auto_parallel dir
      
      * mv to paddle.distributed
      
      * add shard_xx api
      
      * add distributed attrs for var
      
      * add ut, test=develop
      
      * [WIP] Add the auto completion feature and related codes
      
      * [WIP] Improve the auto completion and related codes
      
      * [WIP] Make the auto completion to support data-parallel
      
      * [WIP] Make the completion support mp and dp+mp
      
      * [WIP] Refactor auto completion unit test for MLP
      
      * [WIP] Refactor the implementation of DistributedOperatorImpl
      
      * [WIP] Improve dims_mapping update rule and fix a bug
      
      * [WIP] Support auto completion for one transformer decoder layer
      
      * [WIP] Add a minor change
      
      * [WIP] Fix a bug within the uint test
      
      * Shard XShape tensor, add embedding completion and refactor code
      
      * Add the distributed_operators dir to setup.py.in
      
      * Improve the completion process and add the unittest for gpt
      
      * fix process_mesh ut
      
      * fix process_mesh ut
      
      * update
      
      * update, test=develop
      
      * Add support for automatically completing distributed attrs of special ops
      
      * update
      
      * update
      
      * update
      
      * fix doc sample codes, test=develop
      
      * improve coverage, test=develop
      
      * add static_mode check, test=develop
      
      * Model the cluster for cost model and physical mapping
      
      * update, test=develop
      
      * add set_placement, test=develop
      
      * Add the check to make sure the candidate tensors' size is great than zero
      
      * update doc, test=develop
      
      * update doc, test=develop
      
      * update doc, test=develop
      
      * update doc, test=develop
      
      * update, test=develop
      
      * Auto mark dist attrs annotated by user
      
      * update ndarray to nested list, test=develop
      
      * update, test=develop
      
      * Add auto-completion module for auto-parallel (based on PR#33804)
      
      * Remove unnecessary files
      
      * Remove unrelated files for the auto completion pr
      
      * Update the unit test to improve the coverage
      
      * Modify codes based on reviews
      
      * Minor changes for CI
      
      * Improve some codes based on new comments
      
      * Fix bugs caused by shallow copy in attributes.py
      * Imporve amend_distributed_attr_for_program in context.py
      * Other changes for weihang's comments
      
      * support shard reader
      
      * support shard reader
      
      * add parallel mode
      
      * update process mesh
      
      * add method to compute comm_group
      
      * implement dist_embedding forward func
      
      * implement dist matmul forward func
      
      * implement dist reshape forward func
      
      * add transpiler framework
      
      * add transpiler forward
      
      * implement transpiler forward
      
      * implement transpiler backward & update
      
      * add process
      
      * add unitest
      
      * chmod
      
      * chmod
      
      * chmod
      
      * update unitest
      
      * add unitest for gpt
      
      * remove unused print
      
      * rename transpiler --> partitioner
      
      * rename transpiler --> partitioner
      
      * chmod
      
      * chmod
      
      * bug fixed
      
      * remove amp function
      
      * update case for dp mode
      
      * update case for dp mode
      
      * [Auto Parallel] Integrate all parts with the newest code
      
      * Integrate all parts of auto parallel and improve codes
      
      * Integrate all parts by AutoParallelizer
      * Add unit test for AutoParallelizer
      * Improve auto completion module for pipeline parallel
      * Add support for matmul_v2 in dist_matmul
      * Correct the typo "stratergy" to "strategy"
      
      * Modify distributed_strategy.proto to conform the main stream
      
      * Restore parts of distributed_strategy to conform the develop branch
      Co-authored-by: Nsandyhouse <lilong12@baidu.com>
      Co-authored-by: NJZ-LIANG <jianzhongliang10@gmail.com>
      12155358
    • W
      multiply supports bool · db5fd2a1
      will-jl944 提交于
      multiply supports bool  
      db5fd2a1
    • W
    • N
      modified pool_op for higher performance (#33144) · b95c5ae0
      niuliling123 提交于
      b95c5ae0
    • W
      refactor new executor (#35537) · 0eb7c942
      wanghuancoder 提交于
      * refactor new executor, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      0eb7c942
    • C
      mark WhileOp AsExtra attribute (#35499) · ce7c18f6
      CtfGo 提交于
      * mark WhileOp AsExtra attribute
      
      * revert kX and kOutputs
      ce7c18f6
    • N
    • L
      add clip_by_norm fp16 kernel (#35446) · 7aa4d879
      Leo Chen 提交于
      * add clip_by_norm fp16 kernel
      
      * add ut
      7aa4d879
    • S
      Slice bug (#35357) · 28abd5d8
      Shang Zhizhou 提交于
      * update slice plugin
      
      * add test
      
      * fix code style
      
      * fix trt6
      
      * update test
      
      * fix test
      
      * add timeout
      
      * update trt version
      
      * update cmake
      28abd5d8
    • X
      Intergrate GLOOParallelContext to support Multi-CPU Core for Dygraph DataParallel (#35154) · 51cc73f0
      xiongkun 提交于
      * can pass the fake test
      
      * add files
      
      * modify cmake to pass windows-ci
      
      * for ci pass
      
      * WITH_GLOO=ON
      
      * for pass coverage test
      
      * add cpuonly testcase
      
      * add
      
      * disable nccl when compile with cuda
      
      * change python version in cpuonly
      
      * add backend argument
      
      * add required gpu
      
      * add required:gpu
      51cc73f0
    • H
      Add AsExtra in relu6 Op Maker (#35472) · 692ac3e5
      hong19860320 提交于
      692ac3e5
    • G
      fix bug (#35482) · e133d8ef
      Guoxia Wang 提交于
      e133d8ef
    • Z
      Fix scatter_nd_add and gather bug (#35544) · 3c457a38
      Zeng Jinle 提交于
      * fix scatter_add_nd and gather bug
      
      * fix gather compile error
      3c457a38
    • Z
      Enable program passes on Fleet APIs (#34955) · 5f369881
      Zeng Jinle 提交于
      * add fleet api for program pass
      
      * turn on apply pass for CI test
      
      * fix disable fuse_all_optimizer bug
      
      * try to test ci
      
      * fix CI
      
      * fill unspecified op role
      
      * fix fuse_allreduce
      
      * add ut to improve coverage
      
      * remove useless change
      
      * improve c++ coverage
      
      * follow some comments
      
      * test ir pass pipeline
      
      * update doc
      
      * reduce ut time again
      5f369881
    • Z
      fix the bug of layer_norm when batch_size=1 (#35480) · ad5f7494
      zhangkaihuo 提交于
      The bug is that access to mean and var is incorrect, and the array will be out of bounds: the shape of mean and var is [batch_size], and the range of thread idx is 0~feature_size, so mean[idx] and var[idx] is incorrect.
      
      When batch_size=1, the correct access is mean[0] and var[0], and a unit test with batch_size=1 is added.
      ad5f7494
    • C
      Add FP16 PRelu (#35532) · 4e62af80
      cc 提交于
      4e62af80
    • L
      hidden the auto parallel apis (#35385) · afd1b372
      lilong12 提交于
      * update, test=develop
      afd1b372
    • L
      add checkers for auto parallel apis (#35486) · 39540b0e
      lilong12 提交于
      * update, test=develop
      39540b0e
    • F
      merge CMakeList.txt manual (#35378) · c4a3e8b4
      feng_shuai 提交于
      * merge CMakeList.txt manual
      
      * add platform for changethreadnum
      
      * repair some bugs according to make error
      
      * do nothing just flush CI
      
      * forget change thread num
      
      * add inplace_atol param for check_output_with_place
      
      * Windows
      
      * std:min and std::max should be change because of windows
      c4a3e8b4
    • L
      support weight sharing for pipeline (#35351) · 5199c744
      lilong12 提交于
      * support weight sharing
      5199c744
    • F
      add some file for jetson-op-test-CE (#35431) · 18a963a5
      feng_shuai 提交于
      * add some file for jetson-op-test-CE
      
      * change diff for some case
      
      * rename op_test -> jetson_infer_op
      
      * add some case because too large
      18a963a5
    • L
      [NPU] release gil before op run (#35370) · db6242e9
      Leo Chen 提交于
      * release gil before op run
      
      * support npu grad test
      
      * fix op_test
      db6242e9
    • Z
      3dab2e20
    • L
      Work queue group (#35470) · a53460aa
      liutiexing 提交于
      * Split Tracker and WorkQueue
      
      * add WorkQueueGroup
      
      * add unittest
      
      * fix
      
      * update
      
      * update
      
      * fix compile
      a53460aa
    • W
      add the matmul v2 grad kernel · b3787d1b
      wawltor 提交于
      * add the matmul v2 grad kernel
      
      * relief the test case time
      
      * update the test case for the matmul double grad
      
      * remove the unsed code for the matmul double grad
      
      * update the test case for the double grad matmul
      
      * remove the unused code in dot
      b3787d1b
    • W
      c727ec4a
  3. 07 9月, 2021 3 次提交