1. 20 10月, 2021 3 次提交
    • J
      [FIX] Extend time for test_activation_nn_grad to avoid its timeout issue (#36527) · c285c719
      Jiabin Yang 提交于
      * native commit for triple grad of sigmod
      
      * Updated unittests files
      
      * init functional jacobian api
      
      * Updated trible_test func
      
      * Updated gradient_checker & test_script
      
      * finish test with dtype float32
      
      * add float64 test case
      
      * polish code
      
      * use atol=1e-5 with dtype float64
      
      * fix for ci
      
      * set timeout for test_jacobian
      
      * fix dygraph grad to support high differential
      
      * polish API docstring
      
      * Updated gradient checker and some related files
      
      * fix double grad strip error for high differential
      
      * fix double grad strip error for high differential
      
      * Add Sigmoid triple grad tests
      
      * fix dygraph double grad dtype error when calling for high differential senario
      
      * Updated triple grad teses func
      
      * Use np.random to initialize ddx
      
      * Updated triple_grad_check func
      
      * add todo for gradient checker and refine some comments
      
      * remove additional code
      
      * add test for warnging in backward.py
      
      * add tanh triple grad
      
      * format python code
      
      * refine code
      
      * make test_activation_nn_grad test time to 150s
      Co-authored-by: Nveyron95 <veyron_wu@163.com>
      Co-authored-by: Nlevi131 <limaolin01@baidu.com>
      c285c719
    • J
      [Auto Parallel] Generalization for Partition and Completion (#35735) · 797bd40d
      JZ-LIANG 提交于
      * default dist op
      
      * add dist_attr for dist op
      
      * add unitest
      
      * update inputname
      
      * update function name
      
      * add unitest
      
      * update CMakeLists.txt for CI
      
      * fix dis_matmul
      
      * fix compile error
      
      * update matmul to matmul_v2
      
      * unify api
      
      * unify api
      
      * todo
      
      * update distop forward func
      
      * update distop forward func
      
      * auto parallel backward
      
      * update dist op
      
      * autoparallel backward
      
      * add backward for embedding
      
      * temp1
      
      * temp2
      
      * temp3
      
      * temp4
      
      * backward done1
      
      * backward done2
      
      * backward done3
      
      * dist embedding remove mp mode
      
      * dist matmul remove mp mode
      
      * update dist embedding
      『
      
      * dist op init1
      
      * dist op init 2
      
      * update unitest
      
      * context remove parallel mode
      
      * partitioner remove parallel mode
      
      * update unitest
      
      * a more general method to support varying mesh in pipeline parallel
      
      * support varying mesh in pipeline parallel
      
      * embedding support varying mesh in pipeline parallel
      
      * matmul support varying mesh in pipeline parallel
      
      * default dist op support varying mesh in pipeline parallel
      
      * dist attribute for startup program
      
      * default dist op support varying mesh in pipeline parallel 2
      
      * partitoner support varying mesh in pipeline parallel
      
      * revise logic for auto compeletion
      
      * revise framework.py
      
      * revise reshard unitest
      
      * revise unitest for parallelize
      
      * chmod
      
      * fixed bug for dist embedding name mapping
      Co-authored-by: Nzhaoyingli <zhaoyingli@baidu.com>
      797bd40d
    • 0
      remove no_value using var.name (#36513) · fe01ba6a
      0x45f 提交于
      * remove no_value using var.name
      
      * fix unit test for CI
      
      * fix unit test
      
      * add test case
      
      * fix test case
      
      * add more test case
      fe01ba6a
  2. 19 10月, 2021 11 次提交
  3. 18 10月, 2021 7 次提交
    • H
      [HybridParallel]Support fp16 in dygraph hybrid parallel (#36420) · 10f0a0f6
      Haohongxiang 提交于
      * [HybridParallel]Support fp16 in dygraph hybrid parallel
      
      * update
      
      * update
      
      * update for recompute
      
      * add unittest of pp+fp16
      
      * add unittest of recompute+fp16
      
      * update
      
      * modify ut
      10f0a0f6
    • J
      Added softplus FP32 FWD OneDNN kernel (#36382) · bdac9ff6
      jakpiase 提交于
      * added softplus
      
      * refactored softplus op
      
      * deleted unnecessary file
      
      * added missing file
      
      * added formatting
      
      * disabled tests if GPU is used
      
      * added reviewer suggestion
      
      * unified softplus kernel
      bdac9ff6
    • L
      Lml/vhp (#36146) · 4c0ad772
      levi131 提交于
      * init functional jacobian api
      
      * finish test with dtype float32
      
      * add float64 test case
      
      * polish code
      
      * use atol=1e-5 with dtype float64
      
      * fix for ci
      
      * set timeout for test_jacobian
      
      * init hessian API
      
      * save status
      
      * polish API docstring
      
      * modify docstring
      
      * add utils.py
      
      * save status
      
      * fix dygraph double grad dtype error when calling for high differential senario
      
      * reinvoke ci
      
      * test_hessian.py is ok
      
      * polish hessian API
      
      * init vhp
      
      * Revert "init vhp"
      
      This reverts commit cbd4d3b66abe82b0ac10721b9eddeb7d82e0a1c8.
      
      * init vhp
      
      * finish vhp API logically
      
      * add test for partial_engine.cc
      
      * modify numerical_delta with dtype float32
      
      * merge fix for dtype float64
      
      * spell fix
      
      * save status
      
      * polish code
      
      * rm _stop_gradient_pre_process
      
      * save status
      
      * add example for vhp interface
      
      * add _compute_numerical_vjp and _compute_numerical_vhp
      
      * test is ok
      
      * vhp is ok
      
      * add testVHPFloat64
      
      * modify for comments
      
      * modify format
      
      * modify format
      
      * save status
      
      * test_vhp is ok
      
      * finish code polish
      
      * small modify for v is None
      Co-authored-by: NJiabinYang <360788950@qq.com>
      4c0ad772
    • Q
    • Q
      [NPU] fix dtype for arg_max, test=develop (#36457) · 8757fc5b
      Qi Li 提交于
      8757fc5b
    • T
      [autograd.functional] Fix a bug on handling v=None in vjp and jvp (#36445) · 79dbbcce
      Tongxin Bai 提交于
      * autograd.functional passed pylint checker.
      
      * autograd.functional: fix import errors.
      
      * autograd.functional: fixed unit tests.
      
      * autograd.functional minor format change
      
      * [autograd.functional] Fixed vjp and jvp's v=None bug.
      79dbbcce
    • H
      modify ut of cond (#36475) · e496d1e9
      Haohongxiang 提交于
      e496d1e9
  4. 15 10月, 2021 4 次提交
    • 0
      fix no_grad context error in train mode when using save/load (#36434) · 37257d6a
      0x45f 提交于
      * fix no_grad context error in train mode when using save/load
      
      * change net to train mode in test case
      37257d6a
    • J
      Add BuildCinnPass (#36345) · b3f02c57
      jiangcheng 提交于
      * Add CinnSubgraphSearchPass
      
      * solve CI problem of subgraph order not same
      
      * fix some bug by review advices
      
      * ensure the independently of subgraph, that mean the subgraph should not have link to out-graph
      
      * rename cinn_subgraph_search_pass to build_cinn_pass and delete paddle_to_cinn_pass
      
      * add flag to control wheter append build cinn pass
      
      * remove AppendPass at ParallelExecutorPassBuilder
      
      * rename paddle_to_cinn_pass to build_cinn_pass in build_strategy and close test_run_from_cinn
      b3f02c57
    • J
      [New Feature] Support tanh triple grad (#36225) · 808be657
      Jiabin Yang 提交于
      * native commit for triple grad of sigmod
      
      * Updated unittests files
      
      * init functional jacobian api
      
      * Updated trible_test func
      
      * Updated gradient_checker & test_script
      
      * finish test with dtype float32
      
      * add float64 test case
      
      * polish code
      
      * use atol=1e-5 with dtype float64
      
      * fix for ci
      
      * set timeout for test_jacobian
      
      * fix dygraph grad to support high differential
      
      * polish API docstring
      
      * Updated gradient checker and some related files
      
      * fix double grad strip error for high differential
      
      * fix double grad strip error for high differential
      
      * Add Sigmoid triple grad tests
      
      * fix dygraph double grad dtype error when calling for high differential senario
      
      * Updated triple grad teses func
      
      * Use np.random to initialize ddx
      
      * Updated triple_grad_check func
      
      * add todo for gradient checker and refine some comments
      
      * remove additional code
      
      * add test for warnging in backward.py
      
      * add tanh triple grad
      
      * format python code
      
      * refine code
      Co-authored-by: Nveyron95 <veyron_wu@163.com>
      Co-authored-by: Nlevi131 <limaolin01@baidu.com>
      808be657
    • Z
      fix momentum ops (#36452) · 4dda18a8
      Zeng Jinle 提交于
      4dda18a8
  5. 14 10月, 2021 5 次提交
  6. 13 10月, 2021 10 次提交
    • Y
      [PaddlePaddle hackathon] + ADD CELU (#36088) · d7064f04
      yujun 提交于
      * update
      
      * update
      
      * update
      
      * try make CI pass
      
      * doc typo
      
      * update doc string
      d7064f04
    • L
      Merge lars op (#35476) · 0c31579c
      limingshu 提交于
      * A leap of try for cudaLaunchCooperativeKernel
      
      * fix bugs
      
      * Totally replace the lar cuda kernel
      
      * Fix bugs
      
      * a test for lars merge
      
      * Adding las_op_momentum infer_shape
      
      * Fix codes
      
      * use avg_numel instead of max_numel to acquire grid num
      
      * modify unittest files about lars op
      
      * Finally converge when merged-lars works
      
      * fix ctest files
      
      * add merged_operation kernel when cuda version is older than 11
      
      * Fix code style
      
      * fix ctest failure
      
      * fix error
      
      * fix all ctest error and change lars compute code of cpu
      
      * fix bugs on v100.
      
      * revert python modififation about lars
      
      * revert python modification codes
      0c31579c
    • W
      24418479
    • Z
      Add fp16 for clip_by_norm & clip_by_global_norm (#36198) · 3a869cc5
      zhangbo9674 提交于
      * add fp16 for clip_by_norm api
      
      * support ClipByGlobalNorm for fp16 in dygraph
      
      * add unittest for dygraph clipGlobalNorm
      
      * refine unittest for dygraph clipGlobalNorm for mac and windows
      
      * refine unittest
      
      * add unittest for fp64
      
      * refine unittest for fp64
      3a869cc5
    • G
      support auto parallel data shard (#36055) · 85bb1a85
      Guoxia Wang 提交于
      85bb1a85
    • C
      fix pp comm init bug (#36377) · 817f9ef0
      caozhou 提交于
      817f9ef0
    • J
      [New Feature] Support triple grad in Paddle (#36187) · 2c44ee7e
      Jiabin Yang 提交于
      * native commit for triple grad of sigmod
      
      * Updated unittests files
      
      * init functional jacobian api
      
      * Updated trible_test func
      
      * Updated gradient_checker & test_script
      
      * finish test with dtype float32
      
      * add float64 test case
      
      * polish code
      
      * use atol=1e-5 with dtype float64
      
      * fix for ci
      
      * set timeout for test_jacobian
      
      * fix dygraph grad to support high differential
      
      * polish API docstring
      
      * Updated gradient checker and some related files
      
      * fix double grad strip error for high differential
      
      * fix double grad strip error for high differential
      
      * Add Sigmoid triple grad tests
      
      * fix dygraph double grad dtype error when calling for high differential senario
      
      * Updated triple grad teses func
      
      * Use np.random to initialize ddx
      
      * Updated triple_grad_check func
      
      * add todo for gradient checker and refine some comments
      
      * remove additional code
      
      * add test for warnging in backward.py
      
      * format python code
      Co-authored-by: Nveyron95 <veyron_wu@163.com>
      Co-authored-by: Nlevi131 <limaolin01@baidu.com>
      2c44ee7e
    • W
      [PaddleInference] Pass: add int8 flag for op (#36042) · d7858c99
      Wangzheee 提交于
      * add_int_pass
      
      * add_int8_flag_pass
      
      * add_int8_flag_pass
      
      * fix CMakeLists.txt
      
      * fix test_trt_fc_fuse_quant_dequant_pass.py
      
      * fix python/paddle/fluid/tests/unittests/ir/inference/test_trt_fc_fuse_quant_dequant_pass.py
      
      * fix test_trt_fc_fuse_quant_dequant_pass.py
      d7858c99
    • F
      Set NIGHTLY tag for 'tensordot' UT (#36354) · 90457d8c
      From00 提交于
      90457d8c
    • L
      unify usage of tuple and list (#36368) · 3c2bdaa8
      levi131 提交于
      * modify format
      
      * modify format
      3c2bdaa8