1. 04 1月, 2023 1 次提交
    • H
      [Unify KernelKey] change OpKernelType->KernelKey (#49138) · 4383494f
      HongyuJia 提交于
      * execute use kernel_key first
      
      * change OpKernelType->KernelKey
      
      * fix py3 compile error, remove redundant header files
      
      * fix build_strategy_test
      
      * fix DataType::RAW
      
      * fix custom_type test: operator_test.cc
      
      * fix transform place
      
      * fix backends_are_same_class
      
      * try fix place TransDataDevice
      
      * support all KernelKey
      
      * fix TransformData
      
      * fix place_are_same_class
      
      * fix merge
      
      * fix test_params_no_grad
      
      * fix specific place of GetExpectedKernelType
      
      * fix specific place of GetExpectedKernelType
      
      * fix GetKernelTypeForVar
      
      * fix dtype error
      
      * fix fetch_v2
      
      * change GetKernelTypeForVar
      
      * fix interpreter
      
      * fix typo error
      
      * polish codes
      
      * polish codes
      
      * polish codes
      
      * fix conflict
      4383494f
  2. 28 9月, 2022 1 次提交
    • C
      Remove the declaration of using Tensor in framework/tensor.h (#46432) · e12a905e
      Chen Weihang 提交于
      * remove needless using tensor
      
      * remove needless using tensor
      
      * resolve conflict
      
      * replace tensor using
      
      * fix format error
      
      * revert needless changing
      
      * fix rocm and npu compile error
      
      * fix cinn compile error
      
      * fix format error
      
      * fix mkldnn format error
      
      * fix mkldnn format error
      
      * fix cinn compile error
      
      * fix cinn compile error
      
      * fix cinn compile error
      
      * resolve conflict
      e12a905e
  3. 15 9月, 2022 1 次提交
  4. 16 8月, 2022 1 次提交
    • C
      [Phi] Move amp ops into phi (#45079) · b4f67757
      Chen Weihang 提交于
      * move check finite and unscale kernel into phi
      
      * move infershape into phi
      
      * move update_loss_scaling kernel into phi
      
      * remove original kernels
      
      * move update loss scaling infershape into phi
      
      * add header for xpu and npu
      
      * solve coverage failed
      
      * fix npu test failed
      
      * remove mutable data in cu file
      
      * fix new executor failed
      
      * add valid check for meta tensor output
      b4f67757
  5. 02 7月, 2022 1 次提交
    • L
      unify cpu context, part2 (#44012) · 755438a7
      Leo Chen 提交于
      * fix init()
      
      * delete test_device_context
      
      * replace CPUDeviceContext with CPUContext
      
      * fix test_scalar
      
      * remove dot_op.cc
      
      * fix compile
      755438a7
  6. 26 6月, 2022 1 次提交
  7. 14 6月, 2022 1 次提交
  8. 05 6月, 2022 1 次提交
  9. 28 4月, 2022 1 次提交
  10. 15 9月, 2021 1 次提交
  11. 08 1月, 2021 1 次提交
    • Z
      Support pure fp16 training for AMP API. (#29544) · 7f7dfccf
      Zhen Wang 提交于
      * add cast ops before and after unsupported fp16 ops.
      
      * Keep partial net in FP32 pattern.
      
      * Support check_finite_and_unscale and update_loss_scaling for FP16 calculation mode.
      
      * Add fp16 support for adam op.
      
      * add multi precision attr for adam.
      
      * Fix the bug of test_multi_precision_fp16_train UT.
      
      * Code format for CI.
      
      * Fix the redefine error about MPTypeTrait on windows.
      
      * fix bugs of the _create_accumulators func in Momentum.
      
      * fix bug when inserting post cast op.
      
      * Add the update_loss_scaling op in allow_set of UnusedVarCheck.
      
      * Update for ci coverage.
      
      * Add some doc for OptimizerWithMixedPrecision.
      
      * Fix the code style.
      
      * Imporve the doc of `amp_init`.
      
      * Change for fp16 testing if users have the infer program defined in separate way.
      7f7dfccf
  12. 10 12月, 2020 1 次提交
  13. 09 10月, 2020 1 次提交
  14. 14 9月, 2020 1 次提交
    • Z
      Update amp_check_finite_and_scale_op and add an updating_loss_scaling op for... · d708b210
      Zhen Wang 提交于
      Update amp_check_finite_and_scale_op and add an updating_loss_scaling op for static graph amp training. (#26240)
      
      * update amp_check_finite_and_scale_op for static_amp.
      
      * use amp_check_finite_and_scale in static graph amp.
      
      * update grads to zero when grads own infinite values(as for amp_checkout_finite_and_scale op).
      
      * add update_loss_scaling op in cpp.
      
      * add update_loss_scaling_op unit test.
      
      * update the doc of the check_finite_and_unscale op
      
      * Update the process of gradients updating skipping if the gradients have infinite values.
      
      * update the way to zero grads.
      
      * update test_update_loss_scaling_op.py
      
      * add log info when find infinite grads.
      
      * add the unit test for UpdateLossScaling Layer.
      d708b210