1. 08 9月, 2023 14 次提交
  2. 07 9月, 2023 6 次提交
  3. 06 9月, 2023 16 次提交
  4. 05 9月, 2023 4 次提交
    • L
      Add attributes to support to analyse the stream across interpreters (#56814) · f5497fd0
      lzydev 提交于
      * fix static_build for pp
      
      * add mannual_event to support streams across progs
      
      * revert static_build.sh
      
      * fix coverage-ci
      
      * modify the method to name events
      
      * change code according to review
      f5497fd0
    • C
      fix addn infermeta (#56934) · 99ae88f1
      cyber-pioneer 提交于
      * fix addn infermeta
      
      * fix rule bug
      99ae88f1
    • G
      [Auto Parallel]: Support std::vector<phi::Tensor> input and output for DistTensor. (#56602) · d2fedeac
      Ghost Screaming 提交于
      * [WIP] Support std::vector<phi::Tensor> input and output for DistTensor.
      Concat forward and backward are verified.
      
      * Polish code for new dist tensor implementation.
      
      * Fix bug of DistTensor upgrade. Add support functions for std::vector<Tensor> -> std::vector<Tensor>.
      
      * Add support for DistTensor type of std::vector<phi::Tensor> as input or output of operators.
      Following testcases are passed.
      1. concat: std::vector<phi::Tensor> -> phi::Tensor
      2. unbind: phi::Tensor -> std::vector<phi::Tensor>
      3. broadcast_tensors: std::vector<phi::Tensor> -> std::vector<phi::Tensor>
      
      * Polish code. Remove useless comments.
      
      * Add update_loss_scaling in skip_op_lists.
      
      * Polish code.
      d2fedeac
    • W
      fix some bugs for amp and test case test_tuning_recompute_with_amp.py (#56864) · e9e07a19
      Wennie396 提交于
      * replace amp.use_pure_fp16 with amp.dtype and amp.level
      
      * old api still use use_pure_fp16
      
      * test_fuse_adamw_pass still use use_pure_fp16
      
      * add test case tuning recompute with amp(float16,o2)
      
      * reset new test case properties TIMEOUT 60
      
      * set smaller value of batch_size and batch_num
      
      * deepcopy dist_context fix _rename_input problem
      
      * fix loss name after cast
      
      * set tuning.enable=True and use engine._tune()
      
      * restore some changes in _rename_input()/_rename_output()
      
      * add self.amp_dtype for _cast_loss() in auto_parallel_amp.py
      
      * fix insert op index in _cast_loss()
      e9e07a19