1. 25 10月, 2022 1 次提交
  2. 24 10月, 2022 5 次提交
  3. 23 10月, 2022 1 次提交
  4. 21 10月, 2022 2 次提交
  5. 20 10月, 2022 4 次提交
  6. 19 10月, 2022 6 次提交
  7. 18 10月, 2022 5 次提交
  8. 17 10月, 2022 7 次提交
    • G
      Support BF16 training for sharding (#46846) · 0b39b244
      Ghost Screaming 提交于
      * Fix bug of reduce_sum op. When input.numel() > INT32_MAX, its result
      is wrong.
      
      * support pure bfloat16
      
      * support bf16 linear
      
      * update PR to pass CI
      
      * tiny fix where_grad_kernel.cu
      
      * Support bfloat16 type for reducer and sharding.
      
      * Fix some bug.
      
      * Polish code.
      
      * Polise code.
      
      * Add bfloat16 datatype in fill_grad kernels.
      Co-authored-by: Nsneaxiy <sneaxiy@126.com>
      0b39b244
    • O
    • Y
      [PHI]Modify DataLayout's namespace from paddle::experimental to phi (#46869) · ec749398
      YuanRisheng 提交于
      * namespace modify
      
      * update by comment
      ec749398
    • R
      Fix warning message format error (#47045) · 13284437
      RedContritio 提交于
      13284437
    • O
      [Hackathon 3rd No.22 ] add paddle.incubate.sparse.reshape (#46694) · abb38136
      OccupyMars2025 提交于
      * add sparse reshape
      
      * change the dtype in all test cases to int64
      
      * just one test case
      
      * modify comments
      
      * Update test_sparse_reshape_op.py
      
      * chang the type of "shape"  from  vector<int64_t>  to  IntArray
      
      * check whether sp_out.to_dense() is the cause  of error
      
      * print sp_out
      
      * Update reshape_kernel.cc
      
      * use numpy to generate the equal paddle tensor
      
      * just check dense_tensor.numpy()
      
      * check cpu and cuda versions
      
      * Update test_sparse_reshape_op.py
      
      * supply all test cases for cpu forward coo kernel
      
      * test forward coo cuda kernel
      
      * change configuration of cuda kernel
      
      * keep only one test case
      
      * test coo cpu kernel (forward and backward)
      
      * row major or column major ???
      
      * test cuda coo forward kernel
      
      * complete declaration and registration
      
      * Update __init__.py
      
      * rebuild
      
      * retrigger CI
      
      * add cudaMalloc and cudaMemcpy  in  ReshapeCooKernel  and change back to row major order in a cuda dense tensor
      
      * midify minor error
      
      * test only cpu coo forward kernel
      
      * add all test cases for coo forward kernel  (both cpu and gpu)
      
      * test all forward kernels (coo, csr; cpu, gpu)
      
      * add all test cases for all kinds of kernels
      
      * just retrigger CI
      
      * Update sparse_ops.yaml
      
      * Update sparse_ops.yaml
      
      * Update sparse_ops.yaml
      
      * resolve conflicts
      
      * Update sparse_ops.yaml
      
      * don't specify tensor place
      
      * new shape has -1 or 0 in it
      
      * Update unary_grad_kernel.h
      
      * correct lvalue error
      
      * code style
      
      * Update sparse_backward.yaml
      
      * Update sparse_ops.yaml
      
      * Update unary_kernel.h
      
      * Update unary.py
      
      * Update sparse_backward.yaml
      
      * Update unary.py
      
      * code style
      
      * code style
      
      * code style
      
      * Update unary.py
      
      * specify tensor place explicitly
      
      * do not use numpy array
      
      * use numpy array in unit test again
      
      * modify example code in docstring
      abb38136
    • L
      f9c1cdc1
    • D
      [Custom Device] Add singleton to custom device (#46963) · 73196e5a
      duanyanhui 提交于
      * add singleton to custom device
      
      * Update custom_device.cc
      
      Init device_init_flag_ in default
      73196e5a
  9. 14 10月, 2022 2 次提交
  10. 13 10月, 2022 7 次提交