1. 19 10月, 2022 6 次提交
  2. 18 10月, 2022 5 次提交
  3. 17 10月, 2022 7 次提交
    • G
      Support BF16 training for sharding (#46846) · 0b39b244
      Ghost Screaming 提交于
      * Fix bug of reduce_sum op. When input.numel() > INT32_MAX, its result
      is wrong.
      
      * support pure bfloat16
      
      * support bf16 linear
      
      * update PR to pass CI
      
      * tiny fix where_grad_kernel.cu
      
      * Support bfloat16 type for reducer and sharding.
      
      * Fix some bug.
      
      * Polish code.
      
      * Polise code.
      
      * Add bfloat16 datatype in fill_grad kernels.
      Co-authored-by: Nsneaxiy <sneaxiy@126.com>
      0b39b244
    • O
    • Y
      [PHI]Modify DataLayout's namespace from paddle::experimental to phi (#46869) · ec749398
      YuanRisheng 提交于
      * namespace modify
      
      * update by comment
      ec749398
    • R
      Fix warning message format error (#47045) · 13284437
      RedContritio 提交于
      13284437
    • O
      [Hackathon 3rd No.22 ] add paddle.incubate.sparse.reshape (#46694) · abb38136
      OccupyMars2025 提交于
      * add sparse reshape
      
      * change the dtype in all test cases to int64
      
      * just one test case
      
      * modify comments
      
      * Update test_sparse_reshape_op.py
      
      * chang the type of "shape"  from  vector<int64_t>  to  IntArray
      
      * check whether sp_out.to_dense() is the cause  of error
      
      * print sp_out
      
      * Update reshape_kernel.cc
      
      * use numpy to generate the equal paddle tensor
      
      * just check dense_tensor.numpy()
      
      * check cpu and cuda versions
      
      * Update test_sparse_reshape_op.py
      
      * supply all test cases for cpu forward coo kernel
      
      * test forward coo cuda kernel
      
      * change configuration of cuda kernel
      
      * keep only one test case
      
      * test coo cpu kernel (forward and backward)
      
      * row major or column major ???
      
      * test cuda coo forward kernel
      
      * complete declaration and registration
      
      * Update __init__.py
      
      * rebuild
      
      * retrigger CI
      
      * add cudaMalloc and cudaMemcpy  in  ReshapeCooKernel  and change back to row major order in a cuda dense tensor
      
      * midify minor error
      
      * test only cpu coo forward kernel
      
      * add all test cases for coo forward kernel  (both cpu and gpu)
      
      * test all forward kernels (coo, csr; cpu, gpu)
      
      * add all test cases for all kinds of kernels
      
      * just retrigger CI
      
      * Update sparse_ops.yaml
      
      * Update sparse_ops.yaml
      
      * Update sparse_ops.yaml
      
      * resolve conflicts
      
      * Update sparse_ops.yaml
      
      * don't specify tensor place
      
      * new shape has -1 or 0 in it
      
      * Update unary_grad_kernel.h
      
      * correct lvalue error
      
      * code style
      
      * Update sparse_backward.yaml
      
      * Update sparse_ops.yaml
      
      * Update unary_kernel.h
      
      * Update unary.py
      
      * Update sparse_backward.yaml
      
      * Update unary.py
      
      * code style
      
      * code style
      
      * code style
      
      * Update unary.py
      
      * specify tensor place explicitly
      
      * do not use numpy array
      
      * use numpy array in unit test again
      
      * modify example code in docstring
      abb38136
    • L
      f9c1cdc1
    • D
      [Custom Device] Add singleton to custom device (#46963) · 73196e5a
      duanyanhui 提交于
      * add singleton to custom device
      
      * Update custom_device.cc
      
      Init device_init_flag_ in default
      73196e5a
  4. 14 10月, 2022 2 次提交
  5. 13 10月, 2022 7 次提交
  6. 12 10月, 2022 7 次提交
  7. 11 10月, 2022 4 次提交
  8. 10 10月, 2022 2 次提交
    • Y
      [PHI]Add RNN yaml (#46812) · ab60fd8b
      YuanRisheng 提交于
      * add yaml entry for rnn and rrnn_grad, move infershape function for rnn_grad to phi infer meta
      
      * WIP: move rnn kernrl to phi
      
      * Change the code generation to avoid converting from intializer list to tuple of heterogeneous types.
      This is only triggered when an api has intermediate outputs, and the result of the outputs are of heterogeneous types.
      
      * fix the bug that when none in a vector of tensors requires gradient, the conversion to InferShapeContext to InferMetaContext (a.k.a. BuildInferMetaContext) produces errorous results.
      
      * fix ci bugs
      
      * fix ci bugs
      
      * fix ci bugs
      
      * modify code according comment
      Co-authored-by: Nchenfeiyu <chenfeiyu@baidu.com>
      ab60fd8b
    • R
      remove comment (#46827) · 8a5f17e8
      Rayman 提交于
      8a5f17e8