- 16 6月, 2022 3 次提交
-
-
由 Leo Chen 提交于
* fix xpu kp compilation * add depends
-
由 Leo Chen 提交于
* lazy creating work queue * fix dry_run
-
由 gongweibao 提交于
-
- 15 6月, 2022 17 次提交
-
-
由 huzhiqiang 提交于
-
由 Yiqun Liu 提交于
* Optimize prod's python implementation for dygraph. * Change key_dim to head_dim. * Add comment in unittest. * Disable TF32 in unittest.
-
由 fwenguang 提交于
-
由 Leo Chen 提交于
-
由 fwenguang 提交于
-
由 fwenguang 提交于
-
由 Allen Guo 提交于
* Decoupling ipu sharding and modeling (#665) * feat(shard): decoupling shard setting with modeling. * fix(shard): split test cases to avoid failure. * fix(shard): add function docs and fix typo. * test(shard): add tests. * test(shard): more test case. * fix(): change ipu_index/stage default value to -1. * fix format Co-authored-by: Nczr-gc <96037699+czr-gc@users.noreply.github.com>
-
由 zhangbo9674 提交于
* set_state_dict not use state_dict hook * add ut * refine doc
-
由 zhangkaihuo 提交于
-
由 Guoxia Wang 提交于
-
由 Leo Chen 提交于
-
由 zhaoyingli 提交于
* use tempfile to place temporary files * update * revert test_communicator * fix test_dist_base
-
由 zyfncg 提交于
* rename yaml file * fix merge conflict * fix infrt
-
由 zhouweiwei2014 提交于
* add some kernel(csr*dense->csr, dense*dense->csr) of SparseTensor matmul * fix CI * fix CI * fix comment * fix comment
-
由 Zhong Hui 提交于
-
由 Yiqun Liu 提交于
Use int64_t in GetGpuLaunchConfig1D and ElementwiseKernel as index type to support large tensor. (#43506) * Change some data type from int to int64_t in GetGpuLaunchConfig1D to support large tensor. * Use int64_t in ElementwiseKernel as index type to support large tensor.
-
由 Ruibiao Chen 提交于
* Refactor port.h * Remove some unnecessary code * Fix CI errors
-
- 14 6月, 2022 20 次提交
-
-
由 Li Min 提交于
-
由 Jiabin Yang 提交于
* fix starganv2 * fix starganv2 stop_gradient end error * fix edvr_starganv2 * fix mul kernel to fix optional ddx * fix typo
-
由 Ruibiao Chen 提交于
* Support sequential run for standalone executor * Add UTs * Fix test_standalone_multiply_write * Remove unnecessary UTs
-
由 cambriconhsq 提交于
-
由 zhaoying9105 提交于
* [MLU]: add elementwise_max mlu kernel * [MLU]: add int32 support for elementwise maxk MLU kernel
-
由 zhaoying9105 提交于
-
由 Haohongxiang 提交于
-
由 tianshuo78520a 提交于
-
由 Sing_chan 提交于
-
由 WangZhen 提交于
* Refine ifelse early return
-
由 yaozhixin 提交于
* update paddle.distributed.launch * add sample code * update shell * fix typo * fix typo * update docs * rm code * fix doc 2 * fix doc 3 * fix doc 4 Co-authored-by: Nroot <root@sgjur-pod004-1.ipu.graphcore.cn>
-
由 sneaxiy 提交于
-
由 Yuang Liu 提交于
-
由 Ligoml 提交于
-
由 Sing_chan 提交于
* open inference_ut;test=windows_ci_inference * inference_ut need onnx;test=windows_ci_inference * disable trt_split_converter_test; use higher parallel level * too high parallel will cause ut timeout
-
由 Zhang Jun 提交于
-
由 xiongkun 提交于
* change logic for optimize * modifty * optimize the backward speed of EinsumOp * add cache optimizer for einsum op * EinsumOp: fix new dygraph mode error * fix bug * change Cache->InnerCache * fix code * fix * add nan inf utils for einsum op * add as_extra * memory optimizer for einsum * update code
-
由 Zhang Ting 提交于
* fix the bug that _DataLoaderIterMultiProcess use time to generate the seed * use np.random.randint to generate a base seed
-
由 Sing_chan 提交于
-
由 Sing_chan 提交于
-