未验证 提交 b4626194 编写于 作者: 郭叶军's avatar 郭叶军 提交者: GitHub

zero/mics.py: use on_accelerator instead of cuda only (#3806)

Co-authored-by: NOlatunji Ruwase <olruwase@microsoft.com>
上级 f8551b43
......@@ -17,6 +17,7 @@ from deepspeed.runtime.zero.parameter_offload import (DeepSpeedZeRoOffload, is_z
from deepspeed.runtime.zero.partition_parameters import Init, AllGatherCoalescedHandle, ZeroParamStatus
from deepspeed.runtime.zero.stage3 import DeepSpeedZeroOptimizer_Stage3
from deepspeed.utils import instrument_w_nvtx, log_dist
from deepspeed.accelerator import get_accelerator
from torch import Tensor
from torch.nn import Parameter
......@@ -443,7 +444,7 @@ class MiCS_Optimizer(DeepSpeedZeroOptimizer_Stage3):
if param_repli_size is None or param_repli_size <= 1:
return
if not partitioned_grads_buffers[0].is_cuda:
if not get_accelerator().on_accelerator(partitioned_grads_buffers[0]):
raise RuntimeError("Local sharding has no support for CPU offloading")
if dist.has_all_reduce_coalesced():
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册