未验证 提交 0b96793e 编写于 作者: X xiongkun 提交者: GitHub

[Dygraph TestsFix] Test some tests in new dygraph final_state mode. (#41363)

* fix less than

* fix some tests

* fix additional 3 unittest case
上级 e0d12b8d
......@@ -132,6 +132,7 @@ PD_REGISTER_KERNEL(full_like,
phi::FullLikeKernel,
float,
double,
uint8_t,
int16_t,
int,
int64_t,
......
......@@ -452,11 +452,12 @@ def spawn(func, args=(), nprocs=-1, join=True, daemon=False, **options):
def train(print_result=False):
# 1. initialize parallel environment
dist.init_parallel_env()
group = dist.init_parallel_env()
process_group = group.process_group if group else None
# 2. create data parallel layer & optimizer
layer = LinearNet()
dp_layer = paddle.DataParallel(layer)
dp_layer = paddle.DataParallel(layer, process_group=process_group)
loss_fn = nn.MSELoss()
adam = opt.Adam(
......
......@@ -182,8 +182,8 @@ def equal(x, y, name=None):
y = full(shape=[1], dtype=x.dtype, fill_value=y)
if in_dygraph_mode():
axis = -1
return _C_ops.final_state_equal(x, y, axis)
default_axis = -1
return _C_ops.final_state_equal(x, y, default_axis)
else:
if _in_legacy_dygraph():
return _C_ops.equal(x, y)
......@@ -232,8 +232,8 @@ def greater_equal(x, y, name=None):
print(result1) # result1 = [True False True]
"""
if in_dygraph_mode():
axis = -1
return _C_ops.final_state_greater_equal(x, y, axis)
default_axis = -1
return _C_ops.final_state_greater_equal(x, y, default_axis)
else:
if _in_legacy_dygraph():
return _C_ops.greater_equal(x, y)
......@@ -383,8 +383,8 @@ def less_than(x, y, name=None):
print(result1) # result1 = [False True False]
"""
if in_dygraph_mode():
axis = -1
return _C_ops.final_state_less_than(x, y, axis)
default_axis = -1
return _C_ops.final_state_less_than(x, y, default_axis)
else:
if _in_legacy_dygraph():
return _C_ops.less_than(x, y)
......
......@@ -2668,6 +2668,7 @@ def cumsum(x, axis=None, dtype=None, name=None):
x = cast(x, dtype)
if in_dygraph_mode():
if axis is None: axis = -1
return _C_ops.final_state_cumsum(x, axis, flatten, False, False)
if _in_legacy_dygraph():
if axis is None:
......
......@@ -419,14 +419,14 @@
func : cumprod
backward : cumprod_grad
# cumsum
- api : cumsum
args : (Tensor x, int axis, bool flatten, bool exclusive, bool reverse)
output : Tensor
output : Tensor(out)
infer_meta :
func : CumsumInferMeta
kernel :
func : cumsum
backward : cumsum_grad
- api : depthwise_conv2d_transpose
args : (Tensor x, Tensor filter, int[] strides, int[] paddings, int[] output_padding, int[] output_size, str padding_algorithm, int groups, int[] dilations, str data_format)
......
......@@ -286,6 +286,15 @@
kernel :
func : cumprod_grad
- backward_api : cumsum_grad
forward : cumsum(Tensor x, int axis, bool flatten, bool exclusive, bool reverse) -> Tensor(out)
infer_meta :
func : UnchangedInferMeta
param : [x]
args : (Tensor out_grad, int axis, bool flatten, bool exclusive, bool reverse)
output : Tensor(x_grad)
invoke : cumsum(out_grad, axis, flatten, exclusive, !reverse)
- backward_api : depthwise_conv2d_transpose_grad
forward : depthwise_conv2d_transpose(Tensor x, Tensor filter, int[] strides, int[] paddings, int[] output_padding, int[] output_size, str padding_algorithm, int groups, int[] dilations, str data_format) -> Tensor(out)
args : (Tensor x, Tensor filter, Tensor out_grad, int[] strides, int[] paddings, int[] output_padding, int[] output_size, str padding_algorithm, int groups, int[] dilations, str data_format)
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册