未验证 提交 33afeb31 编写于 作者: S ShenLiang 提交者: GitHub

fix the tanh (#26657)

* fix the tanh

* fix the learning rate
上级 8986a821
...@@ -21,6 +21,7 @@ from collections import defaultdict ...@@ -21,6 +21,7 @@ from collections import defaultdict
from paddle.fluid.distribute_lookup_table import find_distributed_lookup_table from paddle.fluid.distribute_lookup_table import find_distributed_lookup_table
from paddle.fluid.framework import Program, Variable, name_scope, default_main_program, default_startup_program, device_guard from paddle.fluid.framework import Program, Variable, name_scope, default_main_program, default_startup_program, device_guard
import paddle
from ..fluid import framework from ..fluid import framework
from ..fluid import layers from ..fluid import layers
...@@ -308,7 +309,8 @@ class Optimizer(object): ...@@ -308,7 +309,8 @@ class Optimizer(object):
name=unique_name.generate("learning_rate"), name=unique_name.generate("learning_rate"),
shape=[1], shape=[1],
value=float(self._learning_rate), value=float(self._learning_rate),
dtype='float32' if self._dtype is None else self._dtype, dtype=paddle.get_default_dtype()
if self._dtype is None else self._dtype,
persistable=True) persistable=True)
# get learning rate Tensor from LearningRateDecay # get learning rate Tensor from LearningRateDecay
elif isinstance(self._learning_rate, LearningRateDecay): elif isinstance(self._learning_rate, LearningRateDecay):
...@@ -336,7 +338,8 @@ class Optimizer(object): ...@@ -336,7 +338,8 @@ class Optimizer(object):
name=unique_name.generate("learning_rate"), name=unique_name.generate("learning_rate"),
shape=[1], shape=[1],
value=float(self._learning_rate), value=float(self._learning_rate),
dtype='float32' if self._dtype is None else self._dtype, dtype=paddle.get_default_dtype()
if self._dtype is None else self._dtype,
persistable=True) persistable=True)
@framework.dygraph_only @framework.dygraph_only
......
...@@ -2090,6 +2090,7 @@ def tanh(x, name=None): ...@@ -2090,6 +2090,7 @@ def tanh(x, name=None):
return core.ops.tanh(x) return core.ops.tanh(x)
check_variable_and_dtype(x, 'x', ['float16', 'float32', 'float64'], 'tanh') check_variable_and_dtype(x, 'x', ['float16', 'float32', 'float64'], 'tanh')
check_type(x, 'x', (Variable), 'tanh')
helper = LayerHelper('tanh', **locals()) helper = LayerHelper('tanh', **locals())
out = helper.create_variable_for_type_inference(x.dtype) out = helper.create_variable_for_type_inference(x.dtype)
helper.append_op(type='tanh', inputs={'X': x}, outputs={'Out': out}) helper.append_op(type='tanh', inputs={'X': x}, outputs={'Out': out})
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册