未验证 提交 0b823415 编写于 作者: Y Yu Yang 提交者: GitHub

Merge pull request #9496 from reyoung/feature/revert_activation_inplace

Revert "make append activation in place by default (#9417)"
...@@ -398,6 +398,7 @@ class LayerHelper(object): ...@@ -398,6 +398,7 @@ class LayerHelper(object):
return input_var return input_var
if isinstance(act, basestring): if isinstance(act, basestring):
act = {'type': act} act = {'type': act}
tmp = self.create_tmp_variable(dtype=input_var.dtype)
if 'use_mkldnn' in self.kwargs: if 'use_mkldnn' in self.kwargs:
act['use_mkldnn'] = self.kwargs.get('use_mkldnn') act['use_mkldnn'] = self.kwargs.get('use_mkldnn')
...@@ -407,9 +408,9 @@ class LayerHelper(object): ...@@ -407,9 +408,9 @@ class LayerHelper(object):
self.append_op( self.append_op(
type=act_type, type=act_type,
inputs={"X": [input_var]}, inputs={"X": [input_var]},
outputs={"Out": [input_var]}, outputs={"Out": [tmp]},
attrs=act) attrs=act)
return input_var return tmp
def _get_default_initializer(self, dtype): def _get_default_initializer(self, dtype):
if dtype is None or dtype_is_floating(dtype) is True: if dtype is None or dtype_is_floating(dtype) is True:
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册