未验证 提交 dcf8ef04 编写于 作者: Z zxcd 提交者: GitHub

[ASR] Remove fluid api and useless import, test=asr (#2944)

* remove fluid api and useless import.

* fix variable name
上级 e8deb1d8
...@@ -12,7 +12,6 @@ ...@@ -12,7 +12,6 @@
# See the License for the specific language governing permissions and # See the License for the specific language governing permissions and
# limitations under the License. # limitations under the License.
from collections import defaultdict from collections import defaultdict
from turtle import Turtle
from typing import Dict from typing import Dict
from typing import List from typing import List
from typing import Tuple from typing import Tuple
......
...@@ -56,15 +56,15 @@ class ClipGradByGlobalNormWithLog(paddle.nn.ClipGradByGlobalNorm): ...@@ -56,15 +56,15 @@ class ClipGradByGlobalNormWithLog(paddle.nn.ClipGradByGlobalNorm):
if len(sum_square_list) == 0: if len(sum_square_list) == 0:
return params_grads return params_grads
global_norm_var = layers.concat(sum_square_list) global_norm_var = paddle.concat(sum_square_list)
global_norm_var = paddle.sum(global_norm_var) global_norm_var = paddle.sum(global_norm_var)
global_norm_var = paddle.sqrt(global_norm_var) global_norm_var = paddle.sqrt(global_norm_var)
# debug log # debug log
logger.debug(f"Grad Global Norm: {float(global_norm_var)}!!!!") logger.debug(f"Grad Global Norm: {float(global_norm_var)}!!!!")
max_global_norm = layers.fill_constant( max_global_norm = paddle.full(
shape=[1], dtype=global_norm_var.dtype, value=self.clip_norm) shape=[1], dtype=global_norm_var.dtype, fill_value=self.clip_norm)
clip_var = paddle.divide( clip_var = paddle.divide(
x=max_global_norm, x=max_global_norm,
y=paddle.maximum(x=global_norm_var, y=max_global_norm)) y=paddle.maximum(x=global_norm_var, y=max_global_norm))
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册