Update backward appending stragety to support double backward and fix some bug. (#18104)
* Update backward.py: - If there is no input grad var in all outputs of previous ops, do not append this op into graph. - Only apply this stragety when double backward. * Update some double backward op. * Update sum_op to judge whether a tensor is empty by numel or IsInitialized().
Showing
想要评论请 注册 或 登录