未验证 提交 c2c9891b 编写于 作者: H hong 提交者: GitHub

Add no need buffer config (#41605) (#41706)

* Add no need buffer config (#41605)

* add no need buffer

* add no need buffer

* remove determinant no need buffer

* remove useless code
上级 17df5ddf
......@@ -207,6 +207,7 @@
args : (Tensor[] x, Tensor out_grad, Scalar axis = 0)
output : Tensor[](x_grad)
invoke : concat_grad_impl(x, out_grad, axis)
no_need_buffer : x
- backward_api : conj_grad
forward : conj (Tensor x) -> Tensor(out)
......@@ -422,6 +423,7 @@
param : [x]
kernel :
func : expand_as_grad
no_need_buffer : x
- backward_api : expm1_grad
forward : expm1 (Tensor x) -> Tensor(out)
......@@ -445,6 +447,7 @@
data_type: out_grad
backend: out_grad
layout: out_grad
no_need_buffer : x
- backward_api : flip_grad
forward : flip (Tensor x, int[] axis) -> Tensor(out)
......@@ -506,6 +509,7 @@
kernel :
data_type: x
func : gather_grad
no_need_buffer : x
- backward_api : gather_nd_grad
forward : gather_nd (Tensor x, Tensor index) -> Tensor(out)
......@@ -516,6 +520,7 @@
param : [x]
kernel :
func : gather_nd_grad
no_need_buffer : x
- backward_api : gelu_grad
forward : gelu(Tensor x, bool approximate) -> Tensor(out)
......@@ -616,6 +621,7 @@
kernel :
func : index_select_grad
data_type : x
no_need_buffer : x
- backward_api : kldiv_loss_grad
forward : kldiv_loss(Tensor x, Tensor label, str reduction) -> Tensor(out)
......@@ -626,6 +632,7 @@
param: [x]
kernel :
func : kldiv_loss_grad
no_need_buffer : x
- backward_api : kron_grad
forward : kron (Tensor x, Tensor y) -> Tensor(out)
......@@ -789,6 +796,7 @@
kernel :
func : masked_select_grad
data_type : x
no_need_buffer : x
- backward_api : matmul_double_grad
forward : matmul_grad (Tensor x, Tensor y, Tensor grad_out, bool transpose_x=false, bool transpose_y=false) -> Tensor(grad_x), Tensor(grad_y)
......@@ -891,6 +899,7 @@
param: [x]
kernel :
func : mean_grad
no_need_buffer : x
- backward_api : meshgrid_grad
forward : meshgrid (Tensor[] inputs) -> Tensor[](outputs)
......@@ -1172,6 +1181,7 @@
kernel :
func : roll_grad
data_type : x
no_need_buffer : x
- backward_api : round_grad
forward : round(Tensor x) -> Tensor(out)
......@@ -1314,6 +1324,7 @@
param : [input]
kernel :
func : slice_grad
no_need_buffer : input
- backward_api : soft_shrink_grad
forward : soft_shrink (Tensor x, float lambda) -> Tensor(out)
......@@ -1388,6 +1399,7 @@
param : [x]
kernel :
func : strided_slice_grad
no_need_buffer : x
- backward_api : subtract_grad
forward : subtract (Tensor x, Tensor y) -> Tensor(out)
......@@ -1409,6 +1421,7 @@
param : [x]
kernel :
func : sum_grad
no_need_buffer : x
- backward_api : swish_grad
forward : swish (Tensor x, float beta=1.0) -> Tensor(out)
......@@ -1568,3 +1581,4 @@
param : [x, y]
kernel :
func : where_grad
no_need_buffer : x, y
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册