未验证 提交 8173ff94 编写于 作者: M mapingshuo 提交者: GitHub

delete unused log in Mul grad (#3243)

* rm grad code
上级 8c108172
...@@ -155,18 +155,6 @@ class MulGradTester { ...@@ -155,18 +155,6 @@ class MulGradTester {
fill_data_rand(y.data(), -1.f, 1.f, y_dims_.production()); fill_data_rand(y.data(), -1.f, 1.f, y_dims_.production());
this->run_forward(&param_, &kernel_, x, y, out.data()); this->run_forward(&param_, &kernel_, x, y, out.data());
for (int i = 0; i < x_dims_.production(); i++) {
LOG(INFO) << "x_" << i << ": " << x[i];
}
for (int i = 0; i < y_dims_.production(); i++) {
LOG(INFO) << "y_" << i << ": " << y[i];
}
for (int i = 0; i < out_dims_.production(); i++) {
LOG(INFO) << "out_" << i << ": " << out[i];
}
// backward // backward
std::vector<float> out_grad(out_dims_.production()); std::vector<float> out_grad(out_dims_.production());
std::vector<float> x_grad(x_dims_.production()); std::vector<float> x_grad(x_dims_.production());
...@@ -190,12 +178,7 @@ class MulGradTester { ...@@ -190,12 +178,7 @@ class MulGradTester {
float delta = 0.001; float delta = 0.001;
float max_grad_delta = 0.005; float max_grad_delta = 0.005;
for (int i = 0; i < x_dims_.production(); i++) { for (int i = 0; i < x_dims_.production(); i++) {
LOG(INFO) << "--------------------";
LOG(INFO) << "delta: " << delta;
LOG(INFO) << "max_grad_delta: " << max_grad_delta;
for (int j = 0; j < x_dims_.production(); j++) { for (int j = 0; j < x_dims_.production(); j++) {
// x_delta[j] = i == j ? x[j] + delta : x[j];
if (i == j) { if (i == j) {
x_delta[j] = x[j] + delta; x_delta[j] = x[j] + delta;
} else { } else {
...@@ -204,26 +187,12 @@ class MulGradTester { ...@@ -204,26 +187,12 @@ class MulGradTester {
} }
this->run_forward( this->run_forward(
&delta_param_, &delta_kernel_, x_delta, y, out_delta.data()); &delta_param_, &delta_kernel_, x_delta, y, out_delta.data());
for (int j = 0; j < x_dims_.production(); j++) {
LOG(INFO) << "x_" << j << ": " << x[j];
LOG(INFO) << "x_delta_" << j << ": " << x_delta[j];
}
for (int j = 0; j < y_dims_.production(); j++) {
LOG(INFO) << "y_" << j << ": " << y[j];
}
for (int j = 0; j < out_dims_.production(); j++) {
LOG(INFO) << "out_delta_" << j << ": " << out_delta[j];
}
float sum = 0; float sum = 0;
for (int j = 0; j < out_dims_.production(); j++) { for (int j = 0; j < out_dims_.production(); j++) {
sum += (out_delta[j] - out[j]); sum += (out_delta[j] - out[j]);
} }
LOG(INFO) << "x_grad_" << i << ": " << x_grad[i];
LOG(INFO) << "x_grad_num_" << i << ": " << sum / delta;
EXPECT_NEAR(x_grad[i], sum / delta, max_grad_delta); EXPECT_NEAR(x_grad[i], sum / delta, max_grad_delta);
} }
...@@ -237,8 +206,7 @@ class MulGradTester { ...@@ -237,8 +206,7 @@ class MulGradTester {
for (int j = 0; j < out_dims_.production(); j++) { for (int j = 0; j < out_dims_.production(); j++) {
sum += out_delta[j] - out[j]; sum += out_delta[j] - out[j];
} }
LOG(INFO) << "y_grad_" << i << ": " << y_grad[i];
LOG(INFO) << "y_grad_num_" << i << ": " << sum / delta;
EXPECT_NEAR(y_grad[i], sum / delta, max_grad_delta); EXPECT_NEAR(y_grad[i], sum / delta, max_grad_delta);
} }
} }
...@@ -265,8 +233,7 @@ void TestNormalCase(const std::vector<int64_t>& x_dims, ...@@ -265,8 +233,7 @@ void TestNormalCase(const std::vector<int64_t>& x_dims,
DDim(x_dims), DDim(y_dims), x_num_col_dims, y_num_col_dims)); DDim(x_dims), DDim(y_dims), x_num_col_dims, y_num_col_dims));
tester->prepare_kernel(); tester->prepare_kernel();
float delta = 0.001;
float max_grad_delta = 0.005;
tester->check_grad(); tester->check_grad();
} }
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册