提交 e4024962 编写于 作者: C chenweihang

complete unsqueeze op and related unittest.

上级 a1e7f2d5
...@@ -32,42 +32,85 @@ class UnsqueezeOp : public framework::OperatorWithKernel { ...@@ -32,42 +32,85 @@ class UnsqueezeOp : public framework::OperatorWithKernel {
PADDLE_ENFORCE(ctx->HasOutput("Out"), PADDLE_ENFORCE(ctx->HasOutput("Out"),
"Output(Out) of UnsqueezeOp should not be null."); "Output(Out) of UnsqueezeOp should not be null.");
const auto& x_dims = ctx->GetInputDim("X");
const auto& axes = ctx->Attrs().Get<std::vector<int>>("axes"); const auto& axes = ctx->Attrs().Get<std::vector<int>>("axes");
// Check output tensor dims (<9). PADDLE_ENFORCE(!axes.empty(),
PADDLE_ENFORCE_LE(x_dims.size() + axes.size(), 9, "The unsqueeze axes information must be set by Attr(axes).");
"Invalid dimnesions, dynamic dimensions must have "
"between [1, 9] dimensions."); const auto& x_dims = ctx->GetInputDim("X");
// Check the range of unsqueeze aixs. // Validity Check: input tensor dims (<6).
for (int a : axes) { PADDLE_ENFORCE(x_dims.size() < 6,
PADDLE_ENFORCE_LT(a, static_cast<int64_t>(x_dims.size() + axes.size()), "Invalid dimensions, dynamic dimensions should within "
"The axis must be less than output tensor's rank."); "[0, 5] dimensions (Eigen limit).");
// Validity Check: the range of unsqueeze aixs.
// TODO(chenweihang): Don't consider negative axis?.
for (unsigned int idx = 0; idx < axes.size(); ++idx) {
PADDLE_ENFORCE(axes[idx] < 6,
"Invalid dimensions, input axis should within "
"[0, 5] dimensions (Eigen limit).");
} }
auto out_dims = GetOutputShape(axes, x_dims); auto out_dims = GetOutputShape(axes, x_dims);
ctx->SetOutputDim("Out", out_dims); ctx->SetOutputDim("Out", out_dims);
} }
static framework::DDim GetOutputShape(const std::vector<int> unsqueeze_dims, static framework::DDim GetOutputShape(const std::vector<int> unsqz_dims,
const framework::DDim& in_dims) { const framework::DDim& in_dims) {
int out_dims_size = in_dims.size() + unsqueeze_dims.size(); /*
bool should_unsqueeze[9] = {false}; * STL version
* Test Error! don't know why?.
// Determines the dimensions should be unsqueezed in output tensor after. std::vector<int64_t> output_shape;
for (unsigned int idx = 0; idx < unsqueeze_dims.size(); ++idx) {
int current = unsqueeze_dims[idx] < 0 // Contruct base output shape
? unsqueeze_dims[idx] + out_dims_size for(int idx = 0; idx < in_dims.size(); ++idx) {
: unsqueeze_dims[idx]; output_shape.emplace_back(in_dims[idx]);
// Check current index. }
PADDLE_ENFORCE_GE(current, 0, // Validity Check: output dimensions limit.
"Invaild axis, negative axis is out of range."); PADDLE_ENFORCE(unsqz_dims.size() + output_shape.size() < 6,
should_unsqueeze[idx] = true; "The Attr(axes) size is too large. The output shape should "
"be less than 6 (Eigne limit).");
// Insert the unsqueeze axis in turn.
auto it = output_shape.begin();
for (int axis : unsqz_dims) {
int cur = axis < 0 ? (axis + output_shape.size() + 1)
: axis;
// Vaildity Check: the axis bound
PADDLE_ENFORCE(cur >= 0 && cur <= static_cast<int>(output_shape.size()),
"The unsqueeze dims must be within range of current
rank.");
output_shape.emplace(it + axis, 1);
}
*/
unsigned int unsqz_mask = 0;
unsigned int front = 0, back = 0;
int output_dims_size = in_dims.size();
// Simulate insert by bit calc.
for (int axis : unsqz_dims) {
int cur = axis < 0 ? axis + output_dims_size + 1 : axis;
// Vaildity Check: the axis bound
PADDLE_ENFORCE(
cur >= 0 && cur <= output_dims_size,
"The unsqueeze dims must be within range of current rank.");
// Save the front part.
front = unsqz_mask & ((1 << axis) - 1);
// Move the back part.
back = unsqz_mask & ~((1 << axis) - 1);
back <<= 1;
// Merge two part.
back |= (1 << axis);
unsqz_mask = front | back;
// Add the output size.
output_dims_size++;
// Validity Check: rank range.
PADDLE_ENFORCE(output_dims_size < 6,
"The output tensor's rank should be less than 6.");
} }
// Make output dimensions // Make output shape
std::vector<int64_t> output_shape(out_dims_size, 0); std::vector<int64_t> output_shape(output_dims_size, 0);
for (int in_idx = 0, out_idx = 0; out_idx < out_dims_size; ++out_idx) { for (int in_idx = 0, out_idx = 0; out_idx < output_dims_size; ++out_idx) {
if (!should_unsqueeze[out_idx]) { if ((unsqz_mask & (1 << out_idx)) == 0) {
output_shape[out_idx] = in_dims[in_idx++]; output_shape[out_idx] = in_dims[in_idx++];
} else { } else {
output_shape[out_idx] = 1; output_shape[out_idx] = 1;
......
...@@ -18,12 +18,12 @@ limitations under the License. */ ...@@ -18,12 +18,12 @@ limitations under the License. */
namespace ops = paddle::operators; namespace ops = paddle::operators;
REGISTER_OP_CUDA_KERNEL( REGISTER_OP_CUDA_KERNEL(
squeeze, ops::UnsqueezeKernel<paddle::platform::CUDADeviceContext, float>, unsqueeze, ops::UnsqueezeKernel<paddle::platform::CUDADeviceContext, float>,
ops::UnsqueezeKernel<paddle::platform::CUDADeviceContext, double>, ops::UnsqueezeKernel<paddle::platform::CUDADeviceContext, double>,
ops::UnsqueezeKernel<paddle::platform::CUDADeviceContext, int>, ops::UnsqueezeKernel<paddle::platform::CUDADeviceContext, int>,
ops::UnsqueezeKernel<paddle::platform::CUDADeviceContext, int64_t>); ops::UnsqueezeKernel<paddle::platform::CUDADeviceContext, int64_t>);
REGISTER_OP_CUDA_KERNEL( REGISTER_OP_CUDA_KERNEL(
squeeze_grad, unsqueeze_grad,
ops::UnsqueezeGradKernel<paddle::platform::CUDADeviceContext, float>, ops::UnsqueezeGradKernel<paddle::platform::CUDADeviceContext, float>,
ops::UnsqueezeGradKernel<paddle::platform::CUDADeviceContext, double>, ops::UnsqueezeGradKernel<paddle::platform::CUDADeviceContext, double>,
ops::UnsqueezeGradKernel<paddle::platform::CUDADeviceContext, int>, ops::UnsqueezeGradKernel<paddle::platform::CUDADeviceContext, int>,
......
...@@ -19,7 +19,7 @@ from op_test import OpTest ...@@ -19,7 +19,7 @@ from op_test import OpTest
# Correct: General. # Correct: General.
class TestSqueezeOp1(OpTest): class TestUnsqueezeOp(OpTest):
def setUp(self): def setUp(self):
ori_shape = (3, 5) ori_shape = (3, 5)
axes = (0, 2) axes = (0, 2)
...@@ -38,7 +38,7 @@ class TestSqueezeOp1(OpTest): ...@@ -38,7 +38,7 @@ class TestSqueezeOp1(OpTest):
# Correct: There is mins axis. # Correct: There is mins axis.
class TestSqueezeOp2(OpTest): class TestUnsqueezeOp2(OpTest):
def setUp(self): def setUp(self):
ori_shape = (3, 5) ori_shape = (3, 5)
axes = (0, -2) axes = (0, -2)
...@@ -56,6 +56,82 @@ class TestSqueezeOp2(OpTest): ...@@ -56,6 +56,82 @@ class TestSqueezeOp2(OpTest):
self.check_grad(["X"], "Out") self.check_grad(["X"], "Out")
# Correct: There is duplicated axis.
class TestUnsqueezeOp3(OpTest):
def setUp(self):
ori_shape = (3, 2, 5)
axes = (0, 3, 3)
new_shape = (1, 3, 2, 1, 1, 5)
self.op_type = "unsqueeze"
self.inputs = {"X": np.random.random(ori_shape).astype("float32")}
self.attrs = {"axes": axes, "inpalce": False}
self.outputs = {"Out": self.inputs["X"].reshape(new_shape)}
def test_check_output(self):
self.check_output()
def test_check_grad(self):
self.check_grad(["X"], "Out")
# Error: Output dimension is error.
class TestUnsqueezeOp4(OpTest):
def setUp(self):
ori_shape = (3, 2, 5)
axes = (0, 3)
new_shape = (1, 3, 2, 2, 5)
self.op_type = "unsqueeze"
self.inputs = {"X": np.random.random(ori_shape).astype("float32")}
self.attrs = {"axes": axes, "inpalce": False}
self.outputs = {"Out": self.inputs["X"].reshape(new_shape)}
def test_check_output(self):
self.check_output()
def test_check_grad(self):
self.check_grad(["X"], "Out")
# Error: Input axes is invalid case 1.
class TestUnsqueezeOp5(OpTest):
def setUp(self):
ori_shape = (3, 2, 5)
axes = (0, 5)
new_shape = (1, 3, 1, 5)
self.op_type = "unsqueeze"
self.inputs = {"X": np.random.random(ori_shape).astype("float32")}
self.attrs = {"axes": axes, "inpalce": False}
self.outputs = {"Out": self.inputs["X"].reshape(new_shape)}
def test_check_output(self):
self.check_output()
def test_check_grad(self):
self.check_grad(["X"], "Out")
# Error: Input axes is invalid case 2.
class TestUnsqueezeOp5(OpTest):
def setUp(self):
ori_shape = (3, 2, 5)
axes = (0, 2, 10)
new_shape = (1, 3, 1, 5)
self.op_type = "unsqueeze"
self.inputs = {"X": np.random.random(ori_shape).astype("float32")}
self.attrs = {"axes": axes, "inpalce": False}
self.outputs = {"Out": self.inputs["X"].reshape(new_shape)}
def test_check_output(self):
self.check_output()
def test_check_grad(self):
self.check_grad(["X"], "Out")
# Correct: Inplace. # Correct: Inplace.
class TestUnsqueezeOpInplace1(OpTest): class TestUnsqueezeOpInplace1(OpTest):
def setUp(self): def setUp(self):
...@@ -75,12 +151,12 @@ class TestUnsqueezeOpInplace1(OpTest): ...@@ -75,12 +151,12 @@ class TestUnsqueezeOpInplace1(OpTest):
self.check_grad(["X"], "Out") self.check_grad(["X"], "Out")
# Correct: Inplace. There is mins axis. # Correct: Inplace. There is duplicated axis.
class TestUnsqueezeOpInplace2(OpTest): class TestUnsqueezeOpInplace2(OpTest):
def setUp(self): def setUp(self):
ori_shape = (3, 5) ori_shape = (3, 2, 5)
axes = (0, -2) axes = (0, 3, 3)
new_shape = (1, 3, 1, 5) new_shape = (1, 3, 2, 1, 1, 5)
self.op_type = "unsqueeze" self.op_type = "unsqueeze"
self.inputs = {"X": np.random.random(ori_shape).astype("float32")} self.inputs = {"X": np.random.random(ori_shape).astype("float32")}
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册