How to write a new operator¶
Background¶
Here are the base types needed. For details, please refer to the design docs.
framework::OperatorBase
: Operator (Op)base class.framework::OpKernel
: Base class for Op computation.framework::OperatorWithKernel
: Inherited from OperatorBase, describing an operator with computation.class OpProtoAndCheckerMaker
: Describes an Operator’s input, output, attributes and description, mainly used to interface with Python API.
An operator can be differentiated by whether in has kernel methods. An operator with kernel inherits from OperatorWithKernel
while the ones without inherit from OperatorBase
. This tutorial focuses on implementing operators with kernels. In short, an operator includes the following information:
Information | Where is it defined
————– | :———————-
OpProtoMake definition | .cc
files, Backward Op does not need an OpProtoMake interface.
Op definition | .cc
files
Kernel implementation | The kernel methods shared between CPU and GPU are defined in .h
files. CPU-specific kernels live in .cc
files, while GPU-specific kernels are implemented in .cu
files.
Registering the Op | Ops are registered in .cc
files; For Kernel registration, .cc
files contain the CPU implementation, while .cu
files contain the GPU implementation.
New Operator implementations are added to the list paddle/operators, with file names in the format *_op.h
(if applicable), *_op.cc
, *_op.cu
(if applicable).** The system will use the naming scheme to automatically build operators and their corresponding Python extensions. **
Let’s take matrix multiplication operator, MulOp, as an example to introduce the writing of an Operator with Kernel.
Implementing C++ Types¶
1. Defining Class ProtoMaker¶
Matrix Multiplication can be written as $Out = X * Y$, meaning that the operation consists of two inputs and pne output.
First, define ProtoMaker
to describe the Operator’s input, output, and additional comments:
class MulOpMaker : public framework::OpProtoAndCheckerMaker {
public:
MulOpMaker(framework::OpProto *proto, framework::OpAttrChecker *op_checker)
: OpProtoAndCheckerMaker(proto, op_checker) {
AddInput("X", "(Tensor), 2D tensor of size (M x K)");
AddInput("Y", "(Tensor), 2D tensor of size (K x N)");
AddOutput("Out", "(Tensor), 2D tensor of size (M x N)");
AddComment(R"DOC(
Two Element Mul Operator.
The equation is: Out = X * Y
)DOC");
}
};
MulOpMaker
is inherited fromframework::OpProtoAndCheckerMaker
, consisting of 2 variables in the constructor:
framework::OpProto
stores Operator input and variable attribute, used for generating Python API interfaces.framework::OpAttrChecker
is used to validate variable attributes.
The constructor utilizes AddInput
, AddOutput
, and AddComment
, so that the corresponding information will be added to OpProto
.
The code above adds two inputs X
and Y
to MulOp
, an output Out
, and their corresponding descriptions, in accordance to Paddle’s naming convention.
An additional example ScaleOp
is implemented as follows:
template <typename AttrType>
class ScaleOpMaker : public framework::OpProtoAndCheckerMaker {
public:
ScaleOpMaker(framework::OpProto *proto, framework::OpAttrChecker *op_checker)
: OpProtoAndCheckerMaker(proto, op_checker) {
AddInput("X", "The input tensor of scale operator.").NotInGradient();
AddOutput("Out", "The output tensor of scale operator.").NotInGradient();
AddComment(R"DOC(Scale operator
The equation is: Out = scale*X
)DOC");
AddAttr<AttrType>("scale", "scale of scale operator.").SetDefault(1.0);
}
};
There are two changes in this example:
AddInput("X","...").NotInGradient()
expresses that inputX
is not involved inScaleOp
‘s corresponding computation. If an input to an operator is not participating in back-propagation, please explicitly set.NotInGradient()
.AddAttr<AttrType>("scale", "...").SetDefault(1.0);
addsscale
constant as an attribute, and sets the default value to 1.0.
2. Defining Operator¶
The following code defines the interface for MulOp:
class MulOp : public framework::OperatorWithKernel {
public:
using framework::OperatorWithKernel::OperatorWithKernel;
protected:
void InferShape(const framework::InferShapeContext &ctx) const override {
auto dim0 = ctx.Input<Tensor>("X")->dims();
auto dim1 = ctx.Input<Tensor>("Y")->dims();
PADDLE_ENFORCE_EQ(dim0.size(), 2,
"input X(%s) should be a tensor with 2 dims, a matrix",
ctx.op_.Input("X"));
PADDLE_ENFORCE_EQ(dim1.size(), 2,
"input Y(%s) should be a tensor with 2 dims, a matrix",
ctx.op_.Input("Y"));
PADDLE_ENFORCE_EQ(
dim0[1], dim1[0],
"First matrix's width must be equal with second matrix's height.");
ctx.Output<Tensor>("Out")->Resize({dim0[0], dim1[1]});
}
};
MulOp
is inherited from OperatorWithKernel
. Its public
member
using framework::OperatorWithKernel::OperatorWithKernel;
expresses an operator constructor using base class OperatorWithKernel
, alternatively written as
MulOp(const std::string &type, const framework::VariableNameMap &inputs,
const framework::VariableNameMap &outputs,
const framework::AttributeMap &attrs)
: OperatorWithKernel(type, inputs, outputs, attrs) {}
InferShape
interface needs to be re-written.InferShape
is a constant method and cannot modify Op’s member variables, its constant member const framework::InferShapeContext &ctx
can be used to extract input, output, and attributes. It functions to
- 1). validate and error out early: it checks input data dimensions and types.
- 2). configures the tensor shape in the output.
Usually OpProtoMaker
and Op
‘s type definitions are written in .cc
files, which also include the registration methods introduced later.
3. Defining OpKernel¶
MulKernel
inherits framework::OpKernel
, which includes the following templates:
typename Place
denotes device type. When different devices, namely the CPU and the GPU, share the same kernel, this template needs to be added. If they don’t share kernels, this must not be added. An example of a non-sharing kernel isOnehotCrossEntropyOpKernel
.typename T
denotes data type, such asfloat
ordouble
.
MulKernel
types need to rewrite the interface for Compute
.
Compute
takes one input variableconst framework::ExecutionContext& context
.- Compared with
InferShapeContext
,ExecutionContext
includes device types, and can similarly extract input, output, and attribute variables. Compute
implements the computation logics of anOpKernel
.
MulKernel
‘s implementation of Compute
is as follows:
template <typename Place, typename T>
class MulKernel : public framework::OpKernel {
public:
void Compute(const framework::ExecutionContext& context) const override {
auto* X = context.Input<Tensor>("X");
auto* Y = context.Input<Tensor>("Y");
auto* Z = context.Output<Tensor>("Out");
Z->mutable_data<T>(context.GetPlace());
auto* device_context =
const_cast<platform::DeviceContext*>(context.device_context_);
math::matmul<Place, T>(*X, false, *Y, false, 1, Z, 0, device_context);
}
};
Note that different devices (CPU, GPU)share an Op definition; whether or not they share the same OpKernel
depends on whether Compute
calls functions that support both devices.
MulOp
‘s CPU and GPU share the same Kernel
. A non-sharing OpKernel
example can be seen in OnehotCrossEntropyOpKernel
.
To ease the writing of OpKernel
compute, and for reusing code cross-device, Eigen unsupported Tensor
module is used to implement Compute
interface. To learn about how the Eigen library is used in PaddlePaddle, please see usage document.
This concludes the forward implementation of an operator. Next its operation and kernel need to be registered in a .cc
file.
The definition of its corresponding backward operator, if applicable, is similar to that of an forward operator. Note that a backward operator does not include a ProtoMaker
.
4. Registering Operator¶
In
.cc
files, register forward and backward operator classes and the CPU kernel.namespace ops = paddle::operators; REGISTER_OP(mul, ops::MulOp, ops::MulOpMaker, mul_grad, ops::MulOpGrad); REGISTER_OP_CPU_KERNEL(mul, ops::MulKernel<paddle::platform::CPUPlace, float>); REGISTER_OP_CPU_KERNEL(mul_grad, ops::MulGradKernel<paddle::platform::CPUPlace, float>);
In that code block,
REGISTER_OP
registers theops::MulOp
class, type namedmul
, its typeProtoMaker
isops::MulOpMaker
, registeringops::MulOpGrad
asmul_grad
.REGISTER_OP_WITHOUT_GRADIENT
registers an operator without gradient.REGISTER_OP_CPU_KERNEL
registersops::MulKernel
class and specialized template typespaddle::platform::CPUPlace
andfloat
, which also registersops::MulKernel
.
Registering GPU Kernel in
.cu
files- Note that if GPU Kernel is implemented using the
Eigen unsupported
module, then on top of.cu
, a macro definition#define EIGEN_USE_GPU
is needed, such as
// if use Eigen unsupported module before include head files #define EIGEN_USE_GPU namespace ops = paddle::operators; REGISTER_OP_GPU_KERNEL(mul, ops::MulKernel<paddle::platform::GPUPlace, float>); REGISTER_OP_GPU_KERNEL(mul_grad, ops::MulGradKernel<paddle::platform::GPUPlace, float>);
- Note that if GPU Kernel is implemented using the
Python Binding¶
The system will automatically bind to Python and link it to a generated library.
Unit Tests¶
Unit tests include comparing a forward operator’s implementations on different devices, comparing a backward operator’s implementation on different devices, and a scaling test for the backward operator. Here, we introduce the unit tests for MulOp
.