Activations¶
- 
class paddle::ActivationFunction¶
- Activation function is a function that transforms a set of input signals into an output signals. The purpose of the activation function is to introduce non-liearilty into the network. - Note
- Common activation function are provieded, including linear, sigmoid, softmax, sequence_max, relu, brelu, tanh, stanh, softrelu, abs, square, exponential.
 - Subclassed by paddle::IdentityActivation - Public Functions - 
ActivationFunction()¶
 - 
virtual ~ActivationFunction()¶
 - 
virtual void forward(Argument &act) = 0¶
- Foward propagation. - act.value <- f(act.value), where f is the activation function. Suppose that before calling forward(), act.value is x and after forward() is called, act.value is y, then y = f(x). - Usually, act is Layer::output_ 
 - 
virtual void backward(Argument &act) = 0¶
- Backward propagaion. - x and y are defined in the above comment for forward(). - Before calling backward(), act.grad = dE / dy, where E is the error/cost
- After backward() returns, act.grad = dE / dx = (dE/dy) * (dy/dx)
 
 - 
virtual const std::string &getName() const = 0¶
 - Public Static Functions - 
ActivationFunction *create(const std::string &type)¶