# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # TODO: define activation functions of neural network # __all__ = ['brelu', # 'elu', # 'erf', # 'gelu', # 'hard_shrink', # 'hard_sigmoid', # 'hard_swish', # 'hsigmoid', # 'leaky_relu', # 'logsigmoid', # 'maxout', # 'prelu', # 'relu', # 'relu6', # 'selu', # 'sigmoid', # 'soft_relu', # 'softmax', # 'softplus', # 'softshrink', # 'softsign', # 'swish', # 'tanh_shrink', # 'thresholded_relu', # 'log_softmax']