常见的激活函数

1. sigmoid 
函数定义为:
Sigmoid: exists between 0 to 1, it is especially used for models we have to predict the probability as an output.
The function is differentiable and monotonic, but the function's derivative is not, which can cause a neural network to get stuck at the training time.

Softmax function is a more generalized logistic activation function which is used for mutliclass classification.


2、tanh函数 (tangent activation function)

The range of tanh function is from -1 to 1.
The advantage is that the negative inputs will be mapped strongly negative and the zero inputs will be mapped near zero in the tanh graph.

The function is also differentiable and monotonic while its derivative is not monotonic.

The tanh  function is mainly used in classification between two classes.

Both tanh and logistic sigmoid activation functions are used in feed-forward nets.

3. ReLU(Rectified Linear Unit) activation function

f(x) = max(0, x)
The ReLU  is half rectified. 
Range:[0, infinity)

The function and its derivative both are monotonic.

Problems: any negative input given to the ReLU activation function turns value to zero immediately in the graph.

4. Leaky ReLU: to solve the dying ReLU problem
Range: (-infinity, infinity)

Both Leaky and Randomized ReLU function and their derivatives are monotonic in nature.


Comments

Popular posts from this blog

Reading CLIP

Reading CutPaste

OOD-related papers