ReLU

ReLU

Rectified linear units, or reLU, are used as activation function in Deep Neural Networks. They are represented or defined as f(x) = max(x, 0), which means:

If input is negative or zero, output is 0.
If input is positive, output is equal to input.

The advantages of ReLUs over functions like tanh or sigmoid is that they are sparse and they deal well with vanishing gradient problem.

Varients of ReLU

  1. Leaky ReLUs
  2. Parametric ReLU (PReLU)
  3. Smoother softplus approximation
  1. Rectified Linear Units Improve Restricted Boltzmann Machines Paper and Summary
  2. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification Paper and Summary