Its differential is
The differential is undefined at
The ReLU activation function induces sparcity.
Its derivative is the sigmoid function:
The softplus function is a smooth approximation of the ReLU function.
Unlike the ReLU function, Softplus does not induce sparcity.
The error function for neural networks is nearly convex.
Gradients can be become small, and so propagation can be very slow.
Gradients can become too large and not converge.
This addresses the unstable gradient problem.
As parameter space gets bigger, data requirements get bigger.
Sparse models can help. Eg ReLU.
This is where the values in nodes are often