Home
Article
Vanishing Gradient Demo
Number of Layers
3 layers
5 layers
8 layers
10 layers
Activation Function
Sigmoid (max derivative: 0.25)
Tanh (max derivative: 1.0)
ReLU (derivative: 1.0)
Sigmoid
ReLU