Understanding Activation Functions in Neural Networks
Avinash Sharma V

woahh… buddy !! Written like a true prodigy…

Just to clarify.. so in a dying ReLu, the weight gets never updated coz the gradient becomes zero.. could you please explain this part with an intuitive example..

One clap, two clap, three clap, forty?

By clapping more or less, you can signal to us which stories really stand out.