Abstract. Introduces He (Kaiming) initialisation, derived for ReLU networks, and the parametric ReLU activation. Demonstrates that correctly scaled initialisation enables very deep networks to train from scratch without degradation.
Tags:neural-networksinitialisationrelu
This site is currently in Beta. Contact: Chris Paton