Glorot presents in his paper ( Deep Sparse Rectifier Neural Networks ) his idea to penalize activation values with L1 norm in order to to use the relu function for an autoencoder.
Does anybody know how it is done?
[link][1 comment]
Glorot presents in his paper ( Deep Sparse Rectifier Neural Networks ) his idea to penalize activation values with L1 norm in order to to use the relu function for an autoencoder.
Does anybody know how it is done?