Quantcast
Channel: Machine Learning
Viewing all articles
Browse latest Browse all 62750

L2 Regularization for multi-task deep neural network learning

$
0
0

I am trying to train a deep neural network, the hidden layers of which are shared between multiple tasks. My networks are pre-trained with RBM generative pre-training with data from all the tasks.

Though I do get reasonable error rates on my validation set, I see that my model overfits quickly on all the tasks. The overfitting behaviour is consistent amongst all of the tasks.

I tried a blanket L2 regularization for the soft-max layer only for all the tasks. Now my validation errors behave differently for each task. Though they do not necessarily overfit, on one task the classification performance is relatively bad. Is there any literature that talks about L2 regularization for multi-task learning with neural networks?

submitted by speechMachine
[link][1 comment]

Viewing all articles
Browse latest Browse all 62750

Trending Articles