Deep Learning: Assuming a deep neural network is properly regulated, can adding more layers actually make the performance degrade?

Deep Learning: Assuming a deep neural network is properly regulated, can adding more layers actually make the performance degrade?

I found this to be really puzzling. A deeper NN is supposed to be more powerful or at least equal to a shallower NN. I have already used dropout to prevent overfitting. How can the performance be degraded?
Yoshua's Answer
Yoshua Bengio 
Yoshua BengioMy lab has been one of the three that started the deep learning approach, bac...
Upvoted by Prateek TandonRobotics and Strong Artificial Intelligence Researcher• Paul KingComputational Neuroscientist, Technology Entrepreneur • Jack Rae,Google DeepMind Research Engineer
 
If you do not change the size of the layers and just add more layers, capacity should increase, so you could be overfitting. However, you should check whether training error increases or decreases. If it increases (which is also very plausible), it means that adding the layer made the optimization harder, with the optimization methods and initialization that you are using.  That could also explain your problem. However, if training error decreases and test error increases, you are overfitting.
posted @ 2015-08-04 19:55  菜鸡一枚  阅读(299)  评论(0编辑  收藏  举报