Skip to yearly menu bar Skip to main content


Oral

On the Implicit Bias of Dropout

Poorya Mianjy · Raman Arora · Rene Vidal

Abstract:

Algorithmic approaches endow deep learning systems with implicit bias that helps them generalize even in over-parametrized settings. In this paper, we focus on understanding such a bias induced in learning through dropout, a popular technique to avoid overfitting in deep learning. For shallow linear neural networks, we show that dropout tends to make the norm of incoming/outgoing weight vectors of all the hidden nodes equal. We completely characterize the optimization landscape of single hidden-layer linear networks with dropout.

Chat is not available.