Timezone: »

How does Over-Parametrization Lead to Acceleration for Learning a Single Teacher Neuron with Quadratic Activation?
Jun-Kun Wang · Jake Abernethy

Over-parametrization has become a popular technique in deep learning. It is observed that by over-parametrization, a larger neural network needs a fewer training iterations than a smaller one to achieve a certain level of performance --- namely, over-parametrization leads to acceleration in optimization. However, despite that over-parametrization is widely used nowadays, little theory is available to explain the acceleration due to over-parametrization. In this paper, we propose understanding it by studying a simple problem first. Specifically, we consider the setting that there is a single teacher neuron with quadratic activation, where over-parametrization is realized by having multiple student neurons learn the data generated from the teacher neuron. We provably show that over-parametrization helps the iterate generated by gradient descent to enter the neighborhood of a global optimal solution that achieves zero testing error faster.

Author Information

Jun-Kun Wang (Georgia Institute of Technology)
Jake Abernethy (Georgia Institute of Technology)

More from the Same Authors