Talk by Rachel Ward - Weighted Optimization: better generalization by smoother interpolation
Rachel Ward
2020 Talk
in
Workshop: Beyond first order methods in machine learning systems
in
Workshop: Beyond first order methods in machine learning systems
Abstract
We provide a rigorous analysis of how implicit bias towards smooth interpolations leads to low generalization error in the overparameterized setting. We provide the first case study of this connection through a random Fourier series model and weighted least squares. We then argue through this model and numerical experiments that normalization methods in deep learning such as weight normalization improve generalization in overparameterized neural networks by implicitly encouraging smooth interpolants. This is work with Yuege (Gail) Xie, Holger Rauhut, and Hung-Hsu Chou.
Video
Chat is not available.
Successful Page Load