Skip to yearly menu bar Skip to main content


Poster
in
Workshop: HiLD: High-dimensional Learning Dynamics Workshop

Benign Overfitting of Two-Layer Neural Networks under Inputs with Intrinsic Dimension

Shunta Akiyama · Kazusato Oko · Taiji Suzuki


Abstract:

Contrary to classical statistical theory, machine learning models with enormous sizes have shown high performances even when they interpolate the data. Such a phenomenon is called ``benign overfitting'' and has attracted much attention in the theoretical literature. Recent studies have clarified its theoretical perspective mostly in linear models, and there are yet only a few results for neural networks with feature learning. To address this issue, we theoretically investigate the statistical property of two-layer neural networks trained by noisy gradient descent in the setting where inputs have an intrinsic structure with lower dimensionality. We show when a true model is given by another neural network, the trained network can obtain the intrinsic feature of the true model through the gradient based training and eventually achieve benign overfitting.

Chat is not available.