Poster
Geometry of the Loss Landscape in Overparameterized Neural Networks: Symmetries and Invariances
Berfin Simsek · François Ged · Arthur Jacot · Francesco Spadaro · Clement Hongler · Wulfram Gerstner · Johanni Brea
Keywords: [ Theory ] [ Representation Learning ] [ Algorithms ] [ Algorithms -> Large Scale Learning; Applications -> Natural Language Processing; Deep Learning ] [ Efficient Inference Methods; ]
Abstract:
We study how permutation symmetries in overparameterized multi-layer neural
networks generate `symmetry-induced' critical points.
Assuming a network with layers of minimal widths reaches a zero-loss minimum at isolated points that are permutations of one another,
we show that adding one extra neuron to each layer is sufficient to connect all these previously discrete minima into a single manifold.
For a two-layer overparameterized network of width we explicitly describe the manifold of global minima: it consists of affine subspaces of dimension at least that are connected to one another.
For a network of width , we identify the number of affine subspaces containing only symmetry-induced critical points that are related to the critical points of a smaller network of width $r
Chat is not available.