Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Localized Learning: Decentralized Model Updates via Non-Global Objectives

Preventing Dimensional Collapse in Contrastive Local Learning with Subsampling

Louis Fournier · Adeetya Patel · Michael Eickenberg · Edouard Oyallon · Eugene Belilovsky

Keywords: [ Local Learning ] [ self-supervised learning ] [ Representation Collapse ] [ Decoupled Learning ]


Abstract:

This paper presents an investigation of the challenges of training Deep Neural Networks (DNNs) via self-supervised objectives, using local learning as a parallelizable alternative to traditional backpropagation. In our approach, DNN are segmented into distinct blocks, each updated independently via gradients provided by small local auxiliary Neural Networks (NNs). Despite the evident computational benefits, extensive splits often result in performance degradation. Through analysis of a synthetic example, we identify a layer-wise dimensional collapse as a major factor behind such performance losses. To counter this, we propose a novel and straightforward sampling strategy based on blockwise feature-similarity, explicitly designed to evade such dimensional collapse.

Chat is not available.