Poster
 in 
Workshop: 3rd Workshop on High-dimensional Learning Dynamics (HiLD)
                        
                    
                    Reduce and Conquer: Independent Component Analysis at linear sample complexity
Fabiola Ricci · Lorenzo Bardone · Sebastian Goldt
                        
                        
                            [ 
                            Abstract
                         ]
                        
                        
                        
                            [ Project Page  ]
                        
                        
                        
                        
                            
                                
                            
                            
                                
                            
                        
                        
                            
                            
                            
                                
                                     
                                
                                
                                presentation:
                                
                                    3rd Workshop on High-dimensional Learning Dynamics (HiLD)
                                
Fri 18 Jul 9 a.m. PDT — 5:30 p.m. PDT
                    
                    
                    
                                
                
                [
                 OpenReview] 
                
            
                            
                            Fri 18 Jul 9 a.m. PDT — 5:30 p.m. PDT
                        Abstract:
                        
                            Feature learning at the scale of deep neural networks remains poorly understood due to the complexity of deep network dynamics. Independent component analysis (ICA) provides a simple unsupervised model for feature learning, as it learns filters that are similar to deep networks. ICAextracts these features from the higher-order correlations of the inputs, which is a computationallyhard task in high dimensions with a sample complexity of at least $n ≳ D^2$for $D$-dimensionalinputs. In practice, this difficulty is overcome by running ICA in the d-dimensional subspacespanned by the leading principal components of the inputs, which is often taken to be $d = D/4$.However, there exist no theoretical guarantees for this procedure. Here, we first conduct systematicexperiments on ImageNet to demonstrate that running FastICA in a finite subspace of $d ∼ O_D(1)$dimensions yields non-Gaussian directions in the D-dimensional image space. We then introducea “subspace model” for synthetic data, and prove that FastICA does indeed recover the most nonGaussian direction in a sample complexity that is linear in the input dimension. We finally showexperimentally that deep convolutional networks trained on ImageNet exhibit behaviour consistentwith FastICA: during training, they converge to the principal subspace of image patches beforeor when they find non-Gaussian directions. By providing quantitative, rigorous insights into theworking of FastICA, our study thus unveils a plausible feature-learning mechanism in deep convolutional neural networks.
                        
                    
                    
                Chat is not available.
            
        Successful Page Load