[8:30]
Self-Repellent Random Walks on General Graphs - Achieving Minimal Sampling Variance via Nonlinear Markov Chains
[8:38]
Tighter Lower Bounds for Shuffling SGD: Random Permutations and Beyond
[8:46]
Which Features are Learnt by Contrastive Learning? On the Role of Simplicity Bias in Class Collapse and Feature Suppression
[8:54]
Tighter Information-Theoretic Generalization Bounds from Supersamples
[9:02]
Beyond the Universal Law of Robustness: Sharper Laws for Random Features and Neural Tangent Kernels
[9:10]
Bayes-optimal Learning of Deep Random Networks of Extensive-width
[9:18]
Why does Throwing Away Data Improve Worst-Group Error?
[9:26]
Marginalization is not Marginal: No Bad VAE Local Minima when Learning Optimal Sparse Representations
[9:34]
Sharper Bounds for $\ell_p$ Sensitivity Sampling
[9:42]
AdaBoost is not an Optimal Weak to Strong Learner
[9:50]
Generalization on the Unseen, Logic Reasoning and Degree Curriculum