Learning High-Dimensional Parity Functions with Product Networks using Gradient Descent
Guillaume Larue ⋅ Louis-Adrien Dufrène ⋅ Quentin Lampin ⋅ Hadi Ghauch ⋅ Ghaya Rekaya-Ben Othman
Abstract
Parity functions are fundamental Boolean operations with critical applications across machine learning, cryptography, and error correction. Yet, learning high-dimensional parity functions poses significant challenges: in a general setting, standard neural network architectures typically require exponential sample complexity, making gradient-based optimization intractable for large number of inputs $N$. We demonstrate that compact product-based neural architectures combined with stochastic data sparsity (Bernoulli inputs with $p_e \leq 1/N$) and appropriate hyperparameter choice enable efficient parity learning, with theoretical guarantees of convergence. Experiments validate our theory across dimensions up to $N = 10{,}000$, with empirical evidence showing optimal hyperparameter choices for $p_e$ and learning rate $\alpha$, as well as polynomial complexity scaling laws. This work establishes fundamental connections between architectural inductive bias and data sparsity, opening new possibilities for neural arithmetic, structured reasoning, binary neural networks, and machine learning applied to automated protocol discovery.
Successful Page Load