MutAtlas: A PDB-Wide Energy-Guided Atlas of Protein Mutation Effects
Abstract
Predicting protein mutation effects is fundamental to protein engineering and disease variant interpretation, yet experimental mutation data remain accurate but extremely sparse. Large-scale computational augmentation offers scalability, but introduces heterogeneous and poorly calibrated supervision signals arising from distinct modeling paradigms. We construct a PDB-wide, structure-aligned mutation augmentation dataset that exhaustively enumerates single-site substitutions and aligns mutation signals from physics-based energy models, protein language models, and structure-conditioned inverse folding models. Large-scale analysis under a unified mutation preference representation reveals substantial differences in preference structure, confidence, and cross-model agreement, indicating that disagreement is pervasive and reflects conflicting inductive biases rather than random noise. Motivated by these observations, we propose an unsupervised multi-source mutation preference distillation framework that learns from relative mutation preferences while explicitly modeling cross-source disagreement. Without using any experimental mutation labels during training, our approach achieves consistent improvements on the ProteinGym benchmark over zero-shot baselines and naive multi-source fusion strategies. We release the dataset and evaluation pipeline to support reproducible studies of protein mutation effects.