Position: Virtual Cells Need Context, Not Just Scale
Abstract
The intersection of AI and biology has entered a phase of explosive growth, driven by the ambition to build "Virtual Cells" or computational models capable of predicting cellular responses to any perturbation. Following the success of structural biology (e.g., AlphaFold) and large language models, the field has converged on training massive, high-capacity models on large-scale single-cell data. This position paper argues that scaling model capacity is insufficient to solve the Virtual Cell problem because the primary failure mode is a lack of adequate coverage over diverse biological contexts, not insufficient model expressivity. We support this claim by reviewing recent studies showing that simple baselines perform on par with sophisticated architectures within a given biological context, and current models fail to consistently generalize across contexts. We connect this finding to the causal inference literature on transportability and contrast it with domains where scaling has succeeded. We substantiate our argument through analysis of a state-of-the-art model on a 22-million-cell immunology dataset. We conclude that the community faces a causal transport problem that cannot be solved by accumulating more data from the same distributions. Instead, we argue that contextual diversity and causal representation learning deserve increased emphasis, complementing ongoing scaling of model capacity and data volume.