Position: Topological Machine Learning Cannot Progress without Experimental Standards
Abstract
Topological Machine Learning provides strong discriminative power for classification tasks through the use of Topological Data Analysis, and more particularly, Persistent Homology. Although it has strong theoretical appeal, it remains underused by the broader Machine Learning community; criticism often targets the reliance on synthetic data and the absence of shared experimental standards, which makes reported results difficult to compare. Indeed, current empirical evaluations lack a consistent framework for assessing methods: the construction of topological signatures is often opaque, statistical significance testing to validate reported gains, computing times and robustness to perturbations-such as missing data or noise-are often omitted. We assert that progress in Topological Machine Learning depends on establishing clear and consolidated experimental standards that support meaningful comparison across methods, articulated through a transparent and reproducible empirical framework including data processing and performance evaluation. We review current practices, highlight their limitations, and propose a set of principles for conducting rigorous and comparable empirical evaluations. Adopting these standards will enable trustworthy studies, clarify the gains of new methods, and ultimately support the broader adoption of Topological Machine Learning by the Machine Learning community.