One Batch Is Enough: A Unified Dataset Condensation Framework for General Time Series Analysis
Wei Shao ⋅ Ziquan Fang ⋅ Zheqi Lu ⋅ Yongfeng Su ⋅ Yuzhu Wang ⋅ Yunjun Gao
Abstract
Time-series analysis is critical in real-world applications, yet the explosion of time-series data imposes severe burdens on storage and computational resources. Recently, dataset condensation has emerged as a promising data-centric solution by synthesizing compact yet informative datasets to replace large-scale raw data. However, existing methods are largely vision-centric, failing to capture unique temporal properties of time series, or task-specific, tightly coupling the condensed data to a particular downstream objective. As a result, these approaches suffer from feature mismatch and fail to generalize across diverse time-series tasks. To bridge this gap, we propose UniTSC, the first unified dataset condensation framework for general time-series analysis. UniTSC employs a multi-view hybrid encoder to capture task-invariant representations across temporal, spectral, and topological perspectives. Building upon this representation, we design a tri-space alignment paradigm that jointly aligns optimization trajectories, power spectral densities, and multivariate dependency structures, enabling comprehensive information preservation under extreme compression. Extensive experiments show that UniTSC retains up to 97.9\% of downstream performance using as little as 0.01\% of the original training data, revealing that a single batch-equivalent budget ($\textless$ 128 samples) is sufficient to capture the essential dynamics of complex time-series data.
Successful Page Load