Breaking Multi-Task Curse: Reward-Weighted Evolution for Black-Box Many-Task Optimization
Abstract
Evolutionary multi-tasking accelerates black-box optimization via knowledge transfer but falters in scenarios involving many low-similarity tasks. We identify this scalability barrier as the Multi-Task Curse, driven by evaluation budget dispersion and negative transfer. To overcome this, we propose MES-RET (Many-task Evolution Strategy with Reward-weighted Evaluation and Transfer), which combats budget dispersion via a reward-weighted evaluation scheme that guarantees superior expected improvement, while simultaneously mitigating negative transfer through a robust reward-weighted aggregation of mean and covariance statistics, ensuring a safe fallback to independent evolution. Furthermore, to handle neural dimensional mismatches in many-task policy search, we introduce a semantic parameter alignment strategy that bridges heterogeneous state-action spaces. Extensive experiments on synthetic benchmarks, real-world engineering problems, and reinforcement learning tasks demonstrate that MES-RET consistently outperforms state-of-the-art methods, notably enabling skill transfer across morphologically distinct policies.