CodeMamba: Shifting from Target Semantics to Self-Supervised Background Manifold Learning for Singularity Detection in Infrared Sequences
Abstract
Multi-frame infrared small target detection suffers from extreme semantic paucity of targets and representation collapse due to overwhelming class imbalance, resulting in the persistent inability to accurately distinguish point-like targets from dynamic background clutter. To address these issues, we propose CodeMamba, a collaborative dual-stream framework that reframes this task as the complementary mechanisms of background manifold modeling and motion singularity capturing. The implicit stream emphasizes background regularity and anomaly localization, while the explicit stream focuses on motion consistency and spatiotemporal singularity. Finally, we design a Bayesian uncertainty-weighted fusion module that estimates the reliability of each stream by quantifying its observation noise. Extensive experiments on the IRDST and DAUB benchmarks demonstrate that CodeMamba not only outperforms existing methods but also achieves enhanced sensitivity to point-like targets.