MobileFusion: Mobile-Friendly Infrared and Visible Image Fusion via Structural Re-parameterization
Abstract
Deep neural networks have recently advanced infrared and visible image fusion (IVIF), but most existing methods rely on sophisticated yet redundant designs, which hinder real-time deployment on mobile devices with limited compute and memory. In this paper, we present MobileFusion, an extremely lightweight and effective convolutional framework that achieves high-quality fusion under strict resource constraints. MobileFusion leverages a re-parameterizable multi-branch convolution module to promote cross-modal interactions during training while collapsing into a single-path operator for fast inference. It further incorporates a lightweight attention module to enhance context awareness, together with a re-parameterized feed-forward network to improve feature expressiveness. Extensive experiments demonstrate that MobileFusion delivers a favorable trade-off between fusion quality and computational efficiency, enabling real-time and high-quality IVIF on resource-constrained platforms.