Skip to yearly menu bar Skip to main content


Poster

Sparse-to-dense Multimodal Image Registration via Multi-Task Learning

Kaining Zhang · Jiayi Ma


Abstract:

Aligning image pairs captured by different sensors or those undergoing significant appearance changes is crucial for various computer vision and robotics applications. Existing approaches cope with this problem via either Sparse feature Matching (SM) or Dense direct Alignment (DA) paradigms. Sparse methods are efficient but lack accuracy in textureless scenes, while dense ones are more accurate in all scenes but demand for good initialization. In this paper, we propose SDFE, a Sparse-to-Dense Feature Extractor based on a novel multi-task network that simultaneously predicts SM and DA features for robust multimodal image registration. We propose the sparse-to-dense registration paradigm: we first perform initial registration via SM and then refine the result via DA. By using the well-designed SDFE, the sparse-to-dense approach combines the merits from both SM and DA. Extensive experiments on MSCOCO, GoogleEarth, VIS-NIR and VIS-IR datasets demonstrate that our method achieves remarkable performance on both single-modal and multimodal cases.

Live content is unavailable. Log in and register to view live content