RGGT: A Generative-Prior-Guided Transformer for Unified Rigid and Non-Rigid Point Cloud Registration
Abstract
Point cloud registration can be categorized into rigid and non-rigid settings depending on the motion characteristics of the underlying objects. Rigid alignment assumes a single global transformation under which corresponding points remain geometrically consistent across scales, whereas non-rigid alignment involves spatially varying deformations, where geometric similarity holds only locally and semantic correspondence dominates at larger scales. This multi-scale discrepancy creates an optimization gap that has made unified registration particularly challenging. To this end, we propose RGGT, a Generative-Prior-Guided Transformer that unifies rigid and non-rigid registration within a shared optimization space. Through coordinated design at the representation, architecture, and supervision levels, RGGT jointly captures local geometric details and global structural semantics: generative priors enrich point features with unified geometric–semantic cues; a Global–Self–Cross Attention module models long-range structure, local interaction, and bidirectional cross-shape reasoning; and a dual correspondence–reconstruction objective provides consistent supervision for both deformation types. Extensive experiments on rigid (ModelNet40) and non-rigid (4DMatch) benchmarks demonstrate that RGGT achieves state-of-the-art accuracy across both rigid and non-rigid settings within a single unified framework.