Beyond Unidirectional Bias: Reciprocal Perspective Calibration in Scene Graph Generation
Haifeng Zhao ⋅ Wenbo Zhao ⋅ Xuemei Luo ⋅ Lei-Lei Ma ⋅ Dengdi Sun
Abstract
Scene Graph Generation (SGG) paradigms predominantly model relationships as static, unidirectional mappings ($s \to o$), effectively treating objects as passive recipients of actions. This formulation suffers from an inherent \textit{unidirectional bias}, violating the physical reality that visual interactions are intrinsically reciprocal. Consequently, existing models often fail to maintain logical self-consistency when the reasoning anchor shifts from the subject to the object. To rectify this cognitive deficiency, we establish the Mutual-Perspective Inverse Relations (MPIR) principle, positing that a robust visual representation must satisfy logical consistency across dual perspectives. Guided by this principle, we propose the \textbf{Reciprocal Perspective Calibration (RPC)} framework, a model-agnostic framework that operationalizes MPIR via a novel Adaptive Inverse-Relation Augmentation (AIRA) strategy. Furthermore, we introduce Hypernym-Guided Prompts (HGP) to bridge the gap between semantic context and computational efficiency in vision-language models, enabling precise modeling of inverse relations. Extensive experiments demonstrate that RPC not only achieves competitive performance on standard benchmarks but also significantly enhances the model's capability to understand inverse relations, as verified by a new inverse consistency evaluation protocol, demonstrating the cognitive robustness of our method.
Successful Page Load