Knowledge-Enhanced Scene Graph Generation with Multimodal Relation Alignment (Student Abstract)
Ze Fu, Junhao Feng, Changmeng Zheng, Yi Cai
[AAAI-22] Student Abstract and Poster Program
Abstract:
Existing scene graph generation methods suffer the limitations when the image lacks of sufficient visual contexts. To address this limitation, we propose a knowledge-enhanced scene graph generation model with multimodal relation alignment, which supplements the missing visual contexts by well-aligned textual knowledge. First, we represent the textual information into contextualized knowledge which is guided by the visual objects to enhance the contexts. Furthermore, we align the multimodal relation triplets by co-attention module for better semantics fusion. The experimental results show the effectiveness of our method.
Sessions where this paper appears
-
Poster Session 6
Sat, February 26 8:45 AM - 10:30 AM (+00:00)
Red 6
-
Poster Session 11
Mon, February 28 12:45 AM - 2:30 AM (+00:00)
Red 6