Interact, Embed, and EnlargE: Boosting Modality-Specific Representations for Multi-Modal Person Re-Identification
Zi Wang, Chenglong Li, Aihua Zheng, Ran He, Jin Tang
[AAAI-22] Main Track
Abstract:
Multi-modal person Re-ID introduces more complementary information to assist the traditional Re-ID task. Existing multi-modal methods ignore the importance of modality-specific information in the feature fusion stage. To this end, we propose a novel method to boost modality-specific representations for multi-modal person Re-ID: Interact, Embed, and EnlargE (IEEE). First, we propose a cross-modal interacting module to exchange useful information between different modalities in the feature extraction phase. Second, we propose a relation-based embedding module to enhance the richness of feature descriptors by embedding the global feature into the fine-grained local information. Finally, we propose multi-modal margin loss to force the network to learn modality-specific information for each modality by enlarging the intra-class discrepancy. Superior performance on multi-modal Re-ID dataset RGBNT201 and three constructed Re-ID datasets validate the effectiveness of the proposed method comparing with the state-of-the-art approaches.
Introduction Video
Sessions where this paper appears
-
Poster Session 3
Fri, February 25 8:45 AM - 10:30 AM (+00:00)
Red 1
-
Poster Session 8
Sun, February 27 12:45 AM - 2:30 AM (+00:00)
Red 1