TransMEF: A Transformer-Based Multi-Exposure Image Fusion Framework Using Self-Supervised Multi-Task Learning

Linhao Qu, Shaolei Liu, Manning Wang, Zhijian Song

[AAAI-22] Main Track
Abstract: In this paper, we propose TransMEF: a transformer-based multi-exposure image fusion framework via self-supervised multi-task learning. The framework is based on an encoder-decoder network, which can be trained on large natural datasets without the need of ground truth fusion images. We design three self-supervised reconstruction tasks according to the characteristics of multi-exposure images and conduct the three tasks simultaneously through multi-task learning, so that the network can learn the characteristics of multi-exposure images and extract more generalized features. In addition, to compensate for the defect in establishing long-range dependencies in CNN-based architectures, we design an encoder that combines a CNN module with a Transformer module to enable the network to focus on both local and global information. We evaluated our method and compared it to 11 competitive traditional and deep learning-based methods on a latest released multi-exposure image fusion benchmark dataset, and our method achieves the best performance in both subjective and objective evaluations.

Introduction Video

Sessions where this paper appears

  • Poster Session 5

    Sat, February 26 12:45 AM - 2:30 AM (+00:00)
    Red 3
    Add to Calendar

  • Poster Session 12

    Mon, February 28 8:45 AM - 10:30 AM (+00:00)
    Red 3
    Add to Calendar