Authors
Vibashan VS, Jeya Maria Jose Valanarasu, Poojan Oza, Vishal M. Patel
Johns Hopkins University
Portals
Abstract
In image fusion, images obtained from different sensors are fused to generate a single image with enhanced information. In recent years, state-of-the-art methods have adopted Convolution Neural Networks (CNNs) to encode meaningful features for image fusion. Specifically, CNN-based methods perform image fusion by fusing local features. However, they do not consider long-range dependencies that are present in the image. Transformer-based models are designed to overcome this by modeling the long-range dependencies with the help of self-attention mechanism. This motivates us to propose a novel Image Fusion Transformer (IFT) where we develop a transformer-based multi-scale fusion strategy that attends to both local and long-range information (or global context). The proposed method follows a two-stage training approach. In the first stage, we train an auto-encoder to extract deep features at multiple scales. In the second stage, multi-scale features are fused using a Spatio-Transformer (ST) fusion strategy. The ST fusion blocks are comprised of a CNN and a transformer branch which capture local and long-range features, respectively. Extensive experiments on multiple benchmark datasets show that the proposed method performs better than many competitive fusion algorithms. Furthermore, we show the effectiveness of the proposed ST fusion strategy with an ablation analysis. The source code is available at: https://github.com/Vibashan/Image-Fusion-Transformer.
Contribution
- We propose a novel fusion method, called Image Fusion Transformer (IFT), that utilizes both local information and models long-range dependencies to overcome the lack of global contextual understanding that exists in recent image fusion works
- The proposed method utilizes a novel Spatio-Transformer (ST) fusion strategy, where a spatial CNN branch and a transformer branch are employed to utilize both local and global features to fuse the given images better
- The proposed method is evaluated on multiple fusion benchmark datasets, where we achieve competitive results compared to the existing fusion methods.