Affiliation:
1. University of Chinese Academy of Sciences, Beijing 101408, China
2. Beijing Institute of Remote Sensing Equipment, Beijing 100005, China
Abstract
The fusion of infrared and visible images aims to leverage the strengths of both modalities, thereby generating fused images with enhanced visible perception and discrimination capabilities. However, current image fusion methods frequently treat common features between modalities (modality-commonality) and unique features from each modality (modality-distinctiveness) equally during processing, neglecting their distinct characteristics. Therefore, we propose a DDFNet-A for infrared and visible image fusion. DDFNet-A addresses this limitation by decomposing infrared and visible input images into low-frequency features depicting modality-commonality and high-frequency features representing modality-distinctiveness. The extracted low and high features were then fused using distinct methods. In particular, we propose a hybrid attention block (HAB) to improve high-frequency feature extraction ability and a base feature fusion (BFF) module to enhance low-frequency feature fusion ability. Experiments were conducted on public infrared and visible image fusion datasets MSRS, TNO, and VIFB to validate the performance of the proposed network. DDFNet-A achieved competitive results on three datasets, with EN, MI, VIFF, QAB/F, FMI, and Qs metrics reaching the best performance on the TNO dataset, achieving 7.1217, 2.1620, 0.7739, 0.5426, 0.8129, and 0.9079, respectively. These values are 2.06%, 11.95%, 21.04%, 21.52%, 1.04%, and 0.09% higher than those of the second-best methods, respectively. The experimental results confirm that our DDFNet-A achieves better fusion performance than state-of-the-art (SOTA) methods.
Funder
National Natural Science Foundation of China
Fundamental Research Funds for the Central Universities
Reference79 articles.
1. An Adaptive Fusion Algorithm for Visible and Infrared Videos Based on Entropy and the Cumulative Distribution of Gray Levels;Hu;IEEE Trans. Multimedia,2017
2. Multisensor Image Fusion and Enhancement in Spectral Total Variation Domain;Zhao;IEEE Trans. Multimed.,2018
3. Feature Level Image Fusion of Optical Imagery and Synthetic Aperture Radar (SAR) for Invasive Alien Plant Species Detection and Mapping;Rajah;Remote Sens. Appl. Soc. Environ.,2018
4. A Dual-Domain Super-Resolution Image Fusion Method with SIRV and GALCA Model for PolSAR and Panchromatic Images;Liu;IEEE Trans. Geosci. Remote Sens.,2022
5. Zhang, X., Ye, P., Qiao, D., Zhao, J., Peng, S., and Xiao, G. (2019, January 2–5). Object Fusion Tracking Based on Visible and Infrared Images Using Fully Convolutional Siamese Networks. Proceedings of the 2019 22th International Conference on Information Fusion (FUSION), Ottawa, ON, Canada.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献