Indexed by:
Abstract:
Fusing infrared and visible imagery aims to harness their complementary spectral data, enhancing output image quality, sharpness, and content. However, convolutional neural networks (CNNs) are not capable of capturing long-range dependencies, while the Transformer architecture faces the challenge of consuming huge computational resources. To address these critical challenges, this paper proposes a novel framework named MBHFuse, which employs a multi-branch heterogeneous global and local image fusion approach. To effectively extract global and local features, we design a multi-branch heterogeneous encoder module and introduce a differential convolution amplification module (DCAM) to further extract complementary information. Additionally, we devise a new loss function, incorporating multi-branch feature decomposition loss, intensity loss, gradient loss, mean squared error loss, and structural similarity loss, for training the proposed MBHFuse model. Through extensive experiments on public datasets, we demonstrate that the proposed framework outperforms other stateof-the-art (SOTA) methods in both qualitative and quantitative evaluations. Our code will be available at htt ps://github.com/sunyichen1994/MBHFuse.
Keyword:
Reprint Author's Address:
Email:
Source :
OPTICS AND LASER TECHNOLOGY
ISSN: 0030-3992
Year: 2024
Volume: 181
5 . 0 0 0
JCR@2022
Cited Count:
WoS CC Cited Count: 3
SCOPUS Cited Count: 3
ESI Highly Cited Papers on the List: 0 Unfold All
WanFang Cited Count:
Chinese Cited Count:
30 Days PV: 6
Affiliated Colleges: