Loading…
A dual-path residual attention fusion network for infrared and visible images
Infrared image contains typical targets characterized by pixel intensity distribution, while visible image preserves structural details represented by edges and gradient. Infrared and visible image fusion aims to merge their respective key information, and produce a more comprehensive image with sup...
Saved in:
Published in: | Optik (Stuttgart) 2023-10, Vol.290, p.171251, Article 171251 |
---|---|
Main Authors: | , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Infrared image contains typical targets characterized by pixel intensity distribution, while visible image preserves structural details represented by edges and gradient. Infrared and visible image fusion aims to merge their respective key information, and produce a more comprehensive image with superior target perception and better detail representation. Most existing methods cannot reach this goal well, and are easy to generate a limited fused result. Towards this end, a dual-path residual attention fusion network, namely DRAFusion, is proposed in this paper. We construct a multi-scale dense network architecture, which adequately aggregates horizontal and vertical intermediate features at all scales and layers. In the fusion layer, we develop a residual attention fusion module, cascading channel-wise and spatial-wise attention models, which is used to model the global contextual dependencies, and interact with feature maps to more focus on their respective key information. In addition, we design a novel feature adaptation loss function to control the proportion of their key information in the training phase, which can keep a better balance between fused results and source images. Extensive experiments demonstrate that our DRAFusion produces a remarkable performance on the three testing datasets, and is superior to other methods in terms of qualitative visual description and quantitative evaluation validation. |
---|---|
ISSN: | 0030-4026 1618-1336 |
DOI: | 10.1016/j.ijleo.2023.171251 |