Loading…

MDAN: Multilevel dual-branch attention network for infrared and visible image fusion

Infrared and visible image fusion (IVIF) aims to integrate information captured by optical sensors operating in two different modalities, generating a fused image with both salient targets and texture details. Despite significant advancements in IVIF algorithms, the challenge of preserving complete...

Full description

Saved in:
Bibliographic Details
Published in:Optics and lasers in engineering 2024-05, Vol.176, p.108042, Article 108042
Main Authors: Wang, Jiawei, Jiang, Min, Kong, Jun
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c315t-ca7577ebaa16687b2057232febe7a47da8046fc875b6bbfc27393283ba85160a3
cites cdi_FETCH-LOGICAL-c315t-ca7577ebaa16687b2057232febe7a47da8046fc875b6bbfc27393283ba85160a3
container_end_page
container_issue
container_start_page 108042
container_title Optics and lasers in engineering
container_volume 176
creator Wang, Jiawei
Jiang, Min
Kong, Jun
description Infrared and visible image fusion (IVIF) aims to integrate information captured by optical sensors operating in two different modalities, generating a fused image with both salient targets and texture details. Despite significant advancements in IVIF algorithms, the challenge of preserving complete information, especially regarding texture details, still persists. To alleviate this problem, we propose a multilevel dual-branch attention network (MDAN) which comprises an encoder-decoder network and a fusion strategy layer composed of dual-branch fusion block (DBFB). Firstly, the encoder-decoder network is designed to extract multilevel image features and reconstruct the fused images. Secondly, a novel loss function based on singular value decomposition is proposed to constrain the reconstructed images to preserve abundant algebra features which reflect the structure and texture information of the source images. Thirdly, a fusion strategy layer based on spatial-channel attention and feature aggregation block, which consists of DBFB, is proposed to integrate the extracted features. Finally, we evaluate our method through qualitative and quantitative experiments, the results demonstrate that our method exhibits superiority in performance and achieves a remarkable balance between visual perception and objective evaluation metrics when compared to the state-of-the-art (SOTA) methods. •The Multilevel Dual-Branch Attention Network (MDAN) is proposed to extract multi-level features for generating fused images.•A fusion strategy layer based on spatial-channel attention and feature aggregation block is proposed to fused the extracted features.•A loss function based on singular value decomposition is designed to preserve the algebra features of the source images.•A large number of experiments prove the rationality and superiority of the proposed method.
doi_str_mv 10.1016/j.optlaseng.2024.108042
format article
fullrecord <record><control><sourceid>elsevier_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1016_j_optlaseng_2024_108042</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0143816624000228</els_id><sourcerecordid>S0143816624000228</sourcerecordid><originalsourceid>FETCH-LOGICAL-c315t-ca7577ebaa16687b2057232febe7a47da8046fc875b6bbfc27393283ba85160a3</originalsourceid><addsrcrecordid>eNqFkMtOwzAQRS0EEqXwDfgHUvxI4sCu4i1R2JS1NXbGxSU4yHaL-HtcFbFlNdJoztWdQ8g5ZzPOeHuxno2feYCEYTUTTNRl27FaHJAJ75SsmGTikEwYr2XV8bY9JicprVkha84nZLm4mT9f0cVmyH7ALQ6038BQmQjBvlHIGUP2Y6AB89cY36kbI_XBRYjYUwg93frkzYDUf8AKqdukcn1KjhwMCc9-55S83t0urx-qp5f7x-v5U2Ulb3JlQTVKoQEovTplBGuUkMKhQQW16qH80Trbqca0xjgrlLyUopMGuoa3DOSUqH2ujWNKEZ3-jKVH_Nac6Z0cvdZ_cvROjt7LKeR8T2Kpt_UYdbIeg8XeR7RZ96P_N-MHXHhysg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>MDAN: Multilevel dual-branch attention network for infrared and visible image fusion</title><source>Elsevier:Jisc Collections:Elsevier Read and Publish Agreement 2022-2024:Freedom Collection (Reading list)</source><creator>Wang, Jiawei ; Jiang, Min ; Kong, Jun</creator><creatorcontrib>Wang, Jiawei ; Jiang, Min ; Kong, Jun</creatorcontrib><description>Infrared and visible image fusion (IVIF) aims to integrate information captured by optical sensors operating in two different modalities, generating a fused image with both salient targets and texture details. Despite significant advancements in IVIF algorithms, the challenge of preserving complete information, especially regarding texture details, still persists. To alleviate this problem, we propose a multilevel dual-branch attention network (MDAN) which comprises an encoder-decoder network and a fusion strategy layer composed of dual-branch fusion block (DBFB). Firstly, the encoder-decoder network is designed to extract multilevel image features and reconstruct the fused images. Secondly, a novel loss function based on singular value decomposition is proposed to constrain the reconstructed images to preserve abundant algebra features which reflect the structure and texture information of the source images. Thirdly, a fusion strategy layer based on spatial-channel attention and feature aggregation block, which consists of DBFB, is proposed to integrate the extracted features. Finally, we evaluate our method through qualitative and quantitative experiments, the results demonstrate that our method exhibits superiority in performance and achieves a remarkable balance between visual perception and objective evaluation metrics when compared to the state-of-the-art (SOTA) methods. •The Multilevel Dual-Branch Attention Network (MDAN) is proposed to extract multi-level features for generating fused images.•A fusion strategy layer based on spatial-channel attention and feature aggregation block is proposed to fused the extracted features.•A loss function based on singular value decomposition is designed to preserve the algebra features of the source images.•A large number of experiments prove the rationality and superiority of the proposed method.</description><identifier>ISSN: 0143-8166</identifier><identifier>EISSN: 1873-0302</identifier><identifier>DOI: 10.1016/j.optlaseng.2024.108042</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Attention mechanism ; Deep learning ; Fusion strategy ; Infrared and visible image fusion ; Singular value decomposition</subject><ispartof>Optics and lasers in engineering, 2024-05, Vol.176, p.108042, Article 108042</ispartof><rights>2024 Elsevier Ltd</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c315t-ca7577ebaa16687b2057232febe7a47da8046fc875b6bbfc27393283ba85160a3</citedby><cites>FETCH-LOGICAL-c315t-ca7577ebaa16687b2057232febe7a47da8046fc875b6bbfc27393283ba85160a3</cites><orcidid>0000-0003-2551-4748 ; 0000-0003-3826-6405</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids></links><search><creatorcontrib>Wang, Jiawei</creatorcontrib><creatorcontrib>Jiang, Min</creatorcontrib><creatorcontrib>Kong, Jun</creatorcontrib><title>MDAN: Multilevel dual-branch attention network for infrared and visible image fusion</title><title>Optics and lasers in engineering</title><description>Infrared and visible image fusion (IVIF) aims to integrate information captured by optical sensors operating in two different modalities, generating a fused image with both salient targets and texture details. Despite significant advancements in IVIF algorithms, the challenge of preserving complete information, especially regarding texture details, still persists. To alleviate this problem, we propose a multilevel dual-branch attention network (MDAN) which comprises an encoder-decoder network and a fusion strategy layer composed of dual-branch fusion block (DBFB). Firstly, the encoder-decoder network is designed to extract multilevel image features and reconstruct the fused images. Secondly, a novel loss function based on singular value decomposition is proposed to constrain the reconstructed images to preserve abundant algebra features which reflect the structure and texture information of the source images. Thirdly, a fusion strategy layer based on spatial-channel attention and feature aggregation block, which consists of DBFB, is proposed to integrate the extracted features. Finally, we evaluate our method through qualitative and quantitative experiments, the results demonstrate that our method exhibits superiority in performance and achieves a remarkable balance between visual perception and objective evaluation metrics when compared to the state-of-the-art (SOTA) methods. •The Multilevel Dual-Branch Attention Network (MDAN) is proposed to extract multi-level features for generating fused images.•A fusion strategy layer based on spatial-channel attention and feature aggregation block is proposed to fused the extracted features.•A loss function based on singular value decomposition is designed to preserve the algebra features of the source images.•A large number of experiments prove the rationality and superiority of the proposed method.</description><subject>Attention mechanism</subject><subject>Deep learning</subject><subject>Fusion strategy</subject><subject>Infrared and visible image fusion</subject><subject>Singular value decomposition</subject><issn>0143-8166</issn><issn>1873-0302</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNqFkMtOwzAQRS0EEqXwDfgHUvxI4sCu4i1R2JS1NXbGxSU4yHaL-HtcFbFlNdJoztWdQ8g5ZzPOeHuxno2feYCEYTUTTNRl27FaHJAJ75SsmGTikEwYr2XV8bY9JicprVkha84nZLm4mT9f0cVmyH7ALQ6038BQmQjBvlHIGUP2Y6AB89cY36kbI_XBRYjYUwg93frkzYDUf8AKqdukcn1KjhwMCc9-55S83t0urx-qp5f7x-v5U2Ulb3JlQTVKoQEovTplBGuUkMKhQQW16qH80Trbqca0xjgrlLyUopMGuoa3DOSUqH2ujWNKEZ3-jKVH_Nac6Z0cvdZ_cvROjt7LKeR8T2Kpt_UYdbIeg8XeR7RZ96P_N-MHXHhysg</recordid><startdate>202405</startdate><enddate>202405</enddate><creator>Wang, Jiawei</creator><creator>Jiang, Min</creator><creator>Kong, Jun</creator><general>Elsevier Ltd</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0003-2551-4748</orcidid><orcidid>https://orcid.org/0000-0003-3826-6405</orcidid></search><sort><creationdate>202405</creationdate><title>MDAN: Multilevel dual-branch attention network for infrared and visible image fusion</title><author>Wang, Jiawei ; Jiang, Min ; Kong, Jun</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c315t-ca7577ebaa16687b2057232febe7a47da8046fc875b6bbfc27393283ba85160a3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Attention mechanism</topic><topic>Deep learning</topic><topic>Fusion strategy</topic><topic>Infrared and visible image fusion</topic><topic>Singular value decomposition</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Wang, Jiawei</creatorcontrib><creatorcontrib>Jiang, Min</creatorcontrib><creatorcontrib>Kong, Jun</creatorcontrib><collection>CrossRef</collection><jtitle>Optics and lasers in engineering</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Wang, Jiawei</au><au>Jiang, Min</au><au>Kong, Jun</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>MDAN: Multilevel dual-branch attention network for infrared and visible image fusion</atitle><jtitle>Optics and lasers in engineering</jtitle><date>2024-05</date><risdate>2024</risdate><volume>176</volume><spage>108042</spage><pages>108042-</pages><artnum>108042</artnum><issn>0143-8166</issn><eissn>1873-0302</eissn><abstract>Infrared and visible image fusion (IVIF) aims to integrate information captured by optical sensors operating in two different modalities, generating a fused image with both salient targets and texture details. Despite significant advancements in IVIF algorithms, the challenge of preserving complete information, especially regarding texture details, still persists. To alleviate this problem, we propose a multilevel dual-branch attention network (MDAN) which comprises an encoder-decoder network and a fusion strategy layer composed of dual-branch fusion block (DBFB). Firstly, the encoder-decoder network is designed to extract multilevel image features and reconstruct the fused images. Secondly, a novel loss function based on singular value decomposition is proposed to constrain the reconstructed images to preserve abundant algebra features which reflect the structure and texture information of the source images. Thirdly, a fusion strategy layer based on spatial-channel attention and feature aggregation block, which consists of DBFB, is proposed to integrate the extracted features. Finally, we evaluate our method through qualitative and quantitative experiments, the results demonstrate that our method exhibits superiority in performance and achieves a remarkable balance between visual perception and objective evaluation metrics when compared to the state-of-the-art (SOTA) methods. •The Multilevel Dual-Branch Attention Network (MDAN) is proposed to extract multi-level features for generating fused images.•A fusion strategy layer based on spatial-channel attention and feature aggregation block is proposed to fused the extracted features.•A loss function based on singular value decomposition is designed to preserve the algebra features of the source images.•A large number of experiments prove the rationality and superiority of the proposed method.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.optlaseng.2024.108042</doi><orcidid>https://orcid.org/0000-0003-2551-4748</orcidid><orcidid>https://orcid.org/0000-0003-3826-6405</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 0143-8166
ispartof Optics and lasers in engineering, 2024-05, Vol.176, p.108042, Article 108042
issn 0143-8166
1873-0302
language eng
recordid cdi_crossref_primary_10_1016_j_optlaseng_2024_108042
source Elsevier:Jisc Collections:Elsevier Read and Publish Agreement 2022-2024:Freedom Collection (Reading list)
subjects Attention mechanism
Deep learning
Fusion strategy
Infrared and visible image fusion
Singular value decomposition
title MDAN: Multilevel dual-branch attention network for infrared and visible image fusion
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-23T20%3A00%3A06IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=MDAN:%20Multilevel%20dual-branch%20attention%20network%20for%20infrared%20and%20visible%20image%20fusion&rft.jtitle=Optics%20and%20lasers%20in%20engineering&rft.au=Wang,%20Jiawei&rft.date=2024-05&rft.volume=176&rft.spage=108042&rft.pages=108042-&rft.artnum=108042&rft.issn=0143-8166&rft.eissn=1873-0302&rft_id=info:doi/10.1016/j.optlaseng.2024.108042&rft_dat=%3Celsevier_cross%3ES0143816624000228%3C/elsevier_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c315t-ca7577ebaa16687b2057232febe7a47da8046fc875b6bbfc27393283ba85160a3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true