Loading…
MDAN: Multilevel dual-branch attention network for infrared and visible image fusion
Infrared and visible image fusion (IVIF) aims to integrate information captured by optical sensors operating in two different modalities, generating a fused image with both salient targets and texture details. Despite significant advancements in IVIF algorithms, the challenge of preserving complete...
Saved in:
Published in: | Optics and lasers in engineering 2024-05, Vol.176, p.108042, Article 108042 |
---|---|
Main Authors: | , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c315t-ca7577ebaa16687b2057232febe7a47da8046fc875b6bbfc27393283ba85160a3 |
---|---|
cites | cdi_FETCH-LOGICAL-c315t-ca7577ebaa16687b2057232febe7a47da8046fc875b6bbfc27393283ba85160a3 |
container_end_page | |
container_issue | |
container_start_page | 108042 |
container_title | Optics and lasers in engineering |
container_volume | 176 |
creator | Wang, Jiawei Jiang, Min Kong, Jun |
description | Infrared and visible image fusion (IVIF) aims to integrate information captured by optical sensors operating in two different modalities, generating a fused image with both salient targets and texture details. Despite significant advancements in IVIF algorithms, the challenge of preserving complete information, especially regarding texture details, still persists. To alleviate this problem, we propose a multilevel dual-branch attention network (MDAN) which comprises an encoder-decoder network and a fusion strategy layer composed of dual-branch fusion block (DBFB). Firstly, the encoder-decoder network is designed to extract multilevel image features and reconstruct the fused images. Secondly, a novel loss function based on singular value decomposition is proposed to constrain the reconstructed images to preserve abundant algebra features which reflect the structure and texture information of the source images. Thirdly, a fusion strategy layer based on spatial-channel attention and feature aggregation block, which consists of DBFB, is proposed to integrate the extracted features. Finally, we evaluate our method through qualitative and quantitative experiments, the results demonstrate that our method exhibits superiority in performance and achieves a remarkable balance between visual perception and objective evaluation metrics when compared to the state-of-the-art (SOTA) methods.
•The Multilevel Dual-Branch Attention Network (MDAN) is proposed to extract multi-level features for generating fused images.•A fusion strategy layer based on spatial-channel attention and feature aggregation block is proposed to fused the extracted features.•A loss function based on singular value decomposition is designed to preserve the algebra features of the source images.•A large number of experiments prove the rationality and superiority of the proposed method. |
doi_str_mv | 10.1016/j.optlaseng.2024.108042 |
format | article |
fullrecord | <record><control><sourceid>elsevier_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1016_j_optlaseng_2024_108042</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0143816624000228</els_id><sourcerecordid>S0143816624000228</sourcerecordid><originalsourceid>FETCH-LOGICAL-c315t-ca7577ebaa16687b2057232febe7a47da8046fc875b6bbfc27393283ba85160a3</originalsourceid><addsrcrecordid>eNqFkMtOwzAQRS0EEqXwDfgHUvxI4sCu4i1R2JS1NXbGxSU4yHaL-HtcFbFlNdJoztWdQ8g5ZzPOeHuxno2feYCEYTUTTNRl27FaHJAJ75SsmGTikEwYr2XV8bY9JicprVkha84nZLm4mT9f0cVmyH7ALQ6038BQmQjBvlHIGUP2Y6AB89cY36kbI_XBRYjYUwg93frkzYDUf8AKqdukcn1KjhwMCc9-55S83t0urx-qp5f7x-v5U2Ulb3JlQTVKoQEovTplBGuUkMKhQQW16qH80Trbqca0xjgrlLyUopMGuoa3DOSUqH2ujWNKEZ3-jKVH_Nac6Z0cvdZ_cvROjt7LKeR8T2Kpt_UYdbIeg8XeR7RZ96P_N-MHXHhysg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>MDAN: Multilevel dual-branch attention network for infrared and visible image fusion</title><source>Elsevier:Jisc Collections:Elsevier Read and Publish Agreement 2022-2024:Freedom Collection (Reading list)</source><creator>Wang, Jiawei ; Jiang, Min ; Kong, Jun</creator><creatorcontrib>Wang, Jiawei ; Jiang, Min ; Kong, Jun</creatorcontrib><description>Infrared and visible image fusion (IVIF) aims to integrate information captured by optical sensors operating in two different modalities, generating a fused image with both salient targets and texture details. Despite significant advancements in IVIF algorithms, the challenge of preserving complete information, especially regarding texture details, still persists. To alleviate this problem, we propose a multilevel dual-branch attention network (MDAN) which comprises an encoder-decoder network and a fusion strategy layer composed of dual-branch fusion block (DBFB). Firstly, the encoder-decoder network is designed to extract multilevel image features and reconstruct the fused images. Secondly, a novel loss function based on singular value decomposition is proposed to constrain the reconstructed images to preserve abundant algebra features which reflect the structure and texture information of the source images. Thirdly, a fusion strategy layer based on spatial-channel attention and feature aggregation block, which consists of DBFB, is proposed to integrate the extracted features. Finally, we evaluate our method through qualitative and quantitative experiments, the results demonstrate that our method exhibits superiority in performance and achieves a remarkable balance between visual perception and objective evaluation metrics when compared to the state-of-the-art (SOTA) methods.
•The Multilevel Dual-Branch Attention Network (MDAN) is proposed to extract multi-level features for generating fused images.•A fusion strategy layer based on spatial-channel attention and feature aggregation block is proposed to fused the extracted features.•A loss function based on singular value decomposition is designed to preserve the algebra features of the source images.•A large number of experiments prove the rationality and superiority of the proposed method.</description><identifier>ISSN: 0143-8166</identifier><identifier>EISSN: 1873-0302</identifier><identifier>DOI: 10.1016/j.optlaseng.2024.108042</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Attention mechanism ; Deep learning ; Fusion strategy ; Infrared and visible image fusion ; Singular value decomposition</subject><ispartof>Optics and lasers in engineering, 2024-05, Vol.176, p.108042, Article 108042</ispartof><rights>2024 Elsevier Ltd</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c315t-ca7577ebaa16687b2057232febe7a47da8046fc875b6bbfc27393283ba85160a3</citedby><cites>FETCH-LOGICAL-c315t-ca7577ebaa16687b2057232febe7a47da8046fc875b6bbfc27393283ba85160a3</cites><orcidid>0000-0003-2551-4748 ; 0000-0003-3826-6405</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids></links><search><creatorcontrib>Wang, Jiawei</creatorcontrib><creatorcontrib>Jiang, Min</creatorcontrib><creatorcontrib>Kong, Jun</creatorcontrib><title>MDAN: Multilevel dual-branch attention network for infrared and visible image fusion</title><title>Optics and lasers in engineering</title><description>Infrared and visible image fusion (IVIF) aims to integrate information captured by optical sensors operating in two different modalities, generating a fused image with both salient targets and texture details. Despite significant advancements in IVIF algorithms, the challenge of preserving complete information, especially regarding texture details, still persists. To alleviate this problem, we propose a multilevel dual-branch attention network (MDAN) which comprises an encoder-decoder network and a fusion strategy layer composed of dual-branch fusion block (DBFB). Firstly, the encoder-decoder network is designed to extract multilevel image features and reconstruct the fused images. Secondly, a novel loss function based on singular value decomposition is proposed to constrain the reconstructed images to preserve abundant algebra features which reflect the structure and texture information of the source images. Thirdly, a fusion strategy layer based on spatial-channel attention and feature aggregation block, which consists of DBFB, is proposed to integrate the extracted features. Finally, we evaluate our method through qualitative and quantitative experiments, the results demonstrate that our method exhibits superiority in performance and achieves a remarkable balance between visual perception and objective evaluation metrics when compared to the state-of-the-art (SOTA) methods.
•The Multilevel Dual-Branch Attention Network (MDAN) is proposed to extract multi-level features for generating fused images.•A fusion strategy layer based on spatial-channel attention and feature aggregation block is proposed to fused the extracted features.•A loss function based on singular value decomposition is designed to preserve the algebra features of the source images.•A large number of experiments prove the rationality and superiority of the proposed method.</description><subject>Attention mechanism</subject><subject>Deep learning</subject><subject>Fusion strategy</subject><subject>Infrared and visible image fusion</subject><subject>Singular value decomposition</subject><issn>0143-8166</issn><issn>1873-0302</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNqFkMtOwzAQRS0EEqXwDfgHUvxI4sCu4i1R2JS1NXbGxSU4yHaL-HtcFbFlNdJoztWdQ8g5ZzPOeHuxno2feYCEYTUTTNRl27FaHJAJ75SsmGTikEwYr2XV8bY9JicprVkha84nZLm4mT9f0cVmyH7ALQ6038BQmQjBvlHIGUP2Y6AB89cY36kbI_XBRYjYUwg93frkzYDUf8AKqdukcn1KjhwMCc9-55S83t0urx-qp5f7x-v5U2Ulb3JlQTVKoQEovTplBGuUkMKhQQW16qH80Trbqca0xjgrlLyUopMGuoa3DOSUqH2ujWNKEZ3-jKVH_Nac6Z0cvdZ_cvROjt7LKeR8T2Kpt_UYdbIeg8XeR7RZ96P_N-MHXHhysg</recordid><startdate>202405</startdate><enddate>202405</enddate><creator>Wang, Jiawei</creator><creator>Jiang, Min</creator><creator>Kong, Jun</creator><general>Elsevier Ltd</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0003-2551-4748</orcidid><orcidid>https://orcid.org/0000-0003-3826-6405</orcidid></search><sort><creationdate>202405</creationdate><title>MDAN: Multilevel dual-branch attention network for infrared and visible image fusion</title><author>Wang, Jiawei ; Jiang, Min ; Kong, Jun</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c315t-ca7577ebaa16687b2057232febe7a47da8046fc875b6bbfc27393283ba85160a3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Attention mechanism</topic><topic>Deep learning</topic><topic>Fusion strategy</topic><topic>Infrared and visible image fusion</topic><topic>Singular value decomposition</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Wang, Jiawei</creatorcontrib><creatorcontrib>Jiang, Min</creatorcontrib><creatorcontrib>Kong, Jun</creatorcontrib><collection>CrossRef</collection><jtitle>Optics and lasers in engineering</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Wang, Jiawei</au><au>Jiang, Min</au><au>Kong, Jun</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>MDAN: Multilevel dual-branch attention network for infrared and visible image fusion</atitle><jtitle>Optics and lasers in engineering</jtitle><date>2024-05</date><risdate>2024</risdate><volume>176</volume><spage>108042</spage><pages>108042-</pages><artnum>108042</artnum><issn>0143-8166</issn><eissn>1873-0302</eissn><abstract>Infrared and visible image fusion (IVIF) aims to integrate information captured by optical sensors operating in two different modalities, generating a fused image with both salient targets and texture details. Despite significant advancements in IVIF algorithms, the challenge of preserving complete information, especially regarding texture details, still persists. To alleviate this problem, we propose a multilevel dual-branch attention network (MDAN) which comprises an encoder-decoder network and a fusion strategy layer composed of dual-branch fusion block (DBFB). Firstly, the encoder-decoder network is designed to extract multilevel image features and reconstruct the fused images. Secondly, a novel loss function based on singular value decomposition is proposed to constrain the reconstructed images to preserve abundant algebra features which reflect the structure and texture information of the source images. Thirdly, a fusion strategy layer based on spatial-channel attention and feature aggregation block, which consists of DBFB, is proposed to integrate the extracted features. Finally, we evaluate our method through qualitative and quantitative experiments, the results demonstrate that our method exhibits superiority in performance and achieves a remarkable balance between visual perception and objective evaluation metrics when compared to the state-of-the-art (SOTA) methods.
•The Multilevel Dual-Branch Attention Network (MDAN) is proposed to extract multi-level features for generating fused images.•A fusion strategy layer based on spatial-channel attention and feature aggregation block is proposed to fused the extracted features.•A loss function based on singular value decomposition is designed to preserve the algebra features of the source images.•A large number of experiments prove the rationality and superiority of the proposed method.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.optlaseng.2024.108042</doi><orcidid>https://orcid.org/0000-0003-2551-4748</orcidid><orcidid>https://orcid.org/0000-0003-3826-6405</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0143-8166 |
ispartof | Optics and lasers in engineering, 2024-05, Vol.176, p.108042, Article 108042 |
issn | 0143-8166 1873-0302 |
language | eng |
recordid | cdi_crossref_primary_10_1016_j_optlaseng_2024_108042 |
source | Elsevier:Jisc Collections:Elsevier Read and Publish Agreement 2022-2024:Freedom Collection (Reading list) |
subjects | Attention mechanism Deep learning Fusion strategy Infrared and visible image fusion Singular value decomposition |
title | MDAN: Multilevel dual-branch attention network for infrared and visible image fusion |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-23T20%3A00%3A06IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=MDAN:%20Multilevel%20dual-branch%20attention%20network%20for%20infrared%20and%20visible%20image%20fusion&rft.jtitle=Optics%20and%20lasers%20in%20engineering&rft.au=Wang,%20Jiawei&rft.date=2024-05&rft.volume=176&rft.spage=108042&rft.pages=108042-&rft.artnum=108042&rft.issn=0143-8166&rft.eissn=1873-0302&rft_id=info:doi/10.1016/j.optlaseng.2024.108042&rft_dat=%3Celsevier_cross%3ES0143816624000228%3C/elsevier_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c315t-ca7577ebaa16687b2057232febe7a47da8046fc875b6bbfc27393283ba85160a3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |