Loading…

Semantic attention guided low-light image enhancement with multi-scale perception

Low-light environments often lead to complex degradation of captured images. However, most deep learning-based image enhancement methods for low-light conditions only learn a single-channel mapping relationship between the input image in low-light conditions and the desired image in normal light wit...

Full description

Saved in:
Bibliographic Details
Published in:Journal of visual communication and image representation 2024-08, Vol.103, p.104242, Article 104242
Main Authors: Hou, Yongqi, Yang, Bo
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites cdi_FETCH-LOGICAL-c183t-4e1f39412020ed2124dc74ff51dedd43c8b3c6eb5a0468c9c032f7be6fcea7833
container_end_page
container_issue
container_start_page 104242
container_title Journal of visual communication and image representation
container_volume 103
creator Hou, Yongqi
Yang, Bo
description Low-light environments often lead to complex degradation of captured images. However, most deep learning-based image enhancement methods for low-light conditions only learn a single-channel mapping relationship between the input image in low-light conditions and the desired image in normal light without considering semantic priors. This may cause the network to deviate from the original color of the region. In addition, deep network architectures are not suitable for low-light image recovery due to low pixel values. To address these issues, we propose a novel network called SAGNet. It consists of two branches:the main branch extracts global enhancement features at the level of the original image, and the other branch introduces semantic information through region-based feature learning and learns local enhancement features for semantic regions with multi-level perception to maintain color consistency. The extracted features are merged with the global enhancement features for semantic consistency and visualization. We also propose an unsupervised loss function to improve the network’s adaptability to general scenes and reduce the effect of sparse datasets. Extensive experiments and ablation studies show that SAGNet maintains color accuracy better in all cases and keeps natural luminance consistency across the semantic range.
doi_str_mv 10.1016/j.jvcir.2024.104242
format article
fullrecord <record><control><sourceid>elsevier_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1016_j_jvcir_2024_104242</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S1047320324001986</els_id><sourcerecordid>S1047320324001986</sourcerecordid><originalsourceid>FETCH-LOGICAL-c183t-4e1f39412020ed2124dc74ff51dedd43c8b3c6eb5a0468c9c032f7be6fcea7833</originalsourceid><addsrcrecordid>eNp9kMtOAzEMRbMAiVL4Ajb5gSl5dWa6YIEqHpUqIQSso4zjtBnNo0rSVvw9KWXNyleWj319CbnjbMYZL-_bWXsAH2aCCZU7SihxQSZZVIUUTF6R6xhbxphcSDUh7x_YmyF5oCYlzGIc6GbvLVrajcei85ttor43G6Q4bM0A2OcpevRpS_t9l3wRwXRIdxgAdyf8hlw600W8_atT8vX89Ll8LdZvL6vl47oAXstUKOROLhTPNhlawYWyUCnn5jzftkpC3UgosZkbpsoaFsCkcFWDpQM0VS3llMjzXghjjAGd3oVsNHxrzvQpCd3q3yT0KQl9TiJTD2cKs7WDx6AjeMxvWR8Qkraj_5f_ASfdbA4</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Semantic attention guided low-light image enhancement with multi-scale perception</title><source>ScienceDirect Freedom Collection</source><creator>Hou, Yongqi ; Yang, Bo</creator><creatorcontrib>Hou, Yongqi ; Yang, Bo</creatorcontrib><description>Low-light environments often lead to complex degradation of captured images. However, most deep learning-based image enhancement methods for low-light conditions only learn a single-channel mapping relationship between the input image in low-light conditions and the desired image in normal light without considering semantic priors. This may cause the network to deviate from the original color of the region. In addition, deep network architectures are not suitable for low-light image recovery due to low pixel values. To address these issues, we propose a novel network called SAGNet. It consists of two branches:the main branch extracts global enhancement features at the level of the original image, and the other branch introduces semantic information through region-based feature learning and learns local enhancement features for semantic regions with multi-level perception to maintain color consistency. The extracted features are merged with the global enhancement features for semantic consistency and visualization. We also propose an unsupervised loss function to improve the network’s adaptability to general scenes and reduce the effect of sparse datasets. Extensive experiments and ablation studies show that SAGNet maintains color accuracy better in all cases and keeps natural luminance consistency across the semantic range.</description><identifier>ISSN: 1047-3203</identifier><identifier>DOI: 10.1016/j.jvcir.2024.104242</identifier><language>eng</language><publisher>Elsevier Inc</publisher><subject>Attention mechanism ; Deep learning ; Low-light image enhancement ; Semantic guidance</subject><ispartof>Journal of visual communication and image representation, 2024-08, Vol.103, p.104242, Article 104242</ispartof><rights>2024 Elsevier Inc.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c183t-4e1f39412020ed2124dc74ff51dedd43c8b3c6eb5a0468c9c032f7be6fcea7833</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids></links><search><creatorcontrib>Hou, Yongqi</creatorcontrib><creatorcontrib>Yang, Bo</creatorcontrib><title>Semantic attention guided low-light image enhancement with multi-scale perception</title><title>Journal of visual communication and image representation</title><description>Low-light environments often lead to complex degradation of captured images. However, most deep learning-based image enhancement methods for low-light conditions only learn a single-channel mapping relationship between the input image in low-light conditions and the desired image in normal light without considering semantic priors. This may cause the network to deviate from the original color of the region. In addition, deep network architectures are not suitable for low-light image recovery due to low pixel values. To address these issues, we propose a novel network called SAGNet. It consists of two branches:the main branch extracts global enhancement features at the level of the original image, and the other branch introduces semantic information through region-based feature learning and learns local enhancement features for semantic regions with multi-level perception to maintain color consistency. The extracted features are merged with the global enhancement features for semantic consistency and visualization. We also propose an unsupervised loss function to improve the network’s adaptability to general scenes and reduce the effect of sparse datasets. Extensive experiments and ablation studies show that SAGNet maintains color accuracy better in all cases and keeps natural luminance consistency across the semantic range.</description><subject>Attention mechanism</subject><subject>Deep learning</subject><subject>Low-light image enhancement</subject><subject>Semantic guidance</subject><issn>1047-3203</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp9kMtOAzEMRbMAiVL4Ajb5gSl5dWa6YIEqHpUqIQSso4zjtBnNo0rSVvw9KWXNyleWj319CbnjbMYZL-_bWXsAH2aCCZU7SihxQSZZVIUUTF6R6xhbxphcSDUh7x_YmyF5oCYlzGIc6GbvLVrajcei85ttor43G6Q4bM0A2OcpevRpS_t9l3wRwXRIdxgAdyf8hlw600W8_atT8vX89Ll8LdZvL6vl47oAXstUKOROLhTPNhlawYWyUCnn5jzftkpC3UgosZkbpsoaFsCkcFWDpQM0VS3llMjzXghjjAGd3oVsNHxrzvQpCd3q3yT0KQl9TiJTD2cKs7WDx6AjeMxvWR8Qkraj_5f_ASfdbA4</recordid><startdate>202408</startdate><enddate>202408</enddate><creator>Hou, Yongqi</creator><creator>Yang, Bo</creator><general>Elsevier Inc</general><scope>AAYXX</scope><scope>CITATION</scope></search><sort><creationdate>202408</creationdate><title>Semantic attention guided low-light image enhancement with multi-scale perception</title><author>Hou, Yongqi ; Yang, Bo</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c183t-4e1f39412020ed2124dc74ff51dedd43c8b3c6eb5a0468c9c032f7be6fcea7833</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Attention mechanism</topic><topic>Deep learning</topic><topic>Low-light image enhancement</topic><topic>Semantic guidance</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Hou, Yongqi</creatorcontrib><creatorcontrib>Yang, Bo</creatorcontrib><collection>CrossRef</collection><jtitle>Journal of visual communication and image representation</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Hou, Yongqi</au><au>Yang, Bo</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Semantic attention guided low-light image enhancement with multi-scale perception</atitle><jtitle>Journal of visual communication and image representation</jtitle><date>2024-08</date><risdate>2024</risdate><volume>103</volume><spage>104242</spage><pages>104242-</pages><artnum>104242</artnum><issn>1047-3203</issn><abstract>Low-light environments often lead to complex degradation of captured images. However, most deep learning-based image enhancement methods for low-light conditions only learn a single-channel mapping relationship between the input image in low-light conditions and the desired image in normal light without considering semantic priors. This may cause the network to deviate from the original color of the region. In addition, deep network architectures are not suitable for low-light image recovery due to low pixel values. To address these issues, we propose a novel network called SAGNet. It consists of two branches:the main branch extracts global enhancement features at the level of the original image, and the other branch introduces semantic information through region-based feature learning and learns local enhancement features for semantic regions with multi-level perception to maintain color consistency. The extracted features are merged with the global enhancement features for semantic consistency and visualization. We also propose an unsupervised loss function to improve the network’s adaptability to general scenes and reduce the effect of sparse datasets. Extensive experiments and ablation studies show that SAGNet maintains color accuracy better in all cases and keeps natural luminance consistency across the semantic range.</abstract><pub>Elsevier Inc</pub><doi>10.1016/j.jvcir.2024.104242</doi></addata></record>
fulltext fulltext
identifier ISSN: 1047-3203
ispartof Journal of visual communication and image representation, 2024-08, Vol.103, p.104242, Article 104242
issn 1047-3203
language eng
recordid cdi_crossref_primary_10_1016_j_jvcir_2024_104242
source ScienceDirect Freedom Collection
subjects Attention mechanism
Deep learning
Low-light image enhancement
Semantic guidance
title Semantic attention guided low-light image enhancement with multi-scale perception
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-08T02%3A25%3A21IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Semantic%20attention%20guided%20low-light%20image%20enhancement%20with%20multi-scale%20perception&rft.jtitle=Journal%20of%20visual%20communication%20and%20image%20representation&rft.au=Hou,%20Yongqi&rft.date=2024-08&rft.volume=103&rft.spage=104242&rft.pages=104242-&rft.artnum=104242&rft.issn=1047-3203&rft_id=info:doi/10.1016/j.jvcir.2024.104242&rft_dat=%3Celsevier_cross%3ES1047320324001986%3C/elsevier_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c183t-4e1f39412020ed2124dc74ff51dedd43c8b3c6eb5a0468c9c032f7be6fcea7833%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true