Loading…

SACNN: Self-Attention Convolutional Neural Network for Low-Dose CT Denoising With Self-Supervised Perceptual Loss Network

Computed tomography (CT) is a widely used screening and diagnostic tool that allows clinicians to obtain a high-resolution, volumetric image of internal structures in a non-invasive manner. Increasingly, efforts have been made to improve the image quality of low-dose CT (LDCT) to reduce the cumulati...

Full description

Saved in:
Bibliographic Details
Published in:IEEE transactions on medical imaging 2020-07, Vol.39 (7), p.2289-2301
Main Authors: Li, Meng, Hsu, William, Xie, Xiaodong, Cong, Jason, Gao, Wen
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c394t-70b3a6a1b1bdc99e53de6292e0e1f927d92baf7bd69f79949fd3b947f52be33f3
cites cdi_FETCH-LOGICAL-c394t-70b3a6a1b1bdc99e53de6292e0e1f927d92baf7bd69f79949fd3b947f52be33f3
container_end_page 2301
container_issue 7
container_start_page 2289
container_title IEEE transactions on medical imaging
container_volume 39
creator Li, Meng
Hsu, William
Xie, Xiaodong
Cong, Jason
Gao, Wen
description Computed tomography (CT) is a widely used screening and diagnostic tool that allows clinicians to obtain a high-resolution, volumetric image of internal structures in a non-invasive manner. Increasingly, efforts have been made to improve the image quality of low-dose CT (LDCT) to reduce the cumulative radiation exposure of patients undergoing routine screening exams. The resurgence of deep learning has yielded a new approach for noise reduction by training a deep multi-layer convolutional neural networks (CNN) to map the low-dose to normal-dose CT images. However, CNN-based methods heavily rely on convolutional kernels, which use fixed-size filters to process one local neighborhood within the receptive field at a time. As a result, they are not efficient at retrieving structural information across large regions. In this paper, we propose a novel 3D self-attention convolutional neural network for the LDCT denoising problem. Our 3D self-attention module leverages the 3D volume of CT images to capture a wide range of spatial information both within CT slices and between CT slices. With the help of the 3D self-attention module, CNNs are able to leverage pixels with stronger relationships regardless of their distance and achieve better denoising results. In addition, we propose a self-supervised learning scheme to train a domain-specific autoencoder as the perceptual loss function. We combine these two methods and demonstrate their effectiveness on both CNN-based neural networks and WGAN-based neural networks with comprehensive experiments. Tested on the AAPM-Mayo Clinic Low Dose CT Grand Challenge data set, our experiments demonstrate that self-attention (SA) module and autoencoder (AE) perceptual loss function can efficiently enhance traditional CNNs and can achieve comparable or better results than the state-of-the-art methods.
doi_str_mv 10.1109/TMI.2020.2968472
format article
fullrecord <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_proquest_journals_2419494892</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>8964295</ieee_id><sourcerecordid>2419494892</sourcerecordid><originalsourceid>FETCH-LOGICAL-c394t-70b3a6a1b1bdc99e53de6292e0e1f927d92baf7bd69f79949fd3b947f52be33f3</originalsourceid><addsrcrecordid>eNpdkU1PGzEQQC1UVFLaeyUkZIlLL5vaY-_H9BYttEVKU6QE0dtqP8awdLMO9i6If1-HBA49jSW_eZL9GPssxVRKgV9Xvy6nIEBMAZNMp3DAJjKOswhi_ecdmwhIs0iIBI7YB-_vhZA6FvieHSmJWawlTNjzcpYvFt_4kjoTzYaB-qG1Pc9t_2i7cXsuO76g0b2M4cm6v9xYx-f2KTq3nni-4ufU29a3_S2_aYe7nWo5bsg9tp4afkWups0wBsPcev-q-cgOTdl5-rSfx-z6-8Uq_xnNf_-4zGfzqFaohygVlSqTUlayampEilVDCSCQIGkQ0gahKk1aNQmaFFGjaVSFOjUxVKSUUcfsy867cfZhJD8U69bX1HVlT3b0BSidQIaZTAN69h96b0cXfiBQWga3zhACJXZU7cJzHJli49p16Z4LKYptliJkKbZZin2WsHK6F4_Vmpq3hdcOATjZAS0RvV1nmGjAWP0DBa-RWQ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2419494892</pqid></control><display><type>article</type><title>SACNN: Self-Attention Convolutional Neural Network for Low-Dose CT Denoising With Self-Supervised Perceptual Loss Network</title><source>IEEE Electronic Library (IEL) Journals</source><creator>Li, Meng ; Hsu, William ; Xie, Xiaodong ; Cong, Jason ; Gao, Wen</creator><creatorcontrib>Li, Meng ; Hsu, William ; Xie, Xiaodong ; Cong, Jason ; Gao, Wen</creatorcontrib><description>Computed tomography (CT) is a widely used screening and diagnostic tool that allows clinicians to obtain a high-resolution, volumetric image of internal structures in a non-invasive manner. Increasingly, efforts have been made to improve the image quality of low-dose CT (LDCT) to reduce the cumulative radiation exposure of patients undergoing routine screening exams. The resurgence of deep learning has yielded a new approach for noise reduction by training a deep multi-layer convolutional neural networks (CNN) to map the low-dose to normal-dose CT images. However, CNN-based methods heavily rely on convolutional kernels, which use fixed-size filters to process one local neighborhood within the receptive field at a time. As a result, they are not efficient at retrieving structural information across large regions. In this paper, we propose a novel 3D self-attention convolutional neural network for the LDCT denoising problem. Our 3D self-attention module leverages the 3D volume of CT images to capture a wide range of spatial information both within CT slices and between CT slices. With the help of the 3D self-attention module, CNNs are able to leverage pixels with stronger relationships regardless of their distance and achieve better denoising results. In addition, we propose a self-supervised learning scheme to train a domain-specific autoencoder as the perceptual loss function. We combine these two methods and demonstrate their effectiveness on both CNN-based neural networks and WGAN-based neural networks with comprehensive experiments. Tested on the AAPM-Mayo Clinic Low Dose CT Grand Challenge data set, our experiments demonstrate that self-attention (SA) module and autoencoder (AE) perceptual loss function can efficiently enhance traditional CNNs and can achieve comparable or better results than the state-of-the-art methods.</description><identifier>ISSN: 0278-0062</identifier><identifier>EISSN: 1558-254X</identifier><identifier>DOI: 10.1109/TMI.2020.2968472</identifier><identifier>PMID: 31985412</identifier><identifier>CODEN: ITMID4</identifier><language>eng</language><publisher>United States: IEEE</publisher><subject>Artificial neural networks ; autoencoder ; Computed tomography ; Convolutional neural networks ; denoising ; Diagnostic software ; Diagnostic systems ; Feature extraction ; Image quality ; Image reconstruction ; Image resolution ; Information retrieval ; Low-dose CT ; Machine learning ; Medical imaging ; Modules ; Multilayers ; Neural networks ; Noise reduction ; perceptual loss ; Radiation ; Radiation effects ; Receptive field ; self-attention ; Spatial data ; Three-dimensional displays</subject><ispartof>IEEE transactions on medical imaging, 2020-07, Vol.39 (7), p.2289-2301</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2020</rights><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c394t-70b3a6a1b1bdc99e53de6292e0e1f927d92baf7bd69f79949fd3b947f52be33f3</citedby><cites>FETCH-LOGICAL-c394t-70b3a6a1b1bdc99e53de6292e0e1f927d92baf7bd69f79949fd3b947f52be33f3</cites><orcidid>0000-0003-2887-6963 ; 0000-0001-5565-8799 ; 0000-0002-4325-7741 ; 0000-0002-5168-070X</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/8964295$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,776,780,27901,27902,54771</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/31985412$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Li, Meng</creatorcontrib><creatorcontrib>Hsu, William</creatorcontrib><creatorcontrib>Xie, Xiaodong</creatorcontrib><creatorcontrib>Cong, Jason</creatorcontrib><creatorcontrib>Gao, Wen</creatorcontrib><title>SACNN: Self-Attention Convolutional Neural Network for Low-Dose CT Denoising With Self-Supervised Perceptual Loss Network</title><title>IEEE transactions on medical imaging</title><addtitle>TMI</addtitle><addtitle>IEEE Trans Med Imaging</addtitle><description>Computed tomography (CT) is a widely used screening and diagnostic tool that allows clinicians to obtain a high-resolution, volumetric image of internal structures in a non-invasive manner. Increasingly, efforts have been made to improve the image quality of low-dose CT (LDCT) to reduce the cumulative radiation exposure of patients undergoing routine screening exams. The resurgence of deep learning has yielded a new approach for noise reduction by training a deep multi-layer convolutional neural networks (CNN) to map the low-dose to normal-dose CT images. However, CNN-based methods heavily rely on convolutional kernels, which use fixed-size filters to process one local neighborhood within the receptive field at a time. As a result, they are not efficient at retrieving structural information across large regions. In this paper, we propose a novel 3D self-attention convolutional neural network for the LDCT denoising problem. Our 3D self-attention module leverages the 3D volume of CT images to capture a wide range of spatial information both within CT slices and between CT slices. With the help of the 3D self-attention module, CNNs are able to leverage pixels with stronger relationships regardless of their distance and achieve better denoising results. In addition, we propose a self-supervised learning scheme to train a domain-specific autoencoder as the perceptual loss function. We combine these two methods and demonstrate their effectiveness on both CNN-based neural networks and WGAN-based neural networks with comprehensive experiments. Tested on the AAPM-Mayo Clinic Low Dose CT Grand Challenge data set, our experiments demonstrate that self-attention (SA) module and autoencoder (AE) perceptual loss function can efficiently enhance traditional CNNs and can achieve comparable or better results than the state-of-the-art methods.</description><subject>Artificial neural networks</subject><subject>autoencoder</subject><subject>Computed tomography</subject><subject>Convolutional neural networks</subject><subject>denoising</subject><subject>Diagnostic software</subject><subject>Diagnostic systems</subject><subject>Feature extraction</subject><subject>Image quality</subject><subject>Image reconstruction</subject><subject>Image resolution</subject><subject>Information retrieval</subject><subject>Low-dose CT</subject><subject>Machine learning</subject><subject>Medical imaging</subject><subject>Modules</subject><subject>Multilayers</subject><subject>Neural networks</subject><subject>Noise reduction</subject><subject>perceptual loss</subject><subject>Radiation</subject><subject>Radiation effects</subject><subject>Receptive field</subject><subject>self-attention</subject><subject>Spatial data</subject><subject>Three-dimensional displays</subject><issn>0278-0062</issn><issn>1558-254X</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><recordid>eNpdkU1PGzEQQC1UVFLaeyUkZIlLL5vaY-_H9BYttEVKU6QE0dtqP8awdLMO9i6If1-HBA49jSW_eZL9GPssxVRKgV9Xvy6nIEBMAZNMp3DAJjKOswhi_ecdmwhIs0iIBI7YB-_vhZA6FvieHSmJWawlTNjzcpYvFt_4kjoTzYaB-qG1Pc9t_2i7cXsuO76g0b2M4cm6v9xYx-f2KTq3nni-4ufU29a3_S2_aYe7nWo5bsg9tp4afkWups0wBsPcev-q-cgOTdl5-rSfx-z6-8Uq_xnNf_-4zGfzqFaohygVlSqTUlayampEilVDCSCQIGkQ0gahKk1aNQmaFFGjaVSFOjUxVKSUUcfsy867cfZhJD8U69bX1HVlT3b0BSidQIaZTAN69h96b0cXfiBQWga3zhACJXZU7cJzHJli49p16Z4LKYptliJkKbZZin2WsHK6F4_Vmpq3hdcOATjZAS0RvV1nmGjAWP0DBa-RWQ</recordid><startdate>20200701</startdate><enddate>20200701</enddate><creator>Li, Meng</creator><creator>Hsu, William</creator><creator>Xie, Xiaodong</creator><creator>Cong, Jason</creator><creator>Gao, Wen</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QF</scope><scope>7QO</scope><scope>7QQ</scope><scope>7SC</scope><scope>7SE</scope><scope>7SP</scope><scope>7SR</scope><scope>7TA</scope><scope>7TB</scope><scope>7U5</scope><scope>8BQ</scope><scope>8FD</scope><scope>F28</scope><scope>FR3</scope><scope>H8D</scope><scope>JG9</scope><scope>JQ2</scope><scope>KR7</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>NAPCQ</scope><scope>P64</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0003-2887-6963</orcidid><orcidid>https://orcid.org/0000-0001-5565-8799</orcidid><orcidid>https://orcid.org/0000-0002-4325-7741</orcidid><orcidid>https://orcid.org/0000-0002-5168-070X</orcidid></search><sort><creationdate>20200701</creationdate><title>SACNN: Self-Attention Convolutional Neural Network for Low-Dose CT Denoising With Self-Supervised Perceptual Loss Network</title><author>Li, Meng ; Hsu, William ; Xie, Xiaodong ; Cong, Jason ; Gao, Wen</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c394t-70b3a6a1b1bdc99e53de6292e0e1f927d92baf7bd69f79949fd3b947f52be33f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Artificial neural networks</topic><topic>autoencoder</topic><topic>Computed tomography</topic><topic>Convolutional neural networks</topic><topic>denoising</topic><topic>Diagnostic software</topic><topic>Diagnostic systems</topic><topic>Feature extraction</topic><topic>Image quality</topic><topic>Image reconstruction</topic><topic>Image resolution</topic><topic>Information retrieval</topic><topic>Low-dose CT</topic><topic>Machine learning</topic><topic>Medical imaging</topic><topic>Modules</topic><topic>Multilayers</topic><topic>Neural networks</topic><topic>Noise reduction</topic><topic>perceptual loss</topic><topic>Radiation</topic><topic>Radiation effects</topic><topic>Receptive field</topic><topic>self-attention</topic><topic>Spatial data</topic><topic>Three-dimensional displays</topic><toplevel>online_resources</toplevel><creatorcontrib>Li, Meng</creatorcontrib><creatorcontrib>Hsu, William</creatorcontrib><creatorcontrib>Xie, Xiaodong</creatorcontrib><creatorcontrib>Cong, Jason</creatorcontrib><creatorcontrib>Gao, Wen</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Xplore</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Aluminium Industry Abstracts</collection><collection>Biotechnology Research Abstracts</collection><collection>Ceramic Abstracts</collection><collection>Computer and Information Systems Abstracts</collection><collection>Corrosion Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>Materials Business File</collection><collection>Mechanical &amp; Transportation Engineering Abstracts</collection><collection>Solid State and Superconductivity Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>ANTE: Abstracts in New Technology &amp; Engineering</collection><collection>Engineering Research Database</collection><collection>Aerospace Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Civil Engineering Abstracts</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Nursing &amp; Allied Health Premium</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE transactions on medical imaging</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Li, Meng</au><au>Hsu, William</au><au>Xie, Xiaodong</au><au>Cong, Jason</au><au>Gao, Wen</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>SACNN: Self-Attention Convolutional Neural Network for Low-Dose CT Denoising With Self-Supervised Perceptual Loss Network</atitle><jtitle>IEEE transactions on medical imaging</jtitle><stitle>TMI</stitle><addtitle>IEEE Trans Med Imaging</addtitle><date>2020-07-01</date><risdate>2020</risdate><volume>39</volume><issue>7</issue><spage>2289</spage><epage>2301</epage><pages>2289-2301</pages><issn>0278-0062</issn><eissn>1558-254X</eissn><coden>ITMID4</coden><abstract>Computed tomography (CT) is a widely used screening and diagnostic tool that allows clinicians to obtain a high-resolution, volumetric image of internal structures in a non-invasive manner. Increasingly, efforts have been made to improve the image quality of low-dose CT (LDCT) to reduce the cumulative radiation exposure of patients undergoing routine screening exams. The resurgence of deep learning has yielded a new approach for noise reduction by training a deep multi-layer convolutional neural networks (CNN) to map the low-dose to normal-dose CT images. However, CNN-based methods heavily rely on convolutional kernels, which use fixed-size filters to process one local neighborhood within the receptive field at a time. As a result, they are not efficient at retrieving structural information across large regions. In this paper, we propose a novel 3D self-attention convolutional neural network for the LDCT denoising problem. Our 3D self-attention module leverages the 3D volume of CT images to capture a wide range of spatial information both within CT slices and between CT slices. With the help of the 3D self-attention module, CNNs are able to leverage pixels with stronger relationships regardless of their distance and achieve better denoising results. In addition, we propose a self-supervised learning scheme to train a domain-specific autoencoder as the perceptual loss function. We combine these two methods and demonstrate their effectiveness on both CNN-based neural networks and WGAN-based neural networks with comprehensive experiments. Tested on the AAPM-Mayo Clinic Low Dose CT Grand Challenge data set, our experiments demonstrate that self-attention (SA) module and autoencoder (AE) perceptual loss function can efficiently enhance traditional CNNs and can achieve comparable or better results than the state-of-the-art methods.</abstract><cop>United States</cop><pub>IEEE</pub><pmid>31985412</pmid><doi>10.1109/TMI.2020.2968472</doi><tpages>13</tpages><orcidid>https://orcid.org/0000-0003-2887-6963</orcidid><orcidid>https://orcid.org/0000-0001-5565-8799</orcidid><orcidid>https://orcid.org/0000-0002-4325-7741</orcidid><orcidid>https://orcid.org/0000-0002-5168-070X</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 0278-0062
ispartof IEEE transactions on medical imaging, 2020-07, Vol.39 (7), p.2289-2301
issn 0278-0062
1558-254X
language eng
recordid cdi_proquest_journals_2419494892
source IEEE Electronic Library (IEL) Journals
subjects Artificial neural networks
autoencoder
Computed tomography
Convolutional neural networks
denoising
Diagnostic software
Diagnostic systems
Feature extraction
Image quality
Image reconstruction
Image resolution
Information retrieval
Low-dose CT
Machine learning
Medical imaging
Modules
Multilayers
Neural networks
Noise reduction
perceptual loss
Radiation
Radiation effects
Receptive field
self-attention
Spatial data
Three-dimensional displays
title SACNN: Self-Attention Convolutional Neural Network for Low-Dose CT Denoising With Self-Supervised Perceptual Loss Network
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-04T16%3A54%3A22IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=SACNN:%20Self-Attention%20Convolutional%20Neural%20Network%20for%20Low-Dose%20CT%20Denoising%20With%20Self-Supervised%20Perceptual%20Loss%20Network&rft.jtitle=IEEE%20transactions%20on%20medical%20imaging&rft.au=Li,%20Meng&rft.date=2020-07-01&rft.volume=39&rft.issue=7&rft.spage=2289&rft.epage=2301&rft.pages=2289-2301&rft.issn=0278-0062&rft.eissn=1558-254X&rft.coden=ITMID4&rft_id=info:doi/10.1109/TMI.2020.2968472&rft_dat=%3Cproquest_ieee_%3E2419494892%3C/proquest_ieee_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c394t-70b3a6a1b1bdc99e53de6292e0e1f927d92baf7bd69f79949fd3b947f52be33f3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2419494892&rft_id=info:pmid/31985412&rft_ieee_id=8964295&rfr_iscdi=true