Loading…
SACNN: Self-Attention Convolutional Neural Network for Low-Dose CT Denoising With Self-Supervised Perceptual Loss Network
Computed tomography (CT) is a widely used screening and diagnostic tool that allows clinicians to obtain a high-resolution, volumetric image of internal structures in a non-invasive manner. Increasingly, efforts have been made to improve the image quality of low-dose CT (LDCT) to reduce the cumulati...
Saved in:
Published in: | IEEE transactions on medical imaging 2020-07, Vol.39 (7), p.2289-2301 |
---|---|
Main Authors: | , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c394t-70b3a6a1b1bdc99e53de6292e0e1f927d92baf7bd69f79949fd3b947f52be33f3 |
---|---|
cites | cdi_FETCH-LOGICAL-c394t-70b3a6a1b1bdc99e53de6292e0e1f927d92baf7bd69f79949fd3b947f52be33f3 |
container_end_page | 2301 |
container_issue | 7 |
container_start_page | 2289 |
container_title | IEEE transactions on medical imaging |
container_volume | 39 |
creator | Li, Meng Hsu, William Xie, Xiaodong Cong, Jason Gao, Wen |
description | Computed tomography (CT) is a widely used screening and diagnostic tool that allows clinicians to obtain a high-resolution, volumetric image of internal structures in a non-invasive manner. Increasingly, efforts have been made to improve the image quality of low-dose CT (LDCT) to reduce the cumulative radiation exposure of patients undergoing routine screening exams. The resurgence of deep learning has yielded a new approach for noise reduction by training a deep multi-layer convolutional neural networks (CNN) to map the low-dose to normal-dose CT images. However, CNN-based methods heavily rely on convolutional kernels, which use fixed-size filters to process one local neighborhood within the receptive field at a time. As a result, they are not efficient at retrieving structural information across large regions. In this paper, we propose a novel 3D self-attention convolutional neural network for the LDCT denoising problem. Our 3D self-attention module leverages the 3D volume of CT images to capture a wide range of spatial information both within CT slices and between CT slices. With the help of the 3D self-attention module, CNNs are able to leverage pixels with stronger relationships regardless of their distance and achieve better denoising results. In addition, we propose a self-supervised learning scheme to train a domain-specific autoencoder as the perceptual loss function. We combine these two methods and demonstrate their effectiveness on both CNN-based neural networks and WGAN-based neural networks with comprehensive experiments. Tested on the AAPM-Mayo Clinic Low Dose CT Grand Challenge data set, our experiments demonstrate that self-attention (SA) module and autoencoder (AE) perceptual loss function can efficiently enhance traditional CNNs and can achieve comparable or better results than the state-of-the-art methods. |
doi_str_mv | 10.1109/TMI.2020.2968472 |
format | article |
fullrecord | <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_proquest_journals_2419494892</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>8964295</ieee_id><sourcerecordid>2419494892</sourcerecordid><originalsourceid>FETCH-LOGICAL-c394t-70b3a6a1b1bdc99e53de6292e0e1f927d92baf7bd69f79949fd3b947f52be33f3</originalsourceid><addsrcrecordid>eNpdkU1PGzEQQC1UVFLaeyUkZIlLL5vaY-_H9BYttEVKU6QE0dtqP8awdLMO9i6If1-HBA49jSW_eZL9GPssxVRKgV9Xvy6nIEBMAZNMp3DAJjKOswhi_ecdmwhIs0iIBI7YB-_vhZA6FvieHSmJWawlTNjzcpYvFt_4kjoTzYaB-qG1Pc9t_2i7cXsuO76g0b2M4cm6v9xYx-f2KTq3nni-4ufU29a3_S2_aYe7nWo5bsg9tp4afkWups0wBsPcev-q-cgOTdl5-rSfx-z6-8Uq_xnNf_-4zGfzqFaohygVlSqTUlayampEilVDCSCQIGkQ0gahKk1aNQmaFFGjaVSFOjUxVKSUUcfsy867cfZhJD8U69bX1HVlT3b0BSidQIaZTAN69h96b0cXfiBQWga3zhACJXZU7cJzHJli49p16Z4LKYptliJkKbZZin2WsHK6F4_Vmpq3hdcOATjZAS0RvV1nmGjAWP0DBa-RWQ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2419494892</pqid></control><display><type>article</type><title>SACNN: Self-Attention Convolutional Neural Network for Low-Dose CT Denoising With Self-Supervised Perceptual Loss Network</title><source>IEEE Electronic Library (IEL) Journals</source><creator>Li, Meng ; Hsu, William ; Xie, Xiaodong ; Cong, Jason ; Gao, Wen</creator><creatorcontrib>Li, Meng ; Hsu, William ; Xie, Xiaodong ; Cong, Jason ; Gao, Wen</creatorcontrib><description>Computed tomography (CT) is a widely used screening and diagnostic tool that allows clinicians to obtain a high-resolution, volumetric image of internal structures in a non-invasive manner. Increasingly, efforts have been made to improve the image quality of low-dose CT (LDCT) to reduce the cumulative radiation exposure of patients undergoing routine screening exams. The resurgence of deep learning has yielded a new approach for noise reduction by training a deep multi-layer convolutional neural networks (CNN) to map the low-dose to normal-dose CT images. However, CNN-based methods heavily rely on convolutional kernels, which use fixed-size filters to process one local neighborhood within the receptive field at a time. As a result, they are not efficient at retrieving structural information across large regions. In this paper, we propose a novel 3D self-attention convolutional neural network for the LDCT denoising problem. Our 3D self-attention module leverages the 3D volume of CT images to capture a wide range of spatial information both within CT slices and between CT slices. With the help of the 3D self-attention module, CNNs are able to leverage pixels with stronger relationships regardless of their distance and achieve better denoising results. In addition, we propose a self-supervised learning scheme to train a domain-specific autoencoder as the perceptual loss function. We combine these two methods and demonstrate their effectiveness on both CNN-based neural networks and WGAN-based neural networks with comprehensive experiments. Tested on the AAPM-Mayo Clinic Low Dose CT Grand Challenge data set, our experiments demonstrate that self-attention (SA) module and autoencoder (AE) perceptual loss function can efficiently enhance traditional CNNs and can achieve comparable or better results than the state-of-the-art methods.</description><identifier>ISSN: 0278-0062</identifier><identifier>EISSN: 1558-254X</identifier><identifier>DOI: 10.1109/TMI.2020.2968472</identifier><identifier>PMID: 31985412</identifier><identifier>CODEN: ITMID4</identifier><language>eng</language><publisher>United States: IEEE</publisher><subject>Artificial neural networks ; autoencoder ; Computed tomography ; Convolutional neural networks ; denoising ; Diagnostic software ; Diagnostic systems ; Feature extraction ; Image quality ; Image reconstruction ; Image resolution ; Information retrieval ; Low-dose CT ; Machine learning ; Medical imaging ; Modules ; Multilayers ; Neural networks ; Noise reduction ; perceptual loss ; Radiation ; Radiation effects ; Receptive field ; self-attention ; Spatial data ; Three-dimensional displays</subject><ispartof>IEEE transactions on medical imaging, 2020-07, Vol.39 (7), p.2289-2301</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2020</rights><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c394t-70b3a6a1b1bdc99e53de6292e0e1f927d92baf7bd69f79949fd3b947f52be33f3</citedby><cites>FETCH-LOGICAL-c394t-70b3a6a1b1bdc99e53de6292e0e1f927d92baf7bd69f79949fd3b947f52be33f3</cites><orcidid>0000-0003-2887-6963 ; 0000-0001-5565-8799 ; 0000-0002-4325-7741 ; 0000-0002-5168-070X</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/8964295$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,776,780,27901,27902,54771</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/31985412$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Li, Meng</creatorcontrib><creatorcontrib>Hsu, William</creatorcontrib><creatorcontrib>Xie, Xiaodong</creatorcontrib><creatorcontrib>Cong, Jason</creatorcontrib><creatorcontrib>Gao, Wen</creatorcontrib><title>SACNN: Self-Attention Convolutional Neural Network for Low-Dose CT Denoising With Self-Supervised Perceptual Loss Network</title><title>IEEE transactions on medical imaging</title><addtitle>TMI</addtitle><addtitle>IEEE Trans Med Imaging</addtitle><description>Computed tomography (CT) is a widely used screening and diagnostic tool that allows clinicians to obtain a high-resolution, volumetric image of internal structures in a non-invasive manner. Increasingly, efforts have been made to improve the image quality of low-dose CT (LDCT) to reduce the cumulative radiation exposure of patients undergoing routine screening exams. The resurgence of deep learning has yielded a new approach for noise reduction by training a deep multi-layer convolutional neural networks (CNN) to map the low-dose to normal-dose CT images. However, CNN-based methods heavily rely on convolutional kernels, which use fixed-size filters to process one local neighborhood within the receptive field at a time. As a result, they are not efficient at retrieving structural information across large regions. In this paper, we propose a novel 3D self-attention convolutional neural network for the LDCT denoising problem. Our 3D self-attention module leverages the 3D volume of CT images to capture a wide range of spatial information both within CT slices and between CT slices. With the help of the 3D self-attention module, CNNs are able to leverage pixels with stronger relationships regardless of their distance and achieve better denoising results. In addition, we propose a self-supervised learning scheme to train a domain-specific autoencoder as the perceptual loss function. We combine these two methods and demonstrate their effectiveness on both CNN-based neural networks and WGAN-based neural networks with comprehensive experiments. Tested on the AAPM-Mayo Clinic Low Dose CT Grand Challenge data set, our experiments demonstrate that self-attention (SA) module and autoencoder (AE) perceptual loss function can efficiently enhance traditional CNNs and can achieve comparable or better results than the state-of-the-art methods.</description><subject>Artificial neural networks</subject><subject>autoencoder</subject><subject>Computed tomography</subject><subject>Convolutional neural networks</subject><subject>denoising</subject><subject>Diagnostic software</subject><subject>Diagnostic systems</subject><subject>Feature extraction</subject><subject>Image quality</subject><subject>Image reconstruction</subject><subject>Image resolution</subject><subject>Information retrieval</subject><subject>Low-dose CT</subject><subject>Machine learning</subject><subject>Medical imaging</subject><subject>Modules</subject><subject>Multilayers</subject><subject>Neural networks</subject><subject>Noise reduction</subject><subject>perceptual loss</subject><subject>Radiation</subject><subject>Radiation effects</subject><subject>Receptive field</subject><subject>self-attention</subject><subject>Spatial data</subject><subject>Three-dimensional displays</subject><issn>0278-0062</issn><issn>1558-254X</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><recordid>eNpdkU1PGzEQQC1UVFLaeyUkZIlLL5vaY-_H9BYttEVKU6QE0dtqP8awdLMO9i6If1-HBA49jSW_eZL9GPssxVRKgV9Xvy6nIEBMAZNMp3DAJjKOswhi_ecdmwhIs0iIBI7YB-_vhZA6FvieHSmJWawlTNjzcpYvFt_4kjoTzYaB-qG1Pc9t_2i7cXsuO76g0b2M4cm6v9xYx-f2KTq3nni-4ufU29a3_S2_aYe7nWo5bsg9tp4afkWups0wBsPcev-q-cgOTdl5-rSfx-z6-8Uq_xnNf_-4zGfzqFaohygVlSqTUlayampEilVDCSCQIGkQ0gahKk1aNQmaFFGjaVSFOjUxVKSUUcfsy867cfZhJD8U69bX1HVlT3b0BSidQIaZTAN69h96b0cXfiBQWga3zhACJXZU7cJzHJli49p16Z4LKYptliJkKbZZin2WsHK6F4_Vmpq3hdcOATjZAS0RvV1nmGjAWP0DBa-RWQ</recordid><startdate>20200701</startdate><enddate>20200701</enddate><creator>Li, Meng</creator><creator>Hsu, William</creator><creator>Xie, Xiaodong</creator><creator>Cong, Jason</creator><creator>Gao, Wen</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QF</scope><scope>7QO</scope><scope>7QQ</scope><scope>7SC</scope><scope>7SE</scope><scope>7SP</scope><scope>7SR</scope><scope>7TA</scope><scope>7TB</scope><scope>7U5</scope><scope>8BQ</scope><scope>8FD</scope><scope>F28</scope><scope>FR3</scope><scope>H8D</scope><scope>JG9</scope><scope>JQ2</scope><scope>KR7</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>NAPCQ</scope><scope>P64</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0003-2887-6963</orcidid><orcidid>https://orcid.org/0000-0001-5565-8799</orcidid><orcidid>https://orcid.org/0000-0002-4325-7741</orcidid><orcidid>https://orcid.org/0000-0002-5168-070X</orcidid></search><sort><creationdate>20200701</creationdate><title>SACNN: Self-Attention Convolutional Neural Network for Low-Dose CT Denoising With Self-Supervised Perceptual Loss Network</title><author>Li, Meng ; Hsu, William ; Xie, Xiaodong ; Cong, Jason ; Gao, Wen</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c394t-70b3a6a1b1bdc99e53de6292e0e1f927d92baf7bd69f79949fd3b947f52be33f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Artificial neural networks</topic><topic>autoencoder</topic><topic>Computed tomography</topic><topic>Convolutional neural networks</topic><topic>denoising</topic><topic>Diagnostic software</topic><topic>Diagnostic systems</topic><topic>Feature extraction</topic><topic>Image quality</topic><topic>Image reconstruction</topic><topic>Image resolution</topic><topic>Information retrieval</topic><topic>Low-dose CT</topic><topic>Machine learning</topic><topic>Medical imaging</topic><topic>Modules</topic><topic>Multilayers</topic><topic>Neural networks</topic><topic>Noise reduction</topic><topic>perceptual loss</topic><topic>Radiation</topic><topic>Radiation effects</topic><topic>Receptive field</topic><topic>self-attention</topic><topic>Spatial data</topic><topic>Three-dimensional displays</topic><toplevel>online_resources</toplevel><creatorcontrib>Li, Meng</creatorcontrib><creatorcontrib>Hsu, William</creatorcontrib><creatorcontrib>Xie, Xiaodong</creatorcontrib><creatorcontrib>Cong, Jason</creatorcontrib><creatorcontrib>Gao, Wen</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Xplore</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Aluminium Industry Abstracts</collection><collection>Biotechnology Research Abstracts</collection><collection>Ceramic Abstracts</collection><collection>Computer and Information Systems Abstracts</collection><collection>Corrosion Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>Materials Business File</collection><collection>Mechanical & Transportation Engineering Abstracts</collection><collection>Solid State and Superconductivity Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>ANTE: Abstracts in New Technology & Engineering</collection><collection>Engineering Research Database</collection><collection>Aerospace Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Civil Engineering Abstracts</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Nursing & Allied Health Premium</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE transactions on medical imaging</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Li, Meng</au><au>Hsu, William</au><au>Xie, Xiaodong</au><au>Cong, Jason</au><au>Gao, Wen</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>SACNN: Self-Attention Convolutional Neural Network for Low-Dose CT Denoising With Self-Supervised Perceptual Loss Network</atitle><jtitle>IEEE transactions on medical imaging</jtitle><stitle>TMI</stitle><addtitle>IEEE Trans Med Imaging</addtitle><date>2020-07-01</date><risdate>2020</risdate><volume>39</volume><issue>7</issue><spage>2289</spage><epage>2301</epage><pages>2289-2301</pages><issn>0278-0062</issn><eissn>1558-254X</eissn><coden>ITMID4</coden><abstract>Computed tomography (CT) is a widely used screening and diagnostic tool that allows clinicians to obtain a high-resolution, volumetric image of internal structures in a non-invasive manner. Increasingly, efforts have been made to improve the image quality of low-dose CT (LDCT) to reduce the cumulative radiation exposure of patients undergoing routine screening exams. The resurgence of deep learning has yielded a new approach for noise reduction by training a deep multi-layer convolutional neural networks (CNN) to map the low-dose to normal-dose CT images. However, CNN-based methods heavily rely on convolutional kernels, which use fixed-size filters to process one local neighborhood within the receptive field at a time. As a result, they are not efficient at retrieving structural information across large regions. In this paper, we propose a novel 3D self-attention convolutional neural network for the LDCT denoising problem. Our 3D self-attention module leverages the 3D volume of CT images to capture a wide range of spatial information both within CT slices and between CT slices. With the help of the 3D self-attention module, CNNs are able to leverage pixels with stronger relationships regardless of their distance and achieve better denoising results. In addition, we propose a self-supervised learning scheme to train a domain-specific autoencoder as the perceptual loss function. We combine these two methods and demonstrate their effectiveness on both CNN-based neural networks and WGAN-based neural networks with comprehensive experiments. Tested on the AAPM-Mayo Clinic Low Dose CT Grand Challenge data set, our experiments demonstrate that self-attention (SA) module and autoencoder (AE) perceptual loss function can efficiently enhance traditional CNNs and can achieve comparable or better results than the state-of-the-art methods.</abstract><cop>United States</cop><pub>IEEE</pub><pmid>31985412</pmid><doi>10.1109/TMI.2020.2968472</doi><tpages>13</tpages><orcidid>https://orcid.org/0000-0003-2887-6963</orcidid><orcidid>https://orcid.org/0000-0001-5565-8799</orcidid><orcidid>https://orcid.org/0000-0002-4325-7741</orcidid><orcidid>https://orcid.org/0000-0002-5168-070X</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0278-0062 |
ispartof | IEEE transactions on medical imaging, 2020-07, Vol.39 (7), p.2289-2301 |
issn | 0278-0062 1558-254X |
language | eng |
recordid | cdi_proquest_journals_2419494892 |
source | IEEE Electronic Library (IEL) Journals |
subjects | Artificial neural networks autoencoder Computed tomography Convolutional neural networks denoising Diagnostic software Diagnostic systems Feature extraction Image quality Image reconstruction Image resolution Information retrieval Low-dose CT Machine learning Medical imaging Modules Multilayers Neural networks Noise reduction perceptual loss Radiation Radiation effects Receptive field self-attention Spatial data Three-dimensional displays |
title | SACNN: Self-Attention Convolutional Neural Network for Low-Dose CT Denoising With Self-Supervised Perceptual Loss Network |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-04T16%3A54%3A22IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=SACNN:%20Self-Attention%20Convolutional%20Neural%20Network%20for%20Low-Dose%20CT%20Denoising%20With%20Self-Supervised%20Perceptual%20Loss%20Network&rft.jtitle=IEEE%20transactions%20on%20medical%20imaging&rft.au=Li,%20Meng&rft.date=2020-07-01&rft.volume=39&rft.issue=7&rft.spage=2289&rft.epage=2301&rft.pages=2289-2301&rft.issn=0278-0062&rft.eissn=1558-254X&rft.coden=ITMID4&rft_id=info:doi/10.1109/TMI.2020.2968472&rft_dat=%3Cproquest_ieee_%3E2419494892%3C/proquest_ieee_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c394t-70b3a6a1b1bdc99e53de6292e0e1f927d92baf7bd69f79949fd3b947f52be33f3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2419494892&rft_id=info:pmid/31985412&rft_ieee_id=8964295&rfr_iscdi=true |