Loading…
Joint Localization and Classification of Breast Cancer in B-Mode Ultrasound Imaging via Collaborative Learning With Elastography
Convolutional neural networks (CNNs) have been successfully applied in the computer-aided ultrasound diagnosis for breast cancer. Up to now, several CNN-based methods have been proposed. However, most of them consider tumor localization and classification as two separate steps, rather than performin...
Saved in:
Published in: | IEEE journal of biomedical and health informatics 2022-09, Vol.26 (9), p.4474-4485 |
---|---|
Main Authors: | , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c326t-d57a19a86958a530f3a3b1de8b66abe6b4d18529d8527887e623872f6d6c1d8b3 |
---|---|
cites | cdi_FETCH-LOGICAL-c326t-d57a19a86958a530f3a3b1de8b66abe6b4d18529d8527887e623872f6d6c1d8b3 |
container_end_page | 4485 |
container_issue | 9 |
container_start_page | 4474 |
container_title | IEEE journal of biomedical and health informatics |
container_volume | 26 |
creator | Ding, Weichang Wang, Jun Zhou, Weijun Zhou, Shichong Chang, Cai Shi, Jun |
description | Convolutional neural networks (CNNs) have been successfully applied in the computer-aided ultrasound diagnosis for breast cancer. Up to now, several CNN-based methods have been proposed. However, most of them consider tumor localization and classification as two separate steps, rather than performing them simultaneously. Besides, they suffer from the limited diagnosis information in the B-mode ultrasound (BUS) images. In this study, we develop a novel network ResNet-GAP that incorporates both localization and classification into a unified procedure. To enhance the performance of ResNet-GAP, we leverage stiffness information in the elastography ultrasound (EUS) modality by collaborative learning in the training stage. Specifically, a dual-channel ResNet-GAP network is developed, one channel for BUS and the other for EUS. In each channel, multiple class activity maps (CAMs) are generated using a series of convolutional kernels of different sizes. The multi-scale consistency of the CAMs in both channels are further considered in network optimization. Experiments on 264 patients in this study show that the newly developed ResNet-GAP achieves an accuracy of 88.6%, a sensitivity of 95.3%, a specificity of 84.6%, and an AUC of 93.6% on the classification task, and a 1.0NLF of 87.9% on the localization task, which is better than some state-of-the-art approaches. |
doi_str_mv | 10.1109/JBHI.2022.3186933 |
format | article |
fullrecord | <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_proquest_miscellaneous_2681811845</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9809821</ieee_id><sourcerecordid>2712053097</sourcerecordid><originalsourceid>FETCH-LOGICAL-c326t-d57a19a86958a530f3a3b1de8b66abe6b4d18529d8527887e623872f6d6c1d8b3</originalsourceid><addsrcrecordid>eNpdkcFuGyEQhlHVKo6cPEDUC1IuvazLgJeFY71KE0eOekmUI5rdZR2iNTiwtuSe-ujFcppDOcBo5vuHGf2EXAGbATD9_X5xt5xxxvlMgJJaiE_knINUBedMff4Xg55PyGVKrywflVNanpGJKCsp5rI6J3_ug_MjXYUWB_cbRxc8Rd_ResCUXO_aUyr0dBEtppHW6FsbqfN0UTyEztKnYYyYwi6LlhtcO7-me4e0DsOATYhZv7d0ZTH6Y-nZjS_0Jjcfwzri9uVwQb70OCR7-f5OydPPm8f6rlj9ul3WP1ZFK7gci66sEDTmRUuFpWC9QNFAZ1UjJTZWNvMOVMl1l69KqcpKLlTFe9nJFjrViCn5duq7jeFtZ9NoNi61Ng_pbdglw6UCBaDmZUav_0Nfwy76PJ3hFXCWv9dVpuBEtTGkFG1vttFtMB4MMHN0yBwdMkeHzLtDWfP1pHHW2g9eK6YVB_EXPXqKmw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2712053097</pqid></control><display><type>article</type><title>Joint Localization and Classification of Breast Cancer in B-Mode Ultrasound Imaging via Collaborative Learning With Elastography</title><source>IEEE Electronic Library (IEL) Journals</source><creator>Ding, Weichang ; Wang, Jun ; Zhou, Weijun ; Zhou, Shichong ; Chang, Cai ; Shi, Jun</creator><creatorcontrib>Ding, Weichang ; Wang, Jun ; Zhou, Weijun ; Zhou, Shichong ; Chang, Cai ; Shi, Jun</creatorcontrib><description>Convolutional neural networks (CNNs) have been successfully applied in the computer-aided ultrasound diagnosis for breast cancer. Up to now, several CNN-based methods have been proposed. However, most of them consider tumor localization and classification as two separate steps, rather than performing them simultaneously. Besides, they suffer from the limited diagnosis information in the B-mode ultrasound (BUS) images. In this study, we develop a novel network ResNet-GAP that incorporates both localization and classification into a unified procedure. To enhance the performance of ResNet-GAP, we leverage stiffness information in the elastography ultrasound (EUS) modality by collaborative learning in the training stage. Specifically, a dual-channel ResNet-GAP network is developed, one channel for BUS and the other for EUS. In each channel, multiple class activity maps (CAMs) are generated using a series of convolutional kernels of different sizes. The multi-scale consistency of the CAMs in both channels are further considered in network optimization. Experiments on 264 patients in this study show that the newly developed ResNet-GAP achieves an accuracy of 88.6%, a sensitivity of 95.3%, a specificity of 84.6%, and an AUC of 93.6% on the classification task, and a 1.0NLF of 87.9% on the localization task, which is better than some state-of-the-art approaches.</description><identifier>ISSN: 2168-2194</identifier><identifier>EISSN: 2168-2208</identifier><identifier>DOI: 10.1109/JBHI.2022.3186933</identifier><identifier>PMID: 35763467</identifier><identifier>CODEN: IJBHA9</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Artificial neural networks ; Breast cancer ; Cams ; Classification ; Collaborative learning ; collaborative learning with multi-scale consistency ; Convolutional neural networks ; Deep learning ; Diagnosis ; Learning ; Localization ; Location awareness ; Network management systems ; Neural networks ; Optimization ; Stiffness ; tumor classification ; tumor localization ; Tumors ; Ultrasonic imaging ; Ultrasound</subject><ispartof>IEEE journal of biomedical and health informatics, 2022-09, Vol.26 (9), p.4474-4485</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c326t-d57a19a86958a530f3a3b1de8b66abe6b4d18529d8527887e623872f6d6c1d8b3</citedby><cites>FETCH-LOGICAL-c326t-d57a19a86958a530f3a3b1de8b66abe6b4d18529d8527887e623872f6d6c1d8b3</cites><orcidid>0000-0002-3226-3978 ; 0000-0001-9548-0411</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9809821$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,54796</link.rule.ids></links><search><creatorcontrib>Ding, Weichang</creatorcontrib><creatorcontrib>Wang, Jun</creatorcontrib><creatorcontrib>Zhou, Weijun</creatorcontrib><creatorcontrib>Zhou, Shichong</creatorcontrib><creatorcontrib>Chang, Cai</creatorcontrib><creatorcontrib>Shi, Jun</creatorcontrib><title>Joint Localization and Classification of Breast Cancer in B-Mode Ultrasound Imaging via Collaborative Learning With Elastography</title><title>IEEE journal of biomedical and health informatics</title><addtitle>JBHI</addtitle><description>Convolutional neural networks (CNNs) have been successfully applied in the computer-aided ultrasound diagnosis for breast cancer. Up to now, several CNN-based methods have been proposed. However, most of them consider tumor localization and classification as two separate steps, rather than performing them simultaneously. Besides, they suffer from the limited diagnosis information in the B-mode ultrasound (BUS) images. In this study, we develop a novel network ResNet-GAP that incorporates both localization and classification into a unified procedure. To enhance the performance of ResNet-GAP, we leverage stiffness information in the elastography ultrasound (EUS) modality by collaborative learning in the training stage. Specifically, a dual-channel ResNet-GAP network is developed, one channel for BUS and the other for EUS. In each channel, multiple class activity maps (CAMs) are generated using a series of convolutional kernels of different sizes. The multi-scale consistency of the CAMs in both channels are further considered in network optimization. Experiments on 264 patients in this study show that the newly developed ResNet-GAP achieves an accuracy of 88.6%, a sensitivity of 95.3%, a specificity of 84.6%, and an AUC of 93.6% on the classification task, and a 1.0NLF of 87.9% on the localization task, which is better than some state-of-the-art approaches.</description><subject>Artificial neural networks</subject><subject>Breast cancer</subject><subject>Cams</subject><subject>Classification</subject><subject>Collaborative learning</subject><subject>collaborative learning with multi-scale consistency</subject><subject>Convolutional neural networks</subject><subject>Deep learning</subject><subject>Diagnosis</subject><subject>Learning</subject><subject>Localization</subject><subject>Location awareness</subject><subject>Network management systems</subject><subject>Neural networks</subject><subject>Optimization</subject><subject>Stiffness</subject><subject>tumor classification</subject><subject>tumor localization</subject><subject>Tumors</subject><subject>Ultrasonic imaging</subject><subject>Ultrasound</subject><issn>2168-2194</issn><issn>2168-2208</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNpdkcFuGyEQhlHVKo6cPEDUC1IuvazLgJeFY71KE0eOekmUI5rdZR2iNTiwtuSe-ujFcppDOcBo5vuHGf2EXAGbATD9_X5xt5xxxvlMgJJaiE_knINUBedMff4Xg55PyGVKrywflVNanpGJKCsp5rI6J3_ug_MjXYUWB_cbRxc8Rd_ResCUXO_aUyr0dBEtppHW6FsbqfN0UTyEztKnYYyYwi6LlhtcO7-me4e0DsOATYhZv7d0ZTH6Y-nZjS_0Jjcfwzri9uVwQb70OCR7-f5OydPPm8f6rlj9ul3WP1ZFK7gci66sEDTmRUuFpWC9QNFAZ1UjJTZWNvMOVMl1l69KqcpKLlTFe9nJFjrViCn5duq7jeFtZ9NoNi61Ng_pbdglw6UCBaDmZUav_0Nfwy76PJ3hFXCWv9dVpuBEtTGkFG1vttFtMB4MMHN0yBwdMkeHzLtDWfP1pHHW2g9eK6YVB_EXPXqKmw</recordid><startdate>20220901</startdate><enddate>20220901</enddate><creator>Ding, Weichang</creator><creator>Wang, Jun</creator><creator>Zhou, Weijun</creator><creator>Zhou, Shichong</creator><creator>Chang, Cai</creator><creator>Shi, Jun</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QF</scope><scope>7QO</scope><scope>7QQ</scope><scope>7SC</scope><scope>7SE</scope><scope>7SP</scope><scope>7SR</scope><scope>7TA</scope><scope>7TB</scope><scope>7U5</scope><scope>8BQ</scope><scope>8FD</scope><scope>F28</scope><scope>FR3</scope><scope>H8D</scope><scope>JG9</scope><scope>JQ2</scope><scope>K9.</scope><scope>KR7</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>NAPCQ</scope><scope>P64</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0002-3226-3978</orcidid><orcidid>https://orcid.org/0000-0001-9548-0411</orcidid></search><sort><creationdate>20220901</creationdate><title>Joint Localization and Classification of Breast Cancer in B-Mode Ultrasound Imaging via Collaborative Learning With Elastography</title><author>Ding, Weichang ; Wang, Jun ; Zhou, Weijun ; Zhou, Shichong ; Chang, Cai ; Shi, Jun</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c326t-d57a19a86958a530f3a3b1de8b66abe6b4d18529d8527887e623872f6d6c1d8b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Artificial neural networks</topic><topic>Breast cancer</topic><topic>Cams</topic><topic>Classification</topic><topic>Collaborative learning</topic><topic>collaborative learning with multi-scale consistency</topic><topic>Convolutional neural networks</topic><topic>Deep learning</topic><topic>Diagnosis</topic><topic>Learning</topic><topic>Localization</topic><topic>Location awareness</topic><topic>Network management systems</topic><topic>Neural networks</topic><topic>Optimization</topic><topic>Stiffness</topic><topic>tumor classification</topic><topic>tumor localization</topic><topic>Tumors</topic><topic>Ultrasonic imaging</topic><topic>Ultrasound</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Ding, Weichang</creatorcontrib><creatorcontrib>Wang, Jun</creatorcontrib><creatorcontrib>Zhou, Weijun</creatorcontrib><creatorcontrib>Zhou, Shichong</creatorcontrib><creatorcontrib>Chang, Cai</creatorcontrib><creatorcontrib>Shi, Jun</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE/IET Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Aluminium Industry Abstracts</collection><collection>Biotechnology Research Abstracts</collection><collection>Ceramic Abstracts</collection><collection>Computer and Information Systems Abstracts</collection><collection>Corrosion Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>Materials Business File</collection><collection>Mechanical & Transportation Engineering Abstracts</collection><collection>Solid State and Superconductivity Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>ANTE: Abstracts in New Technology & Engineering</collection><collection>Engineering Research Database</collection><collection>Aerospace Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Health & Medical Complete (Alumni)</collection><collection>Civil Engineering Abstracts</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Nursing & Allied Health Premium</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE journal of biomedical and health informatics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Ding, Weichang</au><au>Wang, Jun</au><au>Zhou, Weijun</au><au>Zhou, Shichong</au><au>Chang, Cai</au><au>Shi, Jun</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Joint Localization and Classification of Breast Cancer in B-Mode Ultrasound Imaging via Collaborative Learning With Elastography</atitle><jtitle>IEEE journal of biomedical and health informatics</jtitle><stitle>JBHI</stitle><date>2022-09-01</date><risdate>2022</risdate><volume>26</volume><issue>9</issue><spage>4474</spage><epage>4485</epage><pages>4474-4485</pages><issn>2168-2194</issn><eissn>2168-2208</eissn><coden>IJBHA9</coden><abstract>Convolutional neural networks (CNNs) have been successfully applied in the computer-aided ultrasound diagnosis for breast cancer. Up to now, several CNN-based methods have been proposed. However, most of them consider tumor localization and classification as two separate steps, rather than performing them simultaneously. Besides, they suffer from the limited diagnosis information in the B-mode ultrasound (BUS) images. In this study, we develop a novel network ResNet-GAP that incorporates both localization and classification into a unified procedure. To enhance the performance of ResNet-GAP, we leverage stiffness information in the elastography ultrasound (EUS) modality by collaborative learning in the training stage. Specifically, a dual-channel ResNet-GAP network is developed, one channel for BUS and the other for EUS. In each channel, multiple class activity maps (CAMs) are generated using a series of convolutional kernels of different sizes. The multi-scale consistency of the CAMs in both channels are further considered in network optimization. Experiments on 264 patients in this study show that the newly developed ResNet-GAP achieves an accuracy of 88.6%, a sensitivity of 95.3%, a specificity of 84.6%, and an AUC of 93.6% on the classification task, and a 1.0NLF of 87.9% on the localization task, which is better than some state-of-the-art approaches.</abstract><cop>Piscataway</cop><pub>IEEE</pub><pmid>35763467</pmid><doi>10.1109/JBHI.2022.3186933</doi><tpages>12</tpages><orcidid>https://orcid.org/0000-0002-3226-3978</orcidid><orcidid>https://orcid.org/0000-0001-9548-0411</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2168-2194 |
ispartof | IEEE journal of biomedical and health informatics, 2022-09, Vol.26 (9), p.4474-4485 |
issn | 2168-2194 2168-2208 |
language | eng |
recordid | cdi_proquest_miscellaneous_2681811845 |
source | IEEE Electronic Library (IEL) Journals |
subjects | Artificial neural networks Breast cancer Cams Classification Collaborative learning collaborative learning with multi-scale consistency Convolutional neural networks Deep learning Diagnosis Learning Localization Location awareness Network management systems Neural networks Optimization Stiffness tumor classification tumor localization Tumors Ultrasonic imaging Ultrasound |
title | Joint Localization and Classification of Breast Cancer in B-Mode Ultrasound Imaging via Collaborative Learning With Elastography |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T17%3A43%3A28IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Joint%20Localization%20and%20Classification%20of%20Breast%20Cancer%20in%20B-Mode%20Ultrasound%20Imaging%20via%20Collaborative%20Learning%20With%20Elastography&rft.jtitle=IEEE%20journal%20of%20biomedical%20and%20health%20informatics&rft.au=Ding,%20Weichang&rft.date=2022-09-01&rft.volume=26&rft.issue=9&rft.spage=4474&rft.epage=4485&rft.pages=4474-4485&rft.issn=2168-2194&rft.eissn=2168-2208&rft.coden=IJBHA9&rft_id=info:doi/10.1109/JBHI.2022.3186933&rft_dat=%3Cproquest_ieee_%3E2712053097%3C/proquest_ieee_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c326t-d57a19a86958a530f3a3b1de8b66abe6b4d18529d8527887e623872f6d6c1d8b3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2712053097&rft_id=info:pmid/35763467&rft_ieee_id=9809821&rfr_iscdi=true |