Loading…

Joint Localization and Classification of Breast Cancer in B-Mode Ultrasound Imaging via Collaborative Learning With Elastography

Convolutional neural networks (CNNs) have been successfully applied in the computer-aided ultrasound diagnosis for breast cancer. Up to now, several CNN-based methods have been proposed. However, most of them consider tumor localization and classification as two separate steps, rather than performin...

Full description

Saved in:
Bibliographic Details
Published in:IEEE journal of biomedical and health informatics 2022-09, Vol.26 (9), p.4474-4485
Main Authors: Ding, Weichang, Wang, Jun, Zhou, Weijun, Zhou, Shichong, Chang, Cai, Shi, Jun
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c326t-d57a19a86958a530f3a3b1de8b66abe6b4d18529d8527887e623872f6d6c1d8b3
cites cdi_FETCH-LOGICAL-c326t-d57a19a86958a530f3a3b1de8b66abe6b4d18529d8527887e623872f6d6c1d8b3
container_end_page 4485
container_issue 9
container_start_page 4474
container_title IEEE journal of biomedical and health informatics
container_volume 26
creator Ding, Weichang
Wang, Jun
Zhou, Weijun
Zhou, Shichong
Chang, Cai
Shi, Jun
description Convolutional neural networks (CNNs) have been successfully applied in the computer-aided ultrasound diagnosis for breast cancer. Up to now, several CNN-based methods have been proposed. However, most of them consider tumor localization and classification as two separate steps, rather than performing them simultaneously. Besides, they suffer from the limited diagnosis information in the B-mode ultrasound (BUS) images. In this study, we develop a novel network ResNet-GAP that incorporates both localization and classification into a unified procedure. To enhance the performance of ResNet-GAP, we leverage stiffness information in the elastography ultrasound (EUS) modality by collaborative learning in the training stage. Specifically, a dual-channel ResNet-GAP network is developed, one channel for BUS and the other for EUS. In each channel, multiple class activity maps (CAMs) are generated using a series of convolutional kernels of different sizes. The multi-scale consistency of the CAMs in both channels are further considered in network optimization. Experiments on 264 patients in this study show that the newly developed ResNet-GAP achieves an accuracy of 88.6%, a sensitivity of 95.3%, a specificity of 84.6%, and an AUC of 93.6% on the classification task, and a 1.0NLF of 87.9% on the localization task, which is better than some state-of-the-art approaches.
doi_str_mv 10.1109/JBHI.2022.3186933
format article
fullrecord <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_proquest_miscellaneous_2681811845</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9809821</ieee_id><sourcerecordid>2712053097</sourcerecordid><originalsourceid>FETCH-LOGICAL-c326t-d57a19a86958a530f3a3b1de8b66abe6b4d18529d8527887e623872f6d6c1d8b3</originalsourceid><addsrcrecordid>eNpdkcFuGyEQhlHVKo6cPEDUC1IuvazLgJeFY71KE0eOekmUI5rdZR2iNTiwtuSe-ujFcppDOcBo5vuHGf2EXAGbATD9_X5xt5xxxvlMgJJaiE_knINUBedMff4Xg55PyGVKrywflVNanpGJKCsp5rI6J3_ug_MjXYUWB_cbRxc8Rd_ResCUXO_aUyr0dBEtppHW6FsbqfN0UTyEztKnYYyYwi6LlhtcO7-me4e0DsOATYhZv7d0ZTH6Y-nZjS_0Jjcfwzri9uVwQb70OCR7-f5OydPPm8f6rlj9ul3WP1ZFK7gci66sEDTmRUuFpWC9QNFAZ1UjJTZWNvMOVMl1l69KqcpKLlTFe9nJFjrViCn5duq7jeFtZ9NoNi61Ng_pbdglw6UCBaDmZUav_0Nfwy76PJ3hFXCWv9dVpuBEtTGkFG1vttFtMB4MMHN0yBwdMkeHzLtDWfP1pHHW2g9eK6YVB_EXPXqKmw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2712053097</pqid></control><display><type>article</type><title>Joint Localization and Classification of Breast Cancer in B-Mode Ultrasound Imaging via Collaborative Learning With Elastography</title><source>IEEE Electronic Library (IEL) Journals</source><creator>Ding, Weichang ; Wang, Jun ; Zhou, Weijun ; Zhou, Shichong ; Chang, Cai ; Shi, Jun</creator><creatorcontrib>Ding, Weichang ; Wang, Jun ; Zhou, Weijun ; Zhou, Shichong ; Chang, Cai ; Shi, Jun</creatorcontrib><description>Convolutional neural networks (CNNs) have been successfully applied in the computer-aided ultrasound diagnosis for breast cancer. Up to now, several CNN-based methods have been proposed. However, most of them consider tumor localization and classification as two separate steps, rather than performing them simultaneously. Besides, they suffer from the limited diagnosis information in the B-mode ultrasound (BUS) images. In this study, we develop a novel network ResNet-GAP that incorporates both localization and classification into a unified procedure. To enhance the performance of ResNet-GAP, we leverage stiffness information in the elastography ultrasound (EUS) modality by collaborative learning in the training stage. Specifically, a dual-channel ResNet-GAP network is developed, one channel for BUS and the other for EUS. In each channel, multiple class activity maps (CAMs) are generated using a series of convolutional kernels of different sizes. The multi-scale consistency of the CAMs in both channels are further considered in network optimization. Experiments on 264 patients in this study show that the newly developed ResNet-GAP achieves an accuracy of 88.6%, a sensitivity of 95.3%, a specificity of 84.6%, and an AUC of 93.6% on the classification task, and a 1.0NLF of 87.9% on the localization task, which is better than some state-of-the-art approaches.</description><identifier>ISSN: 2168-2194</identifier><identifier>EISSN: 2168-2208</identifier><identifier>DOI: 10.1109/JBHI.2022.3186933</identifier><identifier>PMID: 35763467</identifier><identifier>CODEN: IJBHA9</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Artificial neural networks ; Breast cancer ; Cams ; Classification ; Collaborative learning ; collaborative learning with multi-scale consistency ; Convolutional neural networks ; Deep learning ; Diagnosis ; Learning ; Localization ; Location awareness ; Network management systems ; Neural networks ; Optimization ; Stiffness ; tumor classification ; tumor localization ; Tumors ; Ultrasonic imaging ; Ultrasound</subject><ispartof>IEEE journal of biomedical and health informatics, 2022-09, Vol.26 (9), p.4474-4485</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c326t-d57a19a86958a530f3a3b1de8b66abe6b4d18529d8527887e623872f6d6c1d8b3</citedby><cites>FETCH-LOGICAL-c326t-d57a19a86958a530f3a3b1de8b66abe6b4d18529d8527887e623872f6d6c1d8b3</cites><orcidid>0000-0002-3226-3978 ; 0000-0001-9548-0411</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9809821$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,54796</link.rule.ids></links><search><creatorcontrib>Ding, Weichang</creatorcontrib><creatorcontrib>Wang, Jun</creatorcontrib><creatorcontrib>Zhou, Weijun</creatorcontrib><creatorcontrib>Zhou, Shichong</creatorcontrib><creatorcontrib>Chang, Cai</creatorcontrib><creatorcontrib>Shi, Jun</creatorcontrib><title>Joint Localization and Classification of Breast Cancer in B-Mode Ultrasound Imaging via Collaborative Learning With Elastography</title><title>IEEE journal of biomedical and health informatics</title><addtitle>JBHI</addtitle><description>Convolutional neural networks (CNNs) have been successfully applied in the computer-aided ultrasound diagnosis for breast cancer. Up to now, several CNN-based methods have been proposed. However, most of them consider tumor localization and classification as two separate steps, rather than performing them simultaneously. Besides, they suffer from the limited diagnosis information in the B-mode ultrasound (BUS) images. In this study, we develop a novel network ResNet-GAP that incorporates both localization and classification into a unified procedure. To enhance the performance of ResNet-GAP, we leverage stiffness information in the elastography ultrasound (EUS) modality by collaborative learning in the training stage. Specifically, a dual-channel ResNet-GAP network is developed, one channel for BUS and the other for EUS. In each channel, multiple class activity maps (CAMs) are generated using a series of convolutional kernels of different sizes. The multi-scale consistency of the CAMs in both channels are further considered in network optimization. Experiments on 264 patients in this study show that the newly developed ResNet-GAP achieves an accuracy of 88.6%, a sensitivity of 95.3%, a specificity of 84.6%, and an AUC of 93.6% on the classification task, and a 1.0NLF of 87.9% on the localization task, which is better than some state-of-the-art approaches.</description><subject>Artificial neural networks</subject><subject>Breast cancer</subject><subject>Cams</subject><subject>Classification</subject><subject>Collaborative learning</subject><subject>collaborative learning with multi-scale consistency</subject><subject>Convolutional neural networks</subject><subject>Deep learning</subject><subject>Diagnosis</subject><subject>Learning</subject><subject>Localization</subject><subject>Location awareness</subject><subject>Network management systems</subject><subject>Neural networks</subject><subject>Optimization</subject><subject>Stiffness</subject><subject>tumor classification</subject><subject>tumor localization</subject><subject>Tumors</subject><subject>Ultrasonic imaging</subject><subject>Ultrasound</subject><issn>2168-2194</issn><issn>2168-2208</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNpdkcFuGyEQhlHVKo6cPEDUC1IuvazLgJeFY71KE0eOekmUI5rdZR2iNTiwtuSe-ujFcppDOcBo5vuHGf2EXAGbATD9_X5xt5xxxvlMgJJaiE_knINUBedMff4Xg55PyGVKrywflVNanpGJKCsp5rI6J3_ug_MjXYUWB_cbRxc8Rd_ResCUXO_aUyr0dBEtppHW6FsbqfN0UTyEztKnYYyYwi6LlhtcO7-me4e0DsOATYhZv7d0ZTH6Y-nZjS_0Jjcfwzri9uVwQb70OCR7-f5OydPPm8f6rlj9ul3WP1ZFK7gci66sEDTmRUuFpWC9QNFAZ1UjJTZWNvMOVMl1l69KqcpKLlTFe9nJFjrViCn5duq7jeFtZ9NoNi61Ng_pbdglw6UCBaDmZUav_0Nfwy76PJ3hFXCWv9dVpuBEtTGkFG1vttFtMB4MMHN0yBwdMkeHzLtDWfP1pHHW2g9eK6YVB_EXPXqKmw</recordid><startdate>20220901</startdate><enddate>20220901</enddate><creator>Ding, Weichang</creator><creator>Wang, Jun</creator><creator>Zhou, Weijun</creator><creator>Zhou, Shichong</creator><creator>Chang, Cai</creator><creator>Shi, Jun</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QF</scope><scope>7QO</scope><scope>7QQ</scope><scope>7SC</scope><scope>7SE</scope><scope>7SP</scope><scope>7SR</scope><scope>7TA</scope><scope>7TB</scope><scope>7U5</scope><scope>8BQ</scope><scope>8FD</scope><scope>F28</scope><scope>FR3</scope><scope>H8D</scope><scope>JG9</scope><scope>JQ2</scope><scope>K9.</scope><scope>KR7</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>NAPCQ</scope><scope>P64</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0002-3226-3978</orcidid><orcidid>https://orcid.org/0000-0001-9548-0411</orcidid></search><sort><creationdate>20220901</creationdate><title>Joint Localization and Classification of Breast Cancer in B-Mode Ultrasound Imaging via Collaborative Learning With Elastography</title><author>Ding, Weichang ; Wang, Jun ; Zhou, Weijun ; Zhou, Shichong ; Chang, Cai ; Shi, Jun</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c326t-d57a19a86958a530f3a3b1de8b66abe6b4d18529d8527887e623872f6d6c1d8b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Artificial neural networks</topic><topic>Breast cancer</topic><topic>Cams</topic><topic>Classification</topic><topic>Collaborative learning</topic><topic>collaborative learning with multi-scale consistency</topic><topic>Convolutional neural networks</topic><topic>Deep learning</topic><topic>Diagnosis</topic><topic>Learning</topic><topic>Localization</topic><topic>Location awareness</topic><topic>Network management systems</topic><topic>Neural networks</topic><topic>Optimization</topic><topic>Stiffness</topic><topic>tumor classification</topic><topic>tumor localization</topic><topic>Tumors</topic><topic>Ultrasonic imaging</topic><topic>Ultrasound</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Ding, Weichang</creatorcontrib><creatorcontrib>Wang, Jun</creatorcontrib><creatorcontrib>Zhou, Weijun</creatorcontrib><creatorcontrib>Zhou, Shichong</creatorcontrib><creatorcontrib>Chang, Cai</creatorcontrib><creatorcontrib>Shi, Jun</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE/IET Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Aluminium Industry Abstracts</collection><collection>Biotechnology Research Abstracts</collection><collection>Ceramic Abstracts</collection><collection>Computer and Information Systems Abstracts</collection><collection>Corrosion Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>Materials Business File</collection><collection>Mechanical &amp; Transportation Engineering Abstracts</collection><collection>Solid State and Superconductivity Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>ANTE: Abstracts in New Technology &amp; Engineering</collection><collection>Engineering Research Database</collection><collection>Aerospace Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Health &amp; Medical Complete (Alumni)</collection><collection>Civil Engineering Abstracts</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Nursing &amp; Allied Health Premium</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE journal of biomedical and health informatics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Ding, Weichang</au><au>Wang, Jun</au><au>Zhou, Weijun</au><au>Zhou, Shichong</au><au>Chang, Cai</au><au>Shi, Jun</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Joint Localization and Classification of Breast Cancer in B-Mode Ultrasound Imaging via Collaborative Learning With Elastography</atitle><jtitle>IEEE journal of biomedical and health informatics</jtitle><stitle>JBHI</stitle><date>2022-09-01</date><risdate>2022</risdate><volume>26</volume><issue>9</issue><spage>4474</spage><epage>4485</epage><pages>4474-4485</pages><issn>2168-2194</issn><eissn>2168-2208</eissn><coden>IJBHA9</coden><abstract>Convolutional neural networks (CNNs) have been successfully applied in the computer-aided ultrasound diagnosis for breast cancer. Up to now, several CNN-based methods have been proposed. However, most of them consider tumor localization and classification as two separate steps, rather than performing them simultaneously. Besides, they suffer from the limited diagnosis information in the B-mode ultrasound (BUS) images. In this study, we develop a novel network ResNet-GAP that incorporates both localization and classification into a unified procedure. To enhance the performance of ResNet-GAP, we leverage stiffness information in the elastography ultrasound (EUS) modality by collaborative learning in the training stage. Specifically, a dual-channel ResNet-GAP network is developed, one channel for BUS and the other for EUS. In each channel, multiple class activity maps (CAMs) are generated using a series of convolutional kernels of different sizes. The multi-scale consistency of the CAMs in both channels are further considered in network optimization. Experiments on 264 patients in this study show that the newly developed ResNet-GAP achieves an accuracy of 88.6%, a sensitivity of 95.3%, a specificity of 84.6%, and an AUC of 93.6% on the classification task, and a 1.0NLF of 87.9% on the localization task, which is better than some state-of-the-art approaches.</abstract><cop>Piscataway</cop><pub>IEEE</pub><pmid>35763467</pmid><doi>10.1109/JBHI.2022.3186933</doi><tpages>12</tpages><orcidid>https://orcid.org/0000-0002-3226-3978</orcidid><orcidid>https://orcid.org/0000-0001-9548-0411</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 2168-2194
ispartof IEEE journal of biomedical and health informatics, 2022-09, Vol.26 (9), p.4474-4485
issn 2168-2194
2168-2208
language eng
recordid cdi_proquest_miscellaneous_2681811845
source IEEE Electronic Library (IEL) Journals
subjects Artificial neural networks
Breast cancer
Cams
Classification
Collaborative learning
collaborative learning with multi-scale consistency
Convolutional neural networks
Deep learning
Diagnosis
Learning
Localization
Location awareness
Network management systems
Neural networks
Optimization
Stiffness
tumor classification
tumor localization
Tumors
Ultrasonic imaging
Ultrasound
title Joint Localization and Classification of Breast Cancer in B-Mode Ultrasound Imaging via Collaborative Learning With Elastography
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T17%3A43%3A28IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Joint%20Localization%20and%20Classification%20of%20Breast%20Cancer%20in%20B-Mode%20Ultrasound%20Imaging%20via%20Collaborative%20Learning%20With%20Elastography&rft.jtitle=IEEE%20journal%20of%20biomedical%20and%20health%20informatics&rft.au=Ding,%20Weichang&rft.date=2022-09-01&rft.volume=26&rft.issue=9&rft.spage=4474&rft.epage=4485&rft.pages=4474-4485&rft.issn=2168-2194&rft.eissn=2168-2208&rft.coden=IJBHA9&rft_id=info:doi/10.1109/JBHI.2022.3186933&rft_dat=%3Cproquest_ieee_%3E2712053097%3C/proquest_ieee_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c326t-d57a19a86958a530f3a3b1de8b66abe6b4d18529d8527887e623872f6d6c1d8b3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2712053097&rft_id=info:pmid/35763467&rft_ieee_id=9809821&rfr_iscdi=true