Loading…
PCANet-Based Structural Representation for Nonrigid Multimodal Medical Image Registration
Nonrigid multimodal image registration remains a challenging task in medical image processing and analysis. The structural representation (SR)-based registration methods have attracted much attention recently. However, the existing SR methods cannot provide satisfactory registration accuracy due to...
Saved in:
Published in: | Sensors (Basel, Switzerland) Switzerland), 2018-05, Vol.18 (5), p.1477 |
---|---|
Main Authors: | , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c535t-7aa79ef9caaaab37e87a42b8f79e65e867652e86d2b0994dcb708465069e188d3 |
---|---|
cites | cdi_FETCH-LOGICAL-c535t-7aa79ef9caaaab37e87a42b8f79e65e867652e86d2b0994dcb708465069e188d3 |
container_end_page | |
container_issue | 5 |
container_start_page | 1477 |
container_title | Sensors (Basel, Switzerland) |
container_volume | 18 |
creator | Zhu, Xingxing Ding, Mingyue Huang, Tao Jin, Xiaomeng Zhang, Xuming |
description | Nonrigid multimodal image registration remains a challenging task in medical image processing and analysis. The structural representation (SR)-based registration methods have attracted much attention recently. However, the existing SR methods cannot provide satisfactory registration accuracy due to the utilization of hand-designed features for structural representation. To address this problem, the structural representation method based on the improved version of the simple deep learning network named PCANet is proposed for medical image registration. In the proposed method, PCANet is firstly trained on numerous medical images to learn convolution kernels for this network. Then, a pair of input medical images to be registered is processed by the learned PCANet. The features extracted by various layers in the PCANet are fused to produce multilevel features. The structural representation images are constructed for two input images based on nonlinear transformation of these multilevel features. The Euclidean distance between structural representation images is calculated and used as the similarity metrics. The objective function defined by the similarity metrics is optimized by L-BFGS method to obtain parameters of the free-form deformation (FFD) model. Extensive experiments on simulated and real multimodal image datasets show that compared with the state-of-the-art registration methods, such as modality-independent neighborhood descriptor (MIND), normalized mutual information (NMI), Weber local descriptor (WLD), and the sum of squared differences on entropy images (ESSD), the proposed method provides better registration performance in terms of target registration error (TRE) and subjective human vision. |
doi_str_mv | 10.3390/s18051477 |
format | article |
fullrecord | <record><control><sourceid>proquest_doaj_</sourceid><recordid>TN_cdi_doaj_primary_oai_doaj_org_article_1344a63dcb884569a81ea52d8b052709</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><doaj_id>oai_doaj_org_article_1344a63dcb884569a81ea52d8b052709</doaj_id><sourcerecordid>2036792364</sourcerecordid><originalsourceid>FETCH-LOGICAL-c535t-7aa79ef9caaaab37e87a42b8f79e65e867652e86d2b0994dcb708465069e188d3</originalsourceid><addsrcrecordid>eNpdkUtv1DAQgCNERUvhwB9AkbjAIdTvxwWprHis1AficeBkOfEkeJXEi-0g8e_xdsuqrS9jjb_5NJ6pqhcYvaVUo7OEFeKYSfmoOsGMsEYRgh7fuR9XT1PaIEQopepJdUy0pIpjclL9_LI6v4LcvLcJXP0tx6XLS7Rj_RW2ERLM2WYf5roPsb4Kc_SDd_XlMmY_BVewS3C-K3E92QFK0eBTjjclz6qj3o4Jnt_G0-rHxw_fV5-bi-tP69X5RdNxynMjrZUaet3ZcloqQUnLSKv6khUclJCCkxIcaZHWzHWtRIoJjoQGrJSjp9V673XBbsw2-snGvyZYb24SIQ7Gxuy7EQymjFlBi0MpxoW2CoPlxKkWcSKRLq53e9d2aSdwXfl-mcU96f2X2f8yQ_hjuFaEiZ3g9a0ght8LpGwmnzoYRztDWJIhiAqpCRWsoK8eoJuwxLmMyhCMlCRY4B31Zk91MaQUoT80g5HZLd8cll_Yl3e7P5D_t03_ARGsqTU</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2108721614</pqid></control><display><type>article</type><title>PCANet-Based Structural Representation for Nonrigid Multimodal Medical Image Registration</title><source>Publicly Available Content Database</source><source>PubMed Central</source><creator>Zhu, Xingxing ; Ding, Mingyue ; Huang, Tao ; Jin, Xiaomeng ; Zhang, Xuming</creator><creatorcontrib>Zhu, Xingxing ; Ding, Mingyue ; Huang, Tao ; Jin, Xiaomeng ; Zhang, Xuming</creatorcontrib><description>Nonrigid multimodal image registration remains a challenging task in medical image processing and analysis. The structural representation (SR)-based registration methods have attracted much attention recently. However, the existing SR methods cannot provide satisfactory registration accuracy due to the utilization of hand-designed features for structural representation. To address this problem, the structural representation method based on the improved version of the simple deep learning network named PCANet is proposed for medical image registration. In the proposed method, PCANet is firstly trained on numerous medical images to learn convolution kernels for this network. Then, a pair of input medical images to be registered is processed by the learned PCANet. The features extracted by various layers in the PCANet are fused to produce multilevel features. The structural representation images are constructed for two input images based on nonlinear transformation of these multilevel features. The Euclidean distance between structural representation images is calculated and used as the similarity metrics. The objective function defined by the similarity metrics is optimized by L-BFGS method to obtain parameters of the free-form deformation (FFD) model. Extensive experiments on simulated and real multimodal image datasets show that compared with the state-of-the-art registration methods, such as modality-independent neighborhood descriptor (MIND), normalized mutual information (NMI), Weber local descriptor (WLD), and the sum of squared differences on entropy images (ESSD), the proposed method provides better registration performance in terms of target registration error (TRE) and subjective human vision.</description><identifier>ISSN: 1424-8220</identifier><identifier>EISSN: 1424-8220</identifier><identifier>DOI: 10.3390/s18051477</identifier><identifier>PMID: 29738512</identifier><language>eng</language><publisher>Switzerland: MDPI AG</publisher><subject>Computer simulation ; Convolution ; Deformation ; Euclidean geometry ; Image processing ; Image registration ; medical image registration ; Medical imaging ; Methods ; PCANet ; Registration ; Representations ; Similarity ; similarity metric ; structural representation ; target registration error</subject><ispartof>Sensors (Basel, Switzerland), 2018-05, Vol.18 (5), p.1477</ispartof><rights>2018. This work is licensed under https://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>2018 by the authors. 2018</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c535t-7aa79ef9caaaab37e87a42b8f79e65e867652e86d2b0994dcb708465069e188d3</citedby><cites>FETCH-LOGICAL-c535t-7aa79ef9caaaab37e87a42b8f79e65e867652e86d2b0994dcb708465069e188d3</cites><orcidid>0000-0003-4332-071X ; 0000-0002-2541-8222</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.proquest.com/docview/2108721614/fulltextPDF?pq-origsite=primo$$EPDF$$P50$$Gproquest$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.proquest.com/docview/2108721614?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>230,314,723,776,780,881,25731,27901,27902,36989,36990,44566,53766,53768,74869</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/29738512$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Zhu, Xingxing</creatorcontrib><creatorcontrib>Ding, Mingyue</creatorcontrib><creatorcontrib>Huang, Tao</creatorcontrib><creatorcontrib>Jin, Xiaomeng</creatorcontrib><creatorcontrib>Zhang, Xuming</creatorcontrib><title>PCANet-Based Structural Representation for Nonrigid Multimodal Medical Image Registration</title><title>Sensors (Basel, Switzerland)</title><addtitle>Sensors (Basel)</addtitle><description>Nonrigid multimodal image registration remains a challenging task in medical image processing and analysis. The structural representation (SR)-based registration methods have attracted much attention recently. However, the existing SR methods cannot provide satisfactory registration accuracy due to the utilization of hand-designed features for structural representation. To address this problem, the structural representation method based on the improved version of the simple deep learning network named PCANet is proposed for medical image registration. In the proposed method, PCANet is firstly trained on numerous medical images to learn convolution kernels for this network. Then, a pair of input medical images to be registered is processed by the learned PCANet. The features extracted by various layers in the PCANet are fused to produce multilevel features. The structural representation images are constructed for two input images based on nonlinear transformation of these multilevel features. The Euclidean distance between structural representation images is calculated and used as the similarity metrics. The objective function defined by the similarity metrics is optimized by L-BFGS method to obtain parameters of the free-form deformation (FFD) model. Extensive experiments on simulated and real multimodal image datasets show that compared with the state-of-the-art registration methods, such as modality-independent neighborhood descriptor (MIND), normalized mutual information (NMI), Weber local descriptor (WLD), and the sum of squared differences on entropy images (ESSD), the proposed method provides better registration performance in terms of target registration error (TRE) and subjective human vision.</description><subject>Computer simulation</subject><subject>Convolution</subject><subject>Deformation</subject><subject>Euclidean geometry</subject><subject>Image processing</subject><subject>Image registration</subject><subject>medical image registration</subject><subject>Medical imaging</subject><subject>Methods</subject><subject>PCANet</subject><subject>Registration</subject><subject>Representations</subject><subject>Similarity</subject><subject>similarity metric</subject><subject>structural representation</subject><subject>target registration error</subject><issn>1424-8220</issn><issn>1424-8220</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2018</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><sourceid>DOA</sourceid><recordid>eNpdkUtv1DAQgCNERUvhwB9AkbjAIdTvxwWprHis1AficeBkOfEkeJXEi-0g8e_xdsuqrS9jjb_5NJ6pqhcYvaVUo7OEFeKYSfmoOsGMsEYRgh7fuR9XT1PaIEQopepJdUy0pIpjclL9_LI6v4LcvLcJXP0tx6XLS7Rj_RW2ERLM2WYf5roPsb4Kc_SDd_XlMmY_BVewS3C-K3E92QFK0eBTjjclz6qj3o4Jnt_G0-rHxw_fV5-bi-tP69X5RdNxynMjrZUaet3ZcloqQUnLSKv6khUclJCCkxIcaZHWzHWtRIoJjoQGrJSjp9V673XBbsw2-snGvyZYb24SIQ7Gxuy7EQymjFlBi0MpxoW2CoPlxKkWcSKRLq53e9d2aSdwXfl-mcU96f2X2f8yQ_hjuFaEiZ3g9a0ght8LpGwmnzoYRztDWJIhiAqpCRWsoK8eoJuwxLmMyhCMlCRY4B31Zk91MaQUoT80g5HZLd8cll_Yl3e7P5D_t03_ARGsqTU</recordid><startdate>20180508</startdate><enddate>20180508</enddate><creator>Zhu, Xingxing</creator><creator>Ding, Mingyue</creator><creator>Huang, Tao</creator><creator>Jin, Xiaomeng</creator><creator>Zhang, Xuming</creator><general>MDPI AG</general><general>MDPI</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7X7</scope><scope>7XB</scope><scope>88E</scope><scope>8FI</scope><scope>8FJ</scope><scope>8FK</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FYUFA</scope><scope>GHDGH</scope><scope>K9.</scope><scope>M0S</scope><scope>M1P</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>7X8</scope><scope>5PM</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0003-4332-071X</orcidid><orcidid>https://orcid.org/0000-0002-2541-8222</orcidid></search><sort><creationdate>20180508</creationdate><title>PCANet-Based Structural Representation for Nonrigid Multimodal Medical Image Registration</title><author>Zhu, Xingxing ; Ding, Mingyue ; Huang, Tao ; Jin, Xiaomeng ; Zhang, Xuming</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c535t-7aa79ef9caaaab37e87a42b8f79e65e867652e86d2b0994dcb708465069e188d3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2018</creationdate><topic>Computer simulation</topic><topic>Convolution</topic><topic>Deformation</topic><topic>Euclidean geometry</topic><topic>Image processing</topic><topic>Image registration</topic><topic>medical image registration</topic><topic>Medical imaging</topic><topic>Methods</topic><topic>PCANet</topic><topic>Registration</topic><topic>Representations</topic><topic>Similarity</topic><topic>similarity metric</topic><topic>structural representation</topic><topic>target registration error</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhu, Xingxing</creatorcontrib><creatorcontrib>Ding, Mingyue</creatorcontrib><creatorcontrib>Huang, Tao</creatorcontrib><creatorcontrib>Jin, Xiaomeng</creatorcontrib><creatorcontrib>Zhang, Xuming</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Health & Medical Collection</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Medical Database (Alumni Edition)</collection><collection>Hospital Premium Collection</collection><collection>Hospital Premium Collection (Alumni Edition)</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>Health Research Premium Collection</collection><collection>Health Research Premium Collection (Alumni)</collection><collection>ProQuest Health & Medical Complete (Alumni)</collection><collection>Health & Medical Collection (Alumni Edition)</collection><collection>Medical Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>Sensors (Basel, Switzerland)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhu, Xingxing</au><au>Ding, Mingyue</au><au>Huang, Tao</au><au>Jin, Xiaomeng</au><au>Zhang, Xuming</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>PCANet-Based Structural Representation for Nonrigid Multimodal Medical Image Registration</atitle><jtitle>Sensors (Basel, Switzerland)</jtitle><addtitle>Sensors (Basel)</addtitle><date>2018-05-08</date><risdate>2018</risdate><volume>18</volume><issue>5</issue><spage>1477</spage><pages>1477-</pages><issn>1424-8220</issn><eissn>1424-8220</eissn><abstract>Nonrigid multimodal image registration remains a challenging task in medical image processing and analysis. The structural representation (SR)-based registration methods have attracted much attention recently. However, the existing SR methods cannot provide satisfactory registration accuracy due to the utilization of hand-designed features for structural representation. To address this problem, the structural representation method based on the improved version of the simple deep learning network named PCANet is proposed for medical image registration. In the proposed method, PCANet is firstly trained on numerous medical images to learn convolution kernels for this network. Then, a pair of input medical images to be registered is processed by the learned PCANet. The features extracted by various layers in the PCANet are fused to produce multilevel features. The structural representation images are constructed for two input images based on nonlinear transformation of these multilevel features. The Euclidean distance between structural representation images is calculated and used as the similarity metrics. The objective function defined by the similarity metrics is optimized by L-BFGS method to obtain parameters of the free-form deformation (FFD) model. Extensive experiments on simulated and real multimodal image datasets show that compared with the state-of-the-art registration methods, such as modality-independent neighborhood descriptor (MIND), normalized mutual information (NMI), Weber local descriptor (WLD), and the sum of squared differences on entropy images (ESSD), the proposed method provides better registration performance in terms of target registration error (TRE) and subjective human vision.</abstract><cop>Switzerland</cop><pub>MDPI AG</pub><pmid>29738512</pmid><doi>10.3390/s18051477</doi><orcidid>https://orcid.org/0000-0003-4332-071X</orcidid><orcidid>https://orcid.org/0000-0002-2541-8222</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1424-8220 |
ispartof | Sensors (Basel, Switzerland), 2018-05, Vol.18 (5), p.1477 |
issn | 1424-8220 1424-8220 |
language | eng |
recordid | cdi_doaj_primary_oai_doaj_org_article_1344a63dcb884569a81ea52d8b052709 |
source | Publicly Available Content Database; PubMed Central |
subjects | Computer simulation Convolution Deformation Euclidean geometry Image processing Image registration medical image registration Medical imaging Methods PCANet Registration Representations Similarity similarity metric structural representation target registration error |
title | PCANet-Based Structural Representation for Nonrigid Multimodal Medical Image Registration |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-09T22%3A32%3A02IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=PCANet-Based%20Structural%20Representation%20for%20Nonrigid%20Multimodal%20Medical%20Image%20Registration&rft.jtitle=Sensors%20(Basel,%20Switzerland)&rft.au=Zhu,%20Xingxing&rft.date=2018-05-08&rft.volume=18&rft.issue=5&rft.spage=1477&rft.pages=1477-&rft.issn=1424-8220&rft.eissn=1424-8220&rft_id=info:doi/10.3390/s18051477&rft_dat=%3Cproquest_doaj_%3E2036792364%3C/proquest_doaj_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c535t-7aa79ef9caaaab37e87a42b8f79e65e867652e86d2b0994dcb708465069e188d3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2108721614&rft_id=info:pmid/29738512&rfr_iscdi=true |