Loading…

MCSTransWnet: A new deep learning process for postoperative corneal topography prediction based on raw multimodal data from the Pentacam HR system

This work provides a new multimodal fusion generative adversarial net (GAN) model, Multiple Conditions Transform W-net (MCSTransWnet), which primarily uses femtosecond laser arcuate keratotomy surgical parameters and preoperative corneal topography to predict postoperative corneal topography in asti...

Full description

Saved in:
Bibliographic Details
Published in:Medicine in novel technology and devices 2024-03, Vol.21, p.100267, Article 100267
Main Authors: Cheng, Nan, Zhang, Zhe, Pan, Jing, Li, Xiao-Na, Chen, Wei-Yi, Zhang, Guang-Hua, Yang, Wei-Hua
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites cdi_FETCH-LOGICAL-c2827-9f89ac936f7044411d81e36c1a99c98ffda0344934ef8d6d0dac5b4312cba573
container_end_page
container_issue
container_start_page 100267
container_title Medicine in novel technology and devices
container_volume 21
creator Cheng, Nan
Zhang, Zhe
Pan, Jing
Li, Xiao-Na
Chen, Wei-Yi
Zhang, Guang-Hua
Yang, Wei-Hua
description This work provides a new multimodal fusion generative adversarial net (GAN) model, Multiple Conditions Transform W-net (MCSTransWnet), which primarily uses femtosecond laser arcuate keratotomy surgical parameters and preoperative corneal topography to predict postoperative corneal topography in astigmatism-corrected patients. The MCSTransWnet model comprises a generator and a discriminator, and the generator is composed of two sub-generators. The first sub-generator extracts features using the U-net model, vision transform (ViT) and a multi-parameter conditional module branch. ​The second sub-generator uses a U-net network for further image denoising. The discriminator uses the pixel discriminator in Pix2Pix. Currently, most GAN models are convolutional neural networks; however, due to their feature extraction locality, it is difficult to comprehend the relationships among global features. Thus, we added a vision Transform network as the model branch to extract the global features. It is normally difficult to train the transformer, and image noise and geometric information loss are likely. Hence, we adopted the standard U-net fusion scheme and transform network as the generator, so that global features, local features, and rich image details could be obtained simultaneously. Our experimental results clearly demonstrate that MCSTransWnet successfully predicts postoperative corneal topographies (structural similarity ​= ​0.765, peak signal-to-noise ratio ​= ​16.012, and Fréchet inception distance ​= ​9.264). Using this technique to obtain the rough shape of the postoperative corneal topography in advance gives clinicians more references and guides changes to surgical planning and improves the success rate of surgery. •Predicting postoperative corneal topography using the generative adversarial network Pix2Pix.•Feasibility of generating adversarial networks to predict postoperative corneal topography.•Predicting postoperative corneal topography using multimodal fusion techniques.•Using multimodal fusion techniques to fuse surgical text data with image data.
doi_str_mv 10.1016/j.medntd.2023.100267
format article
fullrecord <record><control><sourceid>elsevier_doaj_</sourceid><recordid>TN_cdi_doaj_primary_oai_doaj_org_article_329a1ff336be46568290a59e328a7728</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S2590093523000620</els_id><doaj_id>oai_doaj_org_article_329a1ff336be46568290a59e328a7728</doaj_id><sourcerecordid>S2590093523000620</sourcerecordid><originalsourceid>FETCH-LOGICAL-c2827-9f89ac936f7044411d81e36c1a99c98ffda0344934ef8d6d0dac5b4312cba573</originalsourceid><addsrcrecordid>eNp9kd2KFDEQhRtRcFn3DbzIC8yYv053vBCWQd2FFUUHvAw1SWU2Q3fSJHGXeQ2f2Kwt4pVXVRxOfVTV6brXjG4ZZerNaTuji9VtOeWiSZSr4Vl3wXtNN1SL_vk__cvuqpQTbZ6eMabkRffz0-7bPkMs3yPWt-SaRHwkDnEhE0KOIR7JkpPFUohPmSyp1LRghhoekNiUI8JEmpSOGZb7czOjC7aGFMkBCjrSmgyPZP4x1TAn19wOKhCf00zqPZIvGCtYmMnNV1LOpeL8qnvhYSp49adedvsP7_e7m83d54-3u-u7jeUjHzbajxqsFsoPVErJmBsZCmUZaG316L0DKqTUQqIfnXLUge0PUjBuD9AP4rK7XbEuwcksOcyQzyZBML-FlI8Gcg12QiO4Bua9EOqAUvVq5JpCr1HwEYaBj40lV5bNqZSM_i-PUfOUkjmZNSXzlJJZU2pj79YxbGc-BMym2IDRthdmtLUtEv4P-AUt2J5u</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>MCSTransWnet: A new deep learning process for postoperative corneal topography prediction based on raw multimodal data from the Pentacam HR system</title><source>ScienceDirect (Online service)</source><creator>Cheng, Nan ; Zhang, Zhe ; Pan, Jing ; Li, Xiao-Na ; Chen, Wei-Yi ; Zhang, Guang-Hua ; Yang, Wei-Hua</creator><creatorcontrib>Cheng, Nan ; Zhang, Zhe ; Pan, Jing ; Li, Xiao-Na ; Chen, Wei-Yi ; Zhang, Guang-Hua ; Yang, Wei-Hua</creatorcontrib><description>This work provides a new multimodal fusion generative adversarial net (GAN) model, Multiple Conditions Transform W-net (MCSTransWnet), which primarily uses femtosecond laser arcuate keratotomy surgical parameters and preoperative corneal topography to predict postoperative corneal topography in astigmatism-corrected patients. The MCSTransWnet model comprises a generator and a discriminator, and the generator is composed of two sub-generators. The first sub-generator extracts features using the U-net model, vision transform (ViT) and a multi-parameter conditional module branch. ​The second sub-generator uses a U-net network for further image denoising. The discriminator uses the pixel discriminator in Pix2Pix. Currently, most GAN models are convolutional neural networks; however, due to their feature extraction locality, it is difficult to comprehend the relationships among global features. Thus, we added a vision Transform network as the model branch to extract the global features. It is normally difficult to train the transformer, and image noise and geometric information loss are likely. Hence, we adopted the standard U-net fusion scheme and transform network as the generator, so that global features, local features, and rich image details could be obtained simultaneously. Our experimental results clearly demonstrate that MCSTransWnet successfully predicts postoperative corneal topographies (structural similarity ​= ​0.765, peak signal-to-noise ratio ​= ​16.012, and Fréchet inception distance ​= ​9.264). Using this technique to obtain the rough shape of the postoperative corneal topography in advance gives clinicians more references and guides changes to surgical planning and improves the success rate of surgery. •Predicting postoperative corneal topography using the generative adversarial network Pix2Pix.•Feasibility of generating adversarial networks to predict postoperative corneal topography.•Predicting postoperative corneal topography using multimodal fusion techniques.•Using multimodal fusion techniques to fuse surgical text data with image data.</description><identifier>ISSN: 2590-0935</identifier><identifier>EISSN: 2590-0935</identifier><identifier>DOI: 10.1016/j.medntd.2023.100267</identifier><language>eng</language><publisher>Elsevier B.V</publisher><subject>Corneal topography ; Deep learning ; Generative adversarial networks ; Medical imaging ; Multimodal fusion ; Transformer ; U-net ; W-net</subject><ispartof>Medicine in novel technology and devices, 2024-03, Vol.21, p.100267, Article 100267</ispartof><rights>2023</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c2827-9f89ac936f7044411d81e36c1a99c98ffda0344934ef8d6d0dac5b4312cba573</cites><orcidid>0000-0002-7629-0193</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.sciencedirect.com/science/article/pii/S2590093523000620$$EHTML$$P50$$Gelsevier$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,3547,27922,27923,45778</link.rule.ids></links><search><creatorcontrib>Cheng, Nan</creatorcontrib><creatorcontrib>Zhang, Zhe</creatorcontrib><creatorcontrib>Pan, Jing</creatorcontrib><creatorcontrib>Li, Xiao-Na</creatorcontrib><creatorcontrib>Chen, Wei-Yi</creatorcontrib><creatorcontrib>Zhang, Guang-Hua</creatorcontrib><creatorcontrib>Yang, Wei-Hua</creatorcontrib><title>MCSTransWnet: A new deep learning process for postoperative corneal topography prediction based on raw multimodal data from the Pentacam HR system</title><title>Medicine in novel technology and devices</title><description>This work provides a new multimodal fusion generative adversarial net (GAN) model, Multiple Conditions Transform W-net (MCSTransWnet), which primarily uses femtosecond laser arcuate keratotomy surgical parameters and preoperative corneal topography to predict postoperative corneal topography in astigmatism-corrected patients. The MCSTransWnet model comprises a generator and a discriminator, and the generator is composed of two sub-generators. The first sub-generator extracts features using the U-net model, vision transform (ViT) and a multi-parameter conditional module branch. ​The second sub-generator uses a U-net network for further image denoising. The discriminator uses the pixel discriminator in Pix2Pix. Currently, most GAN models are convolutional neural networks; however, due to their feature extraction locality, it is difficult to comprehend the relationships among global features. Thus, we added a vision Transform network as the model branch to extract the global features. It is normally difficult to train the transformer, and image noise and geometric information loss are likely. Hence, we adopted the standard U-net fusion scheme and transform network as the generator, so that global features, local features, and rich image details could be obtained simultaneously. Our experimental results clearly demonstrate that MCSTransWnet successfully predicts postoperative corneal topographies (structural similarity ​= ​0.765, peak signal-to-noise ratio ​= ​16.012, and Fréchet inception distance ​= ​9.264). Using this technique to obtain the rough shape of the postoperative corneal topography in advance gives clinicians more references and guides changes to surgical planning and improves the success rate of surgery. •Predicting postoperative corneal topography using the generative adversarial network Pix2Pix.•Feasibility of generating adversarial networks to predict postoperative corneal topography.•Predicting postoperative corneal topography using multimodal fusion techniques.•Using multimodal fusion techniques to fuse surgical text data with image data.</description><subject>Corneal topography</subject><subject>Deep learning</subject><subject>Generative adversarial networks</subject><subject>Medical imaging</subject><subject>Multimodal fusion</subject><subject>Transformer</subject><subject>U-net</subject><subject>W-net</subject><issn>2590-0935</issn><issn>2590-0935</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>DOA</sourceid><recordid>eNp9kd2KFDEQhRtRcFn3DbzIC8yYv053vBCWQd2FFUUHvAw1SWU2Q3fSJHGXeQ2f2Kwt4pVXVRxOfVTV6brXjG4ZZerNaTuji9VtOeWiSZSr4Vl3wXtNN1SL_vk__cvuqpQTbZ6eMabkRffz0-7bPkMs3yPWt-SaRHwkDnEhE0KOIR7JkpPFUohPmSyp1LRghhoekNiUI8JEmpSOGZb7czOjC7aGFMkBCjrSmgyPZP4x1TAn19wOKhCf00zqPZIvGCtYmMnNV1LOpeL8qnvhYSp49adedvsP7_e7m83d54-3u-u7jeUjHzbajxqsFsoPVErJmBsZCmUZaG316L0DKqTUQqIfnXLUge0PUjBuD9AP4rK7XbEuwcksOcyQzyZBML-FlI8Gcg12QiO4Bua9EOqAUvVq5JpCr1HwEYaBj40lV5bNqZSM_i-PUfOUkjmZNSXzlJJZU2pj79YxbGc-BMym2IDRthdmtLUtEv4P-AUt2J5u</recordid><startdate>202403</startdate><enddate>202403</enddate><creator>Cheng, Nan</creator><creator>Zhang, Zhe</creator><creator>Pan, Jing</creator><creator>Li, Xiao-Na</creator><creator>Chen, Wei-Yi</creator><creator>Zhang, Guang-Hua</creator><creator>Yang, Wei-Hua</creator><general>Elsevier B.V</general><general>Elsevier</general><scope>6I.</scope><scope>AAFTH</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0002-7629-0193</orcidid></search><sort><creationdate>202403</creationdate><title>MCSTransWnet: A new deep learning process for postoperative corneal topography prediction based on raw multimodal data from the Pentacam HR system</title><author>Cheng, Nan ; Zhang, Zhe ; Pan, Jing ; Li, Xiao-Na ; Chen, Wei-Yi ; Zhang, Guang-Hua ; Yang, Wei-Hua</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c2827-9f89ac936f7044411d81e36c1a99c98ffda0344934ef8d6d0dac5b4312cba573</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Corneal topography</topic><topic>Deep learning</topic><topic>Generative adversarial networks</topic><topic>Medical imaging</topic><topic>Multimodal fusion</topic><topic>Transformer</topic><topic>U-net</topic><topic>W-net</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Cheng, Nan</creatorcontrib><creatorcontrib>Zhang, Zhe</creatorcontrib><creatorcontrib>Pan, Jing</creatorcontrib><creatorcontrib>Li, Xiao-Na</creatorcontrib><creatorcontrib>Chen, Wei-Yi</creatorcontrib><creatorcontrib>Zhang, Guang-Hua</creatorcontrib><creatorcontrib>Yang, Wei-Hua</creatorcontrib><collection>ScienceDirect Open Access Titles</collection><collection>Elsevier:ScienceDirect:Open Access</collection><collection>CrossRef</collection><collection>Directory of Open Access Journals</collection><jtitle>Medicine in novel technology and devices</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Cheng, Nan</au><au>Zhang, Zhe</au><au>Pan, Jing</au><au>Li, Xiao-Na</au><au>Chen, Wei-Yi</au><au>Zhang, Guang-Hua</au><au>Yang, Wei-Hua</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>MCSTransWnet: A new deep learning process for postoperative corneal topography prediction based on raw multimodal data from the Pentacam HR system</atitle><jtitle>Medicine in novel technology and devices</jtitle><date>2024-03</date><risdate>2024</risdate><volume>21</volume><spage>100267</spage><pages>100267-</pages><artnum>100267</artnum><issn>2590-0935</issn><eissn>2590-0935</eissn><abstract>This work provides a new multimodal fusion generative adversarial net (GAN) model, Multiple Conditions Transform W-net (MCSTransWnet), which primarily uses femtosecond laser arcuate keratotomy surgical parameters and preoperative corneal topography to predict postoperative corneal topography in astigmatism-corrected patients. The MCSTransWnet model comprises a generator and a discriminator, and the generator is composed of two sub-generators. The first sub-generator extracts features using the U-net model, vision transform (ViT) and a multi-parameter conditional module branch. ​The second sub-generator uses a U-net network for further image denoising. The discriminator uses the pixel discriminator in Pix2Pix. Currently, most GAN models are convolutional neural networks; however, due to their feature extraction locality, it is difficult to comprehend the relationships among global features. Thus, we added a vision Transform network as the model branch to extract the global features. It is normally difficult to train the transformer, and image noise and geometric information loss are likely. Hence, we adopted the standard U-net fusion scheme and transform network as the generator, so that global features, local features, and rich image details could be obtained simultaneously. Our experimental results clearly demonstrate that MCSTransWnet successfully predicts postoperative corneal topographies (structural similarity ​= ​0.765, peak signal-to-noise ratio ​= ​16.012, and Fréchet inception distance ​= ​9.264). Using this technique to obtain the rough shape of the postoperative corneal topography in advance gives clinicians more references and guides changes to surgical planning and improves the success rate of surgery. •Predicting postoperative corneal topography using the generative adversarial network Pix2Pix.•Feasibility of generating adversarial networks to predict postoperative corneal topography.•Predicting postoperative corneal topography using multimodal fusion techniques.•Using multimodal fusion techniques to fuse surgical text data with image data.</abstract><pub>Elsevier B.V</pub><doi>10.1016/j.medntd.2023.100267</doi><orcidid>https://orcid.org/0000-0002-7629-0193</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2590-0935
ispartof Medicine in novel technology and devices, 2024-03, Vol.21, p.100267, Article 100267
issn 2590-0935
2590-0935
language eng
recordid cdi_doaj_primary_oai_doaj_org_article_329a1ff336be46568290a59e328a7728
source ScienceDirect (Online service)
subjects Corneal topography
Deep learning
Generative adversarial networks
Medical imaging
Multimodal fusion
Transformer
U-net
W-net
title MCSTransWnet: A new deep learning process for postoperative corneal topography prediction based on raw multimodal data from the Pentacam HR system
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-14T10%3A25%3A28IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=MCSTransWnet:%20A%20new%20deep%20learning%20process%20for%20postoperative%20corneal%20topography%20prediction%20based%20on%20raw%20multimodal%20data%20from%20the%20Pentacam%20HR%20system&rft.jtitle=Medicine%20in%20novel%20technology%20and%20devices&rft.au=Cheng,%20Nan&rft.date=2024-03&rft.volume=21&rft.spage=100267&rft.pages=100267-&rft.artnum=100267&rft.issn=2590-0935&rft.eissn=2590-0935&rft_id=info:doi/10.1016/j.medntd.2023.100267&rft_dat=%3Celsevier_doaj_%3ES2590093523000620%3C/elsevier_doaj_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c2827-9f89ac936f7044411d81e36c1a99c98ffda0344934ef8d6d0dac5b4312cba573%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true