Loading…
Building a pelvic organ prolapse diagnostic model using vision transformer on multi‐sequence MRI
Background Although the uterus, bladder, and rectum are distinct organs, their muscular fasciae are often interconnected. Clinical experience suggests that they may share common risk factors and associations. When one organ experiences prolapse, it can potentially affect the neighboring organs. Howe...
Saved in:
Published in: | Medical physics (Lancaster) 2025-01, Vol.52 (1), p.553-564 |
---|---|
Main Authors: | , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | cdi_FETCH-LOGICAL-c2461-7b64dc28d4c74086e10d7f54b8fe5384a220751056ddf4b607706e1ef7d332723 |
container_end_page | 564 |
container_issue | 1 |
container_start_page | 553 |
container_title | Medical physics (Lancaster) |
container_volume | 52 |
creator | Zhu, Shaojun Zhu, Xiaoxuan Zheng, Bo Wu, Maonian Li, Qiongshan Qian, Cheng |
description | Background
Although the uterus, bladder, and rectum are distinct organs, their muscular fasciae are often interconnected. Clinical experience suggests that they may share common risk factors and associations. When one organ experiences prolapse, it can potentially affect the neighboring organs. However, the current assessment of disease severity still relies on manual measurements, which can yield varying results depending on the physician, thereby leading to diagnostic inaccuracies.
Purpose
This study aims to develop a multilabel grading model based on deep learning to classify the degree of prolapse of three organs in the female pelvis using stress magnetic resonance imaging (MRI) and provide interpretable result analysis.
Methods
We utilized sagittal MRI sequences taken at rest and during maximum Valsalva maneuver from 662 subjects. The training set included 464 subjects, the validation set included 98 subjects, and the test set included 100 subjects (training set n = 464, validation set n = 98, test set n = 100). We designed a feature extraction module specifically for pelvic floor MRI using the vision transformer architecture and employed label masking training strategy and pre‐training methods to enhance model convergence. The grading results were evaluated using Precision, Kappa, Recall, and Area Under the Curve (AUC). To validate the effectiveness of the model, the designed model was compared with classic grading methods. Finally, we provided interpretability charts illustrating the model's operational principles on the grading task.
Results
In terms of POP grading detection, the model achieved an average Precision, Kappa coefficient, Recall, and AUC of 0.86, 0.77, 0.76, and 0.86, respectively. Compared to existing studies, our model achieved the highest performance metrics. The average time taken to diagnose a patient was 0.38 s.
Conclusions
The proposed model achieved detection accuracy that is comparable to or even exceeds that of physicians, demonstrating the effectiveness of the vision transformer architecture and label masking training strategy for assisting in the grading of POP under static and maximum Valsalva conditions. This offers a promising option for computer‐aided diagnosis and treatment planning of POP. |
doi_str_mv | 10.1002/mp.17441 |
format | article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_3115968577</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3115968577</sourcerecordid><originalsourceid>FETCH-LOGICAL-c2461-7b64dc28d4c74086e10d7f54b8fe5384a220751056ddf4b607706e1ef7d332723</originalsourceid><addsrcrecordid>eNp1kMtOwzAQRS0EoqUg8QXISzYp41ecLKHiUakVCME6cmKnMnIe2E1Rd3wC38iXkNICK1ajqzk6M7oInRIYEwB6UbVjIjkne2hIuWQRp5DuoyFAyiPKQQzQUQgvABAzAYdowFKWCgrxEOVXnXXa1guscGvcyha48QtV49Y3TrXBYG3Vom7Cst9UjTYOd2GDr2ywTY2XXtWhbHxlPO5j1bml_Xz_COa1M3Vh8PxxeowOSuWCOdnNEXq-uX6a3EWz-9vp5HIWFZTHJJJ5zHVBE80LySGJDQEtS8HzpDSCJVxRClIQELHWJc9jkBJ6yJRSM0YlZSN0vvX2r_fXwzKrbCiMc6o2TRcyRohI40RI-YcWvgnBmzJrva2UX2cEsk2jWdVm34326NnO2uWV0b_gT4U9EG2BN-vM-l9RNn_YCr8AghZ_tA</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3115968577</pqid></control><display><type>article</type><title>Building a pelvic organ prolapse diagnostic model using vision transformer on multi‐sequence MRI</title><source>Wiley</source><creator>Zhu, Shaojun ; Zhu, Xiaoxuan ; Zheng, Bo ; Wu, Maonian ; Li, Qiongshan ; Qian, Cheng</creator><creatorcontrib>Zhu, Shaojun ; Zhu, Xiaoxuan ; Zheng, Bo ; Wu, Maonian ; Li, Qiongshan ; Qian, Cheng</creatorcontrib><description>Background
Although the uterus, bladder, and rectum are distinct organs, their muscular fasciae are often interconnected. Clinical experience suggests that they may share common risk factors and associations. When one organ experiences prolapse, it can potentially affect the neighboring organs. However, the current assessment of disease severity still relies on manual measurements, which can yield varying results depending on the physician, thereby leading to diagnostic inaccuracies.
Purpose
This study aims to develop a multilabel grading model based on deep learning to classify the degree of prolapse of three organs in the female pelvis using stress magnetic resonance imaging (MRI) and provide interpretable result analysis.
Methods
We utilized sagittal MRI sequences taken at rest and during maximum Valsalva maneuver from 662 subjects. The training set included 464 subjects, the validation set included 98 subjects, and the test set included 100 subjects (training set n = 464, validation set n = 98, test set n = 100). We designed a feature extraction module specifically for pelvic floor MRI using the vision transformer architecture and employed label masking training strategy and pre‐training methods to enhance model convergence. The grading results were evaluated using Precision, Kappa, Recall, and Area Under the Curve (AUC). To validate the effectiveness of the model, the designed model was compared with classic grading methods. Finally, we provided interpretability charts illustrating the model's operational principles on the grading task.
Results
In terms of POP grading detection, the model achieved an average Precision, Kappa coefficient, Recall, and AUC of 0.86, 0.77, 0.76, and 0.86, respectively. Compared to existing studies, our model achieved the highest performance metrics. The average time taken to diagnose a patient was 0.38 s.
Conclusions
The proposed model achieved detection accuracy that is comparable to or even exceeds that of physicians, demonstrating the effectiveness of the vision transformer architecture and label masking training strategy for assisting in the grading of POP under static and maximum Valsalva conditions. This offers a promising option for computer‐aided diagnosis and treatment planning of POP.</description><identifier>ISSN: 0094-2405</identifier><identifier>ISSN: 2473-4209</identifier><identifier>EISSN: 2473-4209</identifier><identifier>DOI: 10.1002/mp.17441</identifier><identifier>PMID: 39395206</identifier><language>eng</language><publisher>United States</publisher><subject>deep learning ; magnetic resonance imaging ; pelvic organ prolapse ; vision transformer</subject><ispartof>Medical physics (Lancaster), 2025-01, Vol.52 (1), p.553-564</ispartof><rights>2024 American Association of Physicists in Medicine.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c2461-7b64dc28d4c74086e10d7f54b8fe5384a220751056ddf4b607706e1ef7d332723</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27915,27916</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/39395206$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Zhu, Shaojun</creatorcontrib><creatorcontrib>Zhu, Xiaoxuan</creatorcontrib><creatorcontrib>Zheng, Bo</creatorcontrib><creatorcontrib>Wu, Maonian</creatorcontrib><creatorcontrib>Li, Qiongshan</creatorcontrib><creatorcontrib>Qian, Cheng</creatorcontrib><title>Building a pelvic organ prolapse diagnostic model using vision transformer on multi‐sequence MRI</title><title>Medical physics (Lancaster)</title><addtitle>Med Phys</addtitle><description>Background
Although the uterus, bladder, and rectum are distinct organs, their muscular fasciae are often interconnected. Clinical experience suggests that they may share common risk factors and associations. When one organ experiences prolapse, it can potentially affect the neighboring organs. However, the current assessment of disease severity still relies on manual measurements, which can yield varying results depending on the physician, thereby leading to diagnostic inaccuracies.
Purpose
This study aims to develop a multilabel grading model based on deep learning to classify the degree of prolapse of three organs in the female pelvis using stress magnetic resonance imaging (MRI) and provide interpretable result analysis.
Methods
We utilized sagittal MRI sequences taken at rest and during maximum Valsalva maneuver from 662 subjects. The training set included 464 subjects, the validation set included 98 subjects, and the test set included 100 subjects (training set n = 464, validation set n = 98, test set n = 100). We designed a feature extraction module specifically for pelvic floor MRI using the vision transformer architecture and employed label masking training strategy and pre‐training methods to enhance model convergence. The grading results were evaluated using Precision, Kappa, Recall, and Area Under the Curve (AUC). To validate the effectiveness of the model, the designed model was compared with classic grading methods. Finally, we provided interpretability charts illustrating the model's operational principles on the grading task.
Results
In terms of POP grading detection, the model achieved an average Precision, Kappa coefficient, Recall, and AUC of 0.86, 0.77, 0.76, and 0.86, respectively. Compared to existing studies, our model achieved the highest performance metrics. The average time taken to diagnose a patient was 0.38 s.
Conclusions
The proposed model achieved detection accuracy that is comparable to or even exceeds that of physicians, demonstrating the effectiveness of the vision transformer architecture and label masking training strategy for assisting in the grading of POP under static and maximum Valsalva conditions. This offers a promising option for computer‐aided diagnosis and treatment planning of POP.</description><subject>deep learning</subject><subject>magnetic resonance imaging</subject><subject>pelvic organ prolapse</subject><subject>vision transformer</subject><issn>0094-2405</issn><issn>2473-4209</issn><issn>2473-4209</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2025</creationdate><recordtype>article</recordtype><recordid>eNp1kMtOwzAQRS0EoqUg8QXISzYp41ecLKHiUakVCME6cmKnMnIe2E1Rd3wC38iXkNICK1ajqzk6M7oInRIYEwB6UbVjIjkne2hIuWQRp5DuoyFAyiPKQQzQUQgvABAzAYdowFKWCgrxEOVXnXXa1guscGvcyha48QtV49Y3TrXBYG3Vom7Cst9UjTYOd2GDr2ywTY2XXtWhbHxlPO5j1bml_Xz_COa1M3Vh8PxxeowOSuWCOdnNEXq-uX6a3EWz-9vp5HIWFZTHJJJ5zHVBE80LySGJDQEtS8HzpDSCJVxRClIQELHWJc9jkBJ6yJRSM0YlZSN0vvX2r_fXwzKrbCiMc6o2TRcyRohI40RI-YcWvgnBmzJrva2UX2cEsk2jWdVm34326NnO2uWV0b_gT4U9EG2BN-vM-l9RNn_YCr8AghZ_tA</recordid><startdate>202501</startdate><enddate>202501</enddate><creator>Zhu, Shaojun</creator><creator>Zhu, Xiaoxuan</creator><creator>Zheng, Bo</creator><creator>Wu, Maonian</creator><creator>Li, Qiongshan</creator><creator>Qian, Cheng</creator><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope></search><sort><creationdate>202501</creationdate><title>Building a pelvic organ prolapse diagnostic model using vision transformer on multi‐sequence MRI</title><author>Zhu, Shaojun ; Zhu, Xiaoxuan ; Zheng, Bo ; Wu, Maonian ; Li, Qiongshan ; Qian, Cheng</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c2461-7b64dc28d4c74086e10d7f54b8fe5384a220751056ddf4b607706e1ef7d332723</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2025</creationdate><topic>deep learning</topic><topic>magnetic resonance imaging</topic><topic>pelvic organ prolapse</topic><topic>vision transformer</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhu, Shaojun</creatorcontrib><creatorcontrib>Zhu, Xiaoxuan</creatorcontrib><creatorcontrib>Zheng, Bo</creatorcontrib><creatorcontrib>Wu, Maonian</creatorcontrib><creatorcontrib>Li, Qiongshan</creatorcontrib><creatorcontrib>Qian, Cheng</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>Medical physics (Lancaster)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhu, Shaojun</au><au>Zhu, Xiaoxuan</au><au>Zheng, Bo</au><au>Wu, Maonian</au><au>Li, Qiongshan</au><au>Qian, Cheng</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Building a pelvic organ prolapse diagnostic model using vision transformer on multi‐sequence MRI</atitle><jtitle>Medical physics (Lancaster)</jtitle><addtitle>Med Phys</addtitle><date>2025-01</date><risdate>2025</risdate><volume>52</volume><issue>1</issue><spage>553</spage><epage>564</epage><pages>553-564</pages><issn>0094-2405</issn><issn>2473-4209</issn><eissn>2473-4209</eissn><abstract>Background
Although the uterus, bladder, and rectum are distinct organs, their muscular fasciae are often interconnected. Clinical experience suggests that they may share common risk factors and associations. When one organ experiences prolapse, it can potentially affect the neighboring organs. However, the current assessment of disease severity still relies on manual measurements, which can yield varying results depending on the physician, thereby leading to diagnostic inaccuracies.
Purpose
This study aims to develop a multilabel grading model based on deep learning to classify the degree of prolapse of three organs in the female pelvis using stress magnetic resonance imaging (MRI) and provide interpretable result analysis.
Methods
We utilized sagittal MRI sequences taken at rest and during maximum Valsalva maneuver from 662 subjects. The training set included 464 subjects, the validation set included 98 subjects, and the test set included 100 subjects (training set n = 464, validation set n = 98, test set n = 100). We designed a feature extraction module specifically for pelvic floor MRI using the vision transformer architecture and employed label masking training strategy and pre‐training methods to enhance model convergence. The grading results were evaluated using Precision, Kappa, Recall, and Area Under the Curve (AUC). To validate the effectiveness of the model, the designed model was compared with classic grading methods. Finally, we provided interpretability charts illustrating the model's operational principles on the grading task.
Results
In terms of POP grading detection, the model achieved an average Precision, Kappa coefficient, Recall, and AUC of 0.86, 0.77, 0.76, and 0.86, respectively. Compared to existing studies, our model achieved the highest performance metrics. The average time taken to diagnose a patient was 0.38 s.
Conclusions
The proposed model achieved detection accuracy that is comparable to or even exceeds that of physicians, demonstrating the effectiveness of the vision transformer architecture and label masking training strategy for assisting in the grading of POP under static and maximum Valsalva conditions. This offers a promising option for computer‐aided diagnosis and treatment planning of POP.</abstract><cop>United States</cop><pmid>39395206</pmid><doi>10.1002/mp.17441</doi><tpages>12</tpages><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0094-2405 |
ispartof | Medical physics (Lancaster), 2025-01, Vol.52 (1), p.553-564 |
issn | 0094-2405 2473-4209 2473-4209 |
language | eng |
recordid | cdi_proquest_miscellaneous_3115968577 |
source | Wiley |
subjects | deep learning magnetic resonance imaging pelvic organ prolapse vision transformer |
title | Building a pelvic organ prolapse diagnostic model using vision transformer on multi‐sequence MRI |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-14T23%3A26%3A07IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Building%20a%20pelvic%20organ%20prolapse%20diagnostic%20model%20using%20vision%20transformer%20on%20multi%E2%80%90sequence%20MRI&rft.jtitle=Medical%20physics%20(Lancaster)&rft.au=Zhu,%20Shaojun&rft.date=2025-01&rft.volume=52&rft.issue=1&rft.spage=553&rft.epage=564&rft.pages=553-564&rft.issn=0094-2405&rft.eissn=2473-4209&rft_id=info:doi/10.1002/mp.17441&rft_dat=%3Cproquest_cross%3E3115968577%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c2461-7b64dc28d4c74086e10d7f54b8fe5384a220751056ddf4b607706e1ef7d332723%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3115968577&rft_id=info:pmid/39395206&rfr_iscdi=true |