Loading…
Developing a machine learning model to detect diagnostic uncertainty in clinical documentation
Diagnostic uncertainty, when unrecognized or poorly communicated, can result in diagnostic error. However, diagnostic uncertainty is challenging to study due to a lack of validated identification methods. This study aims to identify distinct linguistic patterns associated with diagnostic uncertainty...
Saved in:
Published in: | Journal of hospital medicine 2023-05, Vol.18 (5), p.405-412 |
---|---|
Main Authors: | , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c313t-424ffcb03de797e41d05d4c3036b7419c9ed9dfa4641a707072451e719007d4a3 |
---|---|
cites | cdi_FETCH-LOGICAL-c313t-424ffcb03de797e41d05d4c3036b7419c9ed9dfa4641a707072451e719007d4a3 |
container_end_page | 412 |
container_issue | 5 |
container_start_page | 405 |
container_title | Journal of hospital medicine |
container_volume | 18 |
creator | Marshall, Trisha L Nickels, Lindsay C Brady, Patrick W Edgerton, Ezra J Lee, James J Hagedorn, Philip A |
description | Diagnostic uncertainty, when unrecognized or poorly communicated, can result in diagnostic error. However, diagnostic uncertainty is challenging to study due to a lack of validated identification methods. This study aims to identify distinct linguistic patterns associated with diagnostic uncertainty in clinical documentation.
This case-control study compares the clinical documentation of hospitalized children who received a novel uncertain diagnosis (UD) diagnosis label during their admission to a set of matched controls. Linguistic analyses identified potential linguistic indicators (i.e., words or phrases) of diagnostic uncertainty that were then manually reviewed by a linguist and clinical experts to identify those most relevant to diagnostic uncertainty. A natural language processing program categorized medical terminology into semantic types (i.e., sign or symptom), from which we identified a subset of these semantic types that both categorized reliably and were relevant to diagnostic uncertainty. Finally, a competitive machine learning modeling strategy utilizing the linguistic indicators and semantic types compared different predictive models for identifying diagnostic uncertainty.
Our cohort included 242 UD-labeled patients and 932 matched controls with a combination of 3070 clinical notes. The best-performing model was a random forest, utilizing a combination of linguistic indicators and semantic types, yielding a sensitivity of 89.4% and a positive predictive value of 96.7%.
Expert labeling, natural language processing, and machine learning methods combined with human validation resulted in highly predictive models to detect diagnostic uncertainty in clinical documentation and represent a promising approach to detecting, studying, and ultimately mitigating diagnostic uncertainty in clinical practice. |
doi_str_mv | 10.1002/jhm.13080 |
format | article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_2787213517</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2787213517</sourcerecordid><originalsourceid>FETCH-LOGICAL-c313t-424ffcb03de797e41d05d4c3036b7419c9ed9dfa4641a707072451e719007d4a3</originalsourceid><addsrcrecordid>eNpdkE1LAzEQhoMotlYP_gEJeNFDa7JJNrtHqZ9Q8KJXlzSZbVOySd1khf57t7b1IHOY4eXhZXgQuqRkQgnJ7lbLZkIZKcgRGlIh2FjkJD8-3KLMBugsxhUhnBWCn6IBy0taFjkdos8H-AYX1tYvsMKN0kvrATtQrd9GTTDgcArYQAKdsLFq4UNMVuPOa2iTsj5tsPVYO-utVg6boLsGfFLJBn-OTmrlIlzs9wh9PD2-T1_Gs7fn1-n9bKwZZWnMM17Xek6YAVlK4NQQYbhmhOVzyWmpSzClqRXPOVWS9JNxQUHSkhBpuGIjdLPrXbfhq4OYqsZGDc4pD6GLVSYLmVEmqOzR63_oKnSt77-rsoIUgomC8p663VG6DTG2UFfr1jaq3VSUVFvpVS-9-pXes1f7xm7egPkjD5bZDyY-fGs</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2808535814</pqid></control><display><type>article</type><title>Developing a machine learning model to detect diagnostic uncertainty in clinical documentation</title><source>Wiley</source><creator>Marshall, Trisha L ; Nickels, Lindsay C ; Brady, Patrick W ; Edgerton, Ezra J ; Lee, James J ; Hagedorn, Philip A</creator><creatorcontrib>Marshall, Trisha L ; Nickels, Lindsay C ; Brady, Patrick W ; Edgerton, Ezra J ; Lee, James J ; Hagedorn, Philip A</creatorcontrib><description>Diagnostic uncertainty, when unrecognized or poorly communicated, can result in diagnostic error. However, diagnostic uncertainty is challenging to study due to a lack of validated identification methods. This study aims to identify distinct linguistic patterns associated with diagnostic uncertainty in clinical documentation.
This case-control study compares the clinical documentation of hospitalized children who received a novel uncertain diagnosis (UD) diagnosis label during their admission to a set of matched controls. Linguistic analyses identified potential linguistic indicators (i.e., words or phrases) of diagnostic uncertainty that were then manually reviewed by a linguist and clinical experts to identify those most relevant to diagnostic uncertainty. A natural language processing program categorized medical terminology into semantic types (i.e., sign or symptom), from which we identified a subset of these semantic types that both categorized reliably and were relevant to diagnostic uncertainty. Finally, a competitive machine learning modeling strategy utilizing the linguistic indicators and semantic types compared different predictive models for identifying diagnostic uncertainty.
Our cohort included 242 UD-labeled patients and 932 matched controls with a combination of 3070 clinical notes. The best-performing model was a random forest, utilizing a combination of linguistic indicators and semantic types, yielding a sensitivity of 89.4% and a positive predictive value of 96.7%.
Expert labeling, natural language processing, and machine learning methods combined with human validation resulted in highly predictive models to detect diagnostic uncertainty in clinical documentation and represent a promising approach to detecting, studying, and ultimately mitigating diagnostic uncertainty in clinical practice.</description><identifier>ISSN: 1553-5592</identifier><identifier>EISSN: 1553-5606</identifier><identifier>DOI: 10.1002/jhm.13080</identifier><identifier>PMID: 36919861</identifier><language>eng</language><publisher>United States: Frontline Medical Communications</publisher><subject>Case-Control Studies ; Child ; Documentation ; Humans ; Linguistics ; Machine Learning ; Natural Language Processing ; Semantics ; Terminology ; Uncertainty</subject><ispartof>Journal of hospital medicine, 2023-05, Vol.18 (5), p.405-412</ispartof><rights>2023 Society of Hospital Medicine.</rights><rights>2023 Society of Hospital Medicine</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c313t-424ffcb03de797e41d05d4c3036b7419c9ed9dfa4641a707072451e719007d4a3</citedby><cites>FETCH-LOGICAL-c313t-424ffcb03de797e41d05d4c3036b7419c9ed9dfa4641a707072451e719007d4a3</cites><orcidid>0000-0003-1650-6359 ; 0000-0002-2492-3147</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27903,27904</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/36919861$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Marshall, Trisha L</creatorcontrib><creatorcontrib>Nickels, Lindsay C</creatorcontrib><creatorcontrib>Brady, Patrick W</creatorcontrib><creatorcontrib>Edgerton, Ezra J</creatorcontrib><creatorcontrib>Lee, James J</creatorcontrib><creatorcontrib>Hagedorn, Philip A</creatorcontrib><title>Developing a machine learning model to detect diagnostic uncertainty in clinical documentation</title><title>Journal of hospital medicine</title><addtitle>J Hosp Med</addtitle><description>Diagnostic uncertainty, when unrecognized or poorly communicated, can result in diagnostic error. However, diagnostic uncertainty is challenging to study due to a lack of validated identification methods. This study aims to identify distinct linguistic patterns associated with diagnostic uncertainty in clinical documentation.
This case-control study compares the clinical documentation of hospitalized children who received a novel uncertain diagnosis (UD) diagnosis label during their admission to a set of matched controls. Linguistic analyses identified potential linguistic indicators (i.e., words or phrases) of diagnostic uncertainty that were then manually reviewed by a linguist and clinical experts to identify those most relevant to diagnostic uncertainty. A natural language processing program categorized medical terminology into semantic types (i.e., sign or symptom), from which we identified a subset of these semantic types that both categorized reliably and were relevant to diagnostic uncertainty. Finally, a competitive machine learning modeling strategy utilizing the linguistic indicators and semantic types compared different predictive models for identifying diagnostic uncertainty.
Our cohort included 242 UD-labeled patients and 932 matched controls with a combination of 3070 clinical notes. The best-performing model was a random forest, utilizing a combination of linguistic indicators and semantic types, yielding a sensitivity of 89.4% and a positive predictive value of 96.7%.
Expert labeling, natural language processing, and machine learning methods combined with human validation resulted in highly predictive models to detect diagnostic uncertainty in clinical documentation and represent a promising approach to detecting, studying, and ultimately mitigating diagnostic uncertainty in clinical practice.</description><subject>Case-Control Studies</subject><subject>Child</subject><subject>Documentation</subject><subject>Humans</subject><subject>Linguistics</subject><subject>Machine Learning</subject><subject>Natural Language Processing</subject><subject>Semantics</subject><subject>Terminology</subject><subject>Uncertainty</subject><issn>1553-5592</issn><issn>1553-5606</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNpdkE1LAzEQhoMotlYP_gEJeNFDa7JJNrtHqZ9Q8KJXlzSZbVOySd1khf57t7b1IHOY4eXhZXgQuqRkQgnJ7lbLZkIZKcgRGlIh2FjkJD8-3KLMBugsxhUhnBWCn6IBy0taFjkdos8H-AYX1tYvsMKN0kvrATtQrd9GTTDgcArYQAKdsLFq4UNMVuPOa2iTsj5tsPVYO-utVg6boLsGfFLJBn-OTmrlIlzs9wh9PD2-T1_Gs7fn1-n9bKwZZWnMM17Xek6YAVlK4NQQYbhmhOVzyWmpSzClqRXPOVWS9JNxQUHSkhBpuGIjdLPrXbfhq4OYqsZGDc4pD6GLVSYLmVEmqOzR63_oKnSt77-rsoIUgomC8p663VG6DTG2UFfr1jaq3VSUVFvpVS-9-pXes1f7xm7egPkjD5bZDyY-fGs</recordid><startdate>202305</startdate><enddate>202305</enddate><creator>Marshall, Trisha L</creator><creator>Nickels, Lindsay C</creator><creator>Brady, Patrick W</creator><creator>Edgerton, Ezra J</creator><creator>Lee, James J</creator><creator>Hagedorn, Philip A</creator><general>Frontline Medical Communications</general><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>K9.</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0003-1650-6359</orcidid><orcidid>https://orcid.org/0000-0002-2492-3147</orcidid></search><sort><creationdate>202305</creationdate><title>Developing a machine learning model to detect diagnostic uncertainty in clinical documentation</title><author>Marshall, Trisha L ; Nickels, Lindsay C ; Brady, Patrick W ; Edgerton, Ezra J ; Lee, James J ; Hagedorn, Philip A</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c313t-424ffcb03de797e41d05d4c3036b7419c9ed9dfa4641a707072451e719007d4a3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Case-Control Studies</topic><topic>Child</topic><topic>Documentation</topic><topic>Humans</topic><topic>Linguistics</topic><topic>Machine Learning</topic><topic>Natural Language Processing</topic><topic>Semantics</topic><topic>Terminology</topic><topic>Uncertainty</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Marshall, Trisha L</creatorcontrib><creatorcontrib>Nickels, Lindsay C</creatorcontrib><creatorcontrib>Brady, Patrick W</creatorcontrib><creatorcontrib>Edgerton, Ezra J</creatorcontrib><creatorcontrib>Lee, James J</creatorcontrib><creatorcontrib>Hagedorn, Philip A</creatorcontrib><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>ProQuest Health & Medical Complete (Alumni)</collection><collection>MEDLINE - Academic</collection><jtitle>Journal of hospital medicine</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Marshall, Trisha L</au><au>Nickels, Lindsay C</au><au>Brady, Patrick W</au><au>Edgerton, Ezra J</au><au>Lee, James J</au><au>Hagedorn, Philip A</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Developing a machine learning model to detect diagnostic uncertainty in clinical documentation</atitle><jtitle>Journal of hospital medicine</jtitle><addtitle>J Hosp Med</addtitle><date>2023-05</date><risdate>2023</risdate><volume>18</volume><issue>5</issue><spage>405</spage><epage>412</epage><pages>405-412</pages><issn>1553-5592</issn><eissn>1553-5606</eissn><abstract>Diagnostic uncertainty, when unrecognized or poorly communicated, can result in diagnostic error. However, diagnostic uncertainty is challenging to study due to a lack of validated identification methods. This study aims to identify distinct linguistic patterns associated with diagnostic uncertainty in clinical documentation.
This case-control study compares the clinical documentation of hospitalized children who received a novel uncertain diagnosis (UD) diagnosis label during their admission to a set of matched controls. Linguistic analyses identified potential linguistic indicators (i.e., words or phrases) of diagnostic uncertainty that were then manually reviewed by a linguist and clinical experts to identify those most relevant to diagnostic uncertainty. A natural language processing program categorized medical terminology into semantic types (i.e., sign or symptom), from which we identified a subset of these semantic types that both categorized reliably and were relevant to diagnostic uncertainty. Finally, a competitive machine learning modeling strategy utilizing the linguistic indicators and semantic types compared different predictive models for identifying diagnostic uncertainty.
Our cohort included 242 UD-labeled patients and 932 matched controls with a combination of 3070 clinical notes. The best-performing model was a random forest, utilizing a combination of linguistic indicators and semantic types, yielding a sensitivity of 89.4% and a positive predictive value of 96.7%.
Expert labeling, natural language processing, and machine learning methods combined with human validation resulted in highly predictive models to detect diagnostic uncertainty in clinical documentation and represent a promising approach to detecting, studying, and ultimately mitigating diagnostic uncertainty in clinical practice.</abstract><cop>United States</cop><pub>Frontline Medical Communications</pub><pmid>36919861</pmid><doi>10.1002/jhm.13080</doi><tpages>8</tpages><orcidid>https://orcid.org/0000-0003-1650-6359</orcidid><orcidid>https://orcid.org/0000-0002-2492-3147</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1553-5592 |
ispartof | Journal of hospital medicine, 2023-05, Vol.18 (5), p.405-412 |
issn | 1553-5592 1553-5606 |
language | eng |
recordid | cdi_proquest_miscellaneous_2787213517 |
source | Wiley |
subjects | Case-Control Studies Child Documentation Humans Linguistics Machine Learning Natural Language Processing Semantics Terminology Uncertainty |
title | Developing a machine learning model to detect diagnostic uncertainty in clinical documentation |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-23T00%3A48%3A35IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Developing%20a%20machine%20learning%20model%20to%20detect%20diagnostic%20uncertainty%20in%20clinical%20documentation&rft.jtitle=Journal%20of%20hospital%20medicine&rft.au=Marshall,%20Trisha%20L&rft.date=2023-05&rft.volume=18&rft.issue=5&rft.spage=405&rft.epage=412&rft.pages=405-412&rft.issn=1553-5592&rft.eissn=1553-5606&rft_id=info:doi/10.1002/jhm.13080&rft_dat=%3Cproquest_cross%3E2787213517%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c313t-424ffcb03de797e41d05d4c3036b7419c9ed9dfa4641a707072451e719007d4a3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2808535814&rft_id=info:pmid/36919861&rfr_iscdi=true |