Loading…

Developing a machine learning model to detect diagnostic uncertainty in clinical documentation

Diagnostic uncertainty, when unrecognized or poorly communicated, can result in diagnostic error. However, diagnostic uncertainty is challenging to study due to a lack of validated identification methods. This study aims to identify distinct linguistic patterns associated with diagnostic uncertainty...

Full description

Saved in:
Bibliographic Details
Published in:Journal of hospital medicine 2023-05, Vol.18 (5), p.405-412
Main Authors: Marshall, Trisha L, Nickels, Lindsay C, Brady, Patrick W, Edgerton, Ezra J, Lee, James J, Hagedorn, Philip A
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c313t-424ffcb03de797e41d05d4c3036b7419c9ed9dfa4641a707072451e719007d4a3
cites cdi_FETCH-LOGICAL-c313t-424ffcb03de797e41d05d4c3036b7419c9ed9dfa4641a707072451e719007d4a3
container_end_page 412
container_issue 5
container_start_page 405
container_title Journal of hospital medicine
container_volume 18
creator Marshall, Trisha L
Nickels, Lindsay C
Brady, Patrick W
Edgerton, Ezra J
Lee, James J
Hagedorn, Philip A
description Diagnostic uncertainty, when unrecognized or poorly communicated, can result in diagnostic error. However, diagnostic uncertainty is challenging to study due to a lack of validated identification methods. This study aims to identify distinct linguistic patterns associated with diagnostic uncertainty in clinical documentation. This case-control study compares the clinical documentation of hospitalized children who received a novel uncertain diagnosis (UD) diagnosis label during their admission to a set of matched controls. Linguistic analyses identified potential linguistic indicators (i.e., words or phrases) of diagnostic uncertainty that were then manually reviewed by a linguist and clinical experts to identify those most relevant to diagnostic uncertainty. A natural language processing program categorized medical terminology into semantic types (i.e., sign or symptom), from which we identified a subset of these semantic types that both categorized reliably and were relevant to diagnostic uncertainty. Finally, a competitive machine learning modeling strategy utilizing the linguistic indicators and semantic types compared different predictive models for identifying diagnostic uncertainty. Our cohort included 242 UD-labeled patients and 932 matched controls with a combination of 3070 clinical notes. The best-performing model was a random forest, utilizing a combination of linguistic indicators and semantic types, yielding a sensitivity of 89.4% and a positive predictive value of 96.7%. Expert labeling, natural language processing, and machine learning methods combined with human validation resulted in highly predictive models to detect diagnostic uncertainty in clinical documentation and represent a promising approach to detecting, studying, and ultimately mitigating diagnostic uncertainty in clinical practice.
doi_str_mv 10.1002/jhm.13080
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_2787213517</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2787213517</sourcerecordid><originalsourceid>FETCH-LOGICAL-c313t-424ffcb03de797e41d05d4c3036b7419c9ed9dfa4641a707072451e719007d4a3</originalsourceid><addsrcrecordid>eNpdkE1LAzEQhoMotlYP_gEJeNFDa7JJNrtHqZ9Q8KJXlzSZbVOySd1khf57t7b1IHOY4eXhZXgQuqRkQgnJ7lbLZkIZKcgRGlIh2FjkJD8-3KLMBugsxhUhnBWCn6IBy0taFjkdos8H-AYX1tYvsMKN0kvrATtQrd9GTTDgcArYQAKdsLFq4UNMVuPOa2iTsj5tsPVYO-utVg6boLsGfFLJBn-OTmrlIlzs9wh9PD2-T1_Gs7fn1-n9bKwZZWnMM17Xek6YAVlK4NQQYbhmhOVzyWmpSzClqRXPOVWS9JNxQUHSkhBpuGIjdLPrXbfhq4OYqsZGDc4pD6GLVSYLmVEmqOzR63_oKnSt77-rsoIUgomC8p663VG6DTG2UFfr1jaq3VSUVFvpVS-9-pXes1f7xm7egPkjD5bZDyY-fGs</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2808535814</pqid></control><display><type>article</type><title>Developing a machine learning model to detect diagnostic uncertainty in clinical documentation</title><source>Wiley</source><creator>Marshall, Trisha L ; Nickels, Lindsay C ; Brady, Patrick W ; Edgerton, Ezra J ; Lee, James J ; Hagedorn, Philip A</creator><creatorcontrib>Marshall, Trisha L ; Nickels, Lindsay C ; Brady, Patrick W ; Edgerton, Ezra J ; Lee, James J ; Hagedorn, Philip A</creatorcontrib><description>Diagnostic uncertainty, when unrecognized or poorly communicated, can result in diagnostic error. However, diagnostic uncertainty is challenging to study due to a lack of validated identification methods. This study aims to identify distinct linguistic patterns associated with diagnostic uncertainty in clinical documentation. This case-control study compares the clinical documentation of hospitalized children who received a novel uncertain diagnosis (UD) diagnosis label during their admission to a set of matched controls. Linguistic analyses identified potential linguistic indicators (i.e., words or phrases) of diagnostic uncertainty that were then manually reviewed by a linguist and clinical experts to identify those most relevant to diagnostic uncertainty. A natural language processing program categorized medical terminology into semantic types (i.e., sign or symptom), from which we identified a subset of these semantic types that both categorized reliably and were relevant to diagnostic uncertainty. Finally, a competitive machine learning modeling strategy utilizing the linguistic indicators and semantic types compared different predictive models for identifying diagnostic uncertainty. Our cohort included 242 UD-labeled patients and 932 matched controls with a combination of 3070 clinical notes. The best-performing model was a random forest, utilizing a combination of linguistic indicators and semantic types, yielding a sensitivity of 89.4% and a positive predictive value of 96.7%. Expert labeling, natural language processing, and machine learning methods combined with human validation resulted in highly predictive models to detect diagnostic uncertainty in clinical documentation and represent a promising approach to detecting, studying, and ultimately mitigating diagnostic uncertainty in clinical practice.</description><identifier>ISSN: 1553-5592</identifier><identifier>EISSN: 1553-5606</identifier><identifier>DOI: 10.1002/jhm.13080</identifier><identifier>PMID: 36919861</identifier><language>eng</language><publisher>United States: Frontline Medical Communications</publisher><subject>Case-Control Studies ; Child ; Documentation ; Humans ; Linguistics ; Machine Learning ; Natural Language Processing ; Semantics ; Terminology ; Uncertainty</subject><ispartof>Journal of hospital medicine, 2023-05, Vol.18 (5), p.405-412</ispartof><rights>2023 Society of Hospital Medicine.</rights><rights>2023 Society of Hospital Medicine</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c313t-424ffcb03de797e41d05d4c3036b7419c9ed9dfa4641a707072451e719007d4a3</citedby><cites>FETCH-LOGICAL-c313t-424ffcb03de797e41d05d4c3036b7419c9ed9dfa4641a707072451e719007d4a3</cites><orcidid>0000-0003-1650-6359 ; 0000-0002-2492-3147</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27903,27904</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/36919861$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Marshall, Trisha L</creatorcontrib><creatorcontrib>Nickels, Lindsay C</creatorcontrib><creatorcontrib>Brady, Patrick W</creatorcontrib><creatorcontrib>Edgerton, Ezra J</creatorcontrib><creatorcontrib>Lee, James J</creatorcontrib><creatorcontrib>Hagedorn, Philip A</creatorcontrib><title>Developing a machine learning model to detect diagnostic uncertainty in clinical documentation</title><title>Journal of hospital medicine</title><addtitle>J Hosp Med</addtitle><description>Diagnostic uncertainty, when unrecognized or poorly communicated, can result in diagnostic error. However, diagnostic uncertainty is challenging to study due to a lack of validated identification methods. This study aims to identify distinct linguistic patterns associated with diagnostic uncertainty in clinical documentation. This case-control study compares the clinical documentation of hospitalized children who received a novel uncertain diagnosis (UD) diagnosis label during their admission to a set of matched controls. Linguistic analyses identified potential linguistic indicators (i.e., words or phrases) of diagnostic uncertainty that were then manually reviewed by a linguist and clinical experts to identify those most relevant to diagnostic uncertainty. A natural language processing program categorized medical terminology into semantic types (i.e., sign or symptom), from which we identified a subset of these semantic types that both categorized reliably and were relevant to diagnostic uncertainty. Finally, a competitive machine learning modeling strategy utilizing the linguistic indicators and semantic types compared different predictive models for identifying diagnostic uncertainty. Our cohort included 242 UD-labeled patients and 932 matched controls with a combination of 3070 clinical notes. The best-performing model was a random forest, utilizing a combination of linguistic indicators and semantic types, yielding a sensitivity of 89.4% and a positive predictive value of 96.7%. Expert labeling, natural language processing, and machine learning methods combined with human validation resulted in highly predictive models to detect diagnostic uncertainty in clinical documentation and represent a promising approach to detecting, studying, and ultimately mitigating diagnostic uncertainty in clinical practice.</description><subject>Case-Control Studies</subject><subject>Child</subject><subject>Documentation</subject><subject>Humans</subject><subject>Linguistics</subject><subject>Machine Learning</subject><subject>Natural Language Processing</subject><subject>Semantics</subject><subject>Terminology</subject><subject>Uncertainty</subject><issn>1553-5592</issn><issn>1553-5606</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNpdkE1LAzEQhoMotlYP_gEJeNFDa7JJNrtHqZ9Q8KJXlzSZbVOySd1khf57t7b1IHOY4eXhZXgQuqRkQgnJ7lbLZkIZKcgRGlIh2FjkJD8-3KLMBugsxhUhnBWCn6IBy0taFjkdos8H-AYX1tYvsMKN0kvrATtQrd9GTTDgcArYQAKdsLFq4UNMVuPOa2iTsj5tsPVYO-utVg6boLsGfFLJBn-OTmrlIlzs9wh9PD2-T1_Gs7fn1-n9bKwZZWnMM17Xek6YAVlK4NQQYbhmhOVzyWmpSzClqRXPOVWS9JNxQUHSkhBpuGIjdLPrXbfhq4OYqsZGDc4pD6GLVSYLmVEmqOzR63_oKnSt77-rsoIUgomC8p663VG6DTG2UFfr1jaq3VSUVFvpVS-9-pXes1f7xm7egPkjD5bZDyY-fGs</recordid><startdate>202305</startdate><enddate>202305</enddate><creator>Marshall, Trisha L</creator><creator>Nickels, Lindsay C</creator><creator>Brady, Patrick W</creator><creator>Edgerton, Ezra J</creator><creator>Lee, James J</creator><creator>Hagedorn, Philip A</creator><general>Frontline Medical Communications</general><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>K9.</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0003-1650-6359</orcidid><orcidid>https://orcid.org/0000-0002-2492-3147</orcidid></search><sort><creationdate>202305</creationdate><title>Developing a machine learning model to detect diagnostic uncertainty in clinical documentation</title><author>Marshall, Trisha L ; Nickels, Lindsay C ; Brady, Patrick W ; Edgerton, Ezra J ; Lee, James J ; Hagedorn, Philip A</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c313t-424ffcb03de797e41d05d4c3036b7419c9ed9dfa4641a707072451e719007d4a3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Case-Control Studies</topic><topic>Child</topic><topic>Documentation</topic><topic>Humans</topic><topic>Linguistics</topic><topic>Machine Learning</topic><topic>Natural Language Processing</topic><topic>Semantics</topic><topic>Terminology</topic><topic>Uncertainty</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Marshall, Trisha L</creatorcontrib><creatorcontrib>Nickels, Lindsay C</creatorcontrib><creatorcontrib>Brady, Patrick W</creatorcontrib><creatorcontrib>Edgerton, Ezra J</creatorcontrib><creatorcontrib>Lee, James J</creatorcontrib><creatorcontrib>Hagedorn, Philip A</creatorcontrib><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>ProQuest Health &amp; Medical Complete (Alumni)</collection><collection>MEDLINE - Academic</collection><jtitle>Journal of hospital medicine</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Marshall, Trisha L</au><au>Nickels, Lindsay C</au><au>Brady, Patrick W</au><au>Edgerton, Ezra J</au><au>Lee, James J</au><au>Hagedorn, Philip A</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Developing a machine learning model to detect diagnostic uncertainty in clinical documentation</atitle><jtitle>Journal of hospital medicine</jtitle><addtitle>J Hosp Med</addtitle><date>2023-05</date><risdate>2023</risdate><volume>18</volume><issue>5</issue><spage>405</spage><epage>412</epage><pages>405-412</pages><issn>1553-5592</issn><eissn>1553-5606</eissn><abstract>Diagnostic uncertainty, when unrecognized or poorly communicated, can result in diagnostic error. However, diagnostic uncertainty is challenging to study due to a lack of validated identification methods. This study aims to identify distinct linguistic patterns associated with diagnostic uncertainty in clinical documentation. This case-control study compares the clinical documentation of hospitalized children who received a novel uncertain diagnosis (UD) diagnosis label during their admission to a set of matched controls. Linguistic analyses identified potential linguistic indicators (i.e., words or phrases) of diagnostic uncertainty that were then manually reviewed by a linguist and clinical experts to identify those most relevant to diagnostic uncertainty. A natural language processing program categorized medical terminology into semantic types (i.e., sign or symptom), from which we identified a subset of these semantic types that both categorized reliably and were relevant to diagnostic uncertainty. Finally, a competitive machine learning modeling strategy utilizing the linguistic indicators and semantic types compared different predictive models for identifying diagnostic uncertainty. Our cohort included 242 UD-labeled patients and 932 matched controls with a combination of 3070 clinical notes. The best-performing model was a random forest, utilizing a combination of linguistic indicators and semantic types, yielding a sensitivity of 89.4% and a positive predictive value of 96.7%. Expert labeling, natural language processing, and machine learning methods combined with human validation resulted in highly predictive models to detect diagnostic uncertainty in clinical documentation and represent a promising approach to detecting, studying, and ultimately mitigating diagnostic uncertainty in clinical practice.</abstract><cop>United States</cop><pub>Frontline Medical Communications</pub><pmid>36919861</pmid><doi>10.1002/jhm.13080</doi><tpages>8</tpages><orcidid>https://orcid.org/0000-0003-1650-6359</orcidid><orcidid>https://orcid.org/0000-0002-2492-3147</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 1553-5592
ispartof Journal of hospital medicine, 2023-05, Vol.18 (5), p.405-412
issn 1553-5592
1553-5606
language eng
recordid cdi_proquest_miscellaneous_2787213517
source Wiley
subjects Case-Control Studies
Child
Documentation
Humans
Linguistics
Machine Learning
Natural Language Processing
Semantics
Terminology
Uncertainty
title Developing a machine learning model to detect diagnostic uncertainty in clinical documentation
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-23T00%3A48%3A35IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Developing%20a%20machine%20learning%20model%20to%20detect%20diagnostic%20uncertainty%20in%20clinical%20documentation&rft.jtitle=Journal%20of%20hospital%20medicine&rft.au=Marshall,%20Trisha%20L&rft.date=2023-05&rft.volume=18&rft.issue=5&rft.spage=405&rft.epage=412&rft.pages=405-412&rft.issn=1553-5592&rft.eissn=1553-5606&rft_id=info:doi/10.1002/jhm.13080&rft_dat=%3Cproquest_cross%3E2787213517%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c313t-424ffcb03de797e41d05d4c3036b7419c9ed9dfa4641a707072451e719007d4a3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2808535814&rft_id=info:pmid/36919861&rfr_iscdi=true