Loading…

Clinical Prompt Learning With Frozen Language Models

When the first transformer-based language models were published in the late 2010s, pretraining with general text and then fine-tuning the model on a task-specific dataset often achieved the state-of-the-art performance. However, more recent work suggests that for some tasks, directly prompting the p...

Full description

Saved in:
Bibliographic Details
Published in:IEEE transaction on neural networks and learning systems 2024-11, Vol.35 (11), p.16453-16463
Main Authors: Taylor, Niall, Zhang, Yi, Joyce, Dan W., Gao, Ziming, Kormilitzin, Andrey, Nevado-Holgado, Alejo
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c324t-f3433cf7466f0495d7ea6874e5a62a0a4fde7175cd0b25e30524addf91a604e63
cites cdi_FETCH-LOGICAL-c324t-f3433cf7466f0495d7ea6874e5a62a0a4fde7175cd0b25e30524addf91a604e63
container_end_page 16463
container_issue 11
container_start_page 16453
container_title IEEE transaction on neural networks and learning systems
container_volume 35
creator Taylor, Niall
Zhang, Yi
Joyce, Dan W.
Gao, Ziming
Kormilitzin, Andrey
Nevado-Holgado, Alejo
description When the first transformer-based language models were published in the late 2010s, pretraining with general text and then fine-tuning the model on a task-specific dataset often achieved the state-of-the-art performance. However, more recent work suggests that for some tasks, directly prompting the pretrained model matches or surpasses fine-tuning in performance with few or no model parameter updates required. The use of prompts with language models for natural language processing (NLP) tasks is known as prompt learning. We investigated the viability of prompt learning on clinically meaningful decision tasks and directly compared this with more traditional fine-tuning methods. Results show that prompt learning methods were able to match or surpass the performance of traditional fine-tuning with up to 1000 times fewer trainable parameters, less training time, less training data, and lower computation resource requirements. We argue that these characteristics make prompt learning a very desirable alternative to traditional fine-tuning for clinical tasks, where the computational resources of public health providers are limited, and where data can often not be made available or not be used for fine-tuning due to patient privacy concerns. The complementary code to reproduce the experiments presented in this work can be found at https://github.com/NtaylorOX/Public_Clinical_Prompt .
doi_str_mv 10.1109/TNNLS.2023.3294633
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1109_TNNLS_2023_3294633</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10215061</ieee_id><sourcerecordid>2850311078</sourcerecordid><originalsourceid>FETCH-LOGICAL-c324t-f3433cf7466f0495d7ea6874e5a62a0a4fde7175cd0b25e30524addf91a604e63</originalsourceid><addsrcrecordid>eNpNkDtPwzAUhS0EolXpH0AIZWRJ8dvJiCoKSKEgUQSb5SbXxSiPYidD-fWktFTc5Z7hO2f4EDoneEIITq8X83n2MqGYsgmjKZeMHaEhJZLGlCXJ8SGr9wEah_CJ-5NYSJ6eogFTQvYpGSI-LV3tclNGz76p1m2UgfG1q1fRm2s_oplvvqGOMlOvOrOC6LEpoAxn6MSaMsB4_0fodXa7mN7H2dPdw_Qmi3NGeRtbxhnLreJSWsxTUSgwMlEchJHUYMNtAYookRd4SQUwLCg3RWFTYiTmINkIXe1217756iC0unIhh7I0NTRd0DQRmPU2VNKjdIfmvgnBg9Vr7yrjN5pgvRWmf4XprTC9F9aXLvf73bKC4lD509MDFzvAAcC_RUoEloT9AAOXbdY</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2850311078</pqid></control><display><type>article</type><title>Clinical Prompt Learning With Frozen Language Models</title><source>IEEE Electronic Library (IEL) Journals</source><creator>Taylor, Niall ; Zhang, Yi ; Joyce, Dan W. ; Gao, Ziming ; Kormilitzin, Andrey ; Nevado-Holgado, Alejo</creator><creatorcontrib>Taylor, Niall ; Zhang, Yi ; Joyce, Dan W. ; Gao, Ziming ; Kormilitzin, Andrey ; Nevado-Holgado, Alejo</creatorcontrib><description>When the first transformer-based language models were published in the late 2010s, pretraining with general text and then fine-tuning the model on a task-specific dataset often achieved the state-of-the-art performance. However, more recent work suggests that for some tasks, directly prompting the pretrained model matches or surpasses fine-tuning in performance with few or no model parameter updates required. The use of prompts with language models for natural language processing (NLP) tasks is known as prompt learning. We investigated the viability of prompt learning on clinically meaningful decision tasks and directly compared this with more traditional fine-tuning methods. Results show that prompt learning methods were able to match or surpass the performance of traditional fine-tuning with up to 1000 times fewer trainable parameters, less training time, less training data, and lower computation resource requirements. We argue that these characteristics make prompt learning a very desirable alternative to traditional fine-tuning for clinical tasks, where the computational resources of public health providers are limited, and where data can often not be made available or not be used for fine-tuning due to patient privacy concerns. The complementary code to reproduce the experiments presented in this work can be found at https://github.com/NtaylorOX/Public_Clinical_Prompt .</description><identifier>ISSN: 2162-237X</identifier><identifier>ISSN: 2162-2388</identifier><identifier>EISSN: 2162-2388</identifier><identifier>DOI: 10.1109/TNNLS.2023.3294633</identifier><identifier>PMID: 37566498</identifier><identifier>CODEN: ITNNAL</identifier><language>eng</language><publisher>United States: IEEE</publisher><subject>Adaptation models ; Algorithms ; Bit error rate ; Clinical decision support ; Computer architecture ; few-shot learning ; Humans ; Machine Learning ; Natural Language Processing ; Neural Networks, Computer ; pretrained language models (PLMs) ; prompt learning ; Task analysis ; Training ; Transformers ; Tuning</subject><ispartof>IEEE transaction on neural networks and learning systems, 2024-11, Vol.35 (11), p.16453-16463</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c324t-f3433cf7466f0495d7ea6874e5a62a0a4fde7175cd0b25e30524addf91a604e63</citedby><cites>FETCH-LOGICAL-c324t-f3433cf7466f0495d7ea6874e5a62a0a4fde7175cd0b25e30524addf91a604e63</cites><orcidid>0000-0003-0523-3877 ; 0000-0001-6682-334X ; 0000-0003-3555-9181</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10215061$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,27923,27924,54795</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/37566498$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Taylor, Niall</creatorcontrib><creatorcontrib>Zhang, Yi</creatorcontrib><creatorcontrib>Joyce, Dan W.</creatorcontrib><creatorcontrib>Gao, Ziming</creatorcontrib><creatorcontrib>Kormilitzin, Andrey</creatorcontrib><creatorcontrib>Nevado-Holgado, Alejo</creatorcontrib><title>Clinical Prompt Learning With Frozen Language Models</title><title>IEEE transaction on neural networks and learning systems</title><addtitle>TNNLS</addtitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><description>When the first transformer-based language models were published in the late 2010s, pretraining with general text and then fine-tuning the model on a task-specific dataset often achieved the state-of-the-art performance. However, more recent work suggests that for some tasks, directly prompting the pretrained model matches or surpasses fine-tuning in performance with few or no model parameter updates required. The use of prompts with language models for natural language processing (NLP) tasks is known as prompt learning. We investigated the viability of prompt learning on clinically meaningful decision tasks and directly compared this with more traditional fine-tuning methods. Results show that prompt learning methods were able to match or surpass the performance of traditional fine-tuning with up to 1000 times fewer trainable parameters, less training time, less training data, and lower computation resource requirements. We argue that these characteristics make prompt learning a very desirable alternative to traditional fine-tuning for clinical tasks, where the computational resources of public health providers are limited, and where data can often not be made available or not be used for fine-tuning due to patient privacy concerns. The complementary code to reproduce the experiments presented in this work can be found at https://github.com/NtaylorOX/Public_Clinical_Prompt .</description><subject>Adaptation models</subject><subject>Algorithms</subject><subject>Bit error rate</subject><subject>Clinical decision support</subject><subject>Computer architecture</subject><subject>few-shot learning</subject><subject>Humans</subject><subject>Machine Learning</subject><subject>Natural Language Processing</subject><subject>Neural Networks, Computer</subject><subject>pretrained language models (PLMs)</subject><subject>prompt learning</subject><subject>Task analysis</subject><subject>Training</subject><subject>Transformers</subject><subject>Tuning</subject><issn>2162-237X</issn><issn>2162-2388</issn><issn>2162-2388</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNpNkDtPwzAUhS0EolXpH0AIZWRJ8dvJiCoKSKEgUQSb5SbXxSiPYidD-fWktFTc5Z7hO2f4EDoneEIITq8X83n2MqGYsgmjKZeMHaEhJZLGlCXJ8SGr9wEah_CJ-5NYSJ6eogFTQvYpGSI-LV3tclNGz76p1m2UgfG1q1fRm2s_oplvvqGOMlOvOrOC6LEpoAxn6MSaMsB4_0fodXa7mN7H2dPdw_Qmi3NGeRtbxhnLreJSWsxTUSgwMlEchJHUYMNtAYookRd4SQUwLCg3RWFTYiTmINkIXe1217756iC0unIhh7I0NTRd0DQRmPU2VNKjdIfmvgnBg9Vr7yrjN5pgvRWmf4XprTC9F9aXLvf73bKC4lD509MDFzvAAcC_RUoEloT9AAOXbdY</recordid><startdate>202411</startdate><enddate>202411</enddate><creator>Taylor, Niall</creator><creator>Zhang, Yi</creator><creator>Joyce, Dan W.</creator><creator>Gao, Ziming</creator><creator>Kormilitzin, Andrey</creator><creator>Nevado-Holgado, Alejo</creator><general>IEEE</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0003-0523-3877</orcidid><orcidid>https://orcid.org/0000-0001-6682-334X</orcidid><orcidid>https://orcid.org/0000-0003-3555-9181</orcidid></search><sort><creationdate>202411</creationdate><title>Clinical Prompt Learning With Frozen Language Models</title><author>Taylor, Niall ; Zhang, Yi ; Joyce, Dan W. ; Gao, Ziming ; Kormilitzin, Andrey ; Nevado-Holgado, Alejo</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c324t-f3433cf7466f0495d7ea6874e5a62a0a4fde7175cd0b25e30524addf91a604e63</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Adaptation models</topic><topic>Algorithms</topic><topic>Bit error rate</topic><topic>Clinical decision support</topic><topic>Computer architecture</topic><topic>few-shot learning</topic><topic>Humans</topic><topic>Machine Learning</topic><topic>Natural Language Processing</topic><topic>Neural Networks, Computer</topic><topic>pretrained language models (PLMs)</topic><topic>prompt learning</topic><topic>Task analysis</topic><topic>Training</topic><topic>Transformers</topic><topic>Tuning</topic><toplevel>online_resources</toplevel><creatorcontrib>Taylor, Niall</creatorcontrib><creatorcontrib>Zhang, Yi</creatorcontrib><creatorcontrib>Joyce, Dan W.</creatorcontrib><creatorcontrib>Gao, Ziming</creatorcontrib><creatorcontrib>Kormilitzin, Andrey</creatorcontrib><creatorcontrib>Nevado-Holgado, Alejo</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library Online</collection><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE transaction on neural networks and learning systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Taylor, Niall</au><au>Zhang, Yi</au><au>Joyce, Dan W.</au><au>Gao, Ziming</au><au>Kormilitzin, Andrey</au><au>Nevado-Holgado, Alejo</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Clinical Prompt Learning With Frozen Language Models</atitle><jtitle>IEEE transaction on neural networks and learning systems</jtitle><stitle>TNNLS</stitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><date>2024-11</date><risdate>2024</risdate><volume>35</volume><issue>11</issue><spage>16453</spage><epage>16463</epage><pages>16453-16463</pages><issn>2162-237X</issn><issn>2162-2388</issn><eissn>2162-2388</eissn><coden>ITNNAL</coden><abstract>When the first transformer-based language models were published in the late 2010s, pretraining with general text and then fine-tuning the model on a task-specific dataset often achieved the state-of-the-art performance. However, more recent work suggests that for some tasks, directly prompting the pretrained model matches or surpasses fine-tuning in performance with few or no model parameter updates required. The use of prompts with language models for natural language processing (NLP) tasks is known as prompt learning. We investigated the viability of prompt learning on clinically meaningful decision tasks and directly compared this with more traditional fine-tuning methods. Results show that prompt learning methods were able to match or surpass the performance of traditional fine-tuning with up to 1000 times fewer trainable parameters, less training time, less training data, and lower computation resource requirements. We argue that these characteristics make prompt learning a very desirable alternative to traditional fine-tuning for clinical tasks, where the computational resources of public health providers are limited, and where data can often not be made available or not be used for fine-tuning due to patient privacy concerns. The complementary code to reproduce the experiments presented in this work can be found at https://github.com/NtaylorOX/Public_Clinical_Prompt .</abstract><cop>United States</cop><pub>IEEE</pub><pmid>37566498</pmid><doi>10.1109/TNNLS.2023.3294633</doi><tpages>11</tpages><orcidid>https://orcid.org/0000-0003-0523-3877</orcidid><orcidid>https://orcid.org/0000-0001-6682-334X</orcidid><orcidid>https://orcid.org/0000-0003-3555-9181</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 2162-237X
ispartof IEEE transaction on neural networks and learning systems, 2024-11, Vol.35 (11), p.16453-16463
issn 2162-237X
2162-2388
2162-2388
language eng
recordid cdi_crossref_primary_10_1109_TNNLS_2023_3294633
source IEEE Electronic Library (IEL) Journals
subjects Adaptation models
Algorithms
Bit error rate
Clinical decision support
Computer architecture
few-shot learning
Humans
Machine Learning
Natural Language Processing
Neural Networks, Computer
pretrained language models (PLMs)
prompt learning
Task analysis
Training
Transformers
Tuning
title Clinical Prompt Learning With Frozen Language Models
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-12T09%3A13%3A48IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Clinical%20Prompt%20Learning%20With%20Frozen%20Language%20Models&rft.jtitle=IEEE%20transaction%20on%20neural%20networks%20and%20learning%20systems&rft.au=Taylor,%20Niall&rft.date=2024-11&rft.volume=35&rft.issue=11&rft.spage=16453&rft.epage=16463&rft.pages=16453-16463&rft.issn=2162-237X&rft.eissn=2162-2388&rft.coden=ITNNAL&rft_id=info:doi/10.1109/TNNLS.2023.3294633&rft_dat=%3Cproquest_cross%3E2850311078%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c324t-f3433cf7466f0495d7ea6874e5a62a0a4fde7175cd0b25e30524addf91a604e63%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2850311078&rft_id=info:pmid/37566498&rft_ieee_id=10215061&rfr_iscdi=true