Loading…
Beyond the Stethoscope: Operationalizing Interactive Clinical Reasoning in Large Language Models via Proactive Information Seeking
Assistant agents powered by large language models (LLMs) are adept at answering questions, but they often struggle with asking questions to gather relevant information when required. In high-stakes domains such as medical reasoning, it is crucial to acquire all necessary information by proactively a...
Saved in:
Main Authors: | , , , , , |
---|---|
Format: | Conference Proceeding |
Language: | English |
Subjects: | |
Online Access: | Request full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Assistant agents powered by large language models (LLMs) are adept at answering questions, but they often struggle with asking questions to gather relevant information when required. In high-stakes domains such as medical reasoning, it is crucial to acquire all necessary information by proactively asking information-seeking questions before making a decision. In this paper, we introduce DYNAMED-a framework to simulate realistic, dynamic medical consultations between a Patient System and an Expert System. DYNAMED evaluates LLMs' ability to gather further details about patients' symptoms by asking information-seeking questions. We show that state-ofthe- art LLMs (LlaMA-2-Chat 7B & 13B, GPT-3.5 and GPT- 4) struggle at information-seeking-when given limited initial information and prompted to ask questions if needed, their accuracy drops by 10.1% on average compared to starting with the same limited information and asking no questions. Second, we attempt to improve the information-seeking capabilities of the expert model through self-reflection and medical expertise augmentation, and introduce a synthetic medical conversation dataset, MEDQACHAT. Our best model improves the accuracy of the GPT-3.5 Expert System by 19.0%; however, it still lags by 10.0% compared to when it is given full information upfront. While there is still a lot of work to be done to extend the information-seeking abilities of LLM assistants in critical domains, we aim to provide a flexible framework to facilitate further explorations in this direction. |
---|---|
ISSN: | 2575-2634 |
DOI: | 10.1109/ICHI61247.2024.00090 |