Large Language Models Are Clinical Reasoners: Reasoning-Aware Diagnosis Framework with Prompt-Generated Rationales
DOI:
https://doi.org/10.1609/aaai.v38i16.29802Keywords:
NLP: (Large) Language Models, NLP: Applications, NLP: Interpretability, Analysis, and Evaluation of NLP ModelsAbstract
Machine reasoning has made great progress in recent years owing to large language models (LLMs). In the clinical domain, however, most NLP-driven projects mainly focus on clinical classification or reading comprehension, and under-explore clinical reasoning for disease diagnosis due to the expensive rationale annotation with clinicians. In this work, we present a "reasoning-aware" diagnosis framework that rationalizes the diagnostic process via prompt-based learning in a time- and labor-efficient manner, and learns to reason over the prompt-generated rationales. Specifically, we address the clinical reasoning for disease diagnosis, where the LLM generates diagnostic rationales providing its insight on presented patient data and the reasoning path towards the diagnosis, namely Clinical Chain-of-Thought (Clinical CoT). We empirically demonstrate LLMs/LMs' ability of clinical reasoning via extensive experiments and analyses on both rationale generation and disease diagnosis in various settings. We further propose a novel set of criteria for evaluating machine-generated rationales' potential for real-world clinical settings, facilitating and benefiting future research in this area.Downloads
Published
2024-03-24
How to Cite
Kwon, T., Ong, K. T.- iunn, Kang, D., Moon, S., Lee, J. R., Hwang, D., Sohn, B., Sim, Y., Lee, D., & Yeo, J. (2024). Large Language Models Are Clinical Reasoners: Reasoning-Aware Diagnosis Framework with Prompt-Generated Rationales. Proceedings of the AAAI Conference on Artificial Intelligence, 38(16), 18417-18425. https://doi.org/10.1609/aaai.v38i16.29802
Issue
Section
AAAI Technical Track on Natural Language Processing I