Oral Poster
MDAgents: An Adaptive Collaboration of LLMs for Medical Decision Making
Yubin Kim · Chanwoo Park · Hyewon Jeong · Yik Siu Chan · Xuhai "Orson" Xu · Daniel McDuff · Hyeonhoon Lee · Marzyeh Ghassemi · Cynthia Breazeal · Hae Park
East Exhibit Hall A-C #1007
[
Abstract
]
[ Project Page ]
Oral
presentation:
Oral Session 6A: Machine Learning and Science, Safety
Fri 13 Dec 3:30 p.m. PST — 4:30 p.m. PST
Fri 13 Dec 4:30 p.m. PST
— 7:30 p.m. PST
Fri 13 Dec 3:30 p.m. PST — 4:30 p.m. PST
Abstract:
Foundation models are becoming valuable tools in medicine. Yet despite their promise, the best way to leverage Large Language Models (LLMs) in complex medical tasks remains an open question. We introduce a novel multi-actor framework, named **M**edical **D**ecision-making **Agents** (**MDAgents**) that helps to address this gap by automatically assigning a collaboration structure to a team of LLMs. The assigned solo or group collaboration structure is tailored to the medical task at hand, a simple emulation inspired by the way real-world medical decision-making processes are adapted to tasks of different complexities. We evaluate our framework and baseline methods using state-of-the-art LLMs across a suite of real-world medical knowledge and clinical diagnosis benchmarks. MDAgents achieved the **best performance in seven out of ten** benchmarks on tasks requiring an understanding of medical knowledge and multi-modal reasoning, showing a significant **improvement of up to 6.5\%** ($p$ < 0.05) compared to previous methods' best performances. Ablation studies reveal that MDAgents effectively determines medical complexity to optimize for efficiency and accuracy across diverse medical tasks. Notably, the combination of moderator review and external medical knowledge in group collaboration resulted in an average accuracy improvement of 11.8\%. Our code can be found at https://anonymous.4open.science/r/MDAgents.
Live content is unavailable. Log in and register to view live content