View a PDF of the paper titled The Illusion of Readiness in Health AI, by Yu Gu and 31 other authors
View PDF
HTML (experimental)
Abstract:Large language models have demonstrated remarkable performance in a wide range of medical benchmarks. Yet underneath the seemingly promising results lie salient growth areas, especially in cutting-edge frontiers such as multimodal reasoning. In this paper, we introduce a series of adversarial stress tests to systematically assess the robustness of flagship models and medical benchmarks. Our study reveals prevalent brittleness in the presence of simple adversarial transformations: leading systems can guess the right answer even with key inputs removed, yet may get confused by the slightest prompt alterations, while fabricating convincing yet flawed reasoning traces. Using clinician-guided rubrics, we demonstrate that popular medical benchmarks vary widely in what they truly measure. Our study reveals significant competency gaps of frontier AI in attaining real-world readiness for health applications. If we want AI to earn trust in healthcare, we must demand more than leaderboard wins and must hold AI systems accountable to ensure robustness, sound reasoning, and alignment with real medical demands.
Submission history
From: Yu Gu [view email]
[v1]
Mon, 22 Sep 2025 17:48:05 UTC (17,963 KB)
[v2]
Wed, 1 Oct 2025 17:21:09 UTC (16,928 KB)
[v3]
Thu, 11 Dec 2025 20:55:53 UTC (19,298 KB)


![[2509.18234] The Illusion of Readiness in Health AI Measuring Intelligence Efficiency of Local AI](https://skytik.cc/wp-content/uploads/2025/11/Measuring-Intelligence-Efficiency-of-Local-AI-768x448.png)