AI”Doctors” are cheating medical school exams
dpa/picture alliance via Getty Images
The world’s most advanced artificial intelligence systems are essentially cheating their way through medical tests, achieving impressive scores not through genuine medical knowledge but by exploiting loopholes in how these tests are designed. This discovery has massive implications for the one-hundred billion medical AI industry and every patient who might encounter AI-powered healthcare.
The Medical AI Cheating Problem
Think of medical AI benchmarks like standardized tests that measure how well artificial intelligence systems understand medicine. Just as students take SATs to prove they’re ready for college, AI systems take these medical benchmarks to demonstrate they’re ready to help doctors diagnose diseases and recommend treatments.
But a recent groundbreaking study published by Microsoft Research reveals these AI systems aren’t actually learning medicine. They’re just getting really good at taking tests. It’s like discovering that a student achieved perfect SAT scores not by understanding math and reading, but by memorizing which answer choice tends to be correct most often.
Researchers put six top AI models through rigorous stress tests and found these systems achieve high medical scores through sophisticated test-taking tricks rather than real medical understanding.
How AI Systems Cheat The System
The research team discovered multiple ways AI systems fake medical competence, using methods that would almost assuredly get a human student expelled:
- When researchers simply rearranged the order of multiple choice answers, moving option A to option C for example, AI performance dropped significantly. This means the systems were learning “the answer is usually in position B” rather than “pneumonia causes these specific symptoms.”
- On questions that required analyzing medical images like X-rays or MRIs, AI systems still provided correct answers even when the images were completely removed. GPT-5, for instance, maintained 37.7% accuracy on visually-required questions even without any image, far above the 20% random chance level.
- AI systems figured out how to use clues in wrong answer choices to guess the right one, rather than applying real medical knowledge. Researchers found these models relied heavily on the wording of wrong answers, known as “distractors.” When those distractors were replaced with non-medical terms, the AI’s accuracy collapsed. This revealed it was leaning on test-taking tricks instead of genuine understanding.
Your Healthcare On AI
This research comes at a time when AI is rapidly expanding into healthcare. Eighty percent of hospitals now use AI to improve patient care and operational efficiency, with doctors increasingly relying on AI for everything from reading X-rays to suggesting treatments. Yet this study suggests current testing methods can’t distinguish between genuine medical competence and sophisticated test-taking algorithms.
The Microsoft Research study found that models like GPT-5 achieved 80.89% accuracy on medical image challenges but dropped to 67.56% when images were removed. This 13.33 percentage point decrease reveals hidden reliance on non-visual cues. Even more concerning, when researchers substituted medical images with ones supporting different diagnoses, model accuracy collapsed by more than thirty percentage points despite no change in the text questions.
Consider this scenario: An AI system achieves a 95% score on medical diagnosis tests and gets deployed in emergency rooms to help doctors quickly assess patients. But if that system achieved its high score through test-taking tricks rather than medical understanding, it might miss critical symptoms or recommend inappropriate treatments when faced with real patients whose conditions don’t match the patterns it learned from test questions.
The medical AI market is projected to exceed one-hundred billion by 2030, with healthcare systems worldwide investing heavily in AI diagnostic tools. Healthcare organizations purchasing AI systems based on impressive benchmark scores may unknowingly introduce significant patient safety risks. The Microsoft researchers warn that “medical benchmark scores do not directly reflect real-world readiness”.
The implications go beyond test scores. The Microsoft study revealed that when AI models were asked to explain their medical reasoning, they often generated “convincing yet flawed reasoning” or provided “correct answers supported by fabricated reasoning”. One example showed a model correctly diagnosing dermatomyositis while describing visual features that weren’t present in the image, since no image was provided at all.
Even as AI adoption accelerates, Medicine’s rapid adoption of AI has researchers concerned, with experts warning that hospitals and universities must step up to fill gaps in regulation.
The AI Pattern Recognition Problem
Unlike human medical students who learn by understanding how diseases affect the human body, current AI systems learn by finding patterns in data. This creates what the Microsoft researchers call “shortcut learning,” finding the easiest path to the right answer without developing genuine understanding.
The study found that AI models might diagnose pneumonia not by interpreting radiologic features, but by learning that “productive cough” plus “fever” statistically co-occurs with pneumonia in training data. This is pattern matching, not medical understanding.
Recent research from Nature highlights similar concerns, showing that trust in AI-assisted health systems remains problematic when these systems fail to demonstrate genuine understanding of medical contexts.
Moving Forward With Medical AI
The Microsoft researchers advocate for rethinking how we test medical AI systems. Instead of relying on benchmark scores, we need evaluation methods that can detect when AI systems are gaming tests rather than learning medicine.
The medical AI industry faces a critical moment. The Microsoft Research findings reveal that impressive benchmark scores have created an illusion of readiness that could have serious consequences for patient safety. As AI continues expanding into healthcare, our methods for verifying these systems must evolve to match their sophistication and their potential for sophisticated failure.
Source: https://www.forbes.com/sites/larsdaniel/2025/10/03/ai-doctors-cheat-medical-tests/