Deepfake X-rays are so real even doctors can’t tell the difference
Deepfake X-rays created by AI are now convincing enough to fool both doctors and AI models. In tests, radiologists had limited success identifying fake images, especially when they didn’t know they were being shown. This opens the door to risks like fraudulent medical claims and tampered diagnoses. Experts say stronger safeguards and detection tools are critical as the technology advances.
A new study published on March 24 in Radiology, the journal of the Radiological Society of North America (RSNA), shows that both radiologists and multimodal large language models (LLMs) have difficulty telling real X-rays apart from artificial intelligence (AI)-generated "deepfake" images. The findings raise concerns about the risks posed by synthetic medical images and highlight the need for better tools and training to help protect the accuracy of medical imaging and prepare health care professionals to recognize deepfakes.
A "deepfake" is a video, photo, image or audio file that appears authentic but has been created or altered using AI.
"Our study demonstrates that these deepfake X-rays are realistic enough to deceive radiologists, the most highly trained medical image specialists, even when they were aware that AI-generated images were present," said lead study author Mickael Tordjman, M.D., post-doctoral fellow, Icahn School of Medicine at Mount Sinai, New York. "This creates a high-stakes vulnerability for fraudulent litigation if, for example, a fabricated fracture could be indistinguishable from a real one. There is also a significant cybersecurity risk if hackers were to gain access to a hospital's network and inject synthetic images to manipulate patient diagnoses or cause widespread clinical chaos by undermining the fundamental reliability of the digital medical record."
Study Details and Image Testing
The research included 17 radiologists from 12 institutions across six countries (United States, France, Germany, Turkey, United Kingdom and United Arab Emirates). Their experience ranged from beginners to specialists with up to 40 years in practice. In total, the study examined 264 X-ray images, split evenly between real scans and AI-generated ones.
Participants reviewed two separate image sets with no overlap. One set contained a mix of real images and ChatGPT-generated X-rays from different parts of the body. The second set focused on chest X-rays, with half real and half created using RoentGen, an open-source generative AI diffusion model developed by Stanford Medicine researchers.
Detection Accuracy for Radiologists and AI
When radiologists were not told that fake images were included, only 41% recognized the AI-generated X-rays after evaluating their technical quality. Once they were informed that synthetic images were present, their average accuracy in distinguishing real from fake rose to 75%.
Performance varied widely among individuals. Radiologists correctly identified between 58% and 92% of the ChatGPT-generated images. AI systems showed similar limitations. Four multimodal LLMs -- GPT-4o (OpenAI), GPT-5 (OpenAI), Gemini 2.5 Pro (Google), and Llama 4 Maverick (Meta) -- achieved accuracy rates ranging from 57% to 85%. Even ChatGPT-4o, which was used to generate the deepfake images, did not detect all of them, though it performed better than the other models.
For the RoentGen-generated chest X-rays, radiologists achieved accuracy rates between 62% and 78%, while the AI models ranged from 52% to 89%.
Experience Does Not Guarantee Detection
The study found no link between a radiologist's years of experience and their ability to identify fake X-rays. However, musculoskeletal radiologists performed significantly better than other subspecialists.
Visual Clues in Deepfake X-Rays
Researchers identified several patterns that can appear in synthetic images.
"Deepfake medical images often look too perfect," Dr. Tordjman said. "Bones are overly smooth, spines unnaturally straight, lungs overly symmetrical, blood vessel patterns excessively uniform, and fractures appear unusually clean and consistent, often limited to one side of the bone."
Risks and Safeguards for Medical Imaging
The results highlight serious risks if deepfake X-rays are misused. Fabricated images could be used in legal cases or inserted into hospital systems to influence diagnoses and disrupt care.
To reduce these threats, researchers recommend stronger digital protections. These include invisible watermarks embedded directly into images and cryptographic signatures linked to the technologist at the time of image capture, which can help verify authenticity.
The Future of AI in Medical Imaging
"We are potentially only seeing the tip of the iceberg," Dr. Tordjman said. "The logical next step in this evolution is AI-generation of synthetic 3D images, such as CT and MRI. Establishing educational datasets and detection tools now is critical."
To support education and awareness, the researchers have released a curated deepfake dataset that includes interactive quizzes for training purposes.
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modelHow CoinFello's MinChi Park Built the Trust Layer 500 Million Crypto Users Have Been Waiting For
CoinFello launched publicly at EthCC 2026 with an AI agent that executes DeFi transactions through natural language while keeping private keys on the user's device. The security model uses ERC-7710 scoped delegations — users grant the agent a limited spending permission rather than wallet access, and can revoke it with one action. ETHDenver alpha surfaced two surprises: multilingual demand the team had not anticipated, and developer demand to use CoinFello as an execution layer for third-party agents. The B2B infrastructure angle, enabling Claude Code, Windsurf, and OpenClaw agents to call CoinFello for onchain execution, is now a primary growth thesis alongside the consumer product. Read All
Escaping API Quotas: How I Built a Local 14B Multi-Agent Squad for 16GB VRAM (Qwen3.5 & DeepSeek-R1)
<p>I was building a complex web app prototype using a cloud-based AI IDE. Just as I was getting into the flow, I hit the dreaded wall: <strong>"429 Too Many Requests"</strong>. </p> <p>I was done dealing with subscription anxiety and 6-day quota limits. I wanted to offload the heavy cognitive work to my local machine. But there was a catch: my rig runs on an AMD Radeon RX 6800 with <strong>16GB of VRAM</strong>.</p> <p>Here is how I bypassed the cloud limits and built a fully functional local multi-agent system without melting my GPU.</p> <h3> The "Goldilocks" Zone: Why 14B? </h3> <p>Running a multi-agent system locally is tricky when you have strict hardware limits. Through trial and error, I quickly realized:</p> <ul> <li> <strong>7B/8B models?</strong> They are fast, but too prone to ha
Mastering FastAPI: A Complete Learning Roadmap
<p>FastAPI is a powerful, modern Python web framework that leverages Python type hints and ASGI to build high-performance APIs. To truly master it, you need to understand not only the framework itself but also the ecosystem of tools, protocols, and deployment practices that surround it. Below is a comprehensive guide covering all the layers, concepts, and tools you’ll need to become proficient.</p> <h2> 1. Core FastAPI & Python Fundamentals </h2> <h3> Python (3.8+) </h3> <ul> <li>Type hints – essential for FastAPI’s data validation and editor support.</li> <li>Async/await – understand how async def works and when to use it (I/O-bound operations).</li> <li>Generators & context managers – used for dependencies and middleware.</li> <li>Packaging – pip, venv, poetry, uv for dependency
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in AI Tools

The Korg Handytraxx Play finally got me learning to scratch
About 20 years ago, I bought a turntable with the idea that it would be a staple of my music-making setup. I dreamed of learning to scratch and digging through dollar bin vinyl for samples. Instead, it just got hooked up to a stereo. But a few months ago, I got my hands on the […]
Anthropic's Claude Code source code got accidentally leaked - qz.com
<a href="https://news.google.com/rss/articles/CBMiYEFVX3lxTE82aXdNdDNscjVyMmRYWDExcEdLWEhzdmluY0R3cmVJUW5OZ2xpT0U5TC02MHRpTDlveVZJRlh3T0xBZ2tKRDQyTm9aS0dOLWNEaF9TSk94aWdpZV9vdEJHag?oc=5" target="_blank">Anthropic's Claude Code source code got accidentally leaked</a> <font color="#6f6f6f">qz.com</font>


Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!