DEV Community

CaraComp
CaraComp

Posted on • Originally published at go.caracomp.com

Radiologists Miss 59% of Fake X-Rays on First Look — What That Proves About Your Case Photos

AI-generated radiographs fool medical experts

For developers building computer vision (CV) or biometric systems, the 2026 Radiological Society of North America (RSNA) study is a high-stakes warning about the "trust but verify" paradigm. When trained radiologists miss 59% of synthetic X-rays, the technical implication is clear: visual authenticity is no longer a human-solvable problem. For those of us writing the code that processes, compares, and validates image data, we have to stop treating "human-in-the-loop" as a foolproof safety net and start treating it as a vulnerability.

The technical core of the issue lies in what the researchers call "adversarial perfection." In the study, ChatGPT-generated X-rays were flagged as fake only 41% of the time because they looked "too right"—featuring unnaturally straight spines and perfectly uniform vascular patterns. In the world of facial comparison, we see the same phenomenon. Generative models often produce feature vectors that lack the organic noise and asymmetry found in real human biology.

From Visual Inspection to Euclidean Analysis

For developers working with facial comparison technology, this news reinforces why we must move away from subjective visual evaluation toward rigorous geometric measurement. When we analyze a face, we aren't just "looking" at it; we are mapping 68 or more landmark points and calculating the Euclidean distance between them to generate a reproducible feature vector.

If a developer is relying on a standard compare(img1, img2) API call, they might get a high similarity score, but without analyzing the underlying landmark distribution, they might miss a synthetic forgery. Authentic human faces have specific, messy geometric ratios. AI-generated faces often over-optimize for symmetry. To combat this, our pipelines should be incorporating checks for "geometric variance"—essentially looking for that "too perfect" signature that the radiologists missed.

The Shift from Recognition to Comparison

This study also highlights a critical distinction in our field: the difference between facial recognition and facial comparison. In investigative contexts, facial COMPARISON is the more robust methodology because it allows for the structured, side-by-side analysis of YOUR case photos rather than broad, automated surveillance.

By focusing on facial comparison, developers can implement more granular verification layers:

  • Landmark coordinate consistency across batch uploads.
  • Metadata validation tied to the biometric hash.
  • Euclidean distance reporting that highlights specific geometric anomalies.
  • Court-ready documentation that turns a "match" into a defensible technical report.

The Adversarial Gap

As AI generation tools like GPT-5 and Llama 4 Maverick continue to lower the barrier for creating convincing "fake reality," the burden of proof shifts to the algorithm. We cannot expect a solo private investigator or a detective to spot a 2-pixel shift in a jawline that was adjusted to dodge a match. We have to provide them with the math—the same Euclidean distance analysis used by enterprise-grade tools—that proves the relationship between two images is either legitimate or mathematically impossible.

The RSNA study found no correlation between career length and the ability to spot fakes. A thirty-year veteran performed no better than a resident. This proves that expertise in reading authentic images does not translate into an ability to detect manipulated ones. These are different cognitive tasks. In our codebase, this means we need to stop optimizing for "user confidence" and start optimizing for "measurement accuracy."

At CaraComp, we believe that providing investigators with these technical metrics—at a fraction of the cost of enterprise surveillance tools—is the only way to keep photographic evidence reliable in a post-truth environment. We are moving into an era where "seeing is believing" is a legacy bug. The future of investigation technology is 100% measurement-based.

How are you handling the "adversarial perfection" problem in your computer vision pipelines—are you implementing specific variance checks for synthetic landmarks, or do you still rely on a human-in-the-loop to flag forgeries?

Top comments (0)