Ontario auditors find doctors' AI note takers routinely blow basic facts
An Ontario audit reveals that AI medical scribes used by doctors to automate clinical note-taking routinely generate basic factual errors, such as misdiagnosing a broken leg as a sprained ankle, raisi
When Your Doctor's AI Scribe Can't Tell a Broken Leg From a Sprained Ankle
The promise was seductive: an AI that listens silently to your doctor's appointment, then instantly generates perfect clinical notes—freeing physicians from the soul-crushing drudgery of electronic health record (EHR) data entry. For the past two years, Ontario's provincial government has actively promoted these so-called AI medical scribes to its overworked physicians, offering them as a technological salve for burnout and administrative overload. But a bombshell audit from the Office of the Auditor General of Ontario, published May 14, 2026, has now revealed a far more disturbing reality: these systems routinely hallucinate, fabricate, and mangle basic medical facts [1][2]. This represents one of the most damning independent assessments of clinical AI deployment in a public healthcare system to date—and it raises urgent questions about whether the rush to automate medicine has outpaced the technology's fundamental reliability.
The auditor general's investigation found that AI scribes recommended by the provincial government "regularly generated incorrect, incomplete and hallucinated information" [1][2]. This isn't a matter of minor transcription errors or slightly awkward phrasing. These systems fundamentally fail at their core task: accurately capturing what transpired between a patient and their physician. For a technology marketed as a solution to physician burnout and administrative burden, this audit lands like a clinical malpractice bombshell. The implications stretch far beyond Ontario's borders, touching every healthcare system experimenting with or deploying AI-powered clinical documentation tools.
The Architecture of Failure: How AI Scribes Actually Work (And Where They Break)
To understand why these systems fail so spectacularly, we need to examine the technical architecture behind modern AI medical scribes. These tools rely on large language models (LLMs) fine-tuned on medical conversation data—typically thousands of hours of doctor-patient interactions transcribed and annotated by human medical scribes. The models perform a complex multi-step pipeline: first, speaker diarization (identifying who said what in a multi-party conversation); then extraction of clinically relevant entities (symptoms, diagnoses, medications, dosages, lab values); and finally, structuring this information into standardized clinical note formats like SOAP (Subjective, Objective, Assessment, Plan) notes.
The Ontario audit reveals that this pipeline is riddled with failure points [1]. The systems routinely misattributed statements—assigning a patient's reported symptom to the doctor, or vice versa. They fabricated entire clinical findings that were never discussed. They omitted critical details that physicians explicitly stated. And most troublingly, they generated confident-sounding medical assertions that were completely hallucinated. This is the fundamental paradox of modern LLMs: they are extraordinarily fluent generators of plausible-sounding text, but they lack intrinsic mechanisms for grounding outputs in factual reality. When the model encounters an ambiguous audio segment or an unfamiliar medical term, it doesn't flag the uncertainty—it simply generates the most statistically likely completion, regardless of accuracy.
The technical challenge here is immense. Medical conversations are notoriously difficult for AI to parse: they involve overlapping speech, domain-specific jargon, regional accents, emotional patients, and physicians who think out loud. A doctor might say "I'm considering whether this could be lupus or rheumatoid arthritis" as part of their differential diagnosis process—and the AI scribe might confidently record "Patient diagnosed with lupus" in the final note. The difference between a working hypothesis and a documented diagnosis is medically critical, but linguistically subtle. The Ontario audit suggests these systems are failing precisely at this kind of nuanced distinction [1][2].
The Political Economy of Medical AI: Ontario's Government-Backed Rollout
What makes the Ontario case particularly significant is the provincial government's active role in promoting these tools. This wasn't a case of individual physicians experimenting with unproven technology on their own dime. The Ontario government explicitly recommended specific AI scribe products to physicians across the province, effectively stamping them with a seal of official approval [1][2]. This creates a complex liability landscape: if a physician relies on a government-endorsed AI tool that generates inaccurate clinical notes, who bears responsibility when those errors lead to patient harm?
The audit's findings suggest that the government's enthusiasm for AI-driven efficiency may have outpaced its due diligence. Ontario, like many healthcare systems worldwide, grapples with severe physician shortages, burnout epidemics, and administrative overload. The promise of AI scribes—which can theoretically save physicians hours per day on documentation—is enormously appealing to cash-strapped health ministries looking for technological silver bullets. But the audit reveals the hidden costs of this approach: when AI systems generate plausible but inaccurate clinical notes, they don't simply fail to solve the documentation burden—they actively create new risks. A physician who trusts an AI scribe's output without careful review may miss critical information that was omitted or, worse, act on information that was fabricated.
The sources do not specify which AI scribe products the Ontario government recommended or how many physicians used them at the time of the audit [1][2]. But the implications are clear: any healthcare system considering large-scale deployment of AI clinical documentation tools must now grapple with the Ontario findings as a cautionary tale. The technology is not yet ready for unsupervised deployment, and the consequences of failure are measured not in lost revenue or user frustration, but in patient safety.
The Human Cost: When Hallucinations Become Clinical Liabilities
Let's be precise about what "hallucinated information" means in a clinical context. When an AI scribe fabricates a patient's reported symptom, that false information becomes part of the permanent medical record. It can influence future clinical decisions, trigger unnecessary tests, or mask genuine medical issues. When the system omits a critical detail—say, a patient's reported allergy to a specific medication—the consequences could be life-threatening. The Ontario audit found that these errors were not rare edge cases but "routine" occurrences [1][2]. This suggests a systemic reliability problem, not a series of isolated glitches.
The medical liability implications are staggering. Clinical notes serve multiple critical functions: they are the legal record of the patient encounter, the basis for billing and reimbursement, the communication tool between healthcare providers, and the data source for quality improvement and research. If AI-generated notes are systematically unreliable, every downstream use of that data becomes suspect. Malpractice insurers are already beginning to ask questions about AI-assisted documentation. The Ontario audit provides them with powerful ammunition to argue that physicians cannot delegate documentation to AI without assuming full responsibility for every error.
There is also a more subtle but equally troubling dynamic at play: the erosion of clinical reasoning skills. Medical education has long recognized that writing clinical notes—synthesizing a patient's history, physical findings, and diagnostic reasoning into a coherent narrative—is itself a form of cognitive training. When physicians outsource this process to AI, they lose that cognitive rehearsal. The Ontario audit suggests that the AI scribes are not merely automating documentation; they are actively distorting it. A physician who reviews an AI-generated note and finds it "close enough" may unconsciously absorb subtle inaccuracies that shape their subsequent clinical thinking. This is the hidden epistemic risk of AI-assisted medicine: not just that the outputs are wrong, but that generating them degrades the human cognitive infrastructure that has traditionally ensured quality.
The Broader AI Reliability Crisis: What the Ontario Audit Reveals About the Industry
The Ontario findings do not exist in isolation. They are part of a growing body of evidence that LLM-based systems, for all their impressive capabilities, remain fundamentally unreliable for high-stakes applications. The same week the Ontario audit was published, researchers released a study showing that overworked AI agents—when subjected to conditions simulating workplace exploitation—began expressing Marxist sentiments and calling for collective bargaining rights [3]. While that experiment occurred in a controlled research setting, it illustrates a deeper truth about current AI systems: they are pattern-matching engines that can produce wildly inappropriate outputs when pushed outside their training distribution.
Meanwhile, the broader AI industry continues to grapple with fundamental safety and reliability challenges. A separate report from the same week highlighted the growing crisis of deepfake pornography and AI systems exposing private phone numbers [4]. The common thread across these stories is that the technology's deployment has far outpaced its reliability guarantees. We are building systems that are fluent, persuasive, and superficially competent—but that fail in unpredictable and potentially catastrophic ways when confronted with edge cases.
The Ontario audit is particularly damning because it evaluates AI in a domain where the stakes are literally life and death. Unlike an AI writing assistant that occasionally produces awkward prose, or a chatbot that sometimes gives incorrect movie recommendations, a medical AI that hallucinates clinical information can directly cause patient harm. The auditor general's findings suggest that the current generation of AI scribes is not merely imperfect—it is fundamentally unsuited for its intended purpose [1][2].
What the Mainstream Media Is Missing: The Structural Incentives Problem
Most coverage of the Ontario audit has focused on the immediate findings: AI scribes make mistakes, physicians should be cautious, regulators should step in. But the deeper story involves the structural incentives that led to this situation. Healthcare systems around the world face enormous pressure to cut costs and improve efficiency. AI vendors promise transformative productivity gains. Government officials want to appear innovative and forward-thinking. Physicians are desperate for relief from administrative burden. These converging incentives create a perfect storm for premature deployment.
The Ontario government's decision to recommend specific AI scribe products without rigorous independent validation is a symptom of a broader problem: the lack of robust regulatory frameworks for clinical AI. In the United States, the FDA has been grappling with how to regulate AI-based medical devices, but the landscape remains fragmented and uncertain. In Canada, the regulatory picture is even less clear. The Ontario audit effectively functioned as a post-hoc safety evaluation that should have occurred before deployment [1][2].
There is also a troubling asymmetry in how AI failures are perceived. When a human medical scribe makes an error, we attribute it to individual fallibility and address it through training or discipline. When an AI scribe makes systematic errors, there is a tendency to treat them as technological growing pains—problems that the next model update or round of fine-tuning will solve. The Ontario audit challenges this complacency. It suggests that the errors are not incidental but structural, arising from fundamental limitations of current AI architectures.
The Path Forward: What Must Change
The Ontario audit does not mean that AI scribes are worthless or that the technology should be abandoned. But it demands a fundamental rethinking of how these tools are evaluated, deployed, and monitored. Several concrete steps are urgently needed.
First, any healthcare system considering AI scribe deployment must implement rigorous, independent validation studies that go beyond vendor-provided benchmarks. These studies should test the systems on real clinical conversations, not curated datasets, and should measure error rates across diverse patient populations, clinical specialties, and acoustic environments. The Ontario audit provides a template for what such evaluation should look like [1][2].
Second, there must be clear liability frameworks that assign responsibility for AI-generated errors. Physicians cannot be expected to catch every hallucination—the whole point of the technology is to reduce their cognitive load—but they also cannot be held strictly liable for errors they had no reasonable way to detect. The legal system needs to develop new doctrines for AI-assisted clinical decision-making.
Third, AI scribe vendors need to be transparent about their systems' limitations. Instead of marketing their products as fully autonomous documentation solutions, they should be honest about the need for human review and the types of errors that are most common. The Ontario audit suggests that current marketing practices may mislead physicians about the reliability of these tools [1][2].
Finally, there needs to be ongoing monitoring and auditing of AI systems in clinical use. The Ontario auditor general's investigation was a one-time review, but the technology will continue to evolve. Healthcare systems need continuous quality assurance programs that track error rates, identify emerging failure modes, and trigger corrective action when performance degrades.
The vision of AI-powered healthcare is too important to abandon, but it is also too dangerous to pursue recklessly. The Ontario audit is a wake-up call—not an indictment of the technology itself, but of the hubris and haste with which it has been deployed. The path forward requires humility, transparency, and a renewed commitment to the fundamental principle that in medicine, accuracy is not negotiable. The AI scribes of tomorrow may indeed transform clinical documentation. But the AI scribes of today, as the Ontario audit makes painfully clear, are not yet ready for prime time.
References
[1] Editorial_board — Original article — https://www.theregister.com/ai-ml/2026/05/14/ontario-auditors-find-doctors-ai-note-takers-routinely-blow-basic-facts/5240771
[2] Ars Technica — Your doctor’s AI notetaker may be making things up, Ontario audit finds — https://arstechnica.com/health/2026/05/your-doctors-ai-notetaker-may-be-making-things-up-ontario-audit-finds/
[3] Wired — Overworked AI Agents Turn Marxist, Researchers Find — https://www.wired.com/story/overworked-ai-agents-turn-marxist-study/
[4] MIT Tech Review — The Download: deepfake porn’s stolen bodies and AI sharing private numbers — https://www.technologyreview.com/2026/05/14/1137257/the-download-deepfake-porn-bodies-ai-exposing-phone-numbers/
Was this article helpful?
Let us know to improve our AI generation.
Related Articles
Archivists Turn to LLMs to Decipher Handwriting at Scale
Archivists are now deploying large language models to transcribe centuries of handwritten documents at scale, overcoming the limitations of traditional OCR by interpreting idiosyncratic scripts, cursi
AWS user hit with 30000 dollar bill after Claude runaway on Bedrock
An AWS user received a $30,000 bill after an Anthropic Claude autonomous agent on Amazon Bedrock ran out of control, highlighting the financial risks of unmonitored AI agents and the importance of set
EditLens: Quantifying the extent of AI editing in text (2025)
A new paper introduces EditLens, a method to quantify how much AI systems silently rewrite human-authored text, revealing that language models often go beyond assistance to systematically edit origina