TL;DR
An Ontario audit revealed that AI note-taking systems approved for healthcare providers frequently produce inaccurate, fabricated, or incomplete patient records. The findings highlight concerns over AI reliability in critical medical documentation, with ongoing questions about evaluation standards and safety measures.
The Office of the Auditor General of Ontario has reported that nine out of 20 AI-based medical note systems approved for use by healthcare providers routinely produce inaccurate or fabricated information, raising concerns about patient safety and data reliability.
The audit evaluated 20 AI note-taking systems used across Ontario’s healthcare sector, with findings indicating widespread inaccuracies. Nine systems fabricated information or suggested treatment plans not discussed in original recordings, while 12 inserted incorrect medication data into patient notes. Additionally, 17 systems missed key details about patients’ mental health issues, with six failing to capture these aspects fully or partially.
The evaluation process involved simulated doctor-patient recordings reviewed by medical professionals, who identified these errors. Despite these issues, the report notes that Ontario Ministry of Health officials have not reported any known patient harms directly linked to these AI systems, and more than 5,000 physicians are currently using the technology.
Why It Matters
This report underscores significant concerns about the reliability of AI tools used in critical healthcare documentation. Inaccurate medical records can lead to misdiagnoses, inappropriate treatments, and compromised patient safety. The findings also raise questions about the adequacy of current evaluation standards and oversight for AI systems in healthcare, emphasizing the need for stricter safety protocols and mandatory accuracy checks.

Plaud Note AI Voice Recorder, Note Taker w/Case, App Control, Transcribe & Summarize with AI, Support 112 Languages, for Meetings, Calls, Lectures, Professionals, Teams, Black, Non-Pro Version
Plaud Intelligence: Capture conversations in 112 languages and generate accurate transcripts with the Plaud App and Web. Plaud…
As an affiliate, we earn on qualifying purchases.
As an affiliate, we earn on qualifying purchases.
Background
The use of AI note-taking systems in Ontario began as part of a broader initiative to digitize and streamline healthcare documentation. The AI Scribe program was launched to assist physicians, nurse practitioners, and other healthcare professionals, with evaluations conducted through simulated recordings. Prior to this audit, similar concerns about AI reliability had been raised in other sectors, but this is among the first comprehensive assessments of AI accuracy in a critical medical context within Ontario.
“Inaccurate weightings could result in the selection of vendors whose AI tools may produce inaccurate or biased medical records or lack adequate protection to safeguard sensitive personal health information.”
— Office of the Auditor General of Ontario
“More than 5,000 physicians are participating in the AI Scribe program, and there have been no reports of patient harms associated with the technology so far.”
— Ontario Ministry of Health spokesperson

Medical Transcription: Techniques and Procedures
Used Book in Good Condition
As an affiliate, we earn on qualifying purchases.
As an affiliate, we earn on qualifying purchases.
What Remains Unclear
It is not yet clear how widespread the inaccuracies are in real-world usage beyond the evaluated sample, or whether immediate corrective measures are being implemented. The effectiveness of current oversight and safety protocols remains uncertain, and further investigation is needed to determine the full impact on patient safety.

AI for Nurses: The Practical Guide to HIPAA-Compliant AI Tools, Documentation Workflows, and Ethical Integration for Registered Nurses and Nurse Practitioners (AI for Professionals)
As an affiliate, we earn on qualifying purchases.
As an affiliate, we earn on qualifying purchases.
What’s Next
The Ontario Ministry of Health is expected to review the audit findings and may revise evaluation criteria for AI systems. Additional oversight measures, including mandatory accuracy attestations and security safeguards, are likely to be considered. Further audits and real-world assessments are anticipated to monitor improvements and ensure patient safety.

Unitech EA660 ID Passport Scanner – Touchless Handheld Scanning Solution with VeriScan Basic Software for Age Verification & Visitor Management. Free Charge Dock, Case, Screen Protector & Hand Strap.
Portable ID card scanner equipped with an electronic touch interface
As an affiliate, we earn on qualifying purchases.
As an affiliate, we earn on qualifying purchases.
Key Questions
What specific errors did the Ontario audit find in AI medical notes?
The audit found that 9 out of 20 systems fabricated information, such as treatment suggestions or findings not discussed in recordings, while 12 inserted incorrect drug data, and 17 missed key mental health details.
Are these AI systems currently causing patient harm?
According to the Ontario Ministry of Health, there have been no reported cases of patient harm linked to these AI note-taking systems so far, but the audit raises concerns about potential risks.
How were the AI systems evaluated in the audit?
The evaluation involved simulated doctor-patient recordings reviewed by medical professionals, who assessed the accuracy of the AI-generated notes. However, the report criticizes the scoring criteria used in the evaluation process.
What changes might Ontario implement following this report?
Ontario officials are likely to revise evaluation standards, possibly requiring mandatory accuracy checks, better security measures, and stricter oversight of AI tools used in healthcare documentation.
Could this impact the future adoption of AI in Ontario healthcare?
Yes, the findings may lead to increased scrutiny, tighter regulations, and possible delays in broader AI adoption until reliability and safety are assured.