← Journal · Vol. IIEngineering · No. 08 · 9 min read
Engineering · Clinical AI

Where the ambient scribe accuracy ceiling sits.

2026-02-28

Sarthi Editorial

Three years into ambient deployment, the remaining errors are not random. They cluster in predictable places — and the clustering is what the next generation of systems will have to solve.

If you ask a clinician who has been using an ambient scribe for six months whether it works, the answer you will get is almost always some version of “mostly.” The note is produced. The encounter is captured. The time at the keyboard is reduced. The thing that the product was sold to do, it does.

If you then ask the same clinician what still goes wrong, you will get a much more detailed and much less enthusiastic answer. Three years into serious ambient deployment, the remaining errors are known. They cluster. And the clusters are where the next generation of systems — including ours — has to do the engineering work.

This is a catalog of the edge cases, written from the perspective of what a specialty practice should expect, and what the honest ambient-product vendor should already be working on.

Where the errors cluster

Five clusters recur, specialty after specialty. They are not equally severe, but each one breaks in a characteristic way.

No. 01

Specialty vocabulary and proper nouns.

Device names (Merlin, LATITUDE, CareLink), specialty drug names (rivaroxaban, apixaban, macitentan, nintedanib), and procedure names (bronchoscopic lung volume reduction, left atrial appendage occlusion, pulmonary vein isolation). General-purpose ASR models perform measurably worse on this vocabulary; specialty-tuned models close some but not all of the gap.

No. 02

Multi-speaker and multi-accent encounters.

A single-speaker visit is the easiest case. A visit with a patient plus a spouse plus an interpreter, across regional accents and varying first-language English, is the hardest. Diarization — knowing who said what — remains where many systems break down. The resulting note sometimes attributes history-of-present-illness to the wrong speaker, which is not a spelling error; it is a clinical content error.

No. 03

Quiet consult dictation.

After-the-fact dictation, where the clinician steps out of the exam room and narrates into their phone or mic, is a different speech profile than an ambient in-room capture. Rates of speech, acoustic conditions, and content structure all differ. Systems optimized for in-room ambient sometimes underperform on this flavor of input, which clinicians often assume is the easy case.

No. 04

Numerical and tabular data.

Lab values, vital signs, device parameters, medication doses. Verbally stated numbers are at higher risk of transcription error than free text; a wrong dose in a note is not a typo, it is a potential safety event. Serious systems mitigate with post-processing validators against a drug- and measurement-aware schema, but the failure mode is real.

No. 05

The negatives.

Clinical reasoning depends heavily on what was ruled out, not just what was found. ‘Denies chest pain’ and ‘chest pain’ differ by one word in text and by the entire note in meaning. Errors in negation — missed or inverted negatives — are proportionally rare but disproportionately consequential.

What raises the ceiling

The engineering work to move past the current ceiling runs along several fronts, none of which is glamorous and most of which is the kind of work that does not show up in demo videos.

Specialty-tuned acoustic and language models

The baseline general-purpose model is the wrong starting point for specialty use. The right starting point is a model fine-tuned on clinical audio within the target specialty, with the specialty’s vocabulary — drug names, device names, procedure names — explicitly represented in the training distribution. This is unglamorous data work. It is also where most of the measurable accuracy gain lives.

Diarization that understands clinical structure

Generic diarization asks “which voice is this.” Clinical diarization asks “which voice is this, and what role does this voice play in the visit.” The note-generation logic has to know whether a given utterance was from the patient, a family member, or the clinician — because the same sentence is recorded differently depending on the source. The systems that get this right are measurably better on multi-speaker encounters.

Structured extraction, not free-text generation

One of the biggest accuracy gains available in 2026 comes from structuring the generation problem differently. Rather than asking the model to produce a free-text note in one pass, the more robust approach is to extract structured clinical entities — meds, allergies, problems, findings, plan items — and then render a note from the structured representation, with post-hoc validators applied against drug and lab schemas.

This is more engineering, not less. It is also where the next material rung of the accuracy ladder is.

Negation and temporality handling

A clinically aware system has to handle negation (denies vs. reports), temporality (history of vs. current), and conditionality (at risk for vs. has) as first-class concepts. Current ambient systems handle these adequately in easy cases and unevenly in hard ones. Purpose-built clinical NLP layers — built with enough corpus to see the edge cases — outperform general-purpose summarization here by a clear margin.

Continuous evaluation, not milestone evaluation

Ambient systems that ship with a continuous evaluation harness — a set of known-answer clinical audio fixtures that run on every release, plus production sampling with clinician-annotated ground truth — are the ones that improve. Systems that evaluate at milestones (“we shipped v2; here are the benchmark numbers”) are the ones that plateau.

What a practice should ask its scribe vendor in 2026

If your vendor is an incumbent, these are the questions whose answers distinguish a serious engineering posture from a glossy demo.

  • What is the clinically meaningful error rate on notes in my specialty? Not word-error-rate. Not a general benchmark. The rate at which a clinician in your specialty would want to correct a note produced by the system, measured on a representative sample.
  • How does the system handle multi-speaker encounters with accented English? Ask for evidence. A demo of a single-speaker quiet-room visit does not answer the question.
  • How do you validate medication and dose mentions? A serious answer describes a post-generation schema validator, probably against RxNorm, with a known false-negative and false-positive rate. A weak answer describes the underlying model’s accuracy in the abstract.
  • What happens to the edge-case notes I edit? A serious vendor uses clinician edits as signal for continuous improvement, with appropriate de-identification and consent. A vendor that treats clinician edits as static events is not improving.
  • How do you handle the negation and temporality edge cases? A vague answer — “the model handles context” — is a vague answer. A specific answer references a clinical-NLP layer with measurable behavior.

The honest summary

The ambient scribe is not a solved problem. It is a problem that is solved enough, in enough cases, to be genuinely useful — and unsolved enough, in enough other cases, that the next five years of clinical-documentation AI will be largely about pushing the accuracy ceiling up, not about shipping new form factors.

For practices evaluating the market in 2026, the right frame is not “should we adopt an ambient scribe.” That question was settled. The right frame is “which vendor is actively working on the edge cases that matter for my specialty, and can show me measurable progress on them release over release.”

That is a smaller field than the marketing would suggest.

References
  1. 1. Peterson Health Technology Institute, ambient documentation evaluations (2024–2025).
  2. 2. Published deployment studies from Permanente, Atrium Health, and academic medical centers (2023–2025).
  3. 3. ONC Interoperability Standards Advisory entries on clinical NLP evaluation practices.
Next

A scribe is step one, not the answer.

Sarthi extends past the note — into coding, prior authorization, intake, and follow-up. Thirty minutes, a shared screen, and we walk you through where the ambient layer ends and the workforce begins.

Book a walkthrough