I recently watched a team spend three months evaluating speech to text vendors using word error rate (WER) as their primary metric. They picked the "best" model based on a 0.3% WER advantage. Two weeks into production, their medical transcription system was flagging drug names as errors and missing critical patient information that their "worse" model had captured perfectly.
This scenario plays out constantly across organizations evaluating speech recognition systems. The problem isn't the technology. It's how we measure it.
What Is Word Error Rate?
Word error rate measures the percentage of words a speech recognition system gets wrong compared to a human reference transcript. It calculates errors by counting substitutions (wrong words), insertions (extra words), and deletions (missing words), then dividing by the total word count. WER has been the default evaluation metric for speech recognition evaluation for decades.
The formula looks simple: WER = (S + D + I) / N × 100, where S is substitutions, D is deletions, I is insertions, and N is the total number of words in the reference transcript. A 5% word error rate means five out of every 100 words differ from the human baseline.
The Core Problem: When Better Models Score Worse
Here's the paradox that should concern every team relying on WER for vendor selection: more accurate models often receive higher error rates.
I've seen this firsthand while testing different transcription systems. When I uploaded a 45-minute medical consultation to Scriptivox, it captured quiet patient responses like "mm-hmm" and "okay" that the human transcriber had missed entirely. According to WER calculations, each of these correctly transcribed words counted as an "insertion error."
The model was penalized for being more accurate than the human baseline.
This isn't a theoretical edge case. Modern speech recognition systems, especially those trained on diverse real-world audio, frequently outperform human transcribers on difficult content. Noisy environments, overlapping speakers, accented speech, and rapid conversation all create situations where AI catches words that human ears miss.
WER treats this superior performance as failure.
Three Fatal Flaws in Traditional Speech to Text Accuracy Measurement

Not All Transcription Errors Impact Your Business Equally
WER counts "gonna" transcribed as "going to" the same as "lisinopril" transcribed as "less than a pill." One preserves meaning while changing style. The other transforms a medication name into nonsense that could harm a patient.
In voice agent applications, this distinction becomes even more critical. A substitution (the AI's best guess at an unclear word) keeps the conversation flowing. A deletion (missing the word entirely) can cause the system to hang, waiting for input that never comes. Traditional WER treats both identically.
When transcripts feed directly into large language models for analysis or automation, context preservation matters more than exact word matching. An LLM can work with "gonna" just as effectively as "going to." It cannot extract meaningful insights from a mangled drug name.
Human Reference Transcripts Are Often Wrong
The "ground truth" files used for WER calculations contain systematic errors. Human transcribers miss backchannels ("uh-huh," "right"), quiet affirmations, overlapping speech, false starts, and disfluencies. They're trained to clean up transcripts for readability, not capture every spoken word.
When I tested Scriptivox against human-labeled reference files, the majority of "errors" were cases where the AI correctly identified words the human transcriber had skipped. Playing back the source audio confirmed this repeatedly.
This creates a measurement system where accuracy is defined as "how closely does the AI match flawed human work?" rather than "how accurately does the AI transcribe what was actually spoken?"
Public benchmark datasets compound this problem. Many contain transcription errors that penalize any model capable of capturing the full audio content.
Formatting Differences Masquerade as Accuracy Problems
Modern speech to text systems include language model components that understand grammar and formatting conventions. This capability often produces outputs that differ from human transcripts in style while maintaining identical meaning.
"Healthcare" versus "health care." "All right" versus "alright." "Setup" versus "set up." Each represents a valid formatting choice, but WER calculations treat every instance as an error.
The Whisper Normalizer, an open-source tool most teams use to standardize transcriptions before comparison, catches common variations like "don't" mapping to "do not." But it misses domain-specific equivalences that only matter in context.
These formatting penalties accumulate across longer transcripts, artificially inflating error rates for models that demonstrate stronger grammatical understanding.
Speech to Text Metrics That Actually Measure What Matters
If WER misleads teams about real-world performance, what should you use instead? The answer depends on your specific use case, but four metrics provide clearer signals about production quality.
Semantic Word Error Rate
Semantic WER applies domain-specific normalization before calculating errors. Instead of penalizing every word difference, it categorizes variations: no penalty for equivalent spellings and contractions, minor penalties for single-character errors, and major penalties for meaning-altering mistakes.
This approach works particularly well for applications where transcripts feed into downstream AI systems. If your speech to text output flows into an LLM for analysis or a voice agent for processing, semantic meaning preservation matters more than exact word matching.
Missed Entity Rate
Rather than asking "how many words did you get wrong?", missed entity rate asks "did you capture the words that actually matter?" You define critical entities for your domain (drug names, account numbers, email addresses, product codes) and measure their survival rate through transcription.
For medical applications, a single missed medication name represents system failure regardless of overall word accuracy. For financial services, a correctly transcribed account number matters more than perfect capture of conversational filler.
When testing medical transcription accuracy, I found that models with slightly higher overall WER often achieved zero missed entity rates on critical terminology, while "better" models according to traditional transcription benchmarks failed to capture essential clinical information.
LLM-as-a-Judge Evaluation
This approach uses a large language model to evaluate transcript quality by comparing semantic meaning rather than exact word matches. The LLM judge assigns penalties based on meaning preservation and provides structured feedback about specific error types.
Unlike single-number metrics, LLM evaluation offers diagnostic insights. Instead of just knowing that accuracy dropped 2%, you understand whether the problems stem from speaker identification, technical terminology, or audio quality issues.
This method works especially well for conversational AI applications where maintaining context and intent matters more than perfect word-for-word accuracy.
Choosing the Right Metric for Your Application
Medical and legal teams should prioritize missed entity rate on critical terminology. Voice agent developers need semantic WER combined with emission latency measurements. Meeting transcription products serving human readers still benefit from traditional WER, but only after applying proper normalization.
The most effective approach combines multiple metrics rather than relying on any single measurement.
Building Reliable Speech Recognition Evaluation: A Step-by-Step Approach
Creating evaluation systems that reflect real-world performance requires fixing three fundamental problems: corrupted ground truth, inappropriate metrics, and artificial test conditions.
Step 1: Audit Your Reference Transcripts
Before evaluating any speech recognition system, verify that your human-labeled transcripts accurately represent the source audio. This process reveals systematic gaps in your evaluation dataset.
Upload a sample of your reference files to a modern transcription service and compare the AI output against your human labels. For each disagreement, play back the corresponding audio segment. You'll often find the AI captured words the human transcriber missed.
Document these corrections. Clean reference files improve evaluation accuracy for all future vendor comparisons.
Step 2: Create Domain-Specific Word Lists
Build equivalence lists that normalize valid formatting variations before calculating error rates. "Healthcare" maps to "health care." "Alright" maps to "all right." "Setup" maps to "set up."
These semantic mappings prevent any vendor from being penalized for grammatically correct formatting choices. The normalization should reflect your domain's conventions and the downstream systems that will process the transcripts.
For medical applications, this includes drug name variations and clinical terminology. For financial services, it covers account types and transaction language. For legal work, it encompasses procedural terms and citation formats.
Step 3: Test Under Realistic Conditions
Transcription benchmarks using clean, single-speaker audio don't predict performance on noisy conference calls or overlapping conversations. Your evaluation dataset should match your production audio characteristics.
Include background noise, multiple speakers, accented speech, and domain-specific terminology. Test file formats and audio quality levels that reflect your actual use case.
When I evaluated transcription systems for a customer service application, clean benchmark audio showed minimal differences between vendors. Testing with actual call center recordings revealed dramatic performance gaps that clean audio couldn't predict.
Step 4: Measure Business-Critical Outcomes
Pair speech to text accuracy metrics with downstream performance measurements. For voice agents, track task completion rates and conversation success. For meeting transcription, measure human correction frequency and user satisfaction.
These business metrics often contradict pure accuracy scores. A system with 8% WER that captures 100% of action items may serve meeting participants better than one with 6% WER that misses critical decisions.
Speech Recognition Systems: Vendor Comparison

Based on testing across medical, business, and customer service audio, here's how major providers perform on metrics that matter for production applications:
Scriptivox excels at capturing complete conversations with word-level timestamps. The speaker identification works reliably even with overlapping speech, and the 100-language auto-detection handles multilingual meetings without manual configuration. When I tested it on challenging customer service recordings with background noise, it consistently captured entity-critical information that other systems missed. The built-in audio converter and trimmer tools eliminate preprocessing steps that slow down evaluation workflows.
Otter.ai provides strong real-time transcription with good meeting integration, but struggles with technical terminology and accented speakers. The collaborative editing features help teams correct errors, though high error rates on domain-specific content require significant manual cleanup. For teams prioritizing live meeting notes over accuracy, Otter's real-time capabilities offer value despite transcription limitations.
Rev offers human transcription alongside AI options, making it suitable for high-stakes applications where perfect accuracy justifies higher costs and longer turnaround times. The AI-only service performs well on clear audio but lacks advanced features like speaker identification that modern workflows require.
Descript combines transcription with video editing capabilities, making it valuable for content creators. However, the transcription accuracy lags behind specialized providers, particularly on conversational audio with multiple speakers. The editing integration creates workflow value for video production teams despite accuracy limitations.
Assembly AI provides developer-focused APIs with good accuracy on clear audio. The real-time streaming works well for live applications, but the system struggles with background noise and overlapping speakers in challenging environments.
For most business applications requiring reliable speech to text accuracy at scale, Scriptivox provides the best balance of accuracy, features, and cost-effectiveness.
When Word Error Rate Still Makes Sense
Despite its limitations, WER remains useful in specific contexts. For comparing the same model across different versions or audio conditions, WER provides consistent directional feedback. If you're tracking how a system improves from version 1 to version 2 on identical test sets, WER offers clean comparative data.
Read-speech applications like audiobook narration or news broadcasts also benefit from traditional WER measurement. Clean, single-speaker audio with prepared text reduces the failure modes this analysis describes.
The problem isn't WER itself. It's using WER as the only metric and applying it to domains where its assumptions break down. Think of WER like a thermometer: useful for quick health checks, but insufficient for comprehensive diagnosis.
Modern speech to text evaluation requires multiple metrics matched to specific use cases. Teams building production systems need measurements that reflect real user experiences, not academic benchmarks.
Frequently Asked Questions
Q: What is word error rate and why do teams use it for speech recognition evaluation?
Word error rate measures the percentage of words a speech recognition system gets wrong compared to human reference transcripts, counting substitutions, insertions, and deletions. Teams use WER because it provides a single, comparable number across different systems and has been the industry standard for decades. However, WER treats all word errors equally and can penalize more accurate models that capture words human transcribers missed, making it unreliable for modern speech to text evaluation.
Q: How can I improve speech to text accuracy in my system?
Improve your audio quality first: use higher sample rates (16kHz minimum), reduce background noise, and position microphones closer to speakers. Choose models trained on audio similar to your domain, enable speaker identification for multi-person conversations, and provide custom vocabulary lists for uncommon terminology. However, optimizing purely for lower WER may hurt real-world performance if it causes the system to miss critical entities or produce less meaningful transcripts.
Q: What speech to text metrics should I use instead of word error rate?
Use multiple metrics matched to your use case: Semantic WER applies normalization before calculating errors, removing penalties for valid formatting differences. Missed Entity Rate measures survival of critical terminology like drug names or account numbers. LLM-as-a-Judge evaluation assesses meaning preservation rather than exact word matches. For voice agents, add emission latency and task completion rates. Medical applications should prioritize entity accuracy over overall word count.
Q: How do I build accurate transcription benchmarks for comparing vendors?
Start by auditing your reference transcripts - play back audio segments where AI and human transcripts disagree to identify systematic human transcriber errors. Create domain-specific word lists that normalize equivalent terms before calculating metrics. Test with audio that matches your production conditions, including background noise and multiple speakers. Measure business outcomes like user satisfaction and task completion alongside speech recognition evaluation metrics, since these often contradict pure WER scores.
Q: Why do some speech recognition systems perform worse on benchmarks but better in production?
Benchmark datasets often contain human transcription errors that penalize AI systems for being more accurate than the reference transcript. Modern speech recognition captures quiet speech, backchannels, and overlapping conversations that human transcribers typically miss or clean up for readability. Additionally, models with stronger language understanding may format text differently while preserving meaning, creating artificial "errors" in WER calculations that don't impact real-world usability. This disconnect between academic benchmarks and practical performance explains why production testing reveals different accuracy rankings than published comparisons.
Moving beyond traditional speech to text accuracy measurement requires understanding that perfect word-for-word matching doesn't equal business value. The teams succeeding with speech recognition in 2026 measure what matters for their specific applications, not what academic benchmarks tell them to care about.



