Table of Contents
Speech recognition replaces the traditional multi-stage ASR pipeline with a single neural network trained directly on audio-to-text pairs. Architecture still determines latency, streaming limits, and vocabulary failure patterns in production. This article covers the three dominant speech recognition architectures. It also explains the production failure modes each one creates and how to match model type, vocabulary strategy, and streaming architecture to your workload.
Key Takeaways
Here's the short version before we get into the weeds:
- CTC limits: fast, streamable, and weakest on rare terms.
- Attention models: strongest offline accuracy, but not native streaming.
- RNN-T overview: the usual choice for streaming and real-time use.
- Production WER often gets much worse in noisy conditions.
- Runtime vocabulary prompting can fix many OOV failures without retraining.
What Speech Recognition Actually Means
Speech recognition uses one model to map audio directly to text. That simplifies the stack, but it makes architecture choice central to production behavior.
The Pipeline It Replaced
Traditional ASR systems chain three separately trained components. An acoustic model converts audio features into phoneme probabilities. A pronunciation model maps phonemes to words using a hand-built dictionary. A language model rescoring candidate transcriptions uses statistical n-gram probabilities. Each component introduces its own error surface. Errors compound across stages, and diagnosing failures requires tracing through all three. Deepgram's overview of ASR for business describes this as replacing "a few different models strung together" with "a model that does everything."
What Changes When You Use One Model
A single model learns the mapping from raw audio to text. There's no pronunciation dictionary to maintain and no separate language model to tune. WER becomes a direct function of model architecture and training data. It becomes less about pipeline integration quality. You gain simplicity, but you lose the ability to fix individual pipeline stages independently.
Why the Demo-to-Production Gap Starts Here
Clean benchmark results often don't survive real audio. The architecture you choose determines where the gap widens. A multilingual ASR study found that reverberation alone increases English WER by 24.9 percentage points. That shift moves WER from 6.0% to 31.0%. CTC models degrade on vocabulary. Attention models degrade on latency. Transducers degrade on rare terms. Those trade-offs often decide whether you have a working demo or a working product.
The Three Single-Model Families
The three dominant architectures make different trade-offs between latency, streaming, and accuracy. If you need the simplest rule, choose based on whether you prioritize decoding speed, offline accuracy, or real-time responsiveness.
CTC: Fast Decoding, Independence Assumptions
CTC (Connectionist Temporal Classification) emits token probabilities at every audio frame independently. That conditional independence is both its strength and its weakness. CTC decodes faster than any other architecture because it's non-autoregressive. It streams natively. But without an internal language model, it struggles with proper nouns, rare terms, and context-dependent words. Stanford's SLP3 textbook frames this directly. CTC needs a way to "know about output history" for improved recognition. You can partially compensate with external n-gram language models at decode time.
Attention Encoder-Decoder: Accuracy at the Cost of Latency
Attention encoder-decoders are also called Listen-Attend-Spell or Transformer seq2seq. They use a full autoregressive decoder with cross-attention over the complete audio input. Every output token is conditioned on all previous tokens and the entire acoustic context. This ACL paper reports the strongest offline accuracy, especially on short utterances. The cost is that it isn't natively streaming. The decoder needs the full input sequence before generating output. Making it stream requires architectural modifications like chunk-wise attention. An ACL 2023 paper on hybrid ASR architectures confirms that input-synchronous decoding in CTC and Transducer helps avoid the over-generation and under-generation problems attention decoders face.
Transducer: The Streaming Architecture
RNN-T is the standard choice when you need streaming with context-aware decoding. It balances latency and context better than the other two families.
RNN-T combines a frame-synchronous encoder with a prediction network conditioned on previously emitted tokens. This gives it both streaming capability and context-aware decoding. It's the standard choice for on-device and real-time ASR. Latency budgets can vary with a single Transducer model depending on configuration. The trade-off is that it handles rare terms better than CTC but worse than a full attention decoder in offline mode. That's the gap runtime vocabulary features are designed to close.
Where Speech Recognition Models Break in Production
Most production failures come from three sources: audio input problems, out-of-vocabulary terms, and domain mismatch. If you don't test for those directly, benchmark WER won't tell you much.
Audio Format and Sample Rate Problems
Sending the wrong audio format is the fastest way to get bad results from any streaming API. Telephony audio recorded at 8kHz and upsampled to 16kHz creates an irreducible WER ceiling. One clinical deployment on real telephony audio degraded to 40.94% WER. Researchers attributed that floor partly to this upsampling mismatch. Clipping from misconfigured audio capture pipelines ranks among the most severe perturbations. Applying denoising preprocessing before inference often makes things worse—if you've battled audio quality issues before, this one stings. Two independent studies confirmed that speech enhancement degrades ASR across all tested model-noise configurations. Modern models have internalized noise handling. Preprocessing removes features they expect.
Out-of-Vocabulary Terms and Proper Nouns
OOV failures don't throw errors. They produce plausible-sounding wrong transcriptions, like "try to win" instead of "tretinoin" or inconsistent transliterations for code identifiers like getData(). Word-level systems can't recognize terms absent from training data. Subword tokenization shifts the problem but doesn't eliminate it. CTC personalization is structurally hard at inference time. Its conditional independence assumption prevents context-aware beam path customization. In code-switching scenarios with imbalanced language ratios, OOV problems compound further on the embedded language.
Domain Mismatch and the WER Gap
Clean benchmark performance often degrades sharply in production. Domain mismatch spans more than background noise alone.
You'll usually see much worse performance in noisy production environments. Noise level is only one axis. The multilingual ASR study linked earlier says mismatch spans at least four dimensions: SNR and noise type, reverberation and far-field effects, accent and dialect variation, and domain-specific vocabulary distribution. Each dimension compounds independently. Reliability on one axis is a poor predictor of reliability on another.
Speech Recognition Model Customization Without Retraining
Runtime vocabulary adaptation can close many domain-specific vocabulary gaps without retraining. If terminology is your main problem, this is usually the fastest fix.
Keyterm Prompting lets you adapt a model to domain-specific vocabulary at runtime. No retraining is required for up to 100 terms per request.
Keyterm Prompting vs. Custom Model Training
Keyterm Prompting works by biasing Nova-3 and Flux models toward specific terms you pass in the API request. You can include up to 100 keyterms per request, with a hard ceiling of 500 tokens across all terms. Multi-word phrases are boosted as a single cohesive unit. This differs from the legacy Keywords feature, which boosted individual words rather than phrases and carried an explicit reliability caveat in its documentation. Keyterm Prompting addresses the OOV problem at the inference layer without touching model weights.
When Runtime Adaptation Is Enough
If your vocabulary gaps are predictable, runtime prompting is often enough. It works best when you already know the bounded set of terms the base model misses.
If your vocabulary gaps are predictable, runtime prompting handles them. You're dealing with a bounded set of terms that the base model misrecognizes. Pass them as keyterms. The documented examples show recognition of "tretinoin" improving from "try to win" to the correct term. For contact centers, healthcare documentation, or any domain with a known terminology list under 100 terms, this is the fastest path to production accuracy.
When You Need a Custom Model
Runtime adaptation helps with vocabulary, not acoustics. If your WER gap comes from audio conditions, you need custom model training instead.
Runtime adaptation has limits. If your audio environment is acoustically distinct, the problem isn't vocabulary. It's the acoustic distribution. Custom model training adapts to both vocabulary and acoustic conditions simultaneously. You should consider it when your WER gap persists after adding keyterms, or when your domain vocabulary exceeds the 100-term limit.
Streaming vs. Batch: How Architecture Shapes Your Decision
Choose streaming or batch based on latency budget first. Architecture determines what trade-offs are possible once that budget is set.
Real-time voice agents require streaming architecture with persistent WebSocket connections. Batch workloads favor full-context attention models that trade latency for accuracy.
Streaming Architecture Requirements
Streaming speech recognition runs over persistent, bidirectional WebSocket connections. Audio flows upstream while transcripts return downstream over the same open socket, rather than as a single completed file upload. Streaming design choices affect latency, context, and operational complexity, so you'll want to validate them against your workload.
Batch Processing and Full-Context Accuracy
Batch mode gives you more acoustic context and usually better accuracy. If you don't need subsecond responses, it's often the simpler option.
When you send complete audio files after recording finishes, attention-based architectures can use their full cross-attention mechanism over the complete audio before producing output. The accuracy advantage is measurable. Streaming gap data shows Transducer models with WER gaps between offline mode (5.0%) and streaming with zero look-ahead (9.5%). If your workload is post-call analytics, meeting transcription, or compliance review, batch mode gives you better accuracy without the infrastructure complexity of persistent WebSocket connections.
Choosing Based on Latency Budget, Not Use Case Label
Don't choose streaming or batch based on a use case label. Choose based on your latency budget. Voice agents and live captioning need responses within hundreds of milliseconds, so streaming is the only option. Call analytics that run after the conversation ends can use batch. The gray area is near-real-time dashboards where a few seconds of delay is acceptable. In that case, buffered streaming with larger chunks can split the difference between latency and accuracy.
Validating Your Speech Recognition Setup
You should validate speech recognition with your own production-like audio before choosing an architecture or provider. Benchmark scores are useful for screening, but they won't expose your real failure modes.
Build a Representative Test Set
Collect 50–100 audio samples from your actual production environment. Include your hardest cases: accented speakers, background noise, domain-specific terminology, and the audio formats your pipeline actually produces. Clean benchmark datasets won't reveal the failures you'll hit in production.
Metrics That Match Your Use Case
WER alone doesn't tell the full story. If you're building a voice agent, measure task completion rate under realistic conditions. If you're doing medical transcription, measure accuracy on clinical terms. If you're processing contact center calls, measure per-speaker accuracy with diarization enabled. Pick the metric that maps to your business outcome.
Getting Started with Deepgram Nova-3
Deepgram positions Nova-3 and Flux for different workload types. Nova-3 is positioned for general transcription, while Flux is tuned for voice agent interactions.
Nova-3 delivers a confirmed 5.26% WER on general English batch transcription. Its architecture uses what Deepgram describes as a "simplified architectural approach" with a multi-stage training process. For streaming workloads, the Flux model is tuned for voice agent interactions. The Voice Agent API bundles STT, TTS, and LLM orchestration with predictable pricing. That makes it something infrastructure teams can use to build voice applications without stitching together separate components. You can get started free to test against your own audio before making architecture decisions.
FAQ
Bottom line: architecture choice determines how your speech system fails, and runtime adaptation helps most when the problem is vocabulary rather than acoustics.
What Is the Difference Between Speech Recognition and Traditional ASR?
Traditional ASR chains three models: acoustic, pronunciation, and language. Each requires separate training and maintenance. Single-model systems use one model trained on audio-text pairs directly. The practical difference is fewer integration points to debug, but less ability to fix individual components when accuracy drops.
Which Model Architecture Is Best for Real-Time Voice Agents?
Transducer (RNN-T) is the standard choice. It streams natively and conditions predictions on previously emitted tokens. CTC also streams but lacks that context awareness. Attention encoder-decoders require architectural modifications to stream and add latency even with those changes.
How Does Keyterm Prompting Affect WER on Domain-Specific Vocabulary?
Keyterm Prompting can improve recall on prompted terms. WER improvement depends on how many errors in your audio come from vocabulary misses versus acoustic problems. If most errors are OOV-related, the WER impact can be significant. If they're acoustic, you'll need model customization instead.
Can Speech Recognition Handle Multiple Speakers?
Yes, through speaker diarization layered on top of the base transcription. Diarization labels which speaker said what, with timestamp attribution. It's available in both streaming and batch modes, though accuracy depends on speaker overlap and audio quality.
What Audio Format Produces the Best Results?
Use linear16 (16-bit PCM) at the native sample rate of your recording device. Don't upsample telephony audio from 8kHz to 16kHz because it creates artifacts that degrade accuracy. Don't apply denoising preprocessing. Match the encoding, sample rate, and channel count to what you specify at connection time.

