Big Story

Nicholas Leonard, CEO of VoiceRun, on Why Reliability Defines Enterprise Voice AI

The first thing most teams evaluate in voice AI is how natural the system sounds.  They look at latency, tone, and whether the interaction feels close enough to a human conversation to replace a live operator. That is a useful starting point, but it is also misleading. In production environments, conversation quality is rarely the deciding factor. What determines whether a voice system survives is its ability to consistently complete real tasks under messy, unpredictable conditions. 

The real test begins the moment the system is asked to do something that affects the business. Can it authenticate a user across multiple verification steps without looping or failing? Can it process a payment when the gateway is slow or partially unavailable? Can it update a reservation while syncing across multiple backend systems? Can it detect failure states and recover without forcing a handoff? If the answer to any of these is no, the system behaves like an advanced IVR. It may sound intelligent, but it breaks at the exact point where automation is supposed to create value.

This is why reliability matters. Users are far more tolerant of imperfect phrasing than they are of failed execution. A slightly robotic tone is acceptable. A transaction that fails midway is not. A retry loop during authentication immediately erodes trust. A confirmed booking that is incorrect creates downstream operational issues.

Technically, this is where most voice systems fall apart. Each interaction is not just a response, but a pipeline that must run in real time. Audio comes in and must be cleaned, segmented, and converted to text. The system must determine intent, often using a mix of deterministic logic and LLM-based reasoning. It may trigger API calls, query internal systems, or update state. It then generates a response, converts it back to speech, and returns it to the user, all while managing interruptions, latency constraints, and partial failures. This entire loop must operate reliably across every turn of the conversation.

Most teams underestimate this because they optimize for the demo environment. In a controlled setup, the agent handles a narrow set of flows with predictable inputs and clean integrations. In production, the system is exposed to real variability. Customers interrupt mid-sentence. Background noise affects transcription. APIs return incomplete data. Authentication flows branch unexpectedly. Systems of record introduce latency or inconsistency. 

This is also why we built VoiceRun as a code-first platform. Enterprise teams need direct control over how the system behaves when something goes wrong. If a payment provider times out, a customer interrupts mid-flow, or a backend system returns partial data, the recovery path must be explicit. 

I often compare this to the web. Companies spend years optimizing checkout flows because they know that one failed transaction immediately affects revenue. Voice should be treated the same way. If payments, bookings, or account actions are happening over the phone, the enterprise needs the same level of visibility and control.

The key mistake in evaluating voice AI platforms is focusing on how impressive the demo feels. A smooth conversation in a controlled setting does not reflect the realities of production. The more useful signal is whether the system can handle degraded conditions, edge cases, and integration complexity while still completing the task. That includes how it manages retries, handles partial failures, maintains state across turns, and integrates with existing enterprise infrastructure. Connect with Nicholas to stay up to date on all things voice AI

Market Pulse

  • xAI released standalone Grok STT and TTS APIs, built on the same speech stack used across Grok Voice, Tesla vehicles, and Starlink support systems. Grok STT supports 25+ languages with batch and streaming modes, speaker diarization, word-level timestamps, and inverse text normalization for numbers, dates, and currencies. On phone-call entity recognition benchmarks, xAI reports a 5.0% error rate, compared with ElevenLabs at 12.0%, and AssemblyAI at 21.3%. Grok TTS supports 20+ languages with inline speech controls for pauses, tone shifts, and expressive delivery.

  • VoiceRun recently sponsored two developer hackathons, Enterprise Agent Jam NYC and Boston 311 Hack, both centered on building real-world AI systems. The NYC event brought together developers and enterprise teams to prototype production-ready agent workflows, while the Boston hackathon drew founders, engineers, and public-sector collaborators working on civic applications. Across both gatherings, the focus stayed firmly on execution, with teams shipping working systems under real constraints. VoiceRun’s presence, including opening remarks by CEO Nicholas Leonard, reinforces its position within the core builder community and advances voice and agent infrastructure.

  • Deepgram moved Flux Multilingual (flux-general-multi) to general availability. A single model now supports English, Spanish, French, German, Hindi, Russian, Portuguese, Japanese, Italian, and Dutch with native code-switching and integrated turn detection. The optional language_hint parameter biases recognition without requiring separate model endpoints, and hints can be updated mid-stream without reconnecting the session. This removes per-language routing logic, simplifying multilingual telephony deployments.  

  • Deepgram also released the Deepgram CLI (dg), a unified terminal interface for transcription, speech synthesis, text intelligence, account management, and MCP server integration for AI coding tools. Instead of separate SDK setups across services, teams can test, deploy, and debug voice pipelines from a single command layer, reducing operational overhead during development and production support.

  • Google released Gemini 3.1 Flash TTS in preview with support for 70+ languages, native multi-speaker dialogue generation, and prompt-based style control where developers define pacing, tone, and emphasis in natural language rather than SSML-style parameter tuning. SynthID watermarking is embedded across generated audio, and deployment is available through Gemini API, Google AI Studio, and Vertex AI.

  • Maven introduced a PCI-isolated payment session layer for voice agents. Instead of collecting card data within the agent pipeline, the system creates a separate payment session, transfers the caller to a compliant collection process via voice or DTMF, and returns the user to the original workflow upon completion. Card data never touches the application layer. The system integrates with Vapi, Retell, LiveKit, and Twilio, while supporting Stripe, Authorize.net, Adyen, and Braintree on the payment side.

Unicorn Club is a weekly newsletter for product builders who care about better interfaces. Every Wednesday, Adam curates practical reads across interface craft, team standards, and shipping habits - the sort of stuff you can actually take back into the work that week.

Resources & Events

📅02 ElevenLabs Summit (Warsaw, Poland - June 1, 2026)

The ElevenLabs Summit will take place in Warsaw at Teatr Wielki (Polish National Opera) on June 1. The event will spotlight ElevenLabs’ Poland team and regional innovation in AI voice and agents, showcasing work across Poland, Central and Eastern Europe, and beyond. Details →

📅 2026 Voice AI Symposium and Hackathon (St. Petersburg, FL - May 4-6, 2026)

The 2026 Voice AI Symposium & Hackathon is a global event bringing together researchers, clinicians, and industry leaders to advance the use of voice biomarkers in healthcare. The symposium emphasizes translating AI research into real-world clinical applications through hands-on workshops, research presentations, demos, and startup pitches. Unlike large tech conferences, it is designed to be intimate and collaborative, with a strong focus on practical implementation using shared datasets and tools. Details →

📊 Report Spotlight: AI Index Report 2026 (Stanford HAI)

The latest report from Stanford Institute for Human-Centered Artificial Intelligence shows that while AI adoption continues to grow across industries, most organizations still struggle to move from experimentation to reliable production systems. The gap is driven by integration complexity, governance, and operational readiness. The biggest shift is the move from single-modality systems to multimodal models that natively handle speech alongside text and vision, which is what enables modern voice agents to move beyond IVR-style interactions. Read

For the Commute

Enterprise Voice AI That Actually Works (What's Up with Tech?)

This podcast episode features Fred Fontes discussing why enterprise voice AI is finally becoming usable in regulated industries like banking. The conversation focuses on what changed beyond the models, including the need for controllable systems, strong guardrails, and data sovereignty to meet compliance and security requirements. It highlights practical outcomes such as improved collections performance, faster experimentation through A/B testing, and lower interaction costs, while making clear that the biggest remaining challenge is integrating these systems into existing enterprise workflows and systems of record.

Poll

Keep Reading