InfoQ Homepage News OpenAI’s gpt-realtime Enables Production-Ready Voice Agents with End-to-End Speech Processing
OpenAI’s gpt-realtime Enables Production-Ready Voice Agents with End-to-End Speech Processing
Sep 11, 2025 2 min read
by
Write for InfoQ
Feed your curiosity. Help 550k+ globalsenior developers
each month stay ahead.Get in touch
OpenAI has released gpt-realtime, its most advanced speech-to-speech model, alongside the general availability of the Realtime API. The updates aim to reduce latency, improve speech quality, and give developers stronger tools, such as MCP server support, image input, and Session Initiation Protocol (SIP) phone calling support, for building production-ready AI voice agents.
The combined Realtime API and gpt-realtime is designed to handle end-to-end speech processing within a single system, rather than chaining together separate speech-to-text and text-to-speech models. This architecture cuts response times while preserving nuance in delivery, a critical improvement for real-time agents where even small delays can break conversational flow.
The gpt-realtime was trained to produce higher-quality speech with more natural pacing and intonation, and to respond reliably to style instructions such as “speak empathetically” or “use a professional tone.” Two new synthetic voices, Cedar and Marin, are available, and existing voices have been updated for greater realism.
On comprehension benchmarks, gpt-realtime shows measurable improvements. It can track non-verbal cues, switch languages within a single sentence, and more accurately process alphanumeric sequences (such as phone numbers, VINs, etc) across languages, including Spanish, Chinese, Japanese, and French. Internal testing highlights this jump, with gpt-realtime reaching 82.8% accuracy on Big Bench Audio compared to 65.6% for the previous model. Instruction-following is also sharper, with MultiChallenge audio benchmark scores rising from 20.6% to 30.5%.
Function calling is another area of focus. The model now performs better at identifying relevant functions, calling them at the right time, and supplying the correct arguments. On ComplexFuncBench, accuracy rose to 66.5% from 49.7%. There were updates to asynchronous function calling, allowing the voice agent to continue the conversation while waiting for results, a feature with obvious value for customer support and transactional applications.
The Realtime API has been upgraded to align with production requirements. Developers can now connect remote MCP servers directly into a session, enabling tool calls without manual integration work. Image input is supported, allowing applications to ground conversations in visual context, such as screenshots or photos. SIP support makes it possible to integrate voice agents with existing telephony systems, including PBXs and desk phones. Reusable prompts simplify session management, while full EU data residency support addresses compliance concerns for European deployments.
According to the release notes, early enterprise partners are testing these capabilities in production-like scenarios. Zillow is piloting voice-driven home search, while T-Mobile is exploring customer service use cases where real-time adaptability is essential. Both companies highlight the shift from scripted automation to more flexible, domain-specific expertise delivered through AI agents.
OpenAI has also reinforced safeguards around deployment. The Realtime API incorporates classifiers that can terminate harmful conversations, and developers can add domain-specific guardrails via the Agents SDK. Preset voices in Realtime API are used to reduce impersonation risks.
Both gpt-realtime model and Realtime API are immediately available to all developers. To get started, developers can visit the Realtime API documentation and prompting guide, and test the new gpt-realtime demo in the Playground.
This content is in the AI, ML & Data Engineering topic
Related Topics:
-
Related Editorial
-
Related Sponsors
-
Popular across InfoQ
-
AWS Introduces ECS Managed Instances for Containerized Applications
-
Producing a Better Software Architecture with Residuality Theory
-
GitHub Introduces New Embedding Model to Improve Code Search and Context
-
Google DeepMind Introduces CodeMender, an AI Agent for Automated Code Repair
-
Building Distributed Event-Driven Architectures across Multi-Cloud Boundaries
-
Elena Samuylova on Large Language Model (LLM)-Based Application Evaluation and LLM as a Judge
-
Related Content
The InfoQ Newsletter
A round-up of last week’s content on InfoQ sent out every Tuesday. Join a community of over 250,000 senior developers. View an example