Comparing ElevenLabs Conversational AI and OpenAI Realtime API

Comparing two recent product launches to help you find the best product for your use case

Updated as of October 18th, 2024

There were two major product launches in the world of Conversational AI in the last month - our Conversational AI orchestration platform and OpenAI's Realtime API. We put together this post to help you distinguish between the two and figure out which is best for your use case.

Overview

Both of these products are designed to help you create realtime, conversational voice agents. ElevenLabs Conversational AI makes that possible through an orchestration platform that creates a transcript from speech using Speech to Text, sends that transcript to an LLM of your choice along with a custom knowledge base, and then voices the LLM response using Text to Speech. It's an end to end solution including monitoring and analytics on past calls and will soon offer a testing framework and phone integrations. 

OpenAI's Realtime API is built on a different architecture whereby the model takes audio (speech) as input and provides audio (speech) directly as the output. There is no step by which audio is converted into a written transcript and passed to an LLM, which likely provides latency gains. It’s only available via API and is not an end to end platform. 
FeatureElevenLabs Conv AIOpenAI Realtime
Total Number of Voices3k+6
LLMs SupportedBring your own server or choose from any leading providerOpenAI models only
Call tracking and analyticsYes, built-in dashboardNo, must build using API
Latency1-3 seconds depending on network latency and size of knowledge baseLikely faster due to no transcription step
Price10 cents per minute on business, as low as 2-3 cents per minute on Enterprise with high volume (+LLM cost) 6 cents per minute input, 24 cents per minute output
Voice CloningYes, bring your own voice with a PVCNo voice cloning
API AccessYes, all plansYes, all plans

How they stack up

Understanding Emotion & Pronunciation

When our Conversational AI converts speech into text, some information is lost, including the emotion, tone and pronunciation of the speech. Since OpenAI's Realtime API goes directly from speech to speech, no context is lost. This makes it more adept for certain use cases like correcting someone's pronunciation when learning a new language or identifying and responding to emotion in therapy.

Flexibility

When using the Realtime API, you are using OpenAI's infrastructure for the full conversational experience. It's not possible to integrate another company's LLM, or to bring your own, as the Realtime API only takes audio as input and returns audio as output.

With our Conversational AI platform, you can change the LLM powering your model at any time (including using OpenAI's models). As Anthropic, OpenAI, Google, NVIDIA, and others continue to one up each other in the race to have the most performant LLM, you can update at any time so you are always using state of the art technology.

And for companies that have built their own in house fine-tuned LLM, either for performance or privacy reasons, it's possible to integrate that with ElevenLab’s Conversational AI platform but not with OpenAI’s Realtime API.

Latency

When evaluating any model for latency, there are two important factors to consider

(1) Is the average latency low enough to create a seamless user experience?

(2) How much does latency fluctuate and what does the user experience look like for P90 and P99 latency?

One potential benefit of the OpenAI Realtime API is that because it cuts out the intermediate step of turning speech into text, it is likely to have an overall lower latency.

One potential downside however comes back to the flexibility we discussed earlier. In our testing over the last few weeks, 40-mini was initially the lowest latency LLM to pair with our Conversational AI platform. This week, its latency more than doubled which led our users to switch to Gemini Flash 1.5. With the Realtime API, it's not possible to rotate to a faster LLM.

Also note that the end to end latency for your Conversational AI application will depend not just on your provider, but also on the size of your agent's knowledge base and your network conditions.

Voice Options

OpenAI's Realtime API currently has 6 voice options. Our voice library has over 3.000 voices. You can also use Professional Voice Cloning to use your own custom voice on our platform.  This means the Realtime API won't allow you to pick a voice unique to your brand or content.

Price

In the Realtime API, Audio input is priced at $100 per 1M tokens and output is $200 per 1M tokens. This equates to approximately $0.06 per minute of audio input and $0.24 per minute of audio output. 

ElevenLabs Conversational AI costs 1k credits per minute (+ LLM costs), which is 10 cents per minute (+LLM costs) on our business plan and as low as a few cents per minute for Enterprise customers with high call volumes.

Additional Platform Features

At the end of each call, the Realtime API sends JSON-formatted events containing text and audio chunks including the transcript and recordings of the call and any functional calls made. It's up to you to read, process, report on, and display that information in a way that is useful to your team.

Our platform has built-in functionality for evaluating the success of a call, extracting structure data, and displaying that along with the transcript, summary and recording within our dashboard for your team to review.

Explore more

ElevenLabs

Create with the highest quality AI Audio

Get started free

Already have an account? Log in