Conversational AI
- Custom Dashboard Charts: The Conversational AI Dashboard can now be extended with custom charts displaying the results of evaluation criteria over time. See the new GET and PATCH endpoints for managing dashboard settings.
- Call History Filtering: Added the ability to filter the call history by start date using the new
call_start_before_unix
parameter in the List Conversations endpoint. Try it here. - Server Tools: Added option of making PUT requests in server tools
- Transfer to human: Added call forwarding functionality to support forwarding to operators, see docs here
- Language detection: Fixed an issue where the language detection system tool would trigger on a user replying yes in non-English language.
Usage Analytics
- Custom Aggregation: Added an optional
aggregation_interval
parameter to the Get Usage Metrics endpoint to control the interval over which to aggregate character usage (hour, day, week, month, or cumulative). - New Metric Breakdowns: The Usage Analytics section now supports additional metric breakdowns including
minutes_used
,request_count
,ttfb_avg
, andttfb_p95
, selectable via the newmetric
parameter in the Get Usage Metrics endpoint. Furthermore, you can now get a breakdown and filter byrequest_queue
.
API
View API changes
New Endpoints
- Added 2 new endpoints for managing Conversational AI dashboard settings:
- Get Dashboard Settings - Retrieves custom chart configurations for the ConvAI dashboard.
- Update Dashboard Settings - Updates custom chart configurations for the ConvAI dashboard.
Updated Endpoints
Audio Generation (TTS, S2S, SFX, Voice Design)
- Updated endpoints to support new
output_format
optionpcm_48000
:- Text to Speech (
POST /v1/text-to-speech/{voice_id}
) - Text to Speech with Timestamps (
POST /v1/text-to-speech/{voice_id}/with-timestamps
) - Text to Speech Stream (
POST /v1/text-to-speech/{voice_id}/stream
) - Text to Speech Stream with Timestamps (
POST /v1/text-to-speech/{voice_id}/stream/with-timestamps
) - Speech to Speech (
POST /v1/speech-to-speech/{voice_id}
) - Speech to Speech Stream (
POST /v1/speech-to-speech/{voice_id}/stream
) - Sound Generation (
POST /v1/sound-generation
) - Create Voice Previews (
POST /v1/text-to-voice/create-previews
)
- Text to Speech (
Usage Analytics
- Updated usage metrics endpoint:
- Get Usage Metrics (
GET /v1/usage/character-stats
) - Added optionalaggregation_interval
andmetric
query parameters.
- Get Usage Metrics (
Conversational AI
- Updated conversation listing endpoint:
- List Conversations (
GET /v1/convai/conversations
) - Added optionalcall_start_before_unix
query parameter for filtering by start date.
- List Conversations (
Schema Changes
Conversational AI
- Added detailed LLM usage and pricing information to conversation charging and history models.
- Added
tool_latency_secs
to tool result schemas - Added
access_info
toGET /v1/convai/agents/{agent_id}
Professional Voice Cloning (PVC)
- PVC API: Introduced a comprehensive suite of API endpoints for managing Professional Voice Clones (PVC). You can now programmatically create voices, add/manage/delete audio samples, retrieve audio/waveforms, manage speaker separation, handle verification, and initiate training. For a full list of new endpoints check the API changes summary below or read the PVC API reference here.
Speech to Text
- Enhanced Export Options: Added options to include or exclude timestamps and speaker IDs when exporting Speech to Text results in segmented JSON format via the API.
Conversational AI
- New LLM Models: Added support for new GPT-4.1 models:
gpt-4.1
,gpt-4.1-mini
, andgpt-4.1-nano
here - VAD Score: Added a new client event which sends VAD scores to the client, see reference here
Workspace
- Member Management: Added a new API endpoint to allow administrators to delete workspace members here
API
View API changes
New Endpoints
- Added 16 new endpoints:
- Delete Member - Allows deleting workspace members.
- Create PVC Voice - Creates a new PVC voice.
- Edit PVC Voice - Edits PVC voice metadata.
- Add Samples To PVC Voice - Adds audio samples to a PVC voice.
- Update PVC Voice Sample - Updates a PVC voice sample (noise removal, speaker selection, trimming).
- Delete PVC Voice Sample - Deletes a sample from a PVC voice.
- Retrieve Voice Sample Audio - Retrieves audio for a PVC voice sample.
- Retrieve Voice Sample Visual Waveform - Retrieves the visual waveform for a PVC voice sample.
- Retrieve Speaker Separation Status - Gets the status of speaker separation for a sample.
- Start Speaker Separation - Initiates speaker separation for a sample.
- Retrieve Separated Speaker Audio - Retrieves audio for a specific separated speaker.
- Get PVC Voice Captcha - Gets the captcha for PVC voice verification.
- Verify PVC Voice Captcha - Submits captcha verification for a PVC voice.
- Run PVC Training - Starts the training process for a PVC voice.
- Request Manual Verification - Requests manual verification for a PVC voice.
Updated Endpoints
Speech to Text
- Updated endpoint with changes:
- Create Forced Alignment Task - Added
enabled_spooled_file
parameter to allow streaming large files (POST /v1/forced-alignment
).
- Create Forced Alignment Task - Added
Schema Changes
Conversational AI
GET conversation details
: Addedhas_audio
,has_user_audio
,has_response_audio
boolean fields here
Dubbing
GET dubbing resource
: Addedstatus
field to each render here
Voices
- New PVC flow: Added new flow for Professional Voice Clone creation, try it out here
Conversational AI
- Agent-agent transfer: Added support for agent-to-agent transfers via a new system tool, enabling more complex conversational flows. See the Agent Transfer tool documentation for details.
- Enhanced tool debugging: Improved how tool execution details are displayed in the conversation history for easier debugging.
- Language detection fix: Resolved an issue regarding the forced calling of the language detection tool.
Dubbing
- Render endpoint: Introduced a new endpoint to regenerate audio or video renders for specific languages within a dubbing project. This automatically handles missing transcriptions or translations. See the Render Dub endpoint.
- Increased size limit: Raised the maximum allowed file size for dubbing projects to 1 GiB.
API
View API changes
New Endpoints
- Added render dub endpoint - Regenerate dubs for a specific language.
Updated Endpoints
Pronunciation Dictionaries
- Updated the response for the
GET /v1/pronunciation-dictionaries/{pronunciation_dictionary_id}/
endpoint and related components to include thepermission_on_resource
field.
Speech to Text
- Updated Speech to Text endpoint (
POST /v1/speech-to-text
):- Added
cloud_storage_url
parameter to allow transcription directly from public S3 or GCS URLs (up to 2GB). - Made the
file
parameter optional; exactly one offile
orcloud_storage_url
must now be provided.
- Added
Speech to Speech
- Added optional
file_format
parameter (pcm_s16le_16
orother
) for lower latency with PCM input toPOST /v1/speech-to-speech/{voice_id}
Conversational AI
- Updated components to support agent-agent transfer tool
Voices
- Updated
GET /v1/voices/{voice_id}
samples
field to include optionaltrim_start
andtrim_end
parameters.
AudioNative
- Updated
Get /v1/audio-native/{project_id}/settings
to includestatus
field (processing
orready
).
Speech to text
scribe_v1_experimental
: Launched a new experimental preview of the Scribe v1 model with improvements including improved performance on audio files with multiple languages, reduced hallucinations when audio is interleaved with silence, and improved audio tags. The new model is available via the API under the model namescribe_v1_experimental
Text to speech
- A-law format support: Added a-law format with 8kHz sample rate to enable integration with European telephony systems.
- Fixed quota issues: Fixed a database bug that caused some requests to be mistakenly rejected as exceeding their quota.
Conversational AI
- Document type filtering: Added support for filtering knowledge base documents by their type (file, URL, or text).
- Non-audio agents: Added support for conversational agents that don’t output audio but still send response transcripts and can use tools. Non-audio agents can be enabled by removing the audio client event.
- Improved agent templates: Updated all agent templates with enhanced configurations and prompts. See more about how to improve system prompts here.
- Fixed stuck exports: Fixed an issue that caused exports to be stuck for extended periods.
Studio
- Fixed volume normalization: Fixed issue with streaming project snapshots when volume normalization is enabled.
New API endpoints
- Forced alignment: Added new forced alignment endpoint for aligning audio with text, perfect for subtitle generation.
- Batch calling: Added batch calling endpoint for scheduling calls to multiple recipients
API
View API changes
New Endpoints
- Added Forced alignment endpoint for aligning audio with text
- Added dedicated endpoints for knowledge base document types:
Updated Endpoints
Text to Speech
- Added a-law format (8kHz) to all audio endpoints:
Voices
- Get voices - Added
collection_id
parameter for filtering voices by collection
Knowledge Base
- Get knowledge base - Added
types
parameter for filtering documents by type - General endpoint for creating knowledge base documents marked as deprecated in favor of specialized endpoints
User Subscription
- Get user subscription - Added
professional_voice_slots_used
property to track number of professional voices used in a workspace
Conversational AI
- Added
silence_end_call_timeout
parameter to set maximum wait time before terminating a call - Removed
/v1/convai/agents/{agent_id}/add-secret
endpoint (now handled by workspace secrets endpoints)
Text to speech
- Opus format support: Added support for Opus format with 48kHz sample rate across multiple bitrates (32-192 kbps).
- Improved websocket error handling: Updated TTS websocket API to return more accurate error codes (1011 for internal errors instead of 1008) for better error identification and SLA monitoring.
Conversational AI
- Twilio outbound: Added ability to natively run outbound calls.
- Post-call webhook override: Added ability to override post-call webhook settings at the agent level, providing more flexible configurations.
- Large knowledge base document viewing: Enhanced the knowledge base interface to allow viewing the entire content of large RAG documents.
- Added call SID dynamic variable: Added
system__call_sid
as a system dynamic variable to allow referencing the call ID in prompts and tools.
Studio
- Actor Mode: Added Actor Mode in Studio, allowing you to use your own voice recordings to direct the way speech should sound in Studio projects.
- Improved keyboard shortcuts: Updated keyboard shortcuts for viewing settings and editor shortcuts to avoid conflicts and simplified shortcuts for locking paragraphs.
Dubbing
- Dubbing duplication: Made dubbing duplication feature available to all users.
- Manual mode foreground generation: Added ability to generate foreground audio when using manual mode with a file and CSV.
Voices
- Enhanced voice collections: Improved voice collections with visual upgrades, language-based filtering, navigation breadcrumbs, collection images, and mouse dragging for carousel navigation.
- Locale filtering: Added locale parameter to shared voices endpoint for more precise voice filtering.
API
View API changes
Updated Endpoints
Text to Speech
- Updated Text to Speech endpoints:
- Convert text to speech - Added
apply_language_text_normalization
parameter for improved text pronunciation in supported languages (currently Japanese) - Stream text to speech - Added
apply_language_text_normalization
- Convert with timestamps - Added
apply_language_text_normalization
- Stream with timestamps - Added
apply_language_text_normalization
- Convert text to speech - Added
Audio Format
- Added Opus format support to multiple endpoints:
- Text to speech - Added support for Opus format with 48kHz sample rate at multiple bitrates (32, 64, 96, 128, 192 kbps)
- Stream text to speech - Added Opus format options
- Convert with timestamps - Added Opus format options
- Stream with timestamps - Added Opus format options
- Speech to speech - Added Opus format options
- Stream speech to speech - Added Opus format options
- Create voice previews - Added Opus format options
- Sound generation - Added Opus format options
Conversational AI
- Updated Conversational AI endpoints:
- Delete agent - Changed success response code from 200 to 204
- Updated RAG embedding model options - replaced
gte_Qwen2_15B_instruct
withmultilingual_e5_large_instruct
Voices
- Updated Voice endpoints:
- Get shared voices - Added locale parameter for filtering voices by language region
Dubbing
- Updated Dubbing endpoint:
- Dub a video or audio file - Renamed beta feature
use_replacement_voices_from_library
parameter todisable_voice_cloning
for clarity
- Dub a video or audio file - Renamed beta feature
Voices
- List Voices V2: Added a new V2 voice search endpoint with better search and additional filtering options
Conversational AI
- Native outbound calling: Added native outbound calling for Twilio-configured numbers, eliminating the need for complex setup configurations. Outbound calls are now visible in the Call History page.
- Automatic language detection: Added new system tool for automatic language detection that enables agents to switch languages based on both explicit user requests (“Let’s talk in Spanish”) and implicit language in user audio.
- Pronunciation dictionary improvements: Fixed phoneme tags in pronunciation dictionaries to work correctly with conversational AI.
- Large RAG document viewing: Added ability to view the entire content of large RAG documents in the knowledge base.
- Customizable widget controls: Updated UI to include an optional mute microphone button and made widget icons customizable via slots.
Sound Effects
- Fractional duration support: Fixed an issue where users couldn’t enter fractional values (like 0.5 seconds) for sound effect generation duration.
Speech to Text
- Repetition handling: Improved detection and handling of repetitions in speech-to-text processing.
Studio
- Reader publishing fixes: Added support for mp3_44100_192 output format (high quality) so users below Publisher tier can export audio to Reader.
Mobile
- Core app signup: Added signup endpoints for the new Core mobile app.
API
View API changes
New Endpoints
- Added 5 new endpoints:
- List voices (v2) - Enhanced voice search capabilities with additional filtering options
- Initiate outbound call - New endpoint for making outbound calls via Twilio integration
- Add pronunciation dictionary from rules - Create pronunciation dictionaries directly from rules without file upload
- Get knowledge base document content - Retrieve full document content from the knowledge base
- Get knowledge base document chunk - Retrieve specific chunks from knowledge base documents
Updated Endpoints
Conversational AI
- Updated Conversational AI endpoints:
- Create agent - Added
mic_muting_enabled
property for UI control andworkspace_overrides
property for workspace-specific configurations - Update agent - Added
workspace_overrides
property for customizing agent behavior per workspace - Get agent - Added
workspace_overrides
property to the response - Get widget - Added
mic_muting_enabled
property for controlling microphone muting in the widget UI - Get conversation - Added rag information to view knowledge base content used during conversations
- Create phone number - Replaced generic structure with specific twilio phone number and sip trunk options
- Compute RAG index - Removed
force_reindex
query parameter for more controlled indexing - List knowledge base documents - Changed response structure to support different document types
- Get knowledge base document - Modified to return different response models based on document type
- Create agent - Added
Text to Speech
- Updated Text to Speech endpoints:
- Convert text to speech - Made properties optional, including
stability
andsimilarity
settings - Stream text to speech - Made voice settings properties optional for more flexible streaming requests
- Convert with timestamps - Made settings optional and modified
pronunciation_dictionary_locators
property - Stream with timestamps - Made voice settings properties optional for more flexible requests
- Convert text to speech - Made properties optional, including
Speech to Text
- Updated Speech to Text endpoint:
- Convert speech to text - Removed
biased_keywords
property from form data and improved internal repetition detection algorithm
- Convert speech to text - Removed
Voice Management
- Updated Voice endpoints:
- Get voices - Updated voice settings properties in the response
- Get default voice settings - Made
stability
andsimilarity
properties optional - Get voice settings - Made numeric properties optional for more flexible configuration
- Edit voice settings - Made
stability
andsimilarity
settings optional - Create voice - Modified array properties to accept null values
- Create voice from preview - Updated voice settings model with optional properties
Studio
- Updated Studio endpoints:
- Get project - Added
version_rules_num
to project metadata - Get project snapshot - Removed
status
property - Create pronunciation dictionaries - Modified
pronunciation_dictionary_locators
property and string properties to accept null values
- Get project - Added
Pronunciation Dictionary
- Updated Pronunciation Dictionary endpoints:
- Get all pronunciation dictionaries - Added
sort
andsort_direction
query parameters, pluslatest_version_rules_num
andinteger
properties to response - Get pronunciation dictionary - Added
latest_version_rules_num
andinteger
properties to response - Add from file - Added
version_rules_num
property to response for tracking rules quantity - Add rules - Added
version_rules_num
to response for rules tracking - Remove rules - Added
version_rules_num
to response for rules tracking
- Get all pronunciation dictionaries - Added
Conversational AI
- Default LLM update: Changed the default agent LLM from Gemini 1.5 Flash to Gemini 2.0 Flash for improved performance.
- Fixed incorrect conversation abandons: Improved detection of conversation continuations, preventing premature abandons when users repeat themselves.
- Twilio information in history: Added Twilio call details to conversation history for better tracking.
- Knowledge base redesign: Redesigned the knowledge base interface.
- System dynamic variables: Added system dynamic variables to use time, conversation id, caller id and other system values as dynamic variables in prompts and tools.
- Twilio client initialisation: Adds an agent level override for conversation initiation client data twilio webhook.
- RAG chunks in history: Added retrieved chunks by RAG to the call transcripts in the history view.
Speech to Text
- Reduced pricing: Reduced the pricing of our Scribe model, see more here.
- Improved VAD detection: Enhanced Voice Activity Detection with better pause detection at segment boundaries and improved handling of silent segments.
- Enhanced diarization: Improved speaker clustering with a better ECAPA model, symmetric connectivity matrix, and more selective speaker embedding generation.
- Fixed ASR bugs: Resolved issues with VAD rounding, silence and clustering that affected transcription accuracy.
Studio
- Disable publishing UI: Added ability to disable the publishing interface for specific workspace members to support enterprise workflows.
- Snapshot API improvement: Modified endpoints for project and chapter snapshots to return an empty list instead of throwing errors when snapshots can’t be downloaded.
- Disabled auto-moderation: Turned off automatic moderation based on Text to Speech generations in Studio.
Workspaces
- Fixed API key editing: Resolved an issue where editing workspace API keys would reset character limits to zero, causing the keys to stop working.
- Optimized free subscriptions: Fixed an issue with refreshing free subscription character limits,
API
View API changes
New Endpoints
- Added 3 new endpoints:
Updated Endpoints
Dubbing
- Updated Dubbing endpoints:
- Dub a video or audio file - Added
use_replacement_voices_from_library
property and madesource_path
,target_language
,source_language
nullable - Resource dubbing - Made
language_codes
array nullable - Add language to dubbing resource - Made
language_code
nullable - Add speaker segment - Made
text
nullable - Translate dubbing resource - Made
target_languages
array nullable - Update dubbing segment - Made
start_time
andend_time
nullable
- Dub a video or audio file - Added
Project Management
- Updated Project endpoints:
- Add project - Made
metadata
,project_name
,description
nullable - Create podcast - Made
title
,description
,author
nullable - Get project - Made
last_modified_at
,created_at
,project_name
nullable - Add chapter - Made
chapter_id
,word_count
,statistics
nullable - Update chapter - Made
content
andblocks
properties nullable
- Add project - Made
Conversational AI
- Updated Conversational AI endpoints:
- Update agent - Made
conversation_config
,platform_settings
nullable and addedworkspace_overrides
property - Create agent - Made
agent_name
,prompt
,widget_config
nullable and addedworkspace_overrides
property - Add to knowledge base - Made
document_name
nullable - Get conversation - Added
twilio_call_data
model and madetranscript
,metadata
nullable
- Update agent - Made
Text to Speech
- Updated Text to Speech endpoints:
- Convert text to speech - Made
voice_settings
,text_input
nullable and deprecateduse_pvc_as_ivc
property - Stream text to speech - Made
voice_settings
,text_input
nullable and deprecateduse_pvc_as_ivc
property - Convert with timestamps - Made
character_alignment
andword_alignment
nullable
- Convert text to speech - Made
Voice Management
- Updated Voice endpoints:
- Create voice previews - Added
loudness
,quality
,guidance_scale
properties - Create voice from preview - Added
speaker_separation
properties and madevoice_id
,name
,labels
nullable - Get voice - Added
speaker_boost
,speaker_clarity
,speaker_isolation
properties
- Create voice previews - Added
Speech to Text
- Updated Speech to Text endpoint:
- Convert speech to text - Added
biased_keywords
property
- Convert speech to text - Added
Other Updates
- Download history - Added application/zip content type and 400 response
- Add pronunciation dictionary from file - Made
dictionary_name
anddescription
nullable
Conversational AI
- HIPAA compliance: Conversational AI is now HIPAA compliant on appropriate plans, when a BAA is signed, zero-retention mode is enabled and appropriate LLMs are used. For access please contact sales
- Cascade LLM: Added dynamic dispatch during the LLM step to other LLMs if your default LLM fails. This results in higher latency but prevents the turn failing.
- Better error messages: Added better error messages for websocket failures.
- Audio toggling: Added ability to select only user or agent audio in the conversation playback.
Scribe
- HIPAA compliance: Added a zero retention mode to Scribe to be HIPAA compliant.
- Diarization: Increased time length of audio files that can be transcribed with diarization from 8 minutes to 2 hours.
- Cheaper pricing: Updated Scribe’s pricing to be cheaper, as low as $0.22 per hour for the Business tier.
- Memory usage: Shipped improvements to Scribe’s memory usage.
- Fixed timestamps: Fixed an issue that was causing incorrect timestamps to be returned.
- Biased keywords: Added biased keywords to improve Scribe’s performance.
Text to Speech
- Pronunciation dictionaries: Fixed pronunciation dictionary rule application for replacements that contain symbols.
Dubbing
- Studio support: Added support for creating dubs with
dubbing_studio
enabled, allowing for more advanced dubbing workflows beyond one-off dubs.
Voices
- Verification: Fixed an issue where users on probation could not verify their voice clone.
API
View API changes
New Endpoints
- Added 7 new endpoints:
Updated Endpoints
Studio Projects
- Updated Studio project endpoints to add
source_type
property and deprecatequality_check_on
andquality_check_on_when_bulk_convert
properties:
Voice Management
- Updated Voice endpoints with several property changes:
- Get voice - Made several properties optional and added
preview_url
- Create voice - Made several properties optional and added
preview_url
- Create voice from preview - Made several properties optional and added
preview_url
- Get similar voices - Made
language
,description
,preview_url
, andrate
properties optional
- Get voice - Made several properties optional and added
Conversational AI
- Updated Conversational AI agent endpoints:
- Update agent - Modified
conversation_config
,agent
,platform_settings
, andwidget
properties - Create agent - Modified
conversation_config
,agent
,prompt
, platform_settings, widget properties and addedshareable_page_show_terms
- Get agent - Modified
conversation_config
,agent
,platform_settings
, andwidget
properties - Get widget - Modified
widget_config
property and addedshareable_page_show_terms
- Update agent - Modified
Knowledge Base
- Updated Knowledge Base endpoints to add metadata property:
Other Updates
- Dub a video or audio file - Added
dubbing_studio
property - Convert text to sound effects - Added
output_format
query parameter - Convert speech to text - Added
enable_logging
query parameter - Get secrets - Modified
secrets
andused_by
properties - Get all pronunciation dictionaries - Made
next_cursor
property optional
Removed Endpoints
-
Temporarily removed Conversational AI tools endpoints:
- Get tool
- List tools
- Update tool
- Create tool
- Delete tool
Dubbing
- Scribe for speech recognition: Dubbing Studio now uses Scribe by default for speech recognition to improve accuracy.
Speech to Text
- Fixes: Shipped several fixes improving the stability of Speech to Text.
Conversational AI
- Speed control: Added speed control to an agent’s settings in Conversational AI.
- Post call webhook: Added the option of sending post-call webhooks after conversations are completed.
- Improved error messages: Added better error messages to the Conversational AI websocket.
- Claude 3.7 Sonnet: Added Claude 3.7 Sonnet as a new LLM option in Conversational AI.
API
View API changes
New Endpoints
- Added new Dubbing resource management endpoints:
- for adding languages to dubs
- for retrieving dubbing resources
- for creating segments
- for modifying segments
- for removing segments
- for dubbing segments
- for transcribing segments
- for translating segments
- Added Knowledge Base RAG indexing endpoint
- Added Studio snapshot retrieval endpoints for projects and chapters
Updated Endpoints
- Added
prompt_injectable
property to knowledge base endpoints - Added
name
property to Knowledge Base document creation and retrieval endpoints: - Added
speed
property to agent creation - Removed
secrets
property from agent endpoints (now handled by dedicated secrets endpoints) - Added secret deletion endpoint for removing secrets
- Removed
secrets
property from settings endpoints
Speech to Text
- ElevenLabs launched a new state of the art Speech to Text API available in 99 languages.
Text to Speech
- Speed control: Added speed control to the Text to Speech API.
Studio
- Auto-assigned projects: Increased token limits for auto-assigned projects from 1 month to 3 months worth of tokens, addressing user feedback about working on longer projects.
- Language detection: Added automatic language detection when generating audio for the first time, with suggestions to switch to Eleven Turbo v2.5 for languages not supported by Multilingual v2 (Hungarian, Norwegian, Vietnamese).
- Project export: Enhanced project exporting in ElevenReader with better metadata tracking.
Dubbing
- Clip overlap prevention: Added automatic trimming of overlapping clips in dubbing jobs to ensure clean audio tracks for each speaker and language.
Voice Management
- Instant Voice Cloning: Improved preview generation for Instant Voice Cloning v2, making previews available immediately.
Conversational AI
- Agent ownership: Added display of agent creators in the agent list, improving visibility and management of shared agents.
Web app
- Dark mode: Added dark mode to the web app.
API
View API changes
- Launched /v1/speech-to-text endpoint
- Added
agents.level
property to Conversational AI agents endpoint - Added
platform_settings
to Conversational AI agent endpoint - Added
expandable
variant towidget_config
, with configuration optionsshow_avatar_when_collapsed
anddisable_banner
to Conversational AI agent widget endpoint - Added
webhooks
property andused_by
tosecrets
to secrets endpoint - Added
verified_languages
to voices endpoint - Added
speed
property to voice settings endpoints - Added
verified_languages
,is_added_by_user
tovoices
andmin_notice_period_days
query parameter to shared voices endpoint - Added
verified_languages
,is_added_by_user
tovoices
in similar voices endpoint - Added
search
,show_only_owned_documents
,use_typesense
query parameters to knowledge base endpoint - Added
used_by
to Conversation AI secrets endpoint - Added
invalidate_affected_text
property to Studio pronunciation dictionaries endpoint