Empathic Voice Interface (EVI)
Hume’s Empathic Voice Interface (EVI) is the world’s first emotionally intelligent voice AI.
Hume’s Empathic Voice Interface (EVI) is the world’s first emotionally intelligent voice AI. It accepts live audio input and returns both generated audio and transcripts augmented with measures of vocal expression. By processing the tune, rhythm, and timbre of speech, EVI unlocks a variety of new capabilities, like knowing when to speak and generating more empathic language with the right tone of voice. These features enable smoother and more satisfying voice-based interactions between humans and AI, opening new possibilities for personal AI, customer service, accessibility, robotics, immersive gaming, VR experiences, and much more.
We provide a suite of tools to integrate and customize EVI for your application, including a WebSocket API that handles audio and text transport, a REST API for fetching Chat history and managing EVI configurations,and SDKs to simplify integration into web and Python-based projects. Additionally, we provide open-source examples and a web widget as practical starting points for developers to explore and implement EVI’s capabilities within their own projects.
Overview of EVI features
Building with EVI
The main way to work with EVI is through a WebSocket connection that sends audio and receives responses in real-time. This enables fluid, bidirectional dialogue where users speak, EVI listens and analyzes their expressions, and EVI generates emotionally intelligent responses.
EVI supports two authentication strategies. Learn more about them at the links below:
Both methods require specifying the chosen authentication strategy and providing the corresponding key in the request parameters of the EVI WebSocket endpoint. Learn more about Hume’s authentication strategies here.
You start a conversation by connecting to the WebSocket and streaming the user’s voice input to EVI. You can also send EVI text, and it will speak that text aloud.
EVI will respond with:
- The text of EVI’s reply
- EVI’s expressive audio response
- A transcript of the user’s message along with their vocal expression measures
- Messages if the user interrupts EVI
- A message to let you know if EVI has finished responding
- Error messages if issues arise
Quickstart
Accelerate your project setup with our comprehensive quickstart guides, designed to integrate EVI into your Next.js, TypeScript, or Python applications. Each guide walks you through EVI API integration while demonstrating how to capture user audio and play back EVI’s response audio, helping you get up and running quickly.
Build web applications using our React client SDK in Next.js.
Develop server-side integrations or frontend applications with our TypeScript SDK.
Create integrations in Python using our Python SDK.
API limits
- WebSocket connections limit: limited to up to five (5) concurrent connections.
- WebSocket duration limit: connections are subject to a default timeout after thirty (30) minutes, or after ten (10) minutes of user inactivity. Duration limits may be adjusted by specifying the max_duration and inactivity fields in your EVI configuration.
- WebSocket message payload size limit: messages cannot exceed 16MB in size.
- Request rate limit: HTTP requests (e.g. configs endpoints) are limited to fifty (50) requests per second.
To request an increase in your concurrent connection limit, please submit the “Application to Increase EVI Concurrent Connections” found in the EVI section of the Profile Tab.