EVI Python Quickstart
A quickstart guide for integrating the Empathic Voice Interface (EVI) with Python.
In this guide, you’ll learn how to use Hume’s Python SDK to integrate with EVI, for applications like command-line or desktop applications.
Use the Python SDK for client-side applications like CLIs and desktop apps that run directly on the user’s machine.
For hosted environments (e.g., Streamlit, Gradio, or backend servers), Python won’t be able to access the user’s audio devices. In these cases, use our TypeScript quickstart to build a client-side component that connects to EVI from the browser.
- Environment setup: Download package and system dependencies to run EVI.
- Import statements and helpers: Import needed symbols and define helper functions.
- Authentication: Use your API credentials to authenticate your EVI application.
- Connection: Set up a secure WebSocket connection to interact with EVI.
- Handling incoming messages: Process incoming chat events.
- Audio input: Capture audio data from an input device and send to EVI.
See the complete implementation of this guide on GitHub
Explore or contribute to Hume’s Python SDK on GitHub
Hume’s Python SDK supports EVI using Python versions 3.9 through 3.13 on macOS and Linux platforms. The
evi-python-quickstart example project used in this guide requires Python 3.11+
(requires-python = ">=3.11"). The full SDK specification can be found in the
Python SDK’s readme.
Environment setup
This guide uses the Hume SDK Python package hume with the [microphone] package extra. You can install this with uv (recommended), poetry, or pip. It also uses the python-dotenv package for loading environment variables from an .env file.
uv
poetry
pip
System dependencies
The Hume Python SDK uses the sounddevice library for audio recording and playback, which relies on the PortAudio C Library to be installed on your system. On macOS and Windows, PortAudio is typically included with the sounddevice package, so no additional installation is required. However, on Linux, you will need to manually install PortAudio correctly for your distribution.
Import statements and helpers
First, we import needed symbols from the Python standard library and the Hume SDK, and define some helpers that are useful for printing readable output to the terminal.
Authentication
Log into your Hume AI Account and obtain an API key. Store it as HUME_API_KEY inside your project’s .env file.
Read HUME_API_KEY and use it to instantiate the AsyncHumeClient class. This is the main entry point provided by the Hume Python SDK.
You can specify EVI’s voice and behavior for a chat by Creating a Configuration through the API or the Hume platform web interface. Set HUME_CONFIG_ID in .env or as an environment variable and read it.
Connection
To connect to an EVI chat, use client.empathic_voice.chat.connect(...) with connection arguments passed directly, such
as config_id=HUME_CONFIG_ID. EVI chats are event-based, so iterate over incoming events from the chat socket and
handle each message in your loop (or delegate to a helper function such as on_message).
Handling incoming messages
After you successfully connect to an EVI chat, iterate over messages from the socket and pass each one to your
on_message handler. These are described by the Hume SDK’s SubscribeEvent type.
In this quickstart, the message handler logs chat_metadata, logs final user_message content and top emotions, and
raises an exception if an error event is received. Assistant output and other event types are ignored in this
handler.
Audio input
The Hume SDK provides MicrophoneInterface.start(...), which handles microphone capture and streaming recorded audio
through the WebSocket to EVI. Stream.new() is initialized and passed as byte_stream to
MicrophoneInterface.start(...). This quickstart focuses on mic input and message handling; it does not include
manual audio_output playback queueing.
Pass the chat socket provided by connect(...) in order to use MicrophoneInterface.start:
Specify a microphone device
MicrophoneInterface.start will attempt to use the system’s default audio input device. To specify a specific audio input device, you can pass it via the optional device parameter in MicrophoneInterface.start.
To view a list of available audio devices, run the following command:
If the MacBook Pro Microphone is the desired device, specify device 4 in the Microphone context. For example:
For troubleshooting faulty device detection - particularly with systems using ALSA, the Advanced Linux Sound
Architecture, the device may also be directly specified using the sounddevice library:
Interruption
The allow_user_interrupt parameter on MicrophoneInterface.start(...) controls whether the user can send a
message while the assistant is speaking:
allow_user_interrupt=True: Allows the user to send microphone input even when the assistant is speaking. This enables more fluid, overlapping conversation.allow_user_interrupt=False: Prevents the user from sending microphone input while the assistant is speaking, ensuring that the user does not interrupt the assistant. This is useful in scenarios where clear, uninterrupted communication is important.
Put it all together
Finally, add the following code at the end of your script to run the main function:
View the complete quickstart.py code on GitHub
Next steps
Congratulations! You’ve successfully implemented a real-time conversational application using Hume’s Empathic Voice Interface (EVI).
Next, consider exploring these areas to enhance your EVI application:
See detailed instructions on how you can customize EVI for your application needs.
Learn how you can access and manage conversation transcripts and expression measures.
For further details and practical examples, explore the API Reference and our Hume API Examples on GitHub.

