Review

Hume AI: voice with a theory of emotion

Hume AI is a strong choice for developers who need expressive speech, live voice interaction, and emotion-aware analysis, but its privacy defaults and API-first pricing make it a poor fit for casual creators.

Last updated April 2026 · Pricing and features verified against official documentation

Voice AI usually gets judged on a narrow question: does it sound human enough? Hume AI is trying to answer a harder one: can a voice system change tone, pace, and delivery in ways that actually improve the interaction? That is why the company’s current product surface matters. Hume is no longer just a text-to-speech vendor. It now splits into Octave for expressive speech, EVI for speech-to-speech conversation, and expression measurement for analyzing voice, face, language, images, and text.

That breadth is useful if voice is part of the product, not just a garnish on top of it. Hume’s official docs now show Octave 2 and EVI 4-mini live, and the platform lets developers move from generated narration to live conversation without changing vendors. VentureBeat’s hands-on coverage of EVI 3 described building a custom voice in seconds and found the result more natural than the stock assistants most people know, which is exactly the kind of practical edge Hume is selling.

The case for Hume is strongest when a team needs expressive control, not just speech output. Product builders working on customer-facing agents, brand narration, training tools, or emotionally sensitive interfaces can get real leverage from the combination of voice design, interruptible live speech, and expression analysis in one stack. The public pricing is also unusually legible for this category, which makes it easier to test than many voice products that hide everything behind sales calls.

The case against Hume is just as clear. The product is developer-shaped, the pricing is metered in more than one way, and the privacy defaults for EVI are not conservative enough to ignore. If you want a simpler voice tool with a broader creator focus, ElevenLabs is the easier starting point. Hume is compelling, but it asks buyers to understand exactly what they are buying before they commit.

What the product actually is now

Hume AI should be read as a voice platform with three distinct jobs rather than one monolithic app. The company’s current product docs present Hume as a set of APIs and models for expressive text-to-speech, real-time speech-to-speech, and expression measurement. The public site also frames the company as a research lab and technology company, which matches the way the product has evolved: more infrastructure, more model choice, and more emphasis on measurement and control.

That split matters because each part solves a different problem. Octave is for generating speech with style. EVI is for live interaction. Expression measurement is for analyzing human cues and media. If you are choosing Hume, you are not choosing one feature. You are choosing which of those workflows you actually need.

Strengths

Octave gives speech real direction. Hume’s TTS product is not just a better-read-aloud engine. Octave supports voice design, voice cloning, acting instructions, multilingual output, and streaming playback, which gives teams more control over delivery than the average voice API. That matters when the voice is part of the product identity, not just a utility layer.

Live voice feels like a real interface, not a demo. EVI exists to handle back-and-forth conversation, not just convert text into audio. The current docs highlight interruptibility, back-channeling, and external LLM compatibility, and the newer EVI 4-mini rollout points to lower latency and broader language support. For products that need live turn-taking, that is the right kind of feature set.

The platform covers generation and analysis in one place. Most voice vendors stop at synthesis or transcription. Hume goes further by adding expression measurement alongside speech generation, which is useful for teams that want to understand how people sound, not just what they said. That makes the platform more interesting for coaching, support, and research workflows than a one-dimensional TTS tool.

The product is easier to test than many API companies. Hume’s pricing page is public, the entry tiers are inexpensive, and the platform clearly separates TTS, EVI, and expression measurement. That does not make the product simple, but it does make the cost model visible. In a category where buyers are often forced to guess, that transparency is a real advantage.

Weaknesses

EVI’s default privacy posture is too loose for sensitive work. Hume’s privacy docs say EVI data retention is enabled by default and that anonymized interaction data is used to improve models unless users explicitly opt out. That is a workable default for experimentation, but it is not the default I would want for interviews, customer calls, or any workflow with sensitive content.

The pricing structure is more complicated than it first appears. Hume bills TTS by characters and speech-to-speech by minutes, then layers expression measurement on a separate usage table. That is normal for an API platform, but it means buyers have to model at least three different cost curves before they can understand what the product will actually cost at scale.

The emotional-intelligence pitch can overstate the buying need. If a team just needs a stable voice API or speech infrastructure, Hume’s differentiator may be more conceptually interesting than operationally necessary. Deepgram and AssemblyAI are more direct fits when the job is speech infrastructure first and voice expression second.

Pricing

Hume’s current pricing is public and reasonably granular. The Free plan is $0 per month. Starter is $3 per month. Creator is listed at $7 for the first month and $14 after that. Pro is $70 per month, Scale is $200, Business is $500, and Enterprise is custom.

The important detail is that the plans are not just about access. They are about volume and concurrency. Free includes 10,000 TTS characters and 5 EVI minutes. Starter increases that to 30,000 characters and 40 EVI minutes. Creator includes 140,000 characters and 200 EVI minutes. Pro, Scale, and Business raise both character caps and EVI minutes while adding more concurrent connections and team seats.

The overage model reinforces the same point. Additional TTS characters are priced from $0.15 per 1,000 on Creator down to $0.05 per 1,000 on Business, while extra EVI usage drops from $0.06 per minute on Pro to $0.04 per minute on Business. If you are planning a production workflow, the question is not whether the entry tier is affordable. It is whether your usage pattern is mostly narration, mostly live conversation, or a mix of both.

Expression measurement is priced separately and should be treated that way in budgeting. Hume’s public rates list video with audio at $0.0828 per minute, audio only at $0.0639 per minute, video only at $0.045 per minute, images at $0.00204 each, and text at $0.00024 per word. That split makes sense for the product, but it means buyers need to estimate their workload before they can estimate their bill.

Privacy

Hume’s privacy story is mixed in a way buyers need to notice. For EVI, the docs say users can enable zero data retention and opt out of model training, but those protections are off by default. The same docs say turning off retention also removes access to chat history and the ability to resume chats. That is a meaningful tradeoff, not a checkbox.

Expression measurement is the cleaner side of the house. Hume says files processed by that API are not retained, and it says submitted data is not used to train or improve Hume models. Output data is retained until the user deletes it, which is practical for repeat lookups, but the file-handling posture is much better than the EVI default.

Hume also says it supports SOC 2 Type II, GDPR, and HIPAA on the pricing page, and its privacy docs say customers can request a BAA or DPA. That is enough to make the platform viable for serious work, but not enough to justify a casual upload of sensitive audio. The right reading is simple: Hume has real controls, but users have to activate the ones that matter.

Who it’s best for

The developer building a voice product with personality. If you are shipping an assistant, tutor, support layer, or branded voice experience, Hume gives you expressive speech plus live interaction in one place. That is the buyer most likely to benefit from Octave and EVI together.

The product team that needs voice and analysis in one stack. Hume is a good fit when the same application needs to generate speech and also measure expression in media or text. The platform is more useful here than a plain TTS vendor because it can support both the output layer and the interpretation layer.

The team willing to budget like a platform buyer. Hume makes sense for organizations that can model usage, read privacy settings carefully, and live with API pricing. If your team is comfortable thinking in characters, minutes, and concurrency, the product is easy enough to justify.

Who should look elsewhere

Bottom line

Hume AI is one of the more interesting voice platforms because it treats emotion as part of the product, not just part of the marketing. That makes it unusually attractive for teams building voice experiences where tone, interruption, and delivery matter as much as the words themselves.

The tradeoff is that Hume’s strongest ideas live inside a product that still expects the buyer to do real platform thinking. You have to choose the right API, estimate the right usage pattern, and read the privacy settings carefully. If voice is central to the product, that is a fair price for the control Hume offers. If voice is just a support act, the platform is probably more opinionated than you need.