UStackUStack
HeyGen icon

HeyGen

HeyGen Developers offers an API platform to generate, translate, and lipsync avatar videos with TTS models—built for scalable production workflows.

HeyGen

What is HeyGen?

HeyGen Developers is a developer platform for building production video workflows with APIs. It provides access to a set of video models, including a Video Agent workflow, video generation, video translation, and lipsync, alongside speech generation (TTS).

The core purpose is to let developers generate, transform, and scale avatar and video outputs through API calls (and related tooling like a CLI), with structured responses suitable for integrating into applications and agentic pipelines.

Key Features

  • Video Agent API endpoints: Generate avatar videos from a single prompt, producing finished video outputs without requiring separate avatar selection or scripting in the client workflow.
  • Avatar IV models (Digital Twin and Photo Avatar): Create a lifelike avatar from real video footage (Digital Twin) or animate a talking-head from a single still image (Photo Avatar), then generate speaking videos from a provided script and voice.
  • Video translation in 175+ languages: Translate video into 175+ languages with context-aware, natural lip-sync and gender detection, with output “in your voice.”
  • Translation modes: Support both “Speed” (faster dubbing) and “Precision” (lip-synced dubbing) translation variants within the platform’s translation capabilities.
  • Lipsync with audio replacement: Dub or replace a video’s audio using a provided audio file, with lips re-synced to match the new audio.
  • Voices / Starfish TTS: Generate speech audio from text using HeyGen’s TTS engine.
  • Production-ready developer tooling: The platform highlights its v3 API and an agent-first CLI that wraps v3 capabilities, returning structured JSON and supporting terminal-based workflows.
  • API reference + “Try It” consoles and guides: Documentation includes an authentication/video-creation walkthrough, an endpoint reference (request formats and response schemas), and a “Changelog” for API updates.
  • Security and compliance positioning: The site states SOC 2 Type II and GDPR compliance via independent audit/certification.

How to Use HeyGen

  1. Access the developer documentation for authentication and API usage via the v3 endpoints.
  2. Start with one of the model workflows (e.g., Video Agent, Video Generation, Video Translate, or Lipsync) and call the corresponding API endpoint.
  3. Use your API key in the request header (the site example shows sending x-api-key with a JSON payload).
  4. Supply required inputs for the chosen model (for example, a prompt along with avatar and voice identifiers for Video Agent / avatar-driven generation).
  5. Review structured JSON responses, then use the returned results in your application, CI pipeline, or agent workflow.

Use Cases

  • Create avatar-driven marketing or outreach videos: Send a single prompt to generate polished video output using an avatar workflow without manually selecting an avatar or editing a full script client-side.
  • Turn a person’s photo into social content: Use the Photo Avatar flow to animate a talking-head video from one still image and produce speech-aligned output using your selected voice.
  • Clone a digital presence from real footage: Use the Digital Twin (trained from real video footage) to generate new speaking videos from scripts in supported voices without requiring a camera or studio at generation time.
  • Localize product or training videos: Translate existing video into 175+ languages with lip-synced dubbing, including variants aimed at faster output or higher lip-sync precision.
  • Re-dub or adjust narration for existing footage: Provide an audio file to the Lipsync workflow to replace the video’s audio and automatically re-sync the speaker’s lip movements.

FAQ

How do I authenticate API requests?

The developer docs and examples indicate requests include an API key in the x-api-key header.

What’s the difference between “Speed” and “Precision” for translation and lipsync?

The site describes “Speed” as faster dubbing and “Precision” as lip-synced dubbing; both are available for translation and lipsync workflows.

Which languages are supported for video translation?

HeyGen’s video translation is described as supporting 175+ languages.

Can I generate speech from text without video translation?

Yes. The site lists a Voices / Starfish TTS capability that generates speech audio from text.

Is there a way to use HeyGen from the terminal?

The site describes an agent-first HeyGen CLI that wraps the v3 API so developers and agents can create, poll, and download avatar videos from the command line with structured JSON responses.

Alternatives

  • General-purpose video editing and dubbing workflows: Use tools that focus on manual voiceover, re-timing, and lip-matching as separate steps; compared to HeyGen, these typically require more production effort and tighter manual control.
  • Other developer APIs for dubbing/voice and avatar rendering: Look for platform providers that offer video dubbing or speech-driven avatar generation via APIs; differences are usually in language coverage, lipsync quality controls (speed vs precision), and the availability of avatar-training options (image vs video footage).
  • Offline/locally hosted AI video generation stacks: Some teams may prefer self-hosted pipelines for privacy or operational reasons; compared to HeyGen’s hosted v3 API and CLI, setup and scaling responsibilities move to the user.
  • Agent orchestration platforms with media connectors: If your goal is “agentic video generation,” consider agent platforms that integrate with third-party media generation services; compared to HeyGen’s v3-first approach, integration is often mediated through connectors rather than dedicated video endpoints.