Skip to main content

What Can You Build?

bitHuman turns audio into a real-time talking avatar. Anywhere you need a visual character that speaks — that’s where bitHuman fits.

Customer Support Avatar

Replace hold music with a face. An avatar greets visitors, answers FAQs, and escalates to a human when needed. Architecture: Website embed (iframe) → bitHuman cloud → OpenAI for conversation
avatar = bithuman.AvatarSession(
    avatar_id="YOUR_SUPPORT_AGENT",
    api_secret=os.getenv("BITHUMAN_API_SECRET"),
)
Best deployment: Cloud Plugin for fastest setup, or Website Embed for dropping into existing pages.

AI Tutor / Virtual Teacher

A patient avatar that explains concepts, answers questions, and adapts to the student’s pace. Lip-sync creates presence that text chat can’t match. Architecture: LiveKit room → AI agent (GPT-4 + domain knowledge) → bitHuman avatar Best deployment: Cloud Plugin with custom system prompt via Prompt Guide.

Digital Receptionist / Kiosk

A lobby screen or tablet that greets visitors, provides directions, and handles check-in. Runs on a Raspberry Pi or any Linux machine. Architecture: Kiosk browser → LiveKit → AI agent → bitHuman (self-hosted CPU) Best deployment: Self-Hosted CPU for offline capability, or Raspberry Pi for dedicated hardware.

AI Companion / Virtual Friend

A character with a persistent personality that remembers past conversations. Use context injection to maintain relationship state. Architecture: Mobile app (Flutter/React) → LiveKit → AI agent with memory → bitHuman Best deployment: Cloud Plugin + Flutter integration.

Game NPC / Interactive Character

Non-player characters that respond dynamically to player actions. Use dynamics for gestures (wave, nod, laugh) triggered by game events. Architecture: Game client → WebSocket → AI agent → bitHuman (GPU for custom faces) Best deployment: Self-Hosted GPU for dynamic face generation from character art.

Accessibility Tool

Give a face and voice to text-to-speech output. Visual lip-sync helps hearing-impaired users follow along. Audio output helps visually impaired users interact with content. Architecture: Screen reader / TTS → bitHuman SDK → overlay window Best deployment: Python SDK directly (no LiveKit needed).

Content Creation / Video Generation

Generate talking-head videos from scripts without recording. Batch-process audio files through avatars to create training videos, announcements, or social content. Architecture: Script → TTS → bitHuman SDK → video frames → ffmpeg → MP4
runtime = await AsyncBithuman.create(model_path="presenter.imx", api_secret="...")
await runtime.start()
await runtime.push_audio(tts_audio, sample_rate=16000)
await runtime.flush()

# Collect frames and encode to video
frames = []
async for frame in runtime.run():
    if frame.has_image:
        frames.append(frame.bgr_image)
Best deployment: Python SDK directly with Self-Hosted GPU for custom faces.

Choosing the Right Deployment

Use CaseRecommended DeploymentWhy
Customer supportCloud PluginFast setup, scales automatically
AI tutorCloud PluginLow latency, no infrastructure
Receptionist kioskSelf-Hosted CPUOffline capable, privacy
AI companionCloud Plugin + FlutterMobile-friendly, cross-platform
Game NPCSelf-Hosted GPUDynamic faces, low latency
AccessibilityPython SDKLightweight, no WebRTC overhead
Content creationPython SDK + GPUBatch processing, custom faces

Next Steps