Show HN: Open-Source Framework for Real-Time AI Video Avatars
You can use it to create sales assistants, customer success agents, mock interviewers, language coaches, or even historical characters. It’s modular (choose your STT, LLM, and TTS provider), production-ready, and optimized for ultra-low latency video generation.
Features:
- Real-time speech-to-video avatars (<300ms)
- Native turn detection, VAD, and noise suppression
- Modular pipelines for STT, LLM, TTS, and avatars with real-time model switching
- Built-in RAG + memory for grounding and hallucination resistance
- SDKs for web, mobile, Unity, IoT, and telephony — no glue code needed
- Agent Cloud for infinite scaling with one-click deployments — or self-host with full control
GitHub Repo: https://github.com/videosdk-community/ai-avatar-demo
Full Blog: https://www.videosdk.live/blog/ai-avatar-agent
Would love feedback from anyone working with video, avatars, or real-time conversational AI!
No comments yet